{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.4835579275902206, "eval_steps": 3000, "global_step": 795000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 9.330555937849166e-05, "grad_norm": 93.67707824707031, "learning_rate": 8.166666666666666e-07, "loss": 16.1123, "step": 50 }, { "epoch": 0.00018661111875698333, "grad_norm": 7.377334117889404, "learning_rate": 1.65e-06, "loss": 12.8959, "step": 100 }, { "epoch": 0.00027991667813547503, "grad_norm": 8.839871406555176, "learning_rate": 2.4833333333333334e-06, "loss": 11.8846, "step": 150 }, { "epoch": 0.00037322223751396665, "grad_norm": 16.873918533325195, "learning_rate": 3.3166666666666665e-06, "loss": 11.4614, "step": 200 }, { "epoch": 0.0004665277968924583, "grad_norm": 60.141456604003906, "learning_rate": 4.15e-06, "loss": 10.2624, "step": 250 }, { "epoch": 0.0005598333562709501, "grad_norm": 26.98830223083496, "learning_rate": 4.983333333333334e-06, "loss": 9.1782, "step": 300 }, { "epoch": 0.0006531389156494417, "grad_norm": 5.853992462158203, "learning_rate": 5.816666666666667e-06, "loss": 8.381, "step": 350 }, { "epoch": 0.0007464444750279333, "grad_norm": 38.342830657958984, "learning_rate": 6.650000000000001e-06, "loss": 7.5826, "step": 400 }, { "epoch": 0.000839750034406425, "grad_norm": 19.33097267150879, "learning_rate": 7.483333333333334e-06, "loss": 7.4673, "step": 450 }, { "epoch": 0.0009330555937849167, "grad_norm": 9.538616180419922, "learning_rate": 8.316666666666668e-06, "loss": 7.0077, "step": 500 }, { "epoch": 0.0010263611531634083, "grad_norm": 10.989666938781738, "learning_rate": 9.15e-06, "loss": 7.2275, "step": 550 }, { "epoch": 0.0011196667125419001, "grad_norm": 9.881364822387695, "learning_rate": 9.983333333333333e-06, "loss": 6.8267, "step": 600 }, { "epoch": 0.0012129722719203917, "grad_norm": 12.614446640014648, "learning_rate": 1.0816666666666666e-05, "loss": 7.133, "step": 650 }, { "epoch": 0.0013062778312988835, "grad_norm": 6.586646556854248, "learning_rate": 1.1650000000000002e-05, "loss": 7.0883, "step": 700 }, { "epoch": 0.001399583390677375, "grad_norm": 10.388212203979492, "learning_rate": 1.2483333333333335e-05, "loss": 6.8631, "step": 750 }, { "epoch": 0.0014928889500558666, "grad_norm": 5.741134166717529, "learning_rate": 1.3316666666666666e-05, "loss": 6.6987, "step": 800 }, { "epoch": 0.0015861945094343584, "grad_norm": 6.758159160614014, "learning_rate": 1.415e-05, "loss": 6.8654, "step": 850 }, { "epoch": 0.00167950006881285, "grad_norm": 11.188956260681152, "learning_rate": 1.4983333333333336e-05, "loss": 6.6844, "step": 900 }, { "epoch": 0.0017728056281913417, "grad_norm": 9.388659477233887, "learning_rate": 1.5816666666666667e-05, "loss": 6.6985, "step": 950 }, { "epoch": 0.0018661111875698333, "grad_norm": 17.585216522216797, "learning_rate": 1.665e-05, "loss": 6.5966, "step": 1000 }, { "epoch": 0.001959416746948325, "grad_norm": 5.970058917999268, "learning_rate": 1.7483333333333336e-05, "loss": 6.7767, "step": 1050 }, { "epoch": 0.0020527223063268167, "grad_norm": 13.584476470947266, "learning_rate": 1.8316666666666667e-05, "loss": 6.7776, "step": 1100 }, { "epoch": 0.0021460278657053082, "grad_norm": 8.928284645080566, "learning_rate": 1.915e-05, "loss": 6.5926, "step": 1150 }, { "epoch": 0.0022393334250838002, "grad_norm": 9.114434242248535, "learning_rate": 1.9983333333333336e-05, "loss": 6.687, "step": 1200 }, { "epoch": 0.002332638984462292, "grad_norm": 11.371248245239258, "learning_rate": 2.0816666666666667e-05, "loss": 6.4247, "step": 1250 }, { "epoch": 0.0024259445438407834, "grad_norm": 9.375133514404297, "learning_rate": 2.165e-05, "loss": 6.3891, "step": 1300 }, { "epoch": 0.002519250103219275, "grad_norm": 7.963316440582275, "learning_rate": 2.2483333333333335e-05, "loss": 6.7139, "step": 1350 }, { "epoch": 0.002612555662597767, "grad_norm": 9.820133209228516, "learning_rate": 2.3316666666666666e-05, "loss": 6.5634, "step": 1400 }, { "epoch": 0.0027058612219762585, "grad_norm": 10.639363288879395, "learning_rate": 2.415e-05, "loss": 6.6883, "step": 1450 }, { "epoch": 0.00279916678135475, "grad_norm": 6.419040679931641, "learning_rate": 2.4983333333333335e-05, "loss": 6.6271, "step": 1500 }, { "epoch": 0.0028924723407332416, "grad_norm": 8.272869110107422, "learning_rate": 2.5816666666666666e-05, "loss": 6.4248, "step": 1550 }, { "epoch": 0.002985777900111733, "grad_norm": 7.704101085662842, "learning_rate": 2.6650000000000004e-05, "loss": 6.3276, "step": 1600 }, { "epoch": 0.003079083459490225, "grad_norm": 9.784399032592773, "learning_rate": 2.748333333333333e-05, "loss": 6.2883, "step": 1650 }, { "epoch": 0.0031723890188687168, "grad_norm": 11.014293670654297, "learning_rate": 2.831666666666667e-05, "loss": 6.4566, "step": 1700 }, { "epoch": 0.0032656945782472083, "grad_norm": 10.516816139221191, "learning_rate": 2.915e-05, "loss": 6.2076, "step": 1750 }, { "epoch": 0.0033590001376257, "grad_norm": 11.228531837463379, "learning_rate": 2.9983333333333335e-05, "loss": 6.3539, "step": 1800 }, { "epoch": 0.003452305697004192, "grad_norm": 9.591064453125, "learning_rate": 3.0816666666666666e-05, "loss": 6.418, "step": 1850 }, { "epoch": 0.0035456112563826835, "grad_norm": 10.729142189025879, "learning_rate": 3.1650000000000004e-05, "loss": 6.0638, "step": 1900 }, { "epoch": 0.003638916815761175, "grad_norm": 8.057198524475098, "learning_rate": 3.2483333333333335e-05, "loss": 6.2542, "step": 1950 }, { "epoch": 0.0037322223751396666, "grad_norm": 8.599757194519043, "learning_rate": 3.3316666666666666e-05, "loss": 6.2936, "step": 2000 }, { "epoch": 0.0038255279345181586, "grad_norm": 10.339815139770508, "learning_rate": 3.415e-05, "loss": 6.1887, "step": 2050 }, { "epoch": 0.00391883349389665, "grad_norm": 8.34434700012207, "learning_rate": 3.4983333333333334e-05, "loss": 6.3519, "step": 2100 }, { "epoch": 0.004012139053275142, "grad_norm": 8.931737899780273, "learning_rate": 3.581666666666667e-05, "loss": 5.9798, "step": 2150 }, { "epoch": 0.004105444612653633, "grad_norm": 7.339552402496338, "learning_rate": 3.665e-05, "loss": 5.8089, "step": 2200 }, { "epoch": 0.004198750172032125, "grad_norm": 10.72778606414795, "learning_rate": 3.7483333333333334e-05, "loss": 6.154, "step": 2250 }, { "epoch": 0.0042920557314106164, "grad_norm": 7.218290328979492, "learning_rate": 3.8316666666666665e-05, "loss": 6.2445, "step": 2300 }, { "epoch": 0.004385361290789108, "grad_norm": 7.798749923706055, "learning_rate": 3.915e-05, "loss": 5.9981, "step": 2350 }, { "epoch": 0.0044786668501676005, "grad_norm": 7.12747049331665, "learning_rate": 3.9983333333333334e-05, "loss": 6.0511, "step": 2400 }, { "epoch": 0.004571972409546092, "grad_norm": 6.343996524810791, "learning_rate": 4.081666666666667e-05, "loss": 6.0378, "step": 2450 }, { "epoch": 0.004665277968924584, "grad_norm": 6.697947025299072, "learning_rate": 4.165e-05, "loss": 5.8112, "step": 2500 }, { "epoch": 0.004758583528303075, "grad_norm": 8.327080726623535, "learning_rate": 4.2483333333333334e-05, "loss": 6.2906, "step": 2550 }, { "epoch": 0.004851889087681567, "grad_norm": 5.573576927185059, "learning_rate": 4.3316666666666665e-05, "loss": 5.8131, "step": 2600 }, { "epoch": 0.004945194647060058, "grad_norm": 6.693691253662109, "learning_rate": 4.415e-05, "loss": 5.7015, "step": 2650 }, { "epoch": 0.00503850020643855, "grad_norm": 6.914066314697266, "learning_rate": 4.4983333333333334e-05, "loss": 5.8004, "step": 2700 }, { "epoch": 0.005131805765817041, "grad_norm": 8.469379425048828, "learning_rate": 4.581666666666667e-05, "loss": 5.9158, "step": 2750 }, { "epoch": 0.005225111325195534, "grad_norm": 7.3673295974731445, "learning_rate": 4.665e-05, "loss": 5.969, "step": 2800 }, { "epoch": 0.005318416884574025, "grad_norm": 7.18582010269165, "learning_rate": 4.748333333333333e-05, "loss": 6.0261, "step": 2850 }, { "epoch": 0.005411722443952517, "grad_norm": 5.982076644897461, "learning_rate": 4.831666666666667e-05, "loss": 6.023, "step": 2900 }, { "epoch": 0.0055050280033310086, "grad_norm": 5.787336826324463, "learning_rate": 4.915e-05, "loss": 5.8281, "step": 2950 }, { "epoch": 0.0055983335627095, "grad_norm": 4.3454718589782715, "learning_rate": 4.998333333333334e-05, "loss": 5.9889, "step": 3000 }, { "epoch": 0.0055983335627095, "eval_loss": 6.374537944793701, "eval_runtime": 232.2431, "eval_samples_per_second": 11.23, "eval_steps_per_second": 11.23, "eval_tts_loss": 6.28711945315583, "step": 3000 }, { "epoch": 0.005691639122087992, "grad_norm": 8.272449493408203, "learning_rate": 5.081666666666667e-05, "loss": 6.141, "step": 3050 }, { "epoch": 0.005784944681466483, "grad_norm": 5.653373718261719, "learning_rate": 5.1649999999999995e-05, "loss": 6.1385, "step": 3100 }, { "epoch": 0.005878250240844975, "grad_norm": 4.102393627166748, "learning_rate": 5.248333333333334e-05, "loss": 5.8571, "step": 3150 }, { "epoch": 0.005971555800223466, "grad_norm": 5.258531093597412, "learning_rate": 5.331666666666667e-05, "loss": 5.7932, "step": 3200 }, { "epoch": 0.006064861359601959, "grad_norm": 4.183366775512695, "learning_rate": 5.415e-05, "loss": 5.8642, "step": 3250 }, { "epoch": 0.00615816691898045, "grad_norm": 5.612656593322754, "learning_rate": 5.498333333333333e-05, "loss": 5.8893, "step": 3300 }, { "epoch": 0.006251472478358942, "grad_norm": 5.23427677154541, "learning_rate": 5.581666666666667e-05, "loss": 5.7405, "step": 3350 }, { "epoch": 0.0063447780377374335, "grad_norm": 5.158674240112305, "learning_rate": 5.665e-05, "loss": 5.8673, "step": 3400 }, { "epoch": 0.006438083597115925, "grad_norm": 5.116922378540039, "learning_rate": 5.748333333333333e-05, "loss": 5.7444, "step": 3450 }, { "epoch": 0.006531389156494417, "grad_norm": 5.790384292602539, "learning_rate": 5.831666666666668e-05, "loss": 5.9559, "step": 3500 }, { "epoch": 0.006624694715872908, "grad_norm": 4.801383972167969, "learning_rate": 5.915000000000001e-05, "loss": 6.1995, "step": 3550 }, { "epoch": 0.0067180002752514, "grad_norm": 3.692594528198242, "learning_rate": 5.998333333333334e-05, "loss": 5.733, "step": 3600 }, { "epoch": 0.006811305834629892, "grad_norm": 4.112286567687988, "learning_rate": 6.081666666666666e-05, "loss": 6.0937, "step": 3650 }, { "epoch": 0.006904611394008384, "grad_norm": 4.726688385009766, "learning_rate": 6.165000000000001e-05, "loss": 5.8487, "step": 3700 }, { "epoch": 0.006997916953386875, "grad_norm": 3.719205856323242, "learning_rate": 6.248333333333334e-05, "loss": 5.7952, "step": 3750 }, { "epoch": 0.007091222512765367, "grad_norm": 4.780358791351318, "learning_rate": 6.331666666666667e-05, "loss": 5.9006, "step": 3800 }, { "epoch": 0.0071845280721438585, "grad_norm": 5.655118942260742, "learning_rate": 6.415e-05, "loss": 5.7174, "step": 3850 }, { "epoch": 0.00727783363152235, "grad_norm": 4.822659015655518, "learning_rate": 6.498333333333335e-05, "loss": 5.9248, "step": 3900 }, { "epoch": 0.007371139190900842, "grad_norm": 5.104306697845459, "learning_rate": 6.581666666666668e-05, "loss": 5.8789, "step": 3950 }, { "epoch": 0.007464444750279333, "grad_norm": 4.1900105476379395, "learning_rate": 6.665000000000001e-05, "loss": 5.8902, "step": 4000 }, { "epoch": 0.007557750309657825, "grad_norm": 4.686253547668457, "learning_rate": 6.748333333333334e-05, "loss": 5.6845, "step": 4050 }, { "epoch": 0.007651055869036317, "grad_norm": 4.2270188331604, "learning_rate": 6.831666666666667e-05, "loss": 5.9528, "step": 4100 }, { "epoch": 0.007744361428414809, "grad_norm": 4.093949794769287, "learning_rate": 6.915e-05, "loss": 5.8561, "step": 4150 }, { "epoch": 0.0078376669877933, "grad_norm": 4.820592403411865, "learning_rate": 6.998333333333333e-05, "loss": 5.6467, "step": 4200 }, { "epoch": 0.007930972547171792, "grad_norm": 4.777779579162598, "learning_rate": 7.081666666666668e-05, "loss": 5.8914, "step": 4250 }, { "epoch": 0.008024278106550283, "grad_norm": 2.8841679096221924, "learning_rate": 7.165000000000001e-05, "loss": 5.7689, "step": 4300 }, { "epoch": 0.008117583665928775, "grad_norm": 3.7897868156433105, "learning_rate": 7.248333333333334e-05, "loss": 5.6131, "step": 4350 }, { "epoch": 0.008210889225307267, "grad_norm": 3.63638973236084, "learning_rate": 7.331666666666667e-05, "loss": 5.6802, "step": 4400 }, { "epoch": 0.008304194784685758, "grad_norm": 4.507631778717041, "learning_rate": 7.415000000000001e-05, "loss": 5.9008, "step": 4450 }, { "epoch": 0.00839750034406425, "grad_norm": 4.462691783905029, "learning_rate": 7.498333333333334e-05, "loss": 5.7764, "step": 4500 }, { "epoch": 0.008490805903442741, "grad_norm": 5.587769508361816, "learning_rate": 7.581666666666668e-05, "loss": 5.6366, "step": 4550 }, { "epoch": 0.008584111462821233, "grad_norm": 3.457714319229126, "learning_rate": 7.664999999999999e-05, "loss": 5.7527, "step": 4600 }, { "epoch": 0.008677417022199724, "grad_norm": 4.030301094055176, "learning_rate": 7.748333333333334e-05, "loss": 5.8454, "step": 4650 }, { "epoch": 0.008770722581578216, "grad_norm": 3.3396108150482178, "learning_rate": 7.831666666666667e-05, "loss": 5.9134, "step": 4700 }, { "epoch": 0.00886402814095671, "grad_norm": 3.4558310508728027, "learning_rate": 7.915e-05, "loss": 5.7332, "step": 4750 }, { "epoch": 0.008957333700335201, "grad_norm": 4.230519771575928, "learning_rate": 7.998333333333333e-05, "loss": 5.86, "step": 4800 }, { "epoch": 0.009050639259713692, "grad_norm": 3.126082181930542, "learning_rate": 8.081666666666667e-05, "loss": 5.7412, "step": 4850 }, { "epoch": 0.009143944819092184, "grad_norm": 4.2145609855651855, "learning_rate": 8.165e-05, "loss": 5.9317, "step": 4900 }, { "epoch": 0.009237250378470676, "grad_norm": 3.0938162803649902, "learning_rate": 8.248333333333334e-05, "loss": 5.6891, "step": 4950 }, { "epoch": 0.009330555937849167, "grad_norm": 4.230195045471191, "learning_rate": 8.331666666666668e-05, "loss": 5.7216, "step": 5000 }, { "epoch": 0.009423861497227659, "grad_norm": 3.7208573818206787, "learning_rate": 8.415000000000001e-05, "loss": 5.7874, "step": 5050 }, { "epoch": 0.00951716705660615, "grad_norm": 3.4725289344787598, "learning_rate": 8.498333333333334e-05, "loss": 5.8627, "step": 5100 }, { "epoch": 0.009610472615984642, "grad_norm": 3.3882782459259033, "learning_rate": 8.581666666666666e-05, "loss": 5.767, "step": 5150 }, { "epoch": 0.009703778175363133, "grad_norm": 3.7585625648498535, "learning_rate": 8.665e-05, "loss": 5.6776, "step": 5200 }, { "epoch": 0.009797083734741625, "grad_norm": 4.944695949554443, "learning_rate": 8.748333333333334e-05, "loss": 5.667, "step": 5250 }, { "epoch": 0.009890389294120117, "grad_norm": 3.8249406814575195, "learning_rate": 8.831666666666667e-05, "loss": 5.7428, "step": 5300 }, { "epoch": 0.009983694853498608, "grad_norm": 4.215453147888184, "learning_rate": 8.915e-05, "loss": 5.89, "step": 5350 }, { "epoch": 0.0100770004128771, "grad_norm": 3.0438342094421387, "learning_rate": 8.998333333333334e-05, "loss": 5.639, "step": 5400 }, { "epoch": 0.010170305972255591, "grad_norm": 4.088777542114258, "learning_rate": 9.081666666666667e-05, "loss": 5.5986, "step": 5450 }, { "epoch": 0.010263611531634083, "grad_norm": 3.3326492309570312, "learning_rate": 9.165e-05, "loss": 5.5374, "step": 5500 }, { "epoch": 0.010356917091012574, "grad_norm": 2.9593100547790527, "learning_rate": 9.248333333333334e-05, "loss": 5.4346, "step": 5550 }, { "epoch": 0.010450222650391068, "grad_norm": 3.170647144317627, "learning_rate": 9.331666666666668e-05, "loss": 5.8167, "step": 5600 }, { "epoch": 0.01054352820976956, "grad_norm": 2.9037222862243652, "learning_rate": 9.415e-05, "loss": 5.5978, "step": 5650 }, { "epoch": 0.01063683376914805, "grad_norm": 3.341003894805908, "learning_rate": 9.498333333333333e-05, "loss": 5.6149, "step": 5700 }, { "epoch": 0.010730139328526542, "grad_norm": 2.9810729026794434, "learning_rate": 9.581666666666667e-05, "loss": 5.7028, "step": 5750 }, { "epoch": 0.010823444887905034, "grad_norm": 3.4222352504730225, "learning_rate": 9.665e-05, "loss": 5.5117, "step": 5800 }, { "epoch": 0.010916750447283526, "grad_norm": 3.1818320751190186, "learning_rate": 9.748333333333334e-05, "loss": 5.7761, "step": 5850 }, { "epoch": 0.011010056006662017, "grad_norm": 3.1308321952819824, "learning_rate": 9.831666666666667e-05, "loss": 5.6791, "step": 5900 }, { "epoch": 0.011103361566040509, "grad_norm": 3.701707363128662, "learning_rate": 9.915000000000001e-05, "loss": 5.5769, "step": 5950 }, { "epoch": 0.011196667125419, "grad_norm": 2.931779384613037, "learning_rate": 9.998333333333334e-05, "loss": 5.4739, "step": 6000 }, { "epoch": 0.011196667125419, "eval_loss": 6.0997633934021, "eval_runtime": 229.8318, "eval_samples_per_second": 11.347, "eval_steps_per_second": 11.347, "eval_tts_loss": 6.378738236305035, "step": 6000 }, { "epoch": 0.011289972684797492, "grad_norm": 3.05470871925354, "learning_rate": 0.00010081666666666667, "loss": 5.5917, "step": 6050 }, { "epoch": 0.011383278244175983, "grad_norm": 3.2815566062927246, "learning_rate": 0.00010165, "loss": 5.4096, "step": 6100 }, { "epoch": 0.011476583803554475, "grad_norm": 2.6623780727386475, "learning_rate": 0.00010248333333333334, "loss": 5.4688, "step": 6150 }, { "epoch": 0.011569889362932967, "grad_norm": 3.355912923812866, "learning_rate": 0.00010331666666666667, "loss": 5.6187, "step": 6200 }, { "epoch": 0.011663194922311458, "grad_norm": 2.9381911754608154, "learning_rate": 0.00010415000000000001, "loss": 5.8301, "step": 6250 }, { "epoch": 0.01175650048168995, "grad_norm": 2.482517719268799, "learning_rate": 0.00010498333333333334, "loss": 5.5721, "step": 6300 }, { "epoch": 0.011849806041068441, "grad_norm": 3.363173246383667, "learning_rate": 0.00010581666666666667, "loss": 5.8946, "step": 6350 }, { "epoch": 0.011943111600446933, "grad_norm": 3.2402546405792236, "learning_rate": 0.00010665, "loss": 5.6961, "step": 6400 }, { "epoch": 0.012036417159825426, "grad_norm": 3.1428112983703613, "learning_rate": 0.00010748333333333333, "loss": 5.8315, "step": 6450 }, { "epoch": 0.012129722719203918, "grad_norm": 2.3944103717803955, "learning_rate": 0.00010831666666666667, "loss": 5.6921, "step": 6500 }, { "epoch": 0.01222302827858241, "grad_norm": 2.7807068824768066, "learning_rate": 0.00010915, "loss": 5.563, "step": 6550 }, { "epoch": 0.0123163338379609, "grad_norm": 2.692389965057373, "learning_rate": 0.00010998333333333335, "loss": 5.4729, "step": 6600 }, { "epoch": 0.012409639397339392, "grad_norm": 1.9794119596481323, "learning_rate": 0.00011081666666666669, "loss": 5.7415, "step": 6650 }, { "epoch": 0.012502944956717884, "grad_norm": 1.812415599822998, "learning_rate": 0.00011165000000000002, "loss": 5.4328, "step": 6700 }, { "epoch": 0.012596250516096376, "grad_norm": 3.0957512855529785, "learning_rate": 0.00011248333333333333, "loss": 5.3841, "step": 6750 }, { "epoch": 0.012689556075474867, "grad_norm": 3.101076126098633, "learning_rate": 0.00011331666666666667, "loss": 5.8037, "step": 6800 }, { "epoch": 0.012782861634853359, "grad_norm": 3.04792857170105, "learning_rate": 0.00011415, "loss": 5.7408, "step": 6850 }, { "epoch": 0.01287616719423185, "grad_norm": 3.4043712615966797, "learning_rate": 0.00011498333333333333, "loss": 5.6041, "step": 6900 }, { "epoch": 0.012969472753610342, "grad_norm": 1.666723608970642, "learning_rate": 0.00011581666666666666, "loss": 5.6662, "step": 6950 }, { "epoch": 0.013062778312988833, "grad_norm": 2.3120384216308594, "learning_rate": 0.00011665000000000002, "loss": 5.8835, "step": 7000 }, { "epoch": 0.013156083872367325, "grad_norm": 4.411382675170898, "learning_rate": 0.00011748333333333335, "loss": 5.7121, "step": 7050 }, { "epoch": 0.013249389431745816, "grad_norm": 2.192537546157837, "learning_rate": 0.00011831666666666668, "loss": 5.4821, "step": 7100 }, { "epoch": 0.013342694991124308, "grad_norm": 2.196122646331787, "learning_rate": 0.00011915000000000001, "loss": 5.7267, "step": 7150 }, { "epoch": 0.0134360005505028, "grad_norm": 2.2460808753967285, "learning_rate": 0.00011998333333333334, "loss": 5.5123, "step": 7200 }, { "epoch": 0.013529306109881291, "grad_norm": 2.052151679992676, "learning_rate": 0.00012081666666666667, "loss": 5.8528, "step": 7250 }, { "epoch": 0.013622611669259784, "grad_norm": 2.369872808456421, "learning_rate": 0.00012165, "loss": 5.5378, "step": 7300 }, { "epoch": 0.013715917228638276, "grad_norm": 2.383234977722168, "learning_rate": 0.00012248333333333335, "loss": 5.7572, "step": 7350 }, { "epoch": 0.013809222788016768, "grad_norm": 2.3634586334228516, "learning_rate": 0.0001233166666666667, "loss": 5.3572, "step": 7400 }, { "epoch": 0.01390252834739526, "grad_norm": 2.631533622741699, "learning_rate": 0.00012415, "loss": 5.4715, "step": 7450 }, { "epoch": 0.01399583390677375, "grad_norm": 1.962390661239624, "learning_rate": 0.00012498333333333335, "loss": 5.7204, "step": 7500 }, { "epoch": 0.014089139466152242, "grad_norm": 2.6363272666931152, "learning_rate": 0.00012581666666666667, "loss": 5.6266, "step": 7550 }, { "epoch": 0.014182445025530734, "grad_norm": 2.3159897327423096, "learning_rate": 0.00012665, "loss": 5.6025, "step": 7600 }, { "epoch": 0.014275750584909225, "grad_norm": 2.558704137802124, "learning_rate": 0.00012748333333333333, "loss": 5.622, "step": 7650 }, { "epoch": 0.014369056144287717, "grad_norm": 2.3983957767486572, "learning_rate": 0.00012831666666666665, "loss": 5.677, "step": 7700 }, { "epoch": 0.014462361703666209, "grad_norm": 2.4474925994873047, "learning_rate": 0.00012915000000000002, "loss": 5.735, "step": 7750 }, { "epoch": 0.0145556672630447, "grad_norm": 1.7076762914657593, "learning_rate": 0.00012998333333333334, "loss": 5.6703, "step": 7800 }, { "epoch": 0.014648972822423192, "grad_norm": 2.520663261413574, "learning_rate": 0.00013081666666666668, "loss": 5.6445, "step": 7850 }, { "epoch": 0.014742278381801683, "grad_norm": 1.7655445337295532, "learning_rate": 0.00013165, "loss": 5.5739, "step": 7900 }, { "epoch": 0.014835583941180175, "grad_norm": 2.880577564239502, "learning_rate": 0.00013248333333333335, "loss": 5.5834, "step": 7950 }, { "epoch": 0.014928889500558666, "grad_norm": 2.0654830932617188, "learning_rate": 0.00013331666666666666, "loss": 5.5844, "step": 8000 }, { "epoch": 0.015022195059937158, "grad_norm": 1.5936386585235596, "learning_rate": 0.00013415, "loss": 5.4989, "step": 8050 }, { "epoch": 0.01511550061931565, "grad_norm": 2.4710347652435303, "learning_rate": 0.00013498333333333335, "loss": 5.5193, "step": 8100 }, { "epoch": 0.015208806178694143, "grad_norm": 1.680249571800232, "learning_rate": 0.00013581666666666667, "loss": 5.4661, "step": 8150 }, { "epoch": 0.015302111738072634, "grad_norm": 1.9727201461791992, "learning_rate": 0.00013665000000000001, "loss": 5.7409, "step": 8200 }, { "epoch": 0.015395417297451126, "grad_norm": 2.1253304481506348, "learning_rate": 0.00013748333333333333, "loss": 5.5681, "step": 8250 }, { "epoch": 0.015488722856829618, "grad_norm": 2.3671233654022217, "learning_rate": 0.00013831666666666668, "loss": 5.5692, "step": 8300 }, { "epoch": 0.01558202841620811, "grad_norm": 2.448765516281128, "learning_rate": 0.00013915, "loss": 5.8129, "step": 8350 }, { "epoch": 0.0156753339755866, "grad_norm": 2.6392576694488525, "learning_rate": 0.00013998333333333334, "loss": 5.6357, "step": 8400 }, { "epoch": 0.01576863953496509, "grad_norm": 2.3633005619049072, "learning_rate": 0.00014081666666666666, "loss": 5.6419, "step": 8450 }, { "epoch": 0.015861945094343584, "grad_norm": 2.5643057823181152, "learning_rate": 0.00014165000000000003, "loss": 5.6119, "step": 8500 }, { "epoch": 0.015955250653722074, "grad_norm": 2.252556324005127, "learning_rate": 0.00014248333333333335, "loss": 5.5832, "step": 8550 }, { "epoch": 0.016048556213100567, "grad_norm": 2.5928521156311035, "learning_rate": 0.0001433166666666667, "loss": 5.6567, "step": 8600 }, { "epoch": 0.01614186177247906, "grad_norm": 2.565758228302002, "learning_rate": 0.00014415, "loss": 5.5772, "step": 8650 }, { "epoch": 0.01623516733185755, "grad_norm": 1.9352021217346191, "learning_rate": 0.00014498333333333332, "loss": 5.6935, "step": 8700 }, { "epoch": 0.016328472891236043, "grad_norm": 1.7188228368759155, "learning_rate": 0.00014581666666666667, "loss": 5.5983, "step": 8750 }, { "epoch": 0.016421778450614533, "grad_norm": 1.5790917873382568, "learning_rate": 0.00014665, "loss": 5.5048, "step": 8800 }, { "epoch": 0.016515084009993027, "grad_norm": 2.046977996826172, "learning_rate": 0.00014748333333333336, "loss": 5.4909, "step": 8850 }, { "epoch": 0.016608389569371516, "grad_norm": 2.3554604053497314, "learning_rate": 0.00014831666666666668, "loss": 5.5094, "step": 8900 }, { "epoch": 0.01670169512875001, "grad_norm": 2.019984245300293, "learning_rate": 0.00014915000000000002, "loss": 5.5235, "step": 8950 }, { "epoch": 0.0167950006881285, "grad_norm": 2.2459378242492676, "learning_rate": 0.00014998333333333334, "loss": 5.7603, "step": 9000 }, { "epoch": 0.0167950006881285, "eval_loss": 5.949121475219727, "eval_runtime": 232.0628, "eval_samples_per_second": 11.238, "eval_steps_per_second": 11.238, "eval_tts_loss": 6.626402078707777, "step": 9000 }, { "epoch": 0.016888306247506993, "grad_norm": 1.576752781867981, "learning_rate": 0.00015081666666666668, "loss": 5.425, "step": 9050 }, { "epoch": 0.016981611806885483, "grad_norm": 1.950769066810608, "learning_rate": 0.00015165, "loss": 5.5398, "step": 9100 }, { "epoch": 0.017074917366263976, "grad_norm": 1.870368242263794, "learning_rate": 0.00015248333333333334, "loss": 5.5754, "step": 9150 }, { "epoch": 0.017168222925642466, "grad_norm": 1.75543212890625, "learning_rate": 0.00015331666666666666, "loss": 5.5328, "step": 9200 }, { "epoch": 0.01726152848502096, "grad_norm": 1.626207947731018, "learning_rate": 0.00015415, "loss": 5.6149, "step": 9250 }, { "epoch": 0.01735483404439945, "grad_norm": 2.190242052078247, "learning_rate": 0.00015498333333333335, "loss": 5.4885, "step": 9300 }, { "epoch": 0.017448139603777942, "grad_norm": 1.4450881481170654, "learning_rate": 0.00015581666666666667, "loss": 5.4476, "step": 9350 }, { "epoch": 0.017541445163156432, "grad_norm": 1.9197875261306763, "learning_rate": 0.00015665, "loss": 5.611, "step": 9400 }, { "epoch": 0.017634750722534925, "grad_norm": 2.0102715492248535, "learning_rate": 0.00015748333333333333, "loss": 5.3812, "step": 9450 }, { "epoch": 0.01772805628191342, "grad_norm": 1.8926249742507935, "learning_rate": 0.00015831666666666667, "loss": 5.4527, "step": 9500 }, { "epoch": 0.01782136184129191, "grad_norm": 2.243831157684326, "learning_rate": 0.00015915, "loss": 5.4914, "step": 9550 }, { "epoch": 0.017914667400670402, "grad_norm": 1.8967549800872803, "learning_rate": 0.00015998333333333336, "loss": 5.5251, "step": 9600 }, { "epoch": 0.01800797296004889, "grad_norm": 1.785871982574463, "learning_rate": 0.00016081666666666668, "loss": 5.3667, "step": 9650 }, { "epoch": 0.018101278519427385, "grad_norm": 1.634459376335144, "learning_rate": 0.00016165000000000003, "loss": 5.7341, "step": 9700 }, { "epoch": 0.018194584078805875, "grad_norm": 2.1786787509918213, "learning_rate": 0.00016248333333333334, "loss": 5.4932, "step": 9750 }, { "epoch": 0.018287889638184368, "grad_norm": 2.0094616413116455, "learning_rate": 0.00016331666666666666, "loss": 5.5438, "step": 9800 }, { "epoch": 0.018381195197562858, "grad_norm": 1.5394549369812012, "learning_rate": 0.00016415, "loss": 5.3004, "step": 9850 }, { "epoch": 0.01847450075694135, "grad_norm": 2.034067392349243, "learning_rate": 0.00016498333333333332, "loss": 5.6939, "step": 9900 }, { "epoch": 0.01856780631631984, "grad_norm": 2.557887554168701, "learning_rate": 0.00016581666666666667, "loss": 5.4535, "step": 9950 }, { "epoch": 0.018661111875698334, "grad_norm": 1.7101596593856812, "learning_rate": 0.00016665, "loss": 5.5991, "step": 10000 }, { "epoch": 0.018754417435076824, "grad_norm": 1.1628295183181763, "learning_rate": 0.00016748333333333336, "loss": 5.5027, "step": 10050 }, { "epoch": 0.018847722994455317, "grad_norm": 1.6788240671157837, "learning_rate": 0.00016831666666666667, "loss": 5.3345, "step": 10100 }, { "epoch": 0.018941028553833807, "grad_norm": 2.166861057281494, "learning_rate": 0.00016915000000000002, "loss": 5.6807, "step": 10150 }, { "epoch": 0.0190343341132123, "grad_norm": 1.7033262252807617, "learning_rate": 0.00016998333333333334, "loss": 5.6921, "step": 10200 }, { "epoch": 0.01912763967259079, "grad_norm": 1.5504415035247803, "learning_rate": 0.00017081666666666668, "loss": 5.5782, "step": 10250 }, { "epoch": 0.019220945231969284, "grad_norm": 2.4242970943450928, "learning_rate": 0.00017165, "loss": 5.4636, "step": 10300 }, { "epoch": 0.019314250791347777, "grad_norm": 1.863472819328308, "learning_rate": 0.00017248333333333334, "loss": 5.4955, "step": 10350 }, { "epoch": 0.019407556350726267, "grad_norm": 1.856552004814148, "learning_rate": 0.0001733166666666667, "loss": 5.526, "step": 10400 }, { "epoch": 0.01950086191010476, "grad_norm": 2.124068260192871, "learning_rate": 0.00017415, "loss": 5.5683, "step": 10450 }, { "epoch": 0.01959416746948325, "grad_norm": 1.7942883968353271, "learning_rate": 0.00017498333333333335, "loss": 5.3387, "step": 10500 }, { "epoch": 0.019687473028861743, "grad_norm": 1.7919622659683228, "learning_rate": 0.00017581666666666667, "loss": 5.4876, "step": 10550 }, { "epoch": 0.019780778588240233, "grad_norm": 2.1417160034179688, "learning_rate": 0.00017665, "loss": 5.3508, "step": 10600 }, { "epoch": 0.019874084147618726, "grad_norm": 1.6931318044662476, "learning_rate": 0.00017748333333333333, "loss": 5.6171, "step": 10650 }, { "epoch": 0.019967389706997216, "grad_norm": 1.4930100440979004, "learning_rate": 0.00017831666666666667, "loss": 5.5452, "step": 10700 }, { "epoch": 0.02006069526637571, "grad_norm": 1.6434742212295532, "learning_rate": 0.00017915000000000002, "loss": 5.4924, "step": 10750 }, { "epoch": 0.0201540008257542, "grad_norm": 1.96444571018219, "learning_rate": 0.00017998333333333334, "loss": 5.3689, "step": 10800 }, { "epoch": 0.020247306385132693, "grad_norm": 2.2349231243133545, "learning_rate": 0.00018081666666666668, "loss": 5.3768, "step": 10850 }, { "epoch": 0.020340611944511183, "grad_norm": 1.5882576704025269, "learning_rate": 0.00018165, "loss": 5.747, "step": 10900 }, { "epoch": 0.020433917503889676, "grad_norm": 2.4228999614715576, "learning_rate": 0.00018248333333333334, "loss": 5.7638, "step": 10950 }, { "epoch": 0.020527223063268166, "grad_norm": 2.138207197189331, "learning_rate": 0.00018331666666666666, "loss": 5.7262, "step": 11000 }, { "epoch": 0.02062052862264666, "grad_norm": 1.9708890914916992, "learning_rate": 0.00018415, "loss": 5.5162, "step": 11050 }, { "epoch": 0.02071383418202515, "grad_norm": 1.5397372245788574, "learning_rate": 0.00018498333333333335, "loss": 5.5007, "step": 11100 }, { "epoch": 0.020807139741403642, "grad_norm": 1.579482913017273, "learning_rate": 0.0001858166666666667, "loss": 5.4921, "step": 11150 }, { "epoch": 0.020900445300782135, "grad_norm": 1.2825592756271362, "learning_rate": 0.00018665, "loss": 5.5486, "step": 11200 }, { "epoch": 0.020993750860160625, "grad_norm": 2.018866777420044, "learning_rate": 0.00018748333333333335, "loss": 5.4298, "step": 11250 }, { "epoch": 0.02108705641953912, "grad_norm": 1.8341054916381836, "learning_rate": 0.00018831666666666667, "loss": 5.5407, "step": 11300 }, { "epoch": 0.02118036197891761, "grad_norm": 1.6199865341186523, "learning_rate": 0.00018915000000000002, "loss": 5.3315, "step": 11350 }, { "epoch": 0.0212736675382961, "grad_norm": 1.5176050662994385, "learning_rate": 0.00018998333333333333, "loss": 5.4755, "step": 11400 }, { "epoch": 0.02136697309767459, "grad_norm": 1.8712049722671509, "learning_rate": 0.00019081666666666668, "loss": 5.4285, "step": 11450 }, { "epoch": 0.021460278657053085, "grad_norm": 1.531267762184143, "learning_rate": 0.00019165000000000002, "loss": 5.6335, "step": 11500 }, { "epoch": 0.021553584216431575, "grad_norm": 1.6521434783935547, "learning_rate": 0.00019248333333333334, "loss": 5.3548, "step": 11550 }, { "epoch": 0.021646889775810068, "grad_norm": 1.4107913970947266, "learning_rate": 0.00019331666666666669, "loss": 5.446, "step": 11600 }, { "epoch": 0.021740195335188558, "grad_norm": 1.8727481365203857, "learning_rate": 0.00019415, "loss": 5.5346, "step": 11650 }, { "epoch": 0.02183350089456705, "grad_norm": 2.040088653564453, "learning_rate": 0.00019498333333333335, "loss": 5.7011, "step": 11700 }, { "epoch": 0.02192680645394554, "grad_norm": 1.4047014713287354, "learning_rate": 0.00019581666666666666, "loss": 5.3325, "step": 11750 }, { "epoch": 0.022020112013324034, "grad_norm": 1.2155386209487915, "learning_rate": 0.00019665, "loss": 5.5547, "step": 11800 }, { "epoch": 0.022113417572702524, "grad_norm": 1.2685003280639648, "learning_rate": 0.00019748333333333335, "loss": 5.6444, "step": 11850 }, { "epoch": 0.022206723132081017, "grad_norm": 1.1052175760269165, "learning_rate": 0.00019831666666666667, "loss": 5.4218, "step": 11900 }, { "epoch": 0.022300028691459507, "grad_norm": 1.6609026193618774, "learning_rate": 0.00019915000000000002, "loss": 5.5406, "step": 11950 }, { "epoch": 0.022393334250838, "grad_norm": 1.7019258737564087, "learning_rate": 0.00019998333333333333, "loss": 5.4288, "step": 12000 }, { "epoch": 0.022393334250838, "eval_loss": 5.846170425415039, "eval_runtime": 230.7006, "eval_samples_per_second": 11.305, "eval_steps_per_second": 11.305, "eval_tts_loss": 6.596176531521512, "step": 12000 }, { "epoch": 0.022486639810216494, "grad_norm": 1.391849160194397, "learning_rate": 0.00019999999991525316, "loss": 5.438, "step": 12050 }, { "epoch": 0.022579945369594984, "grad_norm": 1.9549068212509155, "learning_rate": 0.0001999999996540592, "loss": 5.6524, "step": 12100 }, { "epoch": 0.022673250928973477, "grad_norm": 1.765764832496643, "learning_rate": 0.00019999999921638283, "loss": 5.529, "step": 12150 }, { "epoch": 0.022766556488351967, "grad_norm": 1.92975652217865, "learning_rate": 0.00019999999860222407, "loss": 5.1634, "step": 12200 }, { "epoch": 0.02285986204773046, "grad_norm": 1.627753734588623, "learning_rate": 0.00019999999781158292, "loss": 5.6305, "step": 12250 }, { "epoch": 0.02295316760710895, "grad_norm": 1.6781692504882812, "learning_rate": 0.00019999999684445935, "loss": 5.5936, "step": 12300 }, { "epoch": 0.023046473166487443, "grad_norm": 1.200257658958435, "learning_rate": 0.00019999999570085336, "loss": 5.4897, "step": 12350 }, { "epoch": 0.023139778725865933, "grad_norm": 1.7044615745544434, "learning_rate": 0.00019999999438076503, "loss": 5.2909, "step": 12400 }, { "epoch": 0.023233084285244426, "grad_norm": 1.4790157079696655, "learning_rate": 0.00019999999288419424, "loss": 5.3771, "step": 12450 }, { "epoch": 0.023326389844622916, "grad_norm": 1.5862489938735962, "learning_rate": 0.0001999999912111411, "loss": 5.2316, "step": 12500 }, { "epoch": 0.02341969540400141, "grad_norm": 1.9885330200195312, "learning_rate": 0.00019999998936160555, "loss": 5.3982, "step": 12550 }, { "epoch": 0.0235130009633799, "grad_norm": 1.4161193370819092, "learning_rate": 0.00019999998733558766, "loss": 5.4459, "step": 12600 }, { "epoch": 0.023606306522758393, "grad_norm": 1.5754268169403076, "learning_rate": 0.00019999998513308732, "loss": 5.3697, "step": 12650 }, { "epoch": 0.023699612082136882, "grad_norm": 1.0978269577026367, "learning_rate": 0.00019999998275410462, "loss": 5.4439, "step": 12700 }, { "epoch": 0.023792917641515376, "grad_norm": 1.4332829713821411, "learning_rate": 0.0001999999801986396, "loss": 5.2534, "step": 12750 }, { "epoch": 0.023886223200893866, "grad_norm": 1.157253623008728, "learning_rate": 0.00019999997746669216, "loss": 5.4755, "step": 12800 }, { "epoch": 0.02397952876027236, "grad_norm": 1.577779769897461, "learning_rate": 0.00019999997455826237, "loss": 5.4148, "step": 12850 }, { "epoch": 0.024072834319650852, "grad_norm": 1.4218207597732544, "learning_rate": 0.00019999997147335022, "loss": 5.4531, "step": 12900 }, { "epoch": 0.024166139879029342, "grad_norm": 2.003883123397827, "learning_rate": 0.0001999999682119557, "loss": 5.6188, "step": 12950 }, { "epoch": 0.024259445438407835, "grad_norm": 2.01855731010437, "learning_rate": 0.00019999996477407884, "loss": 5.5149, "step": 13000 }, { "epoch": 0.024352750997786325, "grad_norm": 1.6317418813705444, "learning_rate": 0.00019999996115971965, "loss": 5.3213, "step": 13050 }, { "epoch": 0.02444605655716482, "grad_norm": 1.6599922180175781, "learning_rate": 0.0001999999573688781, "loss": 5.4164, "step": 13100 }, { "epoch": 0.02453936211654331, "grad_norm": 1.4030259847640991, "learning_rate": 0.00019999995340155423, "loss": 5.254, "step": 13150 }, { "epoch": 0.0246326676759218, "grad_norm": 1.507149338722229, "learning_rate": 0.00019999994925774806, "loss": 5.2633, "step": 13200 }, { "epoch": 0.02472597323530029, "grad_norm": 1.6153885126113892, "learning_rate": 0.00019999994493745957, "loss": 5.4768, "step": 13250 }, { "epoch": 0.024819278794678785, "grad_norm": 1.2192950248718262, "learning_rate": 0.00019999994044068875, "loss": 5.465, "step": 13300 }, { "epoch": 0.024912584354057275, "grad_norm": 1.3811638355255127, "learning_rate": 0.00019999993576743565, "loss": 5.4459, "step": 13350 }, { "epoch": 0.025005889913435768, "grad_norm": 1.7365858554840088, "learning_rate": 0.00019999993091770026, "loss": 5.4616, "step": 13400 }, { "epoch": 0.025099195472814258, "grad_norm": 1.4087482690811157, "learning_rate": 0.00019999992589148256, "loss": 5.4608, "step": 13450 }, { "epoch": 0.02519250103219275, "grad_norm": 1.7531602382659912, "learning_rate": 0.00019999992068878264, "loss": 5.3973, "step": 13500 }, { "epoch": 0.02528580659157124, "grad_norm": 1.653184175491333, "learning_rate": 0.0001999999153096004, "loss": 5.219, "step": 13550 }, { "epoch": 0.025379112150949734, "grad_norm": 1.3364681005477905, "learning_rate": 0.00019999990975393592, "loss": 5.4375, "step": 13600 }, { "epoch": 0.025472417710328224, "grad_norm": 1.3480881452560425, "learning_rate": 0.0001999999040217892, "loss": 5.4107, "step": 13650 }, { "epoch": 0.025565723269706717, "grad_norm": 1.8013604879379272, "learning_rate": 0.00019999989811316025, "loss": 5.4488, "step": 13700 }, { "epoch": 0.02565902882908521, "grad_norm": 1.5909210443496704, "learning_rate": 0.00019999989202804908, "loss": 5.6496, "step": 13750 }, { "epoch": 0.0257523343884637, "grad_norm": 1.897382140159607, "learning_rate": 0.0001999998857664557, "loss": 5.3464, "step": 13800 }, { "epoch": 0.025845639947842194, "grad_norm": 1.844641089439392, "learning_rate": 0.0001999998793283801, "loss": 5.536, "step": 13850 }, { "epoch": 0.025938945507220684, "grad_norm": 1.41611909866333, "learning_rate": 0.00019999987271382232, "loss": 5.5876, "step": 13900 }, { "epoch": 0.026032251066599177, "grad_norm": 1.5440382957458496, "learning_rate": 0.00019999986592278234, "loss": 5.2344, "step": 13950 }, { "epoch": 0.026125556625977667, "grad_norm": 1.3486498594284058, "learning_rate": 0.0001999998589552602, "loss": 5.4819, "step": 14000 }, { "epoch": 0.02621886218535616, "grad_norm": 1.6134167909622192, "learning_rate": 0.0001999998518112559, "loss": 5.6369, "step": 14050 }, { "epoch": 0.02631216774473465, "grad_norm": 1.950914740562439, "learning_rate": 0.0001999998444907695, "loss": 5.3324, "step": 14100 }, { "epoch": 0.026405473304113143, "grad_norm": 1.515940546989441, "learning_rate": 0.00019999983699380095, "loss": 5.5848, "step": 14150 }, { "epoch": 0.026498778863491633, "grad_norm": 1.8684641122817993, "learning_rate": 0.00019999982932035023, "loss": 5.4585, "step": 14200 }, { "epoch": 0.026592084422870126, "grad_norm": 1.9417451620101929, "learning_rate": 0.00019999982147041748, "loss": 5.3855, "step": 14250 }, { "epoch": 0.026685389982248616, "grad_norm": 1.5790482759475708, "learning_rate": 0.0001999998134440026, "loss": 5.4521, "step": 14300 }, { "epoch": 0.02677869554162711, "grad_norm": 1.9342408180236816, "learning_rate": 0.00019999980524110564, "loss": 5.3272, "step": 14350 }, { "epoch": 0.0268720011010056, "grad_norm": 1.617722988128662, "learning_rate": 0.00019999979686172664, "loss": 5.2174, "step": 14400 }, { "epoch": 0.026965306660384093, "grad_norm": 1.0585445165634155, "learning_rate": 0.00019999978830586557, "loss": 5.1622, "step": 14450 }, { "epoch": 0.027058612219762582, "grad_norm": 1.661190390586853, "learning_rate": 0.00019999977957352248, "loss": 5.1873, "step": 14500 }, { "epoch": 0.027151917779141076, "grad_norm": 1.5143160820007324, "learning_rate": 0.0001999997706646974, "loss": 5.507, "step": 14550 }, { "epoch": 0.02724522333851957, "grad_norm": 1.7400487661361694, "learning_rate": 0.0001999997615793903, "loss": 5.0888, "step": 14600 }, { "epoch": 0.02733852889789806, "grad_norm": 1.344720482826233, "learning_rate": 0.0001999997523176012, "loss": 5.2325, "step": 14650 }, { "epoch": 0.027431834457276552, "grad_norm": 1.390224575996399, "learning_rate": 0.00019999974287933015, "loss": 5.2955, "step": 14700 }, { "epoch": 0.027525140016655042, "grad_norm": 1.685595154762268, "learning_rate": 0.00019999973326457714, "loss": 5.3243, "step": 14750 }, { "epoch": 0.027618445576033535, "grad_norm": 1.574565052986145, "learning_rate": 0.0001999997234733422, "loss": 5.5154, "step": 14800 }, { "epoch": 0.027711751135412025, "grad_norm": 1.8520110845565796, "learning_rate": 0.0001999997135056254, "loss": 5.4337, "step": 14850 }, { "epoch": 0.02780505669479052, "grad_norm": 1.9176346063613892, "learning_rate": 0.0001999997033614266, "loss": 5.2195, "step": 14900 }, { "epoch": 0.027898362254169008, "grad_norm": 1.4428346157073975, "learning_rate": 0.000199999693040746, "loss": 5.1029, "step": 14950 }, { "epoch": 0.0279916678135475, "grad_norm": 1.408713459968567, "learning_rate": 0.0001999996825435835, "loss": 5.4174, "step": 15000 }, { "epoch": 0.0279916678135475, "eval_loss": 5.698703289031982, "eval_runtime": 230.7647, "eval_samples_per_second": 11.302, "eval_steps_per_second": 11.302, "eval_tts_loss": 6.7004290100964115, "step": 15000 }, { "epoch": 0.02808497337292599, "grad_norm": 1.4793498516082764, "learning_rate": 0.00019999967186993915, "loss": 5.5412, "step": 15050 }, { "epoch": 0.028178278932304485, "grad_norm": 1.2378696203231812, "learning_rate": 0.000199999661019813, "loss": 5.3518, "step": 15100 }, { "epoch": 0.028271584491682974, "grad_norm": 1.4232487678527832, "learning_rate": 0.000199999649993205, "loss": 5.4232, "step": 15150 }, { "epoch": 0.028364890051061468, "grad_norm": 1.6765247583389282, "learning_rate": 0.00019999963879011525, "loss": 5.5725, "step": 15200 }, { "epoch": 0.028458195610439958, "grad_norm": 1.599683403968811, "learning_rate": 0.00019999962741054373, "loss": 5.4653, "step": 15250 }, { "epoch": 0.02855150116981845, "grad_norm": 1.5458168983459473, "learning_rate": 0.00019999961585449047, "loss": 5.0618, "step": 15300 }, { "epoch": 0.02864480672919694, "grad_norm": 1.09707510471344, "learning_rate": 0.0001999996041219555, "loss": 5.5357, "step": 15350 }, { "epoch": 0.028738112288575434, "grad_norm": 2.0242269039154053, "learning_rate": 0.00019999959221293878, "loss": 5.3426, "step": 15400 }, { "epoch": 0.028831417847953927, "grad_norm": 1.6951383352279663, "learning_rate": 0.00019999958012744038, "loss": 5.4521, "step": 15450 }, { "epoch": 0.028924723407332417, "grad_norm": 1.8190480470657349, "learning_rate": 0.00019999956786546035, "loss": 5.2914, "step": 15500 }, { "epoch": 0.02901802896671091, "grad_norm": 1.7063592672348022, "learning_rate": 0.00019999955542699863, "loss": 5.2852, "step": 15550 }, { "epoch": 0.0291113345260894, "grad_norm": 1.7153364419937134, "learning_rate": 0.00019999954281205533, "loss": 5.0807, "step": 15600 }, { "epoch": 0.029204640085467894, "grad_norm": 1.677153468132019, "learning_rate": 0.0001999995300206304, "loss": 5.1755, "step": 15650 }, { "epoch": 0.029297945644846383, "grad_norm": 1.5905139446258545, "learning_rate": 0.00019999951705272393, "loss": 5.2782, "step": 15700 }, { "epoch": 0.029391251204224877, "grad_norm": 1.6873728036880493, "learning_rate": 0.0001999995039083359, "loss": 5.51, "step": 15750 }, { "epoch": 0.029484556763603367, "grad_norm": 1.4348547458648682, "learning_rate": 0.00019999949058746632, "loss": 5.2227, "step": 15800 }, { "epoch": 0.02957786232298186, "grad_norm": 1.599501132965088, "learning_rate": 0.00019999947709011524, "loss": 5.4851, "step": 15850 }, { "epoch": 0.02967116788236035, "grad_norm": 1.2430881261825562, "learning_rate": 0.0001999994634162827, "loss": 5.059, "step": 15900 }, { "epoch": 0.029764473441738843, "grad_norm": 1.2983192205429077, "learning_rate": 0.00019999944956596867, "loss": 5.3391, "step": 15950 }, { "epoch": 0.029857779001117333, "grad_norm": 1.8286632299423218, "learning_rate": 0.00019999943553917324, "loss": 5.068, "step": 16000 }, { "epoch": 0.029951084560495826, "grad_norm": 1.7763135433197021, "learning_rate": 0.0001999994213358964, "loss": 5.2494, "step": 16050 }, { "epoch": 0.030044390119874316, "grad_norm": 1.5806595087051392, "learning_rate": 0.00019999940695613816, "loss": 5.3536, "step": 16100 }, { "epoch": 0.03013769567925281, "grad_norm": 1.542901635169983, "learning_rate": 0.00019999939239989857, "loss": 5.2556, "step": 16150 }, { "epoch": 0.0302310012386313, "grad_norm": 1.390985369682312, "learning_rate": 0.00019999937766717764, "loss": 5.155, "step": 16200 }, { "epoch": 0.030324306798009792, "grad_norm": 1.0642412900924683, "learning_rate": 0.0001999993627579754, "loss": 5.3142, "step": 16250 }, { "epoch": 0.030417612357388286, "grad_norm": 1.5385115146636963, "learning_rate": 0.00019999934767229188, "loss": 5.2755, "step": 16300 }, { "epoch": 0.030510917916766776, "grad_norm": 1.719664216041565, "learning_rate": 0.00019999933241012711, "loss": 5.3522, "step": 16350 }, { "epoch": 0.03060422347614527, "grad_norm": 1.6435158252716064, "learning_rate": 0.00019999931697148112, "loss": 5.1949, "step": 16400 }, { "epoch": 0.03069752903552376, "grad_norm": 1.2330780029296875, "learning_rate": 0.00019999930135635395, "loss": 5.0197, "step": 16450 }, { "epoch": 0.030790834594902252, "grad_norm": 1.433413028717041, "learning_rate": 0.0001999992855647456, "loss": 5.4277, "step": 16500 }, { "epoch": 0.030884140154280742, "grad_norm": 1.8009040355682373, "learning_rate": 0.00019999926959665608, "loss": 5.0877, "step": 16550 }, { "epoch": 0.030977445713659235, "grad_norm": 1.1439430713653564, "learning_rate": 0.00019999925345208546, "loss": 5.3289, "step": 16600 }, { "epoch": 0.031070751273037725, "grad_norm": 1.3732225894927979, "learning_rate": 0.00019999923713103374, "loss": 5.1679, "step": 16650 }, { "epoch": 0.03116405683241622, "grad_norm": 1.722701072692871, "learning_rate": 0.000199999220633501, "loss": 5.3267, "step": 16700 }, { "epoch": 0.03125736239179471, "grad_norm": 1.6895197629928589, "learning_rate": 0.00019999920395948718, "loss": 5.1153, "step": 16750 }, { "epoch": 0.0313506679511732, "grad_norm": 1.7803030014038086, "learning_rate": 0.00019999918710899238, "loss": 5.2153, "step": 16800 }, { "epoch": 0.03144397351055169, "grad_norm": 1.187790036201477, "learning_rate": 0.00019999917008201663, "loss": 5.2711, "step": 16850 }, { "epoch": 0.03153727906993018, "grad_norm": 1.3633534908294678, "learning_rate": 0.00019999915287855993, "loss": 5.5233, "step": 16900 }, { "epoch": 0.03163058462930868, "grad_norm": 1.4957891702651978, "learning_rate": 0.00019999913549862233, "loss": 5.2524, "step": 16950 }, { "epoch": 0.03172389018868717, "grad_norm": 1.6286832094192505, "learning_rate": 0.00019999911794220382, "loss": 5.1533, "step": 17000 }, { "epoch": 0.03181719574806566, "grad_norm": 1.2257238626480103, "learning_rate": 0.00019999910020930447, "loss": 5.1066, "step": 17050 }, { "epoch": 0.03191050130744415, "grad_norm": 1.537248134613037, "learning_rate": 0.00019999908229992433, "loss": 5.1558, "step": 17100 }, { "epoch": 0.032003806866822644, "grad_norm": 1.7594759464263916, "learning_rate": 0.00019999906421406337, "loss": 5.2488, "step": 17150 }, { "epoch": 0.032097112426201134, "grad_norm": 2.0060057640075684, "learning_rate": 0.00019999904595172167, "loss": 5.1244, "step": 17200 }, { "epoch": 0.032190417985579624, "grad_norm": 1.306092381477356, "learning_rate": 0.00019999902751289926, "loss": 5.2285, "step": 17250 }, { "epoch": 0.03228372354495812, "grad_norm": 1.605584979057312, "learning_rate": 0.00019999900889759617, "loss": 5.148, "step": 17300 }, { "epoch": 0.03237702910433661, "grad_norm": 1.3685698509216309, "learning_rate": 0.0001999989901058124, "loss": 5.2305, "step": 17350 }, { "epoch": 0.0324703346637151, "grad_norm": 1.7700085639953613, "learning_rate": 0.00019999897113754801, "loss": 5.2642, "step": 17400 }, { "epoch": 0.03256364022309359, "grad_norm": 1.8209742307662964, "learning_rate": 0.00019999895199280306, "loss": 5.2634, "step": 17450 }, { "epoch": 0.03265694578247209, "grad_norm": 1.7181389331817627, "learning_rate": 0.00019999893267157753, "loss": 5.3703, "step": 17500 }, { "epoch": 0.03275025134185058, "grad_norm": 1.6699622869491577, "learning_rate": 0.00019999891317387149, "loss": 5.1961, "step": 17550 }, { "epoch": 0.032843556901229067, "grad_norm": 1.7012041807174683, "learning_rate": 0.00019999889349968494, "loss": 5.2565, "step": 17600 }, { "epoch": 0.032936862460607556, "grad_norm": 1.4097135066986084, "learning_rate": 0.00019999887364901796, "loss": 5.3606, "step": 17650 }, { "epoch": 0.03303016801998605, "grad_norm": 1.8546066284179688, "learning_rate": 0.00019999885362187057, "loss": 5.1426, "step": 17700 }, { "epoch": 0.03312347357936454, "grad_norm": 1.6862845420837402, "learning_rate": 0.00019999883341824282, "loss": 5.4663, "step": 17750 }, { "epoch": 0.03321677913874303, "grad_norm": 1.4715492725372314, "learning_rate": 0.00019999881303813468, "loss": 5.2308, "step": 17800 }, { "epoch": 0.03331008469812152, "grad_norm": 1.50525963306427, "learning_rate": 0.00019999879248154624, "loss": 5.2341, "step": 17850 }, { "epoch": 0.03340339025750002, "grad_norm": 1.6435831785202026, "learning_rate": 0.00019999877174847752, "loss": 5.2826, "step": 17900 }, { "epoch": 0.03349669581687851, "grad_norm": 1.4661362171173096, "learning_rate": 0.00019999875083892858, "loss": 5.2187, "step": 17950 }, { "epoch": 0.033590001376257, "grad_norm": 1.255449652671814, "learning_rate": 0.00019999872975289944, "loss": 5.1868, "step": 18000 }, { "epoch": 0.033590001376257, "eval_loss": 5.565659046173096, "eval_runtime": 233.225, "eval_samples_per_second": 11.182, "eval_steps_per_second": 11.182, "eval_tts_loss": 6.789093455502569, "step": 18000 }, { "epoch": 0.033683306935635496, "grad_norm": 1.6058536767959595, "learning_rate": 0.00019999870849039014, "loss": 5.0283, "step": 18050 }, { "epoch": 0.033776612495013986, "grad_norm": 1.931518793106079, "learning_rate": 0.0001999986870514007, "loss": 5.0369, "step": 18100 }, { "epoch": 0.033869918054392475, "grad_norm": 1.3181180953979492, "learning_rate": 0.00019999866543593116, "loss": 5.0891, "step": 18150 }, { "epoch": 0.033963223613770965, "grad_norm": 1.6189442873001099, "learning_rate": 0.0001999986436439816, "loss": 5.2425, "step": 18200 }, { "epoch": 0.03405652917314946, "grad_norm": 1.4186607599258423, "learning_rate": 0.000199998621675552, "loss": 4.9802, "step": 18250 }, { "epoch": 0.03414983473252795, "grad_norm": 1.727707028388977, "learning_rate": 0.00019999859953064245, "loss": 5.1903, "step": 18300 }, { "epoch": 0.03424314029190644, "grad_norm": 2.1892459392547607, "learning_rate": 0.00019999857720925297, "loss": 5.0998, "step": 18350 }, { "epoch": 0.03433644585128493, "grad_norm": 1.567331075668335, "learning_rate": 0.00019999855471138358, "loss": 5.241, "step": 18400 }, { "epoch": 0.03442975141066343, "grad_norm": 1.5945489406585693, "learning_rate": 0.00019999853203703436, "loss": 5.1888, "step": 18450 }, { "epoch": 0.03452305697004192, "grad_norm": 1.468598484992981, "learning_rate": 0.0001999985091862053, "loss": 5.2305, "step": 18500 }, { "epoch": 0.03461636252942041, "grad_norm": 1.801831603050232, "learning_rate": 0.00019999848615889648, "loss": 5.0736, "step": 18550 }, { "epoch": 0.0347096680887989, "grad_norm": 1.983370065689087, "learning_rate": 0.00019999846295510793, "loss": 5.1701, "step": 18600 }, { "epoch": 0.034802973648177395, "grad_norm": 1.4538954496383667, "learning_rate": 0.00019999843957483967, "loss": 5.1838, "step": 18650 }, { "epoch": 0.034896279207555884, "grad_norm": 2.007187843322754, "learning_rate": 0.0001999984160180918, "loss": 5.1847, "step": 18700 }, { "epoch": 0.034989584766934374, "grad_norm": 1.3527206182479858, "learning_rate": 0.00019999839228486427, "loss": 5.21, "step": 18750 }, { "epoch": 0.035082890326312864, "grad_norm": 0.93910813331604, "learning_rate": 0.0001999983683751572, "loss": 5.217, "step": 18800 }, { "epoch": 0.03517619588569136, "grad_norm": 1.9449275732040405, "learning_rate": 0.00019999834428897063, "loss": 5.2404, "step": 18850 }, { "epoch": 0.03526950144506985, "grad_norm": 1.5186761617660522, "learning_rate": 0.00019999832002630453, "loss": 5.1827, "step": 18900 }, { "epoch": 0.03536280700444834, "grad_norm": 1.8738884925842285, "learning_rate": 0.000199998295587159, "loss": 5.2015, "step": 18950 }, { "epoch": 0.03545611256382684, "grad_norm": 1.6167317628860474, "learning_rate": 0.0001999982709715341, "loss": 5.1333, "step": 19000 }, { "epoch": 0.03554941812320533, "grad_norm": 1.5436670780181885, "learning_rate": 0.00019999824617942985, "loss": 5.2181, "step": 19050 }, { "epoch": 0.03564272368258382, "grad_norm": 1.9840049743652344, "learning_rate": 0.00019999822121084625, "loss": 4.9849, "step": 19100 }, { "epoch": 0.03573602924196231, "grad_norm": 2.000702381134033, "learning_rate": 0.00019999819606578343, "loss": 5.0625, "step": 19150 }, { "epoch": 0.035829334801340804, "grad_norm": 1.5434222221374512, "learning_rate": 0.00019999817074424137, "loss": 5.0164, "step": 19200 }, { "epoch": 0.03592264036071929, "grad_norm": 1.492538332939148, "learning_rate": 0.0001999981452462201, "loss": 5.2329, "step": 19250 }, { "epoch": 0.03601594592009778, "grad_norm": 1.3936853408813477, "learning_rate": 0.00019999811957171975, "loss": 5.2539, "step": 19300 }, { "epoch": 0.03610925147947627, "grad_norm": 1.3118730783462524, "learning_rate": 0.00019999809372074027, "loss": 5.1528, "step": 19350 }, { "epoch": 0.03620255703885477, "grad_norm": 1.0975465774536133, "learning_rate": 0.00019999806769328178, "loss": 5.1746, "step": 19400 }, { "epoch": 0.03629586259823326, "grad_norm": 1.3111509084701538, "learning_rate": 0.0001999980414893443, "loss": 5.1834, "step": 19450 }, { "epoch": 0.03638916815761175, "grad_norm": 1.45402193069458, "learning_rate": 0.00019999801510892784, "loss": 5.2616, "step": 19500 }, { "epoch": 0.03648247371699024, "grad_norm": 1.5854758024215698, "learning_rate": 0.0001999979885520325, "loss": 5.1249, "step": 19550 }, { "epoch": 0.036575779276368736, "grad_norm": 1.6653951406478882, "learning_rate": 0.00019999796181865832, "loss": 5.1441, "step": 19600 }, { "epoch": 0.036669084835747226, "grad_norm": 1.6704020500183105, "learning_rate": 0.0001999979349088053, "loss": 5.1533, "step": 19650 }, { "epoch": 0.036762390395125716, "grad_norm": 1.5533548593521118, "learning_rate": 0.00019999790782247354, "loss": 5.3171, "step": 19700 }, { "epoch": 0.03685569595450421, "grad_norm": 1.6590992212295532, "learning_rate": 0.0001999978805596631, "loss": 5.3232, "step": 19750 }, { "epoch": 0.0369490015138827, "grad_norm": 1.5630754232406616, "learning_rate": 0.00019999785312037392, "loss": 5.138, "step": 19800 }, { "epoch": 0.03704230707326119, "grad_norm": 1.771061897277832, "learning_rate": 0.00019999782550460618, "loss": 5.1169, "step": 19850 }, { "epoch": 0.03713561263263968, "grad_norm": 1.6629877090454102, "learning_rate": 0.00019999779771235983, "loss": 5.207, "step": 19900 }, { "epoch": 0.03722891819201818, "grad_norm": 1.2950810194015503, "learning_rate": 0.00019999776974363503, "loss": 5.1812, "step": 19950 }, { "epoch": 0.03732222375139667, "grad_norm": 0.9732388257980347, "learning_rate": 0.00019999774159843171, "loss": 5.2428, "step": 20000 }, { "epoch": 0.03741552931077516, "grad_norm": 1.3945764303207397, "learning_rate": 0.00019999771327675, "loss": 5.2425, "step": 20050 }, { "epoch": 0.03750883487015365, "grad_norm": 1.880929946899414, "learning_rate": 0.0001999976847785899, "loss": 5.3158, "step": 20100 }, { "epoch": 0.037602140429532145, "grad_norm": 1.3822988271713257, "learning_rate": 0.00019999765610395146, "loss": 5.1829, "step": 20150 }, { "epoch": 0.037695445988910635, "grad_norm": 1.7985529899597168, "learning_rate": 0.00019999762725283479, "loss": 5.0514, "step": 20200 }, { "epoch": 0.037788751548289125, "grad_norm": 1.568518877029419, "learning_rate": 0.0001999975982252399, "loss": 5.2333, "step": 20250 }, { "epoch": 0.037882057107667615, "grad_norm": 1.4257228374481201, "learning_rate": 0.00019999756902116684, "loss": 5.1142, "step": 20300 }, { "epoch": 0.03797536266704611, "grad_norm": 1.0664187669754028, "learning_rate": 0.00019999753964061566, "loss": 5.204, "step": 20350 }, { "epoch": 0.0380686682264246, "grad_norm": 1.2831335067749023, "learning_rate": 0.00019999751008358642, "loss": 5.2272, "step": 20400 }, { "epoch": 0.03816197378580309, "grad_norm": 1.7534427642822266, "learning_rate": 0.00019999748035007917, "loss": 5.1009, "step": 20450 }, { "epoch": 0.03825527934518158, "grad_norm": 1.8466085195541382, "learning_rate": 0.00019999745044009396, "loss": 5.1638, "step": 20500 }, { "epoch": 0.03834858490456008, "grad_norm": 1.4553605318069458, "learning_rate": 0.00019999742035363083, "loss": 5.0138, "step": 20550 }, { "epoch": 0.03844189046393857, "grad_norm": 1.5238099098205566, "learning_rate": 0.00019999739009068988, "loss": 5.1723, "step": 20600 }, { "epoch": 0.03853519602331706, "grad_norm": 1.434646487236023, "learning_rate": 0.0001999973596512711, "loss": 5.1743, "step": 20650 }, { "epoch": 0.038628501582695554, "grad_norm": 1.4512965679168701, "learning_rate": 0.0001999973290353746, "loss": 5.2082, "step": 20700 }, { "epoch": 0.038721807142074044, "grad_norm": 1.4691851139068604, "learning_rate": 0.0001999972982430004, "loss": 5.3011, "step": 20750 }, { "epoch": 0.038815112701452534, "grad_norm": 1.7574752569198608, "learning_rate": 0.00019999726727414855, "loss": 4.8797, "step": 20800 }, { "epoch": 0.038908418260831024, "grad_norm": 1.9155274629592896, "learning_rate": 0.00019999723612881915, "loss": 5.2745, "step": 20850 }, { "epoch": 0.03900172382020952, "grad_norm": 1.9480228424072266, "learning_rate": 0.0001999972048070122, "loss": 5.1093, "step": 20900 }, { "epoch": 0.03909502937958801, "grad_norm": 1.314173698425293, "learning_rate": 0.0001999971733087278, "loss": 4.9488, "step": 20950 }, { "epoch": 0.0391883349389665, "grad_norm": 1.5758183002471924, "learning_rate": 0.00019999714163396597, "loss": 5.0716, "step": 21000 }, { "epoch": 0.0391883349389665, "eval_loss": 5.446030616760254, "eval_runtime": 233.6236, "eval_samples_per_second": 11.163, "eval_steps_per_second": 11.163, "eval_tts_loss": 6.94306072605591, "step": 21000 }, { "epoch": 0.03928164049834499, "grad_norm": 1.767700433731079, "learning_rate": 0.00019999710978272676, "loss": 5.152, "step": 21050 }, { "epoch": 0.03937494605772349, "grad_norm": 1.171859622001648, "learning_rate": 0.00019999707775501026, "loss": 4.9265, "step": 21100 }, { "epoch": 0.039468251617101976, "grad_norm": 1.6448893547058105, "learning_rate": 0.00019999704555081653, "loss": 5.2328, "step": 21150 }, { "epoch": 0.039561557176480466, "grad_norm": 1.2383092641830444, "learning_rate": 0.00019999701317014557, "loss": 5.0509, "step": 21200 }, { "epoch": 0.039654862735858956, "grad_norm": 1.47544264793396, "learning_rate": 0.00019999698061299752, "loss": 5.0847, "step": 21250 }, { "epoch": 0.03974816829523745, "grad_norm": 1.4962610006332397, "learning_rate": 0.00019999694787937238, "loss": 5.1514, "step": 21300 }, { "epoch": 0.03984147385461594, "grad_norm": 1.7069225311279297, "learning_rate": 0.00019999691496927022, "loss": 5.0082, "step": 21350 }, { "epoch": 0.03993477941399443, "grad_norm": 1.970062017440796, "learning_rate": 0.00019999688188269111, "loss": 5.162, "step": 21400 }, { "epoch": 0.04002808497337293, "grad_norm": 1.9130045175552368, "learning_rate": 0.00019999684861963507, "loss": 5.1627, "step": 21450 }, { "epoch": 0.04012139053275142, "grad_norm": 1.7424521446228027, "learning_rate": 0.00019999681518010221, "loss": 5.1727, "step": 21500 }, { "epoch": 0.04021469609212991, "grad_norm": 1.2897844314575195, "learning_rate": 0.00019999678156409258, "loss": 5.0252, "step": 21550 }, { "epoch": 0.0403080016515084, "grad_norm": 1.493475079536438, "learning_rate": 0.0001999967477716062, "loss": 5.029, "step": 21600 }, { "epoch": 0.040401307210886896, "grad_norm": 1.2159661054611206, "learning_rate": 0.00019999671380264316, "loss": 5.0264, "step": 21650 }, { "epoch": 0.040494612770265385, "grad_norm": 1.4600199460983276, "learning_rate": 0.0001999966796572035, "loss": 4.9779, "step": 21700 }, { "epoch": 0.040587918329643875, "grad_norm": 1.526752233505249, "learning_rate": 0.00019999664533528732, "loss": 5.1797, "step": 21750 }, { "epoch": 0.040681223889022365, "grad_norm": 1.6486625671386719, "learning_rate": 0.00019999661083689466, "loss": 5.1361, "step": 21800 }, { "epoch": 0.04077452944840086, "grad_norm": 1.9790832996368408, "learning_rate": 0.00019999657616202556, "loss": 4.9149, "step": 21850 }, { "epoch": 0.04086783500777935, "grad_norm": 1.4455928802490234, "learning_rate": 0.00019999654131068012, "loss": 5.1656, "step": 21900 }, { "epoch": 0.04096114056715784, "grad_norm": 1.556361436843872, "learning_rate": 0.00019999650628285836, "loss": 5.1226, "step": 21950 }, { "epoch": 0.04105444612653633, "grad_norm": 1.2878562211990356, "learning_rate": 0.00019999647107856036, "loss": 5.0742, "step": 22000 }, { "epoch": 0.04114775168591483, "grad_norm": 1.4406765699386597, "learning_rate": 0.0001999964356977862, "loss": 5.167, "step": 22050 }, { "epoch": 0.04124105724529332, "grad_norm": 1.3805707693099976, "learning_rate": 0.0001999964001405359, "loss": 4.9802, "step": 22100 }, { "epoch": 0.04133436280467181, "grad_norm": 1.7562808990478516, "learning_rate": 0.00019999636440680955, "loss": 5.2414, "step": 22150 }, { "epoch": 0.0414276683640503, "grad_norm": 1.5996421575546265, "learning_rate": 0.00019999632849660722, "loss": 5.2813, "step": 22200 }, { "epoch": 0.041520973923428794, "grad_norm": 1.2786633968353271, "learning_rate": 0.00019999629240992898, "loss": 5.3503, "step": 22250 }, { "epoch": 0.041614279482807284, "grad_norm": 1.7056688070297241, "learning_rate": 0.00019999625614677488, "loss": 5.205, "step": 22300 }, { "epoch": 0.041707585042185774, "grad_norm": 1.5653877258300781, "learning_rate": 0.00019999621970714495, "loss": 5.1535, "step": 22350 }, { "epoch": 0.04180089060156427, "grad_norm": 1.6800984144210815, "learning_rate": 0.00019999618309103935, "loss": 5.0171, "step": 22400 }, { "epoch": 0.04189419616094276, "grad_norm": 1.608587384223938, "learning_rate": 0.000199996146298458, "loss": 5.0021, "step": 22450 }, { "epoch": 0.04198750172032125, "grad_norm": 1.1938655376434326, "learning_rate": 0.00019999610932940108, "loss": 4.7871, "step": 22500 }, { "epoch": 0.04208080727969974, "grad_norm": 2.0445282459259033, "learning_rate": 0.00019999607218386864, "loss": 5.2227, "step": 22550 }, { "epoch": 0.04217411283907824, "grad_norm": 1.440500020980835, "learning_rate": 0.0001999960348618607, "loss": 4.9996, "step": 22600 }, { "epoch": 0.04226741839845673, "grad_norm": 1.4636365175247192, "learning_rate": 0.0001999959973633774, "loss": 5.2787, "step": 22650 }, { "epoch": 0.04236072395783522, "grad_norm": 1.6443440914154053, "learning_rate": 0.0001999959596884187, "loss": 5.1626, "step": 22700 }, { "epoch": 0.04245402951721371, "grad_norm": 1.3920623064041138, "learning_rate": 0.00019999592183698477, "loss": 5.1757, "step": 22750 }, { "epoch": 0.0425473350765922, "grad_norm": 1.5439649820327759, "learning_rate": 0.00019999588380907563, "loss": 5.1367, "step": 22800 }, { "epoch": 0.04264064063597069, "grad_norm": 1.575374960899353, "learning_rate": 0.00019999584560469134, "loss": 5.0247, "step": 22850 }, { "epoch": 0.04273394619534918, "grad_norm": 1.206429123878479, "learning_rate": 0.00019999580722383198, "loss": 5.1514, "step": 22900 }, { "epoch": 0.04282725175472767, "grad_norm": 1.5674757957458496, "learning_rate": 0.0001999957686664976, "loss": 5.208, "step": 22950 }, { "epoch": 0.04292055731410617, "grad_norm": 1.2532552480697632, "learning_rate": 0.00019999572993268828, "loss": 4.9993, "step": 23000 }, { "epoch": 0.04301386287348466, "grad_norm": 1.4708222150802612, "learning_rate": 0.00019999569102240413, "loss": 5.121, "step": 23050 }, { "epoch": 0.04310716843286315, "grad_norm": 1.6088930368423462, "learning_rate": 0.00019999565193564513, "loss": 5.0844, "step": 23100 }, { "epoch": 0.043200473992241646, "grad_norm": 1.4088431596755981, "learning_rate": 0.00019999561267241146, "loss": 5.0259, "step": 23150 }, { "epoch": 0.043293779551620136, "grad_norm": 1.430076003074646, "learning_rate": 0.00019999557323270307, "loss": 5.2003, "step": 23200 }, { "epoch": 0.043387085110998626, "grad_norm": 1.7545416355133057, "learning_rate": 0.0001999955336165201, "loss": 4.9939, "step": 23250 }, { "epoch": 0.043480390670377116, "grad_norm": 1.3078407049179077, "learning_rate": 0.00019999549382386262, "loss": 5.2717, "step": 23300 }, { "epoch": 0.04357369622975561, "grad_norm": 1.6095589399337769, "learning_rate": 0.0001999954538547307, "loss": 5.2214, "step": 23350 }, { "epoch": 0.0436670017891341, "grad_norm": 1.4139171838760376, "learning_rate": 0.00019999541370912437, "loss": 5.1498, "step": 23400 }, { "epoch": 0.04376030734851259, "grad_norm": 1.6955764293670654, "learning_rate": 0.00019999537338704376, "loss": 5.2248, "step": 23450 }, { "epoch": 0.04385361290789108, "grad_norm": 1.632716178894043, "learning_rate": 0.0001999953328884889, "loss": 4.9997, "step": 23500 }, { "epoch": 0.04394691846726958, "grad_norm": 1.7285034656524658, "learning_rate": 0.00019999529221345985, "loss": 5.029, "step": 23550 }, { "epoch": 0.04404022402664807, "grad_norm": 1.7824739217758179, "learning_rate": 0.0001999952513619567, "loss": 4.9897, "step": 23600 }, { "epoch": 0.04413352958602656, "grad_norm": 1.517888069152832, "learning_rate": 0.00019999521033397957, "loss": 5.1368, "step": 23650 }, { "epoch": 0.04422683514540505, "grad_norm": 1.3025915622711182, "learning_rate": 0.00019999516912952846, "loss": 5.1239, "step": 23700 }, { "epoch": 0.044320140704783545, "grad_norm": 1.4753954410552979, "learning_rate": 0.00019999512774860348, "loss": 5.1521, "step": 23750 }, { "epoch": 0.044413446264162035, "grad_norm": 1.6234376430511475, "learning_rate": 0.00019999508619120468, "loss": 5.1371, "step": 23800 }, { "epoch": 0.044506751823540525, "grad_norm": 1.5803349018096924, "learning_rate": 0.00019999504445733217, "loss": 5.2971, "step": 23850 }, { "epoch": 0.044600057382919014, "grad_norm": 1.5795341730117798, "learning_rate": 0.00019999500254698596, "loss": 4.8945, "step": 23900 }, { "epoch": 0.04469336294229751, "grad_norm": 1.5893889665603638, "learning_rate": 0.0001999949604601662, "loss": 5.1103, "step": 23950 }, { "epoch": 0.044786668501676, "grad_norm": 1.1029318571090698, "learning_rate": 0.00019999491819687294, "loss": 5.0341, "step": 24000 }, { "epoch": 0.044786668501676, "eval_loss": 5.369950771331787, "eval_runtime": 228.2176, "eval_samples_per_second": 11.428, "eval_steps_per_second": 11.428, "eval_tts_loss": 7.0878655090813805, "step": 24000 }, { "epoch": 0.04487997406105449, "grad_norm": 1.5672223567962646, "learning_rate": 0.00019999487575710623, "loss": 5.1548, "step": 24050 }, { "epoch": 0.04497327962043299, "grad_norm": 1.3560975790023804, "learning_rate": 0.0001999948331408662, "loss": 5.1028, "step": 24100 }, { "epoch": 0.04506658517981148, "grad_norm": 1.365408182144165, "learning_rate": 0.0001999947903481528, "loss": 5.0995, "step": 24150 }, { "epoch": 0.04515989073918997, "grad_norm": 1.3482069969177246, "learning_rate": 0.00019999474737896625, "loss": 4.9473, "step": 24200 }, { "epoch": 0.04525319629856846, "grad_norm": 1.2890368700027466, "learning_rate": 0.00019999470423330654, "loss": 4.9009, "step": 24250 }, { "epoch": 0.045346501857946954, "grad_norm": 1.2973231077194214, "learning_rate": 0.00019999466091117377, "loss": 5.0809, "step": 24300 }, { "epoch": 0.045439807417325444, "grad_norm": 1.5761609077453613, "learning_rate": 0.000199994617412568, "loss": 4.9994, "step": 24350 }, { "epoch": 0.045533112976703934, "grad_norm": 1.3380223512649536, "learning_rate": 0.00019999457373748937, "loss": 4.916, "step": 24400 }, { "epoch": 0.04562641853608242, "grad_norm": 1.7922528982162476, "learning_rate": 0.00019999452988593789, "loss": 5.0799, "step": 24450 }, { "epoch": 0.04571972409546092, "grad_norm": 1.3923864364624023, "learning_rate": 0.00019999448585791369, "loss": 5.0239, "step": 24500 }, { "epoch": 0.04581302965483941, "grad_norm": 1.2720636129379272, "learning_rate": 0.0001999944416534168, "loss": 4.9941, "step": 24550 }, { "epoch": 0.0459063352142179, "grad_norm": 1.602018117904663, "learning_rate": 0.00019999439727244728, "loss": 5.1485, "step": 24600 }, { "epoch": 0.04599964077359639, "grad_norm": 1.5410170555114746, "learning_rate": 0.00019999435271500526, "loss": 5.2976, "step": 24650 }, { "epoch": 0.046092946332974886, "grad_norm": 1.6076123714447021, "learning_rate": 0.00019999430798109082, "loss": 4.9226, "step": 24700 }, { "epoch": 0.046186251892353376, "grad_norm": 1.3308855295181274, "learning_rate": 0.00019999426307070402, "loss": 5.0585, "step": 24750 }, { "epoch": 0.046279557451731866, "grad_norm": 1.0817416906356812, "learning_rate": 0.00019999421798384492, "loss": 4.9678, "step": 24800 }, { "epoch": 0.04637286301111036, "grad_norm": 1.4719314575195312, "learning_rate": 0.00019999417272051364, "loss": 4.734, "step": 24850 }, { "epoch": 0.04646616857048885, "grad_norm": 1.911425232887268, "learning_rate": 0.00019999412728071022, "loss": 5.023, "step": 24900 }, { "epoch": 0.04655947412986734, "grad_norm": 1.2537506818771362, "learning_rate": 0.0001999940816644348, "loss": 5.0473, "step": 24950 }, { "epoch": 0.04665277968924583, "grad_norm": 1.2723119258880615, "learning_rate": 0.00019999403587168737, "loss": 5.0075, "step": 25000 }, { "epoch": 0.04674608524862433, "grad_norm": 1.221137523651123, "learning_rate": 0.00019999398990246808, "loss": 5.0694, "step": 25050 }, { "epoch": 0.04683939080800282, "grad_norm": 1.6972732543945312, "learning_rate": 0.000199993943756777, "loss": 5.0661, "step": 25100 }, { "epoch": 0.04693269636738131, "grad_norm": 1.5249754190444946, "learning_rate": 0.00019999389743461418, "loss": 5.0894, "step": 25150 }, { "epoch": 0.0470260019267598, "grad_norm": 1.5337498188018799, "learning_rate": 0.00019999385093597975, "loss": 5.2557, "step": 25200 }, { "epoch": 0.047119307486138295, "grad_norm": 1.6065733432769775, "learning_rate": 0.00019999380426087378, "loss": 5.2814, "step": 25250 }, { "epoch": 0.047212613045516785, "grad_norm": 1.6499677896499634, "learning_rate": 0.0001999937574092963, "loss": 5.0703, "step": 25300 }, { "epoch": 0.047305918604895275, "grad_norm": 0.9138503670692444, "learning_rate": 0.00019999371038124746, "loss": 5.1375, "step": 25350 }, { "epoch": 0.047399224164273765, "grad_norm": 1.533087968826294, "learning_rate": 0.00019999366317672733, "loss": 5.0459, "step": 25400 }, { "epoch": 0.04749252972365226, "grad_norm": 1.8451659679412842, "learning_rate": 0.00019999361579573592, "loss": 4.9868, "step": 25450 }, { "epoch": 0.04758583528303075, "grad_norm": 1.0893833637237549, "learning_rate": 0.00019999356823827342, "loss": 4.9291, "step": 25500 }, { "epoch": 0.04767914084240924, "grad_norm": 1.099062204360962, "learning_rate": 0.00019999352050433986, "loss": 5.0681, "step": 25550 }, { "epoch": 0.04777244640178773, "grad_norm": 1.3058807849884033, "learning_rate": 0.00019999347259393532, "loss": 4.8777, "step": 25600 }, { "epoch": 0.04786575196116623, "grad_norm": 1.6822357177734375, "learning_rate": 0.0001999934245070599, "loss": 5.2018, "step": 25650 }, { "epoch": 0.04795905752054472, "grad_norm": 1.1984933614730835, "learning_rate": 0.0001999933762437137, "loss": 5.0762, "step": 25700 }, { "epoch": 0.04805236307992321, "grad_norm": 1.4918663501739502, "learning_rate": 0.00019999332780389677, "loss": 5.0357, "step": 25750 }, { "epoch": 0.048145668639301704, "grad_norm": 1.7189278602600098, "learning_rate": 0.0001999932791876092, "loss": 5.0787, "step": 25800 }, { "epoch": 0.048238974198680194, "grad_norm": 1.2158808708190918, "learning_rate": 0.00019999323039485113, "loss": 4.9543, "step": 25850 }, { "epoch": 0.048332279758058684, "grad_norm": 1.298972725868225, "learning_rate": 0.00019999318142562253, "loss": 4.8591, "step": 25900 }, { "epoch": 0.048425585317437174, "grad_norm": 1.8158042430877686, "learning_rate": 0.0001999931322799236, "loss": 4.8411, "step": 25950 }, { "epoch": 0.04851889087681567, "grad_norm": 1.2872735261917114, "learning_rate": 0.00019999308295775442, "loss": 4.9664, "step": 26000 }, { "epoch": 0.04861219643619416, "grad_norm": 1.3019319772720337, "learning_rate": 0.00019999303345911496, "loss": 5.1429, "step": 26050 }, { "epoch": 0.04870550199557265, "grad_norm": 1.3941707611083984, "learning_rate": 0.00019999298378400545, "loss": 5.0969, "step": 26100 }, { "epoch": 0.04879880755495114, "grad_norm": 1.3208707571029663, "learning_rate": 0.0001999929339324259, "loss": 5.0603, "step": 26150 }, { "epoch": 0.04889211311432964, "grad_norm": 1.7119523286819458, "learning_rate": 0.00019999288390437644, "loss": 5.0922, "step": 26200 }, { "epoch": 0.04898541867370813, "grad_norm": 1.4633029699325562, "learning_rate": 0.0001999928336998571, "loss": 5.0348, "step": 26250 }, { "epoch": 0.04907872423308662, "grad_norm": 1.695513367652893, "learning_rate": 0.00019999278331886804, "loss": 4.984, "step": 26300 }, { "epoch": 0.049172029792465106, "grad_norm": 1.6055431365966797, "learning_rate": 0.0001999927327614093, "loss": 5.0255, "step": 26350 }, { "epoch": 0.0492653353518436, "grad_norm": 1.5823806524276733, "learning_rate": 0.00019999268202748092, "loss": 4.9689, "step": 26400 }, { "epoch": 0.04935864091122209, "grad_norm": 1.099482774734497, "learning_rate": 0.00019999263111708312, "loss": 5.1056, "step": 26450 }, { "epoch": 0.04945194647060058, "grad_norm": 1.2539039850234985, "learning_rate": 0.0001999925800302159, "loss": 5.1003, "step": 26500 }, { "epoch": 0.04954525202997908, "grad_norm": 1.8513319492340088, "learning_rate": 0.0001999925287668794, "loss": 5.0544, "step": 26550 }, { "epoch": 0.04963855758935757, "grad_norm": 1.6248927116394043, "learning_rate": 0.00019999247732707364, "loss": 5.0591, "step": 26600 }, { "epoch": 0.04973186314873606, "grad_norm": 1.4035404920578003, "learning_rate": 0.00019999242571079874, "loss": 4.7997, "step": 26650 }, { "epoch": 0.04982516870811455, "grad_norm": 1.5909003019332886, "learning_rate": 0.00019999237391805482, "loss": 5.0233, "step": 26700 }, { "epoch": 0.049918474267493046, "grad_norm": 1.187412977218628, "learning_rate": 0.00019999232194884198, "loss": 4.9776, "step": 26750 }, { "epoch": 0.050011779826871536, "grad_norm": 0.9005943536758423, "learning_rate": 0.00019999226980316025, "loss": 4.9586, "step": 26800 }, { "epoch": 0.050105085386250026, "grad_norm": 1.1468946933746338, "learning_rate": 0.0001999922174810098, "loss": 5.0651, "step": 26850 }, { "epoch": 0.050198390945628515, "grad_norm": 1.2562627792358398, "learning_rate": 0.00019999216498239064, "loss": 4.8625, "step": 26900 }, { "epoch": 0.05029169650500701, "grad_norm": 1.2052425146102905, "learning_rate": 0.00019999211230730294, "loss": 4.9507, "step": 26950 }, { "epoch": 0.0503850020643855, "grad_norm": 1.3357577323913574, "learning_rate": 0.00019999205945574671, "loss": 4.9069, "step": 27000 }, { "epoch": 0.0503850020643855, "eval_loss": 5.300256252288818, "eval_runtime": 230.7705, "eval_samples_per_second": 11.301, "eval_steps_per_second": 11.301, "eval_tts_loss": 6.999222072712689, "step": 27000 }, { "epoch": 0.05047830762376399, "grad_norm": 1.0808535814285278, "learning_rate": 0.00019999200642772214, "loss": 4.9514, "step": 27050 }, { "epoch": 0.05057161318314248, "grad_norm": 1.6835181713104248, "learning_rate": 0.00019999195322322922, "loss": 5.1339, "step": 27100 }, { "epoch": 0.05066491874252098, "grad_norm": 1.3736001253128052, "learning_rate": 0.00019999189984226814, "loss": 4.8656, "step": 27150 }, { "epoch": 0.05075822430189947, "grad_norm": 1.404284954071045, "learning_rate": 0.00019999184628483892, "loss": 4.8601, "step": 27200 }, { "epoch": 0.05085152986127796, "grad_norm": 1.566685676574707, "learning_rate": 0.0001999917925509417, "loss": 4.9994, "step": 27250 }, { "epoch": 0.05094483542065645, "grad_norm": 1.4436407089233398, "learning_rate": 0.00019999173864057658, "loss": 4.9391, "step": 27300 }, { "epoch": 0.051038140980034945, "grad_norm": 1.1985220909118652, "learning_rate": 0.0001999916845537436, "loss": 4.9497, "step": 27350 }, { "epoch": 0.051131446539413435, "grad_norm": 1.4295037984848022, "learning_rate": 0.00019999163029044292, "loss": 5.0714, "step": 27400 }, { "epoch": 0.051224752098791924, "grad_norm": 1.8160730600357056, "learning_rate": 0.00019999157585067458, "loss": 4.9091, "step": 27450 }, { "epoch": 0.05131805765817042, "grad_norm": 1.26568603515625, "learning_rate": 0.00019999152123443872, "loss": 5.0946, "step": 27500 }, { "epoch": 0.05141136321754891, "grad_norm": 1.7634291648864746, "learning_rate": 0.00019999146644173542, "loss": 4.6454, "step": 27550 }, { "epoch": 0.0515046687769274, "grad_norm": 1.3475509881973267, "learning_rate": 0.00019999141147256476, "loss": 4.9264, "step": 27600 }, { "epoch": 0.05159797433630589, "grad_norm": 1.6052281856536865, "learning_rate": 0.00019999135632692687, "loss": 5.0712, "step": 27650 }, { "epoch": 0.05169127989568439, "grad_norm": 1.3022325038909912, "learning_rate": 0.00019999130100482182, "loss": 5.0866, "step": 27700 }, { "epoch": 0.05178458545506288, "grad_norm": 1.5663292407989502, "learning_rate": 0.00019999124550624972, "loss": 5.2636, "step": 27750 }, { "epoch": 0.05187789101444137, "grad_norm": 1.6265084743499756, "learning_rate": 0.00019999118983121065, "loss": 5.0156, "step": 27800 }, { "epoch": 0.05197119657381986, "grad_norm": 1.2428834438323975, "learning_rate": 0.00019999113397970475, "loss": 5.0459, "step": 27850 }, { "epoch": 0.052064502133198354, "grad_norm": 1.7661765813827515, "learning_rate": 0.0001999910779517321, "loss": 4.9641, "step": 27900 }, { "epoch": 0.052157807692576844, "grad_norm": 1.2354525327682495, "learning_rate": 0.00019999102174729277, "loss": 4.8599, "step": 27950 }, { "epoch": 0.05225111325195533, "grad_norm": 1.1708762645721436, "learning_rate": 0.0001999909653663869, "loss": 4.8535, "step": 28000 }, { "epoch": 0.05234441881133382, "grad_norm": 1.4422293901443481, "learning_rate": 0.00019999090880901455, "loss": 4.8957, "step": 28050 }, { "epoch": 0.05243772437071232, "grad_norm": 1.058984637260437, "learning_rate": 0.00019999085207517584, "loss": 5.0249, "step": 28100 }, { "epoch": 0.05253102993009081, "grad_norm": 1.315977692604065, "learning_rate": 0.00019999079516487087, "loss": 5.0057, "step": 28150 }, { "epoch": 0.0526243354894693, "grad_norm": 1.4370442628860474, "learning_rate": 0.00019999073807809974, "loss": 5.0101, "step": 28200 }, { "epoch": 0.05271764104884779, "grad_norm": 1.54827880859375, "learning_rate": 0.00019999068081486256, "loss": 5.054, "step": 28250 }, { "epoch": 0.052810946608226286, "grad_norm": 1.4660943746566772, "learning_rate": 0.00019999062337515945, "loss": 5.1078, "step": 28300 }, { "epoch": 0.052904252167604776, "grad_norm": 1.483725666999817, "learning_rate": 0.00019999056575899043, "loss": 4.935, "step": 28350 }, { "epoch": 0.052997557726983266, "grad_norm": 1.4430571794509888, "learning_rate": 0.00019999050796635568, "loss": 4.9998, "step": 28400 }, { "epoch": 0.05309086328636176, "grad_norm": 1.6983498334884644, "learning_rate": 0.00019999044999725527, "loss": 5.2524, "step": 28450 }, { "epoch": 0.05318416884574025, "grad_norm": 1.6908702850341797, "learning_rate": 0.0001999903918516893, "loss": 5.0378, "step": 28500 }, { "epoch": 0.05327747440511874, "grad_norm": 1.5972379446029663, "learning_rate": 0.00019999033352965793, "loss": 4.9017, "step": 28550 }, { "epoch": 0.05337077996449723, "grad_norm": 1.5607515573501587, "learning_rate": 0.00019999027503116116, "loss": 4.9577, "step": 28600 }, { "epoch": 0.05346408552387573, "grad_norm": 1.344556450843811, "learning_rate": 0.00019999021635619917, "loss": 5.0421, "step": 28650 }, { "epoch": 0.05355739108325422, "grad_norm": 1.3197835683822632, "learning_rate": 0.00019999015750477207, "loss": 5.1273, "step": 28700 }, { "epoch": 0.05365069664263271, "grad_norm": 1.4137849807739258, "learning_rate": 0.00019999009847687988, "loss": 4.8728, "step": 28750 }, { "epoch": 0.0537440022020112, "grad_norm": 1.2608588933944702, "learning_rate": 0.00019999003927252278, "loss": 5.0248, "step": 28800 }, { "epoch": 0.053837307761389695, "grad_norm": 1.033444881439209, "learning_rate": 0.00019998997989170087, "loss": 5.1288, "step": 28850 }, { "epoch": 0.053930613320768185, "grad_norm": 1.6949776411056519, "learning_rate": 0.00019998992033441423, "loss": 5.0369, "step": 28900 }, { "epoch": 0.054023918880146675, "grad_norm": 1.2738184928894043, "learning_rate": 0.00019998986060066297, "loss": 5.0106, "step": 28950 }, { "epoch": 0.054117224439525165, "grad_norm": 1.1779210567474365, "learning_rate": 0.0001999898006904472, "loss": 4.9383, "step": 29000 }, { "epoch": 0.05421052999890366, "grad_norm": 1.4241044521331787, "learning_rate": 0.00019998974060376706, "loss": 5.1722, "step": 29050 }, { "epoch": 0.05430383555828215, "grad_norm": 1.421637773513794, "learning_rate": 0.00019998968034062257, "loss": 5.1138, "step": 29100 }, { "epoch": 0.05439714111766064, "grad_norm": 1.2639997005462646, "learning_rate": 0.00019998961990101393, "loss": 5.1452, "step": 29150 }, { "epoch": 0.05449044667703914, "grad_norm": 1.241308569908142, "learning_rate": 0.0001999895592849412, "loss": 5.0645, "step": 29200 }, { "epoch": 0.05458375223641763, "grad_norm": 1.6818753480911255, "learning_rate": 0.00019998949849240444, "loss": 5.2461, "step": 29250 }, { "epoch": 0.05467705779579612, "grad_norm": 0.8954434990882874, "learning_rate": 0.00019998943752340387, "loss": 4.9691, "step": 29300 }, { "epoch": 0.05477036335517461, "grad_norm": 1.670721411705017, "learning_rate": 0.0001999893763779395, "loss": 5.0303, "step": 29350 }, { "epoch": 0.054863668914553104, "grad_norm": 1.2590948343276978, "learning_rate": 0.00019998931505601147, "loss": 4.961, "step": 29400 }, { "epoch": 0.054956974473931594, "grad_norm": 1.306104302406311, "learning_rate": 0.00019998925355761994, "loss": 4.889, "step": 29450 }, { "epoch": 0.055050280033310084, "grad_norm": 1.4491585493087769, "learning_rate": 0.00019998919188276494, "loss": 4.8524, "step": 29500 }, { "epoch": 0.055143585592688574, "grad_norm": 1.5823389291763306, "learning_rate": 0.0001999891300314466, "loss": 4.994, "step": 29550 }, { "epoch": 0.05523689115206707, "grad_norm": 1.5305702686309814, "learning_rate": 0.00019998906800366506, "loss": 4.8367, "step": 29600 }, { "epoch": 0.05533019671144556, "grad_norm": 1.538707971572876, "learning_rate": 0.0001999890057994204, "loss": 4.9698, "step": 29650 }, { "epoch": 0.05542350227082405, "grad_norm": 1.4228105545043945, "learning_rate": 0.00019998894341871274, "loss": 5.1362, "step": 29700 }, { "epoch": 0.05551680783020254, "grad_norm": 1.1229313611984253, "learning_rate": 0.00019998888086154217, "loss": 4.9158, "step": 29750 }, { "epoch": 0.05561011338958104, "grad_norm": 1.2107417583465576, "learning_rate": 0.00019998881812790883, "loss": 5.0441, "step": 29800 }, { "epoch": 0.05570341894895953, "grad_norm": 1.2901204824447632, "learning_rate": 0.00019998875521781283, "loss": 5.0762, "step": 29850 }, { "epoch": 0.055796724508338016, "grad_norm": 1.3414353132247925, "learning_rate": 0.00019998869213125428, "loss": 4.9442, "step": 29900 }, { "epoch": 0.055890030067716506, "grad_norm": 1.3812546730041504, "learning_rate": 0.00019998862886823327, "loss": 4.8799, "step": 29950 }, { "epoch": 0.055983335627095, "grad_norm": 1.2533543109893799, "learning_rate": 0.00019998856542874992, "loss": 4.8482, "step": 30000 }, { "epoch": 0.055983335627095, "eval_loss": 5.270205020904541, "eval_runtime": 231.0564, "eval_samples_per_second": 11.287, "eval_steps_per_second": 11.287, "eval_tts_loss": 7.119214073906258, "step": 30000 }, { "epoch": 0.05607664118647349, "grad_norm": 1.4817043542861938, "learning_rate": 0.00019998850181280432, "loss": 4.9294, "step": 30050 }, { "epoch": 0.05616994674585198, "grad_norm": 1.3725059032440186, "learning_rate": 0.00019998843802039663, "loss": 4.7408, "step": 30100 }, { "epoch": 0.05626325230523048, "grad_norm": 1.1192152500152588, "learning_rate": 0.00019998837405152696, "loss": 4.975, "step": 30150 }, { "epoch": 0.05635655786460897, "grad_norm": 1.4426796436309814, "learning_rate": 0.0001999883099061954, "loss": 5.1216, "step": 30200 }, { "epoch": 0.05644986342398746, "grad_norm": 1.060693383216858, "learning_rate": 0.00019998824558440203, "loss": 4.8414, "step": 30250 }, { "epoch": 0.05654316898336595, "grad_norm": 1.2075954675674438, "learning_rate": 0.00019998818108614704, "loss": 4.8929, "step": 30300 }, { "epoch": 0.056636474542744446, "grad_norm": 1.2428367137908936, "learning_rate": 0.00019998811641143047, "loss": 5.1015, "step": 30350 }, { "epoch": 0.056729780102122936, "grad_norm": 1.2443081140518188, "learning_rate": 0.0001999880515602525, "loss": 5.0629, "step": 30400 }, { "epoch": 0.056823085661501425, "grad_norm": 1.4452126026153564, "learning_rate": 0.0001999879865326132, "loss": 4.942, "step": 30450 }, { "epoch": 0.056916391220879915, "grad_norm": 0.8212096691131592, "learning_rate": 0.0001999879213285127, "loss": 4.9346, "step": 30500 }, { "epoch": 0.05700969678025841, "grad_norm": 1.2501893043518066, "learning_rate": 0.0001999878559479511, "loss": 4.8447, "step": 30550 }, { "epoch": 0.0571030023396369, "grad_norm": 1.0458223819732666, "learning_rate": 0.00019998779039092854, "loss": 4.929, "step": 30600 }, { "epoch": 0.05719630789901539, "grad_norm": 1.5575039386749268, "learning_rate": 0.0001999877246574451, "loss": 5.0728, "step": 30650 }, { "epoch": 0.05728961345839388, "grad_norm": 0.9992212057113647, "learning_rate": 0.00019998765874750096, "loss": 5.0842, "step": 30700 }, { "epoch": 0.05738291901777238, "grad_norm": 1.637300729751587, "learning_rate": 0.00019998759266109617, "loss": 4.9798, "step": 30750 }, { "epoch": 0.05747622457715087, "grad_norm": 1.372078537940979, "learning_rate": 0.00019998752639823088, "loss": 5.1645, "step": 30800 }, { "epoch": 0.05756953013652936, "grad_norm": 1.0558440685272217, "learning_rate": 0.0001999874599589052, "loss": 4.833, "step": 30850 }, { "epoch": 0.057662835695907855, "grad_norm": 1.3635761737823486, "learning_rate": 0.00019998739334311922, "loss": 4.7573, "step": 30900 }, { "epoch": 0.057756141255286345, "grad_norm": 1.2139973640441895, "learning_rate": 0.00019998732655087312, "loss": 5.1266, "step": 30950 }, { "epoch": 0.057849446814664834, "grad_norm": 1.2504178285598755, "learning_rate": 0.00019998725958216695, "loss": 5.1571, "step": 31000 }, { "epoch": 0.057942752374043324, "grad_norm": 1.4248679876327515, "learning_rate": 0.00019998719243700087, "loss": 5.1027, "step": 31050 }, { "epoch": 0.05803605793342182, "grad_norm": 4.532711029052734, "learning_rate": 0.000199987125115375, "loss": 5.1239, "step": 31100 }, { "epoch": 0.05812936349280031, "grad_norm": 1.717026710510254, "learning_rate": 0.00019998705761728943, "loss": 5.1622, "step": 31150 }, { "epoch": 0.0582226690521788, "grad_norm": 1.3576337099075317, "learning_rate": 0.00019998698994274435, "loss": 4.9084, "step": 31200 }, { "epoch": 0.05831597461155729, "grad_norm": 1.2413636445999146, "learning_rate": 0.00019998692209173977, "loss": 4.791, "step": 31250 }, { "epoch": 0.05840928017093579, "grad_norm": 1.3183064460754395, "learning_rate": 0.00019998685406427588, "loss": 5.1716, "step": 31300 }, { "epoch": 0.05850258573031428, "grad_norm": 1.0643527507781982, "learning_rate": 0.0001999867858603528, "loss": 4.8196, "step": 31350 }, { "epoch": 0.05859589128969277, "grad_norm": 1.1903074979782104, "learning_rate": 0.00019998671747997062, "loss": 4.8136, "step": 31400 }, { "epoch": 0.05868919684907126, "grad_norm": 1.2945053577423096, "learning_rate": 0.00019998664892312948, "loss": 4.9986, "step": 31450 }, { "epoch": 0.058782502408449754, "grad_norm": 1.5691232681274414, "learning_rate": 0.0001999865801898295, "loss": 5.0726, "step": 31500 }, { "epoch": 0.05887580796782824, "grad_norm": 1.3560112714767456, "learning_rate": 0.0001999865112800708, "loss": 4.9443, "step": 31550 }, { "epoch": 0.05896911352720673, "grad_norm": 1.1290074586868286, "learning_rate": 0.0001999864421938535, "loss": 4.9773, "step": 31600 }, { "epoch": 0.05906241908658522, "grad_norm": 1.4924343824386597, "learning_rate": 0.00019998637293117773, "loss": 5.1788, "step": 31650 }, { "epoch": 0.05915572464596372, "grad_norm": 1.2043700218200684, "learning_rate": 0.0001999863034920436, "loss": 4.8067, "step": 31700 }, { "epoch": 0.05924903020534221, "grad_norm": 1.2821531295776367, "learning_rate": 0.00019998623387645122, "loss": 5.01, "step": 31750 }, { "epoch": 0.0593423357647207, "grad_norm": 0.9705774188041687, "learning_rate": 0.00019998616408440075, "loss": 5.0704, "step": 31800 }, { "epoch": 0.059435641324099196, "grad_norm": 1.632771372795105, "learning_rate": 0.00019998609411589228, "loss": 5.1367, "step": 31850 }, { "epoch": 0.059528946883477686, "grad_norm": 1.5198719501495361, "learning_rate": 0.00019998602397092597, "loss": 4.806, "step": 31900 }, { "epoch": 0.059622252442856176, "grad_norm": 1.2326031923294067, "learning_rate": 0.00019998595364950191, "loss": 4.9858, "step": 31950 }, { "epoch": 0.059715558002234666, "grad_norm": 1.4651892185211182, "learning_rate": 0.00019998588315162018, "loss": 4.9011, "step": 32000 }, { "epoch": 0.05980886356161316, "grad_norm": 1.704825520515442, "learning_rate": 0.00019998581247728103, "loss": 5.1689, "step": 32050 }, { "epoch": 0.05990216912099165, "grad_norm": 1.4185510873794556, "learning_rate": 0.0001999857416264845, "loss": 4.984, "step": 32100 }, { "epoch": 0.05999547468037014, "grad_norm": 1.7740285396575928, "learning_rate": 0.00019998567059923073, "loss": 5.0549, "step": 32150 }, { "epoch": 0.06008878023974863, "grad_norm": 1.117332100868225, "learning_rate": 0.00019998559939551983, "loss": 4.9824, "step": 32200 }, { "epoch": 0.06018208579912713, "grad_norm": 1.4967530965805054, "learning_rate": 0.00019998552801535192, "loss": 5.0119, "step": 32250 }, { "epoch": 0.06027539135850562, "grad_norm": 1.1387840509414673, "learning_rate": 0.0001999854564587272, "loss": 5.0042, "step": 32300 }, { "epoch": 0.06036869691788411, "grad_norm": 1.2850226163864136, "learning_rate": 0.0001999853847256457, "loss": 4.7995, "step": 32350 }, { "epoch": 0.0604620024772626, "grad_norm": 1.3279659748077393, "learning_rate": 0.00019998531281610762, "loss": 4.9891, "step": 32400 }, { "epoch": 0.060555308036641095, "grad_norm": 1.3600449562072754, "learning_rate": 0.00019998524073011305, "loss": 4.9334, "step": 32450 }, { "epoch": 0.060648613596019585, "grad_norm": 1.053346872329712, "learning_rate": 0.0001999851684676621, "loss": 4.8668, "step": 32500 }, { "epoch": 0.060741919155398075, "grad_norm": 1.311558485031128, "learning_rate": 0.00019998509602875495, "loss": 4.9149, "step": 32550 }, { "epoch": 0.06083522471477657, "grad_norm": 1.0660821199417114, "learning_rate": 0.00019998502341339166, "loss": 4.8516, "step": 32600 }, { "epoch": 0.06092853027415506, "grad_norm": 1.450120210647583, "learning_rate": 0.00019998495062157243, "loss": 5.0558, "step": 32650 }, { "epoch": 0.06102183583353355, "grad_norm": 1.1899629831314087, "learning_rate": 0.00019998487765329734, "loss": 4.9774, "step": 32700 }, { "epoch": 0.06111514139291204, "grad_norm": 1.3648253679275513, "learning_rate": 0.00019998480450856654, "loss": 4.9348, "step": 32750 }, { "epoch": 0.06120844695229054, "grad_norm": 1.1677742004394531, "learning_rate": 0.00019998473118738015, "loss": 4.9243, "step": 32800 }, { "epoch": 0.06130175251166903, "grad_norm": 1.0506433248519897, "learning_rate": 0.0001999846576897383, "loss": 4.9175, "step": 32850 }, { "epoch": 0.06139505807104752, "grad_norm": 1.6663371324539185, "learning_rate": 0.00019998458401564113, "loss": 5.0618, "step": 32900 }, { "epoch": 0.06148836363042601, "grad_norm": 1.1988695859909058, "learning_rate": 0.00019998451016508875, "loss": 4.9605, "step": 32950 }, { "epoch": 0.061581669189804504, "grad_norm": 1.0045593976974487, "learning_rate": 0.0001999844361380813, "loss": 4.9856, "step": 33000 }, { "epoch": 0.061581669189804504, "eval_loss": 5.222363471984863, "eval_runtime": 230.7018, "eval_samples_per_second": 11.305, "eval_steps_per_second": 11.305, "eval_tts_loss": 7.172317200288908, "step": 33000 }, { "epoch": 0.061674974749182994, "grad_norm": 1.099878191947937, "learning_rate": 0.00019998436193461894, "loss": 4.8131, "step": 33050 }, { "epoch": 0.061768280308561484, "grad_norm": 1.5296263694763184, "learning_rate": 0.00019998428755470176, "loss": 4.8085, "step": 33100 }, { "epoch": 0.061861585867939974, "grad_norm": 1.1066418886184692, "learning_rate": 0.0001999842129983299, "loss": 4.9998, "step": 33150 }, { "epoch": 0.06195489142731847, "grad_norm": 1.0647575855255127, "learning_rate": 0.00019998413826550349, "loss": 4.9304, "step": 33200 }, { "epoch": 0.06204819698669696, "grad_norm": 1.4591343402862549, "learning_rate": 0.0001999840633562227, "loss": 4.7926, "step": 33250 }, { "epoch": 0.06214150254607545, "grad_norm": 1.5085718631744385, "learning_rate": 0.0001999839882704876, "loss": 4.9535, "step": 33300 }, { "epoch": 0.06223480810545394, "grad_norm": 1.4653303623199463, "learning_rate": 0.00019998391300829834, "loss": 5.0559, "step": 33350 }, { "epoch": 0.06232811366483244, "grad_norm": 1.4487497806549072, "learning_rate": 0.00019998383756965508, "loss": 4.7687, "step": 33400 }, { "epoch": 0.062421419224210926, "grad_norm": 1.0603256225585938, "learning_rate": 0.00019998376195455797, "loss": 5.0553, "step": 33450 }, { "epoch": 0.06251472478358942, "grad_norm": 1.2172232866287231, "learning_rate": 0.0001999836861630071, "loss": 5.016, "step": 33500 }, { "epoch": 0.06260803034296791, "grad_norm": 1.0388457775115967, "learning_rate": 0.0001999836101950026, "loss": 4.9595, "step": 33550 }, { "epoch": 0.0627013359023464, "grad_norm": 1.064448595046997, "learning_rate": 0.00019998353405054463, "loss": 4.9284, "step": 33600 }, { "epoch": 0.06279464146172489, "grad_norm": 1.3554415702819824, "learning_rate": 0.0001999834577296333, "loss": 4.9094, "step": 33650 }, { "epoch": 0.06288794702110338, "grad_norm": 1.323470950126648, "learning_rate": 0.00019998338123226878, "loss": 5.2709, "step": 33700 }, { "epoch": 0.06298125258048187, "grad_norm": 1.378735899925232, "learning_rate": 0.00019998330455845114, "loss": 5.002, "step": 33750 }, { "epoch": 0.06307455813986036, "grad_norm": 1.3667746782302856, "learning_rate": 0.00019998322770818064, "loss": 5.029, "step": 33800 }, { "epoch": 0.06316786369923887, "grad_norm": 1.2016565799713135, "learning_rate": 0.0001999831506814573, "loss": 4.7827, "step": 33850 }, { "epoch": 0.06326116925861736, "grad_norm": 1.6369880437850952, "learning_rate": 0.00019998307347828128, "loss": 4.9345, "step": 33900 }, { "epoch": 0.06335447481799585, "grad_norm": 1.6404528617858887, "learning_rate": 0.00019998299609865272, "loss": 4.9889, "step": 33950 }, { "epoch": 0.06344778037737434, "grad_norm": 1.322618007659912, "learning_rate": 0.00019998291854257175, "loss": 4.9924, "step": 34000 }, { "epoch": 0.06354108593675283, "grad_norm": 1.2779415845870972, "learning_rate": 0.00019998284081003856, "loss": 4.7519, "step": 34050 }, { "epoch": 0.06363439149613132, "grad_norm": 1.0011979341506958, "learning_rate": 0.00019998276290105325, "loss": 5.0784, "step": 34100 }, { "epoch": 0.0637276970555098, "grad_norm": 1.2990458011627197, "learning_rate": 0.00019998268481561594, "loss": 4.85, "step": 34150 }, { "epoch": 0.0638210026148883, "grad_norm": 0.8773101568222046, "learning_rate": 0.00019998260655372678, "loss": 4.9803, "step": 34200 }, { "epoch": 0.0639143081742668, "grad_norm": 0.890830934047699, "learning_rate": 0.00019998252811538595, "loss": 5.0072, "step": 34250 }, { "epoch": 0.06400761373364529, "grad_norm": 1.118335485458374, "learning_rate": 0.00019998244950059348, "loss": 4.9289, "step": 34300 }, { "epoch": 0.06410091929302378, "grad_norm": 1.284605860710144, "learning_rate": 0.00019998237070934964, "loss": 5.1104, "step": 34350 }, { "epoch": 0.06419422485240227, "grad_norm": 1.3694264888763428, "learning_rate": 0.00019998229174165446, "loss": 5.0689, "step": 34400 }, { "epoch": 0.06428753041178076, "grad_norm": 1.252570390701294, "learning_rate": 0.00019998221259750814, "loss": 4.9514, "step": 34450 }, { "epoch": 0.06438083597115925, "grad_norm": 1.339966893196106, "learning_rate": 0.00019998213327691085, "loss": 5.0985, "step": 34500 }, { "epoch": 0.06447414153053774, "grad_norm": 1.0073368549346924, "learning_rate": 0.00019998205377986267, "loss": 4.7595, "step": 34550 }, { "epoch": 0.06456744708991624, "grad_norm": 1.4048619270324707, "learning_rate": 0.00019998197410636372, "loss": 4.9045, "step": 34600 }, { "epoch": 0.06466075264929473, "grad_norm": 0.9788991808891296, "learning_rate": 0.00019998189425641423, "loss": 4.4886, "step": 34650 }, { "epoch": 0.06475405820867322, "grad_norm": 1.4313668012619019, "learning_rate": 0.00019998181423001423, "loss": 4.7508, "step": 34700 }, { "epoch": 0.06484736376805171, "grad_norm": 1.3244096040725708, "learning_rate": 0.00019998173402716396, "loss": 5.0475, "step": 34750 }, { "epoch": 0.0649406693274302, "grad_norm": 1.0660874843597412, "learning_rate": 0.00019998165364786352, "loss": 4.9674, "step": 34800 }, { "epoch": 0.06503397488680869, "grad_norm": 1.2351161241531372, "learning_rate": 0.00019998157309211305, "loss": 5.0454, "step": 34850 }, { "epoch": 0.06512728044618718, "grad_norm": 1.4965345859527588, "learning_rate": 0.00019998149235991267, "loss": 4.8981, "step": 34900 }, { "epoch": 0.06522058600556567, "grad_norm": 1.113448143005371, "learning_rate": 0.00019998141145126258, "loss": 5.1095, "step": 34950 }, { "epoch": 0.06531389156494417, "grad_norm": 1.1176252365112305, "learning_rate": 0.0001999813303661629, "loss": 5.0817, "step": 35000 }, { "epoch": 0.06540719712432266, "grad_norm": 1.1764382123947144, "learning_rate": 0.00019998124910461376, "loss": 4.8815, "step": 35050 }, { "epoch": 0.06550050268370115, "grad_norm": 1.3247041702270508, "learning_rate": 0.00019998116766661527, "loss": 4.7398, "step": 35100 }, { "epoch": 0.06559380824307964, "grad_norm": 1.155849575996399, "learning_rate": 0.00019998108605216762, "loss": 4.7499, "step": 35150 }, { "epoch": 0.06568711380245813, "grad_norm": 1.5121906995773315, "learning_rate": 0.000199981004261271, "loss": 4.8656, "step": 35200 }, { "epoch": 0.06578041936183662, "grad_norm": 1.2283687591552734, "learning_rate": 0.00019998092229392544, "loss": 5.0256, "step": 35250 }, { "epoch": 0.06587372492121511, "grad_norm": 1.1128344535827637, "learning_rate": 0.00019998084015013115, "loss": 5.0056, "step": 35300 }, { "epoch": 0.06596703048059362, "grad_norm": 0.7951078414916992, "learning_rate": 0.0001999807578298883, "loss": 4.8792, "step": 35350 }, { "epoch": 0.0660603360399721, "grad_norm": 1.197862148284912, "learning_rate": 0.000199980675333197, "loss": 4.7815, "step": 35400 }, { "epoch": 0.0661536415993506, "grad_norm": 1.3143230676651, "learning_rate": 0.0001999805926600574, "loss": 4.8473, "step": 35450 }, { "epoch": 0.06624694715872909, "grad_norm": 1.100462555885315, "learning_rate": 0.00019998050981046963, "loss": 5.1148, "step": 35500 }, { "epoch": 0.06634025271810758, "grad_norm": 0.8429204225540161, "learning_rate": 0.00019998042678443385, "loss": 5.0559, "step": 35550 }, { "epoch": 0.06643355827748607, "grad_norm": 1.61745023727417, "learning_rate": 0.0001999803435819502, "loss": 4.8, "step": 35600 }, { "epoch": 0.06652686383686456, "grad_norm": 1.1721575260162354, "learning_rate": 0.00019998026020301887, "loss": 4.746, "step": 35650 }, { "epoch": 0.06662016939624305, "grad_norm": 1.6032483577728271, "learning_rate": 0.00019998017664763995, "loss": 5.0657, "step": 35700 }, { "epoch": 0.06671347495562155, "grad_norm": 1.3949735164642334, "learning_rate": 0.0001999800929158136, "loss": 4.8643, "step": 35750 }, { "epoch": 0.06680678051500004, "grad_norm": 1.1598511934280396, "learning_rate": 0.00019998000900754001, "loss": 4.8064, "step": 35800 }, { "epoch": 0.06690008607437853, "grad_norm": 1.3976051807403564, "learning_rate": 0.00019997992492281927, "loss": 4.9548, "step": 35850 }, { "epoch": 0.06699339163375702, "grad_norm": 1.2537288665771484, "learning_rate": 0.0001999798406616516, "loss": 5.0992, "step": 35900 }, { "epoch": 0.06708669719313551, "grad_norm": 0.9580097198486328, "learning_rate": 0.00019997975622403704, "loss": 4.7681, "step": 35950 }, { "epoch": 0.067180002752514, "grad_norm": 1.2984302043914795, "learning_rate": 0.0001999796716099758, "loss": 4.7869, "step": 36000 }, { "epoch": 0.067180002752514, "eval_loss": 5.1954779624938965, "eval_runtime": 233.7952, "eval_samples_per_second": 11.155, "eval_steps_per_second": 11.155, "eval_tts_loss": 7.21724026617337, "step": 36000 }, { "epoch": 0.06727330831189249, "grad_norm": 1.1834443807601929, "learning_rate": 0.00019997958681946808, "loss": 4.9615, "step": 36050 }, { "epoch": 0.06736661387127099, "grad_norm": 1.4487770795822144, "learning_rate": 0.00019997950185251398, "loss": 4.9015, "step": 36100 }, { "epoch": 0.06745991943064948, "grad_norm": 1.003225326538086, "learning_rate": 0.00019997941670911362, "loss": 5.1782, "step": 36150 }, { "epoch": 0.06755322499002797, "grad_norm": 1.2664480209350586, "learning_rate": 0.0001999793313892672, "loss": 4.8188, "step": 36200 }, { "epoch": 0.06764653054940646, "grad_norm": 0.9945356845855713, "learning_rate": 0.00019997924589297482, "loss": 5.017, "step": 36250 }, { "epoch": 0.06773983610878495, "grad_norm": 1.232276201248169, "learning_rate": 0.00019997916022023674, "loss": 4.8427, "step": 36300 }, { "epoch": 0.06783314166816344, "grad_norm": 1.5773524045944214, "learning_rate": 0.00019997907437105297, "loss": 4.8048, "step": 36350 }, { "epoch": 0.06792644722754193, "grad_norm": 1.3550626039505005, "learning_rate": 0.00019997898834542372, "loss": 4.9573, "step": 36400 }, { "epoch": 0.06801975278692042, "grad_norm": 1.5464149713516235, "learning_rate": 0.0001999789021433492, "loss": 4.9219, "step": 36450 }, { "epoch": 0.06811305834629892, "grad_norm": 1.3397859334945679, "learning_rate": 0.00019997881576482945, "loss": 5.1099, "step": 36500 }, { "epoch": 0.06820636390567741, "grad_norm": 1.3835015296936035, "learning_rate": 0.0001999787292098647, "loss": 4.9419, "step": 36550 }, { "epoch": 0.0682996694650559, "grad_norm": 0.9244956374168396, "learning_rate": 0.0001999786424784551, "loss": 5.0481, "step": 36600 }, { "epoch": 0.0683929750244344, "grad_norm": 1.0417383909225464, "learning_rate": 0.00019997855557060077, "loss": 4.8458, "step": 36650 }, { "epoch": 0.06848628058381288, "grad_norm": 1.2598869800567627, "learning_rate": 0.0001999784684863019, "loss": 4.8585, "step": 36700 }, { "epoch": 0.06857958614319137, "grad_norm": 1.0697765350341797, "learning_rate": 0.00019997838122555862, "loss": 4.7396, "step": 36750 }, { "epoch": 0.06867289170256986, "grad_norm": 1.1050491333007812, "learning_rate": 0.0001999782937883711, "loss": 4.8543, "step": 36800 }, { "epoch": 0.06876619726194837, "grad_norm": 1.0602041482925415, "learning_rate": 0.00019997820617473946, "loss": 4.8054, "step": 36850 }, { "epoch": 0.06885950282132686, "grad_norm": 1.0901793241500854, "learning_rate": 0.0001999781183846639, "loss": 4.943, "step": 36900 }, { "epoch": 0.06895280838070535, "grad_norm": 1.1608269214630127, "learning_rate": 0.00019997803041814453, "loss": 4.8669, "step": 36950 }, { "epoch": 0.06904611394008384, "grad_norm": 1.428856611251831, "learning_rate": 0.00019997794227518152, "loss": 4.9257, "step": 37000 }, { "epoch": 0.06913941949946233, "grad_norm": 1.2141627073287964, "learning_rate": 0.00019997785395577507, "loss": 4.8884, "step": 37050 }, { "epoch": 0.06923272505884082, "grad_norm": 0.917578935623169, "learning_rate": 0.00019997776545992526, "loss": 4.7185, "step": 37100 }, { "epoch": 0.0693260306182193, "grad_norm": 1.1432592868804932, "learning_rate": 0.00019997767678763234, "loss": 5.1003, "step": 37150 }, { "epoch": 0.0694193361775978, "grad_norm": 1.5270577669143677, "learning_rate": 0.00019997758793889635, "loss": 4.7651, "step": 37200 }, { "epoch": 0.0695126417369763, "grad_norm": 0.8794492483139038, "learning_rate": 0.00019997749891371752, "loss": 4.9522, "step": 37250 }, { "epoch": 0.06960594729635479, "grad_norm": 1.5430872440338135, "learning_rate": 0.000199977409712096, "loss": 4.8259, "step": 37300 }, { "epoch": 0.06969925285573328, "grad_norm": 1.2974966764450073, "learning_rate": 0.000199977320334032, "loss": 5.0415, "step": 37350 }, { "epoch": 0.06979255841511177, "grad_norm": 1.3754868507385254, "learning_rate": 0.00019997723077952556, "loss": 4.9111, "step": 37400 }, { "epoch": 0.06988586397449026, "grad_norm": 1.1165885925292969, "learning_rate": 0.00019997714104857693, "loss": 5.0136, "step": 37450 }, { "epoch": 0.06997916953386875, "grad_norm": 1.127763032913208, "learning_rate": 0.0001999770511411862, "loss": 4.6376, "step": 37500 }, { "epoch": 0.07007247509324724, "grad_norm": 1.1348410844802856, "learning_rate": 0.00019997696105735358, "loss": 4.7574, "step": 37550 }, { "epoch": 0.07016578065262573, "grad_norm": 1.012832522392273, "learning_rate": 0.00019997687079707926, "loss": 4.9673, "step": 37600 }, { "epoch": 0.07025908621200423, "grad_norm": 0.8529366254806519, "learning_rate": 0.0001999767803603633, "loss": 4.9508, "step": 37650 }, { "epoch": 0.07035239177138272, "grad_norm": 1.2480065822601318, "learning_rate": 0.00019997668974720595, "loss": 4.8772, "step": 37700 }, { "epoch": 0.07044569733076121, "grad_norm": 1.318292498588562, "learning_rate": 0.00019997659895760732, "loss": 4.8596, "step": 37750 }, { "epoch": 0.0705390028901397, "grad_norm": 1.3873937129974365, "learning_rate": 0.0001999765079915676, "loss": 4.8429, "step": 37800 }, { "epoch": 0.07063230844951819, "grad_norm": 1.1860456466674805, "learning_rate": 0.0001999764168490869, "loss": 4.8605, "step": 37850 }, { "epoch": 0.07072561400889668, "grad_norm": 1.0608043670654297, "learning_rate": 0.00019997632553016544, "loss": 5.0347, "step": 37900 }, { "epoch": 0.07081891956827517, "grad_norm": 1.168198585510254, "learning_rate": 0.00019997623403480335, "loss": 4.7621, "step": 37950 }, { "epoch": 0.07091222512765367, "grad_norm": 0.9395182132720947, "learning_rate": 0.0001999761423630008, "loss": 4.9286, "step": 38000 }, { "epoch": 0.07100553068703216, "grad_norm": 1.2564831972122192, "learning_rate": 0.00019997605051475794, "loss": 4.9218, "step": 38050 }, { "epoch": 0.07109883624641065, "grad_norm": 1.497332215309143, "learning_rate": 0.00019997595849007496, "loss": 5.0628, "step": 38100 }, { "epoch": 0.07119214180578914, "grad_norm": 1.2820568084716797, "learning_rate": 0.00019997586628895197, "loss": 4.8333, "step": 38150 }, { "epoch": 0.07128544736516763, "grad_norm": 1.5082170963287354, "learning_rate": 0.0001999757739113892, "loss": 5.0072, "step": 38200 }, { "epoch": 0.07137875292454612, "grad_norm": 1.0686591863632202, "learning_rate": 0.00019997568135738677, "loss": 4.8872, "step": 38250 }, { "epoch": 0.07147205848392461, "grad_norm": 1.14346444606781, "learning_rate": 0.00019997558862694484, "loss": 4.8167, "step": 38300 }, { "epoch": 0.0715653640433031, "grad_norm": 0.6724060773849487, "learning_rate": 0.00019997549572006365, "loss": 4.8189, "step": 38350 }, { "epoch": 0.07165866960268161, "grad_norm": 1.5035301446914673, "learning_rate": 0.00019997540263674322, "loss": 4.9308, "step": 38400 }, { "epoch": 0.0717519751620601, "grad_norm": 1.1256290674209595, "learning_rate": 0.00019997530937698382, "loss": 4.8645, "step": 38450 }, { "epoch": 0.07184528072143859, "grad_norm": 0.9030829668045044, "learning_rate": 0.00019997521594078562, "loss": 4.9792, "step": 38500 }, { "epoch": 0.07193858628081708, "grad_norm": 1.1105825901031494, "learning_rate": 0.00019997512232814875, "loss": 4.9466, "step": 38550 }, { "epoch": 0.07203189184019557, "grad_norm": 1.272040843963623, "learning_rate": 0.00019997502853907336, "loss": 4.9014, "step": 38600 }, { "epoch": 0.07212519739957406, "grad_norm": 1.1471481323242188, "learning_rate": 0.00019997493457355963, "loss": 4.9466, "step": 38650 }, { "epoch": 0.07221850295895255, "grad_norm": 1.134016990661621, "learning_rate": 0.00019997484043160775, "loss": 4.8908, "step": 38700 }, { "epoch": 0.07231180851833105, "grad_norm": 1.214316964149475, "learning_rate": 0.00019997474611321787, "loss": 4.9729, "step": 38750 }, { "epoch": 0.07240511407770954, "grad_norm": 0.9872856140136719, "learning_rate": 0.00019997465161839014, "loss": 4.9633, "step": 38800 }, { "epoch": 0.07249841963708803, "grad_norm": 1.3392361402511597, "learning_rate": 0.00019997455694712474, "loss": 5.0015, "step": 38850 }, { "epoch": 0.07259172519646652, "grad_norm": 1.0813117027282715, "learning_rate": 0.00019997446209942186, "loss": 5.0432, "step": 38900 }, { "epoch": 0.07268503075584501, "grad_norm": 1.2037447690963745, "learning_rate": 0.0001999743670752816, "loss": 4.8759, "step": 38950 }, { "epoch": 0.0727783363152235, "grad_norm": 1.0193177461624146, "learning_rate": 0.00019997427187470423, "loss": 4.9456, "step": 39000 }, { "epoch": 0.0727783363152235, "eval_loss": 5.16133975982666, "eval_runtime": 230.6318, "eval_samples_per_second": 11.308, "eval_steps_per_second": 11.308, "eval_tts_loss": 7.213367073268178, "step": 39000 }, { "epoch": 0.07287164187460199, "grad_norm": 1.2097446918487549, "learning_rate": 0.00019997417649768982, "loss": 5.0498, "step": 39050 }, { "epoch": 0.07296494743398048, "grad_norm": 1.0760691165924072, "learning_rate": 0.00019997408094423862, "loss": 4.6899, "step": 39100 }, { "epoch": 0.07305825299335898, "grad_norm": 1.1207256317138672, "learning_rate": 0.0001999739852143507, "loss": 4.9026, "step": 39150 }, { "epoch": 0.07315155855273747, "grad_norm": 0.801072359085083, "learning_rate": 0.00019997388930802635, "loss": 4.8494, "step": 39200 }, { "epoch": 0.07324486411211596, "grad_norm": 1.1047637462615967, "learning_rate": 0.00019997379322526563, "loss": 4.7812, "step": 39250 }, { "epoch": 0.07333816967149445, "grad_norm": 1.152491807937622, "learning_rate": 0.00019997369696606878, "loss": 4.9095, "step": 39300 }, { "epoch": 0.07343147523087294, "grad_norm": 1.1126306056976318, "learning_rate": 0.00019997360053043592, "loss": 4.9798, "step": 39350 }, { "epoch": 0.07352478079025143, "grad_norm": 1.156316876411438, "learning_rate": 0.00019997350391836726, "loss": 4.9773, "step": 39400 }, { "epoch": 0.07361808634962992, "grad_norm": 1.0668503046035767, "learning_rate": 0.00019997340712986296, "loss": 5.1012, "step": 39450 }, { "epoch": 0.07371139190900843, "grad_norm": 1.3585624694824219, "learning_rate": 0.0001999733101649232, "loss": 4.6893, "step": 39500 }, { "epoch": 0.07380469746838692, "grad_norm": 1.5277858972549438, "learning_rate": 0.00019997321302354812, "loss": 4.8931, "step": 39550 }, { "epoch": 0.0738980030277654, "grad_norm": 1.2020436525344849, "learning_rate": 0.00019997311570573792, "loss": 5.1182, "step": 39600 }, { "epoch": 0.0739913085871439, "grad_norm": 1.1372538805007935, "learning_rate": 0.00019997301821149274, "loss": 4.9238, "step": 39650 }, { "epoch": 0.07408461414652238, "grad_norm": 0.7125515341758728, "learning_rate": 0.0001999729205408128, "loss": 4.8202, "step": 39700 }, { "epoch": 0.07417791970590087, "grad_norm": 0.9166264533996582, "learning_rate": 0.00019997282269369824, "loss": 4.8832, "step": 39750 }, { "epoch": 0.07427122526527936, "grad_norm": 1.099700927734375, "learning_rate": 0.00019997272467014922, "loss": 4.9073, "step": 39800 }, { "epoch": 0.07436453082465785, "grad_norm": 1.3260917663574219, "learning_rate": 0.00019997262647016597, "loss": 4.8271, "step": 39850 }, { "epoch": 0.07445783638403636, "grad_norm": 1.2185454368591309, "learning_rate": 0.0001999725280937486, "loss": 4.8525, "step": 39900 }, { "epoch": 0.07455114194341485, "grad_norm": 1.0137861967086792, "learning_rate": 0.0001999724295408973, "loss": 4.749, "step": 39950 }, { "epoch": 0.07464444750279334, "grad_norm": 0.8680387139320374, "learning_rate": 0.00019997233081161226, "loss": 5.0129, "step": 40000 }, { "epoch": 0.07473775306217183, "grad_norm": 1.314736008644104, "learning_rate": 0.00019997223190589366, "loss": 5.0179, "step": 40050 }, { "epoch": 0.07483105862155032, "grad_norm": 1.209912657737732, "learning_rate": 0.00019997213282374165, "loss": 5.0474, "step": 40100 }, { "epoch": 0.0749243641809288, "grad_norm": 0.9931349158287048, "learning_rate": 0.0001999720335651564, "loss": 4.8296, "step": 40150 }, { "epoch": 0.0750176697403073, "grad_norm": 0.9877680540084839, "learning_rate": 0.00019997193413013814, "loss": 4.7287, "step": 40200 }, { "epoch": 0.0751109752996858, "grad_norm": 1.2788828611373901, "learning_rate": 0.00019997183451868697, "loss": 4.7418, "step": 40250 }, { "epoch": 0.07520428085906429, "grad_norm": 1.1528019905090332, "learning_rate": 0.00019997173473080312, "loss": 4.7367, "step": 40300 }, { "epoch": 0.07529758641844278, "grad_norm": 1.3090732097625732, "learning_rate": 0.00019997163476648676, "loss": 4.8197, "step": 40350 }, { "epoch": 0.07539089197782127, "grad_norm": 1.2664506435394287, "learning_rate": 0.00019997153462573804, "loss": 4.6739, "step": 40400 }, { "epoch": 0.07548419753719976, "grad_norm": 1.1950277090072632, "learning_rate": 0.00019997143430855715, "loss": 5.007, "step": 40450 }, { "epoch": 0.07557750309657825, "grad_norm": 0.9949437975883484, "learning_rate": 0.00019997133381494426, "loss": 4.9069, "step": 40500 }, { "epoch": 0.07567080865595674, "grad_norm": 1.1204107999801636, "learning_rate": 0.00019997123314489956, "loss": 4.8598, "step": 40550 }, { "epoch": 0.07576411421533523, "grad_norm": 1.2911077737808228, "learning_rate": 0.00019997113229842323, "loss": 4.9724, "step": 40600 }, { "epoch": 0.07585741977471373, "grad_norm": 0.9205583930015564, "learning_rate": 0.00019997103127551548, "loss": 4.8516, "step": 40650 }, { "epoch": 0.07595072533409222, "grad_norm": 1.0593533515930176, "learning_rate": 0.0001999709300761764, "loss": 4.8309, "step": 40700 }, { "epoch": 0.07604403089347071, "grad_norm": 1.1807661056518555, "learning_rate": 0.00019997082870040625, "loss": 4.9056, "step": 40750 }, { "epoch": 0.0761373364528492, "grad_norm": 1.105985164642334, "learning_rate": 0.00019997072714820514, "loss": 4.8024, "step": 40800 }, { "epoch": 0.07623064201222769, "grad_norm": 1.2069274187088013, "learning_rate": 0.00019997062541957333, "loss": 4.9473, "step": 40850 }, { "epoch": 0.07632394757160618, "grad_norm": 1.3253543376922607, "learning_rate": 0.00019997052351451093, "loss": 4.9783, "step": 40900 }, { "epoch": 0.07641725313098467, "grad_norm": 1.22372567653656, "learning_rate": 0.00019997042143301815, "loss": 4.9517, "step": 40950 }, { "epoch": 0.07651055869036316, "grad_norm": 1.091848611831665, "learning_rate": 0.00019997031917509515, "loss": 4.7118, "step": 41000 }, { "epoch": 0.07660386424974167, "grad_norm": 1.0672491788864136, "learning_rate": 0.00019997021674074213, "loss": 4.9462, "step": 41050 }, { "epoch": 0.07669716980912016, "grad_norm": 1.0972808599472046, "learning_rate": 0.0001999701141299593, "loss": 4.8651, "step": 41100 }, { "epoch": 0.07679047536849865, "grad_norm": 1.0306609869003296, "learning_rate": 0.0001999700113427468, "loss": 4.9256, "step": 41150 }, { "epoch": 0.07688378092787714, "grad_norm": 1.1696428060531616, "learning_rate": 0.0001999699083791048, "loss": 4.7898, "step": 41200 }, { "epoch": 0.07697708648725562, "grad_norm": 1.2328746318817139, "learning_rate": 0.00019996980523903353, "loss": 4.942, "step": 41250 }, { "epoch": 0.07707039204663411, "grad_norm": 1.0950030088424683, "learning_rate": 0.00019996970192253313, "loss": 4.8404, "step": 41300 }, { "epoch": 0.0771636976060126, "grad_norm": 1.0388070344924927, "learning_rate": 0.00019996959842960379, "loss": 5.0406, "step": 41350 }, { "epoch": 0.07725700316539111, "grad_norm": 1.3518260717391968, "learning_rate": 0.0001999694947602457, "loss": 4.7539, "step": 41400 }, { "epoch": 0.0773503087247696, "grad_norm": 0.8718711733818054, "learning_rate": 0.00019996939091445904, "loss": 4.8804, "step": 41450 }, { "epoch": 0.07744361428414809, "grad_norm": 1.1106172800064087, "learning_rate": 0.00019996928689224402, "loss": 4.9871, "step": 41500 }, { "epoch": 0.07753691984352658, "grad_norm": 1.3443810939788818, "learning_rate": 0.00019996918269360076, "loss": 4.7499, "step": 41550 }, { "epoch": 0.07763022540290507, "grad_norm": 0.968607485294342, "learning_rate": 0.00019996907831852951, "loss": 4.7705, "step": 41600 }, { "epoch": 0.07772353096228356, "grad_norm": 1.1009609699249268, "learning_rate": 0.0001999689737670304, "loss": 4.8043, "step": 41650 }, { "epoch": 0.07781683652166205, "grad_norm": 0.9938008785247803, "learning_rate": 0.00019996886903910367, "loss": 4.9927, "step": 41700 }, { "epoch": 0.07791014208104054, "grad_norm": 1.0750339031219482, "learning_rate": 0.00019996876413474947, "loss": 4.9449, "step": 41750 }, { "epoch": 0.07800344764041904, "grad_norm": 1.2172019481658936, "learning_rate": 0.000199968659053968, "loss": 5.0049, "step": 41800 }, { "epoch": 0.07809675319979753, "grad_norm": 1.288533329963684, "learning_rate": 0.0001999685537967594, "loss": 4.6487, "step": 41850 }, { "epoch": 0.07819005875917602, "grad_norm": 1.087883710861206, "learning_rate": 0.00019996844836312394, "loss": 4.8504, "step": 41900 }, { "epoch": 0.07828336431855451, "grad_norm": 1.2223601341247559, "learning_rate": 0.00019996834275306174, "loss": 4.8588, "step": 41950 }, { "epoch": 0.078376669877933, "grad_norm": 0.8469075560569763, "learning_rate": 0.00019996823696657299, "loss": 4.9408, "step": 42000 }, { "epoch": 0.078376669877933, "eval_loss": 5.132350444793701, "eval_runtime": 230.8355, "eval_samples_per_second": 11.298, "eval_steps_per_second": 11.298, "eval_tts_loss": 7.264853509248697, "step": 42000 }, { "epoch": 0.07846997543731149, "grad_norm": 1.1458607912063599, "learning_rate": 0.0001999681310036579, "loss": 4.9473, "step": 42050 }, { "epoch": 0.07856328099668998, "grad_norm": 0.9783209562301636, "learning_rate": 0.00019996802486431665, "loss": 4.9335, "step": 42100 }, { "epoch": 0.07865658655606848, "grad_norm": 1.3118724822998047, "learning_rate": 0.00019996791854854943, "loss": 4.9134, "step": 42150 }, { "epoch": 0.07874989211544697, "grad_norm": 1.4106804132461548, "learning_rate": 0.0001999678120563564, "loss": 4.9419, "step": 42200 }, { "epoch": 0.07884319767482546, "grad_norm": 1.3338770866394043, "learning_rate": 0.0001999677053877378, "loss": 4.9377, "step": 42250 }, { "epoch": 0.07893650323420395, "grad_norm": 1.1981934309005737, "learning_rate": 0.00019996759854269377, "loss": 4.9804, "step": 42300 }, { "epoch": 0.07902980879358244, "grad_norm": 1.1797490119934082, "learning_rate": 0.00019996749152122455, "loss": 4.86, "step": 42350 }, { "epoch": 0.07912311435296093, "grad_norm": 0.7988554239273071, "learning_rate": 0.0001999673843233303, "loss": 5.0499, "step": 42400 }, { "epoch": 0.07921641991233942, "grad_norm": 0.9866506457328796, "learning_rate": 0.00019996727694901117, "loss": 4.8936, "step": 42450 }, { "epoch": 0.07930972547171791, "grad_norm": 1.507678747177124, "learning_rate": 0.0001999671693982674, "loss": 4.7716, "step": 42500 }, { "epoch": 0.07940303103109642, "grad_norm": 1.2090741395950317, "learning_rate": 0.00019996706167109917, "loss": 4.7974, "step": 42550 }, { "epoch": 0.0794963365904749, "grad_norm": 0.9674336314201355, "learning_rate": 0.00019996695376750667, "loss": 4.8457, "step": 42600 }, { "epoch": 0.0795896421498534, "grad_norm": 0.8843348622322083, "learning_rate": 0.00019996684568749008, "loss": 4.8981, "step": 42650 }, { "epoch": 0.07968294770923189, "grad_norm": 1.0419225692749023, "learning_rate": 0.00019996673743104956, "loss": 4.6177, "step": 42700 }, { "epoch": 0.07977625326861038, "grad_norm": 0.8469592928886414, "learning_rate": 0.0001999666289981854, "loss": 4.8767, "step": 42750 }, { "epoch": 0.07986955882798887, "grad_norm": 0.8583770394325256, "learning_rate": 0.00019996652038889768, "loss": 4.9546, "step": 42800 }, { "epoch": 0.07996286438736736, "grad_norm": 0.9213618636131287, "learning_rate": 0.00019996641160318666, "loss": 4.5851, "step": 42850 }, { "epoch": 0.08005616994674586, "grad_norm": 1.247279405593872, "learning_rate": 0.00019996630264105253, "loss": 4.8712, "step": 42900 }, { "epoch": 0.08014947550612435, "grad_norm": 1.3103673458099365, "learning_rate": 0.00019996619350249546, "loss": 4.9501, "step": 42950 }, { "epoch": 0.08024278106550284, "grad_norm": 0.9502874612808228, "learning_rate": 0.00019996608418751564, "loss": 4.9463, "step": 43000 }, { "epoch": 0.08033608662488133, "grad_norm": 1.1887316703796387, "learning_rate": 0.0001999659746961133, "loss": 4.7633, "step": 43050 }, { "epoch": 0.08042939218425982, "grad_norm": 1.132875680923462, "learning_rate": 0.00019996586502828856, "loss": 4.7873, "step": 43100 }, { "epoch": 0.08052269774363831, "grad_norm": 1.1734470129013062, "learning_rate": 0.00019996575518404166, "loss": 4.9746, "step": 43150 }, { "epoch": 0.0806160033030168, "grad_norm": 1.2493457794189453, "learning_rate": 0.0001999656451633728, "loss": 4.7152, "step": 43200 }, { "epoch": 0.08070930886239529, "grad_norm": 1.2277987003326416, "learning_rate": 0.00019996553496628216, "loss": 4.8031, "step": 43250 }, { "epoch": 0.08080261442177379, "grad_norm": 0.9987390041351318, "learning_rate": 0.00019996542459276996, "loss": 4.6355, "step": 43300 }, { "epoch": 0.08089591998115228, "grad_norm": 0.8293343186378479, "learning_rate": 0.0001999653140428364, "loss": 4.845, "step": 43350 }, { "epoch": 0.08098922554053077, "grad_norm": 1.0015592575073242, "learning_rate": 0.0001999652033164816, "loss": 4.8523, "step": 43400 }, { "epoch": 0.08108253109990926, "grad_norm": 1.1842169761657715, "learning_rate": 0.0001999650924137058, "loss": 5.0227, "step": 43450 }, { "epoch": 0.08117583665928775, "grad_norm": 1.0604639053344727, "learning_rate": 0.00019996498133450924, "loss": 4.9266, "step": 43500 }, { "epoch": 0.08126914221866624, "grad_norm": 1.1228868961334229, "learning_rate": 0.00019996487007889206, "loss": 4.9329, "step": 43550 }, { "epoch": 0.08136244777804473, "grad_norm": 1.149283766746521, "learning_rate": 0.0001999647586468545, "loss": 4.6841, "step": 43600 }, { "epoch": 0.08145575333742322, "grad_norm": 1.1383576393127441, "learning_rate": 0.00019996464703839667, "loss": 4.9698, "step": 43650 }, { "epoch": 0.08154905889680172, "grad_norm": 1.0777636766433716, "learning_rate": 0.00019996453525351887, "loss": 4.7305, "step": 43700 }, { "epoch": 0.08164236445618021, "grad_norm": 1.4025020599365234, "learning_rate": 0.00019996442329222125, "loss": 4.9614, "step": 43750 }, { "epoch": 0.0817356700155587, "grad_norm": 1.1779139041900635, "learning_rate": 0.00019996431115450403, "loss": 4.8422, "step": 43800 }, { "epoch": 0.0818289755749372, "grad_norm": 0.9133360981941223, "learning_rate": 0.00019996419884036733, "loss": 5.0592, "step": 43850 }, { "epoch": 0.08192228113431568, "grad_norm": 1.0266224145889282, "learning_rate": 0.00019996408634981148, "loss": 4.9319, "step": 43900 }, { "epoch": 0.08201558669369417, "grad_norm": 0.882277250289917, "learning_rate": 0.00019996397368283656, "loss": 4.8332, "step": 43950 }, { "epoch": 0.08210889225307266, "grad_norm": 1.1271204948425293, "learning_rate": 0.00019996386083944283, "loss": 5.1547, "step": 44000 }, { "epoch": 0.08220219781245117, "grad_norm": 1.0688495635986328, "learning_rate": 0.0001999637478196305, "loss": 5.0013, "step": 44050 }, { "epoch": 0.08229550337182966, "grad_norm": 1.0313043594360352, "learning_rate": 0.0001999636346233997, "loss": 4.9081, "step": 44100 }, { "epoch": 0.08238880893120815, "grad_norm": 0.9132486581802368, "learning_rate": 0.00019996352125075067, "loss": 4.5762, "step": 44150 }, { "epoch": 0.08248211449058664, "grad_norm": 1.090657353401184, "learning_rate": 0.00019996340770168363, "loss": 4.9538, "step": 44200 }, { "epoch": 0.08257542004996513, "grad_norm": 1.004305124282837, "learning_rate": 0.0001999632939761988, "loss": 4.9216, "step": 44250 }, { "epoch": 0.08266872560934362, "grad_norm": 1.309733510017395, "learning_rate": 0.00019996318007429632, "loss": 4.8428, "step": 44300 }, { "epoch": 0.0827620311687221, "grad_norm": 1.159638524055481, "learning_rate": 0.0001999630659959764, "loss": 4.9621, "step": 44350 }, { "epoch": 0.0828553367281006, "grad_norm": 1.147161602973938, "learning_rate": 0.00019996295174123927, "loss": 4.7534, "step": 44400 }, { "epoch": 0.0829486422874791, "grad_norm": 0.9718197584152222, "learning_rate": 0.00019996283731008513, "loss": 4.8755, "step": 44450 }, { "epoch": 0.08304194784685759, "grad_norm": 0.9277188181877136, "learning_rate": 0.00019996272270251416, "loss": 4.8344, "step": 44500 }, { "epoch": 0.08313525340623608, "grad_norm": 1.071356177330017, "learning_rate": 0.00019996260791852656, "loss": 4.8703, "step": 44550 }, { "epoch": 0.08322855896561457, "grad_norm": 1.029938817024231, "learning_rate": 0.00019996249295812257, "loss": 4.8064, "step": 44600 }, { "epoch": 0.08332186452499306, "grad_norm": 1.397524356842041, "learning_rate": 0.00019996237782130236, "loss": 4.8744, "step": 44650 }, { "epoch": 0.08341517008437155, "grad_norm": 1.0888935327529907, "learning_rate": 0.00019996226250806615, "loss": 4.8386, "step": 44700 }, { "epoch": 0.08350847564375004, "grad_norm": 0.727277934551239, "learning_rate": 0.0001999621470184141, "loss": 4.8482, "step": 44750 }, { "epoch": 0.08360178120312854, "grad_norm": 1.148036241531372, "learning_rate": 0.00019996203135234648, "loss": 4.8495, "step": 44800 }, { "epoch": 0.08369508676250703, "grad_norm": 1.073508858680725, "learning_rate": 0.00019996191550986347, "loss": 4.8289, "step": 44850 }, { "epoch": 0.08378839232188552, "grad_norm": 0.7582327723503113, "learning_rate": 0.00019996179949096528, "loss": 4.7814, "step": 44900 }, { "epoch": 0.08388169788126401, "grad_norm": 0.9114712476730347, "learning_rate": 0.00019996168329565207, "loss": 4.8783, "step": 44950 }, { "epoch": 0.0839750034406425, "grad_norm": 1.14763343334198, "learning_rate": 0.00019996156692392408, "loss": 4.6919, "step": 45000 }, { "epoch": 0.0839750034406425, "eval_loss": 5.113004684448242, "eval_runtime": 231.8555, "eval_samples_per_second": 11.248, "eval_steps_per_second": 11.248, "eval_tts_loss": 7.2781233883782805, "step": 45000 }, { "epoch": 0.08406830900002099, "grad_norm": 1.0536847114562988, "learning_rate": 0.00019996145037578153, "loss": 4.741, "step": 45050 }, { "epoch": 0.08416161455939948, "grad_norm": 1.0102460384368896, "learning_rate": 0.0001999613336512246, "loss": 4.9963, "step": 45100 }, { "epoch": 0.08425492011877797, "grad_norm": 1.3230600357055664, "learning_rate": 0.0001999612167502535, "loss": 4.8845, "step": 45150 }, { "epoch": 0.08434822567815647, "grad_norm": 1.181218147277832, "learning_rate": 0.00019996109967286845, "loss": 4.7497, "step": 45200 }, { "epoch": 0.08444153123753496, "grad_norm": 1.2599127292633057, "learning_rate": 0.00019996098241906963, "loss": 4.9527, "step": 45250 }, { "epoch": 0.08453483679691345, "grad_norm": 1.0429056882858276, "learning_rate": 0.0001999608649888573, "loss": 4.8438, "step": 45300 }, { "epoch": 0.08462814235629194, "grad_norm": 0.9174293279647827, "learning_rate": 0.00019996074738223157, "loss": 4.885, "step": 45350 }, { "epoch": 0.08472144791567043, "grad_norm": 0.7659981846809387, "learning_rate": 0.00019996062959919277, "loss": 4.8716, "step": 45400 }, { "epoch": 0.08481475347504892, "grad_norm": 1.2928154468536377, "learning_rate": 0.00019996051163974102, "loss": 4.9047, "step": 45450 }, { "epoch": 0.08490805903442741, "grad_norm": 1.145384669303894, "learning_rate": 0.00019996039350387657, "loss": 4.9983, "step": 45500 }, { "epoch": 0.08500136459380592, "grad_norm": 1.0447450876235962, "learning_rate": 0.0001999602751915996, "loss": 4.7498, "step": 45550 }, { "epoch": 0.0850946701531844, "grad_norm": 1.108702301979065, "learning_rate": 0.00019996015670291036, "loss": 5.0015, "step": 45600 }, { "epoch": 0.0851879757125629, "grad_norm": 1.2329564094543457, "learning_rate": 0.00019996003803780902, "loss": 4.7922, "step": 45650 }, { "epoch": 0.08528128127194139, "grad_norm": 1.1614669561386108, "learning_rate": 0.00019995991919629578, "loss": 5.0659, "step": 45700 }, { "epoch": 0.08537458683131988, "grad_norm": 1.1069921255111694, "learning_rate": 0.00019995980017837087, "loss": 5.134, "step": 45750 }, { "epoch": 0.08546789239069837, "grad_norm": 1.1076229810714722, "learning_rate": 0.00019995968098403453, "loss": 4.7386, "step": 45800 }, { "epoch": 0.08556119795007686, "grad_norm": 0.8227840662002563, "learning_rate": 0.00019995956161328692, "loss": 4.854, "step": 45850 }, { "epoch": 0.08565450350945535, "grad_norm": 0.9372345805168152, "learning_rate": 0.0001999594420661283, "loss": 4.8924, "step": 45900 }, { "epoch": 0.08574780906883385, "grad_norm": 1.0700907707214355, "learning_rate": 0.00019995932234255883, "loss": 4.7795, "step": 45950 }, { "epoch": 0.08584111462821234, "grad_norm": 1.173011064529419, "learning_rate": 0.00019995920244257872, "loss": 4.8779, "step": 46000 }, { "epoch": 0.08593442018759083, "grad_norm": 1.0453685522079468, "learning_rate": 0.00019995908236618824, "loss": 4.7893, "step": 46050 }, { "epoch": 0.08602772574696932, "grad_norm": 1.3147802352905273, "learning_rate": 0.00019995896211338757, "loss": 4.8669, "step": 46100 }, { "epoch": 0.08612103130634781, "grad_norm": 1.0655571222305298, "learning_rate": 0.00019995884168417692, "loss": 4.8534, "step": 46150 }, { "epoch": 0.0862143368657263, "grad_norm": 1.1845111846923828, "learning_rate": 0.00019995872107855648, "loss": 4.8177, "step": 46200 }, { "epoch": 0.08630764242510479, "grad_norm": 1.0939445495605469, "learning_rate": 0.0001999586002965265, "loss": 4.7689, "step": 46250 }, { "epoch": 0.08640094798448329, "grad_norm": 1.0770061016082764, "learning_rate": 0.00019995847933808718, "loss": 4.7746, "step": 46300 }, { "epoch": 0.08649425354386178, "grad_norm": 1.2694488763809204, "learning_rate": 0.00019995835820323872, "loss": 4.7283, "step": 46350 }, { "epoch": 0.08658755910324027, "grad_norm": 1.181217908859253, "learning_rate": 0.00019995823689198137, "loss": 4.9545, "step": 46400 }, { "epoch": 0.08668086466261876, "grad_norm": 1.1785484552383423, "learning_rate": 0.0001999581154043153, "loss": 4.7946, "step": 46450 }, { "epoch": 0.08677417022199725, "grad_norm": 0.8478593826293945, "learning_rate": 0.00019995799374024077, "loss": 4.5888, "step": 46500 }, { "epoch": 0.08686747578137574, "grad_norm": 0.7958502173423767, "learning_rate": 0.00019995787189975794, "loss": 4.7813, "step": 46550 }, { "epoch": 0.08696078134075423, "grad_norm": 1.2592016458511353, "learning_rate": 0.00019995774988286707, "loss": 4.7929, "step": 46600 }, { "epoch": 0.08705408690013272, "grad_norm": 0.9072364568710327, "learning_rate": 0.00019995762768956833, "loss": 4.7389, "step": 46650 }, { "epoch": 0.08714739245951122, "grad_norm": 0.8660083413124084, "learning_rate": 0.000199957505319862, "loss": 4.9873, "step": 46700 }, { "epoch": 0.08724069801888971, "grad_norm": 1.458755373954773, "learning_rate": 0.00019995738277374825, "loss": 4.851, "step": 46750 }, { "epoch": 0.0873340035782682, "grad_norm": 1.0809298753738403, "learning_rate": 0.0001999572600512273, "loss": 4.8688, "step": 46800 }, { "epoch": 0.0874273091376467, "grad_norm": 1.176781415939331, "learning_rate": 0.00019995713715229937, "loss": 4.848, "step": 46850 }, { "epoch": 0.08752061469702518, "grad_norm": 1.1064727306365967, "learning_rate": 0.0001999570140769647, "loss": 4.9472, "step": 46900 }, { "epoch": 0.08761392025640367, "grad_norm": 1.0191171169281006, "learning_rate": 0.00019995689082522348, "loss": 5.0196, "step": 46950 }, { "epoch": 0.08770722581578216, "grad_norm": 1.0902825593948364, "learning_rate": 0.00019995676739707597, "loss": 4.8434, "step": 47000 }, { "epoch": 0.08780053137516065, "grad_norm": 1.0799312591552734, "learning_rate": 0.0001999566437925223, "loss": 4.7215, "step": 47050 }, { "epoch": 0.08789383693453916, "grad_norm": 0.7429112792015076, "learning_rate": 0.00019995652001156278, "loss": 4.7088, "step": 47100 }, { "epoch": 0.08798714249391765, "grad_norm": 1.215327262878418, "learning_rate": 0.00019995639605419757, "loss": 4.7287, "step": 47150 }, { "epoch": 0.08808044805329614, "grad_norm": 0.9942157864570618, "learning_rate": 0.0001999562719204269, "loss": 4.8316, "step": 47200 }, { "epoch": 0.08817375361267463, "grad_norm": 1.1454399824142456, "learning_rate": 0.00019995614761025098, "loss": 4.9424, "step": 47250 }, { "epoch": 0.08826705917205312, "grad_norm": 1.0674301385879517, "learning_rate": 0.0001999560231236701, "loss": 4.8136, "step": 47300 }, { "epoch": 0.0883603647314316, "grad_norm": 1.0504412651062012, "learning_rate": 0.0001999558984606844, "loss": 4.8764, "step": 47350 }, { "epoch": 0.0884536702908101, "grad_norm": 1.0923010110855103, "learning_rate": 0.00019995577362129412, "loss": 4.8347, "step": 47400 }, { "epoch": 0.0885469758501886, "grad_norm": 0.7745321989059448, "learning_rate": 0.0001999556486054995, "loss": 4.9508, "step": 47450 }, { "epoch": 0.08864028140956709, "grad_norm": 0.992335319519043, "learning_rate": 0.00019995552341330074, "loss": 4.8245, "step": 47500 }, { "epoch": 0.08873358696894558, "grad_norm": 1.1324732303619385, "learning_rate": 0.00019995539804469807, "loss": 4.7327, "step": 47550 }, { "epoch": 0.08882689252832407, "grad_norm": 1.0124810934066772, "learning_rate": 0.00019995527249969174, "loss": 4.7606, "step": 47600 }, { "epoch": 0.08892019808770256, "grad_norm": 1.1177074909210205, "learning_rate": 0.00019995514677828192, "loss": 4.9257, "step": 47650 }, { "epoch": 0.08901350364708105, "grad_norm": 1.0227179527282715, "learning_rate": 0.00019995502088046885, "loss": 4.7303, "step": 47700 }, { "epoch": 0.08910680920645954, "grad_norm": 0.9861817359924316, "learning_rate": 0.00019995489480625277, "loss": 4.932, "step": 47750 }, { "epoch": 0.08920011476583803, "grad_norm": 1.1230319738388062, "learning_rate": 0.00019995476855563384, "loss": 4.7706, "step": 47800 }, { "epoch": 0.08929342032521653, "grad_norm": 0.8386600613594055, "learning_rate": 0.00019995464212861237, "loss": 4.7139, "step": 47850 }, { "epoch": 0.08938672588459502, "grad_norm": 1.322847604751587, "learning_rate": 0.00019995451552518853, "loss": 4.8769, "step": 47900 }, { "epoch": 0.08948003144397351, "grad_norm": 1.3985576629638672, "learning_rate": 0.00019995438874536258, "loss": 4.9127, "step": 47950 }, { "epoch": 0.089573337003352, "grad_norm": 1.1547143459320068, "learning_rate": 0.0001999542617891347, "loss": 4.8883, "step": 48000 }, { "epoch": 0.089573337003352, "eval_loss": 5.081268310546875, "eval_runtime": 231.9586, "eval_samples_per_second": 11.243, "eval_steps_per_second": 11.243, "eval_tts_loss": 7.271998688881428, "step": 48000 }, { "epoch": 0.08966664256273049, "grad_norm": 0.7190650701522827, "learning_rate": 0.00019995413465650515, "loss": 4.971, "step": 48050 }, { "epoch": 0.08975994812210898, "grad_norm": 1.1452605724334717, "learning_rate": 0.00019995400734747413, "loss": 4.9515, "step": 48100 }, { "epoch": 0.08985325368148747, "grad_norm": 0.7332000136375427, "learning_rate": 0.00019995387986204187, "loss": 4.7901, "step": 48150 }, { "epoch": 0.08994655924086598, "grad_norm": 1.2664438486099243, "learning_rate": 0.00019995375220020862, "loss": 4.7153, "step": 48200 }, { "epoch": 0.09003986480024447, "grad_norm": 0.9586982727050781, "learning_rate": 0.00019995362436197459, "loss": 4.8614, "step": 48250 }, { "epoch": 0.09013317035962295, "grad_norm": 1.139235019683838, "learning_rate": 0.00019995349634733994, "loss": 4.8637, "step": 48300 }, { "epoch": 0.09022647591900144, "grad_norm": 1.2145622968673706, "learning_rate": 0.000199953368156305, "loss": 4.8257, "step": 48350 }, { "epoch": 0.09031978147837993, "grad_norm": 1.1296002864837646, "learning_rate": 0.00019995323978886997, "loss": 4.8457, "step": 48400 }, { "epoch": 0.09041308703775842, "grad_norm": 0.7904325723648071, "learning_rate": 0.00019995311124503505, "loss": 4.8158, "step": 48450 }, { "epoch": 0.09050639259713691, "grad_norm": 1.2136282920837402, "learning_rate": 0.00019995298252480046, "loss": 4.7482, "step": 48500 }, { "epoch": 0.0905996981565154, "grad_norm": 1.312458872795105, "learning_rate": 0.00019995285362816643, "loss": 4.7005, "step": 48550 }, { "epoch": 0.09069300371589391, "grad_norm": 1.1623774766921997, "learning_rate": 0.00019995272455513324, "loss": 4.8313, "step": 48600 }, { "epoch": 0.0907863092752724, "grad_norm": 1.2712230682373047, "learning_rate": 0.00019995259530570104, "loss": 5.055, "step": 48650 }, { "epoch": 0.09087961483465089, "grad_norm": 1.1107847690582275, "learning_rate": 0.00019995246587987012, "loss": 4.9028, "step": 48700 }, { "epoch": 0.09097292039402938, "grad_norm": 1.1616772413253784, "learning_rate": 0.00019995233627764067, "loss": 4.7912, "step": 48750 }, { "epoch": 0.09106622595340787, "grad_norm": 0.872082531452179, "learning_rate": 0.00019995220649901297, "loss": 4.8532, "step": 48800 }, { "epoch": 0.09115953151278636, "grad_norm": 0.9554139971733093, "learning_rate": 0.00019995207654398714, "loss": 4.7248, "step": 48850 }, { "epoch": 0.09125283707216485, "grad_norm": 1.1058579683303833, "learning_rate": 0.00019995194641256355, "loss": 4.6829, "step": 48900 }, { "epoch": 0.09134614263154335, "grad_norm": 1.0344289541244507, "learning_rate": 0.00019995181610474238, "loss": 4.9281, "step": 48950 }, { "epoch": 0.09143944819092184, "grad_norm": 1.0128055810928345, "learning_rate": 0.00019995168562052379, "loss": 4.8823, "step": 49000 }, { "epoch": 0.09153275375030033, "grad_norm": 1.189562201499939, "learning_rate": 0.00019995155495990805, "loss": 4.8325, "step": 49050 }, { "epoch": 0.09162605930967882, "grad_norm": 1.1200377941131592, "learning_rate": 0.00019995142412289544, "loss": 5.0455, "step": 49100 }, { "epoch": 0.09171936486905731, "grad_norm": 1.298416256904602, "learning_rate": 0.00019995129310948612, "loss": 4.8515, "step": 49150 }, { "epoch": 0.0918126704284358, "grad_norm": 1.1552833318710327, "learning_rate": 0.00019995116191968038, "loss": 4.8612, "step": 49200 }, { "epoch": 0.09190597598781429, "grad_norm": 1.0328021049499512, "learning_rate": 0.00019995103055347842, "loss": 5.0486, "step": 49250 }, { "epoch": 0.09199928154719278, "grad_norm": 0.7570762038230896, "learning_rate": 0.00019995089901088048, "loss": 4.7099, "step": 49300 }, { "epoch": 0.09209258710657128, "grad_norm": 0.9117104411125183, "learning_rate": 0.00019995076729188677, "loss": 4.8546, "step": 49350 }, { "epoch": 0.09218589266594977, "grad_norm": 1.0194332599639893, "learning_rate": 0.00019995063539649758, "loss": 4.8198, "step": 49400 }, { "epoch": 0.09227919822532826, "grad_norm": 0.7058835625648499, "learning_rate": 0.00019995050332471308, "loss": 4.8416, "step": 49450 }, { "epoch": 0.09237250378470675, "grad_norm": 1.046873688697815, "learning_rate": 0.00019995037107653355, "loss": 4.7131, "step": 49500 }, { "epoch": 0.09246580934408524, "grad_norm": 1.1325424909591675, "learning_rate": 0.00019995023865195918, "loss": 4.8868, "step": 49550 }, { "epoch": 0.09255911490346373, "grad_norm": 0.933991551399231, "learning_rate": 0.00019995010605099024, "loss": 4.8612, "step": 49600 }, { "epoch": 0.09265242046284222, "grad_norm": 1.2743916511535645, "learning_rate": 0.0001999499732736269, "loss": 4.7381, "step": 49650 }, { "epoch": 0.09274572602222073, "grad_norm": 0.9663635492324829, "learning_rate": 0.0001999498403198695, "loss": 4.8429, "step": 49700 }, { "epoch": 0.09283903158159922, "grad_norm": 1.0755882263183594, "learning_rate": 0.0001999497071897182, "loss": 4.8957, "step": 49750 }, { "epoch": 0.0929323371409777, "grad_norm": 1.0110505819320679, "learning_rate": 0.0001999495738831733, "loss": 5.0335, "step": 49800 }, { "epoch": 0.0930256427003562, "grad_norm": 1.1059298515319824, "learning_rate": 0.0001999494404002349, "loss": 4.9632, "step": 49850 }, { "epoch": 0.09311894825973469, "grad_norm": 1.1788870096206665, "learning_rate": 0.00019994930674090337, "loss": 5.0718, "step": 49900 }, { "epoch": 0.09321225381911317, "grad_norm": 1.0631682872772217, "learning_rate": 0.0001999491729051789, "loss": 4.6773, "step": 49950 }, { "epoch": 0.09330555937849166, "grad_norm": 0.9250020384788513, "learning_rate": 0.00019994903889306172, "loss": 4.7364, "step": 50000 }, { "epoch": 0.09339886493787015, "grad_norm": 1.3172533512115479, "learning_rate": 0.0001999489047045521, "loss": 4.8955, "step": 50050 }, { "epoch": 0.09349217049724866, "grad_norm": 1.0621302127838135, "learning_rate": 0.00019994877033965023, "loss": 4.7293, "step": 50100 }, { "epoch": 0.09358547605662715, "grad_norm": 1.2123887538909912, "learning_rate": 0.00019994863579835638, "loss": 4.7187, "step": 50150 }, { "epoch": 0.09367878161600564, "grad_norm": 1.2413897514343262, "learning_rate": 0.00019994850108067075, "loss": 4.9169, "step": 50200 }, { "epoch": 0.09377208717538413, "grad_norm": 1.3045127391815186, "learning_rate": 0.0001999483661865936, "loss": 4.8983, "step": 50250 }, { "epoch": 0.09386539273476262, "grad_norm": 1.0260050296783447, "learning_rate": 0.00019994823111612518, "loss": 4.8541, "step": 50300 }, { "epoch": 0.09395869829414111, "grad_norm": 0.8468745350837708, "learning_rate": 0.00019994809586926576, "loss": 4.7221, "step": 50350 }, { "epoch": 0.0940520038535196, "grad_norm": 0.558533251285553, "learning_rate": 0.00019994796044601549, "loss": 4.8051, "step": 50400 }, { "epoch": 0.09414530941289809, "grad_norm": 1.0962811708450317, "learning_rate": 0.00019994782484637465, "loss": 5.0269, "step": 50450 }, { "epoch": 0.09423861497227659, "grad_norm": 1.1004583835601807, "learning_rate": 0.00019994768907034352, "loss": 4.6689, "step": 50500 }, { "epoch": 0.09433192053165508, "grad_norm": 0.585641086101532, "learning_rate": 0.00019994755311792227, "loss": 4.913, "step": 50550 }, { "epoch": 0.09442522609103357, "grad_norm": 1.0821267366409302, "learning_rate": 0.0001999474169891112, "loss": 4.8242, "step": 50600 }, { "epoch": 0.09451853165041206, "grad_norm": 0.7899723649024963, "learning_rate": 0.0001999472806839105, "loss": 4.8794, "step": 50650 }, { "epoch": 0.09461183720979055, "grad_norm": 1.1543865203857422, "learning_rate": 0.00019994714420232046, "loss": 4.9543, "step": 50700 }, { "epoch": 0.09470514276916904, "grad_norm": 0.8708122372627258, "learning_rate": 0.00019994700754434128, "loss": 4.6829, "step": 50750 }, { "epoch": 0.09479844832854753, "grad_norm": 1.07050359249115, "learning_rate": 0.00019994687070997325, "loss": 4.8473, "step": 50800 }, { "epoch": 0.09489175388792603, "grad_norm": 0.9256060123443604, "learning_rate": 0.00019994673369921656, "loss": 4.6233, "step": 50850 }, { "epoch": 0.09498505944730452, "grad_norm": 1.1149330139160156, "learning_rate": 0.00019994659651207143, "loss": 4.88, "step": 50900 }, { "epoch": 0.09507836500668301, "grad_norm": 1.1947017908096313, "learning_rate": 0.00019994645914853817, "loss": 4.7873, "step": 50950 }, { "epoch": 0.0951716705660615, "grad_norm": 0.9794861674308777, "learning_rate": 0.00019994632160861704, "loss": 4.8523, "step": 51000 }, { "epoch": 0.0951716705660615, "eval_loss": 5.061774730682373, "eval_runtime": 231.5203, "eval_samples_per_second": 11.265, "eval_steps_per_second": 11.265, "eval_tts_loss": 7.302327345133947, "step": 51000 }, { "epoch": 0.09526497612543999, "grad_norm": 1.0766355991363525, "learning_rate": 0.0001999461838923082, "loss": 4.662, "step": 51050 }, { "epoch": 0.09535828168481848, "grad_norm": 1.242478609085083, "learning_rate": 0.00019994604599961194, "loss": 5.0057, "step": 51100 }, { "epoch": 0.09545158724419697, "grad_norm": 1.0459357500076294, "learning_rate": 0.0001999459079305285, "loss": 4.8375, "step": 51150 }, { "epoch": 0.09554489280357546, "grad_norm": 1.0175572633743286, "learning_rate": 0.0001999457696850581, "loss": 4.7238, "step": 51200 }, { "epoch": 0.09563819836295397, "grad_norm": 1.033231258392334, "learning_rate": 0.000199945631263201, "loss": 5.1248, "step": 51250 }, { "epoch": 0.09573150392233246, "grad_norm": 1.0977270603179932, "learning_rate": 0.00019994549266495745, "loss": 4.7472, "step": 51300 }, { "epoch": 0.09582480948171095, "grad_norm": 1.0846363306045532, "learning_rate": 0.00019994535389032772, "loss": 4.7239, "step": 51350 }, { "epoch": 0.09591811504108944, "grad_norm": 0.9102774858474731, "learning_rate": 0.000199945214939312, "loss": 4.9151, "step": 51400 }, { "epoch": 0.09601142060046793, "grad_norm": 1.0620869398117065, "learning_rate": 0.00019994507581191058, "loss": 4.5926, "step": 51450 }, { "epoch": 0.09610472615984642, "grad_norm": 1.1233978271484375, "learning_rate": 0.00019994493650812368, "loss": 5.0843, "step": 51500 }, { "epoch": 0.0961980317192249, "grad_norm": 1.2142945528030396, "learning_rate": 0.00019994479702795155, "loss": 4.874, "step": 51550 }, { "epoch": 0.09629133727860341, "grad_norm": 1.21846342086792, "learning_rate": 0.00019994465737139443, "loss": 4.888, "step": 51600 }, { "epoch": 0.0963846428379819, "grad_norm": 1.1796367168426514, "learning_rate": 0.00019994451753845257, "loss": 4.9034, "step": 51650 }, { "epoch": 0.09647794839736039, "grad_norm": 0.9840373992919922, "learning_rate": 0.00019994437752912623, "loss": 4.759, "step": 51700 }, { "epoch": 0.09657125395673888, "grad_norm": 1.2641807794570923, "learning_rate": 0.00019994423734341567, "loss": 4.7452, "step": 51750 }, { "epoch": 0.09666455951611737, "grad_norm": 1.1618438959121704, "learning_rate": 0.0001999440969813211, "loss": 4.6881, "step": 51800 }, { "epoch": 0.09675786507549586, "grad_norm": 0.8106015920639038, "learning_rate": 0.0001999439564428428, "loss": 5.0901, "step": 51850 }, { "epoch": 0.09685117063487435, "grad_norm": 1.3673611879348755, "learning_rate": 0.00019994381572798096, "loss": 4.9073, "step": 51900 }, { "epoch": 0.09694447619425284, "grad_norm": 0.9756952524185181, "learning_rate": 0.0001999436748367359, "loss": 4.9472, "step": 51950 }, { "epoch": 0.09703778175363134, "grad_norm": 1.3207862377166748, "learning_rate": 0.00019994353376910783, "loss": 4.7792, "step": 52000 }, { "epoch": 0.09713108731300983, "grad_norm": 1.0687203407287598, "learning_rate": 0.00019994339252509704, "loss": 4.7979, "step": 52050 }, { "epoch": 0.09722439287238832, "grad_norm": 1.2489949464797974, "learning_rate": 0.00019994325110470371, "loss": 4.8534, "step": 52100 }, { "epoch": 0.09731769843176681, "grad_norm": 1.4121779203414917, "learning_rate": 0.00019994310950792815, "loss": 4.8141, "step": 52150 }, { "epoch": 0.0974110039911453, "grad_norm": 1.0948699712753296, "learning_rate": 0.00019994296773477056, "loss": 4.9626, "step": 52200 }, { "epoch": 0.09750430955052379, "grad_norm": 0.8050052523612976, "learning_rate": 0.00019994282578523126, "loss": 4.838, "step": 52250 }, { "epoch": 0.09759761510990228, "grad_norm": 1.2219855785369873, "learning_rate": 0.00019994268365931044, "loss": 5.0231, "step": 52300 }, { "epoch": 0.09769092066928078, "grad_norm": 1.0798107385635376, "learning_rate": 0.00019994254135700837, "loss": 4.9951, "step": 52350 }, { "epoch": 0.09778422622865927, "grad_norm": 1.1264784336090088, "learning_rate": 0.0001999423988783253, "loss": 4.9835, "step": 52400 }, { "epoch": 0.09787753178803776, "grad_norm": 1.0738651752471924, "learning_rate": 0.00019994225622326143, "loss": 4.6703, "step": 52450 }, { "epoch": 0.09797083734741625, "grad_norm": 0.9834763407707214, "learning_rate": 0.0001999421133918171, "loss": 4.8748, "step": 52500 }, { "epoch": 0.09806414290679474, "grad_norm": 0.9889846444129944, "learning_rate": 0.00019994197038399256, "loss": 5.0041, "step": 52550 }, { "epoch": 0.09815744846617323, "grad_norm": 0.6598125696182251, "learning_rate": 0.000199941827199788, "loss": 4.7931, "step": 52600 }, { "epoch": 0.09825075402555172, "grad_norm": 0.7698989510536194, "learning_rate": 0.00019994168383920372, "loss": 4.6956, "step": 52650 }, { "epoch": 0.09834405958493021, "grad_norm": 0.9986603260040283, "learning_rate": 0.00019994154030223994, "loss": 4.7647, "step": 52700 }, { "epoch": 0.09843736514430872, "grad_norm": 0.9387907385826111, "learning_rate": 0.0001999413965888969, "loss": 4.7919, "step": 52750 }, { "epoch": 0.0985306707036872, "grad_norm": 0.997485339641571, "learning_rate": 0.0001999412526991749, "loss": 4.864, "step": 52800 }, { "epoch": 0.0986239762630657, "grad_norm": 1.0263787508010864, "learning_rate": 0.0001999411086330742, "loss": 5.0764, "step": 52850 }, { "epoch": 0.09871728182244419, "grad_norm": 1.1817213296890259, "learning_rate": 0.000199940964390595, "loss": 4.9443, "step": 52900 }, { "epoch": 0.09881058738182268, "grad_norm": 1.2216764688491821, "learning_rate": 0.00019994081997173758, "loss": 4.9013, "step": 52950 }, { "epoch": 0.09890389294120117, "grad_norm": 1.020540475845337, "learning_rate": 0.00019994067537650221, "loss": 4.9208, "step": 53000 }, { "epoch": 0.09899719850057966, "grad_norm": 1.07760488986969, "learning_rate": 0.00019994053060488911, "loss": 4.8805, "step": 53050 }, { "epoch": 0.09909050405995816, "grad_norm": 0.9263754487037659, "learning_rate": 0.00019994038565689857, "loss": 4.6882, "step": 53100 }, { "epoch": 0.09918380961933665, "grad_norm": 1.0581586360931396, "learning_rate": 0.00019994024053253084, "loss": 4.7098, "step": 53150 }, { "epoch": 0.09927711517871514, "grad_norm": 0.9992542266845703, "learning_rate": 0.00019994009523178617, "loss": 4.8672, "step": 53200 }, { "epoch": 0.09937042073809363, "grad_norm": 0.9917731285095215, "learning_rate": 0.00019993994975466483, "loss": 4.7594, "step": 53250 }, { "epoch": 0.09946372629747212, "grad_norm": 0.8953468799591064, "learning_rate": 0.00019993980410116705, "loss": 4.9375, "step": 53300 }, { "epoch": 0.09955703185685061, "grad_norm": 1.0633111000061035, "learning_rate": 0.0001999396582712931, "loss": 4.7604, "step": 53350 }, { "epoch": 0.0996503374162291, "grad_norm": 0.9700153470039368, "learning_rate": 0.00019993951226504323, "loss": 4.8346, "step": 53400 }, { "epoch": 0.09974364297560759, "grad_norm": 1.198554277420044, "learning_rate": 0.00019993936608241771, "loss": 4.7809, "step": 53450 }, { "epoch": 0.09983694853498609, "grad_norm": 0.8726997971534729, "learning_rate": 0.0001999392197234168, "loss": 4.9039, "step": 53500 }, { "epoch": 0.09993025409436458, "grad_norm": 1.296087384223938, "learning_rate": 0.0001999390731880408, "loss": 4.7456, "step": 53550 }, { "epoch": 0.10002355965374307, "grad_norm": 0.9055943489074707, "learning_rate": 0.00019993892647628987, "loss": 4.964, "step": 53600 }, { "epoch": 0.10011686521312156, "grad_norm": 0.900660514831543, "learning_rate": 0.0001999387795881643, "loss": 4.92, "step": 53650 }, { "epoch": 0.10021017077250005, "grad_norm": 1.2182655334472656, "learning_rate": 0.0001999386325236644, "loss": 4.8738, "step": 53700 }, { "epoch": 0.10030347633187854, "grad_norm": 1.0977277755737305, "learning_rate": 0.0001999384852827904, "loss": 4.4823, "step": 53750 }, { "epoch": 0.10039678189125703, "grad_norm": 0.7490143775939941, "learning_rate": 0.00019993833786554252, "loss": 4.8447, "step": 53800 }, { "epoch": 0.10049008745063552, "grad_norm": 1.216125726699829, "learning_rate": 0.0001999381902719211, "loss": 4.7476, "step": 53850 }, { "epoch": 0.10058339301001402, "grad_norm": 0.9461069703102112, "learning_rate": 0.00019993804250192633, "loss": 4.7481, "step": 53900 }, { "epoch": 0.10067669856939251, "grad_norm": 0.9821112751960754, "learning_rate": 0.00019993789455555853, "loss": 4.6914, "step": 53950 }, { "epoch": 0.100770004128771, "grad_norm": 1.069184422492981, "learning_rate": 0.00019993774643281792, "loss": 4.8143, "step": 54000 }, { "epoch": 0.100770004128771, "eval_loss": 5.050092697143555, "eval_runtime": 233.5139, "eval_samples_per_second": 11.169, "eval_steps_per_second": 11.169, "eval_tts_loss": 7.309272037400504, "step": 54000 }, { "epoch": 0.1008633096881495, "grad_norm": 0.8678284883499146, "learning_rate": 0.00019993759813370475, "loss": 4.7695, "step": 54050 }, { "epoch": 0.10095661524752798, "grad_norm": 1.0255569219589233, "learning_rate": 0.00019993744965821932, "loss": 4.7333, "step": 54100 }, { "epoch": 0.10104992080690647, "grad_norm": 0.9796172976493835, "learning_rate": 0.0001999373010063619, "loss": 4.9283, "step": 54150 }, { "epoch": 0.10114322636628496, "grad_norm": 0.9378018975257874, "learning_rate": 0.00019993715217813272, "loss": 4.7371, "step": 54200 }, { "epoch": 0.10123653192566347, "grad_norm": 1.0029010772705078, "learning_rate": 0.00019993700317353201, "loss": 4.6258, "step": 54250 }, { "epoch": 0.10132983748504196, "grad_norm": 0.922479510307312, "learning_rate": 0.0001999368539925601, "loss": 5.0026, "step": 54300 }, { "epoch": 0.10142314304442045, "grad_norm": 1.0813497304916382, "learning_rate": 0.00019993670463521723, "loss": 4.8323, "step": 54350 }, { "epoch": 0.10151644860379894, "grad_norm": 1.0883569717407227, "learning_rate": 0.00019993655510150367, "loss": 4.8957, "step": 54400 }, { "epoch": 0.10160975416317743, "grad_norm": 0.792852520942688, "learning_rate": 0.00019993640539141966, "loss": 5.0127, "step": 54450 }, { "epoch": 0.10170305972255592, "grad_norm": 0.8241887092590332, "learning_rate": 0.0001999362555049655, "loss": 4.7821, "step": 54500 }, { "epoch": 0.1017963652819344, "grad_norm": 1.088869571685791, "learning_rate": 0.0001999361054421414, "loss": 4.8543, "step": 54550 }, { "epoch": 0.1018896708413129, "grad_norm": 1.0283081531524658, "learning_rate": 0.00019993595520294765, "loss": 4.7311, "step": 54600 }, { "epoch": 0.1019829764006914, "grad_norm": 0.7650019526481628, "learning_rate": 0.00019993580478738456, "loss": 4.7729, "step": 54650 }, { "epoch": 0.10207628196006989, "grad_norm": 0.8107566237449646, "learning_rate": 0.00019993565419545236, "loss": 5.0273, "step": 54700 }, { "epoch": 0.10216958751944838, "grad_norm": 1.2203528881072998, "learning_rate": 0.00019993550342715127, "loss": 4.6369, "step": 54750 }, { "epoch": 0.10226289307882687, "grad_norm": 0.9938797354698181, "learning_rate": 0.00019993535248248164, "loss": 4.7657, "step": 54800 }, { "epoch": 0.10235619863820536, "grad_norm": 1.1002684831619263, "learning_rate": 0.0001999352013614437, "loss": 4.8818, "step": 54850 }, { "epoch": 0.10244950419758385, "grad_norm": 0.9734085202217102, "learning_rate": 0.0001999350500640377, "loss": 4.8906, "step": 54900 }, { "epoch": 0.10254280975696234, "grad_norm": 0.9943709373474121, "learning_rate": 0.0001999348985902639, "loss": 4.8368, "step": 54950 }, { "epoch": 0.10263611531634084, "grad_norm": 0.8544298410415649, "learning_rate": 0.00019993474694012263, "loss": 4.62, "step": 55000 }, { "epoch": 0.10272942087571933, "grad_norm": 0.7274461388587952, "learning_rate": 0.0001999345951136141, "loss": 4.7548, "step": 55050 }, { "epoch": 0.10282272643509782, "grad_norm": 1.0700294971466064, "learning_rate": 0.00019993444311073859, "loss": 4.7764, "step": 55100 }, { "epoch": 0.10291603199447631, "grad_norm": 1.211059808731079, "learning_rate": 0.00019993429093149637, "loss": 4.7714, "step": 55150 }, { "epoch": 0.1030093375538548, "grad_norm": 1.0298805236816406, "learning_rate": 0.00019993413857588772, "loss": 4.8341, "step": 55200 }, { "epoch": 0.10310264311323329, "grad_norm": 1.2335697412490845, "learning_rate": 0.0001999339860439129, "loss": 4.7979, "step": 55250 }, { "epoch": 0.10319594867261178, "grad_norm": 1.05863618850708, "learning_rate": 0.00019993383333557216, "loss": 5.0299, "step": 55300 }, { "epoch": 0.10328925423199027, "grad_norm": 0.8664848804473877, "learning_rate": 0.0001999336804508658, "loss": 4.7784, "step": 55350 }, { "epoch": 0.10338255979136877, "grad_norm": 1.1443610191345215, "learning_rate": 0.00019993352738979407, "loss": 4.93, "step": 55400 }, { "epoch": 0.10347586535074726, "grad_norm": 1.0898007154464722, "learning_rate": 0.00019993337415235725, "loss": 4.6334, "step": 55450 }, { "epoch": 0.10356917091012575, "grad_norm": 0.9437859654426575, "learning_rate": 0.00019993322073855562, "loss": 4.8104, "step": 55500 }, { "epoch": 0.10366247646950424, "grad_norm": 1.0893656015396118, "learning_rate": 0.00019993306714838943, "loss": 4.8592, "step": 55550 }, { "epoch": 0.10375578202888273, "grad_norm": 1.0553542375564575, "learning_rate": 0.00019993291338185897, "loss": 4.8595, "step": 55600 }, { "epoch": 0.10384908758826122, "grad_norm": 0.9406054019927979, "learning_rate": 0.00019993275943896448, "loss": 4.7845, "step": 55650 }, { "epoch": 0.10394239314763971, "grad_norm": 0.9452017545700073, "learning_rate": 0.00019993260531970627, "loss": 4.6355, "step": 55700 }, { "epoch": 0.10403569870701822, "grad_norm": 0.8077996969223022, "learning_rate": 0.00019993245102408459, "loss": 4.9368, "step": 55750 }, { "epoch": 0.10412900426639671, "grad_norm": 0.9253317713737488, "learning_rate": 0.00019993229655209974, "loss": 4.6939, "step": 55800 }, { "epoch": 0.1042223098257752, "grad_norm": 1.2367135286331177, "learning_rate": 0.0001999321419037519, "loss": 5.0054, "step": 55850 }, { "epoch": 0.10431561538515369, "grad_norm": 0.9509602189064026, "learning_rate": 0.00019993198707904148, "loss": 4.7925, "step": 55900 }, { "epoch": 0.10440892094453218, "grad_norm": 0.9606555700302124, "learning_rate": 0.00019993183207796868, "loss": 4.7034, "step": 55950 }, { "epoch": 0.10450222650391067, "grad_norm": 0.7478787899017334, "learning_rate": 0.00019993167690053376, "loss": 4.9531, "step": 56000 }, { "epoch": 0.10459553206328916, "grad_norm": 0.9634212851524353, "learning_rate": 0.00019993152154673704, "loss": 4.9075, "step": 56050 }, { "epoch": 0.10468883762266765, "grad_norm": 0.83271723985672, "learning_rate": 0.00019993136601657873, "loss": 4.6116, "step": 56100 }, { "epoch": 0.10478214318204615, "grad_norm": 1.1285144090652466, "learning_rate": 0.0001999312103100592, "loss": 4.6186, "step": 56150 }, { "epoch": 0.10487544874142464, "grad_norm": 0.8766820430755615, "learning_rate": 0.00019993105442717862, "loss": 4.7467, "step": 56200 }, { "epoch": 0.10496875430080313, "grad_norm": 1.0494366884231567, "learning_rate": 0.00019993089836793732, "loss": 4.7853, "step": 56250 }, { "epoch": 0.10506205986018162, "grad_norm": 0.9477689862251282, "learning_rate": 0.00019993074213233556, "loss": 4.8802, "step": 56300 }, { "epoch": 0.10515536541956011, "grad_norm": 1.2510788440704346, "learning_rate": 0.00019993058572037364, "loss": 4.8521, "step": 56350 }, { "epoch": 0.1052486709789386, "grad_norm": 1.0136951208114624, "learning_rate": 0.0001999304291320518, "loss": 4.6612, "step": 56400 }, { "epoch": 0.10534197653831709, "grad_norm": 1.1237655878067017, "learning_rate": 0.00019993027236737038, "loss": 5.043, "step": 56450 }, { "epoch": 0.10543528209769558, "grad_norm": 0.8546143770217896, "learning_rate": 0.00019993011542632956, "loss": 4.6913, "step": 56500 }, { "epoch": 0.10552858765707408, "grad_norm": 1.2368345260620117, "learning_rate": 0.00019992995830892972, "loss": 4.9539, "step": 56550 }, { "epoch": 0.10562189321645257, "grad_norm": 1.055706262588501, "learning_rate": 0.00019992980101517102, "loss": 4.8411, "step": 56600 }, { "epoch": 0.10571519877583106, "grad_norm": 0.8533573746681213, "learning_rate": 0.00019992964354505387, "loss": 4.6453, "step": 56650 }, { "epoch": 0.10580850433520955, "grad_norm": 1.2660565376281738, "learning_rate": 0.00019992948589857843, "loss": 4.9287, "step": 56700 }, { "epoch": 0.10590180989458804, "grad_norm": 1.2691367864608765, "learning_rate": 0.00019992932807574507, "loss": 4.7645, "step": 56750 }, { "epoch": 0.10599511545396653, "grad_norm": 1.0898069143295288, "learning_rate": 0.00019992917007655403, "loss": 4.924, "step": 56800 }, { "epoch": 0.10608842101334502, "grad_norm": 1.1141910552978516, "learning_rate": 0.00019992901190100555, "loss": 4.7186, "step": 56850 }, { "epoch": 0.10618172657272353, "grad_norm": 1.047428846359253, "learning_rate": 0.00019992885354909998, "loss": 4.5938, "step": 56900 }, { "epoch": 0.10627503213210202, "grad_norm": 1.3936903476715088, "learning_rate": 0.00019992869502083755, "loss": 4.661, "step": 56950 }, { "epoch": 0.1063683376914805, "grad_norm": 1.2814745903015137, "learning_rate": 0.00019992853631621858, "loss": 5.1075, "step": 57000 }, { "epoch": 0.1063683376914805, "eval_loss": 5.03396463394165, "eval_runtime": 231.8405, "eval_samples_per_second": 11.249, "eval_steps_per_second": 11.249, "eval_tts_loss": 7.338024720789556, "step": 57000 }, { "epoch": 0.106461643250859, "grad_norm": 0.9817240834236145, "learning_rate": 0.0001999283774352433, "loss": 4.6815, "step": 57050 }, { "epoch": 0.10655494881023748, "grad_norm": 0.8881374001502991, "learning_rate": 0.000199928218377912, "loss": 4.7038, "step": 57100 }, { "epoch": 0.10664825436961597, "grad_norm": 0.8835790157318115, "learning_rate": 0.00019992805914422501, "loss": 4.805, "step": 57150 }, { "epoch": 0.10674155992899446, "grad_norm": 1.0517323017120361, "learning_rate": 0.0001999278997341826, "loss": 4.8259, "step": 57200 }, { "epoch": 0.10683486548837295, "grad_norm": 0.9587489366531372, "learning_rate": 0.000199927740147785, "loss": 4.7815, "step": 57250 }, { "epoch": 0.10692817104775146, "grad_norm": 1.1249165534973145, "learning_rate": 0.00019992758038503249, "loss": 4.5363, "step": 57300 }, { "epoch": 0.10702147660712995, "grad_norm": 1.0124765634536743, "learning_rate": 0.00019992742044592542, "loss": 4.6893, "step": 57350 }, { "epoch": 0.10711478216650844, "grad_norm": 0.969052255153656, "learning_rate": 0.00019992726033046405, "loss": 4.7008, "step": 57400 }, { "epoch": 0.10720808772588693, "grad_norm": 1.0661418437957764, "learning_rate": 0.00019992710003864864, "loss": 4.9145, "step": 57450 }, { "epoch": 0.10730139328526542, "grad_norm": 0.9736279845237732, "learning_rate": 0.00019992693957047948, "loss": 4.5799, "step": 57500 }, { "epoch": 0.10739469884464391, "grad_norm": 0.8981762528419495, "learning_rate": 0.0001999267789259568, "loss": 4.9882, "step": 57550 }, { "epoch": 0.1074880044040224, "grad_norm": 1.3107281923294067, "learning_rate": 0.000199926618105081, "loss": 4.5749, "step": 57600 }, { "epoch": 0.1075813099634009, "grad_norm": 0.8268171548843384, "learning_rate": 0.00019992645710785228, "loss": 4.7451, "step": 57650 }, { "epoch": 0.10767461552277939, "grad_norm": 0.8941455483436584, "learning_rate": 0.00019992629593427097, "loss": 4.7114, "step": 57700 }, { "epoch": 0.10776792108215788, "grad_norm": 1.0125592947006226, "learning_rate": 0.0001999261345843373, "loss": 4.7143, "step": 57750 }, { "epoch": 0.10786122664153637, "grad_norm": 1.0623835325241089, "learning_rate": 0.0001999259730580516, "loss": 4.8096, "step": 57800 }, { "epoch": 0.10795453220091486, "grad_norm": 0.9189125299453735, "learning_rate": 0.00019992581135541414, "loss": 4.5356, "step": 57850 }, { "epoch": 0.10804783776029335, "grad_norm": 1.0694884061813354, "learning_rate": 0.0001999256494764252, "loss": 4.902, "step": 57900 }, { "epoch": 0.10814114331967184, "grad_norm": 1.200424313545227, "learning_rate": 0.00019992548742108507, "loss": 4.6708, "step": 57950 }, { "epoch": 0.10823444887905033, "grad_norm": 1.0127471685409546, "learning_rate": 0.00019992532518939404, "loss": 4.6113, "step": 58000 }, { "epoch": 0.10832775443842883, "grad_norm": 1.081398367881775, "learning_rate": 0.0001999251627813524, "loss": 4.7227, "step": 58050 }, { "epoch": 0.10842105999780732, "grad_norm": 1.0527719259262085, "learning_rate": 0.00019992500019696042, "loss": 4.7023, "step": 58100 }, { "epoch": 0.10851436555718581, "grad_norm": 0.9371632933616638, "learning_rate": 0.0001999248374362184, "loss": 4.7177, "step": 58150 }, { "epoch": 0.1086076711165643, "grad_norm": 1.2352396249771118, "learning_rate": 0.00019992467449912664, "loss": 4.7167, "step": 58200 }, { "epoch": 0.10870097667594279, "grad_norm": 1.295750379562378, "learning_rate": 0.0001999245113856854, "loss": 4.6888, "step": 58250 }, { "epoch": 0.10879428223532128, "grad_norm": 0.9914165735244751, "learning_rate": 0.00019992434809589494, "loss": 4.6626, "step": 58300 }, { "epoch": 0.10888758779469977, "grad_norm": 1.038947582244873, "learning_rate": 0.00019992418462975566, "loss": 4.8624, "step": 58350 }, { "epoch": 0.10898089335407828, "grad_norm": 0.9562987685203552, "learning_rate": 0.00019992402098726776, "loss": 4.8079, "step": 58400 }, { "epoch": 0.10907419891345677, "grad_norm": 1.1640965938568115, "learning_rate": 0.0001999238571684315, "loss": 4.749, "step": 58450 }, { "epoch": 0.10916750447283526, "grad_norm": 0.8574897646903992, "learning_rate": 0.00019992369317324725, "loss": 4.5478, "step": 58500 }, { "epoch": 0.10926081003221375, "grad_norm": 1.1693881750106812, "learning_rate": 0.00019992352900171523, "loss": 4.8086, "step": 58550 }, { "epoch": 0.10935411559159224, "grad_norm": 1.0062192678451538, "learning_rate": 0.0001999233646538358, "loss": 4.7885, "step": 58600 }, { "epoch": 0.10944742115097073, "grad_norm": 1.1878728866577148, "learning_rate": 0.0001999232001296092, "loss": 4.7585, "step": 58650 }, { "epoch": 0.10954072671034921, "grad_norm": 0.6779685020446777, "learning_rate": 0.00019992303542903575, "loss": 4.8446, "step": 58700 }, { "epoch": 0.1096340322697277, "grad_norm": 1.022136926651001, "learning_rate": 0.0001999228705521157, "loss": 4.7668, "step": 58750 }, { "epoch": 0.10972733782910621, "grad_norm": 0.8543698787689209, "learning_rate": 0.00019992270549884936, "loss": 4.9229, "step": 58800 }, { "epoch": 0.1098206433884847, "grad_norm": 1.0954478979110718, "learning_rate": 0.00019992254026923704, "loss": 4.7663, "step": 58850 }, { "epoch": 0.10991394894786319, "grad_norm": 1.051382303237915, "learning_rate": 0.00019992237486327904, "loss": 4.9512, "step": 58900 }, { "epoch": 0.11000725450724168, "grad_norm": 1.2128174304962158, "learning_rate": 0.0001999222092809756, "loss": 4.677, "step": 58950 }, { "epoch": 0.11010056006662017, "grad_norm": 0.8626731038093567, "learning_rate": 0.00019992204352232707, "loss": 4.6943, "step": 59000 }, { "epoch": 0.11019386562599866, "grad_norm": 1.2140212059020996, "learning_rate": 0.00019992187758733368, "loss": 4.8939, "step": 59050 }, { "epoch": 0.11028717118537715, "grad_norm": 1.127719759941101, "learning_rate": 0.00019992171147599578, "loss": 4.8684, "step": 59100 }, { "epoch": 0.11038047674475565, "grad_norm": 1.0840810537338257, "learning_rate": 0.00019992154518831364, "loss": 4.947, "step": 59150 }, { "epoch": 0.11047378230413414, "grad_norm": 1.0713996887207031, "learning_rate": 0.00019992137872428754, "loss": 4.7424, "step": 59200 }, { "epoch": 0.11056708786351263, "grad_norm": 0.8163297176361084, "learning_rate": 0.0001999212120839178, "loss": 4.8426, "step": 59250 }, { "epoch": 0.11066039342289112, "grad_norm": 0.9037017226219177, "learning_rate": 0.0001999210452672047, "loss": 5.0381, "step": 59300 }, { "epoch": 0.11075369898226961, "grad_norm": 1.091113567352295, "learning_rate": 0.00019992087827414853, "loss": 4.5411, "step": 59350 }, { "epoch": 0.1108470045416481, "grad_norm": 0.872543454170227, "learning_rate": 0.0001999207111047496, "loss": 4.6965, "step": 59400 }, { "epoch": 0.11094031010102659, "grad_norm": 0.9569900035858154, "learning_rate": 0.0001999205437590082, "loss": 5.0155, "step": 59450 }, { "epoch": 0.11103361566040508, "grad_norm": 0.8814337253570557, "learning_rate": 0.00019992037623692461, "loss": 4.9238, "step": 59500 }, { "epoch": 0.11112692121978358, "grad_norm": 1.0400547981262207, "learning_rate": 0.00019992020853849916, "loss": 4.7072, "step": 59550 }, { "epoch": 0.11122022677916207, "grad_norm": 1.0725575685501099, "learning_rate": 0.0001999200406637321, "loss": 4.7132, "step": 59600 }, { "epoch": 0.11131353233854056, "grad_norm": 1.1183878183364868, "learning_rate": 0.00019991987261262377, "loss": 4.8993, "step": 59650 }, { "epoch": 0.11140683789791905, "grad_norm": 1.005928874015808, "learning_rate": 0.0001999197043851744, "loss": 4.7833, "step": 59700 }, { "epoch": 0.11150014345729754, "grad_norm": 0.9358251690864563, "learning_rate": 0.0001999195359813844, "loss": 4.6399, "step": 59750 }, { "epoch": 0.11159344901667603, "grad_norm": 0.9199875593185425, "learning_rate": 0.00019991936740125396, "loss": 4.9291, "step": 59800 }, { "epoch": 0.11168675457605452, "grad_norm": 1.1673732995986938, "learning_rate": 0.00019991919864478344, "loss": 4.8272, "step": 59850 }, { "epoch": 0.11178006013543301, "grad_norm": 1.0224826335906982, "learning_rate": 0.0001999190297119731, "loss": 4.6601, "step": 59900 }, { "epoch": 0.11187336569481152, "grad_norm": 1.0551832914352417, "learning_rate": 0.00019991886060282324, "loss": 4.9511, "step": 59950 }, { "epoch": 0.11196667125419, "grad_norm": 0.7255629301071167, "learning_rate": 0.00019991869131733418, "loss": 4.8392, "step": 60000 }, { "epoch": 0.11196667125419, "eval_loss": 5.020105838775635, "eval_runtime": 233.6602, "eval_samples_per_second": 11.162, "eval_steps_per_second": 11.162, "eval_tts_loss": 7.413437219613741, "step": 60000 }, { "epoch": 0.1120599768135685, "grad_norm": 1.0283355712890625, "learning_rate": 0.00019991852185550623, "loss": 4.8832, "step": 60050 }, { "epoch": 0.11215328237294699, "grad_norm": 0.8838055729866028, "learning_rate": 0.00019991835221733965, "loss": 4.6478, "step": 60100 }, { "epoch": 0.11224658793232548, "grad_norm": 1.2853875160217285, "learning_rate": 0.00019991818240283477, "loss": 4.7012, "step": 60150 }, { "epoch": 0.11233989349170397, "grad_norm": 1.3384721279144287, "learning_rate": 0.0001999180124119919, "loss": 4.8132, "step": 60200 }, { "epoch": 0.11243319905108246, "grad_norm": 1.0799473524093628, "learning_rate": 0.00019991784224481128, "loss": 4.793, "step": 60250 }, { "epoch": 0.11252650461046096, "grad_norm": 0.8523072004318237, "learning_rate": 0.00019991767190129326, "loss": 4.9316, "step": 60300 }, { "epoch": 0.11261981016983945, "grad_norm": 0.9551059007644653, "learning_rate": 0.00019991750138143813, "loss": 4.8224, "step": 60350 }, { "epoch": 0.11271311572921794, "grad_norm": 1.0353354215621948, "learning_rate": 0.0001999173306852462, "loss": 4.6761, "step": 60400 }, { "epoch": 0.11280642128859643, "grad_norm": 0.6829168796539307, "learning_rate": 0.00019991715981271778, "loss": 4.8235, "step": 60450 }, { "epoch": 0.11289972684797492, "grad_norm": 0.7613895535469055, "learning_rate": 0.00019991698876385314, "loss": 4.759, "step": 60500 }, { "epoch": 0.11299303240735341, "grad_norm": 0.9478106498718262, "learning_rate": 0.0001999168175386526, "loss": 4.6671, "step": 60550 }, { "epoch": 0.1130863379667319, "grad_norm": 0.9536855220794678, "learning_rate": 0.00019991664613711643, "loss": 4.8579, "step": 60600 }, { "epoch": 0.11317964352611039, "grad_norm": 1.144857406616211, "learning_rate": 0.00019991647455924497, "loss": 4.9233, "step": 60650 }, { "epoch": 0.11327294908548889, "grad_norm": 1.1187453269958496, "learning_rate": 0.00019991630280503852, "loss": 4.7061, "step": 60700 }, { "epoch": 0.11336625464486738, "grad_norm": 1.1460309028625488, "learning_rate": 0.0001999161308744974, "loss": 4.8344, "step": 60750 }, { "epoch": 0.11345956020424587, "grad_norm": 0.9266796112060547, "learning_rate": 0.00019991595876762185, "loss": 4.6502, "step": 60800 }, { "epoch": 0.11355286576362436, "grad_norm": 0.9658082127571106, "learning_rate": 0.00019991578648441224, "loss": 4.9107, "step": 60850 }, { "epoch": 0.11364617132300285, "grad_norm": 0.9839101433753967, "learning_rate": 0.00019991561402486885, "loss": 4.6822, "step": 60900 }, { "epoch": 0.11373947688238134, "grad_norm": 1.0026122331619263, "learning_rate": 0.00019991544138899195, "loss": 4.6888, "step": 60950 }, { "epoch": 0.11383278244175983, "grad_norm": 1.139105200767517, "learning_rate": 0.0001999152685767819, "loss": 4.6663, "step": 61000 }, { "epoch": 0.11392608800113833, "grad_norm": 0.9084557294845581, "learning_rate": 0.00019991509558823898, "loss": 4.77, "step": 61050 }, { "epoch": 0.11401939356051682, "grad_norm": 1.2279844284057617, "learning_rate": 0.0001999149224233635, "loss": 4.8361, "step": 61100 }, { "epoch": 0.11411269911989531, "grad_norm": 0.8564290404319763, "learning_rate": 0.00019991474908215574, "loss": 4.8206, "step": 61150 }, { "epoch": 0.1142060046792738, "grad_norm": 0.92253577709198, "learning_rate": 0.00019991457556461605, "loss": 4.8973, "step": 61200 }, { "epoch": 0.1142993102386523, "grad_norm": 1.0969637632369995, "learning_rate": 0.0001999144018707447, "loss": 4.9673, "step": 61250 }, { "epoch": 0.11439261579803078, "grad_norm": 1.1410199403762817, "learning_rate": 0.00019991422800054204, "loss": 4.6975, "step": 61300 }, { "epoch": 0.11448592135740927, "grad_norm": 1.0572748184204102, "learning_rate": 0.00019991405395400835, "loss": 4.7044, "step": 61350 }, { "epoch": 0.11457922691678776, "grad_norm": 1.115541696548462, "learning_rate": 0.0001999138797311439, "loss": 4.7546, "step": 61400 }, { "epoch": 0.11467253247616627, "grad_norm": 0.9594268798828125, "learning_rate": 0.00019991370533194904, "loss": 4.7057, "step": 61450 }, { "epoch": 0.11476583803554476, "grad_norm": 1.0258264541625977, "learning_rate": 0.00019991353075642408, "loss": 4.7853, "step": 61500 }, { "epoch": 0.11485914359492325, "grad_norm": 1.0233960151672363, "learning_rate": 0.00019991335600456932, "loss": 4.7917, "step": 61550 }, { "epoch": 0.11495244915430174, "grad_norm": 0.901678204536438, "learning_rate": 0.00019991318107638506, "loss": 4.9225, "step": 61600 }, { "epoch": 0.11504575471368023, "grad_norm": 0.7477099299430847, "learning_rate": 0.00019991300597187165, "loss": 4.7892, "step": 61650 }, { "epoch": 0.11513906027305872, "grad_norm": 1.0212405920028687, "learning_rate": 0.00019991283069102932, "loss": 4.789, "step": 61700 }, { "epoch": 0.1152323658324372, "grad_norm": 0.9930858016014099, "learning_rate": 0.00019991265523385842, "loss": 4.7956, "step": 61750 }, { "epoch": 0.11532567139181571, "grad_norm": 0.9863542318344116, "learning_rate": 0.0001999124796003593, "loss": 4.7178, "step": 61800 }, { "epoch": 0.1154189769511942, "grad_norm": 0.8697104454040527, "learning_rate": 0.0001999123037905322, "loss": 4.7792, "step": 61850 }, { "epoch": 0.11551228251057269, "grad_norm": 1.0396831035614014, "learning_rate": 0.0001999121278043775, "loss": 4.8915, "step": 61900 }, { "epoch": 0.11560558806995118, "grad_norm": 1.0583255290985107, "learning_rate": 0.00019991195164189546, "loss": 4.7479, "step": 61950 }, { "epoch": 0.11569889362932967, "grad_norm": 0.8265174627304077, "learning_rate": 0.0001999117753030864, "loss": 4.7255, "step": 62000 }, { "epoch": 0.11579219918870816, "grad_norm": 1.1002154350280762, "learning_rate": 0.00019991159878795064, "loss": 4.8505, "step": 62050 }, { "epoch": 0.11588550474808665, "grad_norm": 1.1061955690383911, "learning_rate": 0.00019991142209648848, "loss": 4.6528, "step": 62100 }, { "epoch": 0.11597881030746514, "grad_norm": 1.06044340133667, "learning_rate": 0.00019991124522870024, "loss": 4.8613, "step": 62150 }, { "epoch": 0.11607211586684364, "grad_norm": 1.1321583986282349, "learning_rate": 0.00019991106818458622, "loss": 4.6643, "step": 62200 }, { "epoch": 0.11616542142622213, "grad_norm": 1.0335965156555176, "learning_rate": 0.00019991089096414678, "loss": 4.7951, "step": 62250 }, { "epoch": 0.11625872698560062, "grad_norm": 1.1257051229476929, "learning_rate": 0.00019991071356738216, "loss": 4.8458, "step": 62300 }, { "epoch": 0.11635203254497911, "grad_norm": 0.9923104047775269, "learning_rate": 0.00019991053599429273, "loss": 4.7476, "step": 62350 }, { "epoch": 0.1164453381043576, "grad_norm": 1.1084821224212646, "learning_rate": 0.00019991035824487878, "loss": 4.6966, "step": 62400 }, { "epoch": 0.11653864366373609, "grad_norm": 0.9355782866477966, "learning_rate": 0.00019991018031914064, "loss": 4.7396, "step": 62450 }, { "epoch": 0.11663194922311458, "grad_norm": 1.1987919807434082, "learning_rate": 0.0001999100022170786, "loss": 4.7583, "step": 62500 }, { "epoch": 0.11672525478249308, "grad_norm": 1.2184990644454956, "learning_rate": 0.000199909823938693, "loss": 5.0186, "step": 62550 }, { "epoch": 0.11681856034187157, "grad_norm": 1.0517642498016357, "learning_rate": 0.00019990964548398412, "loss": 4.9411, "step": 62600 }, { "epoch": 0.11691186590125006, "grad_norm": 1.2433594465255737, "learning_rate": 0.00019990946685295228, "loss": 4.7959, "step": 62650 }, { "epoch": 0.11700517146062855, "grad_norm": 1.1244089603424072, "learning_rate": 0.00019990928804559786, "loss": 4.6779, "step": 62700 }, { "epoch": 0.11709847702000704, "grad_norm": 1.0606389045715332, "learning_rate": 0.00019990910906192107, "loss": 4.8322, "step": 62750 }, { "epoch": 0.11719178257938553, "grad_norm": 1.1183406114578247, "learning_rate": 0.00019990892990192233, "loss": 4.7931, "step": 62800 }, { "epoch": 0.11728508813876402, "grad_norm": 0.8485277891159058, "learning_rate": 0.0001999087505656019, "loss": 4.8164, "step": 62850 }, { "epoch": 0.11737839369814251, "grad_norm": 0.9464021921157837, "learning_rate": 0.0001999085710529601, "loss": 4.7665, "step": 62900 }, { "epoch": 0.11747169925752102, "grad_norm": 0.7202494144439697, "learning_rate": 0.00019990839136399725, "loss": 4.7638, "step": 62950 }, { "epoch": 0.11756500481689951, "grad_norm": 1.0719358921051025, "learning_rate": 0.00019990821149871364, "loss": 4.8894, "step": 63000 }, { "epoch": 0.11756500481689951, "eval_loss": 5.0077104568481445, "eval_runtime": 233.0923, "eval_samples_per_second": 11.189, "eval_steps_per_second": 11.189, "eval_tts_loss": 7.356751820346786, "step": 63000 }, { "epoch": 0.117658310376278, "grad_norm": 1.0613762140274048, "learning_rate": 0.00019990803145710965, "loss": 4.8346, "step": 63050 }, { "epoch": 0.11775161593565649, "grad_norm": 0.9940460324287415, "learning_rate": 0.00019990785123918557, "loss": 4.8047, "step": 63100 }, { "epoch": 0.11784492149503498, "grad_norm": 1.0382481813430786, "learning_rate": 0.0001999076708449417, "loss": 4.6179, "step": 63150 }, { "epoch": 0.11793822705441347, "grad_norm": 1.2504392862319946, "learning_rate": 0.00019990749027437837, "loss": 4.7242, "step": 63200 }, { "epoch": 0.11803153261379196, "grad_norm": 1.3020495176315308, "learning_rate": 0.00019990730952749587, "loss": 4.7754, "step": 63250 }, { "epoch": 0.11812483817317045, "grad_norm": 0.9937262535095215, "learning_rate": 0.00019990712860429457, "loss": 4.6131, "step": 63300 }, { "epoch": 0.11821814373254895, "grad_norm": 0.7087301015853882, "learning_rate": 0.0001999069475047748, "loss": 4.8183, "step": 63350 }, { "epoch": 0.11831144929192744, "grad_norm": 1.2310640811920166, "learning_rate": 0.0001999067662289368, "loss": 4.7773, "step": 63400 }, { "epoch": 0.11840475485130593, "grad_norm": 1.2428086996078491, "learning_rate": 0.00019990658477678097, "loss": 4.7297, "step": 63450 }, { "epoch": 0.11849806041068442, "grad_norm": 0.7657245397567749, "learning_rate": 0.00019990640314830756, "loss": 4.9017, "step": 63500 }, { "epoch": 0.11859136597006291, "grad_norm": 1.0270503759384155, "learning_rate": 0.00019990622134351694, "loss": 4.7397, "step": 63550 }, { "epoch": 0.1186846715294414, "grad_norm": 0.7919912338256836, "learning_rate": 0.00019990603936240943, "loss": 4.8849, "step": 63600 }, { "epoch": 0.11877797708881989, "grad_norm": 1.0639008283615112, "learning_rate": 0.00019990585720498536, "loss": 4.8832, "step": 63650 }, { "epoch": 0.11887128264819839, "grad_norm": 1.0238081216812134, "learning_rate": 0.00019990567487124502, "loss": 4.6403, "step": 63700 }, { "epoch": 0.11896458820757688, "grad_norm": 1.0058157444000244, "learning_rate": 0.00019990549236118872, "loss": 4.5824, "step": 63750 }, { "epoch": 0.11905789376695537, "grad_norm": 0.7168673276901245, "learning_rate": 0.00019990530967481683, "loss": 4.8433, "step": 63800 }, { "epoch": 0.11915119932633386, "grad_norm": 0.9038125276565552, "learning_rate": 0.0001999051268121296, "loss": 4.8536, "step": 63850 }, { "epoch": 0.11924450488571235, "grad_norm": 0.9761154651641846, "learning_rate": 0.00019990494377312745, "loss": 4.7876, "step": 63900 }, { "epoch": 0.11933781044509084, "grad_norm": 1.0363209247589111, "learning_rate": 0.00019990476055781065, "loss": 4.8438, "step": 63950 }, { "epoch": 0.11943111600446933, "grad_norm": 0.9672982096672058, "learning_rate": 0.0001999045771661795, "loss": 4.7065, "step": 64000 }, { "epoch": 0.11952442156384782, "grad_norm": 1.009544014930725, "learning_rate": 0.00019990439359823437, "loss": 4.8137, "step": 64050 }, { "epoch": 0.11961772712322633, "grad_norm": 0.8747443556785583, "learning_rate": 0.00019990420985397552, "loss": 4.6859, "step": 64100 }, { "epoch": 0.11971103268260481, "grad_norm": 1.121934413909912, "learning_rate": 0.0001999040259334034, "loss": 4.8379, "step": 64150 }, { "epoch": 0.1198043382419833, "grad_norm": 0.8262189030647278, "learning_rate": 0.0001999038418365182, "loss": 4.7602, "step": 64200 }, { "epoch": 0.1198976438013618, "grad_norm": 0.9570197463035583, "learning_rate": 0.0001999036575633203, "loss": 4.8409, "step": 64250 }, { "epoch": 0.11999094936074028, "grad_norm": 0.7180740833282471, "learning_rate": 0.00019990347311381004, "loss": 4.9055, "step": 64300 }, { "epoch": 0.12008425492011877, "grad_norm": 0.7505408525466919, "learning_rate": 0.0001999032884879877, "loss": 4.8526, "step": 64350 }, { "epoch": 0.12017756047949726, "grad_norm": 0.9490662813186646, "learning_rate": 0.00019990310368585366, "loss": 4.6729, "step": 64400 }, { "epoch": 0.12027086603887577, "grad_norm": 1.0772948265075684, "learning_rate": 0.0001999029187074082, "loss": 4.6781, "step": 64450 }, { "epoch": 0.12036417159825426, "grad_norm": 0.6764106154441833, "learning_rate": 0.00019990273355265168, "loss": 4.8261, "step": 64500 }, { "epoch": 0.12045747715763275, "grad_norm": 0.8827162981033325, "learning_rate": 0.00019990254822158443, "loss": 4.8375, "step": 64550 }, { "epoch": 0.12055078271701124, "grad_norm": 0.9409701228141785, "learning_rate": 0.00019990236271420674, "loss": 4.8701, "step": 64600 }, { "epoch": 0.12064408827638973, "grad_norm": 0.8678632974624634, "learning_rate": 0.00019990217703051894, "loss": 4.9678, "step": 64650 }, { "epoch": 0.12073739383576822, "grad_norm": 1.1174262762069702, "learning_rate": 0.00019990199117052141, "loss": 4.8362, "step": 64700 }, { "epoch": 0.1208306993951467, "grad_norm": 0.8138838410377502, "learning_rate": 0.00019990180513421445, "loss": 4.9131, "step": 64750 }, { "epoch": 0.1209240049545252, "grad_norm": 1.0765495300292969, "learning_rate": 0.00019990161892159835, "loss": 4.8251, "step": 64800 }, { "epoch": 0.1210173105139037, "grad_norm": 1.0338910818099976, "learning_rate": 0.0001999014325326735, "loss": 4.4701, "step": 64850 }, { "epoch": 0.12111061607328219, "grad_norm": 0.7882349491119385, "learning_rate": 0.00019990124596744018, "loss": 4.6385, "step": 64900 }, { "epoch": 0.12120392163266068, "grad_norm": 1.1244629621505737, "learning_rate": 0.00019990105922589876, "loss": 4.6897, "step": 64950 }, { "epoch": 0.12129722719203917, "grad_norm": 1.0500319004058838, "learning_rate": 0.0001999008723080495, "loss": 4.6929, "step": 65000 }, { "epoch": 0.12139053275141766, "grad_norm": 0.9561808705329895, "learning_rate": 0.0001999006852138928, "loss": 4.866, "step": 65050 }, { "epoch": 0.12148383831079615, "grad_norm": 1.0104138851165771, "learning_rate": 0.000199900497943429, "loss": 4.8589, "step": 65100 }, { "epoch": 0.12157714387017464, "grad_norm": 0.975612998008728, "learning_rate": 0.00019990031049665837, "loss": 4.9448, "step": 65150 }, { "epoch": 0.12167044942955314, "grad_norm": 0.8425891399383545, "learning_rate": 0.0001999001228735813, "loss": 4.6693, "step": 65200 }, { "epoch": 0.12176375498893163, "grad_norm": 1.091673493385315, "learning_rate": 0.00019989993507419805, "loss": 4.7452, "step": 65250 }, { "epoch": 0.12185706054831012, "grad_norm": 1.0499212741851807, "learning_rate": 0.00019989974709850903, "loss": 4.9866, "step": 65300 }, { "epoch": 0.12195036610768861, "grad_norm": 0.9292690753936768, "learning_rate": 0.0001998995589465145, "loss": 4.9873, "step": 65350 }, { "epoch": 0.1220436716670671, "grad_norm": 0.9211663603782654, "learning_rate": 0.00019989937061821485, "loss": 4.8513, "step": 65400 }, { "epoch": 0.12213697722644559, "grad_norm": 1.0245435237884521, "learning_rate": 0.00019989918211361037, "loss": 4.813, "step": 65450 }, { "epoch": 0.12223028278582408, "grad_norm": 0.8180379867553711, "learning_rate": 0.00019989899343270145, "loss": 4.8255, "step": 65500 }, { "epoch": 0.12232358834520257, "grad_norm": 1.1613210439682007, "learning_rate": 0.00019989880457548832, "loss": 5.14, "step": 65550 }, { "epoch": 0.12241689390458108, "grad_norm": 1.1680033206939697, "learning_rate": 0.00019989861554197143, "loss": 4.7817, "step": 65600 }, { "epoch": 0.12251019946395957, "grad_norm": 1.1515231132507324, "learning_rate": 0.00019989842633215106, "loss": 4.8963, "step": 65650 }, { "epoch": 0.12260350502333806, "grad_norm": 0.9831023216247559, "learning_rate": 0.00019989823694602753, "loss": 4.613, "step": 65700 }, { "epoch": 0.12269681058271655, "grad_norm": 1.052646279335022, "learning_rate": 0.0001998980473836012, "loss": 4.7509, "step": 65750 }, { "epoch": 0.12279011614209503, "grad_norm": 0.8736461400985718, "learning_rate": 0.00019989785764487237, "loss": 4.8964, "step": 65800 }, { "epoch": 0.12288342170147352, "grad_norm": 0.9044458866119385, "learning_rate": 0.0001998976677298414, "loss": 4.8151, "step": 65850 }, { "epoch": 0.12297672726085201, "grad_norm": 0.8736430406570435, "learning_rate": 0.00019989747763850865, "loss": 4.6468, "step": 65900 }, { "epoch": 0.12307003282023052, "grad_norm": 1.0747185945510864, "learning_rate": 0.0001998972873708744, "loss": 4.8394, "step": 65950 }, { "epoch": 0.12316333837960901, "grad_norm": 1.0368112325668335, "learning_rate": 0.00019989709692693906, "loss": 4.7296, "step": 66000 }, { "epoch": 0.12316333837960901, "eval_loss": 4.993486404418945, "eval_runtime": 231.2413, "eval_samples_per_second": 11.278, "eval_steps_per_second": 11.278, "eval_tts_loss": 7.411762504489875, "step": 66000 }, { "epoch": 0.1232566439389875, "grad_norm": 1.0206661224365234, "learning_rate": 0.0001998969063067029, "loss": 4.8384, "step": 66050 }, { "epoch": 0.12334994949836599, "grad_norm": 1.2967489957809448, "learning_rate": 0.00019989671551016625, "loss": 4.6992, "step": 66100 }, { "epoch": 0.12344325505774448, "grad_norm": 0.8370565176010132, "learning_rate": 0.00019989652453732948, "loss": 4.7676, "step": 66150 }, { "epoch": 0.12353656061712297, "grad_norm": 1.1219327449798584, "learning_rate": 0.00019989633338819295, "loss": 4.913, "step": 66200 }, { "epoch": 0.12362986617650146, "grad_norm": 1.0357359647750854, "learning_rate": 0.00019989614206275696, "loss": 4.7011, "step": 66250 }, { "epoch": 0.12372317173587995, "grad_norm": 1.0418895483016968, "learning_rate": 0.00019989595056102183, "loss": 4.858, "step": 66300 }, { "epoch": 0.12381647729525845, "grad_norm": 0.9631625413894653, "learning_rate": 0.00019989575888298795, "loss": 4.9325, "step": 66350 }, { "epoch": 0.12390978285463694, "grad_norm": 1.2581260204315186, "learning_rate": 0.00019989556702865565, "loss": 4.7477, "step": 66400 }, { "epoch": 0.12400308841401543, "grad_norm": 0.7702711224555969, "learning_rate": 0.0001998953749980252, "loss": 4.9714, "step": 66450 }, { "epoch": 0.12409639397339392, "grad_norm": 1.0711308717727661, "learning_rate": 0.00019989518279109705, "loss": 4.836, "step": 66500 }, { "epoch": 0.12418969953277241, "grad_norm": 1.082642912864685, "learning_rate": 0.00019989499040787146, "loss": 4.7201, "step": 66550 }, { "epoch": 0.1242830050921509, "grad_norm": 1.1112911701202393, "learning_rate": 0.00019989479784834877, "loss": 4.6956, "step": 66600 }, { "epoch": 0.12437631065152939, "grad_norm": 0.9402649998664856, "learning_rate": 0.00019989460511252936, "loss": 4.6172, "step": 66650 }, { "epoch": 0.12446961621090788, "grad_norm": 1.1066251993179321, "learning_rate": 0.00019989441220041353, "loss": 4.7511, "step": 66700 }, { "epoch": 0.12456292177028638, "grad_norm": 1.2885663509368896, "learning_rate": 0.00019989421911200162, "loss": 4.8006, "step": 66750 }, { "epoch": 0.12465622732966487, "grad_norm": 0.717598557472229, "learning_rate": 0.00019989402584729405, "loss": 4.6387, "step": 66800 }, { "epoch": 0.12474953288904336, "grad_norm": 1.0635030269622803, "learning_rate": 0.00019989383240629106, "loss": 4.6984, "step": 66850 }, { "epoch": 0.12484283844842185, "grad_norm": 1.0408589839935303, "learning_rate": 0.00019989363878899306, "loss": 4.696, "step": 66900 }, { "epoch": 0.12493614400780034, "grad_norm": 0.8848728537559509, "learning_rate": 0.00019989344499540032, "loss": 4.6144, "step": 66950 }, { "epoch": 0.12502944956717885, "grad_norm": 0.9865880608558655, "learning_rate": 0.00019989325102551326, "loss": 4.8072, "step": 67000 }, { "epoch": 0.12512275512655732, "grad_norm": 2.2153608798980713, "learning_rate": 0.00019989305687933217, "loss": 4.7434, "step": 67050 }, { "epoch": 0.12521606068593583, "grad_norm": 0.9598948955535889, "learning_rate": 0.00019989286255685742, "loss": 4.7729, "step": 67100 }, { "epoch": 0.1253093662453143, "grad_norm": 1.0703397989273071, "learning_rate": 0.00019989266805808936, "loss": 4.9212, "step": 67150 }, { "epoch": 0.1254026718046928, "grad_norm": 0.954094648361206, "learning_rate": 0.0001998924733830283, "loss": 4.7929, "step": 67200 }, { "epoch": 0.12549597736407128, "grad_norm": 0.7662084102630615, "learning_rate": 0.0001998922785316746, "loss": 4.9457, "step": 67250 }, { "epoch": 0.12558928292344979, "grad_norm": 1.1596436500549316, "learning_rate": 0.0001998920835040286, "loss": 4.658, "step": 67300 }, { "epoch": 0.1256825884828283, "grad_norm": 1.0783514976501465, "learning_rate": 0.00019989188830009068, "loss": 4.9466, "step": 67350 }, { "epoch": 0.12577589404220677, "grad_norm": 0.8832924365997314, "learning_rate": 0.00019989169291986114, "loss": 4.7456, "step": 67400 }, { "epoch": 0.12586919960158527, "grad_norm": 0.9739583730697632, "learning_rate": 0.00019989149736334032, "loss": 4.9385, "step": 67450 }, { "epoch": 0.12596250516096374, "grad_norm": 1.0089017152786255, "learning_rate": 0.0001998913016305286, "loss": 4.6565, "step": 67500 }, { "epoch": 0.12605581072034225, "grad_norm": 0.9786615371704102, "learning_rate": 0.0001998911057214263, "loss": 4.7881, "step": 67550 }, { "epoch": 0.12614911627972072, "grad_norm": 0.9184353351593018, "learning_rate": 0.00019989090963603377, "loss": 4.779, "step": 67600 }, { "epoch": 0.12624242183909923, "grad_norm": 0.8026026487350464, "learning_rate": 0.00019989071337435136, "loss": 4.6157, "step": 67650 }, { "epoch": 0.12633572739847773, "grad_norm": 0.8393520712852478, "learning_rate": 0.00019989051693637943, "loss": 4.5899, "step": 67700 }, { "epoch": 0.1264290329578562, "grad_norm": 0.9405346512794495, "learning_rate": 0.00019989032032211834, "loss": 4.5016, "step": 67750 }, { "epoch": 0.1265223385172347, "grad_norm": 1.0097512006759644, "learning_rate": 0.00019989012353156834, "loss": 4.8901, "step": 67800 }, { "epoch": 0.1266156440766132, "grad_norm": 1.1094014644622803, "learning_rate": 0.0001998899265647299, "loss": 4.8235, "step": 67850 }, { "epoch": 0.1267089496359917, "grad_norm": 1.045980453491211, "learning_rate": 0.00019988972942160333, "loss": 4.8714, "step": 67900 }, { "epoch": 0.12680225519537017, "grad_norm": 0.897870659828186, "learning_rate": 0.0001998895321021889, "loss": 4.78, "step": 67950 }, { "epoch": 0.12689556075474867, "grad_norm": 0.8801417946815491, "learning_rate": 0.0001998893346064871, "loss": 4.7189, "step": 68000 }, { "epoch": 0.12698886631412717, "grad_norm": 0.8801950812339783, "learning_rate": 0.00019988913693449816, "loss": 4.817, "step": 68050 }, { "epoch": 0.12708217187350565, "grad_norm": 1.0675042867660522, "learning_rate": 0.00019988893908622245, "loss": 4.9578, "step": 68100 }, { "epoch": 0.12717547743288415, "grad_norm": 0.8822778463363647, "learning_rate": 0.00019988874106166037, "loss": 4.9018, "step": 68150 }, { "epoch": 0.12726878299226263, "grad_norm": 0.8060069680213928, "learning_rate": 0.00019988854286081225, "loss": 4.6307, "step": 68200 }, { "epoch": 0.12736208855164113, "grad_norm": 1.0484051704406738, "learning_rate": 0.00019988834448367838, "loss": 4.928, "step": 68250 }, { "epoch": 0.1274553941110196, "grad_norm": 1.0150614976882935, "learning_rate": 0.0001998881459302592, "loss": 4.7119, "step": 68300 }, { "epoch": 0.1275486996703981, "grad_norm": 0.8359283208847046, "learning_rate": 0.000199887947200555, "loss": 4.6128, "step": 68350 }, { "epoch": 0.1276420052297766, "grad_norm": 0.9599153399467468, "learning_rate": 0.00019988774829456617, "loss": 4.7391, "step": 68400 }, { "epoch": 0.1277353107891551, "grad_norm": 1.055501103401184, "learning_rate": 0.000199887549212293, "loss": 4.7485, "step": 68450 }, { "epoch": 0.1278286163485336, "grad_norm": 0.6903030872344971, "learning_rate": 0.0001998873499537359, "loss": 4.8311, "step": 68500 }, { "epoch": 0.12792192190791207, "grad_norm": 1.1893553733825684, "learning_rate": 0.0001998871505188952, "loss": 4.8558, "step": 68550 }, { "epoch": 0.12801522746729058, "grad_norm": 1.0602223873138428, "learning_rate": 0.00019988695090777126, "loss": 4.8572, "step": 68600 }, { "epoch": 0.12810853302666905, "grad_norm": 1.0478551387786865, "learning_rate": 0.00019988675112036442, "loss": 4.8788, "step": 68650 }, { "epoch": 0.12820183858604756, "grad_norm": 0.6287881731987, "learning_rate": 0.00019988655115667507, "loss": 4.8575, "step": 68700 }, { "epoch": 0.12829514414542603, "grad_norm": 1.2025055885314941, "learning_rate": 0.0001998863510167035, "loss": 4.5946, "step": 68750 }, { "epoch": 0.12838844970480454, "grad_norm": 1.10360586643219, "learning_rate": 0.0001998861507004501, "loss": 4.8477, "step": 68800 }, { "epoch": 0.12848175526418304, "grad_norm": 1.1787991523742676, "learning_rate": 0.00019988595020791524, "loss": 4.5538, "step": 68850 }, { "epoch": 0.12857506082356152, "grad_norm": 1.1096701622009277, "learning_rate": 0.00019988574953909923, "loss": 4.8392, "step": 68900 }, { "epoch": 0.12866836638294002, "grad_norm": 1.0045229196548462, "learning_rate": 0.00019988554869400247, "loss": 4.8437, "step": 68950 }, { "epoch": 0.1287616719423185, "grad_norm": 0.7254719734191895, "learning_rate": 0.00019988534767262527, "loss": 4.6926, "step": 69000 }, { "epoch": 0.1287616719423185, "eval_loss": 4.980347156524658, "eval_runtime": 233.3792, "eval_samples_per_second": 11.175, "eval_steps_per_second": 11.175, "eval_tts_loss": 7.437576909478114, "step": 69000 }, { "epoch": 0.128854977501697, "grad_norm": 1.043537974357605, "learning_rate": 0.00019988514647496802, "loss": 4.7274, "step": 69050 }, { "epoch": 0.12894828306107547, "grad_norm": 1.02115797996521, "learning_rate": 0.00019988494510103104, "loss": 4.7884, "step": 69100 }, { "epoch": 0.12904158862045398, "grad_norm": 1.1048721075057983, "learning_rate": 0.00019988474355081475, "loss": 4.4432, "step": 69150 }, { "epoch": 0.12913489417983248, "grad_norm": 0.9006786942481995, "learning_rate": 0.0001998845418243194, "loss": 4.703, "step": 69200 }, { "epoch": 0.12922819973921096, "grad_norm": 0.9208207130432129, "learning_rate": 0.00019988433992154545, "loss": 4.6751, "step": 69250 }, { "epoch": 0.12932150529858946, "grad_norm": 0.9180870652198792, "learning_rate": 0.00019988413784249322, "loss": 4.5272, "step": 69300 }, { "epoch": 0.12941481085796794, "grad_norm": 1.102773904800415, "learning_rate": 0.00019988393558716306, "loss": 4.7114, "step": 69350 }, { "epoch": 0.12950811641734644, "grad_norm": 0.9914569854736328, "learning_rate": 0.00019988373315555532, "loss": 4.8268, "step": 69400 }, { "epoch": 0.12960142197672492, "grad_norm": 0.9935908913612366, "learning_rate": 0.00019988353054767037, "loss": 4.8687, "step": 69450 }, { "epoch": 0.12969472753610342, "grad_norm": 0.8525168895721436, "learning_rate": 0.00019988332776350856, "loss": 4.7935, "step": 69500 }, { "epoch": 0.12978803309548193, "grad_norm": 0.8857572674751282, "learning_rate": 0.00019988312480307026, "loss": 4.7981, "step": 69550 }, { "epoch": 0.1298813386548604, "grad_norm": 1.1465469598770142, "learning_rate": 0.0001998829216663558, "loss": 4.6701, "step": 69600 }, { "epoch": 0.1299746442142389, "grad_norm": 0.91568523645401, "learning_rate": 0.00019988271835336558, "loss": 4.7749, "step": 69650 }, { "epoch": 0.13006794977361738, "grad_norm": 0.9274788498878479, "learning_rate": 0.00019988251486409992, "loss": 4.6415, "step": 69700 }, { "epoch": 0.13016125533299588, "grad_norm": 1.1520894765853882, "learning_rate": 0.0001998823111985592, "loss": 4.5849, "step": 69750 }, { "epoch": 0.13025456089237436, "grad_norm": 1.3699066638946533, "learning_rate": 0.0001998821073567438, "loss": 4.7611, "step": 69800 }, { "epoch": 0.13034786645175286, "grad_norm": 0.9259943962097168, "learning_rate": 0.00019988190333865402, "loss": 4.8313, "step": 69850 }, { "epoch": 0.13044117201113134, "grad_norm": 0.8618115782737732, "learning_rate": 0.00019988169914429026, "loss": 4.6103, "step": 69900 }, { "epoch": 0.13053447757050984, "grad_norm": 1.0242559909820557, "learning_rate": 0.0001998814947736529, "loss": 4.9478, "step": 69950 }, { "epoch": 0.13062778312988835, "grad_norm": 1.2048753499984741, "learning_rate": 0.00019988129022674228, "loss": 4.9015, "step": 70000 }, { "epoch": 0.13072108868926682, "grad_norm": 0.8471056818962097, "learning_rate": 0.0001998810855035587, "loss": 4.5826, "step": 70050 }, { "epoch": 0.13081439424864533, "grad_norm": 0.9561209082603455, "learning_rate": 0.00019988088060410265, "loss": 4.7941, "step": 70100 }, { "epoch": 0.1309076998080238, "grad_norm": 0.6860484480857849, "learning_rate": 0.00019988067552837439, "loss": 4.6454, "step": 70150 }, { "epoch": 0.1310010053674023, "grad_norm": 1.028444528579712, "learning_rate": 0.0001998804702763743, "loss": 4.8362, "step": 70200 }, { "epoch": 0.13109431092678078, "grad_norm": 0.9688596129417419, "learning_rate": 0.00019988026484810275, "loss": 4.8212, "step": 70250 }, { "epoch": 0.1311876164861593, "grad_norm": 1.0312488079071045, "learning_rate": 0.00019988005924356013, "loss": 4.8123, "step": 70300 }, { "epoch": 0.1312809220455378, "grad_norm": 0.8787838220596313, "learning_rate": 0.0001998798534627468, "loss": 4.6147, "step": 70350 }, { "epoch": 0.13137422760491627, "grad_norm": 1.1073654890060425, "learning_rate": 0.00019987964750566305, "loss": 4.7933, "step": 70400 }, { "epoch": 0.13146753316429477, "grad_norm": 1.085288643836975, "learning_rate": 0.00019987944137230934, "loss": 4.8248, "step": 70450 }, { "epoch": 0.13156083872367325, "grad_norm": 1.2170988321304321, "learning_rate": 0.00019987923506268595, "loss": 4.783, "step": 70500 }, { "epoch": 0.13165414428305175, "grad_norm": 1.0648763179779053, "learning_rate": 0.00019987902857679333, "loss": 4.9425, "step": 70550 }, { "epoch": 0.13174744984243023, "grad_norm": 1.1931159496307373, "learning_rate": 0.00019987882191463177, "loss": 4.6804, "step": 70600 }, { "epoch": 0.13184075540180873, "grad_norm": 0.9888035655021667, "learning_rate": 0.00019987861507620168, "loss": 4.7612, "step": 70650 }, { "epoch": 0.13193406096118723, "grad_norm": 0.9140545725822449, "learning_rate": 0.0001998784080615034, "loss": 4.6891, "step": 70700 }, { "epoch": 0.1320273665205657, "grad_norm": 1.096142053604126, "learning_rate": 0.00019987820087053728, "loss": 4.9494, "step": 70750 }, { "epoch": 0.1321206720799442, "grad_norm": 1.1193305253982544, "learning_rate": 0.00019987799350330374, "loss": 4.8832, "step": 70800 }, { "epoch": 0.1322139776393227, "grad_norm": 1.2808094024658203, "learning_rate": 0.0001998777859598031, "loss": 4.9684, "step": 70850 }, { "epoch": 0.1323072831987012, "grad_norm": 0.852861225605011, "learning_rate": 0.00019987757824003577, "loss": 4.911, "step": 70900 }, { "epoch": 0.13240058875807967, "grad_norm": 0.8314254283905029, "learning_rate": 0.00019987737034400204, "loss": 4.6268, "step": 70950 }, { "epoch": 0.13249389431745817, "grad_norm": 1.0888328552246094, "learning_rate": 0.00019987716227170234, "loss": 4.837, "step": 71000 }, { "epoch": 0.13258719987683665, "grad_norm": 0.9562662243843079, "learning_rate": 0.00019987695402313706, "loss": 4.8852, "step": 71050 }, { "epoch": 0.13268050543621515, "grad_norm": 1.0483977794647217, "learning_rate": 0.00019987674559830652, "loss": 4.6314, "step": 71100 }, { "epoch": 0.13277381099559366, "grad_norm": 1.0416483879089355, "learning_rate": 0.00019987653699721106, "loss": 4.6485, "step": 71150 }, { "epoch": 0.13286711655497213, "grad_norm": 0.9139002561569214, "learning_rate": 0.00019987632821985113, "loss": 4.4885, "step": 71200 }, { "epoch": 0.13296042211435063, "grad_norm": 1.0126796960830688, "learning_rate": 0.00019987611926622703, "loss": 4.7368, "step": 71250 }, { "epoch": 0.1330537276737291, "grad_norm": 0.8071941137313843, "learning_rate": 0.00019987591013633912, "loss": 4.7529, "step": 71300 }, { "epoch": 0.13314703323310761, "grad_norm": 0.8318772912025452, "learning_rate": 0.00019987570083018788, "loss": 4.686, "step": 71350 }, { "epoch": 0.1332403387924861, "grad_norm": 1.0917198657989502, "learning_rate": 0.00019987549134777355, "loss": 4.6129, "step": 71400 }, { "epoch": 0.1333336443518646, "grad_norm": 1.0430625677108765, "learning_rate": 0.00019987528168909655, "loss": 4.5787, "step": 71450 }, { "epoch": 0.1334269499112431, "grad_norm": 1.307949423789978, "learning_rate": 0.00019987507185415727, "loss": 4.7455, "step": 71500 }, { "epoch": 0.13352025547062157, "grad_norm": 0.9685766696929932, "learning_rate": 0.00019987486184295606, "loss": 4.7564, "step": 71550 }, { "epoch": 0.13361356103000008, "grad_norm": 0.9537215232849121, "learning_rate": 0.00019987465165549327, "loss": 4.7439, "step": 71600 }, { "epoch": 0.13370686658937855, "grad_norm": 0.8062765598297119, "learning_rate": 0.00019987444129176933, "loss": 4.6103, "step": 71650 }, { "epoch": 0.13380017214875706, "grad_norm": 1.0107309818267822, "learning_rate": 0.00019987423075178455, "loss": 4.8355, "step": 71700 }, { "epoch": 0.13389347770813553, "grad_norm": 0.8485686779022217, "learning_rate": 0.00019987402003553934, "loss": 4.7971, "step": 71750 }, { "epoch": 0.13398678326751404, "grad_norm": 1.2249001264572144, "learning_rate": 0.00019987380914303403, "loss": 4.823, "step": 71800 }, { "epoch": 0.13408008882689254, "grad_norm": 1.0090044736862183, "learning_rate": 0.00019987359807426905, "loss": 4.9343, "step": 71850 }, { "epoch": 0.13417339438627102, "grad_norm": 1.018228530883789, "learning_rate": 0.00019987338682924476, "loss": 4.8365, "step": 71900 }, { "epoch": 0.13426669994564952, "grad_norm": 0.947043240070343, "learning_rate": 0.00019987317540796147, "loss": 4.7334, "step": 71950 }, { "epoch": 0.134360005505028, "grad_norm": 1.1589524745941162, "learning_rate": 0.00019987296381041963, "loss": 4.6444, "step": 72000 }, { "epoch": 0.134360005505028, "eval_loss": 4.982792377471924, "eval_runtime": 233.4346, "eval_samples_per_second": 11.172, "eval_steps_per_second": 11.172, "eval_tts_loss": 7.417768215367777, "step": 72000 }, { "epoch": 0.1344533110644065, "grad_norm": 1.0631495714187622, "learning_rate": 0.00019987275203661956, "loss": 4.8445, "step": 72050 }, { "epoch": 0.13454661662378498, "grad_norm": 1.056353211402893, "learning_rate": 0.00019987254008656165, "loss": 4.8737, "step": 72100 }, { "epoch": 0.13463992218316348, "grad_norm": 1.192908525466919, "learning_rate": 0.0001998723279602463, "loss": 4.7873, "step": 72150 }, { "epoch": 0.13473322774254198, "grad_norm": 1.1714168787002563, "learning_rate": 0.00019987211565767384, "loss": 4.7542, "step": 72200 }, { "epoch": 0.13482653330192046, "grad_norm": 0.7499253153800964, "learning_rate": 0.00019987190317884469, "loss": 4.8558, "step": 72250 }, { "epoch": 0.13491983886129896, "grad_norm": 1.1434695720672607, "learning_rate": 0.0001998716905237592, "loss": 4.684, "step": 72300 }, { "epoch": 0.13501314442067744, "grad_norm": 0.9762905836105347, "learning_rate": 0.00019987147769241775, "loss": 4.6689, "step": 72350 }, { "epoch": 0.13510644998005594, "grad_norm": 1.05728018283844, "learning_rate": 0.00019987126468482073, "loss": 4.7129, "step": 72400 }, { "epoch": 0.13519975553943442, "grad_norm": 1.0053168535232544, "learning_rate": 0.00019987105150096849, "loss": 4.6524, "step": 72450 }, { "epoch": 0.13529306109881292, "grad_norm": 0.9576205611228943, "learning_rate": 0.0001998708381408614, "loss": 4.6339, "step": 72500 }, { "epoch": 0.1353863666581914, "grad_norm": 1.0552775859832764, "learning_rate": 0.00019987062460449984, "loss": 4.6316, "step": 72550 }, { "epoch": 0.1354796722175699, "grad_norm": 1.1562297344207764, "learning_rate": 0.00019987041089188426, "loss": 4.7064, "step": 72600 }, { "epoch": 0.1355729777769484, "grad_norm": 1.0458017587661743, "learning_rate": 0.00019987019700301494, "loss": 4.8765, "step": 72650 }, { "epoch": 0.13566628333632688, "grad_norm": 0.8299524784088135, "learning_rate": 0.00019986998293789233, "loss": 4.8115, "step": 72700 }, { "epoch": 0.13575958889570539, "grad_norm": 1.1313036680221558, "learning_rate": 0.0001998697686965167, "loss": 4.7096, "step": 72750 }, { "epoch": 0.13585289445508386, "grad_norm": 1.0760210752487183, "learning_rate": 0.00019986955427888855, "loss": 4.886, "step": 72800 }, { "epoch": 0.13594620001446237, "grad_norm": 1.1115272045135498, "learning_rate": 0.0001998693396850082, "loss": 4.8623, "step": 72850 }, { "epoch": 0.13603950557384084, "grad_norm": 0.6817601919174194, "learning_rate": 0.00019986912491487603, "loss": 4.7888, "step": 72900 }, { "epoch": 0.13613281113321934, "grad_norm": 1.2269861698150635, "learning_rate": 0.00019986890996849245, "loss": 4.8229, "step": 72950 }, { "epoch": 0.13622611669259785, "grad_norm": 1.13284170627594, "learning_rate": 0.00019986869484585782, "loss": 4.7043, "step": 73000 }, { "epoch": 0.13631942225197632, "grad_norm": 1.085242748260498, "learning_rate": 0.0001998684795469725, "loss": 4.8635, "step": 73050 }, { "epoch": 0.13641272781135483, "grad_norm": 0.9383891224861145, "learning_rate": 0.00019986826407183687, "loss": 4.7206, "step": 73100 }, { "epoch": 0.1365060333707333, "grad_norm": 0.9130173921585083, "learning_rate": 0.00019986804842045134, "loss": 4.783, "step": 73150 }, { "epoch": 0.1365993389301118, "grad_norm": 0.9875626564025879, "learning_rate": 0.0001998678325928163, "loss": 4.6706, "step": 73200 }, { "epoch": 0.13669264448949028, "grad_norm": 0.775663435459137, "learning_rate": 0.00019986761658893207, "loss": 4.5753, "step": 73250 }, { "epoch": 0.1367859500488688, "grad_norm": 0.9947422742843628, "learning_rate": 0.00019986740040879908, "loss": 4.59, "step": 73300 }, { "epoch": 0.1368792556082473, "grad_norm": 0.7354251742362976, "learning_rate": 0.00019986718405241773, "loss": 4.5268, "step": 73350 }, { "epoch": 0.13697256116762577, "grad_norm": 0.9858702421188354, "learning_rate": 0.00019986696751978834, "loss": 4.7284, "step": 73400 }, { "epoch": 0.13706586672700427, "grad_norm": 1.1084253787994385, "learning_rate": 0.00019986675081091134, "loss": 4.864, "step": 73450 }, { "epoch": 0.13715917228638275, "grad_norm": 1.0216785669326782, "learning_rate": 0.00019986653392578706, "loss": 4.6343, "step": 73500 }, { "epoch": 0.13725247784576125, "grad_norm": 1.0096267461776733, "learning_rate": 0.00019986631686441596, "loss": 4.7856, "step": 73550 }, { "epoch": 0.13734578340513973, "grad_norm": 0.9455108642578125, "learning_rate": 0.0001998660996267984, "loss": 4.6775, "step": 73600 }, { "epoch": 0.13743908896451823, "grad_norm": 1.4153708219528198, "learning_rate": 0.0001998658822129347, "loss": 4.6638, "step": 73650 }, { "epoch": 0.13753239452389673, "grad_norm": 0.8518409729003906, "learning_rate": 0.00019986566462282533, "loss": 4.9213, "step": 73700 }, { "epoch": 0.1376257000832752, "grad_norm": 1.1150931119918823, "learning_rate": 0.0001998654468564706, "loss": 4.7676, "step": 73750 }, { "epoch": 0.1377190056426537, "grad_norm": 0.9590360522270203, "learning_rate": 0.00019986522891387093, "loss": 4.9231, "step": 73800 }, { "epoch": 0.1378123112020322, "grad_norm": 0.7029619216918945, "learning_rate": 0.0001998650107950267, "loss": 4.692, "step": 73850 }, { "epoch": 0.1379056167614107, "grad_norm": 0.8161832690238953, "learning_rate": 0.00019986479249993832, "loss": 4.7669, "step": 73900 }, { "epoch": 0.13799892232078917, "grad_norm": 1.2208623886108398, "learning_rate": 0.00019986457402860618, "loss": 4.774, "step": 73950 }, { "epoch": 0.13809222788016767, "grad_norm": 0.6693832278251648, "learning_rate": 0.00019986435538103058, "loss": 4.6447, "step": 74000 }, { "epoch": 0.13818553343954615, "grad_norm": 1.1013367176055908, "learning_rate": 0.000199864136557212, "loss": 4.9094, "step": 74050 }, { "epoch": 0.13827883899892465, "grad_norm": 0.6470247507095337, "learning_rate": 0.00019986391755715075, "loss": 4.7323, "step": 74100 }, { "epoch": 0.13837214455830316, "grad_norm": 1.0333960056304932, "learning_rate": 0.0001998636983808473, "loss": 4.6899, "step": 74150 }, { "epoch": 0.13846545011768163, "grad_norm": 1.0309245586395264, "learning_rate": 0.00019986347902830196, "loss": 4.8887, "step": 74200 }, { "epoch": 0.13855875567706014, "grad_norm": 1.0553001165390015, "learning_rate": 0.00019986325949951516, "loss": 4.654, "step": 74250 }, { "epoch": 0.1386520612364386, "grad_norm": 1.3011856079101562, "learning_rate": 0.0001998630397944873, "loss": 4.8259, "step": 74300 }, { "epoch": 0.13874536679581712, "grad_norm": 1.04082190990448, "learning_rate": 0.00019986281991321873, "loss": 4.9014, "step": 74350 }, { "epoch": 0.1388386723551956, "grad_norm": 0.9634092450141907, "learning_rate": 0.00019986259985570987, "loss": 4.7242, "step": 74400 }, { "epoch": 0.1389319779145741, "grad_norm": 1.2973626852035522, "learning_rate": 0.00019986237962196107, "loss": 4.7606, "step": 74450 }, { "epoch": 0.1390252834739526, "grad_norm": 0.8108509182929993, "learning_rate": 0.00019986215921197277, "loss": 4.7773, "step": 74500 }, { "epoch": 0.13911858903333107, "grad_norm": 1.048374891281128, "learning_rate": 0.0001998619386257453, "loss": 4.8176, "step": 74550 }, { "epoch": 0.13921189459270958, "grad_norm": 0.9757275581359863, "learning_rate": 0.0001998617178632791, "loss": 4.5193, "step": 74600 }, { "epoch": 0.13930520015208805, "grad_norm": 1.088984727859497, "learning_rate": 0.00019986149692457453, "loss": 4.738, "step": 74650 }, { "epoch": 0.13939850571146656, "grad_norm": 1.0206042528152466, "learning_rate": 0.00019986127580963199, "loss": 4.7663, "step": 74700 }, { "epoch": 0.13949181127084503, "grad_norm": 0.859443724155426, "learning_rate": 0.00019986105451845186, "loss": 4.7207, "step": 74750 }, { "epoch": 0.13958511683022354, "grad_norm": 0.9786943793296814, "learning_rate": 0.00019986083305103457, "loss": 4.8639, "step": 74800 }, { "epoch": 0.13967842238960204, "grad_norm": 1.2023348808288574, "learning_rate": 0.00019986061140738043, "loss": 4.7091, "step": 74850 }, { "epoch": 0.13977172794898052, "grad_norm": 1.1484274864196777, "learning_rate": 0.00019986038958748993, "loss": 4.6658, "step": 74900 }, { "epoch": 0.13986503350835902, "grad_norm": 1.0033177137374878, "learning_rate": 0.0001998601675913634, "loss": 4.8025, "step": 74950 }, { "epoch": 0.1399583390677375, "grad_norm": 1.1093839406967163, "learning_rate": 0.00019985994541900124, "loss": 4.7705, "step": 75000 }, { "epoch": 0.1399583390677375, "eval_loss": 4.963533878326416, "eval_runtime": 229.2936, "eval_samples_per_second": 11.374, "eval_steps_per_second": 11.374, "eval_tts_loss": 7.357835572455764, "step": 75000 }, { "epoch": 0.140051644627116, "grad_norm": 0.8389537930488586, "learning_rate": 0.00019985972307040383, "loss": 4.8843, "step": 75050 }, { "epoch": 0.14014495018649448, "grad_norm": 1.009690761566162, "learning_rate": 0.00019985950054557164, "loss": 4.6147, "step": 75100 }, { "epoch": 0.14023825574587298, "grad_norm": 1.0867621898651123, "learning_rate": 0.00019985927784450493, "loss": 4.7178, "step": 75150 }, { "epoch": 0.14033156130525146, "grad_norm": 0.7208017706871033, "learning_rate": 0.0001998590549672042, "loss": 4.918, "step": 75200 }, { "epoch": 0.14042486686462996, "grad_norm": 0.9712778925895691, "learning_rate": 0.0001998588319136698, "loss": 4.8301, "step": 75250 }, { "epoch": 0.14051817242400846, "grad_norm": 0.7589443922042847, "learning_rate": 0.00019985860868390213, "loss": 4.7432, "step": 75300 }, { "epoch": 0.14061147798338694, "grad_norm": 0.7034851312637329, "learning_rate": 0.0001998583852779016, "loss": 4.6202, "step": 75350 }, { "epoch": 0.14070478354276544, "grad_norm": 1.2004735469818115, "learning_rate": 0.0001998581616956686, "loss": 4.4097, "step": 75400 }, { "epoch": 0.14079808910214392, "grad_norm": 1.1603310108184814, "learning_rate": 0.0001998579379372035, "loss": 4.8106, "step": 75450 }, { "epoch": 0.14089139466152242, "grad_norm": 0.9474809169769287, "learning_rate": 0.00019985771400250668, "loss": 4.6189, "step": 75500 }, { "epoch": 0.1409847002209009, "grad_norm": 0.9576531052589417, "learning_rate": 0.0001998574898915786, "loss": 4.6265, "step": 75550 }, { "epoch": 0.1410780057802794, "grad_norm": 0.7832580804824829, "learning_rate": 0.00019985726560441962, "loss": 4.656, "step": 75600 }, { "epoch": 0.1411713113396579, "grad_norm": 1.3878700733184814, "learning_rate": 0.00019985704114103015, "loss": 4.5926, "step": 75650 }, { "epoch": 0.14126461689903638, "grad_norm": 0.8612073659896851, "learning_rate": 0.00019985681650141053, "loss": 4.8437, "step": 75700 }, { "epoch": 0.1413579224584149, "grad_norm": 1.2866979837417603, "learning_rate": 0.00019985659168556124, "loss": 4.9809, "step": 75750 }, { "epoch": 0.14145122801779336, "grad_norm": 1.1674981117248535, "learning_rate": 0.0001998563666934826, "loss": 4.7633, "step": 75800 }, { "epoch": 0.14154453357717187, "grad_norm": 1.0047423839569092, "learning_rate": 0.0001998561415251751, "loss": 4.6555, "step": 75850 }, { "epoch": 0.14163783913655034, "grad_norm": 0.8399094343185425, "learning_rate": 0.00019985591618063903, "loss": 5.0474, "step": 75900 }, { "epoch": 0.14173114469592885, "grad_norm": 1.3186641931533813, "learning_rate": 0.00019985569065987484, "loss": 4.7141, "step": 75950 }, { "epoch": 0.14182445025530735, "grad_norm": 1.2126065492630005, "learning_rate": 0.00019985546496288295, "loss": 4.5208, "step": 76000 }, { "epoch": 0.14191775581468583, "grad_norm": 1.0247917175292969, "learning_rate": 0.0001998552390896637, "loss": 4.6857, "step": 76050 }, { "epoch": 0.14201106137406433, "grad_norm": 0.8075127005577087, "learning_rate": 0.00019985501304021756, "loss": 4.6593, "step": 76100 }, { "epoch": 0.1421043669334428, "grad_norm": 0.8528395295143127, "learning_rate": 0.00019985478681454487, "loss": 4.7271, "step": 76150 }, { "epoch": 0.1421976724928213, "grad_norm": 0.9807457327842712, "learning_rate": 0.00019985456041264605, "loss": 4.5915, "step": 76200 }, { "epoch": 0.14229097805219978, "grad_norm": 1.1197192668914795, "learning_rate": 0.0001998543338345215, "loss": 4.4433, "step": 76250 }, { "epoch": 0.1423842836115783, "grad_norm": 1.044595718383789, "learning_rate": 0.00019985410708017165, "loss": 4.7313, "step": 76300 }, { "epoch": 0.1424775891709568, "grad_norm": 0.8226972818374634, "learning_rate": 0.00019985388014959684, "loss": 4.5888, "step": 76350 }, { "epoch": 0.14257089473033527, "grad_norm": 0.9106892943382263, "learning_rate": 0.00019985365304279752, "loss": 4.7119, "step": 76400 }, { "epoch": 0.14266420028971377, "grad_norm": 1.2753225564956665, "learning_rate": 0.00019985342575977406, "loss": 4.7376, "step": 76450 }, { "epoch": 0.14275750584909225, "grad_norm": 0.9028472304344177, "learning_rate": 0.00019985319830052688, "loss": 4.6413, "step": 76500 }, { "epoch": 0.14285081140847075, "grad_norm": 1.0299334526062012, "learning_rate": 0.00019985297066505637, "loss": 4.6844, "step": 76550 }, { "epoch": 0.14294411696784923, "grad_norm": 1.1871273517608643, "learning_rate": 0.00019985274285336293, "loss": 4.8803, "step": 76600 }, { "epoch": 0.14303742252722773, "grad_norm": 0.5859182476997375, "learning_rate": 0.000199852514865447, "loss": 4.627, "step": 76650 }, { "epoch": 0.1431307280866062, "grad_norm": 0.8444706201553345, "learning_rate": 0.00019985228670130894, "loss": 4.7149, "step": 76700 }, { "epoch": 0.1432240336459847, "grad_norm": 0.889392614364624, "learning_rate": 0.0001998520583609491, "loss": 4.6281, "step": 76750 }, { "epoch": 0.14331733920536321, "grad_norm": 0.9411704540252686, "learning_rate": 0.00019985182984436802, "loss": 4.6086, "step": 76800 }, { "epoch": 0.1434106447647417, "grad_norm": 0.8570734262466431, "learning_rate": 0.000199851601151566, "loss": 4.7565, "step": 76850 }, { "epoch": 0.1435039503241202, "grad_norm": 0.8784626722335815, "learning_rate": 0.0001998513722825435, "loss": 4.891, "step": 76900 }, { "epoch": 0.14359725588349867, "grad_norm": 1.0024194717407227, "learning_rate": 0.00019985114323730087, "loss": 4.7093, "step": 76950 }, { "epoch": 0.14369056144287717, "grad_norm": 1.1379250288009644, "learning_rate": 0.00019985091401583855, "loss": 4.8616, "step": 77000 }, { "epoch": 0.14378386700225565, "grad_norm": 0.9281609058380127, "learning_rate": 0.00019985068461815693, "loss": 4.7413, "step": 77050 }, { "epoch": 0.14387717256163415, "grad_norm": 1.4726687669754028, "learning_rate": 0.00019985045504425643, "loss": 4.7395, "step": 77100 }, { "epoch": 0.14397047812101266, "grad_norm": 1.1294130086898804, "learning_rate": 0.00019985022529413744, "loss": 4.8413, "step": 77150 }, { "epoch": 0.14406378368039113, "grad_norm": 0.7833269834518433, "learning_rate": 0.00019984999536780038, "loss": 4.7995, "step": 77200 }, { "epoch": 0.14415708923976964, "grad_norm": 0.861925482749939, "learning_rate": 0.00019984976526524567, "loss": 4.7661, "step": 77250 }, { "epoch": 0.1442503947991481, "grad_norm": 1.15879225730896, "learning_rate": 0.00019984953498647365, "loss": 4.8043, "step": 77300 }, { "epoch": 0.14434370035852662, "grad_norm": 0.854045569896698, "learning_rate": 0.0001998493045314848, "loss": 4.6388, "step": 77350 }, { "epoch": 0.1444370059179051, "grad_norm": 1.0074321031570435, "learning_rate": 0.00019984907390027944, "loss": 4.7215, "step": 77400 }, { "epoch": 0.1445303114772836, "grad_norm": 0.9455142617225647, "learning_rate": 0.0001998488430928581, "loss": 4.7325, "step": 77450 }, { "epoch": 0.1446236170366621, "grad_norm": 1.0225214958190918, "learning_rate": 0.00019984861210922109, "loss": 4.7662, "step": 77500 }, { "epoch": 0.14471692259604058, "grad_norm": 0.8790587186813354, "learning_rate": 0.00019984838094936884, "loss": 4.7582, "step": 77550 }, { "epoch": 0.14481022815541908, "grad_norm": 0.7881041169166565, "learning_rate": 0.00019984814961330178, "loss": 4.7529, "step": 77600 }, { "epoch": 0.14490353371479756, "grad_norm": 0.9009997844696045, "learning_rate": 0.00019984791810102032, "loss": 4.8476, "step": 77650 }, { "epoch": 0.14499683927417606, "grad_norm": 1.1333011388778687, "learning_rate": 0.00019984768641252484, "loss": 4.7743, "step": 77700 }, { "epoch": 0.14509014483355454, "grad_norm": 1.0094295740127563, "learning_rate": 0.00019984745454781577, "loss": 4.7856, "step": 77750 }, { "epoch": 0.14518345039293304, "grad_norm": 1.1273713111877441, "learning_rate": 0.00019984722250689348, "loss": 4.6921, "step": 77800 }, { "epoch": 0.14527675595231151, "grad_norm": 1.0432093143463135, "learning_rate": 0.00019984699028975845, "loss": 4.6715, "step": 77850 }, { "epoch": 0.14537006151169002, "grad_norm": 1.0306214094161987, "learning_rate": 0.00019984675789641102, "loss": 4.9655, "step": 77900 }, { "epoch": 0.14546336707106852, "grad_norm": 0.5953212380409241, "learning_rate": 0.00019984652532685165, "loss": 4.6442, "step": 77950 }, { "epoch": 0.145556672630447, "grad_norm": 0.900357723236084, "learning_rate": 0.00019984629258108073, "loss": 4.6404, "step": 78000 }, { "epoch": 0.145556672630447, "eval_loss": 4.966691017150879, "eval_runtime": 229.295, "eval_samples_per_second": 11.374, "eval_steps_per_second": 11.374, "eval_tts_loss": 7.41557245793541, "step": 78000 }, { "epoch": 0.1456499781898255, "grad_norm": 0.9374669194221497, "learning_rate": 0.00019984605965909865, "loss": 4.6393, "step": 78050 }, { "epoch": 0.14574328374920398, "grad_norm": 0.7869539856910706, "learning_rate": 0.00019984582656090586, "loss": 4.5879, "step": 78100 }, { "epoch": 0.14583658930858248, "grad_norm": 0.6721176505088806, "learning_rate": 0.00019984559328650278, "loss": 4.8905, "step": 78150 }, { "epoch": 0.14592989486796096, "grad_norm": 0.9904831647872925, "learning_rate": 0.00019984535983588977, "loss": 4.8523, "step": 78200 }, { "epoch": 0.14602320042733946, "grad_norm": 0.9806076884269714, "learning_rate": 0.00019984512620906727, "loss": 4.8636, "step": 78250 }, { "epoch": 0.14611650598671796, "grad_norm": 1.2187144756317139, "learning_rate": 0.0001998448924060357, "loss": 4.7684, "step": 78300 }, { "epoch": 0.14620981154609644, "grad_norm": 0.9885228276252747, "learning_rate": 0.00019984465842679545, "loss": 4.743, "step": 78350 }, { "epoch": 0.14630311710547494, "grad_norm": 0.8706178665161133, "learning_rate": 0.00019984442427134694, "loss": 4.7663, "step": 78400 }, { "epoch": 0.14639642266485342, "grad_norm": 0.9276357889175415, "learning_rate": 0.0001998441899396906, "loss": 4.9867, "step": 78450 }, { "epoch": 0.14648972822423192, "grad_norm": 1.1178739070892334, "learning_rate": 0.00019984395543182687, "loss": 4.7049, "step": 78500 }, { "epoch": 0.1465830337836104, "grad_norm": 0.9989400506019592, "learning_rate": 0.00019984372074775605, "loss": 4.6692, "step": 78550 }, { "epoch": 0.1466763393429889, "grad_norm": 1.0636882781982422, "learning_rate": 0.0001998434858874787, "loss": 4.8389, "step": 78600 }, { "epoch": 0.1467696449023674, "grad_norm": 1.0234490633010864, "learning_rate": 0.00019984325085099515, "loss": 4.7642, "step": 78650 }, { "epoch": 0.14686295046174588, "grad_norm": 0.9406095743179321, "learning_rate": 0.00019984301563830579, "loss": 4.8147, "step": 78700 }, { "epoch": 0.1469562560211244, "grad_norm": 1.0061767101287842, "learning_rate": 0.0001998427802494111, "loss": 4.8082, "step": 78750 }, { "epoch": 0.14704956158050286, "grad_norm": 0.9918909668922424, "learning_rate": 0.00019984254468431148, "loss": 4.7704, "step": 78800 }, { "epoch": 0.14714286713988137, "grad_norm": 1.0421675443649292, "learning_rate": 0.00019984230894300731, "loss": 4.916, "step": 78850 }, { "epoch": 0.14723617269925984, "grad_norm": 1.1399683952331543, "learning_rate": 0.00019984207302549907, "loss": 4.7578, "step": 78900 }, { "epoch": 0.14732947825863835, "grad_norm": 1.0678455829620361, "learning_rate": 0.0001998418369317871, "loss": 4.838, "step": 78950 }, { "epoch": 0.14742278381801685, "grad_norm": 0.9881036877632141, "learning_rate": 0.00019984160066187188, "loss": 4.8102, "step": 79000 }, { "epoch": 0.14751608937739533, "grad_norm": 1.053001880645752, "learning_rate": 0.0001998413642157538, "loss": 4.7473, "step": 79050 }, { "epoch": 0.14760939493677383, "grad_norm": 0.9560543298721313, "learning_rate": 0.00019984112759343327, "loss": 4.7536, "step": 79100 }, { "epoch": 0.1477027004961523, "grad_norm": 0.8099980354309082, "learning_rate": 0.00019984089079491075, "loss": 4.5886, "step": 79150 }, { "epoch": 0.1477960060555308, "grad_norm": 0.8475550413131714, "learning_rate": 0.0001998406538201866, "loss": 4.628, "step": 79200 }, { "epoch": 0.14788931161490929, "grad_norm": 1.233866572380066, "learning_rate": 0.00019984041666926125, "loss": 4.6666, "step": 79250 }, { "epoch": 0.1479826171742878, "grad_norm": 1.088966727256775, "learning_rate": 0.00019984017934213515, "loss": 4.7315, "step": 79300 }, { "epoch": 0.14807592273366627, "grad_norm": 1.3375529050827026, "learning_rate": 0.0001998399418388087, "loss": 4.6797, "step": 79350 }, { "epoch": 0.14816922829304477, "grad_norm": 0.8874202370643616, "learning_rate": 0.00019983970415928234, "loss": 4.8035, "step": 79400 }, { "epoch": 0.14826253385242327, "grad_norm": 0.9835208058357239, "learning_rate": 0.00019983946630355642, "loss": 4.8454, "step": 79450 }, { "epoch": 0.14835583941180175, "grad_norm": 1.30135977268219, "learning_rate": 0.00019983922827163144, "loss": 4.9588, "step": 79500 }, { "epoch": 0.14844914497118025, "grad_norm": 0.8822344541549683, "learning_rate": 0.00019983899006350777, "loss": 4.6737, "step": 79550 }, { "epoch": 0.14854245053055873, "grad_norm": 0.812828779220581, "learning_rate": 0.0001998387516791859, "loss": 4.7829, "step": 79600 }, { "epoch": 0.14863575608993723, "grad_norm": 0.9404298663139343, "learning_rate": 0.00019983851311866616, "loss": 4.4065, "step": 79650 }, { "epoch": 0.1487290616493157, "grad_norm": 1.5870780944824219, "learning_rate": 0.000199838274381949, "loss": 4.8494, "step": 79700 }, { "epoch": 0.1488223672086942, "grad_norm": 0.9650738835334778, "learning_rate": 0.0001998380354690349, "loss": 4.7546, "step": 79750 }, { "epoch": 0.14891567276807272, "grad_norm": 1.1705528497695923, "learning_rate": 0.00019983779637992418, "loss": 4.9383, "step": 79800 }, { "epoch": 0.1490089783274512, "grad_norm": 1.1494523286819458, "learning_rate": 0.00019983755711461734, "loss": 4.7803, "step": 79850 }, { "epoch": 0.1491022838868297, "grad_norm": 0.7461511492729187, "learning_rate": 0.00019983731767311477, "loss": 4.7261, "step": 79900 }, { "epoch": 0.14919558944620817, "grad_norm": 1.043479323387146, "learning_rate": 0.0001998370780554169, "loss": 4.8492, "step": 79950 }, { "epoch": 0.14928889500558667, "grad_norm": 1.0631600618362427, "learning_rate": 0.0001998368382615242, "loss": 4.7619, "step": 80000 }, { "epoch": 0.14938220056496515, "grad_norm": 1.0687527656555176, "learning_rate": 0.00019983659829143698, "loss": 4.9802, "step": 80050 }, { "epoch": 0.14947550612434365, "grad_norm": 0.7543812394142151, "learning_rate": 0.0001998363581451558, "loss": 5.0034, "step": 80100 }, { "epoch": 0.14956881168372216, "grad_norm": 0.6798197627067566, "learning_rate": 0.00019983611782268095, "loss": 4.7076, "step": 80150 }, { "epoch": 0.14966211724310063, "grad_norm": 0.8588387370109558, "learning_rate": 0.00019983587732401293, "loss": 4.8063, "step": 80200 }, { "epoch": 0.14975542280247914, "grad_norm": 1.1029499769210815, "learning_rate": 0.0001998356366491522, "loss": 4.6521, "step": 80250 }, { "epoch": 0.1498487283618576, "grad_norm": 0.8040598034858704, "learning_rate": 0.0001998353957980991, "loss": 4.6823, "step": 80300 }, { "epoch": 0.14994203392123612, "grad_norm": 1.1279397010803223, "learning_rate": 0.00019983515477085408, "loss": 4.7486, "step": 80350 }, { "epoch": 0.1500353394806146, "grad_norm": 1.1090079545974731, "learning_rate": 0.00019983491356741762, "loss": 4.7731, "step": 80400 }, { "epoch": 0.1501286450399931, "grad_norm": 1.3547964096069336, "learning_rate": 0.00019983467218779007, "loss": 4.8618, "step": 80450 }, { "epoch": 0.1502219505993716, "grad_norm": 1.263022780418396, "learning_rate": 0.00019983443063197192, "loss": 4.8745, "step": 80500 }, { "epoch": 0.15031525615875008, "grad_norm": 1.056709885597229, "learning_rate": 0.0001998341888999635, "loss": 4.7292, "step": 80550 }, { "epoch": 0.15040856171812858, "grad_norm": 1.0344963073730469, "learning_rate": 0.0001998339469917654, "loss": 4.5219, "step": 80600 }, { "epoch": 0.15050186727750706, "grad_norm": 0.7098541259765625, "learning_rate": 0.0001998337049073779, "loss": 4.6419, "step": 80650 }, { "epoch": 0.15059517283688556, "grad_norm": 1.0774918794631958, "learning_rate": 0.00019983346264680147, "loss": 4.7146, "step": 80700 }, { "epoch": 0.15068847839626404, "grad_norm": 0.8869603872299194, "learning_rate": 0.00019983322021003655, "loss": 4.8143, "step": 80750 }, { "epoch": 0.15078178395564254, "grad_norm": 1.0924726724624634, "learning_rate": 0.00019983297759708357, "loss": 4.7092, "step": 80800 }, { "epoch": 0.15087508951502102, "grad_norm": 1.0143225193023682, "learning_rate": 0.00019983273480794295, "loss": 4.6599, "step": 80850 }, { "epoch": 0.15096839507439952, "grad_norm": 1.0644515752792358, "learning_rate": 0.0001998324918426151, "loss": 4.6655, "step": 80900 }, { "epoch": 0.15106170063377802, "grad_norm": 1.2146971225738525, "learning_rate": 0.00019983224870110048, "loss": 4.7266, "step": 80950 }, { "epoch": 0.1511550061931565, "grad_norm": 0.9861900210380554, "learning_rate": 0.00019983200538339953, "loss": 4.8545, "step": 81000 }, { "epoch": 0.1511550061931565, "eval_loss": 4.951722145080566, "eval_runtime": 230.9557, "eval_samples_per_second": 11.292, "eval_steps_per_second": 11.292, "eval_tts_loss": 7.390448370325648, "step": 81000 }, { "epoch": 0.151248311752535, "grad_norm": 1.09833824634552, "learning_rate": 0.00019983176188951264, "loss": 4.7289, "step": 81050 }, { "epoch": 0.15134161731191348, "grad_norm": 1.1554728746414185, "learning_rate": 0.00019983151821944027, "loss": 4.7042, "step": 81100 }, { "epoch": 0.15143492287129198, "grad_norm": 0.9898795485496521, "learning_rate": 0.00019983127437318282, "loss": 4.7149, "step": 81150 }, { "epoch": 0.15152822843067046, "grad_norm": 0.9649356603622437, "learning_rate": 0.00019983103035074072, "loss": 4.9477, "step": 81200 }, { "epoch": 0.15162153399004896, "grad_norm": 0.9268784523010254, "learning_rate": 0.00019983078615211442, "loss": 4.6484, "step": 81250 }, { "epoch": 0.15171483954942747, "grad_norm": 1.1820745468139648, "learning_rate": 0.00019983054177730435, "loss": 4.6602, "step": 81300 }, { "epoch": 0.15180814510880594, "grad_norm": 0.8073734045028687, "learning_rate": 0.00019983029722631096, "loss": 4.6794, "step": 81350 }, { "epoch": 0.15190145066818445, "grad_norm": 0.8954252600669861, "learning_rate": 0.00019983005249913467, "loss": 4.6629, "step": 81400 }, { "epoch": 0.15199475622756292, "grad_norm": 1.1588051319122314, "learning_rate": 0.0001998298075957759, "loss": 4.9659, "step": 81450 }, { "epoch": 0.15208806178694143, "grad_norm": 1.1768730878829956, "learning_rate": 0.00019982956251623506, "loss": 4.7299, "step": 81500 }, { "epoch": 0.1521813673463199, "grad_norm": 0.9447085857391357, "learning_rate": 0.00019982931726051262, "loss": 4.7048, "step": 81550 }, { "epoch": 0.1522746729056984, "grad_norm": 0.8608984351158142, "learning_rate": 0.000199829071828609, "loss": 4.7718, "step": 81600 }, { "epoch": 0.1523679784650769, "grad_norm": 0.7421265840530396, "learning_rate": 0.0001998288262205246, "loss": 4.8859, "step": 81650 }, { "epoch": 0.15246128402445538, "grad_norm": 0.9788615107536316, "learning_rate": 0.00019982858043625994, "loss": 4.8217, "step": 81700 }, { "epoch": 0.1525545895838339, "grad_norm": 0.8593958616256714, "learning_rate": 0.00019982833447581538, "loss": 4.918, "step": 81750 }, { "epoch": 0.15264789514321236, "grad_norm": 1.2187819480895996, "learning_rate": 0.0001998280883391914, "loss": 4.6866, "step": 81800 }, { "epoch": 0.15274120070259087, "grad_norm": 1.0192663669586182, "learning_rate": 0.0001998278420263884, "loss": 4.7098, "step": 81850 }, { "epoch": 0.15283450626196934, "grad_norm": 1.1101956367492676, "learning_rate": 0.00019982759553740677, "loss": 4.6928, "step": 81900 }, { "epoch": 0.15292781182134785, "grad_norm": 0.8426311016082764, "learning_rate": 0.00019982734887224703, "loss": 4.7077, "step": 81950 }, { "epoch": 0.15302111738072632, "grad_norm": 1.0216084718704224, "learning_rate": 0.00019982710203090962, "loss": 4.822, "step": 82000 }, { "epoch": 0.15311442294010483, "grad_norm": 1.1248880624771118, "learning_rate": 0.00019982685501339492, "loss": 4.7781, "step": 82050 }, { "epoch": 0.15320772849948333, "grad_norm": 0.8825070261955261, "learning_rate": 0.00019982660781970335, "loss": 4.7426, "step": 82100 }, { "epoch": 0.1533010340588618, "grad_norm": 1.2149336338043213, "learning_rate": 0.00019982636044983543, "loss": 4.7091, "step": 82150 }, { "epoch": 0.1533943396182403, "grad_norm": 1.1825565099716187, "learning_rate": 0.00019982611290379152, "loss": 4.7934, "step": 82200 }, { "epoch": 0.1534876451776188, "grad_norm": 0.9500203728675842, "learning_rate": 0.0001998258651815721, "loss": 4.5296, "step": 82250 }, { "epoch": 0.1535809507369973, "grad_norm": 1.058222770690918, "learning_rate": 0.00019982561728317758, "loss": 4.7336, "step": 82300 }, { "epoch": 0.15367425629637577, "grad_norm": 1.1481443643569946, "learning_rate": 0.00019982536920860842, "loss": 4.6843, "step": 82350 }, { "epoch": 0.15376756185575427, "grad_norm": 1.009384036064148, "learning_rate": 0.00019982512095786503, "loss": 4.6789, "step": 82400 }, { "epoch": 0.15386086741513277, "grad_norm": 1.0169785022735596, "learning_rate": 0.00019982487253094791, "loss": 4.8183, "step": 82450 }, { "epoch": 0.15395417297451125, "grad_norm": 1.4280728101730347, "learning_rate": 0.0001998246239278574, "loss": 4.7322, "step": 82500 }, { "epoch": 0.15404747853388975, "grad_norm": 0.9956715106964111, "learning_rate": 0.00019982437514859402, "loss": 4.467, "step": 82550 }, { "epoch": 0.15414078409326823, "grad_norm": 0.9345201253890991, "learning_rate": 0.0001998241261931582, "loss": 4.8818, "step": 82600 }, { "epoch": 0.15423408965264673, "grad_norm": 0.8383892774581909, "learning_rate": 0.00019982387706155035, "loss": 5.0506, "step": 82650 }, { "epoch": 0.1543273952120252, "grad_norm": 1.0288281440734863, "learning_rate": 0.00019982362775377094, "loss": 4.7276, "step": 82700 }, { "epoch": 0.1544207007714037, "grad_norm": 1.5287654399871826, "learning_rate": 0.00019982337826982035, "loss": 4.7515, "step": 82750 }, { "epoch": 0.15451400633078222, "grad_norm": 1.1659990549087524, "learning_rate": 0.00019982312860969908, "loss": 4.7312, "step": 82800 }, { "epoch": 0.1546073118901607, "grad_norm": 1.0841422080993652, "learning_rate": 0.00019982287877340756, "loss": 4.6208, "step": 82850 }, { "epoch": 0.1547006174495392, "grad_norm": 0.9620239734649658, "learning_rate": 0.0001998226287609462, "loss": 4.7568, "step": 82900 }, { "epoch": 0.15479392300891767, "grad_norm": 1.0376754999160767, "learning_rate": 0.0001998223785723155, "loss": 4.9583, "step": 82950 }, { "epoch": 0.15488722856829618, "grad_norm": 0.8780238032341003, "learning_rate": 0.00019982212820751584, "loss": 4.7517, "step": 83000 }, { "epoch": 0.15498053412767465, "grad_norm": 0.9946292042732239, "learning_rate": 0.00019982187766654768, "loss": 4.7237, "step": 83050 }, { "epoch": 0.15507383968705316, "grad_norm": 0.7709692716598511, "learning_rate": 0.00019982162694941148, "loss": 4.9062, "step": 83100 }, { "epoch": 0.15516714524643166, "grad_norm": 1.010310411453247, "learning_rate": 0.00019982137605610768, "loss": 4.7754, "step": 83150 }, { "epoch": 0.15526045080581014, "grad_norm": 0.752420961856842, "learning_rate": 0.00019982112498663672, "loss": 4.762, "step": 83200 }, { "epoch": 0.15535375636518864, "grad_norm": 0.9181608557701111, "learning_rate": 0.00019982087374099906, "loss": 4.6671, "step": 83250 }, { "epoch": 0.15544706192456711, "grad_norm": 1.0234555006027222, "learning_rate": 0.00019982062231919509, "loss": 4.963, "step": 83300 }, { "epoch": 0.15554036748394562, "grad_norm": 1.1121389865875244, "learning_rate": 0.00019982037072122525, "loss": 4.5522, "step": 83350 }, { "epoch": 0.1556336730433241, "grad_norm": 1.1406534910202026, "learning_rate": 0.00019982011894709007, "loss": 4.8627, "step": 83400 }, { "epoch": 0.1557269786027026, "grad_norm": 0.9664844870567322, "learning_rate": 0.00019981986699678993, "loss": 4.6783, "step": 83450 }, { "epoch": 0.15582028416208107, "grad_norm": 0.931291937828064, "learning_rate": 0.0001998196148703253, "loss": 4.6384, "step": 83500 }, { "epoch": 0.15591358972145958, "grad_norm": 1.102245569229126, "learning_rate": 0.0001998193625676966, "loss": 4.5632, "step": 83550 }, { "epoch": 0.15600689528083808, "grad_norm": 0.9392174482345581, "learning_rate": 0.00019981911008890428, "loss": 4.71, "step": 83600 }, { "epoch": 0.15610020084021656, "grad_norm": 0.7961680293083191, "learning_rate": 0.00019981885743394884, "loss": 4.7738, "step": 83650 }, { "epoch": 0.15619350639959506, "grad_norm": 0.9825389385223389, "learning_rate": 0.00019981860460283063, "loss": 4.4743, "step": 83700 }, { "epoch": 0.15628681195897354, "grad_norm": 0.8377106189727783, "learning_rate": 0.00019981835159555016, "loss": 4.7894, "step": 83750 }, { "epoch": 0.15638011751835204, "grad_norm": 0.9133685231208801, "learning_rate": 0.00019981809841210788, "loss": 4.7747, "step": 83800 }, { "epoch": 0.15647342307773052, "grad_norm": 0.8195970058441162, "learning_rate": 0.0001998178450525042, "loss": 4.592, "step": 83850 }, { "epoch": 0.15656672863710902, "grad_norm": 1.0992616415023804, "learning_rate": 0.0001998175915167396, "loss": 4.8078, "step": 83900 }, { "epoch": 0.15666003419648752, "grad_norm": 1.0138832330703735, "learning_rate": 0.0001998173378048145, "loss": 4.6467, "step": 83950 }, { "epoch": 0.156753339755866, "grad_norm": 0.8786870837211609, "learning_rate": 0.0001998170839167294, "loss": 4.7446, "step": 84000 }, { "epoch": 0.156753339755866, "eval_loss": 4.946982383728027, "eval_runtime": 229.0908, "eval_samples_per_second": 11.384, "eval_steps_per_second": 11.384, "eval_tts_loss": 7.394625851078862, "step": 84000 }, { "epoch": 0.1568466453152445, "grad_norm": 1.0230584144592285, "learning_rate": 0.00019981682985248465, "loss": 4.6526, "step": 84050 }, { "epoch": 0.15693995087462298, "grad_norm": 0.7539162039756775, "learning_rate": 0.0001998165756120808, "loss": 4.6949, "step": 84100 }, { "epoch": 0.15703325643400148, "grad_norm": 1.1452572345733643, "learning_rate": 0.00019981632119551822, "loss": 4.6745, "step": 84150 }, { "epoch": 0.15712656199337996, "grad_norm": 0.9784168601036072, "learning_rate": 0.00019981606660279745, "loss": 4.9249, "step": 84200 }, { "epoch": 0.15721986755275846, "grad_norm": 1.1359683275222778, "learning_rate": 0.00019981581183391885, "loss": 4.761, "step": 84250 }, { "epoch": 0.15731317311213697, "grad_norm": 1.1261489391326904, "learning_rate": 0.0001998155568888829, "loss": 4.6123, "step": 84300 }, { "epoch": 0.15740647867151544, "grad_norm": 1.306156039237976, "learning_rate": 0.00019981530176769007, "loss": 4.6882, "step": 84350 }, { "epoch": 0.15749978423089395, "grad_norm": 0.861440896987915, "learning_rate": 0.0001998150464703408, "loss": 4.6409, "step": 84400 }, { "epoch": 0.15759308979027242, "grad_norm": 0.6180737018585205, "learning_rate": 0.00019981479099683553, "loss": 4.6318, "step": 84450 }, { "epoch": 0.15768639534965093, "grad_norm": 1.0346062183380127, "learning_rate": 0.00019981453534717475, "loss": 4.8614, "step": 84500 }, { "epoch": 0.1577797009090294, "grad_norm": 1.3572213649749756, "learning_rate": 0.00019981427952135884, "loss": 4.7108, "step": 84550 }, { "epoch": 0.1578730064684079, "grad_norm": 0.7147801518440247, "learning_rate": 0.0001998140235193883, "loss": 4.6901, "step": 84600 }, { "epoch": 0.15796631202778638, "grad_norm": 1.1168407201766968, "learning_rate": 0.00019981376734126355, "loss": 4.7444, "step": 84650 }, { "epoch": 0.15805961758716489, "grad_norm": 1.0936293601989746, "learning_rate": 0.00019981351098698509, "loss": 4.8869, "step": 84700 }, { "epoch": 0.1581529231465434, "grad_norm": 0.9468141198158264, "learning_rate": 0.00019981325445655331, "loss": 4.7834, "step": 84750 }, { "epoch": 0.15824622870592187, "grad_norm": 1.188549518585205, "learning_rate": 0.00019981299774996875, "loss": 4.7705, "step": 84800 }, { "epoch": 0.15833953426530037, "grad_norm": 1.0123662948608398, "learning_rate": 0.0001998127408672318, "loss": 4.7251, "step": 84850 }, { "epoch": 0.15843283982467884, "grad_norm": 1.1224631071090698, "learning_rate": 0.0001998124838083429, "loss": 4.7438, "step": 84900 }, { "epoch": 0.15852614538405735, "grad_norm": 0.9828144311904907, "learning_rate": 0.00019981222657330255, "loss": 4.7185, "step": 84950 }, { "epoch": 0.15861945094343582, "grad_norm": 0.932032585144043, "learning_rate": 0.00019981196916211115, "loss": 4.7946, "step": 85000 }, { "epoch": 0.15871275650281433, "grad_norm": 1.082952618598938, "learning_rate": 0.0001998117115747692, "loss": 4.7448, "step": 85050 }, { "epoch": 0.15880606206219283, "grad_norm": 1.2148911952972412, "learning_rate": 0.00019981145381127715, "loss": 4.6534, "step": 85100 }, { "epoch": 0.1588993676215713, "grad_norm": 0.8082234859466553, "learning_rate": 0.00019981119587163547, "loss": 4.6324, "step": 85150 }, { "epoch": 0.1589926731809498, "grad_norm": 1.083858847618103, "learning_rate": 0.00019981093775584457, "loss": 4.6805, "step": 85200 }, { "epoch": 0.1590859787403283, "grad_norm": 1.0355510711669922, "learning_rate": 0.00019981067946390493, "loss": 5.026, "step": 85250 }, { "epoch": 0.1591792842997068, "grad_norm": 1.0597902536392212, "learning_rate": 0.00019981042099581698, "loss": 4.8371, "step": 85300 }, { "epoch": 0.15927258985908527, "grad_norm": 0.8655616641044617, "learning_rate": 0.0001998101623515812, "loss": 4.8269, "step": 85350 }, { "epoch": 0.15936589541846377, "grad_norm": 1.1475633382797241, "learning_rate": 0.00019980990353119809, "loss": 4.6525, "step": 85400 }, { "epoch": 0.15945920097784227, "grad_norm": 0.9865615963935852, "learning_rate": 0.00019980964453466803, "loss": 4.6893, "step": 85450 }, { "epoch": 0.15955250653722075, "grad_norm": 0.8663655519485474, "learning_rate": 0.00019980938536199151, "loss": 4.736, "step": 85500 }, { "epoch": 0.15964581209659925, "grad_norm": 0.9567622542381287, "learning_rate": 0.000199809126013169, "loss": 4.5879, "step": 85550 }, { "epoch": 0.15973911765597773, "grad_norm": 0.9500153660774231, "learning_rate": 0.00019980886648820093, "loss": 4.7534, "step": 85600 }, { "epoch": 0.15983242321535623, "grad_norm": 0.9333567023277283, "learning_rate": 0.0001998086067870878, "loss": 4.671, "step": 85650 }, { "epoch": 0.1599257287747347, "grad_norm": 0.9707118272781372, "learning_rate": 0.00019980834690983, "loss": 4.7252, "step": 85700 }, { "epoch": 0.1600190343341132, "grad_norm": 0.801749587059021, "learning_rate": 0.00019980808685642805, "loss": 4.7634, "step": 85750 }, { "epoch": 0.16011233989349172, "grad_norm": 1.0327675342559814, "learning_rate": 0.0001998078266268824, "loss": 4.7954, "step": 85800 }, { "epoch": 0.1602056454528702, "grad_norm": 0.9381875991821289, "learning_rate": 0.00019980756622119347, "loss": 4.6373, "step": 85850 }, { "epoch": 0.1602989510122487, "grad_norm": 0.9350003600120544, "learning_rate": 0.00019980730563936175, "loss": 4.7086, "step": 85900 }, { "epoch": 0.16039225657162717, "grad_norm": 1.1545840501785278, "learning_rate": 0.00019980704488138776, "loss": 4.7647, "step": 85950 }, { "epoch": 0.16048556213100568, "grad_norm": 1.1162843704223633, "learning_rate": 0.00019980678394727184, "loss": 4.701, "step": 86000 }, { "epoch": 0.16057886769038415, "grad_norm": 0.9487619400024414, "learning_rate": 0.0001998065228370145, "loss": 4.847, "step": 86050 }, { "epoch": 0.16067217324976266, "grad_norm": 1.1424440145492554, "learning_rate": 0.00019980626155061623, "loss": 4.7775, "step": 86100 }, { "epoch": 0.16076547880914113, "grad_norm": 1.0836292505264282, "learning_rate": 0.00019980600008807746, "loss": 4.7517, "step": 86150 }, { "epoch": 0.16085878436851964, "grad_norm": 1.026230812072754, "learning_rate": 0.00019980573844939866, "loss": 4.6033, "step": 86200 }, { "epoch": 0.16095208992789814, "grad_norm": 0.7633348107337952, "learning_rate": 0.0001998054766345803, "loss": 4.7064, "step": 86250 }, { "epoch": 0.16104539548727662, "grad_norm": 0.9830362796783447, "learning_rate": 0.00019980521464362284, "loss": 4.6753, "step": 86300 }, { "epoch": 0.16113870104665512, "grad_norm": 0.7929043173789978, "learning_rate": 0.00019980495247652673, "loss": 4.4158, "step": 86350 }, { "epoch": 0.1612320066060336, "grad_norm": 0.906753420829773, "learning_rate": 0.00019980469013329243, "loss": 4.5392, "step": 86400 }, { "epoch": 0.1613253121654121, "grad_norm": 0.841719388961792, "learning_rate": 0.00019980442761392041, "loss": 4.7356, "step": 86450 }, { "epoch": 0.16141861772479058, "grad_norm": 0.9761390686035156, "learning_rate": 0.00019980416491841116, "loss": 4.8662, "step": 86500 }, { "epoch": 0.16151192328416908, "grad_norm": 1.2850239276885986, "learning_rate": 0.0001998039020467651, "loss": 4.6927, "step": 86550 }, { "epoch": 0.16160522884354758, "grad_norm": 0.9129481911659241, "learning_rate": 0.00019980363899898277, "loss": 4.855, "step": 86600 }, { "epoch": 0.16169853440292606, "grad_norm": 1.1330456733703613, "learning_rate": 0.0001998033757750645, "loss": 5.0053, "step": 86650 }, { "epoch": 0.16179183996230456, "grad_norm": 1.0761449337005615, "learning_rate": 0.00019980311237501085, "loss": 4.6144, "step": 86700 }, { "epoch": 0.16188514552168304, "grad_norm": 0.9955719709396362, "learning_rate": 0.0001998028487988223, "loss": 4.7949, "step": 86750 }, { "epoch": 0.16197845108106154, "grad_norm": 1.190447211265564, "learning_rate": 0.00019980258504649927, "loss": 4.7975, "step": 86800 }, { "epoch": 0.16207175664044002, "grad_norm": 1.0047932863235474, "learning_rate": 0.00019980232111804224, "loss": 4.7366, "step": 86850 }, { "epoch": 0.16216506219981852, "grad_norm": 0.9887164831161499, "learning_rate": 0.00019980205701345164, "loss": 4.7679, "step": 86900 }, { "epoch": 0.16225836775919703, "grad_norm": 1.305762529373169, "learning_rate": 0.00019980179273272803, "loss": 4.8636, "step": 86950 }, { "epoch": 0.1623516733185755, "grad_norm": 1.0135672092437744, "learning_rate": 0.00019980152827587177, "loss": 4.7288, "step": 87000 }, { "epoch": 0.1623516733185755, "eval_loss": 4.931732654571533, "eval_runtime": 230.4634, "eval_samples_per_second": 11.316, "eval_steps_per_second": 11.316, "eval_tts_loss": 7.40806279529566, "step": 87000 }, { "epoch": 0.162444978877954, "grad_norm": 1.1869860887527466, "learning_rate": 0.0001998012636428834, "loss": 4.6017, "step": 87050 }, { "epoch": 0.16253828443733248, "grad_norm": 0.9774503707885742, "learning_rate": 0.00019980099883376334, "loss": 4.7084, "step": 87100 }, { "epoch": 0.16263158999671098, "grad_norm": 0.8578800559043884, "learning_rate": 0.00019980073384851207, "loss": 4.9125, "step": 87150 }, { "epoch": 0.16272489555608946, "grad_norm": 1.0579946041107178, "learning_rate": 0.00019980046868713013, "loss": 4.7317, "step": 87200 }, { "epoch": 0.16281820111546796, "grad_norm": 0.6333214044570923, "learning_rate": 0.00019980020334961783, "loss": 4.5953, "step": 87250 }, { "epoch": 0.16291150667484644, "grad_norm": 1.0423671007156372, "learning_rate": 0.0001997999378359758, "loss": 4.7238, "step": 87300 }, { "epoch": 0.16300481223422494, "grad_norm": 0.7020459175109863, "learning_rate": 0.0001997996721462044, "loss": 4.6665, "step": 87350 }, { "epoch": 0.16309811779360345, "grad_norm": 0.9732704162597656, "learning_rate": 0.00019979940628030415, "loss": 4.6746, "step": 87400 }, { "epoch": 0.16319142335298192, "grad_norm": 0.8247692584991455, "learning_rate": 0.00019979914023827552, "loss": 4.8856, "step": 87450 }, { "epoch": 0.16328472891236043, "grad_norm": 1.1054822206497192, "learning_rate": 0.00019979887402011898, "loss": 4.639, "step": 87500 }, { "epoch": 0.1633780344717389, "grad_norm": 1.0070406198501587, "learning_rate": 0.00019979860762583496, "loss": 4.7724, "step": 87550 }, { "epoch": 0.1634713400311174, "grad_norm": 0.9917934536933899, "learning_rate": 0.00019979834105542397, "loss": 4.8537, "step": 87600 }, { "epoch": 0.16356464559049588, "grad_norm": 1.4356805086135864, "learning_rate": 0.00019979807430888644, "loss": 4.816, "step": 87650 }, { "epoch": 0.1636579511498744, "grad_norm": 1.1016113758087158, "learning_rate": 0.0001997978073862229, "loss": 4.8199, "step": 87700 }, { "epoch": 0.1637512567092529, "grad_norm": 1.1834055185317993, "learning_rate": 0.00019979754028743378, "loss": 4.5439, "step": 87750 }, { "epoch": 0.16384456226863137, "grad_norm": 1.2742072343826294, "learning_rate": 0.00019979727301251959, "loss": 4.7865, "step": 87800 }, { "epoch": 0.16393786782800987, "grad_norm": 1.1467939615249634, "learning_rate": 0.0001997970055614807, "loss": 4.848, "step": 87850 }, { "epoch": 0.16403117338738835, "grad_norm": 0.9691332578659058, "learning_rate": 0.00019979673793431773, "loss": 4.866, "step": 87900 }, { "epoch": 0.16412447894676685, "grad_norm": 1.061712622642517, "learning_rate": 0.00019979647013103105, "loss": 4.5297, "step": 87950 }, { "epoch": 0.16421778450614533, "grad_norm": 0.9618381857872009, "learning_rate": 0.00019979620215162115, "loss": 4.6956, "step": 88000 }, { "epoch": 0.16431109006552383, "grad_norm": 1.0956761837005615, "learning_rate": 0.0001997959339960885, "loss": 4.7234, "step": 88050 }, { "epoch": 0.16440439562490233, "grad_norm": 1.304779291152954, "learning_rate": 0.00019979566566443363, "loss": 4.7132, "step": 88100 }, { "epoch": 0.1644977011842808, "grad_norm": 0.8119908571243286, "learning_rate": 0.00019979539715665693, "loss": 4.7707, "step": 88150 }, { "epoch": 0.1645910067436593, "grad_norm": 0.9809343218803406, "learning_rate": 0.00019979512847275897, "loss": 4.6912, "step": 88200 }, { "epoch": 0.1646843123030378, "grad_norm": 1.1159480810165405, "learning_rate": 0.00019979485961274008, "loss": 4.7362, "step": 88250 }, { "epoch": 0.1647776178624163, "grad_norm": 1.0415050983428955, "learning_rate": 0.0001997945905766009, "loss": 4.8182, "step": 88300 }, { "epoch": 0.16487092342179477, "grad_norm": 0.856876015663147, "learning_rate": 0.00019979432136434179, "loss": 4.475, "step": 88350 }, { "epoch": 0.16496422898117327, "grad_norm": 0.9265991449356079, "learning_rate": 0.0001997940519759633, "loss": 4.5788, "step": 88400 }, { "epoch": 0.16505753454055178, "grad_norm": 1.082276701927185, "learning_rate": 0.00019979378241146583, "loss": 4.658, "step": 88450 }, { "epoch": 0.16515084009993025, "grad_norm": 0.916746199131012, "learning_rate": 0.0001997935126708499, "loss": 4.7118, "step": 88500 }, { "epoch": 0.16524414565930876, "grad_norm": 0.7613538503646851, "learning_rate": 0.00019979324275411597, "loss": 4.6433, "step": 88550 }, { "epoch": 0.16533745121868723, "grad_norm": 0.9194437861442566, "learning_rate": 0.00019979297266126453, "loss": 4.8783, "step": 88600 }, { "epoch": 0.16543075677806574, "grad_norm": 0.907049298286438, "learning_rate": 0.00019979270239229606, "loss": 5.006, "step": 88650 }, { "epoch": 0.1655240623374442, "grad_norm": 0.8706242442131042, "learning_rate": 0.00019979243194721103, "loss": 4.6975, "step": 88700 }, { "epoch": 0.16561736789682271, "grad_norm": 1.136927843093872, "learning_rate": 0.0001997921613260099, "loss": 4.7098, "step": 88750 }, { "epoch": 0.1657106734562012, "grad_norm": 0.9624984264373779, "learning_rate": 0.0001997918905286932, "loss": 4.7689, "step": 88800 }, { "epoch": 0.1658039790155797, "grad_norm": 1.003145694732666, "learning_rate": 0.00019979161955526134, "loss": 4.6874, "step": 88850 }, { "epoch": 0.1658972845749582, "grad_norm": 0.9634168148040771, "learning_rate": 0.00019979134840571484, "loss": 4.7637, "step": 88900 }, { "epoch": 0.16599059013433667, "grad_norm": 1.0608584880828857, "learning_rate": 0.00019979107708005417, "loss": 4.6663, "step": 88950 }, { "epoch": 0.16608389569371518, "grad_norm": 0.7554059028625488, "learning_rate": 0.00019979080557827982, "loss": 4.7785, "step": 89000 }, { "epoch": 0.16617720125309365, "grad_norm": 1.0240588188171387, "learning_rate": 0.00019979053390039225, "loss": 4.7544, "step": 89050 }, { "epoch": 0.16627050681247216, "grad_norm": 1.253793478012085, "learning_rate": 0.00019979026204639192, "loss": 4.5589, "step": 89100 }, { "epoch": 0.16636381237185063, "grad_norm": 0.729597270488739, "learning_rate": 0.00019978999001627937, "loss": 4.5692, "step": 89150 }, { "epoch": 0.16645711793122914, "grad_norm": 0.9666997194290161, "learning_rate": 0.00019978971781005504, "loss": 4.7101, "step": 89200 }, { "epoch": 0.16655042349060764, "grad_norm": 1.3623840808868408, "learning_rate": 0.0001997894454277194, "loss": 4.6743, "step": 89250 }, { "epoch": 0.16664372904998612, "grad_norm": 0.9426568746566772, "learning_rate": 0.00019978917286927295, "loss": 4.6291, "step": 89300 }, { "epoch": 0.16673703460936462, "grad_norm": 0.9860442876815796, "learning_rate": 0.00019978890013471617, "loss": 4.7023, "step": 89350 }, { "epoch": 0.1668303401687431, "grad_norm": 0.9578588008880615, "learning_rate": 0.00019978862722404956, "loss": 4.8009, "step": 89400 }, { "epoch": 0.1669236457281216, "grad_norm": 1.0634305477142334, "learning_rate": 0.00019978835413727358, "loss": 4.8068, "step": 89450 }, { "epoch": 0.16701695128750008, "grad_norm": 0.5965625643730164, "learning_rate": 0.0001997880808743887, "loss": 4.5365, "step": 89500 }, { "epoch": 0.16711025684687858, "grad_norm": 0.9046986699104309, "learning_rate": 0.0001997878074353954, "loss": 4.7543, "step": 89550 }, { "epoch": 0.16720356240625708, "grad_norm": 1.0531772375106812, "learning_rate": 0.0001997875338202942, "loss": 4.7865, "step": 89600 }, { "epoch": 0.16729686796563556, "grad_norm": 1.0829286575317383, "learning_rate": 0.00019978726002908555, "loss": 4.5878, "step": 89650 }, { "epoch": 0.16739017352501406, "grad_norm": 0.9807512164115906, "learning_rate": 0.00019978698606176994, "loss": 4.7111, "step": 89700 }, { "epoch": 0.16748347908439254, "grad_norm": 0.7498921155929565, "learning_rate": 0.00019978671191834787, "loss": 4.5111, "step": 89750 }, { "epoch": 0.16757678464377104, "grad_norm": 1.3092355728149414, "learning_rate": 0.0001997864375988198, "loss": 4.6699, "step": 89800 }, { "epoch": 0.16767009020314952, "grad_norm": 1.1617666482925415, "learning_rate": 0.00019978616310318623, "loss": 4.7742, "step": 89850 }, { "epoch": 0.16776339576252802, "grad_norm": 1.0421655178070068, "learning_rate": 0.00019978588843144767, "loss": 4.8136, "step": 89900 }, { "epoch": 0.16785670132190653, "grad_norm": 1.0589338541030884, "learning_rate": 0.00019978561358360454, "loss": 4.961, "step": 89950 }, { "epoch": 0.167950006881285, "grad_norm": 0.8682652115821838, "learning_rate": 0.00019978533855965735, "loss": 4.6367, "step": 90000 }, { "epoch": 0.167950006881285, "eval_loss": 4.925217628479004, "eval_runtime": 229.1885, "eval_samples_per_second": 11.379, "eval_steps_per_second": 11.379, "eval_tts_loss": 7.469426870473198, "step": 90000 }, { "epoch": 0.1680433124406635, "grad_norm": 1.0179359912872314, "learning_rate": 0.00019978506335960664, "loss": 4.7742, "step": 90050 }, { "epoch": 0.16813661800004198, "grad_norm": 1.0394161939620972, "learning_rate": 0.00019978478798345282, "loss": 4.7354, "step": 90100 }, { "epoch": 0.16822992355942049, "grad_norm": 1.118033766746521, "learning_rate": 0.0001997845124311964, "loss": 4.6653, "step": 90150 }, { "epoch": 0.16832322911879896, "grad_norm": 0.7835417985916138, "learning_rate": 0.00019978423670283788, "loss": 4.7592, "step": 90200 }, { "epoch": 0.16841653467817747, "grad_norm": 1.0521881580352783, "learning_rate": 0.00019978396079837777, "loss": 4.7853, "step": 90250 }, { "epoch": 0.16850984023755594, "grad_norm": 0.9498222470283508, "learning_rate": 0.0001997836847178165, "loss": 4.7828, "step": 90300 }, { "epoch": 0.16860314579693444, "grad_norm": 0.9034340381622314, "learning_rate": 0.00019978340846115456, "loss": 4.6206, "step": 90350 }, { "epoch": 0.16869645135631295, "grad_norm": 1.0530623197555542, "learning_rate": 0.00019978313202839249, "loss": 4.6888, "step": 90400 }, { "epoch": 0.16878975691569142, "grad_norm": 1.2802088260650635, "learning_rate": 0.00019978285541953077, "loss": 4.665, "step": 90450 }, { "epoch": 0.16888306247506993, "grad_norm": 0.7782902717590332, "learning_rate": 0.00019978257863456983, "loss": 4.7043, "step": 90500 }, { "epoch": 0.1689763680344484, "grad_norm": 0.930324912071228, "learning_rate": 0.00019978230167351022, "loss": 4.7948, "step": 90550 }, { "epoch": 0.1690696735938269, "grad_norm": 1.2882273197174072, "learning_rate": 0.00019978202453635238, "loss": 4.8561, "step": 90600 }, { "epoch": 0.16916297915320538, "grad_norm": 0.7437873482704163, "learning_rate": 0.00019978174722309685, "loss": 4.8459, "step": 90650 }, { "epoch": 0.1692562847125839, "grad_norm": 0.9222044944763184, "learning_rate": 0.00019978146973374407, "loss": 4.581, "step": 90700 }, { "epoch": 0.1693495902719624, "grad_norm": 1.0944164991378784, "learning_rate": 0.00019978119206829454, "loss": 4.86, "step": 90750 }, { "epoch": 0.16944289583134087, "grad_norm": 0.8687542080879211, "learning_rate": 0.00019978091422674882, "loss": 4.7375, "step": 90800 }, { "epoch": 0.16953620139071937, "grad_norm": 1.013555884361267, "learning_rate": 0.0001997806362091073, "loss": 4.8501, "step": 90850 }, { "epoch": 0.16962950695009785, "grad_norm": 0.9529797434806824, "learning_rate": 0.00019978035801537054, "loss": 4.9599, "step": 90900 }, { "epoch": 0.16972281250947635, "grad_norm": 1.0254067182540894, "learning_rate": 0.00019978007964553899, "loss": 4.6487, "step": 90950 }, { "epoch": 0.16981611806885483, "grad_norm": 1.2346069812774658, "learning_rate": 0.00019977980109961313, "loss": 4.9191, "step": 91000 }, { "epoch": 0.16990942362823333, "grad_norm": 0.7727384567260742, "learning_rate": 0.00019977952237759352, "loss": 4.676, "step": 91050 }, { "epoch": 0.17000272918761183, "grad_norm": 0.8196634650230408, "learning_rate": 0.00019977924347948058, "loss": 4.7945, "step": 91100 }, { "epoch": 0.1700960347469903, "grad_norm": 0.937310516834259, "learning_rate": 0.00019977896440527486, "loss": 4.8049, "step": 91150 }, { "epoch": 0.1701893403063688, "grad_norm": 0.6780749559402466, "learning_rate": 0.0001997786851549768, "loss": 4.9489, "step": 91200 }, { "epoch": 0.1702826458657473, "grad_norm": 1.1923000812530518, "learning_rate": 0.00019977840572858691, "loss": 4.6109, "step": 91250 }, { "epoch": 0.1703759514251258, "grad_norm": 1.0224426984786987, "learning_rate": 0.00019977812612610572, "loss": 4.7127, "step": 91300 }, { "epoch": 0.17046925698450427, "grad_norm": 0.8383764624595642, "learning_rate": 0.0001997778463475337, "loss": 4.6056, "step": 91350 }, { "epoch": 0.17056256254388277, "grad_norm": 1.2207326889038086, "learning_rate": 0.00019977756639287132, "loss": 4.7927, "step": 91400 }, { "epoch": 0.17065586810326125, "grad_norm": 1.1539119482040405, "learning_rate": 0.0001997772862621191, "loss": 4.7997, "step": 91450 }, { "epoch": 0.17074917366263975, "grad_norm": 0.9894577860832214, "learning_rate": 0.0001997770059552775, "loss": 4.7092, "step": 91500 }, { "epoch": 0.17084247922201826, "grad_norm": 1.1220641136169434, "learning_rate": 0.00019977672547234705, "loss": 4.7147, "step": 91550 }, { "epoch": 0.17093578478139673, "grad_norm": 0.8501759171485901, "learning_rate": 0.00019977644481332824, "loss": 4.7409, "step": 91600 }, { "epoch": 0.17102909034077524, "grad_norm": 0.859946608543396, "learning_rate": 0.00019977616397822158, "loss": 4.5836, "step": 91650 }, { "epoch": 0.1711223959001537, "grad_norm": 0.7580205798149109, "learning_rate": 0.00019977588296702753, "loss": 4.6373, "step": 91700 }, { "epoch": 0.17121570145953222, "grad_norm": 1.3309026956558228, "learning_rate": 0.0001997756017797466, "loss": 4.6238, "step": 91750 }, { "epoch": 0.1713090070189107, "grad_norm": 0.965422511100769, "learning_rate": 0.0001997753204163793, "loss": 4.8401, "step": 91800 }, { "epoch": 0.1714023125782892, "grad_norm": 0.9164948463439941, "learning_rate": 0.0001997750388769261, "loss": 4.8671, "step": 91850 }, { "epoch": 0.1714956181376677, "grad_norm": 0.8471038937568665, "learning_rate": 0.00019977475716138752, "loss": 4.7976, "step": 91900 }, { "epoch": 0.17158892369704618, "grad_norm": 1.1867235898971558, "learning_rate": 0.00019977447526976406, "loss": 4.8716, "step": 91950 }, { "epoch": 0.17168222925642468, "grad_norm": 1.0447583198547363, "learning_rate": 0.00019977419320205616, "loss": 4.5796, "step": 92000 }, { "epoch": 0.17177553481580315, "grad_norm": 0.894728422164917, "learning_rate": 0.00019977391095826443, "loss": 4.7049, "step": 92050 }, { "epoch": 0.17186884037518166, "grad_norm": 1.03791344165802, "learning_rate": 0.0001997736285383893, "loss": 4.6183, "step": 92100 }, { "epoch": 0.17196214593456013, "grad_norm": 0.7112425565719604, "learning_rate": 0.00019977334594243124, "loss": 4.5619, "step": 92150 }, { "epoch": 0.17205545149393864, "grad_norm": 0.7274053692817688, "learning_rate": 0.00019977306317039078, "loss": 4.7356, "step": 92200 }, { "epoch": 0.17214875705331714, "grad_norm": 0.9669164419174194, "learning_rate": 0.0001997727802222684, "loss": 4.5581, "step": 92250 }, { "epoch": 0.17224206261269562, "grad_norm": 0.9389071464538574, "learning_rate": 0.00019977249709806464, "loss": 4.7506, "step": 92300 }, { "epoch": 0.17233536817207412, "grad_norm": 0.9951257109642029, "learning_rate": 0.00019977221379778, "loss": 4.9339, "step": 92350 }, { "epoch": 0.1724286737314526, "grad_norm": 1.1136345863342285, "learning_rate": 0.00019977193032141494, "loss": 4.8545, "step": 92400 }, { "epoch": 0.1725219792908311, "grad_norm": 0.7983369827270508, "learning_rate": 0.00019977164666896998, "loss": 4.6367, "step": 92450 }, { "epoch": 0.17261528485020958, "grad_norm": 0.905382513999939, "learning_rate": 0.00019977136284044563, "loss": 4.4375, "step": 92500 }, { "epoch": 0.17270859040958808, "grad_norm": 1.1273249387741089, "learning_rate": 0.00019977107883584235, "loss": 4.777, "step": 92550 }, { "epoch": 0.17280189596896658, "grad_norm": 0.7400512099266052, "learning_rate": 0.00019977079465516068, "loss": 4.6677, "step": 92600 }, { "epoch": 0.17289520152834506, "grad_norm": 1.0424542427062988, "learning_rate": 0.00019977051029840116, "loss": 4.6719, "step": 92650 }, { "epoch": 0.17298850708772356, "grad_norm": 1.0340840816497803, "learning_rate": 0.00019977022576556417, "loss": 4.7587, "step": 92700 }, { "epoch": 0.17308181264710204, "grad_norm": 1.0534294843673706, "learning_rate": 0.00019976994105665033, "loss": 4.7536, "step": 92750 }, { "epoch": 0.17317511820648054, "grad_norm": 0.754612922668457, "learning_rate": 0.0001997696561716601, "loss": 4.5673, "step": 92800 }, { "epoch": 0.17326842376585902, "grad_norm": 1.1167819499969482, "learning_rate": 0.00019976937111059395, "loss": 4.735, "step": 92850 }, { "epoch": 0.17336172932523752, "grad_norm": 1.0311816930770874, "learning_rate": 0.00019976908587345245, "loss": 4.8155, "step": 92900 }, { "epoch": 0.173455034884616, "grad_norm": 1.6865220069885254, "learning_rate": 0.00019976880046023606, "loss": 4.5625, "step": 92950 }, { "epoch": 0.1735483404439945, "grad_norm": 1.0632766485214233, "learning_rate": 0.0001997685148709453, "loss": 4.6878, "step": 93000 }, { "epoch": 0.1735483404439945, "eval_loss": 4.916426658630371, "eval_runtime": 229.9701, "eval_samples_per_second": 11.341, "eval_steps_per_second": 11.341, "eval_tts_loss": 7.440112192040892, "step": 93000 }, { "epoch": 0.173641646003373, "grad_norm": 0.9786635637283325, "learning_rate": 0.00019976822910558063, "loss": 4.7599, "step": 93050 }, { "epoch": 0.17373495156275148, "grad_norm": 0.9855509400367737, "learning_rate": 0.00019976794316414265, "loss": 4.4347, "step": 93100 }, { "epoch": 0.17382825712213, "grad_norm": 0.8809786438941956, "learning_rate": 0.00019976765704663176, "loss": 4.6688, "step": 93150 }, { "epoch": 0.17392156268150846, "grad_norm": 1.0891468524932861, "learning_rate": 0.00019976737075304851, "loss": 4.7232, "step": 93200 }, { "epoch": 0.17401486824088697, "grad_norm": 1.038535237312317, "learning_rate": 0.00019976708428339342, "loss": 4.7486, "step": 93250 }, { "epoch": 0.17410817380026544, "grad_norm": 1.0178803205490112, "learning_rate": 0.00019976679763766697, "loss": 4.6426, "step": 93300 }, { "epoch": 0.17420147935964395, "grad_norm": 0.7876224517822266, "learning_rate": 0.00019976651081586968, "loss": 4.7747, "step": 93350 }, { "epoch": 0.17429478491902245, "grad_norm": 0.6796728372573853, "learning_rate": 0.00019976622381800207, "loss": 4.8263, "step": 93400 }, { "epoch": 0.17438809047840093, "grad_norm": 0.796763002872467, "learning_rate": 0.0001997659366440646, "loss": 4.5471, "step": 93450 }, { "epoch": 0.17448139603777943, "grad_norm": 0.7914952039718628, "learning_rate": 0.00019976564929405783, "loss": 4.8249, "step": 93500 }, { "epoch": 0.1745747015971579, "grad_norm": 0.7674325704574585, "learning_rate": 0.00019976536176798224, "loss": 4.6139, "step": 93550 }, { "epoch": 0.1746680071565364, "grad_norm": 0.8790803551673889, "learning_rate": 0.00019976507406583834, "loss": 4.6569, "step": 93600 }, { "epoch": 0.17476131271591488, "grad_norm": 1.362616777420044, "learning_rate": 0.00019976478618762662, "loss": 4.7218, "step": 93650 }, { "epoch": 0.1748546182752934, "grad_norm": 1.1918220520019531, "learning_rate": 0.00019976449813334762, "loss": 4.9003, "step": 93700 }, { "epoch": 0.1749479238346719, "grad_norm": 0.9692675471305847, "learning_rate": 0.00019976420990300183, "loss": 4.6867, "step": 93750 }, { "epoch": 0.17504122939405037, "grad_norm": 1.2216622829437256, "learning_rate": 0.00019976392149658977, "loss": 4.6528, "step": 93800 }, { "epoch": 0.17513453495342887, "grad_norm": 0.9711513519287109, "learning_rate": 0.00019976363291411195, "loss": 4.7519, "step": 93850 }, { "epoch": 0.17522784051280735, "grad_norm": 1.0663491487503052, "learning_rate": 0.00019976334415556886, "loss": 4.5517, "step": 93900 }, { "epoch": 0.17532114607218585, "grad_norm": 0.9361185431480408, "learning_rate": 0.00019976305522096101, "loss": 4.7161, "step": 93950 }, { "epoch": 0.17541445163156433, "grad_norm": 0.7986595630645752, "learning_rate": 0.00019976276611028895, "loss": 4.7228, "step": 94000 }, { "epoch": 0.17550775719094283, "grad_norm": 0.9233583211898804, "learning_rate": 0.00019976247682355316, "loss": 4.8464, "step": 94050 }, { "epoch": 0.1756010627503213, "grad_norm": 1.2323817014694214, "learning_rate": 0.00019976218736075414, "loss": 4.5892, "step": 94100 }, { "epoch": 0.1756943683096998, "grad_norm": 1.0528844594955444, "learning_rate": 0.00019976189772189242, "loss": 4.8237, "step": 94150 }, { "epoch": 0.17578767386907831, "grad_norm": 1.1424999237060547, "learning_rate": 0.00019976160790696848, "loss": 4.6996, "step": 94200 }, { "epoch": 0.1758809794284568, "grad_norm": 0.8401368856430054, "learning_rate": 0.00019976131791598286, "loss": 4.8017, "step": 94250 }, { "epoch": 0.1759742849878353, "grad_norm": 0.7714797854423523, "learning_rate": 0.00019976102774893608, "loss": 4.5253, "step": 94300 }, { "epoch": 0.17606759054721377, "grad_norm": 1.129595398902893, "learning_rate": 0.00019976073740582864, "loss": 4.8271, "step": 94350 }, { "epoch": 0.17616089610659227, "grad_norm": 0.9854161739349365, "learning_rate": 0.00019976044688666106, "loss": 4.6552, "step": 94400 }, { "epoch": 0.17625420166597075, "grad_norm": 1.1386884450912476, "learning_rate": 0.00019976015619143384, "loss": 4.6417, "step": 94450 }, { "epoch": 0.17634750722534925, "grad_norm": 1.0306202173233032, "learning_rate": 0.0001997598653201475, "loss": 4.7774, "step": 94500 }, { "epoch": 0.17644081278472776, "grad_norm": 1.3340245485305786, "learning_rate": 0.00019975957427280255, "loss": 4.9127, "step": 94550 }, { "epoch": 0.17653411834410623, "grad_norm": 1.3886502981185913, "learning_rate": 0.0001997592830493995, "loss": 4.5873, "step": 94600 }, { "epoch": 0.17662742390348474, "grad_norm": 1.117301344871521, "learning_rate": 0.00019975899164993886, "loss": 4.7842, "step": 94650 }, { "epoch": 0.1767207294628632, "grad_norm": 1.071517825126648, "learning_rate": 0.00019975870007442117, "loss": 4.6684, "step": 94700 }, { "epoch": 0.17681403502224172, "grad_norm": 1.0764856338500977, "learning_rate": 0.0001997584083228469, "loss": 4.6823, "step": 94750 }, { "epoch": 0.1769073405816202, "grad_norm": 0.9086782932281494, "learning_rate": 0.00019975811639521662, "loss": 4.863, "step": 94800 }, { "epoch": 0.1770006461409987, "grad_norm": 0.9636552929878235, "learning_rate": 0.00019975782429153082, "loss": 4.685, "step": 94850 }, { "epoch": 0.1770939517003772, "grad_norm": 0.620227038860321, "learning_rate": 0.00019975753201179, "loss": 4.6816, "step": 94900 }, { "epoch": 0.17718725725975568, "grad_norm": 0.8294452428817749, "learning_rate": 0.00019975723955599468, "loss": 4.6858, "step": 94950 }, { "epoch": 0.17728056281913418, "grad_norm": 1.1841310262680054, "learning_rate": 0.0001997569469241454, "loss": 4.6845, "step": 95000 }, { "epoch": 0.17737386837851266, "grad_norm": 1.1353343725204468, "learning_rate": 0.00019975665411624266, "loss": 4.9758, "step": 95050 }, { "epoch": 0.17746717393789116, "grad_norm": 1.098388910293579, "learning_rate": 0.00019975636113228696, "loss": 4.807, "step": 95100 }, { "epoch": 0.17756047949726964, "grad_norm": 1.1415269374847412, "learning_rate": 0.00019975606797227886, "loss": 4.6393, "step": 95150 }, { "epoch": 0.17765378505664814, "grad_norm": 1.01333487033844, "learning_rate": 0.00019975577463621883, "loss": 4.7383, "step": 95200 }, { "epoch": 0.17774709061602664, "grad_norm": 0.9192553758621216, "learning_rate": 0.00019975548112410743, "loss": 4.7246, "step": 95250 }, { "epoch": 0.17784039617540512, "grad_norm": 1.233699917793274, "learning_rate": 0.00019975518743594516, "loss": 4.6757, "step": 95300 }, { "epoch": 0.17793370173478362, "grad_norm": 1.0001840591430664, "learning_rate": 0.00019975489357173252, "loss": 4.6488, "step": 95350 }, { "epoch": 0.1780270072941621, "grad_norm": 0.9523444771766663, "learning_rate": 0.00019975459953147005, "loss": 4.6393, "step": 95400 }, { "epoch": 0.1781203128535406, "grad_norm": 1.081119179725647, "learning_rate": 0.00019975430531515826, "loss": 4.583, "step": 95450 }, { "epoch": 0.17821361841291908, "grad_norm": 0.9086116552352905, "learning_rate": 0.00019975401092279767, "loss": 4.659, "step": 95500 }, { "epoch": 0.17830692397229758, "grad_norm": 0.9077746868133545, "learning_rate": 0.00019975371635438884, "loss": 4.5218, "step": 95550 }, { "epoch": 0.17840022953167606, "grad_norm": 1.0118244886398315, "learning_rate": 0.00019975342160993224, "loss": 4.9855, "step": 95600 }, { "epoch": 0.17849353509105456, "grad_norm": 0.7483125329017639, "learning_rate": 0.0001997531266894284, "loss": 4.5592, "step": 95650 }, { "epoch": 0.17858684065043307, "grad_norm": 0.9252362847328186, "learning_rate": 0.00019975283159287782, "loss": 4.6356, "step": 95700 }, { "epoch": 0.17868014620981154, "grad_norm": 0.7764643430709839, "learning_rate": 0.00019975253632028105, "loss": 4.6681, "step": 95750 }, { "epoch": 0.17877345176919004, "grad_norm": 0.9048830270767212, "learning_rate": 0.00019975224087163861, "loss": 4.5292, "step": 95800 }, { "epoch": 0.17886675732856852, "grad_norm": 0.9486568570137024, "learning_rate": 0.00019975194524695102, "loss": 4.7868, "step": 95850 }, { "epoch": 0.17896006288794702, "grad_norm": 0.8098161816596985, "learning_rate": 0.0001997516494462188, "loss": 4.6501, "step": 95900 }, { "epoch": 0.1790533684473255, "grad_norm": 1.2166732549667358, "learning_rate": 0.00019975135346944248, "loss": 4.4709, "step": 95950 }, { "epoch": 0.179146674006704, "grad_norm": 0.9364936351776123, "learning_rate": 0.00019975105731662255, "loss": 4.592, "step": 96000 }, { "epoch": 0.179146674006704, "eval_loss": 4.917341232299805, "eval_runtime": 231.2993, "eval_samples_per_second": 11.275, "eval_steps_per_second": 11.275, "eval_tts_loss": 7.391886383147933, "step": 96000 }, { "epoch": 0.1792399795660825, "grad_norm": 1.0819836854934692, "learning_rate": 0.00019975076098775957, "loss": 4.6746, "step": 96050 }, { "epoch": 0.17933328512546098, "grad_norm": 1.1346968412399292, "learning_rate": 0.00019975046448285407, "loss": 4.7887, "step": 96100 }, { "epoch": 0.1794265906848395, "grad_norm": 1.0858914852142334, "learning_rate": 0.00019975016780190652, "loss": 4.5072, "step": 96150 }, { "epoch": 0.17951989624421796, "grad_norm": 1.002199649810791, "learning_rate": 0.0001997498709449175, "loss": 4.7632, "step": 96200 }, { "epoch": 0.17961320180359647, "grad_norm": 0.8308196067810059, "learning_rate": 0.00019974957391188746, "loss": 4.6145, "step": 96250 }, { "epoch": 0.17970650736297494, "grad_norm": 0.9789503216743469, "learning_rate": 0.00019974927670281703, "loss": 4.6052, "step": 96300 }, { "epoch": 0.17979981292235345, "grad_norm": 1.1850559711456299, "learning_rate": 0.00019974897931770663, "loss": 4.6967, "step": 96350 }, { "epoch": 0.17989311848173195, "grad_norm": 1.0554805994033813, "learning_rate": 0.00019974868175655686, "loss": 4.8473, "step": 96400 }, { "epoch": 0.17998642404111043, "grad_norm": 1.0979632139205933, "learning_rate": 0.00019974838401936822, "loss": 4.8164, "step": 96450 }, { "epoch": 0.18007972960048893, "grad_norm": 1.1470582485198975, "learning_rate": 0.00019974808610614123, "loss": 4.674, "step": 96500 }, { "epoch": 0.1801730351598674, "grad_norm": 0.7371369004249573, "learning_rate": 0.00019974778801687643, "loss": 4.7965, "step": 96550 }, { "epoch": 0.1802663407192459, "grad_norm": 0.9566873908042908, "learning_rate": 0.00019974748975157432, "loss": 4.8604, "step": 96600 }, { "epoch": 0.18035964627862439, "grad_norm": 1.2158596515655518, "learning_rate": 0.00019974719131023543, "loss": 4.7986, "step": 96650 }, { "epoch": 0.1804529518380029, "grad_norm": 1.0148197412490845, "learning_rate": 0.0001997468926928603, "loss": 4.9964, "step": 96700 }, { "epoch": 0.1805462573973814, "grad_norm": 1.0894036293029785, "learning_rate": 0.00019974659389944947, "loss": 4.7985, "step": 96750 }, { "epoch": 0.18063956295675987, "grad_norm": 1.0293515920639038, "learning_rate": 0.00019974629493000348, "loss": 4.7646, "step": 96800 }, { "epoch": 0.18073286851613837, "grad_norm": 1.054319977760315, "learning_rate": 0.00019974599578452277, "loss": 4.6976, "step": 96850 }, { "epoch": 0.18082617407551685, "grad_norm": 1.062825083732605, "learning_rate": 0.00019974569646300799, "loss": 4.8446, "step": 96900 }, { "epoch": 0.18091947963489535, "grad_norm": 0.7005245089530945, "learning_rate": 0.00019974539696545956, "loss": 4.5875, "step": 96950 }, { "epoch": 0.18101278519427383, "grad_norm": 1.2943309545516968, "learning_rate": 0.0001997450972918781, "loss": 4.6447, "step": 97000 }, { "epoch": 0.18110609075365233, "grad_norm": 0.8457721471786499, "learning_rate": 0.00019974479744226404, "loss": 4.8779, "step": 97050 }, { "epoch": 0.1811993963130308, "grad_norm": 1.1781150102615356, "learning_rate": 0.000199744497416618, "loss": 4.7332, "step": 97100 }, { "epoch": 0.1812927018724093, "grad_norm": 0.9106166958808899, "learning_rate": 0.00019974419721494044, "loss": 4.5838, "step": 97150 }, { "epoch": 0.18138600743178782, "grad_norm": 1.1041089296340942, "learning_rate": 0.00019974389683723195, "loss": 4.5042, "step": 97200 }, { "epoch": 0.1814793129911663, "grad_norm": 0.9294777512550354, "learning_rate": 0.00019974359628349304, "loss": 4.8125, "step": 97250 }, { "epoch": 0.1815726185505448, "grad_norm": 1.158308506011963, "learning_rate": 0.0001997432955537242, "loss": 4.9712, "step": 97300 }, { "epoch": 0.18166592410992327, "grad_norm": 0.9689257740974426, "learning_rate": 0.00019974299464792602, "loss": 4.7698, "step": 97350 }, { "epoch": 0.18175922966930178, "grad_norm": 1.2290277481079102, "learning_rate": 0.00019974269356609902, "loss": 4.585, "step": 97400 }, { "epoch": 0.18185253522868025, "grad_norm": 1.2505154609680176, "learning_rate": 0.00019974239230824368, "loss": 4.7192, "step": 97450 }, { "epoch": 0.18194584078805875, "grad_norm": 1.1231937408447266, "learning_rate": 0.00019974209087436057, "loss": 4.6296, "step": 97500 }, { "epoch": 0.18203914634743726, "grad_norm": 1.0043781995773315, "learning_rate": 0.00019974178926445023, "loss": 4.7468, "step": 97550 }, { "epoch": 0.18213245190681573, "grad_norm": 0.986024796962738, "learning_rate": 0.0001997414874785132, "loss": 4.6579, "step": 97600 }, { "epoch": 0.18222575746619424, "grad_norm": 0.9532257914543152, "learning_rate": 0.00019974118551654998, "loss": 4.8355, "step": 97650 }, { "epoch": 0.18231906302557271, "grad_norm": 1.6420865058898926, "learning_rate": 0.0001997408833785611, "loss": 4.793, "step": 97700 }, { "epoch": 0.18241236858495122, "grad_norm": 1.155208706855774, "learning_rate": 0.00019974058106454713, "loss": 4.7287, "step": 97750 }, { "epoch": 0.1825056741443297, "grad_norm": 0.7323354482650757, "learning_rate": 0.00019974027857450858, "loss": 4.7684, "step": 97800 }, { "epoch": 0.1825989797037082, "grad_norm": 0.9008813500404358, "learning_rate": 0.000199739975908446, "loss": 4.858, "step": 97850 }, { "epoch": 0.1826922852630867, "grad_norm": 1.181888461112976, "learning_rate": 0.0001997396730663599, "loss": 4.7723, "step": 97900 }, { "epoch": 0.18278559082246518, "grad_norm": 0.7667856812477112, "learning_rate": 0.00019973937004825082, "loss": 4.5993, "step": 97950 }, { "epoch": 0.18287889638184368, "grad_norm": 1.1447443962097168, "learning_rate": 0.00019973906685411932, "loss": 4.5683, "step": 98000 }, { "epoch": 0.18297220194122216, "grad_norm": 0.7537021040916443, "learning_rate": 0.00019973876348396589, "loss": 4.7089, "step": 98050 }, { "epoch": 0.18306550750060066, "grad_norm": 0.937552809715271, "learning_rate": 0.00019973845993779112, "loss": 4.8778, "step": 98100 }, { "epoch": 0.18315881305997914, "grad_norm": 1.2293739318847656, "learning_rate": 0.0001997381562155955, "loss": 4.762, "step": 98150 }, { "epoch": 0.18325211861935764, "grad_norm": 1.0932244062423706, "learning_rate": 0.0001997378523173796, "loss": 4.8538, "step": 98200 }, { "epoch": 0.18334542417873612, "grad_norm": 1.1234608888626099, "learning_rate": 0.00019973754824314393, "loss": 4.6881, "step": 98250 }, { "epoch": 0.18343872973811462, "grad_norm": 0.9274517893791199, "learning_rate": 0.00019973724399288903, "loss": 4.8478, "step": 98300 }, { "epoch": 0.18353203529749312, "grad_norm": 0.9792783260345459, "learning_rate": 0.00019973693956661545, "loss": 4.5753, "step": 98350 }, { "epoch": 0.1836253408568716, "grad_norm": 1.3715656995773315, "learning_rate": 0.00019973663496432374, "loss": 4.6483, "step": 98400 }, { "epoch": 0.1837186464162501, "grad_norm": 0.7978438138961792, "learning_rate": 0.00019973633018601438, "loss": 4.6874, "step": 98450 }, { "epoch": 0.18381195197562858, "grad_norm": 0.9657984375953674, "learning_rate": 0.00019973602523168797, "loss": 4.7344, "step": 98500 }, { "epoch": 0.18390525753500708, "grad_norm": 1.0131194591522217, "learning_rate": 0.00019973572010134502, "loss": 4.6663, "step": 98550 }, { "epoch": 0.18399856309438556, "grad_norm": 1.2889825105667114, "learning_rate": 0.00019973541479498608, "loss": 4.6407, "step": 98600 }, { "epoch": 0.18409186865376406, "grad_norm": 1.0131268501281738, "learning_rate": 0.00019973510931261168, "loss": 4.8937, "step": 98650 }, { "epoch": 0.18418517421314257, "grad_norm": 0.9884664416313171, "learning_rate": 0.00019973480365422235, "loss": 4.7954, "step": 98700 }, { "epoch": 0.18427847977252104, "grad_norm": 1.0592337846755981, "learning_rate": 0.00019973449781981865, "loss": 4.7603, "step": 98750 }, { "epoch": 0.18437178533189955, "grad_norm": 0.9447229504585266, "learning_rate": 0.00019973419180940112, "loss": 4.5579, "step": 98800 }, { "epoch": 0.18446509089127802, "grad_norm": 1.186189889907837, "learning_rate": 0.00019973388562297026, "loss": 4.6064, "step": 98850 }, { "epoch": 0.18455839645065653, "grad_norm": 1.156333088874817, "learning_rate": 0.00019973357926052665, "loss": 4.6359, "step": 98900 }, { "epoch": 0.184651702010035, "grad_norm": 1.2002015113830566, "learning_rate": 0.00019973327272207082, "loss": 4.5993, "step": 98950 }, { "epoch": 0.1847450075694135, "grad_norm": 1.1841974258422852, "learning_rate": 0.00019973296600760333, "loss": 4.6125, "step": 99000 }, { "epoch": 0.1847450075694135, "eval_loss": 4.913722991943359, "eval_runtime": 229.7391, "eval_samples_per_second": 11.352, "eval_steps_per_second": 11.352, "eval_tts_loss": 7.435562890975742, "step": 99000 }, { "epoch": 0.184838313128792, "grad_norm": 1.0555998086929321, "learning_rate": 0.00019973265911712472, "loss": 4.7896, "step": 99050 }, { "epoch": 0.18493161868817048, "grad_norm": 1.1442583799362183, "learning_rate": 0.00019973235205063546, "loss": 4.695, "step": 99100 }, { "epoch": 0.185024924247549, "grad_norm": 1.149874210357666, "learning_rate": 0.0001997320448081362, "loss": 4.7661, "step": 99150 }, { "epoch": 0.18511822980692746, "grad_norm": 0.9762941002845764, "learning_rate": 0.00019973173738962742, "loss": 4.6591, "step": 99200 }, { "epoch": 0.18521153536630597, "grad_norm": 1.0889984369277954, "learning_rate": 0.00019973142979510963, "loss": 4.6857, "step": 99250 }, { "epoch": 0.18530484092568444, "grad_norm": 1.1313124895095825, "learning_rate": 0.00019973112202458345, "loss": 4.5218, "step": 99300 }, { "epoch": 0.18539814648506295, "grad_norm": 0.9728614687919617, "learning_rate": 0.0001997308140780494, "loss": 4.6974, "step": 99350 }, { "epoch": 0.18549145204444145, "grad_norm": 1.0745929479599, "learning_rate": 0.000199730505955508, "loss": 4.6735, "step": 99400 }, { "epoch": 0.18558475760381993, "grad_norm": 0.6565387845039368, "learning_rate": 0.00019973019765695977, "loss": 4.6161, "step": 99450 }, { "epoch": 0.18567806316319843, "grad_norm": 1.0615601539611816, "learning_rate": 0.00019972988918240533, "loss": 4.8634, "step": 99500 }, { "epoch": 0.1857713687225769, "grad_norm": 0.948047399520874, "learning_rate": 0.00019972958053184517, "loss": 4.8276, "step": 99550 }, { "epoch": 0.1858646742819554, "grad_norm": 1.0971624851226807, "learning_rate": 0.00019972927170527983, "loss": 4.9053, "step": 99600 }, { "epoch": 0.1859579798413339, "grad_norm": 0.9675891399383545, "learning_rate": 0.0001997289627027099, "loss": 4.7391, "step": 99650 }, { "epoch": 0.1860512854007124, "grad_norm": 0.8968614339828491, "learning_rate": 0.0001997286535241359, "loss": 4.6259, "step": 99700 }, { "epoch": 0.18614459096009087, "grad_norm": 0.772883951663971, "learning_rate": 0.00019972834416955837, "loss": 4.6393, "step": 99750 }, { "epoch": 0.18623789651946937, "grad_norm": 0.9643422365188599, "learning_rate": 0.00019972803463897784, "loss": 4.793, "step": 99800 }, { "epoch": 0.18633120207884787, "grad_norm": 1.0831654071807861, "learning_rate": 0.00019972772493239492, "loss": 4.7557, "step": 99850 }, { "epoch": 0.18642450763822635, "grad_norm": 1.0079478025436401, "learning_rate": 0.00019972741504981008, "loss": 4.4274, "step": 99900 }, { "epoch": 0.18651781319760485, "grad_norm": 0.9231737852096558, "learning_rate": 0.00019972710499122388, "loss": 4.8235, "step": 99950 }, { "epoch": 0.18661111875698333, "grad_norm": 1.2413074970245361, "learning_rate": 0.00019972679475663694, "loss": 4.676, "step": 100000 }, { "epoch": 0.18670442431636183, "grad_norm": 1.0384629964828491, "learning_rate": 0.0001997264843460497, "loss": 4.7109, "step": 100050 }, { "epoch": 0.1867977298757403, "grad_norm": 1.113856554031372, "learning_rate": 0.0001997261737594628, "loss": 4.5818, "step": 100100 }, { "epoch": 0.1868910354351188, "grad_norm": 1.0480083227157593, "learning_rate": 0.00019972586299687675, "loss": 4.6197, "step": 100150 }, { "epoch": 0.18698434099449732, "grad_norm": 1.1027523279190063, "learning_rate": 0.00019972555205829208, "loss": 4.714, "step": 100200 }, { "epoch": 0.1870776465538758, "grad_norm": 1.1065394878387451, "learning_rate": 0.00019972524094370936, "loss": 4.6669, "step": 100250 }, { "epoch": 0.1871709521132543, "grad_norm": 1.1278355121612549, "learning_rate": 0.00019972492965312916, "loss": 4.7749, "step": 100300 }, { "epoch": 0.18726425767263277, "grad_norm": 0.9157338738441467, "learning_rate": 0.00019972461818655197, "loss": 4.886, "step": 100350 }, { "epoch": 0.18735756323201128, "grad_norm": 0.8030517101287842, "learning_rate": 0.0001997243065439784, "loss": 4.7571, "step": 100400 }, { "epoch": 0.18745086879138975, "grad_norm": 0.9438421726226807, "learning_rate": 0.000199723994725409, "loss": 4.8845, "step": 100450 }, { "epoch": 0.18754417435076826, "grad_norm": 1.0755733251571655, "learning_rate": 0.00019972368273084426, "loss": 4.8225, "step": 100500 }, { "epoch": 0.18763747991014676, "grad_norm": 1.0262365341186523, "learning_rate": 0.00019972337056028475, "loss": 4.7597, "step": 100550 }, { "epoch": 0.18773078546952524, "grad_norm": 0.8555694222450256, "learning_rate": 0.00019972305821373108, "loss": 4.6938, "step": 100600 }, { "epoch": 0.18782409102890374, "grad_norm": 0.9016522169113159, "learning_rate": 0.00019972274569118374, "loss": 4.5223, "step": 100650 }, { "epoch": 0.18791739658828222, "grad_norm": 1.187323808670044, "learning_rate": 0.0001997224329926433, "loss": 4.7224, "step": 100700 }, { "epoch": 0.18801070214766072, "grad_norm": 1.1199045181274414, "learning_rate": 0.0001997221201181103, "loss": 4.5454, "step": 100750 }, { "epoch": 0.1881040077070392, "grad_norm": 0.8698449730873108, "learning_rate": 0.00019972180706758535, "loss": 4.7941, "step": 100800 }, { "epoch": 0.1881973132664177, "grad_norm": 1.17020583152771, "learning_rate": 0.00019972149384106893, "loss": 4.5653, "step": 100850 }, { "epoch": 0.18829061882579617, "grad_norm": 1.278030514717102, "learning_rate": 0.00019972118043856162, "loss": 4.4897, "step": 100900 }, { "epoch": 0.18838392438517468, "grad_norm": 0.8267766237258911, "learning_rate": 0.00019972086686006395, "loss": 4.5364, "step": 100950 }, { "epoch": 0.18847722994455318, "grad_norm": 1.0478912591934204, "learning_rate": 0.00019972055310557652, "loss": 4.8908, "step": 101000 }, { "epoch": 0.18857053550393166, "grad_norm": 1.0098049640655518, "learning_rate": 0.00019972023917509986, "loss": 4.6964, "step": 101050 }, { "epoch": 0.18866384106331016, "grad_norm": 0.9657384753227234, "learning_rate": 0.00019971992506863452, "loss": 4.5589, "step": 101100 }, { "epoch": 0.18875714662268864, "grad_norm": 1.05173921585083, "learning_rate": 0.00019971961078618108, "loss": 4.5888, "step": 101150 }, { "epoch": 0.18885045218206714, "grad_norm": 0.9653334617614746, "learning_rate": 0.00019971929632774002, "loss": 4.4529, "step": 101200 }, { "epoch": 0.18894375774144562, "grad_norm": 0.8235491514205933, "learning_rate": 0.000199718981693312, "loss": 4.6389, "step": 101250 }, { "epoch": 0.18903706330082412, "grad_norm": 0.921193540096283, "learning_rate": 0.0001997186668828975, "loss": 4.7759, "step": 101300 }, { "epoch": 0.18913036886020262, "grad_norm": 0.6879051327705383, "learning_rate": 0.0001997183518964971, "loss": 4.9006, "step": 101350 }, { "epoch": 0.1892236744195811, "grad_norm": 1.0606837272644043, "learning_rate": 0.00019971803673411135, "loss": 4.7399, "step": 101400 }, { "epoch": 0.1893169799789596, "grad_norm": 0.8906022906303406, "learning_rate": 0.00019971772139574083, "loss": 4.5369, "step": 101450 }, { "epoch": 0.18941028553833808, "grad_norm": 0.6738720536231995, "learning_rate": 0.0001997174058813861, "loss": 4.7231, "step": 101500 }, { "epoch": 0.18950359109771658, "grad_norm": 1.0936765670776367, "learning_rate": 0.00019971709019104764, "loss": 4.6172, "step": 101550 }, { "epoch": 0.18959689665709506, "grad_norm": 0.9944111108779907, "learning_rate": 0.0001997167743247261, "loss": 4.7858, "step": 101600 }, { "epoch": 0.18969020221647356, "grad_norm": 0.7380058765411377, "learning_rate": 0.00019971645828242195, "loss": 4.8101, "step": 101650 }, { "epoch": 0.18978350777585207, "grad_norm": 1.133588194847107, "learning_rate": 0.00019971614206413586, "loss": 4.5909, "step": 101700 }, { "epoch": 0.18987681333523054, "grad_norm": 0.9205106496810913, "learning_rate": 0.00019971582566986827, "loss": 4.4921, "step": 101750 }, { "epoch": 0.18997011889460905, "grad_norm": 1.0584709644317627, "learning_rate": 0.0001997155090996198, "loss": 4.7779, "step": 101800 }, { "epoch": 0.19006342445398752, "grad_norm": 1.113847017288208, "learning_rate": 0.00019971519235339105, "loss": 4.8153, "step": 101850 }, { "epoch": 0.19015673001336603, "grad_norm": 1.1247295141220093, "learning_rate": 0.00019971487543118247, "loss": 4.8936, "step": 101900 }, { "epoch": 0.1902500355727445, "grad_norm": 1.1859774589538574, "learning_rate": 0.00019971455833299473, "loss": 4.5792, "step": 101950 }, { "epoch": 0.190343341132123, "grad_norm": 0.8291439414024353, "learning_rate": 0.00019971424105882831, "loss": 4.5756, "step": 102000 }, { "epoch": 0.190343341132123, "eval_loss": 4.920861721038818, "eval_runtime": 229.1235, "eval_samples_per_second": 11.383, "eval_steps_per_second": 11.383, "eval_tts_loss": 7.3611460567746265, "step": 102000 }, { "epoch": 0.1904366466915015, "grad_norm": 0.6810999512672424, "learning_rate": 0.0001997139236086838, "loss": 4.8005, "step": 102050 }, { "epoch": 0.19052995225087999, "grad_norm": 1.0597984790802002, "learning_rate": 0.00019971360598256178, "loss": 4.7247, "step": 102100 }, { "epoch": 0.1906232578102585, "grad_norm": 0.9405797719955444, "learning_rate": 0.00019971328818046277, "loss": 4.628, "step": 102150 }, { "epoch": 0.19071656336963697, "grad_norm": 1.037429928779602, "learning_rate": 0.00019971297020238737, "loss": 4.8803, "step": 102200 }, { "epoch": 0.19080986892901547, "grad_norm": 1.3440206050872803, "learning_rate": 0.00019971265204833612, "loss": 4.7855, "step": 102250 }, { "epoch": 0.19090317448839395, "grad_norm": 1.0140005350112915, "learning_rate": 0.00019971233371830958, "loss": 4.6445, "step": 102300 }, { "epoch": 0.19099648004777245, "grad_norm": 1.2359912395477295, "learning_rate": 0.0001997120152123083, "loss": 4.8863, "step": 102350 }, { "epoch": 0.19108978560715092, "grad_norm": 0.9746628403663635, "learning_rate": 0.00019971169653033288, "loss": 4.7716, "step": 102400 }, { "epoch": 0.19118309116652943, "grad_norm": 1.2258669137954712, "learning_rate": 0.00019971137767238386, "loss": 4.9284, "step": 102450 }, { "epoch": 0.19127639672590793, "grad_norm": 0.9276184439659119, "learning_rate": 0.0001997110586384618, "loss": 4.5106, "step": 102500 }, { "epoch": 0.1913697022852864, "grad_norm": 1.0533397197723389, "learning_rate": 0.00019971073942856723, "loss": 4.8912, "step": 102550 }, { "epoch": 0.1914630078446649, "grad_norm": 1.0632177591323853, "learning_rate": 0.00019971042004270081, "loss": 4.764, "step": 102600 }, { "epoch": 0.1915563134040434, "grad_norm": 1.2086039781570435, "learning_rate": 0.000199710100480863, "loss": 4.8496, "step": 102650 }, { "epoch": 0.1916496189634219, "grad_norm": 1.030631184577942, "learning_rate": 0.00019970978074305444, "loss": 4.6516, "step": 102700 }, { "epoch": 0.19174292452280037, "grad_norm": 1.213368535041809, "learning_rate": 0.00019970946082927565, "loss": 4.5534, "step": 102750 }, { "epoch": 0.19183623008217887, "grad_norm": 0.9166072607040405, "learning_rate": 0.0001997091407395272, "loss": 4.6505, "step": 102800 }, { "epoch": 0.19192953564155738, "grad_norm": 1.179938793182373, "learning_rate": 0.00019970882047380965, "loss": 4.8351, "step": 102850 }, { "epoch": 0.19202284120093585, "grad_norm": 0.961173415184021, "learning_rate": 0.00019970850003212358, "loss": 4.7488, "step": 102900 }, { "epoch": 0.19211614676031435, "grad_norm": 0.8187737464904785, "learning_rate": 0.00019970817941446955, "loss": 4.8423, "step": 102950 }, { "epoch": 0.19220945231969283, "grad_norm": 1.1100882291793823, "learning_rate": 0.00019970785862084814, "loss": 4.7505, "step": 103000 }, { "epoch": 0.19230275787907133, "grad_norm": 0.842174232006073, "learning_rate": 0.00019970753765125992, "loss": 4.4804, "step": 103050 }, { "epoch": 0.1923960634384498, "grad_norm": 1.1668494939804077, "learning_rate": 0.0001997072165057054, "loss": 4.6164, "step": 103100 }, { "epoch": 0.19248936899782831, "grad_norm": 1.0573556423187256, "learning_rate": 0.0001997068951841852, "loss": 4.7643, "step": 103150 }, { "epoch": 0.19258267455720682, "grad_norm": 1.0298389196395874, "learning_rate": 0.00019970657368669989, "loss": 4.5873, "step": 103200 }, { "epoch": 0.1926759801165853, "grad_norm": 1.0277085304260254, "learning_rate": 0.00019970625201325002, "loss": 4.6273, "step": 103250 }, { "epoch": 0.1927692856759638, "grad_norm": 0.9712157249450684, "learning_rate": 0.00019970593016383617, "loss": 4.4374, "step": 103300 }, { "epoch": 0.19286259123534227, "grad_norm": 0.8751323223114014, "learning_rate": 0.00019970560813845886, "loss": 4.6056, "step": 103350 }, { "epoch": 0.19295589679472078, "grad_norm": 1.0707224607467651, "learning_rate": 0.00019970528593711874, "loss": 4.9276, "step": 103400 }, { "epoch": 0.19304920235409925, "grad_norm": 1.120894193649292, "learning_rate": 0.00019970496355981628, "loss": 4.5104, "step": 103450 }, { "epoch": 0.19314250791347776, "grad_norm": 0.9309062361717224, "learning_rate": 0.00019970464100655213, "loss": 4.775, "step": 103500 }, { "epoch": 0.19323581347285623, "grad_norm": 1.098646879196167, "learning_rate": 0.00019970431827732684, "loss": 4.6697, "step": 103550 }, { "epoch": 0.19332911903223474, "grad_norm": 1.0255775451660156, "learning_rate": 0.00019970399537214096, "loss": 4.7094, "step": 103600 }, { "epoch": 0.19342242459161324, "grad_norm": 0.9653803110122681, "learning_rate": 0.00019970367229099506, "loss": 4.6016, "step": 103650 }, { "epoch": 0.19351573015099172, "grad_norm": 1.1465661525726318, "learning_rate": 0.00019970334903388976, "loss": 4.892, "step": 103700 }, { "epoch": 0.19360903571037022, "grad_norm": 1.0665031671524048, "learning_rate": 0.00019970302560082557, "loss": 4.8088, "step": 103750 }, { "epoch": 0.1937023412697487, "grad_norm": 1.180091381072998, "learning_rate": 0.00019970270199180307, "loss": 4.7068, "step": 103800 }, { "epoch": 0.1937956468291272, "grad_norm": 0.9207305312156677, "learning_rate": 0.00019970237820682285, "loss": 4.6804, "step": 103850 }, { "epoch": 0.19388895238850568, "grad_norm": 0.6405048370361328, "learning_rate": 0.0001997020542458855, "loss": 4.6396, "step": 103900 }, { "epoch": 0.19398225794788418, "grad_norm": 0.9103243947029114, "learning_rate": 0.00019970173010899154, "loss": 4.7101, "step": 103950 }, { "epoch": 0.19407556350726268, "grad_norm": 1.086352825164795, "learning_rate": 0.00019970140579614158, "loss": 4.8022, "step": 104000 }, { "epoch": 0.19416886906664116, "grad_norm": 0.9352219104766846, "learning_rate": 0.0001997010813073362, "loss": 4.759, "step": 104050 }, { "epoch": 0.19426217462601966, "grad_norm": 0.8962497115135193, "learning_rate": 0.00019970075664257593, "loss": 4.7071, "step": 104100 }, { "epoch": 0.19435548018539814, "grad_norm": 1.1224924325942993, "learning_rate": 0.00019970043180186139, "loss": 4.8363, "step": 104150 }, { "epoch": 0.19444878574477664, "grad_norm": 0.8941819071769714, "learning_rate": 0.00019970010678519308, "loss": 4.7968, "step": 104200 }, { "epoch": 0.19454209130415512, "grad_norm": 1.0338774919509888, "learning_rate": 0.00019969978159257168, "loss": 4.7946, "step": 104250 }, { "epoch": 0.19463539686353362, "grad_norm": 1.0611014366149902, "learning_rate": 0.0001996994562239977, "loss": 4.7344, "step": 104300 }, { "epoch": 0.19472870242291213, "grad_norm": 0.9688970446586609, "learning_rate": 0.00019969913067947168, "loss": 4.6926, "step": 104350 }, { "epoch": 0.1948220079822906, "grad_norm": 0.716977059841156, "learning_rate": 0.00019969880495899426, "loss": 4.775, "step": 104400 }, { "epoch": 0.1949153135416691, "grad_norm": 1.210655927658081, "learning_rate": 0.000199698479062566, "loss": 4.7247, "step": 104450 }, { "epoch": 0.19500861910104758, "grad_norm": 1.2916840314865112, "learning_rate": 0.0001996981529901875, "loss": 4.6479, "step": 104500 }, { "epoch": 0.19510192466042608, "grad_norm": 1.0678060054779053, "learning_rate": 0.00019969782674185923, "loss": 4.7438, "step": 104550 }, { "epoch": 0.19519523021980456, "grad_norm": 0.8926165699958801, "learning_rate": 0.0001996975003175819, "loss": 4.7003, "step": 104600 }, { "epoch": 0.19528853577918306, "grad_norm": 0.8851156234741211, "learning_rate": 0.000199697173717356, "loss": 4.7207, "step": 104650 }, { "epoch": 0.19538184133856157, "grad_norm": 0.8675005435943604, "learning_rate": 0.0001996968469411821, "loss": 4.4109, "step": 104700 }, { "epoch": 0.19547514689794004, "grad_norm": 0.9853196740150452, "learning_rate": 0.00019969651998906084, "loss": 4.6939, "step": 104750 }, { "epoch": 0.19556845245731855, "grad_norm": 0.8662623167037964, "learning_rate": 0.0001996961928609928, "loss": 4.763, "step": 104800 }, { "epoch": 0.19566175801669702, "grad_norm": 1.1778194904327393, "learning_rate": 0.00019969586555697845, "loss": 4.6761, "step": 104850 }, { "epoch": 0.19575506357607553, "grad_norm": 1.2361232042312622, "learning_rate": 0.0001996955380770185, "loss": 4.7453, "step": 104900 }, { "epoch": 0.195848369135454, "grad_norm": 1.2205801010131836, "learning_rate": 0.00019969521042111343, "loss": 4.8602, "step": 104950 }, { "epoch": 0.1959416746948325, "grad_norm": 1.0522173643112183, "learning_rate": 0.00019969488258926384, "loss": 4.8686, "step": 105000 }, { "epoch": 0.1959416746948325, "eval_loss": 4.90353536605835, "eval_runtime": 230.4752, "eval_samples_per_second": 11.316, "eval_steps_per_second": 11.316, "eval_tts_loss": 7.433045758504386, "step": 105000 }, { "epoch": 0.19603498025421098, "grad_norm": 0.8923590183258057, "learning_rate": 0.00019969455458147036, "loss": 4.6476, "step": 105050 }, { "epoch": 0.1961282858135895, "grad_norm": 1.0533981323242188, "learning_rate": 0.00019969422639773353, "loss": 4.7786, "step": 105100 }, { "epoch": 0.196221591372968, "grad_norm": 1.0477826595306396, "learning_rate": 0.00019969389803805391, "loss": 4.6004, "step": 105150 }, { "epoch": 0.19631489693234647, "grad_norm": 1.089428186416626, "learning_rate": 0.00019969356950243214, "loss": 4.6546, "step": 105200 }, { "epoch": 0.19640820249172497, "grad_norm": 0.7010145783424377, "learning_rate": 0.00019969324079086872, "loss": 4.6761, "step": 105250 }, { "epoch": 0.19650150805110345, "grad_norm": 0.9401247501373291, "learning_rate": 0.0001996929119033643, "loss": 4.5199, "step": 105300 }, { "epoch": 0.19659481361048195, "grad_norm": 1.1464163064956665, "learning_rate": 0.00019969258283991943, "loss": 4.7447, "step": 105350 }, { "epoch": 0.19668811916986043, "grad_norm": 0.9506011605262756, "learning_rate": 0.0001996922536005347, "loss": 4.9001, "step": 105400 }, { "epoch": 0.19678142472923893, "grad_norm": 1.0676487684249878, "learning_rate": 0.00019969192418521064, "loss": 4.5418, "step": 105450 }, { "epoch": 0.19687473028861743, "grad_norm": 1.0197359323501587, "learning_rate": 0.0001996915945939479, "loss": 4.706, "step": 105500 }, { "epoch": 0.1969680358479959, "grad_norm": 1.120069980621338, "learning_rate": 0.00019969126482674706, "loss": 4.6804, "step": 105550 }, { "epoch": 0.1970613414073744, "grad_norm": 0.9666004776954651, "learning_rate": 0.00019969093488360863, "loss": 4.792, "step": 105600 }, { "epoch": 0.1971546469667529, "grad_norm": 1.1865931749343872, "learning_rate": 0.0001996906047645333, "loss": 4.6457, "step": 105650 }, { "epoch": 0.1972479525261314, "grad_norm": 0.9605655074119568, "learning_rate": 0.00019969027446952155, "loss": 4.8505, "step": 105700 }, { "epoch": 0.19734125808550987, "grad_norm": 0.6585858464241028, "learning_rate": 0.00019968994399857404, "loss": 4.5811, "step": 105750 }, { "epoch": 0.19743456364488837, "grad_norm": 1.0499614477157593, "learning_rate": 0.0001996896133516913, "loss": 4.661, "step": 105800 }, { "epoch": 0.19752786920426688, "grad_norm": 1.0251386165618896, "learning_rate": 0.0001996892825288739, "loss": 4.8558, "step": 105850 }, { "epoch": 0.19762117476364535, "grad_norm": 1.2540901899337769, "learning_rate": 0.00019968895153012252, "loss": 4.5407, "step": 105900 }, { "epoch": 0.19771448032302386, "grad_norm": 0.6922102570533752, "learning_rate": 0.00019968862035543765, "loss": 4.5388, "step": 105950 }, { "epoch": 0.19780778588240233, "grad_norm": 1.0659449100494385, "learning_rate": 0.00019968828900481992, "loss": 4.6475, "step": 106000 }, { "epoch": 0.19790109144178084, "grad_norm": 0.8389386534690857, "learning_rate": 0.00019968795747826986, "loss": 4.555, "step": 106050 }, { "epoch": 0.1979943970011593, "grad_norm": 0.8430420160293579, "learning_rate": 0.00019968762577578814, "loss": 4.7839, "step": 106100 }, { "epoch": 0.19808770256053781, "grad_norm": 1.0360347032546997, "learning_rate": 0.00019968729389737528, "loss": 4.7264, "step": 106150 }, { "epoch": 0.19818100811991632, "grad_norm": 0.9396027326583862, "learning_rate": 0.0001996869618430319, "loss": 4.5931, "step": 106200 }, { "epoch": 0.1982743136792948, "grad_norm": 0.8887374997138977, "learning_rate": 0.00019968662961275855, "loss": 4.6601, "step": 106250 }, { "epoch": 0.1983676192386733, "grad_norm": 1.1754348278045654, "learning_rate": 0.00019968629720655586, "loss": 4.7664, "step": 106300 }, { "epoch": 0.19846092479805177, "grad_norm": 0.6698839068412781, "learning_rate": 0.0001996859646244244, "loss": 4.5077, "step": 106350 }, { "epoch": 0.19855423035743028, "grad_norm": 0.944202184677124, "learning_rate": 0.00019968563186636474, "loss": 4.6335, "step": 106400 }, { "epoch": 0.19864753591680875, "grad_norm": 1.1213611364364624, "learning_rate": 0.00019968529893237745, "loss": 4.6235, "step": 106450 }, { "epoch": 0.19874084147618726, "grad_norm": 1.068108081817627, "learning_rate": 0.0001996849658224632, "loss": 4.6891, "step": 106500 }, { "epoch": 0.19883414703556573, "grad_norm": 1.1877700090408325, "learning_rate": 0.0001996846325366225, "loss": 4.5616, "step": 106550 }, { "epoch": 0.19892745259494424, "grad_norm": 0.9596836566925049, "learning_rate": 0.00019968429907485597, "loss": 4.612, "step": 106600 }, { "epoch": 0.19902075815432274, "grad_norm": 0.8319620490074158, "learning_rate": 0.0001996839654371642, "loss": 4.6712, "step": 106650 }, { "epoch": 0.19911406371370122, "grad_norm": 0.921432375907898, "learning_rate": 0.00019968363162354774, "loss": 4.7998, "step": 106700 }, { "epoch": 0.19920736927307972, "grad_norm": 1.089003324508667, "learning_rate": 0.00019968329763400726, "loss": 4.5917, "step": 106750 }, { "epoch": 0.1993006748324582, "grad_norm": 1.233829140663147, "learning_rate": 0.00019968296346854327, "loss": 4.8402, "step": 106800 }, { "epoch": 0.1993939803918367, "grad_norm": 1.0184001922607422, "learning_rate": 0.0001996826291271564, "loss": 4.7715, "step": 106850 }, { "epoch": 0.19948728595121518, "grad_norm": 1.058369517326355, "learning_rate": 0.0001996822946098472, "loss": 4.6193, "step": 106900 }, { "epoch": 0.19958059151059368, "grad_norm": 1.2557109594345093, "learning_rate": 0.00019968195991661633, "loss": 4.9265, "step": 106950 }, { "epoch": 0.19967389706997218, "grad_norm": 0.7956662178039551, "learning_rate": 0.00019968162504746434, "loss": 4.8568, "step": 107000 }, { "epoch": 0.19976720262935066, "grad_norm": 0.9638285040855408, "learning_rate": 0.00019968129000239179, "loss": 4.6949, "step": 107050 }, { "epoch": 0.19986050818872916, "grad_norm": 1.0805922746658325, "learning_rate": 0.0001996809547813993, "loss": 4.7621, "step": 107100 }, { "epoch": 0.19995381374810764, "grad_norm": 0.9019108414649963, "learning_rate": 0.00019968061938448748, "loss": 4.4901, "step": 107150 }, { "epoch": 0.20004711930748614, "grad_norm": 0.7746037840843201, "learning_rate": 0.0001996802838116569, "loss": 4.4827, "step": 107200 }, { "epoch": 0.20014042486686462, "grad_norm": 1.065016508102417, "learning_rate": 0.00019967994806290818, "loss": 4.8291, "step": 107250 }, { "epoch": 0.20023373042624312, "grad_norm": 1.0848886966705322, "learning_rate": 0.00019967961213824187, "loss": 4.7838, "step": 107300 }, { "epoch": 0.20032703598562163, "grad_norm": 0.9382323026657104, "learning_rate": 0.0001996792760376586, "loss": 4.6825, "step": 107350 }, { "epoch": 0.2004203415450001, "grad_norm": 1.0932012796401978, "learning_rate": 0.00019967893976115893, "loss": 4.8105, "step": 107400 }, { "epoch": 0.2005136471043786, "grad_norm": 0.8511140942573547, "learning_rate": 0.00019967860330874344, "loss": 4.6117, "step": 107450 }, { "epoch": 0.20060695266375708, "grad_norm": 1.1722475290298462, "learning_rate": 0.00019967826668041281, "loss": 4.6226, "step": 107500 }, { "epoch": 0.20070025822313559, "grad_norm": 1.120883822441101, "learning_rate": 0.00019967792987616757, "loss": 4.8838, "step": 107550 }, { "epoch": 0.20079356378251406, "grad_norm": 0.9528012871742249, "learning_rate": 0.0001996775928960083, "loss": 4.7664, "step": 107600 }, { "epoch": 0.20088686934189257, "grad_norm": 0.9668023586273193, "learning_rate": 0.0001996772557399356, "loss": 4.6156, "step": 107650 }, { "epoch": 0.20098017490127104, "grad_norm": 0.8092545866966248, "learning_rate": 0.00019967691840795014, "loss": 4.8754, "step": 107700 }, { "epoch": 0.20107348046064955, "grad_norm": 1.1513761281967163, "learning_rate": 0.0001996765809000524, "loss": 4.7403, "step": 107750 }, { "epoch": 0.20116678602002805, "grad_norm": 1.2313563823699951, "learning_rate": 0.00019967624321624307, "loss": 4.7554, "step": 107800 }, { "epoch": 0.20126009157940652, "grad_norm": 1.2222445011138916, "learning_rate": 0.0001996759053565227, "loss": 4.6705, "step": 107850 }, { "epoch": 0.20135339713878503, "grad_norm": 1.0731276273727417, "learning_rate": 0.00019967556732089189, "loss": 4.7103, "step": 107900 }, { "epoch": 0.2014467026981635, "grad_norm": 0.8914058804512024, "learning_rate": 0.00019967522910935127, "loss": 4.7109, "step": 107950 }, { "epoch": 0.201540008257542, "grad_norm": 0.9752347469329834, "learning_rate": 0.00019967489072190137, "loss": 4.8808, "step": 108000 }, { "epoch": 0.201540008257542, "eval_loss": 4.898900508880615, "eval_runtime": 231.0932, "eval_samples_per_second": 11.285, "eval_steps_per_second": 11.285, "eval_tts_loss": 7.3902602159986674, "step": 108000 }, { "epoch": 0.20163331381692048, "grad_norm": 1.1584863662719727, "learning_rate": 0.00019967455215854285, "loss": 4.6683, "step": 108050 }, { "epoch": 0.201726619376299, "grad_norm": 1.2001816034317017, "learning_rate": 0.00019967421341927628, "loss": 4.6347, "step": 108100 }, { "epoch": 0.2018199249356775, "grad_norm": 1.0059622526168823, "learning_rate": 0.00019967387450410225, "loss": 4.9153, "step": 108150 }, { "epoch": 0.20191323049505597, "grad_norm": 0.6907665729522705, "learning_rate": 0.00019967353541302138, "loss": 4.6052, "step": 108200 }, { "epoch": 0.20200653605443447, "grad_norm": 0.9193375706672668, "learning_rate": 0.00019967319614603427, "loss": 4.7746, "step": 108250 }, { "epoch": 0.20209984161381295, "grad_norm": 0.8793426752090454, "learning_rate": 0.00019967285670314147, "loss": 4.6992, "step": 108300 }, { "epoch": 0.20219314717319145, "grad_norm": 1.1513392925262451, "learning_rate": 0.00019967251708434365, "loss": 4.9056, "step": 108350 }, { "epoch": 0.20228645273256993, "grad_norm": 1.0788843631744385, "learning_rate": 0.00019967217728964135, "loss": 4.6675, "step": 108400 }, { "epoch": 0.20237975829194843, "grad_norm": 0.9058928489685059, "learning_rate": 0.0001996718373190352, "loss": 4.7488, "step": 108450 }, { "epoch": 0.20247306385132693, "grad_norm": 0.7392362952232361, "learning_rate": 0.0001996714971725258, "loss": 4.6251, "step": 108500 }, { "epoch": 0.2025663694107054, "grad_norm": 0.9372434616088867, "learning_rate": 0.00019967115685011377, "loss": 4.9667, "step": 108550 }, { "epoch": 0.20265967497008391, "grad_norm": 1.1257874965667725, "learning_rate": 0.00019967081635179968, "loss": 4.7603, "step": 108600 }, { "epoch": 0.2027529805294624, "grad_norm": 0.8157814741134644, "learning_rate": 0.0001996704756775841, "loss": 4.5966, "step": 108650 }, { "epoch": 0.2028462860888409, "grad_norm": 1.0224437713623047, "learning_rate": 0.0001996701348274677, "loss": 4.6797, "step": 108700 }, { "epoch": 0.20293959164821937, "grad_norm": 1.2594748735427856, "learning_rate": 0.00019966979380145103, "loss": 4.6246, "step": 108750 }, { "epoch": 0.20303289720759787, "grad_norm": 0.936982274055481, "learning_rate": 0.00019966945259953473, "loss": 4.7162, "step": 108800 }, { "epoch": 0.20312620276697638, "grad_norm": 1.1596720218658447, "learning_rate": 0.00019966911122171936, "loss": 4.6277, "step": 108850 }, { "epoch": 0.20321950832635485, "grad_norm": 0.9620667099952698, "learning_rate": 0.00019966876966800556, "loss": 4.7468, "step": 108900 }, { "epoch": 0.20331281388573336, "grad_norm": 1.2302641868591309, "learning_rate": 0.00019966842793839394, "loss": 4.6094, "step": 108950 }, { "epoch": 0.20340611944511183, "grad_norm": 1.131081461906433, "learning_rate": 0.00019966808603288504, "loss": 4.628, "step": 109000 }, { "epoch": 0.20349942500449034, "grad_norm": 0.9959622621536255, "learning_rate": 0.00019966774395147955, "loss": 4.8578, "step": 109050 }, { "epoch": 0.2035927305638688, "grad_norm": 1.1797744035720825, "learning_rate": 0.00019966740169417803, "loss": 4.8179, "step": 109100 }, { "epoch": 0.20368603612324732, "grad_norm": 1.0603262186050415, "learning_rate": 0.00019966705926098102, "loss": 4.6757, "step": 109150 }, { "epoch": 0.2037793416826258, "grad_norm": 1.2537606954574585, "learning_rate": 0.00019966671665188924, "loss": 4.5657, "step": 109200 }, { "epoch": 0.2038726472420043, "grad_norm": 0.8469638228416443, "learning_rate": 0.0001996663738669032, "loss": 4.5299, "step": 109250 }, { "epoch": 0.2039659528013828, "grad_norm": 1.1526724100112915, "learning_rate": 0.0001996660309060236, "loss": 4.6851, "step": 109300 }, { "epoch": 0.20405925836076128, "grad_norm": 0.9065315127372742, "learning_rate": 0.00019966568776925097, "loss": 4.6405, "step": 109350 }, { "epoch": 0.20415256392013978, "grad_norm": 1.0660070180892944, "learning_rate": 0.00019966534445658591, "loss": 4.8246, "step": 109400 }, { "epoch": 0.20424586947951825, "grad_norm": 0.9208210110664368, "learning_rate": 0.0001996650009680291, "loss": 4.8364, "step": 109450 }, { "epoch": 0.20433917503889676, "grad_norm": 1.1117137670516968, "learning_rate": 0.00019966465730358106, "loss": 4.6232, "step": 109500 }, { "epoch": 0.20443248059827523, "grad_norm": 0.7682850360870361, "learning_rate": 0.00019966431346324246, "loss": 4.6112, "step": 109550 }, { "epoch": 0.20452578615765374, "grad_norm": 0.9791049957275391, "learning_rate": 0.0001996639694470139, "loss": 4.739, "step": 109600 }, { "epoch": 0.20461909171703224, "grad_norm": 0.7025151252746582, "learning_rate": 0.00019966362525489592, "loss": 4.625, "step": 109650 }, { "epoch": 0.20471239727641072, "grad_norm": 0.9386808276176453, "learning_rate": 0.00019966328088688923, "loss": 4.7254, "step": 109700 }, { "epoch": 0.20480570283578922, "grad_norm": 0.9682460427284241, "learning_rate": 0.00019966293634299435, "loss": 4.6955, "step": 109750 }, { "epoch": 0.2048990083951677, "grad_norm": 0.9403902292251587, "learning_rate": 0.00019966259162321196, "loss": 4.648, "step": 109800 }, { "epoch": 0.2049923139545462, "grad_norm": 0.8865294456481934, "learning_rate": 0.00019966224672754259, "loss": 4.6593, "step": 109850 }, { "epoch": 0.20508561951392468, "grad_norm": 0.9685643315315247, "learning_rate": 0.0001996619016559869, "loss": 4.8078, "step": 109900 }, { "epoch": 0.20517892507330318, "grad_norm": 1.1610651016235352, "learning_rate": 0.00019966155640854552, "loss": 4.7153, "step": 109950 }, { "epoch": 0.20527223063268168, "grad_norm": 1.3731842041015625, "learning_rate": 0.000199661210985219, "loss": 4.6649, "step": 110000 }, { "epoch": 0.20536553619206016, "grad_norm": 0.9859686493873596, "learning_rate": 0.00019966086538600797, "loss": 4.649, "step": 110050 }, { "epoch": 0.20545884175143866, "grad_norm": 0.9828516244888306, "learning_rate": 0.00019966051961091305, "loss": 4.6669, "step": 110100 }, { "epoch": 0.20555214731081714, "grad_norm": 1.1860650777816772, "learning_rate": 0.00019966017365993487, "loss": 4.6617, "step": 110150 }, { "epoch": 0.20564545287019564, "grad_norm": 0.7568049430847168, "learning_rate": 0.00019965982753307403, "loss": 4.5716, "step": 110200 }, { "epoch": 0.20573875842957412, "grad_norm": 1.4290099143981934, "learning_rate": 0.0001996594812303311, "loss": 4.7415, "step": 110250 }, { "epoch": 0.20583206398895262, "grad_norm": 1.090683937072754, "learning_rate": 0.00019965913475170673, "loss": 4.7548, "step": 110300 }, { "epoch": 0.2059253695483311, "grad_norm": 1.1351553201675415, "learning_rate": 0.00019965878809720153, "loss": 4.6024, "step": 110350 }, { "epoch": 0.2060186751077096, "grad_norm": 0.8278642296791077, "learning_rate": 0.0001996584412668161, "loss": 4.7692, "step": 110400 }, { "epoch": 0.2061119806670881, "grad_norm": 0.9527729749679565, "learning_rate": 0.00019965809426055104, "loss": 4.6836, "step": 110450 }, { "epoch": 0.20620528622646658, "grad_norm": 0.9294393062591553, "learning_rate": 0.000199657747078407, "loss": 4.7799, "step": 110500 }, { "epoch": 0.2062985917858451, "grad_norm": 0.8470353484153748, "learning_rate": 0.00019965739972038456, "loss": 4.6028, "step": 110550 }, { "epoch": 0.20639189734522356, "grad_norm": 1.0058228969573975, "learning_rate": 0.00019965705218648436, "loss": 4.7502, "step": 110600 }, { "epoch": 0.20648520290460207, "grad_norm": 1.0412211418151855, "learning_rate": 0.000199656704476707, "loss": 4.7988, "step": 110650 }, { "epoch": 0.20657850846398054, "grad_norm": 1.1575844287872314, "learning_rate": 0.00019965635659105306, "loss": 4.905, "step": 110700 }, { "epoch": 0.20667181402335905, "grad_norm": 1.1507608890533447, "learning_rate": 0.0001996560085295232, "loss": 4.708, "step": 110750 }, { "epoch": 0.20676511958273755, "grad_norm": 1.1214720010757446, "learning_rate": 0.00019965566029211807, "loss": 4.7185, "step": 110800 }, { "epoch": 0.20685842514211603, "grad_norm": 0.8996137380599976, "learning_rate": 0.00019965531187883816, "loss": 4.6913, "step": 110850 }, { "epoch": 0.20695173070149453, "grad_norm": 1.0486303567886353, "learning_rate": 0.0001996549632896842, "loss": 4.9886, "step": 110900 }, { "epoch": 0.207045036260873, "grad_norm": 0.9184983968734741, "learning_rate": 0.00019965461452465676, "loss": 4.4322, "step": 110950 }, { "epoch": 0.2071383418202515, "grad_norm": 1.004927396774292, "learning_rate": 0.00019965426558375647, "loss": 4.7627, "step": 111000 }, { "epoch": 0.2071383418202515, "eval_loss": 4.893270015716553, "eval_runtime": 228.2153, "eval_samples_per_second": 11.428, "eval_steps_per_second": 11.428, "eval_tts_loss": 7.3785387509319635, "step": 111000 }, { "epoch": 0.20723164737962999, "grad_norm": 1.0662178993225098, "learning_rate": 0.0001996539164669839, "loss": 4.5831, "step": 111050 }, { "epoch": 0.2073249529390085, "grad_norm": 1.0260896682739258, "learning_rate": 0.00019965356717433971, "loss": 4.6671, "step": 111100 }, { "epoch": 0.207418258498387, "grad_norm": 0.8488427400588989, "learning_rate": 0.00019965321770582455, "loss": 4.8138, "step": 111150 }, { "epoch": 0.20751156405776547, "grad_norm": 0.9575690627098083, "learning_rate": 0.00019965286806143898, "loss": 4.5885, "step": 111200 }, { "epoch": 0.20760486961714397, "grad_norm": 0.9709866046905518, "learning_rate": 0.00019965251824118361, "loss": 4.721, "step": 111250 }, { "epoch": 0.20769817517652245, "grad_norm": 1.058888554573059, "learning_rate": 0.00019965216824505908, "loss": 4.681, "step": 111300 }, { "epoch": 0.20779148073590095, "grad_norm": 0.9510217905044556, "learning_rate": 0.00019965181807306604, "loss": 4.84, "step": 111350 }, { "epoch": 0.20788478629527943, "grad_norm": 1.249549388885498, "learning_rate": 0.00019965146772520505, "loss": 4.5905, "step": 111400 }, { "epoch": 0.20797809185465793, "grad_norm": 0.9404703378677368, "learning_rate": 0.00019965111720147673, "loss": 4.7374, "step": 111450 }, { "epoch": 0.20807139741403644, "grad_norm": 0.9950361251831055, "learning_rate": 0.00019965076650188177, "loss": 4.7764, "step": 111500 }, { "epoch": 0.2081647029734149, "grad_norm": 1.015062689781189, "learning_rate": 0.0001996504156264207, "loss": 4.7205, "step": 111550 }, { "epoch": 0.20825800853279341, "grad_norm": 1.0537135601043701, "learning_rate": 0.00019965006457509422, "loss": 4.7283, "step": 111600 }, { "epoch": 0.2083513140921719, "grad_norm": 0.9111753106117249, "learning_rate": 0.00019964971334790287, "loss": 4.5574, "step": 111650 }, { "epoch": 0.2084446196515504, "grad_norm": 1.2378709316253662, "learning_rate": 0.00019964936194484737, "loss": 4.5554, "step": 111700 }, { "epoch": 0.20853792521092887, "grad_norm": 0.8984060883522034, "learning_rate": 0.00019964901036592823, "loss": 4.6334, "step": 111750 }, { "epoch": 0.20863123077030737, "grad_norm": 0.6208471655845642, "learning_rate": 0.0001996486586111461, "loss": 4.5307, "step": 111800 }, { "epoch": 0.20872453632968585, "grad_norm": 1.0021463632583618, "learning_rate": 0.00019964830668050168, "loss": 4.8079, "step": 111850 }, { "epoch": 0.20881784188906435, "grad_norm": 0.6217372417449951, "learning_rate": 0.00019964795457399549, "loss": 4.5041, "step": 111900 }, { "epoch": 0.20891114744844286, "grad_norm": 1.1446421146392822, "learning_rate": 0.00019964760229162823, "loss": 4.656, "step": 111950 }, { "epoch": 0.20900445300782133, "grad_norm": 0.9120539426803589, "learning_rate": 0.00019964724983340045, "loss": 4.6973, "step": 112000 }, { "epoch": 0.20909775856719984, "grad_norm": 0.6660919189453125, "learning_rate": 0.0001996468971993128, "loss": 4.584, "step": 112050 }, { "epoch": 0.2091910641265783, "grad_norm": 1.0424081087112427, "learning_rate": 0.00019964654438936594, "loss": 4.7912, "step": 112100 }, { "epoch": 0.20928436968595682, "grad_norm": 0.8505585789680481, "learning_rate": 0.00019964619140356047, "loss": 4.6274, "step": 112150 }, { "epoch": 0.2093776752453353, "grad_norm": 1.0397758483886719, "learning_rate": 0.00019964583824189699, "loss": 4.4355, "step": 112200 }, { "epoch": 0.2094709808047138, "grad_norm": 1.0277235507965088, "learning_rate": 0.00019964548490437612, "loss": 4.7227, "step": 112250 }, { "epoch": 0.2095642863640923, "grad_norm": 1.0164071321487427, "learning_rate": 0.0001996451313909985, "loss": 4.6742, "step": 112300 }, { "epoch": 0.20965759192347078, "grad_norm": 1.1814101934432983, "learning_rate": 0.0001996447777017648, "loss": 4.5611, "step": 112350 }, { "epoch": 0.20975089748284928, "grad_norm": 0.9306598901748657, "learning_rate": 0.00019964442383667557, "loss": 4.749, "step": 112400 }, { "epoch": 0.20984420304222776, "grad_norm": 0.7609649896621704, "learning_rate": 0.0001996440697957315, "loss": 4.8032, "step": 112450 }, { "epoch": 0.20993750860160626, "grad_norm": 0.7726313471794128, "learning_rate": 0.00019964371557893313, "loss": 4.6712, "step": 112500 }, { "epoch": 0.21003081416098474, "grad_norm": 0.718001127243042, "learning_rate": 0.00019964336118628116, "loss": 4.3786, "step": 112550 }, { "epoch": 0.21012411972036324, "grad_norm": 1.221449375152588, "learning_rate": 0.00019964300661777618, "loss": 4.6134, "step": 112600 }, { "epoch": 0.21021742527974174, "grad_norm": 0.9702039361000061, "learning_rate": 0.0001996426518734188, "loss": 4.4521, "step": 112650 }, { "epoch": 0.21031073083912022, "grad_norm": 1.0901730060577393, "learning_rate": 0.00019964229695320973, "loss": 4.7553, "step": 112700 }, { "epoch": 0.21040403639849872, "grad_norm": 1.4701701402664185, "learning_rate": 0.00019964194185714953, "loss": 4.4546, "step": 112750 }, { "epoch": 0.2104973419578772, "grad_norm": 1.0534309148788452, "learning_rate": 0.0001996415865852388, "loss": 4.618, "step": 112800 }, { "epoch": 0.2105906475172557, "grad_norm": 0.875619649887085, "learning_rate": 0.00019964123113747824, "loss": 4.7289, "step": 112850 }, { "epoch": 0.21068395307663418, "grad_norm": 1.1053240299224854, "learning_rate": 0.00019964087551386841, "loss": 4.6377, "step": 112900 }, { "epoch": 0.21077725863601268, "grad_norm": 1.1777266263961792, "learning_rate": 0.00019964051971440997, "loss": 4.9208, "step": 112950 }, { "epoch": 0.21087056419539116, "grad_norm": 0.9478060007095337, "learning_rate": 0.00019964016373910358, "loss": 4.5935, "step": 113000 }, { "epoch": 0.21096386975476966, "grad_norm": 1.104618787765503, "learning_rate": 0.0001996398075879498, "loss": 4.7408, "step": 113050 }, { "epoch": 0.21105717531414817, "grad_norm": 0.9139821529388428, "learning_rate": 0.00019963945126094928, "loss": 4.8311, "step": 113100 }, { "epoch": 0.21115048087352664, "grad_norm": 1.1723382472991943, "learning_rate": 0.00019963909475810267, "loss": 4.7919, "step": 113150 }, { "epoch": 0.21124378643290515, "grad_norm": 1.1692183017730713, "learning_rate": 0.0001996387380794106, "loss": 4.8049, "step": 113200 }, { "epoch": 0.21133709199228362, "grad_norm": 0.8972973823547363, "learning_rate": 0.0001996383812248737, "loss": 4.6703, "step": 113250 }, { "epoch": 0.21143039755166212, "grad_norm": 0.8258083462715149, "learning_rate": 0.00019963802419449258, "loss": 4.6485, "step": 113300 }, { "epoch": 0.2115237031110406, "grad_norm": 1.101173758506775, "learning_rate": 0.00019963766698826785, "loss": 4.5944, "step": 113350 }, { "epoch": 0.2116170086704191, "grad_norm": 0.9222469925880432, "learning_rate": 0.00019963730960620023, "loss": 4.6801, "step": 113400 }, { "epoch": 0.2117103142297976, "grad_norm": 0.8024877309799194, "learning_rate": 0.00019963695204829022, "loss": 4.6408, "step": 113450 }, { "epoch": 0.21180361978917608, "grad_norm": 1.1959446668624878, "learning_rate": 0.00019963659431453858, "loss": 4.7532, "step": 113500 }, { "epoch": 0.2118969253485546, "grad_norm": 1.03960382938385, "learning_rate": 0.00019963623640494586, "loss": 4.531, "step": 113550 }, { "epoch": 0.21199023090793306, "grad_norm": 0.9562681317329407, "learning_rate": 0.0001996358783195127, "loss": 4.525, "step": 113600 }, { "epoch": 0.21208353646731157, "grad_norm": 0.9814481735229492, "learning_rate": 0.00019963552005823975, "loss": 4.5483, "step": 113650 }, { "epoch": 0.21217684202669004, "grad_norm": 0.948824942111969, "learning_rate": 0.00019963516162112765, "loss": 4.8581, "step": 113700 }, { "epoch": 0.21227014758606855, "grad_norm": 1.0382654666900635, "learning_rate": 0.00019963480300817704, "loss": 4.7724, "step": 113750 }, { "epoch": 0.21236345314544705, "grad_norm": 1.0841362476348877, "learning_rate": 0.00019963444421938848, "loss": 4.8569, "step": 113800 }, { "epoch": 0.21245675870482553, "grad_norm": 1.138662338256836, "learning_rate": 0.00019963408525476272, "loss": 4.828, "step": 113850 }, { "epoch": 0.21255006426420403, "grad_norm": 0.9034907817840576, "learning_rate": 0.00019963372611430028, "loss": 4.7168, "step": 113900 }, { "epoch": 0.2126433698235825, "grad_norm": 0.9675794243812561, "learning_rate": 0.0001996333667980019, "loss": 4.9587, "step": 113950 }, { "epoch": 0.212736675382961, "grad_norm": 0.9928723573684692, "learning_rate": 0.0001996330073058681, "loss": 4.6761, "step": 114000 }, { "epoch": 0.212736675382961, "eval_loss": 4.887726783752441, "eval_runtime": 228.3892, "eval_samples_per_second": 11.419, "eval_steps_per_second": 11.419, "eval_tts_loss": 7.461525186266396, "step": 114000 }, { "epoch": 0.2128299809423395, "grad_norm": 0.9984357953071594, "learning_rate": 0.0001996326476378996, "loss": 4.7151, "step": 114050 }, { "epoch": 0.212923286501718, "grad_norm": 0.9879376888275146, "learning_rate": 0.000199632287794097, "loss": 4.5944, "step": 114100 }, { "epoch": 0.2130165920610965, "grad_norm": 0.9942005276679993, "learning_rate": 0.00019963192777446094, "loss": 4.676, "step": 114150 }, { "epoch": 0.21310989762047497, "grad_norm": 1.000113606452942, "learning_rate": 0.00019963156757899205, "loss": 4.5756, "step": 114200 }, { "epoch": 0.21320320317985347, "grad_norm": 1.1187829971313477, "learning_rate": 0.000199631207207691, "loss": 4.5309, "step": 114250 }, { "epoch": 0.21329650873923195, "grad_norm": 1.0895246267318726, "learning_rate": 0.0001996308466605584, "loss": 4.4908, "step": 114300 }, { "epoch": 0.21338981429861045, "grad_norm": 0.7784030437469482, "learning_rate": 0.00019963048593759486, "loss": 4.4745, "step": 114350 }, { "epoch": 0.21348311985798893, "grad_norm": 1.0729563236236572, "learning_rate": 0.00019963012503880107, "loss": 4.5451, "step": 114400 }, { "epoch": 0.21357642541736743, "grad_norm": 1.1839163303375244, "learning_rate": 0.00019962976396417764, "loss": 4.7851, "step": 114450 }, { "epoch": 0.2136697309767459, "grad_norm": 1.1860408782958984, "learning_rate": 0.00019962940271372518, "loss": 4.6238, "step": 114500 }, { "epoch": 0.2137630365361244, "grad_norm": 0.9122569561004639, "learning_rate": 0.00019962904128744438, "loss": 4.6063, "step": 114550 }, { "epoch": 0.21385634209550292, "grad_norm": 0.8689755797386169, "learning_rate": 0.00019962867968533585, "loss": 4.5032, "step": 114600 }, { "epoch": 0.2139496476548814, "grad_norm": 0.8875397443771362, "learning_rate": 0.0001996283179074002, "loss": 4.7033, "step": 114650 }, { "epoch": 0.2140429532142599, "grad_norm": 0.7156913876533508, "learning_rate": 0.00019962795595363813, "loss": 4.8056, "step": 114700 }, { "epoch": 0.21413625877363837, "grad_norm": 1.1040927171707153, "learning_rate": 0.00019962759382405025, "loss": 4.6738, "step": 114750 }, { "epoch": 0.21422956433301688, "grad_norm": 0.9362927675247192, "learning_rate": 0.00019962723151863718, "loss": 4.8287, "step": 114800 }, { "epoch": 0.21432286989239535, "grad_norm": 0.8515648245811462, "learning_rate": 0.00019962686903739958, "loss": 4.4636, "step": 114850 }, { "epoch": 0.21441617545177385, "grad_norm": 1.0025912523269653, "learning_rate": 0.0001996265063803381, "loss": 4.7396, "step": 114900 }, { "epoch": 0.21450948101115236, "grad_norm": 0.7344082593917847, "learning_rate": 0.00019962614354745336, "loss": 4.5112, "step": 114950 }, { "epoch": 0.21460278657053083, "grad_norm": 1.0541155338287354, "learning_rate": 0.00019962578053874598, "loss": 4.6548, "step": 115000 }, { "epoch": 0.21469609212990934, "grad_norm": 0.9491979479789734, "learning_rate": 0.00019962541735421666, "loss": 4.7796, "step": 115050 }, { "epoch": 0.21478939768928781, "grad_norm": 0.8249531388282776, "learning_rate": 0.000199625053993866, "loss": 4.7001, "step": 115100 }, { "epoch": 0.21488270324866632, "grad_norm": 0.862076461315155, "learning_rate": 0.00019962469045769465, "loss": 4.4974, "step": 115150 }, { "epoch": 0.2149760088080448, "grad_norm": 1.1294341087341309, "learning_rate": 0.00019962432674570325, "loss": 4.6363, "step": 115200 }, { "epoch": 0.2150693143674233, "grad_norm": 0.7202677130699158, "learning_rate": 0.0001996239628578924, "loss": 4.6216, "step": 115250 }, { "epoch": 0.2151626199268018, "grad_norm": 0.8325316905975342, "learning_rate": 0.00019962359879426285, "loss": 4.6454, "step": 115300 }, { "epoch": 0.21525592548618028, "grad_norm": 1.1335790157318115, "learning_rate": 0.00019962323455481514, "loss": 4.6516, "step": 115350 }, { "epoch": 0.21534923104555878, "grad_norm": 0.8559627532958984, "learning_rate": 0.00019962287013954998, "loss": 4.7461, "step": 115400 }, { "epoch": 0.21544253660493726, "grad_norm": 1.0715285539627075, "learning_rate": 0.00019962250554846796, "loss": 4.7282, "step": 115450 }, { "epoch": 0.21553584216431576, "grad_norm": 1.0772089958190918, "learning_rate": 0.00019962214078156975, "loss": 4.808, "step": 115500 }, { "epoch": 0.21562914772369424, "grad_norm": 0.8694863319396973, "learning_rate": 0.000199621775838856, "loss": 4.6563, "step": 115550 }, { "epoch": 0.21572245328307274, "grad_norm": 0.7618568539619446, "learning_rate": 0.00019962141072032733, "loss": 4.5163, "step": 115600 }, { "epoch": 0.21581575884245124, "grad_norm": 0.7407212257385254, "learning_rate": 0.0001996210454259844, "loss": 4.7637, "step": 115650 }, { "epoch": 0.21590906440182972, "grad_norm": 0.86847984790802, "learning_rate": 0.00019962067995582788, "loss": 4.6896, "step": 115700 }, { "epoch": 0.21600236996120822, "grad_norm": 1.2252628803253174, "learning_rate": 0.00019962031430985838, "loss": 4.5515, "step": 115750 }, { "epoch": 0.2160956755205867, "grad_norm": 0.8714708089828491, "learning_rate": 0.00019961994848807652, "loss": 4.699, "step": 115800 }, { "epoch": 0.2161889810799652, "grad_norm": 0.9541438221931458, "learning_rate": 0.000199619582490483, "loss": 4.7368, "step": 115850 }, { "epoch": 0.21628228663934368, "grad_norm": 1.139961838722229, "learning_rate": 0.00019961921631707845, "loss": 4.7307, "step": 115900 }, { "epoch": 0.21637559219872218, "grad_norm": 1.154788613319397, "learning_rate": 0.00019961884996786352, "loss": 4.7677, "step": 115950 }, { "epoch": 0.21646889775810066, "grad_norm": 0.9261074662208557, "learning_rate": 0.00019961848344283883, "loss": 4.5874, "step": 116000 }, { "epoch": 0.21656220331747916, "grad_norm": 1.3883506059646606, "learning_rate": 0.00019961811674200505, "loss": 4.6135, "step": 116050 }, { "epoch": 0.21665550887685767, "grad_norm": 1.012757658958435, "learning_rate": 0.0001996177498653628, "loss": 4.7373, "step": 116100 }, { "epoch": 0.21674881443623614, "grad_norm": 1.0129855871200562, "learning_rate": 0.0001996173828129128, "loss": 4.7325, "step": 116150 }, { "epoch": 0.21684211999561465, "grad_norm": 1.0152562856674194, "learning_rate": 0.00019961701558465563, "loss": 4.5226, "step": 116200 }, { "epoch": 0.21693542555499312, "grad_norm": 1.094165563583374, "learning_rate": 0.00019961664818059194, "loss": 4.7992, "step": 116250 }, { "epoch": 0.21702873111437163, "grad_norm": 1.0749988555908203, "learning_rate": 0.0001996162806007224, "loss": 4.7296, "step": 116300 }, { "epoch": 0.2171220366737501, "grad_norm": 0.8733508586883545, "learning_rate": 0.00019961591284504767, "loss": 4.8671, "step": 116350 }, { "epoch": 0.2172153422331286, "grad_norm": 0.7462573647499084, "learning_rate": 0.00019961554491356836, "loss": 4.8757, "step": 116400 }, { "epoch": 0.2173086477925071, "grad_norm": 0.8342740535736084, "learning_rate": 0.00019961517680628515, "loss": 4.5841, "step": 116450 }, { "epoch": 0.21740195335188559, "grad_norm": 1.1010075807571411, "learning_rate": 0.00019961480852319867, "loss": 4.5726, "step": 116500 }, { "epoch": 0.2174952589112641, "grad_norm": 0.8100841045379639, "learning_rate": 0.0001996144400643096, "loss": 4.6927, "step": 116550 }, { "epoch": 0.21758856447064256, "grad_norm": 1.0698412656784058, "learning_rate": 0.00019961407142961854, "loss": 4.8108, "step": 116600 }, { "epoch": 0.21768187003002107, "grad_norm": 0.9264890551567078, "learning_rate": 0.00019961370261912623, "loss": 4.555, "step": 116650 }, { "epoch": 0.21777517558939954, "grad_norm": 0.8985602855682373, "learning_rate": 0.00019961333363283318, "loss": 4.7173, "step": 116700 }, { "epoch": 0.21786848114877805, "grad_norm": 0.9962376356124878, "learning_rate": 0.0001996129644707402, "loss": 4.5897, "step": 116750 }, { "epoch": 0.21796178670815655, "grad_norm": 1.108918309211731, "learning_rate": 0.00019961259513284781, "loss": 4.7698, "step": 116800 }, { "epoch": 0.21805509226753503, "grad_norm": 1.3311587572097778, "learning_rate": 0.00019961222561915677, "loss": 4.6247, "step": 116850 }, { "epoch": 0.21814839782691353, "grad_norm": 0.602816104888916, "learning_rate": 0.00019961185592966762, "loss": 4.9767, "step": 116900 }, { "epoch": 0.218241703386292, "grad_norm": 1.00581693649292, "learning_rate": 0.0001996114860643811, "loss": 4.5662, "step": 116950 }, { "epoch": 0.2183350089456705, "grad_norm": 0.9467529654502869, "learning_rate": 0.00019961111602329786, "loss": 4.6823, "step": 117000 }, { "epoch": 0.2183350089456705, "eval_loss": 4.885964393615723, "eval_runtime": 228.2045, "eval_samples_per_second": 11.428, "eval_steps_per_second": 11.428, "eval_tts_loss": 7.423409835071205, "step": 117000 }, { "epoch": 0.218428314505049, "grad_norm": 0.9018577933311462, "learning_rate": 0.00019961074580641846, "loss": 4.7069, "step": 117050 }, { "epoch": 0.2185216200644275, "grad_norm": 1.015038251876831, "learning_rate": 0.0001996103754137437, "loss": 4.772, "step": 117100 }, { "epoch": 0.21861492562380597, "grad_norm": 1.082078456878662, "learning_rate": 0.00019961000484527413, "loss": 4.5817, "step": 117150 }, { "epoch": 0.21870823118318447, "grad_norm": 1.0951112508773804, "learning_rate": 0.0001996096341010104, "loss": 4.76, "step": 117200 }, { "epoch": 0.21880153674256297, "grad_norm": 1.023581862449646, "learning_rate": 0.00019960926318095322, "loss": 4.547, "step": 117250 }, { "epoch": 0.21889484230194145, "grad_norm": 0.8084960579872131, "learning_rate": 0.00019960889208510318, "loss": 4.8218, "step": 117300 }, { "epoch": 0.21898814786131995, "grad_norm": 1.2182049751281738, "learning_rate": 0.000199608520813461, "loss": 4.6656, "step": 117350 }, { "epoch": 0.21908145342069843, "grad_norm": 0.9085681438446045, "learning_rate": 0.00019960814936602732, "loss": 4.8121, "step": 117400 }, { "epoch": 0.21917475898007693, "grad_norm": 0.9307880997657776, "learning_rate": 0.00019960777774280278, "loss": 4.6564, "step": 117450 }, { "epoch": 0.2192680645394554, "grad_norm": 0.8649754524230957, "learning_rate": 0.00019960740594378803, "loss": 4.7026, "step": 117500 }, { "epoch": 0.2193613700988339, "grad_norm": 0.7863887548446655, "learning_rate": 0.00019960703396898371, "loss": 4.5972, "step": 117550 }, { "epoch": 0.21945467565821242, "grad_norm": 0.9783292412757874, "learning_rate": 0.00019960666181839055, "loss": 4.6181, "step": 117600 }, { "epoch": 0.2195479812175909, "grad_norm": 1.023344874382019, "learning_rate": 0.00019960628949200913, "loss": 4.5983, "step": 117650 }, { "epoch": 0.2196412867769694, "grad_norm": 0.9733097553253174, "learning_rate": 0.0001996059169898401, "loss": 4.8129, "step": 117700 }, { "epoch": 0.21973459233634787, "grad_norm": 0.9498526453971863, "learning_rate": 0.00019960554431188422, "loss": 4.5947, "step": 117750 }, { "epoch": 0.21982789789572638, "grad_norm": 1.1250981092453003, "learning_rate": 0.00019960517145814203, "loss": 4.4665, "step": 117800 }, { "epoch": 0.21992120345510485, "grad_norm": 0.7923647165298462, "learning_rate": 0.00019960479842861428, "loss": 4.5954, "step": 117850 }, { "epoch": 0.22001450901448336, "grad_norm": 1.0662282705307007, "learning_rate": 0.00019960442522330154, "loss": 4.755, "step": 117900 }, { "epoch": 0.22010781457386186, "grad_norm": 1.0039044618606567, "learning_rate": 0.00019960405184220456, "loss": 4.7219, "step": 117950 }, { "epoch": 0.22020112013324034, "grad_norm": 1.210990309715271, "learning_rate": 0.00019960367828532392, "loss": 4.6545, "step": 118000 }, { "epoch": 0.22029442569261884, "grad_norm": 0.9507876634597778, "learning_rate": 0.00019960330455266035, "loss": 4.7404, "step": 118050 }, { "epoch": 0.22038773125199732, "grad_norm": 1.1549371480941772, "learning_rate": 0.00019960293064421445, "loss": 4.8518, "step": 118100 }, { "epoch": 0.22048103681137582, "grad_norm": 1.0183074474334717, "learning_rate": 0.0001996025565599869, "loss": 4.7184, "step": 118150 }, { "epoch": 0.2205743423707543, "grad_norm": 1.1634994745254517, "learning_rate": 0.00019960218229997837, "loss": 4.6093, "step": 118200 }, { "epoch": 0.2206676479301328, "grad_norm": 0.969548225402832, "learning_rate": 0.00019960180786418951, "loss": 4.8916, "step": 118250 }, { "epoch": 0.2207609534895113, "grad_norm": 1.2633224725723267, "learning_rate": 0.000199601433252621, "loss": 4.6768, "step": 118300 }, { "epoch": 0.22085425904888978, "grad_norm": 0.9347729682922363, "learning_rate": 0.00019960105846527346, "loss": 4.6532, "step": 118350 }, { "epoch": 0.22094756460826828, "grad_norm": 1.0598022937774658, "learning_rate": 0.0001996006835021476, "loss": 4.6961, "step": 118400 }, { "epoch": 0.22104087016764676, "grad_norm": 1.2343388795852661, "learning_rate": 0.00019960030836324404, "loss": 4.6699, "step": 118450 }, { "epoch": 0.22113417572702526, "grad_norm": 0.871600866317749, "learning_rate": 0.00019959993304856347, "loss": 4.6907, "step": 118500 }, { "epoch": 0.22122748128640374, "grad_norm": 0.9813231825828552, "learning_rate": 0.00019959955755810652, "loss": 4.6616, "step": 118550 }, { "epoch": 0.22132078684578224, "grad_norm": 1.049269199371338, "learning_rate": 0.00019959918189187392, "loss": 4.6005, "step": 118600 }, { "epoch": 0.22141409240516072, "grad_norm": 0.7890821695327759, "learning_rate": 0.00019959880604986628, "loss": 4.685, "step": 118650 }, { "epoch": 0.22150739796453922, "grad_norm": 0.8772531151771545, "learning_rate": 0.00019959843003208424, "loss": 4.7188, "step": 118700 }, { "epoch": 0.22160070352391772, "grad_norm": 1.0186238288879395, "learning_rate": 0.00019959805383852852, "loss": 4.8515, "step": 118750 }, { "epoch": 0.2216940090832962, "grad_norm": 1.081235647201538, "learning_rate": 0.00019959767746919975, "loss": 4.7645, "step": 118800 }, { "epoch": 0.2217873146426747, "grad_norm": 0.9641054272651672, "learning_rate": 0.0001995973009240986, "loss": 4.6056, "step": 118850 }, { "epoch": 0.22188062020205318, "grad_norm": 0.8466959595680237, "learning_rate": 0.00019959692420322577, "loss": 4.7974, "step": 118900 }, { "epoch": 0.22197392576143168, "grad_norm": 1.1167409420013428, "learning_rate": 0.00019959654730658185, "loss": 4.8381, "step": 118950 }, { "epoch": 0.22206723132081016, "grad_norm": 0.8874496221542358, "learning_rate": 0.0001995961702341676, "loss": 4.5208, "step": 119000 }, { "epoch": 0.22216053688018866, "grad_norm": 1.0184682607650757, "learning_rate": 0.0001995957929859836, "loss": 4.766, "step": 119050 }, { "epoch": 0.22225384243956717, "grad_norm": 0.8084706664085388, "learning_rate": 0.00019959541556203054, "loss": 4.6748, "step": 119100 }, { "epoch": 0.22234714799894564, "grad_norm": 0.9491915106773376, "learning_rate": 0.0001995950379623091, "loss": 4.7225, "step": 119150 }, { "epoch": 0.22244045355832415, "grad_norm": 0.9508479833602905, "learning_rate": 0.00019959466018682, "loss": 4.785, "step": 119200 }, { "epoch": 0.22253375911770262, "grad_norm": 0.8194633722305298, "learning_rate": 0.0001995942822355638, "loss": 4.7103, "step": 119250 }, { "epoch": 0.22262706467708113, "grad_norm": 0.9935794472694397, "learning_rate": 0.0001995939041085412, "loss": 4.7287, "step": 119300 }, { "epoch": 0.2227203702364596, "grad_norm": 0.7997922897338867, "learning_rate": 0.0001995935258057529, "loss": 4.621, "step": 119350 }, { "epoch": 0.2228136757958381, "grad_norm": 0.9721159338951111, "learning_rate": 0.00019959314732719956, "loss": 4.7279, "step": 119400 }, { "epoch": 0.2229069813552166, "grad_norm": 1.006213665008545, "learning_rate": 0.0001995927686728818, "loss": 4.8168, "step": 119450 }, { "epoch": 0.22300028691459509, "grad_norm": 0.976470947265625, "learning_rate": 0.00019959238984280038, "loss": 4.6645, "step": 119500 }, { "epoch": 0.2230935924739736, "grad_norm": 0.9182271957397461, "learning_rate": 0.00019959201083695588, "loss": 4.6798, "step": 119550 }, { "epoch": 0.22318689803335207, "grad_norm": 0.873684287071228, "learning_rate": 0.00019959163165534903, "loss": 4.4998, "step": 119600 }, { "epoch": 0.22328020359273057, "grad_norm": 1.1776008605957031, "learning_rate": 0.00019959125229798046, "loss": 4.7635, "step": 119650 }, { "epoch": 0.22337350915210905, "grad_norm": 1.159077763557434, "learning_rate": 0.00019959087276485087, "loss": 4.6824, "step": 119700 }, { "epoch": 0.22346681471148755, "grad_norm": 1.2135952711105347, "learning_rate": 0.0001995904930559609, "loss": 4.7218, "step": 119750 }, { "epoch": 0.22356012027086603, "grad_norm": 0.8527218699455261, "learning_rate": 0.0001995901131713112, "loss": 4.7581, "step": 119800 }, { "epoch": 0.22365342583024453, "grad_norm": 1.1286325454711914, "learning_rate": 0.0001995897331109025, "loss": 4.6019, "step": 119850 }, { "epoch": 0.22374673138962303, "grad_norm": 0.7209294438362122, "learning_rate": 0.00019958935287473547, "loss": 4.4614, "step": 119900 }, { "epoch": 0.2238400369490015, "grad_norm": 1.012035608291626, "learning_rate": 0.00019958897246281074, "loss": 4.743, "step": 119950 }, { "epoch": 0.22393334250838, "grad_norm": 0.9427865147590637, "learning_rate": 0.000199588591875129, "loss": 4.5508, "step": 120000 }, { "epoch": 0.22393334250838, "eval_loss": 4.883612632751465, "eval_runtime": 229.5565, "eval_samples_per_second": 11.361, "eval_steps_per_second": 11.361, "eval_tts_loss": 7.450549049201519, "step": 120000 }, { "epoch": 0.2240266480677585, "grad_norm": 1.0907127857208252, "learning_rate": 0.0001995882111116909, "loss": 4.6219, "step": 120050 }, { "epoch": 0.224119953627137, "grad_norm": 1.2542510032653809, "learning_rate": 0.00019958783017249714, "loss": 4.5626, "step": 120100 }, { "epoch": 0.22421325918651547, "grad_norm": 0.9580556750297546, "learning_rate": 0.00019958744905754837, "loss": 4.7883, "step": 120150 }, { "epoch": 0.22430656474589397, "grad_norm": 1.1819099187850952, "learning_rate": 0.00019958706776684532, "loss": 4.7903, "step": 120200 }, { "epoch": 0.22439987030527248, "grad_norm": 0.9004278182983398, "learning_rate": 0.00019958668630038856, "loss": 4.6342, "step": 120250 }, { "epoch": 0.22449317586465095, "grad_norm": 0.898798942565918, "learning_rate": 0.00019958630465817886, "loss": 4.9032, "step": 120300 }, { "epoch": 0.22458648142402945, "grad_norm": 1.324995994567871, "learning_rate": 0.00019958592284021685, "loss": 4.7004, "step": 120350 }, { "epoch": 0.22467978698340793, "grad_norm": 0.9683493971824646, "learning_rate": 0.00019958554084650317, "loss": 4.7933, "step": 120400 }, { "epoch": 0.22477309254278643, "grad_norm": 0.9748978614807129, "learning_rate": 0.00019958515867703858, "loss": 4.8761, "step": 120450 }, { "epoch": 0.2248663981021649, "grad_norm": 1.192099690437317, "learning_rate": 0.00019958477633182372, "loss": 4.6939, "step": 120500 }, { "epoch": 0.22495970366154341, "grad_norm": 1.2291393280029297, "learning_rate": 0.0001995843938108592, "loss": 4.743, "step": 120550 }, { "epoch": 0.22505300922092192, "grad_norm": 1.2293897867202759, "learning_rate": 0.00019958401111414578, "loss": 4.8312, "step": 120600 }, { "epoch": 0.2251463147803004, "grad_norm": 1.0013765096664429, "learning_rate": 0.0001995836282416841, "loss": 4.71, "step": 120650 }, { "epoch": 0.2252396203396789, "grad_norm": 1.1996088027954102, "learning_rate": 0.00019958324519347485, "loss": 4.8127, "step": 120700 }, { "epoch": 0.22533292589905737, "grad_norm": 0.9375223517417908, "learning_rate": 0.00019958286196951867, "loss": 4.6523, "step": 120750 }, { "epoch": 0.22542623145843588, "grad_norm": 0.7946412563323975, "learning_rate": 0.00019958247856981628, "loss": 4.5345, "step": 120800 }, { "epoch": 0.22551953701781435, "grad_norm": 0.9169266223907471, "learning_rate": 0.00019958209499436832, "loss": 4.5954, "step": 120850 }, { "epoch": 0.22561284257719286, "grad_norm": 1.0124138593673706, "learning_rate": 0.0001995817112431755, "loss": 4.7285, "step": 120900 }, { "epoch": 0.22570614813657136, "grad_norm": 0.8469855189323425, "learning_rate": 0.00019958132731623843, "loss": 4.5862, "step": 120950 }, { "epoch": 0.22579945369594984, "grad_norm": 0.9922230839729309, "learning_rate": 0.00019958094321355791, "loss": 4.7852, "step": 121000 }, { "epoch": 0.22589275925532834, "grad_norm": 1.1002434492111206, "learning_rate": 0.00019958055893513451, "loss": 4.7454, "step": 121050 }, { "epoch": 0.22598606481470682, "grad_norm": 1.2257192134857178, "learning_rate": 0.00019958017448096896, "loss": 4.7457, "step": 121100 }, { "epoch": 0.22607937037408532, "grad_norm": 1.063035249710083, "learning_rate": 0.0001995797898510619, "loss": 4.5511, "step": 121150 }, { "epoch": 0.2261726759334638, "grad_norm": 0.769403874874115, "learning_rate": 0.00019957940504541406, "loss": 4.6497, "step": 121200 }, { "epoch": 0.2262659814928423, "grad_norm": 0.9175217151641846, "learning_rate": 0.0001995790200640261, "loss": 4.6908, "step": 121250 }, { "epoch": 0.22635928705222078, "grad_norm": 1.0887384414672852, "learning_rate": 0.00019957863490689865, "loss": 4.9673, "step": 121300 }, { "epoch": 0.22645259261159928, "grad_norm": 1.0084506273269653, "learning_rate": 0.00019957824957403245, "loss": 4.6811, "step": 121350 }, { "epoch": 0.22654589817097778, "grad_norm": 1.035234808921814, "learning_rate": 0.00019957786406542817, "loss": 4.5367, "step": 121400 }, { "epoch": 0.22663920373035626, "grad_norm": 1.082025408744812, "learning_rate": 0.0001995774783810865, "loss": 4.8276, "step": 121450 }, { "epoch": 0.22673250928973476, "grad_norm": 0.8627831339836121, "learning_rate": 0.00019957709252100807, "loss": 4.5622, "step": 121500 }, { "epoch": 0.22682581484911324, "grad_norm": 0.6274713277816772, "learning_rate": 0.0001995767064851936, "loss": 4.6064, "step": 121550 }, { "epoch": 0.22691912040849174, "grad_norm": 0.6322067975997925, "learning_rate": 0.00019957632027364376, "loss": 4.6496, "step": 121600 }, { "epoch": 0.22701242596787022, "grad_norm": 0.7151324152946472, "learning_rate": 0.00019957593388635923, "loss": 4.4901, "step": 121650 }, { "epoch": 0.22710573152724872, "grad_norm": 0.8204741477966309, "learning_rate": 0.00019957554732334071, "loss": 4.8227, "step": 121700 }, { "epoch": 0.22719903708662723, "grad_norm": 1.0234726667404175, "learning_rate": 0.0001995751605845889, "loss": 4.6125, "step": 121750 }, { "epoch": 0.2272923426460057, "grad_norm": 1.1819202899932861, "learning_rate": 0.0001995747736701044, "loss": 4.7185, "step": 121800 }, { "epoch": 0.2273856482053842, "grad_norm": 0.9932510256767273, "learning_rate": 0.00019957438657988797, "loss": 4.9552, "step": 121850 }, { "epoch": 0.22747895376476268, "grad_norm": 1.0149807929992676, "learning_rate": 0.00019957399931394028, "loss": 4.7491, "step": 121900 }, { "epoch": 0.22757225932414119, "grad_norm": 0.8650890588760376, "learning_rate": 0.00019957361187226199, "loss": 4.7261, "step": 121950 }, { "epoch": 0.22766556488351966, "grad_norm": 0.8572404980659485, "learning_rate": 0.00019957322425485376, "loss": 4.6074, "step": 122000 }, { "epoch": 0.22775887044289816, "grad_norm": 0.8673201203346252, "learning_rate": 0.00019957283646171634, "loss": 4.5425, "step": 122050 }, { "epoch": 0.22785217600227667, "grad_norm": 1.1532055139541626, "learning_rate": 0.0001995724484928504, "loss": 4.6619, "step": 122100 }, { "epoch": 0.22794548156165514, "grad_norm": 1.0270742177963257, "learning_rate": 0.00019957206034825658, "loss": 4.7036, "step": 122150 }, { "epoch": 0.22803878712103365, "grad_norm": 1.0229623317718506, "learning_rate": 0.0001995716720279356, "loss": 4.5638, "step": 122200 }, { "epoch": 0.22813209268041212, "grad_norm": 1.0269454717636108, "learning_rate": 0.00019957128353188816, "loss": 4.4661, "step": 122250 }, { "epoch": 0.22822539823979063, "grad_norm": 1.051167368888855, "learning_rate": 0.0001995708948601149, "loss": 4.5614, "step": 122300 }, { "epoch": 0.2283187037991691, "grad_norm": 0.9413858652114868, "learning_rate": 0.00019957050601261656, "loss": 4.5145, "step": 122350 }, { "epoch": 0.2284120093585476, "grad_norm": 0.7322555184364319, "learning_rate": 0.00019957011698939377, "loss": 4.7114, "step": 122400 }, { "epoch": 0.2285053149179261, "grad_norm": 0.9706824421882629, "learning_rate": 0.00019956972779044727, "loss": 4.5615, "step": 122450 }, { "epoch": 0.2285986204773046, "grad_norm": 1.0096715688705444, "learning_rate": 0.0001995693384157777, "loss": 4.7756, "step": 122500 }, { "epoch": 0.2286919260366831, "grad_norm": 0.901025652885437, "learning_rate": 0.00019956894886538574, "loss": 4.7176, "step": 122550 }, { "epoch": 0.22878523159606157, "grad_norm": 0.9558797478675842, "learning_rate": 0.00019956855913927214, "loss": 4.8155, "step": 122600 }, { "epoch": 0.22887853715544007, "grad_norm": 0.6272359490394592, "learning_rate": 0.00019956816923743755, "loss": 4.7247, "step": 122650 }, { "epoch": 0.22897184271481855, "grad_norm": 1.197523832321167, "learning_rate": 0.00019956777915988269, "loss": 4.6914, "step": 122700 }, { "epoch": 0.22906514827419705, "grad_norm": 0.9652473330497742, "learning_rate": 0.00019956738890660817, "loss": 4.6038, "step": 122750 }, { "epoch": 0.22915845383357553, "grad_norm": 1.0007468461990356, "learning_rate": 0.00019956699847761472, "loss": 4.6983, "step": 122800 }, { "epoch": 0.22925175939295403, "grad_norm": 0.9335245490074158, "learning_rate": 0.0001995666078729031, "loss": 4.631, "step": 122850 }, { "epoch": 0.22934506495233253, "grad_norm": 0.9081475734710693, "learning_rate": 0.00019956621709247386, "loss": 4.8258, "step": 122900 }, { "epoch": 0.229438370511711, "grad_norm": 0.7586884498596191, "learning_rate": 0.00019956582613632785, "loss": 4.5945, "step": 122950 }, { "epoch": 0.2295316760710895, "grad_norm": 0.909521222114563, "learning_rate": 0.0001995654350044656, "loss": 4.6523, "step": 123000 }, { "epoch": 0.2295316760710895, "eval_loss": 4.8781867027282715, "eval_runtime": 230.72, "eval_samples_per_second": 11.304, "eval_steps_per_second": 11.304, "eval_tts_loss": 7.44082480463045, "step": 123000 }, { "epoch": 0.229624981630468, "grad_norm": 1.208822250366211, "learning_rate": 0.00019956504369688792, "loss": 4.9531, "step": 123050 }, { "epoch": 0.2297182871898465, "grad_norm": 1.0203379392623901, "learning_rate": 0.00019956465221359544, "loss": 4.6531, "step": 123100 }, { "epoch": 0.22981159274922497, "grad_norm": 0.9764498472213745, "learning_rate": 0.00019956426055458887, "loss": 4.5436, "step": 123150 }, { "epoch": 0.22990489830860347, "grad_norm": 1.1145259141921997, "learning_rate": 0.00019956386871986892, "loss": 4.5113, "step": 123200 }, { "epoch": 0.22999820386798198, "grad_norm": 0.9552769064903259, "learning_rate": 0.00019956347670943624, "loss": 4.7847, "step": 123250 }, { "epoch": 0.23009150942736045, "grad_norm": 1.0565723180770874, "learning_rate": 0.00019956308452329154, "loss": 4.8575, "step": 123300 }, { "epoch": 0.23018481498673896, "grad_norm": 1.123518705368042, "learning_rate": 0.0001995626921614355, "loss": 4.6688, "step": 123350 }, { "epoch": 0.23027812054611743, "grad_norm": 0.911044716835022, "learning_rate": 0.00019956229962386885, "loss": 4.7472, "step": 123400 }, { "epoch": 0.23037142610549594, "grad_norm": 1.0663899183273315, "learning_rate": 0.00019956190691059223, "loss": 4.7779, "step": 123450 }, { "epoch": 0.2304647316648744, "grad_norm": 0.9321340918540955, "learning_rate": 0.0001995615140216064, "loss": 4.6964, "step": 123500 }, { "epoch": 0.23055803722425292, "grad_norm": 0.8135243058204651, "learning_rate": 0.00019956112095691203, "loss": 4.6287, "step": 123550 }, { "epoch": 0.23065134278363142, "grad_norm": 1.0988374948501587, "learning_rate": 0.00019956072771650977, "loss": 4.8658, "step": 123600 }, { "epoch": 0.2307446483430099, "grad_norm": 0.8417083024978638, "learning_rate": 0.00019956033430040034, "loss": 4.7001, "step": 123650 }, { "epoch": 0.2308379539023884, "grad_norm": 1.0932660102844238, "learning_rate": 0.00019955994070858447, "loss": 4.5109, "step": 123700 }, { "epoch": 0.23093125946176687, "grad_norm": 1.016459584236145, "learning_rate": 0.00019955954694106278, "loss": 4.7203, "step": 123750 }, { "epoch": 0.23102456502114538, "grad_norm": 1.1212247610092163, "learning_rate": 0.000199559152997836, "loss": 4.5223, "step": 123800 }, { "epoch": 0.23111787058052385, "grad_norm": 0.849577784538269, "learning_rate": 0.00019955875887890488, "loss": 4.6992, "step": 123850 }, { "epoch": 0.23121117613990236, "grad_norm": 0.7211328744888306, "learning_rate": 0.00019955836458427007, "loss": 4.554, "step": 123900 }, { "epoch": 0.23130448169928083, "grad_norm": 0.7936710715293884, "learning_rate": 0.00019955797011393222, "loss": 4.7927, "step": 123950 }, { "epoch": 0.23139778725865934, "grad_norm": 0.9986937642097473, "learning_rate": 0.0001995575754678921, "loss": 4.4808, "step": 124000 }, { "epoch": 0.23149109281803784, "grad_norm": 0.8931700587272644, "learning_rate": 0.00019955718064615036, "loss": 4.7492, "step": 124050 }, { "epoch": 0.23158439837741632, "grad_norm": 1.1811097860336304, "learning_rate": 0.00019955678564870772, "loss": 4.5177, "step": 124100 }, { "epoch": 0.23167770393679482, "grad_norm": 0.7191671133041382, "learning_rate": 0.0001995563904755649, "loss": 4.8179, "step": 124150 }, { "epoch": 0.2317710094961733, "grad_norm": 1.1839138269424438, "learning_rate": 0.00019955599512672253, "loss": 4.6699, "step": 124200 }, { "epoch": 0.2318643150555518, "grad_norm": 0.9712040424346924, "learning_rate": 0.00019955559960218135, "loss": 4.6827, "step": 124250 }, { "epoch": 0.23195762061493028, "grad_norm": 1.1487324237823486, "learning_rate": 0.00019955520390194207, "loss": 4.5614, "step": 124300 }, { "epoch": 0.23205092617430878, "grad_norm": 1.104325532913208, "learning_rate": 0.00019955480802600538, "loss": 4.6895, "step": 124350 }, { "epoch": 0.23214423173368728, "grad_norm": 1.0275545120239258, "learning_rate": 0.00019955441197437191, "loss": 4.6102, "step": 124400 }, { "epoch": 0.23223753729306576, "grad_norm": 1.1445039510726929, "learning_rate": 0.00019955401574704248, "loss": 4.6854, "step": 124450 }, { "epoch": 0.23233084285244426, "grad_norm": 0.9036419987678528, "learning_rate": 0.00019955361934401774, "loss": 4.6985, "step": 124500 }, { "epoch": 0.23242414841182274, "grad_norm": 0.825995683670044, "learning_rate": 0.00019955322276529836, "loss": 4.8375, "step": 124550 }, { "epoch": 0.23251745397120124, "grad_norm": 1.3338888883590698, "learning_rate": 0.00019955282601088502, "loss": 4.851, "step": 124600 }, { "epoch": 0.23261075953057972, "grad_norm": 1.4017682075500488, "learning_rate": 0.0001995524290807785, "loss": 4.7548, "step": 124650 }, { "epoch": 0.23270406508995822, "grad_norm": 0.6777539849281311, "learning_rate": 0.00019955203197497947, "loss": 4.6612, "step": 124700 }, { "epoch": 0.23279737064933673, "grad_norm": 0.94838947057724, "learning_rate": 0.0001995516346934886, "loss": 4.8645, "step": 124750 }, { "epoch": 0.2328906762087152, "grad_norm": 1.0013059377670288, "learning_rate": 0.0001995512372363066, "loss": 4.6178, "step": 124800 }, { "epoch": 0.2329839817680937, "grad_norm": 0.8147088885307312, "learning_rate": 0.0001995508396034342, "loss": 4.4805, "step": 124850 }, { "epoch": 0.23307728732747218, "grad_norm": 0.6364634037017822, "learning_rate": 0.0001995504417948721, "loss": 4.614, "step": 124900 }, { "epoch": 0.23317059288685069, "grad_norm": 1.0096015930175781, "learning_rate": 0.00019955004381062096, "loss": 4.8126, "step": 124950 }, { "epoch": 0.23326389844622916, "grad_norm": 1.0549802780151367, "learning_rate": 0.00019954964565068147, "loss": 4.5155, "step": 125000 }, { "epoch": 0.23335720400560767, "grad_norm": 1.0715384483337402, "learning_rate": 0.00019954924731505443, "loss": 4.8361, "step": 125050 }, { "epoch": 0.23345050956498617, "grad_norm": 1.089919090270996, "learning_rate": 0.00019954884880374048, "loss": 4.4423, "step": 125100 }, { "epoch": 0.23354381512436465, "grad_norm": 1.1438995599746704, "learning_rate": 0.00019954845011674032, "loss": 4.5797, "step": 125150 }, { "epoch": 0.23363712068374315, "grad_norm": 0.8875765800476074, "learning_rate": 0.00019954805125405468, "loss": 4.7633, "step": 125200 }, { "epoch": 0.23373042624312163, "grad_norm": 1.0435429811477661, "learning_rate": 0.00019954765221568424, "loss": 4.5587, "step": 125250 }, { "epoch": 0.23382373180250013, "grad_norm": 0.8705243468284607, "learning_rate": 0.00019954725300162966, "loss": 4.7671, "step": 125300 }, { "epoch": 0.2339170373618786, "grad_norm": 1.100628137588501, "learning_rate": 0.00019954685361189177, "loss": 4.5598, "step": 125350 }, { "epoch": 0.2340103429212571, "grad_norm": 0.7905378341674805, "learning_rate": 0.00019954645404647115, "loss": 4.6633, "step": 125400 }, { "epoch": 0.23410364848063558, "grad_norm": 1.0382624864578247, "learning_rate": 0.00019954605430536857, "loss": 4.8594, "step": 125450 }, { "epoch": 0.2341969540400141, "grad_norm": 1.0554344654083252, "learning_rate": 0.0001995456543885847, "loss": 4.8501, "step": 125500 }, { "epoch": 0.2342902595993926, "grad_norm": 0.6847901344299316, "learning_rate": 0.0001995452542961203, "loss": 4.686, "step": 125550 }, { "epoch": 0.23438356515877107, "grad_norm": 0.9976589679718018, "learning_rate": 0.000199544854027976, "loss": 4.7711, "step": 125600 }, { "epoch": 0.23447687071814957, "grad_norm": 1.0457299947738647, "learning_rate": 0.0001995444535841526, "loss": 4.5277, "step": 125650 }, { "epoch": 0.23457017627752805, "grad_norm": 0.7479730844497681, "learning_rate": 0.00019954405296465072, "loss": 4.7013, "step": 125700 }, { "epoch": 0.23466348183690655, "grad_norm": 1.141799807548523, "learning_rate": 0.0001995436521694711, "loss": 4.8921, "step": 125750 }, { "epoch": 0.23475678739628503, "grad_norm": 0.865399181842804, "learning_rate": 0.00019954325119861442, "loss": 4.6746, "step": 125800 }, { "epoch": 0.23485009295566353, "grad_norm": 1.0194836854934692, "learning_rate": 0.00019954285005208147, "loss": 4.8391, "step": 125850 }, { "epoch": 0.23494339851504203, "grad_norm": 0.7085402011871338, "learning_rate": 0.00019954244872987287, "loss": 4.7155, "step": 125900 }, { "epoch": 0.2350367040744205, "grad_norm": 0.9328494071960449, "learning_rate": 0.0001995420472319894, "loss": 4.7267, "step": 125950 }, { "epoch": 0.23513000963379901, "grad_norm": 0.9679514169692993, "learning_rate": 0.0001995416455584317, "loss": 4.6913, "step": 126000 }, { "epoch": 0.23513000963379901, "eval_loss": 4.869121551513672, "eval_runtime": 230.4744, "eval_samples_per_second": 11.316, "eval_steps_per_second": 11.316, "eval_tts_loss": 7.37138597899048, "step": 126000 }, { "epoch": 0.2352233151931775, "grad_norm": 1.0236223936080933, "learning_rate": 0.00019954124370920054, "loss": 4.6602, "step": 126050 }, { "epoch": 0.235316620752556, "grad_norm": 0.9021424651145935, "learning_rate": 0.0001995408416842966, "loss": 4.3545, "step": 126100 }, { "epoch": 0.23540992631193447, "grad_norm": 1.059244155883789, "learning_rate": 0.00019954043948372055, "loss": 4.7123, "step": 126150 }, { "epoch": 0.23550323187131297, "grad_norm": 0.8630461096763611, "learning_rate": 0.00019954003710747316, "loss": 4.7759, "step": 126200 }, { "epoch": 0.23559653743069148, "grad_norm": 1.192002534866333, "learning_rate": 0.00019953963455555514, "loss": 4.7526, "step": 126250 }, { "epoch": 0.23568984299006995, "grad_norm": 1.1485189199447632, "learning_rate": 0.00019953923182796713, "loss": 4.7111, "step": 126300 }, { "epoch": 0.23578314854944846, "grad_norm": 0.9921421408653259, "learning_rate": 0.00019953882892470993, "loss": 4.6672, "step": 126350 }, { "epoch": 0.23587645410882693, "grad_norm": 1.0147913694381714, "learning_rate": 0.0001995384258457842, "loss": 4.8141, "step": 126400 }, { "epoch": 0.23596975966820544, "grad_norm": 1.1433029174804688, "learning_rate": 0.00019953802259119066, "loss": 4.6498, "step": 126450 }, { "epoch": 0.2360630652275839, "grad_norm": 0.8153554201126099, "learning_rate": 0.00019953761916093003, "loss": 4.501, "step": 126500 }, { "epoch": 0.23615637078696242, "grad_norm": 0.8511623740196228, "learning_rate": 0.00019953721555500303, "loss": 4.6241, "step": 126550 }, { "epoch": 0.2362496763463409, "grad_norm": 0.9093624353408813, "learning_rate": 0.00019953681177341034, "loss": 4.8207, "step": 126600 }, { "epoch": 0.2363429819057194, "grad_norm": 1.1618552207946777, "learning_rate": 0.0001995364078161527, "loss": 4.7207, "step": 126650 }, { "epoch": 0.2364362874650979, "grad_norm": 1.0167028903961182, "learning_rate": 0.0001995360036832308, "loss": 4.7409, "step": 126700 }, { "epoch": 0.23652959302447638, "grad_norm": 1.2850828170776367, "learning_rate": 0.00019953559937464535, "loss": 4.8472, "step": 126750 }, { "epoch": 0.23662289858385488, "grad_norm": 0.9665905237197876, "learning_rate": 0.00019953519489039715, "loss": 4.658, "step": 126800 }, { "epoch": 0.23671620414323336, "grad_norm": 0.9830486178398132, "learning_rate": 0.0001995347902304868, "loss": 4.6645, "step": 126850 }, { "epoch": 0.23680950970261186, "grad_norm": 1.1448428630828857, "learning_rate": 0.00019953438539491505, "loss": 4.7673, "step": 126900 }, { "epoch": 0.23690281526199033, "grad_norm": 1.1206557750701904, "learning_rate": 0.00019953398038368266, "loss": 4.8454, "step": 126950 }, { "epoch": 0.23699612082136884, "grad_norm": 0.9602530002593994, "learning_rate": 0.00019953357519679026, "loss": 4.588, "step": 127000 }, { "epoch": 0.23708942638074734, "grad_norm": 0.816632091999054, "learning_rate": 0.00019953316983423863, "loss": 4.6631, "step": 127050 }, { "epoch": 0.23718273194012582, "grad_norm": 0.996062159538269, "learning_rate": 0.00019953276429602847, "loss": 4.6397, "step": 127100 }, { "epoch": 0.23727603749950432, "grad_norm": 1.116538405418396, "learning_rate": 0.0001995323585821605, "loss": 4.7699, "step": 127150 }, { "epoch": 0.2373693430588828, "grad_norm": 1.0206782817840576, "learning_rate": 0.00019953195269263544, "loss": 4.568, "step": 127200 }, { "epoch": 0.2374626486182613, "grad_norm": 0.8355110287666321, "learning_rate": 0.00019953154662745397, "loss": 4.5723, "step": 127250 }, { "epoch": 0.23755595417763978, "grad_norm": 0.7962259650230408, "learning_rate": 0.00019953114038661685, "loss": 4.6924, "step": 127300 }, { "epoch": 0.23764925973701828, "grad_norm": 1.0387051105499268, "learning_rate": 0.0001995307339701248, "loss": 4.6205, "step": 127350 }, { "epoch": 0.23774256529639679, "grad_norm": 0.9013943672180176, "learning_rate": 0.0001995303273779785, "loss": 4.705, "step": 127400 }, { "epoch": 0.23783587085577526, "grad_norm": 0.9052708148956299, "learning_rate": 0.00019952992061017866, "loss": 4.6817, "step": 127450 }, { "epoch": 0.23792917641515376, "grad_norm": 1.073463797569275, "learning_rate": 0.00019952951366672606, "loss": 4.8822, "step": 127500 }, { "epoch": 0.23802248197453224, "grad_norm": 0.8236823081970215, "learning_rate": 0.00019952910654762136, "loss": 4.7687, "step": 127550 }, { "epoch": 0.23811578753391074, "grad_norm": 1.148167371749878, "learning_rate": 0.0001995286992528653, "loss": 4.6313, "step": 127600 }, { "epoch": 0.23820909309328922, "grad_norm": 0.5891329050064087, "learning_rate": 0.00019952829178245863, "loss": 4.6894, "step": 127650 }, { "epoch": 0.23830239865266772, "grad_norm": 0.9606289267539978, "learning_rate": 0.00019952788413640196, "loss": 4.7802, "step": 127700 }, { "epoch": 0.23839570421204623, "grad_norm": 0.8213685750961304, "learning_rate": 0.00019952747631469617, "loss": 4.6354, "step": 127750 }, { "epoch": 0.2384890097714247, "grad_norm": 0.9802691340446472, "learning_rate": 0.00019952706831734187, "loss": 4.5251, "step": 127800 }, { "epoch": 0.2385823153308032, "grad_norm": 0.8711889386177063, "learning_rate": 0.00019952666014433977, "loss": 4.6781, "step": 127850 }, { "epoch": 0.23867562089018168, "grad_norm": 1.1614457368850708, "learning_rate": 0.0001995262517956907, "loss": 4.5533, "step": 127900 }, { "epoch": 0.2387689264495602, "grad_norm": 0.9102240204811096, "learning_rate": 0.00019952584327139524, "loss": 4.6598, "step": 127950 }, { "epoch": 0.23886223200893866, "grad_norm": 0.8396139144897461, "learning_rate": 0.00019952543457145422, "loss": 4.4948, "step": 128000 }, { "epoch": 0.23895553756831717, "grad_norm": 0.927024781703949, "learning_rate": 0.0001995250256958683, "loss": 4.5239, "step": 128050 }, { "epoch": 0.23904884312769564, "grad_norm": 1.0152112245559692, "learning_rate": 0.00019952461664463823, "loss": 4.6425, "step": 128100 }, { "epoch": 0.23914214868707415, "grad_norm": 1.4888454675674438, "learning_rate": 0.0001995242074177647, "loss": 4.7832, "step": 128150 }, { "epoch": 0.23923545424645265, "grad_norm": 1.1071109771728516, "learning_rate": 0.0001995237980152485, "loss": 4.6113, "step": 128200 }, { "epoch": 0.23932875980583113, "grad_norm": 0.9352720379829407, "learning_rate": 0.00019952338843709028, "loss": 4.6275, "step": 128250 }, { "epoch": 0.23942206536520963, "grad_norm": 0.9158457517623901, "learning_rate": 0.00019952297868329079, "loss": 4.6196, "step": 128300 }, { "epoch": 0.2395153709245881, "grad_norm": 0.46742984652519226, "learning_rate": 0.00019952256875385076, "loss": 4.6612, "step": 128350 }, { "epoch": 0.2396086764839666, "grad_norm": 1.0727510452270508, "learning_rate": 0.0001995221586487709, "loss": 4.7063, "step": 128400 }, { "epoch": 0.23970198204334509, "grad_norm": 0.9847069382667542, "learning_rate": 0.00019952174836805197, "loss": 4.9514, "step": 128450 }, { "epoch": 0.2397952876027236, "grad_norm": 1.0134080648422241, "learning_rate": 0.00019952133791169464, "loss": 4.4549, "step": 128500 }, { "epoch": 0.2398885931621021, "grad_norm": 0.7838501334190369, "learning_rate": 0.00019952092727969965, "loss": 4.348, "step": 128550 }, { "epoch": 0.23998189872148057, "grad_norm": 0.9687092900276184, "learning_rate": 0.00019952051647206776, "loss": 4.5199, "step": 128600 }, { "epoch": 0.24007520428085907, "grad_norm": 0.833861231803894, "learning_rate": 0.00019952010548879968, "loss": 4.812, "step": 128650 }, { "epoch": 0.24016850984023755, "grad_norm": 1.0054713487625122, "learning_rate": 0.00019951969432989607, "loss": 4.6566, "step": 128700 }, { "epoch": 0.24026181539961605, "grad_norm": 0.8803058862686157, "learning_rate": 0.00019951928299535775, "loss": 4.4073, "step": 128750 }, { "epoch": 0.24035512095899453, "grad_norm": 1.0025216341018677, "learning_rate": 0.0001995188714851854, "loss": 4.6065, "step": 128800 }, { "epoch": 0.24044842651837303, "grad_norm": 0.9606624245643616, "learning_rate": 0.0001995184597993798, "loss": 4.6934, "step": 128850 }, { "epoch": 0.24054173207775154, "grad_norm": 0.8765124678611755, "learning_rate": 0.00019951804793794155, "loss": 4.6806, "step": 128900 }, { "epoch": 0.24063503763713, "grad_norm": 0.9700292944908142, "learning_rate": 0.0001995176359008715, "loss": 4.7322, "step": 128950 }, { "epoch": 0.24072834319650852, "grad_norm": 0.7968749403953552, "learning_rate": 0.0001995172236881703, "loss": 4.648, "step": 129000 }, { "epoch": 0.24072834319650852, "eval_loss": 4.864802837371826, "eval_runtime": 229.1704, "eval_samples_per_second": 11.38, "eval_steps_per_second": 11.38, "eval_tts_loss": 7.457818870576944, "step": 129000 }, { "epoch": 0.240821648755887, "grad_norm": 0.8393458724021912, "learning_rate": 0.00019951681129983876, "loss": 4.6002, "step": 129050 }, { "epoch": 0.2409149543152655, "grad_norm": 1.1570461988449097, "learning_rate": 0.00019951639873587752, "loss": 4.5618, "step": 129100 }, { "epoch": 0.24100825987464397, "grad_norm": 1.0129081010818481, "learning_rate": 0.00019951598599628735, "loss": 4.4886, "step": 129150 }, { "epoch": 0.24110156543402247, "grad_norm": 0.7932485342025757, "learning_rate": 0.00019951557308106897, "loss": 4.6398, "step": 129200 }, { "epoch": 0.24119487099340095, "grad_norm": 0.9829528331756592, "learning_rate": 0.00019951515999022315, "loss": 4.6797, "step": 129250 }, { "epoch": 0.24128817655277945, "grad_norm": 0.748111367225647, "learning_rate": 0.00019951474672375054, "loss": 4.5282, "step": 129300 }, { "epoch": 0.24138148211215796, "grad_norm": 0.9890366792678833, "learning_rate": 0.00019951433328165195, "loss": 4.7415, "step": 129350 }, { "epoch": 0.24147478767153643, "grad_norm": 0.8397785425186157, "learning_rate": 0.00019951391966392806, "loss": 4.704, "step": 129400 }, { "epoch": 0.24156809323091494, "grad_norm": 0.9623748064041138, "learning_rate": 0.0001995135058705796, "loss": 4.6239, "step": 129450 }, { "epoch": 0.2416613987902934, "grad_norm": 0.6656234860420227, "learning_rate": 0.00019951309190160732, "loss": 4.5185, "step": 129500 }, { "epoch": 0.24175470434967192, "grad_norm": 1.2473630905151367, "learning_rate": 0.00019951267775701194, "loss": 4.6633, "step": 129550 }, { "epoch": 0.2418480099090504, "grad_norm": 1.0328094959259033, "learning_rate": 0.00019951226343679419, "loss": 4.6325, "step": 129600 }, { "epoch": 0.2419413154684289, "grad_norm": 0.8722690343856812, "learning_rate": 0.00019951184894095481, "loss": 4.7876, "step": 129650 }, { "epoch": 0.2420346210278074, "grad_norm": 0.8723517060279846, "learning_rate": 0.00019951143426949453, "loss": 4.6523, "step": 129700 }, { "epoch": 0.24212792658718588, "grad_norm": 0.9116722941398621, "learning_rate": 0.00019951101942241407, "loss": 4.6544, "step": 129750 }, { "epoch": 0.24222123214656438, "grad_norm": 0.9696924686431885, "learning_rate": 0.0001995106043997142, "loss": 4.5401, "step": 129800 }, { "epoch": 0.24231453770594286, "grad_norm": 0.9570209383964539, "learning_rate": 0.00019951018920139556, "loss": 4.6208, "step": 129850 }, { "epoch": 0.24240784326532136, "grad_norm": 1.2670087814331055, "learning_rate": 0.00019950977382745898, "loss": 4.6398, "step": 129900 }, { "epoch": 0.24250114882469984, "grad_norm": 0.9587291479110718, "learning_rate": 0.00019950935827790515, "loss": 4.6175, "step": 129950 }, { "epoch": 0.24259445438407834, "grad_norm": 0.772126317024231, "learning_rate": 0.00019950894255273485, "loss": 4.4946, "step": 130000 }, { "epoch": 0.24268775994345684, "grad_norm": 1.0134305953979492, "learning_rate": 0.00019950852665194874, "loss": 4.7511, "step": 130050 }, { "epoch": 0.24278106550283532, "grad_norm": 0.9733072519302368, "learning_rate": 0.0001995081105755476, "loss": 4.7236, "step": 130100 }, { "epoch": 0.24287437106221382, "grad_norm": 0.7325246334075928, "learning_rate": 0.00019950769432353215, "loss": 4.783, "step": 130150 }, { "epoch": 0.2429676766215923, "grad_norm": 1.1216601133346558, "learning_rate": 0.00019950727789590312, "loss": 4.8321, "step": 130200 }, { "epoch": 0.2430609821809708, "grad_norm": 1.0440672636032104, "learning_rate": 0.00019950686129266128, "loss": 4.7706, "step": 130250 }, { "epoch": 0.24315428774034928, "grad_norm": 0.9024654030799866, "learning_rate": 0.0001995064445138073, "loss": 4.7164, "step": 130300 }, { "epoch": 0.24324759329972778, "grad_norm": 0.8911654949188232, "learning_rate": 0.00019950602755934197, "loss": 4.5, "step": 130350 }, { "epoch": 0.24334089885910629, "grad_norm": 0.9491771459579468, "learning_rate": 0.00019950561042926602, "loss": 4.6981, "step": 130400 }, { "epoch": 0.24343420441848476, "grad_norm": 0.9986402988433838, "learning_rate": 0.0001995051931235802, "loss": 4.8321, "step": 130450 }, { "epoch": 0.24352750997786327, "grad_norm": 0.9581927061080933, "learning_rate": 0.0001995047756422852, "loss": 4.7905, "step": 130500 }, { "epoch": 0.24362081553724174, "grad_norm": 0.8729482293128967, "learning_rate": 0.00019950435798538174, "loss": 4.5298, "step": 130550 }, { "epoch": 0.24371412109662025, "grad_norm": 1.103448510169983, "learning_rate": 0.00019950394015287064, "loss": 4.5703, "step": 130600 }, { "epoch": 0.24380742665599872, "grad_norm": 1.0000076293945312, "learning_rate": 0.00019950352214475258, "loss": 4.4457, "step": 130650 }, { "epoch": 0.24390073221537723, "grad_norm": 0.767560601234436, "learning_rate": 0.0001995031039610283, "loss": 5.0027, "step": 130700 }, { "epoch": 0.2439940377747557, "grad_norm": 0.7368218898773193, "learning_rate": 0.00019950268560169857, "loss": 4.6973, "step": 130750 }, { "epoch": 0.2440873433341342, "grad_norm": 0.9467607140541077, "learning_rate": 0.0001995022670667641, "loss": 4.4799, "step": 130800 }, { "epoch": 0.2441806488935127, "grad_norm": 0.8919101357460022, "learning_rate": 0.00019950184835622564, "loss": 4.563, "step": 130850 }, { "epoch": 0.24427395445289118, "grad_norm": 0.9824000000953674, "learning_rate": 0.00019950142947008394, "loss": 4.3836, "step": 130900 }, { "epoch": 0.2443672600122697, "grad_norm": 0.9101470112800598, "learning_rate": 0.00019950101040833973, "loss": 4.5129, "step": 130950 }, { "epoch": 0.24446056557164816, "grad_norm": 0.9340775609016418, "learning_rate": 0.00019950059117099372, "loss": 4.5774, "step": 131000 }, { "epoch": 0.24455387113102667, "grad_norm": 0.7431853413581848, "learning_rate": 0.00019950017175804664, "loss": 4.6977, "step": 131050 }, { "epoch": 0.24464717669040514, "grad_norm": 1.0146548748016357, "learning_rate": 0.00019949975216949933, "loss": 4.6435, "step": 131100 }, { "epoch": 0.24474048224978365, "grad_norm": 1.0537569522857666, "learning_rate": 0.00019949933240535244, "loss": 4.6641, "step": 131150 }, { "epoch": 0.24483378780916215, "grad_norm": 1.1558393239974976, "learning_rate": 0.00019949891246560673, "loss": 4.7927, "step": 131200 }, { "epoch": 0.24492709336854063, "grad_norm": 0.8795566558837891, "learning_rate": 0.00019949849235026296, "loss": 4.7991, "step": 131250 }, { "epoch": 0.24502039892791913, "grad_norm": 1.0229332447052002, "learning_rate": 0.00019949807205932185, "loss": 4.6722, "step": 131300 }, { "epoch": 0.2451137044872976, "grad_norm": 0.985419750213623, "learning_rate": 0.00019949765159278413, "loss": 4.6676, "step": 131350 }, { "epoch": 0.2452070100466761, "grad_norm": 0.8589661121368408, "learning_rate": 0.00019949723095065062, "loss": 4.5871, "step": 131400 }, { "epoch": 0.2453003156060546, "grad_norm": 1.0366382598876953, "learning_rate": 0.00019949681013292194, "loss": 4.7374, "step": 131450 }, { "epoch": 0.2453936211654331, "grad_norm": 1.379646897315979, "learning_rate": 0.00019949638913959892, "loss": 4.7077, "step": 131500 }, { "epoch": 0.2454869267248116, "grad_norm": 1.0151523351669312, "learning_rate": 0.0001994959679706823, "loss": 4.6403, "step": 131550 }, { "epoch": 0.24558023228419007, "grad_norm": 0.8096948266029358, "learning_rate": 0.00019949554662617277, "loss": 4.6464, "step": 131600 }, { "epoch": 0.24567353784356857, "grad_norm": 1.1682924032211304, "learning_rate": 0.00019949512510607115, "loss": 4.6433, "step": 131650 }, { "epoch": 0.24576684340294705, "grad_norm": 1.0436276197433472, "learning_rate": 0.00019949470341037808, "loss": 4.6682, "step": 131700 }, { "epoch": 0.24586014896232555, "grad_norm": 1.0496689081192017, "learning_rate": 0.00019949428153909443, "loss": 4.7054, "step": 131750 }, { "epoch": 0.24595345452170403, "grad_norm": 1.0121614933013916, "learning_rate": 0.00019949385949222083, "loss": 4.656, "step": 131800 }, { "epoch": 0.24604676008108253, "grad_norm": 0.8388519287109375, "learning_rate": 0.00019949343726975808, "loss": 4.5378, "step": 131850 }, { "epoch": 0.24614006564046104, "grad_norm": 0.9778211116790771, "learning_rate": 0.00019949301487170694, "loss": 4.6026, "step": 131900 }, { "epoch": 0.2462333711998395, "grad_norm": 1.1076475381851196, "learning_rate": 0.0001994925922980681, "loss": 4.7711, "step": 131950 }, { "epoch": 0.24632667675921802, "grad_norm": 1.092414379119873, "learning_rate": 0.0001994921695488424, "loss": 4.6508, "step": 132000 }, { "epoch": 0.24632667675921802, "eval_loss": 4.861928939819336, "eval_runtime": 233.537, "eval_samples_per_second": 11.167, "eval_steps_per_second": 11.167, "eval_tts_loss": 7.461830320030999, "step": 132000 }, { "epoch": 0.2464199823185965, "grad_norm": 1.5649731159210205, "learning_rate": 0.0001994917466240305, "loss": 4.7985, "step": 132050 }, { "epoch": 0.246513287877975, "grad_norm": 0.7559811472892761, "learning_rate": 0.00019949132352363317, "loss": 4.545, "step": 132100 }, { "epoch": 0.24660659343735347, "grad_norm": 1.1505687236785889, "learning_rate": 0.00019949090024765114, "loss": 4.7491, "step": 132150 }, { "epoch": 0.24669989899673198, "grad_norm": 0.856060802936554, "learning_rate": 0.0001994904767960852, "loss": 4.6556, "step": 132200 }, { "epoch": 0.24679320455611045, "grad_norm": 1.4347225427627563, "learning_rate": 0.00019949005316893608, "loss": 4.6111, "step": 132250 }, { "epoch": 0.24688651011548896, "grad_norm": 1.2799773216247559, "learning_rate": 0.0001994896293662045, "loss": 4.8259, "step": 132300 }, { "epoch": 0.24697981567486746, "grad_norm": 0.7314300537109375, "learning_rate": 0.00019948920538789125, "loss": 4.5923, "step": 132350 }, { "epoch": 0.24707312123424593, "grad_norm": 1.4587212800979614, "learning_rate": 0.00019948878123399703, "loss": 4.6689, "step": 132400 }, { "epoch": 0.24716642679362444, "grad_norm": 0.7531588077545166, "learning_rate": 0.00019948835690452264, "loss": 4.6059, "step": 132450 }, { "epoch": 0.24725973235300291, "grad_norm": 1.1337380409240723, "learning_rate": 0.00019948793239946881, "loss": 4.6628, "step": 132500 }, { "epoch": 0.24735303791238142, "grad_norm": 0.6850484609603882, "learning_rate": 0.00019948750771883628, "loss": 4.7061, "step": 132550 }, { "epoch": 0.2474463434717599, "grad_norm": 1.0581914186477661, "learning_rate": 0.00019948708286262581, "loss": 4.7761, "step": 132600 }, { "epoch": 0.2475396490311384, "grad_norm": 0.8422567844390869, "learning_rate": 0.00019948665783083814, "loss": 4.8732, "step": 132650 }, { "epoch": 0.2476329545905169, "grad_norm": 1.0814651250839233, "learning_rate": 0.000199486232623474, "loss": 4.8206, "step": 132700 }, { "epoch": 0.24772626014989538, "grad_norm": 1.1069611310958862, "learning_rate": 0.00019948580724053422, "loss": 4.8655, "step": 132750 }, { "epoch": 0.24781956570927388, "grad_norm": 1.0033888816833496, "learning_rate": 0.00019948538168201944, "loss": 4.6903, "step": 132800 }, { "epoch": 0.24791287126865236, "grad_norm": 0.9167322516441345, "learning_rate": 0.00019948495594793048, "loss": 4.4695, "step": 132850 }, { "epoch": 0.24800617682803086, "grad_norm": 0.970410943031311, "learning_rate": 0.00019948453003826807, "loss": 4.647, "step": 132900 }, { "epoch": 0.24809948238740934, "grad_norm": 1.0341490507125854, "learning_rate": 0.00019948410395303302, "loss": 4.6685, "step": 132950 }, { "epoch": 0.24819278794678784, "grad_norm": 0.9720880389213562, "learning_rate": 0.000199483677692226, "loss": 4.6414, "step": 133000 }, { "epoch": 0.24828609350616634, "grad_norm": 0.8272822499275208, "learning_rate": 0.00019948325125584778, "loss": 4.8563, "step": 133050 }, { "epoch": 0.24837939906554482, "grad_norm": 1.2242487668991089, "learning_rate": 0.0001994828246438991, "loss": 4.6999, "step": 133100 }, { "epoch": 0.24847270462492332, "grad_norm": 0.9121090769767761, "learning_rate": 0.0001994823978563808, "loss": 4.752, "step": 133150 }, { "epoch": 0.2485660101843018, "grad_norm": 0.7485866546630859, "learning_rate": 0.00019948197089329352, "loss": 4.6905, "step": 133200 }, { "epoch": 0.2486593157436803, "grad_norm": 0.6626511812210083, "learning_rate": 0.0001994815437546381, "loss": 4.5935, "step": 133250 }, { "epoch": 0.24875262130305878, "grad_norm": 1.1593636274337769, "learning_rate": 0.00019948111644041525, "loss": 4.9399, "step": 133300 }, { "epoch": 0.24884592686243728, "grad_norm": 1.0495115518569946, "learning_rate": 0.00019948068895062575, "loss": 4.6613, "step": 133350 }, { "epoch": 0.24893923242181576, "grad_norm": 1.1008858680725098, "learning_rate": 0.00019948026128527034, "loss": 4.9013, "step": 133400 }, { "epoch": 0.24903253798119426, "grad_norm": 0.784603476524353, "learning_rate": 0.00019947983344434974, "loss": 4.6166, "step": 133450 }, { "epoch": 0.24912584354057277, "grad_norm": 0.9832925796508789, "learning_rate": 0.00019947940542786477, "loss": 4.7428, "step": 133500 }, { "epoch": 0.24921914909995124, "grad_norm": 0.9270418286323547, "learning_rate": 0.0001994789772358161, "loss": 4.4206, "step": 133550 }, { "epoch": 0.24931245465932975, "grad_norm": 1.2622601985931396, "learning_rate": 0.00019947854886820458, "loss": 4.6319, "step": 133600 }, { "epoch": 0.24940576021870822, "grad_norm": 0.7396588325500488, "learning_rate": 0.00019947812032503092, "loss": 4.786, "step": 133650 }, { "epoch": 0.24949906577808673, "grad_norm": 0.9550930261611938, "learning_rate": 0.00019947769160629587, "loss": 4.7091, "step": 133700 }, { "epoch": 0.2495923713374652, "grad_norm": 1.0118082761764526, "learning_rate": 0.00019947726271200025, "loss": 4.7579, "step": 133750 }, { "epoch": 0.2496856768968437, "grad_norm": 0.838410496711731, "learning_rate": 0.0001994768336421447, "loss": 4.7016, "step": 133800 }, { "epoch": 0.2497789824562222, "grad_norm": 0.8363553881645203, "learning_rate": 0.00019947640439673008, "loss": 4.6482, "step": 133850 }, { "epoch": 0.24987228801560069, "grad_norm": 1.0128004550933838, "learning_rate": 0.0001994759749757571, "loss": 4.8391, "step": 133900 }, { "epoch": 0.2499655935749792, "grad_norm": 0.8684887290000916, "learning_rate": 0.00019947554537922652, "loss": 4.5435, "step": 133950 }, { "epoch": 0.2500588991343577, "grad_norm": 0.8165358901023865, "learning_rate": 0.0001994751156071391, "loss": 4.6167, "step": 134000 }, { "epoch": 0.25015220469373617, "grad_norm": 0.9833562970161438, "learning_rate": 0.00019947468565949562, "loss": 4.7929, "step": 134050 }, { "epoch": 0.25024551025311464, "grad_norm": 0.8654858469963074, "learning_rate": 0.00019947425553629682, "loss": 4.8782, "step": 134100 }, { "epoch": 0.2503388158124931, "grad_norm": 0.8595501184463501, "learning_rate": 0.00019947382523754348, "loss": 4.6882, "step": 134150 }, { "epoch": 0.25043212137187165, "grad_norm": 0.9764409065246582, "learning_rate": 0.0001994733947632363, "loss": 4.723, "step": 134200 }, { "epoch": 0.25052542693125013, "grad_norm": 1.361682415008545, "learning_rate": 0.0001994729641133761, "loss": 4.9449, "step": 134250 }, { "epoch": 0.2506187324906286, "grad_norm": 1.0243170261383057, "learning_rate": 0.00019947253328796364, "loss": 4.8382, "step": 134300 }, { "epoch": 0.25071203805000714, "grad_norm": 0.9603681564331055, "learning_rate": 0.00019947210228699963, "loss": 4.7538, "step": 134350 }, { "epoch": 0.2508053436093856, "grad_norm": 0.8127009272575378, "learning_rate": 0.0001994716711104849, "loss": 4.74, "step": 134400 }, { "epoch": 0.2508986491687641, "grad_norm": 1.0043110847473145, "learning_rate": 0.00019947123975842012, "loss": 4.6537, "step": 134450 }, { "epoch": 0.25099195472814256, "grad_norm": 0.88910973072052, "learning_rate": 0.00019947080823080616, "loss": 4.6681, "step": 134500 }, { "epoch": 0.2510852602875211, "grad_norm": 1.072106957435608, "learning_rate": 0.00019947037652764369, "loss": 4.6476, "step": 134550 }, { "epoch": 0.25117856584689957, "grad_norm": 1.160860538482666, "learning_rate": 0.0001994699446489335, "loss": 4.8706, "step": 134600 }, { "epoch": 0.25127187140627805, "grad_norm": 1.0427870750427246, "learning_rate": 0.0001994695125946764, "loss": 4.7866, "step": 134650 }, { "epoch": 0.2513651769656566, "grad_norm": 0.8088847398757935, "learning_rate": 0.00019946908036487307, "loss": 4.5036, "step": 134700 }, { "epoch": 0.25145848252503505, "grad_norm": 0.8402940034866333, "learning_rate": 0.00019946864795952434, "loss": 4.4938, "step": 134750 }, { "epoch": 0.25155178808441353, "grad_norm": 0.9996683597564697, "learning_rate": 0.00019946821537863095, "loss": 4.6048, "step": 134800 }, { "epoch": 0.251645093643792, "grad_norm": 0.9219832420349121, "learning_rate": 0.00019946778262219366, "loss": 4.6674, "step": 134850 }, { "epoch": 0.25173839920317054, "grad_norm": 0.9935362935066223, "learning_rate": 0.00019946734969021322, "loss": 4.8539, "step": 134900 }, { "epoch": 0.251831704762549, "grad_norm": 1.0310291051864624, "learning_rate": 0.00019946691658269042, "loss": 4.5868, "step": 134950 }, { "epoch": 0.2519250103219275, "grad_norm": 0.9530249238014221, "learning_rate": 0.00019946648329962602, "loss": 4.6397, "step": 135000 }, { "epoch": 0.2519250103219275, "eval_loss": 4.854243755340576, "eval_runtime": 229.3514, "eval_samples_per_second": 11.371, "eval_steps_per_second": 11.371, "eval_tts_loss": 7.440514219263454, "step": 135000 }, { "epoch": 0.252018315881306, "grad_norm": 0.7310658097267151, "learning_rate": 0.00019946604984102077, "loss": 4.6717, "step": 135050 }, { "epoch": 0.2521116214406845, "grad_norm": 0.7537361979484558, "learning_rate": 0.00019946561620687544, "loss": 4.6167, "step": 135100 }, { "epoch": 0.252204927000063, "grad_norm": 0.9830604195594788, "learning_rate": 0.00019946518239719082, "loss": 4.4883, "step": 135150 }, { "epoch": 0.25229823255944145, "grad_norm": 0.8440564870834351, "learning_rate": 0.00019946474841196765, "loss": 4.7197, "step": 135200 }, { "epoch": 0.25239153811882, "grad_norm": 1.2416318655014038, "learning_rate": 0.00019946431425120665, "loss": 4.5972, "step": 135250 }, { "epoch": 0.25248484367819846, "grad_norm": 1.1739615201950073, "learning_rate": 0.00019946387991490871, "loss": 4.4314, "step": 135300 }, { "epoch": 0.25257814923757693, "grad_norm": 1.079363226890564, "learning_rate": 0.0001994634454030745, "loss": 4.7442, "step": 135350 }, { "epoch": 0.25267145479695546, "grad_norm": 1.0851373672485352, "learning_rate": 0.00019946301071570478, "loss": 4.7188, "step": 135400 }, { "epoch": 0.25276476035633394, "grad_norm": 0.8519002199172974, "learning_rate": 0.00019946257585280036, "loss": 4.5734, "step": 135450 }, { "epoch": 0.2528580659157124, "grad_norm": 1.175642728805542, "learning_rate": 0.00019946214081436204, "loss": 4.8021, "step": 135500 }, { "epoch": 0.2529513714750909, "grad_norm": 0.984249472618103, "learning_rate": 0.0001994617056003905, "loss": 4.7219, "step": 135550 }, { "epoch": 0.2530446770344694, "grad_norm": 1.4463765621185303, "learning_rate": 0.00019946127021088655, "loss": 4.5299, "step": 135600 }, { "epoch": 0.2531379825938479, "grad_norm": 1.2375439405441284, "learning_rate": 0.00019946083464585096, "loss": 4.6073, "step": 135650 }, { "epoch": 0.2532312881532264, "grad_norm": 1.073440432548523, "learning_rate": 0.0001994603989052845, "loss": 4.5566, "step": 135700 }, { "epoch": 0.2533245937126049, "grad_norm": 1.0174309015274048, "learning_rate": 0.00019945996298918793, "loss": 4.6937, "step": 135750 }, { "epoch": 0.2534178992719834, "grad_norm": 0.9919219613075256, "learning_rate": 0.00019945952689756204, "loss": 4.665, "step": 135800 }, { "epoch": 0.25351120483136186, "grad_norm": 0.8121524453163147, "learning_rate": 0.0001994590906304076, "loss": 4.882, "step": 135850 }, { "epoch": 0.25360451039074033, "grad_norm": 0.5900788903236389, "learning_rate": 0.00019945865418772533, "loss": 4.7844, "step": 135900 }, { "epoch": 0.25369781595011887, "grad_norm": 1.1510095596313477, "learning_rate": 0.00019945821756951605, "loss": 4.729, "step": 135950 }, { "epoch": 0.25379112150949734, "grad_norm": 0.9832910895347595, "learning_rate": 0.0001994577807757805, "loss": 4.7004, "step": 136000 }, { "epoch": 0.2538844270688758, "grad_norm": 0.8559871315956116, "learning_rate": 0.0001994573438065195, "loss": 4.7026, "step": 136050 }, { "epoch": 0.25397773262825435, "grad_norm": 0.9402391910552979, "learning_rate": 0.00019945690666173376, "loss": 4.5792, "step": 136100 }, { "epoch": 0.2540710381876328, "grad_norm": 0.9144654870033264, "learning_rate": 0.00019945646934142408, "loss": 4.6403, "step": 136150 }, { "epoch": 0.2541643437470113, "grad_norm": 1.0262720584869385, "learning_rate": 0.00019945603184559125, "loss": 4.766, "step": 136200 }, { "epoch": 0.2542576493063898, "grad_norm": 0.9860953092575073, "learning_rate": 0.00019945559417423604, "loss": 4.7118, "step": 136250 }, { "epoch": 0.2543509548657683, "grad_norm": 0.9816266298294067, "learning_rate": 0.00019945515632735917, "loss": 4.6694, "step": 136300 }, { "epoch": 0.2544442604251468, "grad_norm": 1.0225788354873657, "learning_rate": 0.00019945471830496144, "loss": 4.6733, "step": 136350 }, { "epoch": 0.25453756598452526, "grad_norm": 0.8615061640739441, "learning_rate": 0.00019945428010704368, "loss": 4.4516, "step": 136400 }, { "epoch": 0.2546308715439038, "grad_norm": 0.9485636353492737, "learning_rate": 0.00019945384173360658, "loss": 4.7265, "step": 136450 }, { "epoch": 0.25472417710328227, "grad_norm": 0.9879016876220703, "learning_rate": 0.00019945340318465097, "loss": 4.6516, "step": 136500 }, { "epoch": 0.25481748266266074, "grad_norm": 0.886114776134491, "learning_rate": 0.00019945296446017755, "loss": 4.7834, "step": 136550 }, { "epoch": 0.2549107882220392, "grad_norm": 0.901164174079895, "learning_rate": 0.00019945252556018722, "loss": 4.7377, "step": 136600 }, { "epoch": 0.25500409378141775, "grad_norm": 1.0700454711914062, "learning_rate": 0.00019945208648468064, "loss": 4.629, "step": 136650 }, { "epoch": 0.2550973993407962, "grad_norm": 0.761396586894989, "learning_rate": 0.00019945164723365863, "loss": 4.5535, "step": 136700 }, { "epoch": 0.2551907049001747, "grad_norm": 0.9955791234970093, "learning_rate": 0.00019945120780712197, "loss": 4.8143, "step": 136750 }, { "epoch": 0.2552840104595532, "grad_norm": 0.7975375652313232, "learning_rate": 0.00019945076820507144, "loss": 4.6698, "step": 136800 }, { "epoch": 0.2553773160189317, "grad_norm": 0.8464979529380798, "learning_rate": 0.00019945032842750776, "loss": 4.6989, "step": 136850 }, { "epoch": 0.2554706215783102, "grad_norm": 0.9093438386917114, "learning_rate": 0.0001994498884744318, "loss": 4.7043, "step": 136900 }, { "epoch": 0.25556392713768866, "grad_norm": 0.7675898671150208, "learning_rate": 0.00019944944834584426, "loss": 4.8434, "step": 136950 }, { "epoch": 0.2556572326970672, "grad_norm": 1.0677566528320312, "learning_rate": 0.00019944900804174593, "loss": 4.7249, "step": 137000 }, { "epoch": 0.25575053825644567, "grad_norm": 0.7189279794692993, "learning_rate": 0.00019944856756213763, "loss": 4.5614, "step": 137050 }, { "epoch": 0.25584384381582415, "grad_norm": 1.0350892543792725, "learning_rate": 0.0001994481269070201, "loss": 4.963, "step": 137100 }, { "epoch": 0.2559371493752026, "grad_norm": 0.7921010255813599, "learning_rate": 0.0001994476860763941, "loss": 4.5003, "step": 137150 }, { "epoch": 0.25603045493458115, "grad_norm": 1.237960934638977, "learning_rate": 0.00019944724507026045, "loss": 4.6863, "step": 137200 }, { "epoch": 0.25612376049395963, "grad_norm": 0.9462816119194031, "learning_rate": 0.0001994468038886199, "loss": 4.6557, "step": 137250 }, { "epoch": 0.2562170660533381, "grad_norm": 1.1308284997940063, "learning_rate": 0.00019944636253147325, "loss": 4.5256, "step": 137300 }, { "epoch": 0.25631037161271664, "grad_norm": 0.8221685886383057, "learning_rate": 0.00019944592099882126, "loss": 4.4556, "step": 137350 }, { "epoch": 0.2564036771720951, "grad_norm": 0.8622978925704956, "learning_rate": 0.00019944547929066472, "loss": 4.5635, "step": 137400 }, { "epoch": 0.2564969827314736, "grad_norm": 1.0377954244613647, "learning_rate": 0.0001994450374070044, "loss": 4.7431, "step": 137450 }, { "epoch": 0.25659028829085206, "grad_norm": 0.9822484254837036, "learning_rate": 0.00019944459534784113, "loss": 4.5507, "step": 137500 }, { "epoch": 0.2566835938502306, "grad_norm": 1.063193917274475, "learning_rate": 0.0001994441531131756, "loss": 4.6944, "step": 137550 }, { "epoch": 0.25677689940960907, "grad_norm": 0.9671196937561035, "learning_rate": 0.00019944371070300865, "loss": 4.7503, "step": 137600 }, { "epoch": 0.25687020496898755, "grad_norm": 1.0232065916061401, "learning_rate": 0.00019944326811734104, "loss": 4.4999, "step": 137650 }, { "epoch": 0.2569635105283661, "grad_norm": 0.7935104966163635, "learning_rate": 0.00019944282535617358, "loss": 4.7047, "step": 137700 }, { "epoch": 0.25705681608774456, "grad_norm": 1.06540846824646, "learning_rate": 0.00019944238241950704, "loss": 4.8058, "step": 137750 }, { "epoch": 0.25715012164712303, "grad_norm": 1.0669400691986084, "learning_rate": 0.00019944193930734216, "loss": 4.6499, "step": 137800 }, { "epoch": 0.2572434272065015, "grad_norm": 0.9878630042076111, "learning_rate": 0.00019944149601967977, "loss": 4.5647, "step": 137850 }, { "epoch": 0.25733673276588004, "grad_norm": 1.0921598672866821, "learning_rate": 0.00019944105255652065, "loss": 4.6141, "step": 137900 }, { "epoch": 0.2574300383252585, "grad_norm": 0.7695943117141724, "learning_rate": 0.00019944060891786554, "loss": 4.5943, "step": 137950 }, { "epoch": 0.257523343884637, "grad_norm": 1.0480443239212036, "learning_rate": 0.00019944016510371526, "loss": 4.7454, "step": 138000 }, { "epoch": 0.257523343884637, "eval_loss": 4.861461162567139, "eval_runtime": 230.715, "eval_samples_per_second": 11.304, "eval_steps_per_second": 11.304, "eval_tts_loss": 7.436345217680913, "step": 138000 }, { "epoch": 0.2576166494440155, "grad_norm": 1.0323460102081299, "learning_rate": 0.0001994397211140706, "loss": 4.7979, "step": 138050 }, { "epoch": 0.257709955003394, "grad_norm": 1.2108174562454224, "learning_rate": 0.0001994392769489323, "loss": 4.5913, "step": 138100 }, { "epoch": 0.2578032605627725, "grad_norm": 0.9902898073196411, "learning_rate": 0.0001994388326083012, "loss": 4.4259, "step": 138150 }, { "epoch": 0.25789656612215095, "grad_norm": 1.0889242887496948, "learning_rate": 0.00019943838809217805, "loss": 4.6503, "step": 138200 }, { "epoch": 0.2579898716815295, "grad_norm": 0.991397500038147, "learning_rate": 0.00019943794340056362, "loss": 4.6364, "step": 138250 }, { "epoch": 0.25808317724090796, "grad_norm": 1.141694188117981, "learning_rate": 0.00019943749853345875, "loss": 4.7025, "step": 138300 }, { "epoch": 0.25817648280028643, "grad_norm": 0.8368604183197021, "learning_rate": 0.0001994370534908642, "loss": 4.6701, "step": 138350 }, { "epoch": 0.25826978835966496, "grad_norm": 1.0592689514160156, "learning_rate": 0.00019943660827278072, "loss": 4.569, "step": 138400 }, { "epoch": 0.25836309391904344, "grad_norm": 0.7820909023284912, "learning_rate": 0.00019943616287920914, "loss": 4.7223, "step": 138450 }, { "epoch": 0.2584563994784219, "grad_norm": 1.0248945951461792, "learning_rate": 0.0001994357173101502, "loss": 4.6945, "step": 138500 }, { "epoch": 0.2585497050378004, "grad_norm": 0.9100630283355713, "learning_rate": 0.00019943527156560475, "loss": 4.5899, "step": 138550 }, { "epoch": 0.2586430105971789, "grad_norm": 1.1249892711639404, "learning_rate": 0.00019943482564557352, "loss": 4.6453, "step": 138600 }, { "epoch": 0.2587363161565574, "grad_norm": 1.033903956413269, "learning_rate": 0.00019943437955005733, "loss": 4.7613, "step": 138650 }, { "epoch": 0.2588296217159359, "grad_norm": 1.0414884090423584, "learning_rate": 0.00019943393327905695, "loss": 4.6118, "step": 138700 }, { "epoch": 0.2589229272753144, "grad_norm": 0.7779209017753601, "learning_rate": 0.00019943348683257317, "loss": 4.7519, "step": 138750 }, { "epoch": 0.2590162328346929, "grad_norm": 1.0751841068267822, "learning_rate": 0.0001994330402106068, "loss": 4.5788, "step": 138800 }, { "epoch": 0.25910953839407136, "grad_norm": 0.7722923159599304, "learning_rate": 0.00019943259341315857, "loss": 4.4937, "step": 138850 }, { "epoch": 0.25920284395344984, "grad_norm": 0.8846758008003235, "learning_rate": 0.00019943214644022934, "loss": 4.7343, "step": 138900 }, { "epoch": 0.25929614951282837, "grad_norm": 1.157002329826355, "learning_rate": 0.00019943169929181987, "loss": 4.8432, "step": 138950 }, { "epoch": 0.25938945507220684, "grad_norm": 0.8861170411109924, "learning_rate": 0.00019943125196793094, "loss": 4.7967, "step": 139000 }, { "epoch": 0.2594827606315853, "grad_norm": 0.9354442954063416, "learning_rate": 0.00019943080446856333, "loss": 4.6719, "step": 139050 }, { "epoch": 0.25957606619096385, "grad_norm": 1.0083680152893066, "learning_rate": 0.00019943035679371787, "loss": 4.6639, "step": 139100 }, { "epoch": 0.2596693717503423, "grad_norm": 1.0523858070373535, "learning_rate": 0.0001994299089433953, "loss": 4.6234, "step": 139150 }, { "epoch": 0.2597626773097208, "grad_norm": 0.9569562077522278, "learning_rate": 0.00019942946091759648, "loss": 4.4507, "step": 139200 }, { "epoch": 0.2598559828690993, "grad_norm": 1.1821283102035522, "learning_rate": 0.00019942901271632213, "loss": 4.7459, "step": 139250 }, { "epoch": 0.2599492884284778, "grad_norm": 0.9408050179481506, "learning_rate": 0.00019942856433957304, "loss": 4.6761, "step": 139300 }, { "epoch": 0.2600425939878563, "grad_norm": 0.939137876033783, "learning_rate": 0.00019942811578735006, "loss": 4.6466, "step": 139350 }, { "epoch": 0.26013589954723476, "grad_norm": 0.8856296539306641, "learning_rate": 0.00019942766705965395, "loss": 4.9821, "step": 139400 }, { "epoch": 0.26022920510661324, "grad_norm": 0.6106601357460022, "learning_rate": 0.00019942721815648548, "loss": 4.5543, "step": 139450 }, { "epoch": 0.26032251066599177, "grad_norm": 0.8569189310073853, "learning_rate": 0.0001994267690778455, "loss": 4.6647, "step": 139500 }, { "epoch": 0.26041581622537024, "grad_norm": 1.102795958518982, "learning_rate": 0.00019942631982373474, "loss": 4.4512, "step": 139550 }, { "epoch": 0.2605091217847487, "grad_norm": 0.9730409979820251, "learning_rate": 0.00019942587039415403, "loss": 4.9565, "step": 139600 }, { "epoch": 0.26060242734412725, "grad_norm": 0.7721739411354065, "learning_rate": 0.00019942542078910415, "loss": 4.7068, "step": 139650 }, { "epoch": 0.26069573290350573, "grad_norm": 0.70539790391922, "learning_rate": 0.0001994249710085859, "loss": 4.5592, "step": 139700 }, { "epoch": 0.2607890384628842, "grad_norm": 1.0219758749008179, "learning_rate": 0.00019942452105260007, "loss": 4.6835, "step": 139750 }, { "epoch": 0.2608823440222627, "grad_norm": 1.2075634002685547, "learning_rate": 0.00019942407092114745, "loss": 4.5386, "step": 139800 }, { "epoch": 0.2609756495816412, "grad_norm": 0.8083966970443726, "learning_rate": 0.00019942362061422885, "loss": 4.6374, "step": 139850 }, { "epoch": 0.2610689551410197, "grad_norm": 0.7901577353477478, "learning_rate": 0.00019942317013184506, "loss": 4.7071, "step": 139900 }, { "epoch": 0.26116226070039816, "grad_norm": 1.2326618432998657, "learning_rate": 0.00019942271947399683, "loss": 4.6927, "step": 139950 }, { "epoch": 0.2612555662597767, "grad_norm": 0.7666860818862915, "learning_rate": 0.00019942226864068502, "loss": 4.8306, "step": 140000 }, { "epoch": 0.26134887181915517, "grad_norm": 0.9988517761230469, "learning_rate": 0.00019942181763191042, "loss": 4.6557, "step": 140050 }, { "epoch": 0.26144217737853365, "grad_norm": 0.7240771651268005, "learning_rate": 0.0001994213664476738, "loss": 4.4553, "step": 140100 }, { "epoch": 0.2615354829379121, "grad_norm": 1.13454270362854, "learning_rate": 0.00019942091508797595, "loss": 4.5737, "step": 140150 }, { "epoch": 0.26162878849729065, "grad_norm": 0.8832876682281494, "learning_rate": 0.00019942046355281767, "loss": 4.7567, "step": 140200 }, { "epoch": 0.26172209405666913, "grad_norm": 1.089231014251709, "learning_rate": 0.00019942001184219976, "loss": 4.5035, "step": 140250 }, { "epoch": 0.2618153996160476, "grad_norm": 1.0818380117416382, "learning_rate": 0.00019941955995612305, "loss": 4.8077, "step": 140300 }, { "epoch": 0.26190870517542614, "grad_norm": 1.0916963815689087, "learning_rate": 0.00019941910789458825, "loss": 4.7335, "step": 140350 }, { "epoch": 0.2620020107348046, "grad_norm": 1.029809594154358, "learning_rate": 0.00019941865565759628, "loss": 4.7261, "step": 140400 }, { "epoch": 0.2620953162941831, "grad_norm": 0.786016047000885, "learning_rate": 0.00019941820324514785, "loss": 4.6044, "step": 140450 }, { "epoch": 0.26218862185356157, "grad_norm": 0.9927809834480286, "learning_rate": 0.0001994177506572438, "loss": 4.5903, "step": 140500 }, { "epoch": 0.2622819274129401, "grad_norm": 1.1766144037246704, "learning_rate": 0.0001994172978938849, "loss": 4.4297, "step": 140550 }, { "epoch": 0.2623752329723186, "grad_norm": 0.8965726494789124, "learning_rate": 0.00019941684495507198, "loss": 4.7621, "step": 140600 }, { "epoch": 0.26246853853169705, "grad_norm": 0.9904941916465759, "learning_rate": 0.00019941639184080582, "loss": 4.5519, "step": 140650 }, { "epoch": 0.2625618440910756, "grad_norm": 0.9716530442237854, "learning_rate": 0.0001994159385510872, "loss": 4.6065, "step": 140700 }, { "epoch": 0.26265514965045406, "grad_norm": 0.7707858085632324, "learning_rate": 0.00019941548508591695, "loss": 4.6938, "step": 140750 }, { "epoch": 0.26274845520983253, "grad_norm": 1.0510951280593872, "learning_rate": 0.00019941503144529587, "loss": 4.8028, "step": 140800 }, { "epoch": 0.262841760769211, "grad_norm": 0.8451001048088074, "learning_rate": 0.00019941457762922475, "loss": 4.4955, "step": 140850 }, { "epoch": 0.26293506632858954, "grad_norm": 0.8992056250572205, "learning_rate": 0.00019941412363770438, "loss": 4.6839, "step": 140900 }, { "epoch": 0.263028371887968, "grad_norm": 0.8834390640258789, "learning_rate": 0.00019941366947073559, "loss": 4.6869, "step": 140950 }, { "epoch": 0.2631216774473465, "grad_norm": 1.1744734048843384, "learning_rate": 0.00019941321512831914, "loss": 4.6612, "step": 141000 }, { "epoch": 0.2631216774473465, "eval_loss": 4.854221820831299, "eval_runtime": 230.4865, "eval_samples_per_second": 11.315, "eval_steps_per_second": 11.315, "eval_tts_loss": 7.429972851733202, "step": 141000 }, { "epoch": 0.263214983006725, "grad_norm": 1.01349675655365, "learning_rate": 0.00019941276061045588, "loss": 4.7109, "step": 141050 }, { "epoch": 0.2633082885661035, "grad_norm": 0.7250033617019653, "learning_rate": 0.0001994123059171466, "loss": 4.6136, "step": 141100 }, { "epoch": 0.263401594125482, "grad_norm": 0.8900704979896545, "learning_rate": 0.00019941185104839208, "loss": 4.4201, "step": 141150 }, { "epoch": 0.26349489968486045, "grad_norm": 0.9894741773605347, "learning_rate": 0.00019941139600419315, "loss": 4.5358, "step": 141200 }, { "epoch": 0.263588205244239, "grad_norm": 0.728403627872467, "learning_rate": 0.00019941094078455057, "loss": 4.4723, "step": 141250 }, { "epoch": 0.26368151080361746, "grad_norm": 0.9740440249443054, "learning_rate": 0.00019941048538946516, "loss": 4.9644, "step": 141300 }, { "epoch": 0.26377481636299593, "grad_norm": 0.9161767959594727, "learning_rate": 0.00019941002981893778, "loss": 4.7106, "step": 141350 }, { "epoch": 0.26386812192237447, "grad_norm": 1.2737964391708374, "learning_rate": 0.00019940957407296918, "loss": 4.569, "step": 141400 }, { "epoch": 0.26396142748175294, "grad_norm": 0.902451753616333, "learning_rate": 0.00019940911815156012, "loss": 4.6479, "step": 141450 }, { "epoch": 0.2640547330411314, "grad_norm": 0.8375920653343201, "learning_rate": 0.00019940866205471152, "loss": 4.7673, "step": 141500 }, { "epoch": 0.2641480386005099, "grad_norm": 0.6975705027580261, "learning_rate": 0.0001994082057824241, "loss": 4.638, "step": 141550 }, { "epoch": 0.2642413441598884, "grad_norm": 1.0885804891586304, "learning_rate": 0.00019940774933469867, "loss": 4.7292, "step": 141600 }, { "epoch": 0.2643346497192669, "grad_norm": 1.1714972257614136, "learning_rate": 0.00019940729271153608, "loss": 4.7405, "step": 141650 }, { "epoch": 0.2644279552786454, "grad_norm": 1.126559853553772, "learning_rate": 0.00019940683591293713, "loss": 4.58, "step": 141700 }, { "epoch": 0.2645212608380239, "grad_norm": 1.1203880310058594, "learning_rate": 0.00019940637893890256, "loss": 4.6255, "step": 141750 }, { "epoch": 0.2646145663974024, "grad_norm": 0.9070851802825928, "learning_rate": 0.00019940592178943324, "loss": 4.8198, "step": 141800 }, { "epoch": 0.26470787195678086, "grad_norm": 0.8276423215866089, "learning_rate": 0.00019940546446452998, "loss": 4.6199, "step": 141850 }, { "epoch": 0.26480117751615934, "grad_norm": 1.2383900880813599, "learning_rate": 0.0001994050069641935, "loss": 4.6207, "step": 141900 }, { "epoch": 0.26489448307553787, "grad_norm": 0.9433544278144836, "learning_rate": 0.00019940454928842475, "loss": 4.7679, "step": 141950 }, { "epoch": 0.26498778863491634, "grad_norm": 1.12177574634552, "learning_rate": 0.00019940409143722444, "loss": 4.8803, "step": 142000 }, { "epoch": 0.2650810941942948, "grad_norm": 0.8541791439056396, "learning_rate": 0.00019940363341059336, "loss": 4.7773, "step": 142050 }, { "epoch": 0.2651743997536733, "grad_norm": 1.2318682670593262, "learning_rate": 0.0001994031752085324, "loss": 4.7115, "step": 142100 }, { "epoch": 0.2652677053130518, "grad_norm": 1.1669272184371948, "learning_rate": 0.0001994027168310423, "loss": 4.6872, "step": 142150 }, { "epoch": 0.2653610108724303, "grad_norm": 0.7266700267791748, "learning_rate": 0.0001994022582781239, "loss": 4.4811, "step": 142200 }, { "epoch": 0.2654543164318088, "grad_norm": 0.6903916001319885, "learning_rate": 0.000199401799549778, "loss": 4.7914, "step": 142250 }, { "epoch": 0.2655476219911873, "grad_norm": 1.0394556522369385, "learning_rate": 0.00019940134064600547, "loss": 4.686, "step": 142300 }, { "epoch": 0.2656409275505658, "grad_norm": 1.02165687084198, "learning_rate": 0.00019940088156680702, "loss": 4.5818, "step": 142350 }, { "epoch": 0.26573423310994426, "grad_norm": 0.8454614281654358, "learning_rate": 0.0001994004223121835, "loss": 4.6726, "step": 142400 }, { "epoch": 0.26582753866932274, "grad_norm": 1.206411600112915, "learning_rate": 0.00019939996288213574, "loss": 4.6144, "step": 142450 }, { "epoch": 0.26592084422870127, "grad_norm": 0.9744541049003601, "learning_rate": 0.00019939950327666452, "loss": 4.902, "step": 142500 }, { "epoch": 0.26601414978807975, "grad_norm": 0.7370103001594543, "learning_rate": 0.00019939904349577069, "loss": 4.5981, "step": 142550 }, { "epoch": 0.2661074553474582, "grad_norm": 0.9200401902198792, "learning_rate": 0.000199398583539455, "loss": 4.5697, "step": 142600 }, { "epoch": 0.26620076090683675, "grad_norm": 1.0769352912902832, "learning_rate": 0.00019939812340771833, "loss": 4.7748, "step": 142650 }, { "epoch": 0.26629406646621523, "grad_norm": 1.1400402784347534, "learning_rate": 0.00019939766310056144, "loss": 4.7504, "step": 142700 }, { "epoch": 0.2663873720255937, "grad_norm": 0.8837327361106873, "learning_rate": 0.0001993972026179852, "loss": 4.7623, "step": 142750 }, { "epoch": 0.2664806775849722, "grad_norm": 0.9852148294448853, "learning_rate": 0.00019939674195999034, "loss": 4.7179, "step": 142800 }, { "epoch": 0.2665739831443507, "grad_norm": 1.1082642078399658, "learning_rate": 0.00019939628112657776, "loss": 4.6306, "step": 142850 }, { "epoch": 0.2666672887037292, "grad_norm": 1.2532259225845337, "learning_rate": 0.0001993958201177482, "loss": 4.7182, "step": 142900 }, { "epoch": 0.26676059426310766, "grad_norm": 1.113945484161377, "learning_rate": 0.00019939535893350252, "loss": 4.7235, "step": 142950 }, { "epoch": 0.2668538998224862, "grad_norm": 0.853884220123291, "learning_rate": 0.00019939489757384154, "loss": 4.7894, "step": 143000 }, { "epoch": 0.26694720538186467, "grad_norm": 1.0670225620269775, "learning_rate": 0.00019939443603876603, "loss": 4.6292, "step": 143050 }, { "epoch": 0.26704051094124315, "grad_norm": 1.090808629989624, "learning_rate": 0.00019939397432827685, "loss": 4.7517, "step": 143100 }, { "epoch": 0.2671338165006216, "grad_norm": 0.8925207853317261, "learning_rate": 0.0001993935124423748, "loss": 4.5034, "step": 143150 }, { "epoch": 0.26722712206000016, "grad_norm": 0.7665544748306274, "learning_rate": 0.00019939305038106064, "loss": 4.8442, "step": 143200 }, { "epoch": 0.26732042761937863, "grad_norm": 1.0686653852462769, "learning_rate": 0.00019939258814433527, "loss": 4.5459, "step": 143250 }, { "epoch": 0.2674137331787571, "grad_norm": 1.0229724645614624, "learning_rate": 0.00019939212573219948, "loss": 4.7637, "step": 143300 }, { "epoch": 0.26750703873813564, "grad_norm": 1.1056723594665527, "learning_rate": 0.00019939166314465406, "loss": 4.4069, "step": 143350 }, { "epoch": 0.2676003442975141, "grad_norm": 0.7907326221466064, "learning_rate": 0.00019939120038169983, "loss": 4.6325, "step": 143400 }, { "epoch": 0.2676936498568926, "grad_norm": 1.0753734111785889, "learning_rate": 0.00019939073744333765, "loss": 4.6872, "step": 143450 }, { "epoch": 0.26778695541627107, "grad_norm": 1.119105339050293, "learning_rate": 0.00019939027432956828, "loss": 4.64, "step": 143500 }, { "epoch": 0.2678802609756496, "grad_norm": 0.839352011680603, "learning_rate": 0.00019938981104039257, "loss": 4.6258, "step": 143550 }, { "epoch": 0.2679735665350281, "grad_norm": 1.0468686819076538, "learning_rate": 0.00019938934757581132, "loss": 4.6909, "step": 143600 }, { "epoch": 0.26806687209440655, "grad_norm": 1.086549162864685, "learning_rate": 0.0001993888839358254, "loss": 4.6586, "step": 143650 }, { "epoch": 0.2681601776537851, "grad_norm": 0.9070858359336853, "learning_rate": 0.0001993884201204356, "loss": 4.8447, "step": 143700 }, { "epoch": 0.26825348321316356, "grad_norm": 1.0874264240264893, "learning_rate": 0.00019938795612964267, "loss": 4.6228, "step": 143750 }, { "epoch": 0.26834678877254203, "grad_norm": 0.9141344428062439, "learning_rate": 0.0001993874919634475, "loss": 4.6042, "step": 143800 }, { "epoch": 0.2684400943319205, "grad_norm": 1.0117084980010986, "learning_rate": 0.0001993870276218509, "loss": 4.5892, "step": 143850 }, { "epoch": 0.26853339989129904, "grad_norm": 1.1900421380996704, "learning_rate": 0.0001993865631048537, "loss": 4.6517, "step": 143900 }, { "epoch": 0.2686267054506775, "grad_norm": 1.108844518661499, "learning_rate": 0.0001993860984124567, "loss": 4.6209, "step": 143950 }, { "epoch": 0.268720011010056, "grad_norm": 0.9162120223045349, "learning_rate": 0.00019938563354466072, "loss": 4.7472, "step": 144000 }, { "epoch": 0.268720011010056, "eval_loss": 4.846428394317627, "eval_runtime": 231.517, "eval_samples_per_second": 11.265, "eval_steps_per_second": 11.265, "eval_tts_loss": 7.418669227682099, "step": 144000 }, { "epoch": 0.2688133165694345, "grad_norm": 0.8355111479759216, "learning_rate": 0.00019938516850146655, "loss": 4.5845, "step": 144050 }, { "epoch": 0.268906622128813, "grad_norm": 0.6543982625007629, "learning_rate": 0.0001993847032828751, "loss": 4.5369, "step": 144100 }, { "epoch": 0.2689999276881915, "grad_norm": 0.947825014591217, "learning_rate": 0.0001993842378888871, "loss": 4.6122, "step": 144150 }, { "epoch": 0.26909323324756995, "grad_norm": 1.1164684295654297, "learning_rate": 0.0001993837723195034, "loss": 4.6844, "step": 144200 }, { "epoch": 0.2691865388069485, "grad_norm": 0.8201596736907959, "learning_rate": 0.00019938330657472486, "loss": 4.6992, "step": 144250 }, { "epoch": 0.26927984436632696, "grad_norm": 1.0226540565490723, "learning_rate": 0.00019938284065455225, "loss": 4.8901, "step": 144300 }, { "epoch": 0.26937314992570544, "grad_norm": 0.8372309803962708, "learning_rate": 0.0001993823745589864, "loss": 4.776, "step": 144350 }, { "epoch": 0.26946645548508397, "grad_norm": 0.6700859069824219, "learning_rate": 0.0001993819082880282, "loss": 4.5808, "step": 144400 }, { "epoch": 0.26955976104446244, "grad_norm": 0.9853873252868652, "learning_rate": 0.00019938144184167836, "loss": 4.7041, "step": 144450 }, { "epoch": 0.2696530666038409, "grad_norm": 1.0174555778503418, "learning_rate": 0.0001993809752199378, "loss": 4.6368, "step": 144500 }, { "epoch": 0.2697463721632194, "grad_norm": 1.0446439981460571, "learning_rate": 0.00019938050842280728, "loss": 4.8199, "step": 144550 }, { "epoch": 0.2698396777225979, "grad_norm": 1.1562541723251343, "learning_rate": 0.00019938004145028768, "loss": 4.8258, "step": 144600 }, { "epoch": 0.2699329832819764, "grad_norm": 1.0433145761489868, "learning_rate": 0.00019937957430237976, "loss": 4.5957, "step": 144650 }, { "epoch": 0.2700262888413549, "grad_norm": 0.8568206429481506, "learning_rate": 0.0001993791069790844, "loss": 4.6801, "step": 144700 }, { "epoch": 0.2701195944007334, "grad_norm": 1.0583845376968384, "learning_rate": 0.0001993786394804024, "loss": 4.7475, "step": 144750 }, { "epoch": 0.2702128999601119, "grad_norm": 1.1356602907180786, "learning_rate": 0.00019937817180633457, "loss": 4.5852, "step": 144800 }, { "epoch": 0.27030620551949036, "grad_norm": 1.091326117515564, "learning_rate": 0.00019937770395688179, "loss": 4.6252, "step": 144850 }, { "epoch": 0.27039951107886884, "grad_norm": 1.135559320449829, "learning_rate": 0.0001993772359320448, "loss": 4.7506, "step": 144900 }, { "epoch": 0.27049281663824737, "grad_norm": 0.910708487033844, "learning_rate": 0.0001993767677318245, "loss": 4.6904, "step": 144950 }, { "epoch": 0.27058612219762584, "grad_norm": 1.0501621961593628, "learning_rate": 0.0001993762993562217, "loss": 4.8702, "step": 145000 }, { "epoch": 0.2706794277570043, "grad_norm": 0.691247820854187, "learning_rate": 0.0001993758308052372, "loss": 4.666, "step": 145050 }, { "epoch": 0.2707727333163828, "grad_norm": 0.9486430287361145, "learning_rate": 0.00019937536207887183, "loss": 4.7462, "step": 145100 }, { "epoch": 0.27086603887576133, "grad_norm": 0.9459395408630371, "learning_rate": 0.00019937489317712646, "loss": 4.8276, "step": 145150 }, { "epoch": 0.2709593444351398, "grad_norm": 1.2889015674591064, "learning_rate": 0.00019937442410000185, "loss": 4.6273, "step": 145200 }, { "epoch": 0.2710526499945183, "grad_norm": 1.114570140838623, "learning_rate": 0.00019937395484749892, "loss": 4.5248, "step": 145250 }, { "epoch": 0.2711459555538968, "grad_norm": 1.1005711555480957, "learning_rate": 0.00019937348541961843, "loss": 4.6502, "step": 145300 }, { "epoch": 0.2712392611132753, "grad_norm": 0.806015133857727, "learning_rate": 0.00019937301581636122, "loss": 4.6222, "step": 145350 }, { "epoch": 0.27133256667265376, "grad_norm": 0.9318636059761047, "learning_rate": 0.00019937254603772812, "loss": 4.5528, "step": 145400 }, { "epoch": 0.27142587223203224, "grad_norm": 0.697272539138794, "learning_rate": 0.00019937207608371996, "loss": 4.7631, "step": 145450 }, { "epoch": 0.27151917779141077, "grad_norm": 0.9297815561294556, "learning_rate": 0.00019937160595433757, "loss": 4.5448, "step": 145500 }, { "epoch": 0.27161248335078925, "grad_norm": 1.144652247428894, "learning_rate": 0.00019937113564958176, "loss": 4.6709, "step": 145550 }, { "epoch": 0.2717057889101677, "grad_norm": 0.9392854571342468, "learning_rate": 0.0001993706651694534, "loss": 4.7134, "step": 145600 }, { "epoch": 0.27179909446954625, "grad_norm": 0.9400250315666199, "learning_rate": 0.00019937019451395332, "loss": 4.5493, "step": 145650 }, { "epoch": 0.27189240002892473, "grad_norm": 1.1566085815429688, "learning_rate": 0.0001993697236830823, "loss": 4.939, "step": 145700 }, { "epoch": 0.2719857055883032, "grad_norm": 0.7615442872047424, "learning_rate": 0.00019936925267684122, "loss": 4.7826, "step": 145750 }, { "epoch": 0.2720790111476817, "grad_norm": 0.8968701362609863, "learning_rate": 0.0001993687814952309, "loss": 4.5258, "step": 145800 }, { "epoch": 0.2721723167070602, "grad_norm": 1.0293611288070679, "learning_rate": 0.00019936831013825214, "loss": 4.6597, "step": 145850 }, { "epoch": 0.2722656222664387, "grad_norm": 0.9446884989738464, "learning_rate": 0.00019936783860590582, "loss": 4.6033, "step": 145900 }, { "epoch": 0.27235892782581717, "grad_norm": 1.0046056509017944, "learning_rate": 0.00019936736689819275, "loss": 4.5402, "step": 145950 }, { "epoch": 0.2724522333851957, "grad_norm": 0.7611010074615479, "learning_rate": 0.00019936689501511375, "loss": 4.6937, "step": 146000 }, { "epoch": 0.2725455389445742, "grad_norm": 0.9381779432296753, "learning_rate": 0.00019936642295666966, "loss": 4.5509, "step": 146050 }, { "epoch": 0.27263884450395265, "grad_norm": 0.862014889717102, "learning_rate": 0.00019936595072286134, "loss": 4.6046, "step": 146100 }, { "epoch": 0.2727321500633311, "grad_norm": 1.1331309080123901, "learning_rate": 0.00019936547831368956, "loss": 4.4339, "step": 146150 }, { "epoch": 0.27282545562270966, "grad_norm": 0.9985886812210083, "learning_rate": 0.00019936500572915523, "loss": 4.7718, "step": 146200 }, { "epoch": 0.27291876118208813, "grad_norm": 1.0175446271896362, "learning_rate": 0.00019936453296925914, "loss": 4.5348, "step": 146250 }, { "epoch": 0.2730120667414666, "grad_norm": 1.1573448181152344, "learning_rate": 0.00019936406003400215, "loss": 4.5544, "step": 146300 }, { "epoch": 0.27310537230084514, "grad_norm": 0.7975154519081116, "learning_rate": 0.00019936358692338506, "loss": 4.6486, "step": 146350 }, { "epoch": 0.2731986778602236, "grad_norm": 0.9577890038490295, "learning_rate": 0.0001993631136374087, "loss": 4.661, "step": 146400 }, { "epoch": 0.2732919834196021, "grad_norm": 1.0218697786331177, "learning_rate": 0.00019936264017607398, "loss": 4.6003, "step": 146450 }, { "epoch": 0.27338528897898057, "grad_norm": 1.3306013345718384, "learning_rate": 0.00019936216653938162, "loss": 4.7019, "step": 146500 }, { "epoch": 0.2734785945383591, "grad_norm": 1.0977168083190918, "learning_rate": 0.00019936169272733256, "loss": 4.7338, "step": 146550 }, { "epoch": 0.2735719000977376, "grad_norm": 1.1367696523666382, "learning_rate": 0.0001993612187399276, "loss": 4.9661, "step": 146600 }, { "epoch": 0.27366520565711605, "grad_norm": 1.0126125812530518, "learning_rate": 0.00019936074457716757, "loss": 4.7346, "step": 146650 }, { "epoch": 0.2737585112164946, "grad_norm": 0.9203944802284241, "learning_rate": 0.0001993602702390533, "loss": 4.7422, "step": 146700 }, { "epoch": 0.27385181677587306, "grad_norm": 0.8814552426338196, "learning_rate": 0.00019935979572558563, "loss": 4.5732, "step": 146750 }, { "epoch": 0.27394512233525153, "grad_norm": 0.9684024453163147, "learning_rate": 0.00019935932103676542, "loss": 4.6788, "step": 146800 }, { "epoch": 0.27403842789463, "grad_norm": 0.8495442271232605, "learning_rate": 0.00019935884617259349, "loss": 4.4585, "step": 146850 }, { "epoch": 0.27413173345400854, "grad_norm": 0.5965443849563599, "learning_rate": 0.00019935837113307066, "loss": 4.4931, "step": 146900 }, { "epoch": 0.274225039013387, "grad_norm": 0.9253427386283875, "learning_rate": 0.00019935789591819778, "loss": 4.7332, "step": 146950 }, { "epoch": 0.2743183445727655, "grad_norm": 0.959456205368042, "learning_rate": 0.0001993574205279757, "loss": 4.6347, "step": 147000 }, { "epoch": 0.2743183445727655, "eval_loss": 4.84540319442749, "eval_runtime": 228.6502, "eval_samples_per_second": 11.406, "eval_steps_per_second": 11.406, "eval_tts_loss": 7.421910372788395, "step": 147000 }, { "epoch": 0.274411650132144, "grad_norm": 0.8804351687431335, "learning_rate": 0.0001993569449624053, "loss": 4.5913, "step": 147050 }, { "epoch": 0.2745049556915225, "grad_norm": 1.257071852684021, "learning_rate": 0.00019935646922148733, "loss": 4.6649, "step": 147100 }, { "epoch": 0.274598261250901, "grad_norm": 1.0313574075698853, "learning_rate": 0.0001993559933052227, "loss": 4.6562, "step": 147150 }, { "epoch": 0.27469156681027945, "grad_norm": 0.9633307456970215, "learning_rate": 0.0001993555172136122, "loss": 4.6321, "step": 147200 }, { "epoch": 0.274784872369658, "grad_norm": 0.7918700575828552, "learning_rate": 0.0001993550409466567, "loss": 4.7728, "step": 147250 }, { "epoch": 0.27487817792903646, "grad_norm": 0.6969190239906311, "learning_rate": 0.00019935456450435703, "loss": 4.7156, "step": 147300 }, { "epoch": 0.27497148348841494, "grad_norm": 1.4013220071792603, "learning_rate": 0.00019935408788671405, "loss": 4.5602, "step": 147350 }, { "epoch": 0.27506478904779347, "grad_norm": 1.0779790878295898, "learning_rate": 0.00019935361109372859, "loss": 4.7721, "step": 147400 }, { "epoch": 0.27515809460717194, "grad_norm": 1.0877084732055664, "learning_rate": 0.00019935313412540147, "loss": 4.7129, "step": 147450 }, { "epoch": 0.2752514001665504, "grad_norm": 1.1705236434936523, "learning_rate": 0.00019935265698173358, "loss": 4.7317, "step": 147500 }, { "epoch": 0.2753447057259289, "grad_norm": 1.0509397983551025, "learning_rate": 0.0001993521796627257, "loss": 4.7795, "step": 147550 }, { "epoch": 0.2754380112853074, "grad_norm": 1.1017162799835205, "learning_rate": 0.00019935170216837873, "loss": 4.7745, "step": 147600 }, { "epoch": 0.2755313168446859, "grad_norm": 0.9443593621253967, "learning_rate": 0.0001993512244986935, "loss": 4.6332, "step": 147650 }, { "epoch": 0.2756246224040644, "grad_norm": 1.0674967765808105, "learning_rate": 0.00019935074665367077, "loss": 4.721, "step": 147700 }, { "epoch": 0.27571792796344285, "grad_norm": 0.976577877998352, "learning_rate": 0.00019935026863331153, "loss": 4.5112, "step": 147750 }, { "epoch": 0.2758112335228214, "grad_norm": 0.9510547518730164, "learning_rate": 0.0001993497904376165, "loss": 4.8486, "step": 147800 }, { "epoch": 0.27590453908219986, "grad_norm": 1.0854049921035767, "learning_rate": 0.00019934931206658658, "loss": 4.4141, "step": 147850 }, { "epoch": 0.27599784464157834, "grad_norm": 1.5401407480239868, "learning_rate": 0.0001993488335202226, "loss": 4.741, "step": 147900 }, { "epoch": 0.27609115020095687, "grad_norm": 1.0566352605819702, "learning_rate": 0.00019934835479852546, "loss": 4.9034, "step": 147950 }, { "epoch": 0.27618445576033535, "grad_norm": 1.0547597408294678, "learning_rate": 0.0001993478759014959, "loss": 4.6906, "step": 148000 }, { "epoch": 0.2762777613197138, "grad_norm": 0.8163520097732544, "learning_rate": 0.00019934739682913485, "loss": 4.8077, "step": 148050 }, { "epoch": 0.2763710668790923, "grad_norm": 1.1130653619766235, "learning_rate": 0.00019934691758144308, "loss": 4.7433, "step": 148100 }, { "epoch": 0.27646437243847083, "grad_norm": 0.769314706325531, "learning_rate": 0.00019934643815842153, "loss": 4.6217, "step": 148150 }, { "epoch": 0.2765576779978493, "grad_norm": 1.1348406076431274, "learning_rate": 0.000199345958560071, "loss": 4.5468, "step": 148200 }, { "epoch": 0.2766509835572278, "grad_norm": 0.9563210010528564, "learning_rate": 0.00019934547878639228, "loss": 4.6116, "step": 148250 }, { "epoch": 0.2767442891166063, "grad_norm": 0.9107294082641602, "learning_rate": 0.00019934499883738632, "loss": 4.8486, "step": 148300 }, { "epoch": 0.2768375946759848, "grad_norm": 1.0701485872268677, "learning_rate": 0.0001993445187130539, "loss": 4.7653, "step": 148350 }, { "epoch": 0.27693090023536326, "grad_norm": 1.1094465255737305, "learning_rate": 0.00019934403841339587, "loss": 4.451, "step": 148400 }, { "epoch": 0.27702420579474174, "grad_norm": 1.283995270729065, "learning_rate": 0.00019934355793841313, "loss": 4.8218, "step": 148450 }, { "epoch": 0.27711751135412027, "grad_norm": 0.9863646030426025, "learning_rate": 0.00019934307728810646, "loss": 4.831, "step": 148500 }, { "epoch": 0.27721081691349875, "grad_norm": 1.0133435726165771, "learning_rate": 0.00019934259646247674, "loss": 4.4905, "step": 148550 }, { "epoch": 0.2773041224728772, "grad_norm": 0.9534834623336792, "learning_rate": 0.00019934211546152484, "loss": 4.7916, "step": 148600 }, { "epoch": 0.27739742803225576, "grad_norm": 1.2967839241027832, "learning_rate": 0.00019934163428525156, "loss": 4.7967, "step": 148650 }, { "epoch": 0.27749073359163423, "grad_norm": 0.869767427444458, "learning_rate": 0.00019934115293365778, "loss": 4.7234, "step": 148700 }, { "epoch": 0.2775840391510127, "grad_norm": 0.9812904000282288, "learning_rate": 0.00019934067140674436, "loss": 4.4395, "step": 148750 }, { "epoch": 0.2776773447103912, "grad_norm": 0.9434531331062317, "learning_rate": 0.0001993401897045121, "loss": 4.6072, "step": 148800 }, { "epoch": 0.2777706502697697, "grad_norm": 0.6807860732078552, "learning_rate": 0.00019933970782696193, "loss": 4.6445, "step": 148850 }, { "epoch": 0.2778639558291482, "grad_norm": 0.8833329677581787, "learning_rate": 0.00019933922577409463, "loss": 4.4624, "step": 148900 }, { "epoch": 0.27795726138852667, "grad_norm": 1.2914021015167236, "learning_rate": 0.0001993387435459111, "loss": 4.8027, "step": 148950 }, { "epoch": 0.2780505669479052, "grad_norm": 1.0718110799789429, "learning_rate": 0.00019933826114241213, "loss": 4.6699, "step": 149000 }, { "epoch": 0.2781438725072837, "grad_norm": 0.8196322321891785, "learning_rate": 0.00019933777856359863, "loss": 4.5286, "step": 149050 }, { "epoch": 0.27823717806666215, "grad_norm": 1.0307577848434448, "learning_rate": 0.00019933729580947143, "loss": 4.6294, "step": 149100 }, { "epoch": 0.2783304836260406, "grad_norm": 0.874259889125824, "learning_rate": 0.00019933681288003137, "loss": 4.6817, "step": 149150 }, { "epoch": 0.27842378918541916, "grad_norm": 0.8164685964584351, "learning_rate": 0.0001993363297752793, "loss": 4.6156, "step": 149200 }, { "epoch": 0.27851709474479763, "grad_norm": 1.4516046047210693, "learning_rate": 0.0001993358464952161, "loss": 4.8161, "step": 149250 }, { "epoch": 0.2786104003041761, "grad_norm": 1.4525372982025146, "learning_rate": 0.00019933536303984263, "loss": 4.6891, "step": 149300 }, { "epoch": 0.27870370586355464, "grad_norm": 1.0817906856536865, "learning_rate": 0.0001993348794091597, "loss": 4.9178, "step": 149350 }, { "epoch": 0.2787970114229331, "grad_norm": 0.8734576106071472, "learning_rate": 0.0001993343956031682, "loss": 4.7184, "step": 149400 }, { "epoch": 0.2788903169823116, "grad_norm": 0.8996809720993042, "learning_rate": 0.00019933391162186897, "loss": 4.5792, "step": 149450 }, { "epoch": 0.27898362254169007, "grad_norm": 1.0781534910202026, "learning_rate": 0.00019933342746526286, "loss": 4.6787, "step": 149500 }, { "epoch": 0.2790769281010686, "grad_norm": 0.9181458353996277, "learning_rate": 0.00019933294313335072, "loss": 4.6463, "step": 149550 }, { "epoch": 0.2791702336604471, "grad_norm": 1.0423184633255005, "learning_rate": 0.00019933245862613344, "loss": 4.5832, "step": 149600 }, { "epoch": 0.27926353921982555, "grad_norm": 0.9878393411636353, "learning_rate": 0.00019933197394361181, "loss": 4.5294, "step": 149650 }, { "epoch": 0.2793568447792041, "grad_norm": 0.9291845560073853, "learning_rate": 0.00019933148908578678, "loss": 4.7486, "step": 149700 }, { "epoch": 0.27945015033858256, "grad_norm": 1.0207691192626953, "learning_rate": 0.0001993310040526591, "loss": 4.7574, "step": 149750 }, { "epoch": 0.27954345589796104, "grad_norm": 1.0741710662841797, "learning_rate": 0.00019933051884422968, "loss": 4.6548, "step": 149800 }, { "epoch": 0.2796367614573395, "grad_norm": 1.0750080347061157, "learning_rate": 0.0001993300334604994, "loss": 4.6506, "step": 149850 }, { "epoch": 0.27973006701671804, "grad_norm": 1.0452051162719727, "learning_rate": 0.00019932954790146908, "loss": 4.7486, "step": 149900 }, { "epoch": 0.2798233725760965, "grad_norm": 0.8522201776504517, "learning_rate": 0.0001993290621671396, "loss": 4.6639, "step": 149950 }, { "epoch": 0.279916678135475, "grad_norm": 0.8773214817047119, "learning_rate": 0.00019932857625751176, "loss": 4.7012, "step": 150000 }, { "epoch": 0.279916678135475, "eval_loss": 4.83031702041626, "eval_runtime": 229.2208, "eval_samples_per_second": 11.378, "eval_steps_per_second": 11.378, "eval_tts_loss": 7.386615539569568, "step": 150000 }, { "epoch": 0.2800099836948535, "grad_norm": 0.9581308364868164, "learning_rate": 0.00019932809017258647, "loss": 4.8594, "step": 150050 }, { "epoch": 0.280103289254232, "grad_norm": 0.9009203314781189, "learning_rate": 0.00019932760391236457, "loss": 4.6758, "step": 150100 }, { "epoch": 0.2801965948136105, "grad_norm": 0.9050132632255554, "learning_rate": 0.00019932711747684695, "loss": 4.5833, "step": 150150 }, { "epoch": 0.28028990037298895, "grad_norm": 1.0961129665374756, "learning_rate": 0.00019932663086603446, "loss": 4.7339, "step": 150200 }, { "epoch": 0.2803832059323675, "grad_norm": 1.4378759860992432, "learning_rate": 0.00019932614407992792, "loss": 4.8097, "step": 150250 }, { "epoch": 0.28047651149174596, "grad_norm": 0.941024899482727, "learning_rate": 0.0001993256571185282, "loss": 4.5834, "step": 150300 }, { "epoch": 0.28056981705112444, "grad_norm": 0.8255406618118286, "learning_rate": 0.0001993251699818362, "loss": 4.5948, "step": 150350 }, { "epoch": 0.2806631226105029, "grad_norm": 0.9121527671813965, "learning_rate": 0.00019932468266985274, "loss": 4.8421, "step": 150400 }, { "epoch": 0.28075642816988144, "grad_norm": 1.2012547254562378, "learning_rate": 0.0001993241951825787, "loss": 4.7063, "step": 150450 }, { "epoch": 0.2808497337292599, "grad_norm": 1.0221003293991089, "learning_rate": 0.0001993237075200149, "loss": 4.5613, "step": 150500 }, { "epoch": 0.2809430392886384, "grad_norm": 0.6838630437850952, "learning_rate": 0.00019932321968216225, "loss": 4.53, "step": 150550 }, { "epoch": 0.28103634484801693, "grad_norm": 0.9109449982643127, "learning_rate": 0.0001993227316690216, "loss": 4.7778, "step": 150600 }, { "epoch": 0.2811296504073954, "grad_norm": 1.0373647212982178, "learning_rate": 0.00019932224348059383, "loss": 4.7198, "step": 150650 }, { "epoch": 0.2812229559667739, "grad_norm": 1.202540397644043, "learning_rate": 0.00019932175511687975, "loss": 4.5706, "step": 150700 }, { "epoch": 0.28131626152615236, "grad_norm": 1.1028730869293213, "learning_rate": 0.00019932126657788024, "loss": 4.6268, "step": 150750 }, { "epoch": 0.2814095670855309, "grad_norm": 0.88832688331604, "learning_rate": 0.00019932077786359616, "loss": 4.7166, "step": 150800 }, { "epoch": 0.28150287264490936, "grad_norm": 1.0212421417236328, "learning_rate": 0.00019932028897402844, "loss": 4.7143, "step": 150850 }, { "epoch": 0.28159617820428784, "grad_norm": 0.8306617736816406, "learning_rate": 0.00019931979990917782, "loss": 4.7177, "step": 150900 }, { "epoch": 0.28168948376366637, "grad_norm": 1.1900157928466797, "learning_rate": 0.00019931931066904528, "loss": 4.6889, "step": 150950 }, { "epoch": 0.28178278932304485, "grad_norm": 0.8949201703071594, "learning_rate": 0.00019931882125363162, "loss": 4.72, "step": 151000 }, { "epoch": 0.2818760948824233, "grad_norm": 1.0268856287002563, "learning_rate": 0.00019931833166293768, "loss": 4.7837, "step": 151050 }, { "epoch": 0.2819694004418018, "grad_norm": 1.0028334856033325, "learning_rate": 0.00019931784189696438, "loss": 4.4346, "step": 151100 }, { "epoch": 0.28206270600118033, "grad_norm": 0.6958338618278503, "learning_rate": 0.00019931735195571258, "loss": 4.599, "step": 151150 }, { "epoch": 0.2821560115605588, "grad_norm": 1.0011975765228271, "learning_rate": 0.00019931686183918314, "loss": 4.6019, "step": 151200 }, { "epoch": 0.2822493171199373, "grad_norm": 1.0134965181350708, "learning_rate": 0.00019931637154737688, "loss": 4.5689, "step": 151250 }, { "epoch": 0.2823426226793158, "grad_norm": 1.1758054494857788, "learning_rate": 0.00019931588108029473, "loss": 4.863, "step": 151300 }, { "epoch": 0.2824359282386943, "grad_norm": 0.6513984203338623, "learning_rate": 0.0001993153904379375, "loss": 4.8059, "step": 151350 }, { "epoch": 0.28252923379807277, "grad_norm": 0.9191519618034363, "learning_rate": 0.0001993148996203061, "loss": 4.5621, "step": 151400 }, { "epoch": 0.28262253935745124, "grad_norm": 1.069206953048706, "learning_rate": 0.00019931440862740136, "loss": 4.54, "step": 151450 }, { "epoch": 0.2827158449168298, "grad_norm": 0.8649135828018188, "learning_rate": 0.00019931391745922416, "loss": 4.4516, "step": 151500 }, { "epoch": 0.28280915047620825, "grad_norm": 0.8605431914329529, "learning_rate": 0.0001993134261157754, "loss": 4.6004, "step": 151550 }, { "epoch": 0.2829024560355867, "grad_norm": 1.1754326820373535, "learning_rate": 0.00019931293459705588, "loss": 4.7162, "step": 151600 }, { "epoch": 0.28299576159496526, "grad_norm": 0.9741180539131165, "learning_rate": 0.0001993124429030665, "loss": 4.8181, "step": 151650 }, { "epoch": 0.28308906715434373, "grad_norm": 1.0868324041366577, "learning_rate": 0.00019931195103380819, "loss": 4.602, "step": 151700 }, { "epoch": 0.2831823727137222, "grad_norm": 0.9874230027198792, "learning_rate": 0.00019931145898928173, "loss": 4.6277, "step": 151750 }, { "epoch": 0.2832756782731007, "grad_norm": 1.13186776638031, "learning_rate": 0.00019931096676948802, "loss": 4.5465, "step": 151800 }, { "epoch": 0.2833689838324792, "grad_norm": 0.8239442110061646, "learning_rate": 0.00019931047437442794, "loss": 4.7412, "step": 151850 }, { "epoch": 0.2834622893918577, "grad_norm": 1.0195013284683228, "learning_rate": 0.00019930998180410232, "loss": 4.4785, "step": 151900 }, { "epoch": 0.28355559495123617, "grad_norm": 1.3029369115829468, "learning_rate": 0.00019930948905851206, "loss": 4.8227, "step": 151950 }, { "epoch": 0.2836489005106147, "grad_norm": 0.8565980792045593, "learning_rate": 0.00019930899613765805, "loss": 4.5811, "step": 152000 }, { "epoch": 0.2837422060699932, "grad_norm": 0.9807513356208801, "learning_rate": 0.0001993085030415411, "loss": 4.733, "step": 152050 }, { "epoch": 0.28383551162937165, "grad_norm": 1.0454181432724, "learning_rate": 0.00019930800977016215, "loss": 4.703, "step": 152100 }, { "epoch": 0.2839288171887501, "grad_norm": 1.0807571411132812, "learning_rate": 0.00019930751632352202, "loss": 4.4999, "step": 152150 }, { "epoch": 0.28402212274812866, "grad_norm": 1.1262049674987793, "learning_rate": 0.0001993070227016216, "loss": 4.6597, "step": 152200 }, { "epoch": 0.28411542830750713, "grad_norm": 0.8939054012298584, "learning_rate": 0.00019930652890446176, "loss": 4.6779, "step": 152250 }, { "epoch": 0.2842087338668856, "grad_norm": 0.9293799996376038, "learning_rate": 0.00019930603493204337, "loss": 4.551, "step": 152300 }, { "epoch": 0.28430203942626414, "grad_norm": 1.1074103116989136, "learning_rate": 0.0001993055407843673, "loss": 4.805, "step": 152350 }, { "epoch": 0.2843953449856426, "grad_norm": 1.147536039352417, "learning_rate": 0.0001993050464614344, "loss": 4.7777, "step": 152400 }, { "epoch": 0.2844886505450211, "grad_norm": 0.8999912142753601, "learning_rate": 0.00019930455196324562, "loss": 4.7226, "step": 152450 }, { "epoch": 0.28458195610439957, "grad_norm": 0.8882901072502136, "learning_rate": 0.00019930405728980176, "loss": 4.5275, "step": 152500 }, { "epoch": 0.2846752616637781, "grad_norm": 0.958745539188385, "learning_rate": 0.00019930356244110369, "loss": 4.5713, "step": 152550 }, { "epoch": 0.2847685672231566, "grad_norm": 0.8023204803466797, "learning_rate": 0.00019930306741715233, "loss": 4.6426, "step": 152600 }, { "epoch": 0.28486187278253505, "grad_norm": 0.7979447245597839, "learning_rate": 0.0001993025722179485, "loss": 4.81, "step": 152650 }, { "epoch": 0.2849551783419136, "grad_norm": 0.8460389375686646, "learning_rate": 0.00019930207684349313, "loss": 4.7702, "step": 152700 }, { "epoch": 0.28504848390129206, "grad_norm": 0.905491054058075, "learning_rate": 0.00019930158129378707, "loss": 4.6398, "step": 152750 }, { "epoch": 0.28514178946067054, "grad_norm": 0.5452645421028137, "learning_rate": 0.00019930108556883116, "loss": 4.4575, "step": 152800 }, { "epoch": 0.285235095020049, "grad_norm": 1.016485333442688, "learning_rate": 0.00019930058966862634, "loss": 4.6945, "step": 152850 }, { "epoch": 0.28532840057942754, "grad_norm": 0.6815197467803955, "learning_rate": 0.00019930009359317343, "loss": 4.6492, "step": 152900 }, { "epoch": 0.285421706138806, "grad_norm": 0.9308417439460754, "learning_rate": 0.00019929959734247331, "loss": 4.6306, "step": 152950 }, { "epoch": 0.2855150116981845, "grad_norm": 0.9671070575714111, "learning_rate": 0.0001992991009165269, "loss": 4.6504, "step": 153000 }, { "epoch": 0.2855150116981845, "eval_loss": 4.835600852966309, "eval_runtime": 230.5993, "eval_samples_per_second": 11.31, "eval_steps_per_second": 11.31, "eval_tts_loss": 7.43258083978685, "step": 153000 }, { "epoch": 0.28560831725756297, "grad_norm": 0.8776199221611023, "learning_rate": 0.00019929860431533504, "loss": 4.5989, "step": 153050 }, { "epoch": 0.2857016228169415, "grad_norm": 0.8226851224899292, "learning_rate": 0.00019929810753889865, "loss": 4.3491, "step": 153100 }, { "epoch": 0.28579492837632, "grad_norm": 1.3178420066833496, "learning_rate": 0.00019929761058721853, "loss": 4.5217, "step": 153150 }, { "epoch": 0.28588823393569845, "grad_norm": 1.0944714546203613, "learning_rate": 0.0001992971134602956, "loss": 4.7149, "step": 153200 }, { "epoch": 0.285981539495077, "grad_norm": 0.8839026689529419, "learning_rate": 0.00019929661615813077, "loss": 4.5549, "step": 153250 }, { "epoch": 0.28607484505445546, "grad_norm": 0.9436131119728088, "learning_rate": 0.00019929611868072487, "loss": 4.7396, "step": 153300 }, { "epoch": 0.28616815061383394, "grad_norm": 0.8555207252502441, "learning_rate": 0.00019929562102807878, "loss": 4.6732, "step": 153350 }, { "epoch": 0.2862614561732124, "grad_norm": 0.8614400625228882, "learning_rate": 0.0001992951232001934, "loss": 4.5358, "step": 153400 }, { "epoch": 0.28635476173259095, "grad_norm": 0.7893537282943726, "learning_rate": 0.00019929462519706958, "loss": 4.7613, "step": 153450 }, { "epoch": 0.2864480672919694, "grad_norm": 1.0133888721466064, "learning_rate": 0.00019929412701870822, "loss": 4.7695, "step": 153500 }, { "epoch": 0.2865413728513479, "grad_norm": 1.0339868068695068, "learning_rate": 0.00019929362866511022, "loss": 4.6753, "step": 153550 }, { "epoch": 0.28663467841072643, "grad_norm": 0.8841448426246643, "learning_rate": 0.00019929313013627642, "loss": 4.5938, "step": 153600 }, { "epoch": 0.2867279839701049, "grad_norm": 0.8612989783287048, "learning_rate": 0.00019929263143220774, "loss": 4.4805, "step": 153650 }, { "epoch": 0.2868212895294834, "grad_norm": 0.9694012403488159, "learning_rate": 0.000199292132552905, "loss": 4.8494, "step": 153700 }, { "epoch": 0.28691459508886186, "grad_norm": 1.2090984582901, "learning_rate": 0.00019929163349836915, "loss": 4.7134, "step": 153750 }, { "epoch": 0.2870079006482404, "grad_norm": 1.0495941638946533, "learning_rate": 0.00019929113426860103, "loss": 4.7221, "step": 153800 }, { "epoch": 0.28710120620761886, "grad_norm": 0.8493571281433105, "learning_rate": 0.00019929063486360152, "loss": 4.4686, "step": 153850 }, { "epoch": 0.28719451176699734, "grad_norm": 1.2186065912246704, "learning_rate": 0.0001992901352833715, "loss": 4.7176, "step": 153900 }, { "epoch": 0.28728781732637587, "grad_norm": 1.0181117057800293, "learning_rate": 0.00019928963552791188, "loss": 4.484, "step": 153950 }, { "epoch": 0.28738112288575435, "grad_norm": 1.2028471231460571, "learning_rate": 0.0001992891355972235, "loss": 4.8875, "step": 154000 }, { "epoch": 0.2874744284451328, "grad_norm": 1.1655298471450806, "learning_rate": 0.00019928863549130733, "loss": 4.5263, "step": 154050 }, { "epoch": 0.2875677340045113, "grad_norm": 1.0849800109863281, "learning_rate": 0.00019928813521016413, "loss": 4.4706, "step": 154100 }, { "epoch": 0.28766103956388983, "grad_norm": 0.9207818508148193, "learning_rate": 0.00019928763475379486, "loss": 4.605, "step": 154150 }, { "epoch": 0.2877543451232683, "grad_norm": 0.8862490653991699, "learning_rate": 0.00019928713412220036, "loss": 4.5169, "step": 154200 }, { "epoch": 0.2878476506826468, "grad_norm": 1.0077502727508545, "learning_rate": 0.00019928663331538154, "loss": 4.7856, "step": 154250 }, { "epoch": 0.2879409562420253, "grad_norm": 1.1006149053573608, "learning_rate": 0.00019928613233333934, "loss": 4.7136, "step": 154300 }, { "epoch": 0.2880342618014038, "grad_norm": 1.0672188997268677, "learning_rate": 0.00019928563117607454, "loss": 4.6588, "step": 154350 }, { "epoch": 0.28812756736078227, "grad_norm": 1.0344702005386353, "learning_rate": 0.00019928512984358808, "loss": 4.5211, "step": 154400 }, { "epoch": 0.28822087292016074, "grad_norm": 0.9514538645744324, "learning_rate": 0.00019928462833588081, "loss": 4.586, "step": 154450 }, { "epoch": 0.2883141784795393, "grad_norm": 0.9031883478164673, "learning_rate": 0.00019928412665295368, "loss": 4.6923, "step": 154500 }, { "epoch": 0.28840748403891775, "grad_norm": 0.9801226258277893, "learning_rate": 0.00019928362479480752, "loss": 4.6889, "step": 154550 }, { "epoch": 0.2885007895982962, "grad_norm": 0.9183834791183472, "learning_rate": 0.00019928312276144326, "loss": 4.5817, "step": 154600 }, { "epoch": 0.28859409515767476, "grad_norm": 0.8800172209739685, "learning_rate": 0.0001992826205528617, "loss": 4.6275, "step": 154650 }, { "epoch": 0.28868740071705323, "grad_norm": 1.0279831886291504, "learning_rate": 0.00019928211816906384, "loss": 4.6246, "step": 154700 }, { "epoch": 0.2887807062764317, "grad_norm": 0.9106327295303345, "learning_rate": 0.00019928161561005045, "loss": 4.5679, "step": 154750 }, { "epoch": 0.2888740118358102, "grad_norm": 0.9070572257041931, "learning_rate": 0.0001992811128758225, "loss": 4.6825, "step": 154800 }, { "epoch": 0.2889673173951887, "grad_norm": 0.9388306140899658, "learning_rate": 0.00019928060996638093, "loss": 4.4394, "step": 154850 }, { "epoch": 0.2890606229545672, "grad_norm": 0.9213298559188843, "learning_rate": 0.00019928010688172645, "loss": 4.6287, "step": 154900 }, { "epoch": 0.28915392851394567, "grad_norm": 0.9049386382102966, "learning_rate": 0.00019927960362186008, "loss": 4.7666, "step": 154950 }, { "epoch": 0.2892472340733242, "grad_norm": 0.9011629223823547, "learning_rate": 0.0001992791001867827, "loss": 4.5634, "step": 155000 }, { "epoch": 0.2893405396327027, "grad_norm": 1.0195951461791992, "learning_rate": 0.00019927859657649518, "loss": 4.6655, "step": 155050 }, { "epoch": 0.28943384519208115, "grad_norm": 0.7028415203094482, "learning_rate": 0.00019927809279099838, "loss": 4.4955, "step": 155100 }, { "epoch": 0.2895271507514596, "grad_norm": 0.9944257140159607, "learning_rate": 0.0001992775888302932, "loss": 4.9458, "step": 155150 }, { "epoch": 0.28962045631083816, "grad_norm": 1.0102462768554688, "learning_rate": 0.0001992770846943806, "loss": 4.748, "step": 155200 }, { "epoch": 0.28971376187021664, "grad_norm": 0.8883141875267029, "learning_rate": 0.00019927658038326138, "loss": 4.4985, "step": 155250 }, { "epoch": 0.2898070674295951, "grad_norm": 0.8198468089103699, "learning_rate": 0.0001992760758969365, "loss": 4.6353, "step": 155300 }, { "epoch": 0.28990037298897364, "grad_norm": 0.7499061226844788, "learning_rate": 0.00019927557123540677, "loss": 4.5579, "step": 155350 }, { "epoch": 0.2899936785483521, "grad_norm": 0.7764875888824463, "learning_rate": 0.0001992750663986731, "loss": 4.4778, "step": 155400 }, { "epoch": 0.2900869841077306, "grad_norm": 0.8831614255905151, "learning_rate": 0.00019927456138673646, "loss": 4.403, "step": 155450 }, { "epoch": 0.29018028966710907, "grad_norm": 1.0157040357589722, "learning_rate": 0.00019927405619959767, "loss": 4.6915, "step": 155500 }, { "epoch": 0.2902735952264876, "grad_norm": 1.1390964984893799, "learning_rate": 0.00019927355083725762, "loss": 4.5186, "step": 155550 }, { "epoch": 0.2903669007858661, "grad_norm": 1.154977560043335, "learning_rate": 0.00019927304529971727, "loss": 4.7172, "step": 155600 }, { "epoch": 0.29046020634524455, "grad_norm": 1.095638632774353, "learning_rate": 0.00019927253958697743, "loss": 4.8124, "step": 155650 }, { "epoch": 0.29055351190462303, "grad_norm": 0.9511516094207764, "learning_rate": 0.00019927203369903901, "loss": 4.4195, "step": 155700 }, { "epoch": 0.29064681746400156, "grad_norm": 1.0165811777114868, "learning_rate": 0.00019927152763590296, "loss": 4.6677, "step": 155750 }, { "epoch": 0.29074012302338004, "grad_norm": 0.8672209978103638, "learning_rate": 0.0001992710213975701, "loss": 4.7442, "step": 155800 }, { "epoch": 0.2908334285827585, "grad_norm": 1.1338589191436768, "learning_rate": 0.00019927051498404134, "loss": 4.6876, "step": 155850 }, { "epoch": 0.29092673414213704, "grad_norm": 1.0696067810058594, "learning_rate": 0.00019927000839531763, "loss": 4.5303, "step": 155900 }, { "epoch": 0.2910200397015155, "grad_norm": 0.9519825577735901, "learning_rate": 0.00019926950163139977, "loss": 4.581, "step": 155950 }, { "epoch": 0.291113345260894, "grad_norm": 1.0160168409347534, "learning_rate": 0.00019926899469228877, "loss": 4.589, "step": 156000 }, { "epoch": 0.291113345260894, "eval_loss": 4.831135272979736, "eval_runtime": 231.6301, "eval_samples_per_second": 11.259, "eval_steps_per_second": 11.259, "eval_tts_loss": 7.382750091658608, "step": 156000 }, { "epoch": 0.2912066508202725, "grad_norm": 1.5486959218978882, "learning_rate": 0.00019926848757798542, "loss": 4.5012, "step": 156050 }, { "epoch": 0.291299956379651, "grad_norm": 1.007653832435608, "learning_rate": 0.00019926798028849064, "loss": 4.5957, "step": 156100 }, { "epoch": 0.2913932619390295, "grad_norm": 0.7587629556655884, "learning_rate": 0.00019926747282380538, "loss": 4.5975, "step": 156150 }, { "epoch": 0.29148656749840796, "grad_norm": 0.8650146126747131, "learning_rate": 0.0001992669651839305, "loss": 4.5823, "step": 156200 }, { "epoch": 0.2915798730577865, "grad_norm": 0.8940017819404602, "learning_rate": 0.00019926645736886687, "loss": 4.5223, "step": 156250 }, { "epoch": 0.29167317861716496, "grad_norm": 1.1759480237960815, "learning_rate": 0.0001992659493786154, "loss": 4.5113, "step": 156300 }, { "epoch": 0.29176648417654344, "grad_norm": 0.9425550699234009, "learning_rate": 0.000199265441213177, "loss": 4.62, "step": 156350 }, { "epoch": 0.2918597897359219, "grad_norm": 1.2574522495269775, "learning_rate": 0.00019926493287255258, "loss": 4.7219, "step": 156400 }, { "epoch": 0.29195309529530045, "grad_norm": 0.6988387703895569, "learning_rate": 0.00019926442435674302, "loss": 4.5019, "step": 156450 }, { "epoch": 0.2920464008546789, "grad_norm": 1.215088963508606, "learning_rate": 0.0001992639156657492, "loss": 4.7895, "step": 156500 }, { "epoch": 0.2921397064140574, "grad_norm": 1.1339821815490723, "learning_rate": 0.00019926340679957206, "loss": 4.7092, "step": 156550 }, { "epoch": 0.29223301197343593, "grad_norm": 1.184877872467041, "learning_rate": 0.00019926289775821246, "loss": 4.848, "step": 156600 }, { "epoch": 0.2923263175328144, "grad_norm": 1.0538442134857178, "learning_rate": 0.0001992623885416713, "loss": 4.7515, "step": 156650 }, { "epoch": 0.2924196230921929, "grad_norm": 0.875169038772583, "learning_rate": 0.0001992618791499495, "loss": 4.5665, "step": 156700 }, { "epoch": 0.29251292865157136, "grad_norm": 0.8852382898330688, "learning_rate": 0.00019926136958304795, "loss": 4.4925, "step": 156750 }, { "epoch": 0.2926062342109499, "grad_norm": 0.683463454246521, "learning_rate": 0.00019926085984096754, "loss": 4.5527, "step": 156800 }, { "epoch": 0.29269953977032837, "grad_norm": 0.848481297492981, "learning_rate": 0.00019926034992370922, "loss": 4.5793, "step": 156850 }, { "epoch": 0.29279284532970684, "grad_norm": 0.9897000193595886, "learning_rate": 0.0001992598398312738, "loss": 4.6143, "step": 156900 }, { "epoch": 0.2928861508890854, "grad_norm": 0.9709441065788269, "learning_rate": 0.00019925932956366222, "loss": 4.5603, "step": 156950 }, { "epoch": 0.29297945644846385, "grad_norm": 0.9135065674781799, "learning_rate": 0.00019925881912087544, "loss": 4.7722, "step": 157000 }, { "epoch": 0.2930727620078423, "grad_norm": 0.9769490361213684, "learning_rate": 0.00019925830850291428, "loss": 4.534, "step": 157050 }, { "epoch": 0.2931660675672208, "grad_norm": 1.041882872581482, "learning_rate": 0.00019925779770977963, "loss": 4.6919, "step": 157100 }, { "epoch": 0.29325937312659933, "grad_norm": 0.6851228475570679, "learning_rate": 0.0001992572867414725, "loss": 4.5882, "step": 157150 }, { "epoch": 0.2933526786859778, "grad_norm": 1.007688283920288, "learning_rate": 0.00019925677559799368, "loss": 4.5231, "step": 157200 }, { "epoch": 0.2934459842453563, "grad_norm": 1.0649223327636719, "learning_rate": 0.00019925626427934413, "loss": 4.7773, "step": 157250 }, { "epoch": 0.2935392898047348, "grad_norm": 0.9082051515579224, "learning_rate": 0.00019925575278552474, "loss": 4.5285, "step": 157300 }, { "epoch": 0.2936325953641133, "grad_norm": 1.0188795328140259, "learning_rate": 0.00019925524111653642, "loss": 4.7059, "step": 157350 }, { "epoch": 0.29372590092349177, "grad_norm": 0.9674282670021057, "learning_rate": 0.00019925472927238006, "loss": 4.5048, "step": 157400 }, { "epoch": 0.29381920648287024, "grad_norm": 0.9318031072616577, "learning_rate": 0.00019925421725305656, "loss": 4.5211, "step": 157450 }, { "epoch": 0.2939125120422488, "grad_norm": 1.3290448188781738, "learning_rate": 0.00019925370505856683, "loss": 4.6717, "step": 157500 }, { "epoch": 0.29400581760162725, "grad_norm": 0.9480525851249695, "learning_rate": 0.00019925319268891177, "loss": 4.4553, "step": 157550 }, { "epoch": 0.2940991231610057, "grad_norm": 0.9672073721885681, "learning_rate": 0.00019925268014409227, "loss": 4.8206, "step": 157600 }, { "epoch": 0.29419242872038426, "grad_norm": 0.8860331177711487, "learning_rate": 0.0001992521674241093, "loss": 4.6907, "step": 157650 }, { "epoch": 0.29428573427976273, "grad_norm": 0.9548688530921936, "learning_rate": 0.0001992516545289637, "loss": 4.6107, "step": 157700 }, { "epoch": 0.2943790398391412, "grad_norm": 0.9516497850418091, "learning_rate": 0.0001992511414586564, "loss": 4.5693, "step": 157750 }, { "epoch": 0.2944723453985197, "grad_norm": 1.0647341012954712, "learning_rate": 0.0001992506282131883, "loss": 4.5256, "step": 157800 }, { "epoch": 0.2945656509578982, "grad_norm": 0.9625911712646484, "learning_rate": 0.0001992501147925603, "loss": 4.592, "step": 157850 }, { "epoch": 0.2946589565172767, "grad_norm": 1.109753131866455, "learning_rate": 0.00019924960119677328, "loss": 4.3946, "step": 157900 }, { "epoch": 0.29475226207665517, "grad_norm": 0.7586527466773987, "learning_rate": 0.0001992490874258282, "loss": 4.6777, "step": 157950 }, { "epoch": 0.2948455676360337, "grad_norm": 1.0405488014221191, "learning_rate": 0.00019924857347972593, "loss": 4.6913, "step": 158000 }, { "epoch": 0.2949388731954122, "grad_norm": 0.7991945743560791, "learning_rate": 0.00019924805935846746, "loss": 4.6677, "step": 158050 }, { "epoch": 0.29503217875479065, "grad_norm": 1.0138425827026367, "learning_rate": 0.00019924754506205354, "loss": 4.7778, "step": 158100 }, { "epoch": 0.29512548431416913, "grad_norm": 0.9910553693771362, "learning_rate": 0.00019924703059048522, "loss": 4.5255, "step": 158150 }, { "epoch": 0.29521878987354766, "grad_norm": 1.0533186197280884, "learning_rate": 0.00019924651594376334, "loss": 4.779, "step": 158200 }, { "epoch": 0.29531209543292614, "grad_norm": 0.8881279230117798, "learning_rate": 0.0001992460011218888, "loss": 4.8608, "step": 158250 }, { "epoch": 0.2954054009923046, "grad_norm": 0.9535214304924011, "learning_rate": 0.00019924548612486258, "loss": 4.6377, "step": 158300 }, { "epoch": 0.2954987065516831, "grad_norm": 1.084489107131958, "learning_rate": 0.00019924497095268547, "loss": 4.6459, "step": 158350 }, { "epoch": 0.2955920121110616, "grad_norm": 1.0378689765930176, "learning_rate": 0.0001992444556053585, "loss": 4.7632, "step": 158400 }, { "epoch": 0.2956853176704401, "grad_norm": 0.878700315952301, "learning_rate": 0.00019924394008288254, "loss": 4.4921, "step": 158450 }, { "epoch": 0.29577862322981857, "grad_norm": 1.0042437314987183, "learning_rate": 0.00019924342438525848, "loss": 4.7155, "step": 158500 }, { "epoch": 0.2958719287891971, "grad_norm": 1.3455873727798462, "learning_rate": 0.00019924290851248722, "loss": 4.6473, "step": 158550 }, { "epoch": 0.2959652343485756, "grad_norm": 0.9573052525520325, "learning_rate": 0.0001992423924645697, "loss": 4.8455, "step": 158600 }, { "epoch": 0.29605853990795405, "grad_norm": 1.201541781425476, "learning_rate": 0.00019924187624150677, "loss": 4.6125, "step": 158650 }, { "epoch": 0.29615184546733253, "grad_norm": 1.1214284896850586, "learning_rate": 0.00019924135984329946, "loss": 4.5251, "step": 158700 }, { "epoch": 0.29624515102671106, "grad_norm": 0.8579356670379639, "learning_rate": 0.00019924084326994858, "loss": 4.6285, "step": 158750 }, { "epoch": 0.29633845658608954, "grad_norm": 1.1231144666671753, "learning_rate": 0.00019924032652145508, "loss": 4.5755, "step": 158800 }, { "epoch": 0.296431762145468, "grad_norm": 0.8243206739425659, "learning_rate": 0.00019923980959781984, "loss": 4.5558, "step": 158850 }, { "epoch": 0.29652506770484655, "grad_norm": 1.0935394763946533, "learning_rate": 0.00019923929249904385, "loss": 4.913, "step": 158900 }, { "epoch": 0.296618373264225, "grad_norm": 0.9777284264564514, "learning_rate": 0.00019923877522512795, "loss": 4.6809, "step": 158950 }, { "epoch": 0.2967116788236035, "grad_norm": 0.9716470241546631, "learning_rate": 0.00019923825777607305, "loss": 4.4853, "step": 159000 }, { "epoch": 0.2967116788236035, "eval_loss": 4.834835529327393, "eval_runtime": 229.1261, "eval_samples_per_second": 11.382, "eval_steps_per_second": 11.382, "eval_tts_loss": 7.493346262103278, "step": 159000 }, { "epoch": 0.296804984382982, "grad_norm": 1.0158995389938354, "learning_rate": 0.00019923774015188014, "loss": 4.6832, "step": 159050 }, { "epoch": 0.2968982899423605, "grad_norm": 1.0746111869812012, "learning_rate": 0.00019923722235255, "loss": 4.9126, "step": 159100 }, { "epoch": 0.296991595501739, "grad_norm": 0.8816997408866882, "learning_rate": 0.00019923670437808368, "loss": 4.6653, "step": 159150 }, { "epoch": 0.29708490106111746, "grad_norm": 0.8043944835662842, "learning_rate": 0.00019923618622848203, "loss": 4.6432, "step": 159200 }, { "epoch": 0.297178206620496, "grad_norm": 1.2232578992843628, "learning_rate": 0.00019923566790374598, "loss": 4.7212, "step": 159250 }, { "epoch": 0.29727151217987446, "grad_norm": 0.991773784160614, "learning_rate": 0.0001992351494038764, "loss": 4.7941, "step": 159300 }, { "epoch": 0.29736481773925294, "grad_norm": 0.9849087595939636, "learning_rate": 0.00019923463072887428, "loss": 4.6429, "step": 159350 }, { "epoch": 0.2974581232986314, "grad_norm": 1.1423635482788086, "learning_rate": 0.0001992341118787405, "loss": 4.875, "step": 159400 }, { "epoch": 0.29755142885800995, "grad_norm": 0.6247135996818542, "learning_rate": 0.00019923359285347594, "loss": 4.4517, "step": 159450 }, { "epoch": 0.2976447344173884, "grad_norm": 0.9761686325073242, "learning_rate": 0.00019923307365308154, "loss": 4.5767, "step": 159500 }, { "epoch": 0.2977380399767669, "grad_norm": 0.8199406266212463, "learning_rate": 0.00019923255427755826, "loss": 4.7449, "step": 159550 }, { "epoch": 0.29783134553614543, "grad_norm": 1.3415900468826294, "learning_rate": 0.00019923203472690698, "loss": 4.5032, "step": 159600 }, { "epoch": 0.2979246510955239, "grad_norm": 0.806204617023468, "learning_rate": 0.00019923151500112862, "loss": 4.6618, "step": 159650 }, { "epoch": 0.2980179566549024, "grad_norm": 1.2007147073745728, "learning_rate": 0.0001992309951002241, "loss": 4.8032, "step": 159700 }, { "epoch": 0.29811126221428086, "grad_norm": 0.580830991268158, "learning_rate": 0.00019923047502419432, "loss": 4.5456, "step": 159750 }, { "epoch": 0.2982045677736594, "grad_norm": 1.0669281482696533, "learning_rate": 0.0001992299547730402, "loss": 4.4379, "step": 159800 }, { "epoch": 0.29829787333303787, "grad_norm": 0.7621221542358398, "learning_rate": 0.0001992294343467627, "loss": 4.698, "step": 159850 }, { "epoch": 0.29839117889241634, "grad_norm": 1.0081877708435059, "learning_rate": 0.00019922891374536268, "loss": 4.4803, "step": 159900 }, { "epoch": 0.2984844844517949, "grad_norm": 1.1483149528503418, "learning_rate": 0.00019922839296884114, "loss": 4.7129, "step": 159950 }, { "epoch": 0.29857779001117335, "grad_norm": 1.0426714420318604, "learning_rate": 0.00019922787201719888, "loss": 4.5651, "step": 160000 }, { "epoch": 0.2986710955705518, "grad_norm": 0.8986678123474121, "learning_rate": 0.00019922735089043691, "loss": 4.779, "step": 160050 }, { "epoch": 0.2987644011299303, "grad_norm": 0.9667002558708191, "learning_rate": 0.00019922682958855613, "loss": 4.6677, "step": 160100 }, { "epoch": 0.29885770668930883, "grad_norm": 0.9710828065872192, "learning_rate": 0.00019922630811155747, "loss": 4.7131, "step": 160150 }, { "epoch": 0.2989510122486873, "grad_norm": 1.078286051750183, "learning_rate": 0.00019922578645944182, "loss": 4.6679, "step": 160200 }, { "epoch": 0.2990443178080658, "grad_norm": 0.694128692150116, "learning_rate": 0.00019922526463221008, "loss": 4.5504, "step": 160250 }, { "epoch": 0.2991376233674443, "grad_norm": 1.0354713201522827, "learning_rate": 0.00019922474262986325, "loss": 4.4172, "step": 160300 }, { "epoch": 0.2992309289268228, "grad_norm": 0.9620189666748047, "learning_rate": 0.0001992242204524022, "loss": 4.6703, "step": 160350 }, { "epoch": 0.29932423448620127, "grad_norm": 0.9035593271255493, "learning_rate": 0.0001992236980998279, "loss": 4.6122, "step": 160400 }, { "epoch": 0.29941754004557974, "grad_norm": 0.9520270824432373, "learning_rate": 0.00019922317557214116, "loss": 4.7732, "step": 160450 }, { "epoch": 0.2995108456049583, "grad_norm": 0.8644099831581116, "learning_rate": 0.00019922265286934302, "loss": 4.5148, "step": 160500 }, { "epoch": 0.29960415116433675, "grad_norm": 0.7828960418701172, "learning_rate": 0.00019922212999143435, "loss": 4.7247, "step": 160550 }, { "epoch": 0.2996974567237152, "grad_norm": 0.9005388617515564, "learning_rate": 0.00019922160693841608, "loss": 4.6693, "step": 160600 }, { "epoch": 0.29979076228309376, "grad_norm": 0.6794205904006958, "learning_rate": 0.0001992210837102891, "loss": 4.5782, "step": 160650 }, { "epoch": 0.29988406784247224, "grad_norm": 1.0133854150772095, "learning_rate": 0.0001992205603070544, "loss": 4.562, "step": 160700 }, { "epoch": 0.2999773734018507, "grad_norm": 0.9719075560569763, "learning_rate": 0.00019922003672871285, "loss": 4.7801, "step": 160750 }, { "epoch": 0.3000706789612292, "grad_norm": 0.9305253028869629, "learning_rate": 0.00019921951297526542, "loss": 4.4637, "step": 160800 }, { "epoch": 0.3001639845206077, "grad_norm": 0.8835601210594177, "learning_rate": 0.00019921898904671297, "loss": 4.6572, "step": 160850 }, { "epoch": 0.3002572900799862, "grad_norm": 1.0033226013183594, "learning_rate": 0.0001992184649430565, "loss": 4.4474, "step": 160900 }, { "epoch": 0.30035059563936467, "grad_norm": 1.0143111944198608, "learning_rate": 0.00019921794066429688, "loss": 4.6021, "step": 160950 }, { "epoch": 0.3004439011987432, "grad_norm": 1.0978728532791138, "learning_rate": 0.00019921741621043503, "loss": 4.7851, "step": 161000 }, { "epoch": 0.3005372067581217, "grad_norm": 0.6397398710250854, "learning_rate": 0.00019921689158147195, "loss": 4.7239, "step": 161050 }, { "epoch": 0.30063051231750015, "grad_norm": 1.0616174936294556, "learning_rate": 0.00019921636677740846, "loss": 4.7843, "step": 161100 }, { "epoch": 0.30072381787687863, "grad_norm": 1.0118709802627563, "learning_rate": 0.00019921584179824557, "loss": 4.2376, "step": 161150 }, { "epoch": 0.30081712343625716, "grad_norm": 1.0035194158554077, "learning_rate": 0.00019921531664398418, "loss": 4.3126, "step": 161200 }, { "epoch": 0.30091042899563564, "grad_norm": 1.0038213729858398, "learning_rate": 0.0001992147913146252, "loss": 4.673, "step": 161250 }, { "epoch": 0.3010037345550141, "grad_norm": 1.1764312982559204, "learning_rate": 0.00019921426581016958, "loss": 4.6918, "step": 161300 }, { "epoch": 0.3010970401143926, "grad_norm": 0.8991124629974365, "learning_rate": 0.00019921374013061823, "loss": 4.4333, "step": 161350 }, { "epoch": 0.3011903456737711, "grad_norm": 0.8967214226722717, "learning_rate": 0.0001992132142759721, "loss": 4.5326, "step": 161400 }, { "epoch": 0.3012836512331496, "grad_norm": 1.1561260223388672, "learning_rate": 0.0001992126882462321, "loss": 4.8165, "step": 161450 }, { "epoch": 0.3013769567925281, "grad_norm": 1.163358211517334, "learning_rate": 0.00019921216204139914, "loss": 4.48, "step": 161500 }, { "epoch": 0.3014702623519066, "grad_norm": 0.9330498576164246, "learning_rate": 0.0001992116356614742, "loss": 4.5813, "step": 161550 }, { "epoch": 0.3015635679112851, "grad_norm": 0.8502509593963623, "learning_rate": 0.00019921110910645816, "loss": 4.8287, "step": 161600 }, { "epoch": 0.30165687347066356, "grad_norm": 0.6995953917503357, "learning_rate": 0.000199210582376352, "loss": 4.683, "step": 161650 }, { "epoch": 0.30175017903004203, "grad_norm": 1.1503678560256958, "learning_rate": 0.0001992100554711566, "loss": 4.7569, "step": 161700 }, { "epoch": 0.30184348458942056, "grad_norm": 0.8800851702690125, "learning_rate": 0.0001992095283908729, "loss": 4.6177, "step": 161750 }, { "epoch": 0.30193679014879904, "grad_norm": 0.9621112942695618, "learning_rate": 0.00019920900113550185, "loss": 4.6983, "step": 161800 }, { "epoch": 0.3020300957081775, "grad_norm": 1.1952325105667114, "learning_rate": 0.00019920847370504435, "loss": 4.7141, "step": 161850 }, { "epoch": 0.30212340126755605, "grad_norm": 0.8587202429771423, "learning_rate": 0.0001992079460995014, "loss": 4.7418, "step": 161900 }, { "epoch": 0.3022167068269345, "grad_norm": 0.8307145237922668, "learning_rate": 0.0001992074183188738, "loss": 4.7172, "step": 161950 }, { "epoch": 0.302310012386313, "grad_norm": 0.9355490207672119, "learning_rate": 0.00019920689036316262, "loss": 4.7346, "step": 162000 }, { "epoch": 0.302310012386313, "eval_loss": 4.827625274658203, "eval_runtime": 232.9861, "eval_samples_per_second": 11.194, "eval_steps_per_second": 11.194, "eval_tts_loss": 7.458007696548911, "step": 162000 }, { "epoch": 0.3024033179456915, "grad_norm": 1.3112372159957886, "learning_rate": 0.00019920636223236872, "loss": 4.6245, "step": 162050 }, { "epoch": 0.30249662350507, "grad_norm": 0.8583440780639648, "learning_rate": 0.00019920583392649306, "loss": 4.6192, "step": 162100 }, { "epoch": 0.3025899290644485, "grad_norm": 0.9936934113502502, "learning_rate": 0.00019920530544553653, "loss": 4.7136, "step": 162150 }, { "epoch": 0.30268323462382696, "grad_norm": 0.9335197806358337, "learning_rate": 0.00019920477678950013, "loss": 4.6699, "step": 162200 }, { "epoch": 0.3027765401832055, "grad_norm": 0.9630725383758545, "learning_rate": 0.0001992042479583847, "loss": 4.6637, "step": 162250 }, { "epoch": 0.30286984574258397, "grad_norm": 0.7474913597106934, "learning_rate": 0.00019920371895219128, "loss": 4.7166, "step": 162300 }, { "epoch": 0.30296315130196244, "grad_norm": 0.7774944305419922, "learning_rate": 0.00019920318977092075, "loss": 4.8446, "step": 162350 }, { "epoch": 0.3030564568613409, "grad_norm": 1.0125433206558228, "learning_rate": 0.00019920266041457401, "loss": 4.6259, "step": 162400 }, { "epoch": 0.30314976242071945, "grad_norm": 1.1934956312179565, "learning_rate": 0.00019920213088315202, "loss": 4.6795, "step": 162450 }, { "epoch": 0.3032430679800979, "grad_norm": 0.9211152195930481, "learning_rate": 0.00019920160117665577, "loss": 4.6288, "step": 162500 }, { "epoch": 0.3033363735394764, "grad_norm": 0.9939976930618286, "learning_rate": 0.00019920107129508613, "loss": 4.7214, "step": 162550 }, { "epoch": 0.30342967909885493, "grad_norm": 0.8724084496498108, "learning_rate": 0.00019920054123844408, "loss": 4.4638, "step": 162600 }, { "epoch": 0.3035229846582334, "grad_norm": 0.994696855545044, "learning_rate": 0.00019920001100673046, "loss": 4.8186, "step": 162650 }, { "epoch": 0.3036162902176119, "grad_norm": 0.9215883612632751, "learning_rate": 0.00019919948059994634, "loss": 4.5902, "step": 162700 }, { "epoch": 0.30370959577699036, "grad_norm": 0.9584205746650696, "learning_rate": 0.00019919895001809258, "loss": 4.6455, "step": 162750 }, { "epoch": 0.3038029013363689, "grad_norm": 0.8970978856086731, "learning_rate": 0.00019919841926117012, "loss": 4.5378, "step": 162800 }, { "epoch": 0.30389620689574737, "grad_norm": 0.7979338765144348, "learning_rate": 0.0001991978883291799, "loss": 4.6594, "step": 162850 }, { "epoch": 0.30398951245512584, "grad_norm": 0.7578598856925964, "learning_rate": 0.00019919735722212284, "loss": 4.832, "step": 162900 }, { "epoch": 0.3040828180145044, "grad_norm": 1.0004535913467407, "learning_rate": 0.00019919682593999994, "loss": 4.695, "step": 162950 }, { "epoch": 0.30417612357388285, "grad_norm": 0.9248453974723816, "learning_rate": 0.00019919629448281209, "loss": 4.3721, "step": 163000 }, { "epoch": 0.3042694291332613, "grad_norm": 1.3004566431045532, "learning_rate": 0.00019919576285056022, "loss": 4.7349, "step": 163050 }, { "epoch": 0.3043627346926398, "grad_norm": 0.8813239336013794, "learning_rate": 0.0001991952310432453, "loss": 4.6711, "step": 163100 }, { "epoch": 0.30445604025201833, "grad_norm": 1.0755641460418701, "learning_rate": 0.00019919469906086822, "loss": 4.8154, "step": 163150 }, { "epoch": 0.3045493458113968, "grad_norm": 0.9873396158218384, "learning_rate": 0.00019919416690343, "loss": 4.3667, "step": 163200 }, { "epoch": 0.3046426513707753, "grad_norm": 0.7341350317001343, "learning_rate": 0.0001991936345709315, "loss": 4.5203, "step": 163250 }, { "epoch": 0.3047359569301538, "grad_norm": 1.094083547592163, "learning_rate": 0.00019919310206337366, "loss": 4.6661, "step": 163300 }, { "epoch": 0.3048292624895323, "grad_norm": 1.1205472946166992, "learning_rate": 0.00019919256938075747, "loss": 4.6381, "step": 163350 }, { "epoch": 0.30492256804891077, "grad_norm": 0.9988998174667358, "learning_rate": 0.00019919203652308385, "loss": 4.7326, "step": 163400 }, { "epoch": 0.30501587360828925, "grad_norm": 0.9494795203208923, "learning_rate": 0.00019919150349035375, "loss": 4.4937, "step": 163450 }, { "epoch": 0.3051091791676678, "grad_norm": 0.9424142241477966, "learning_rate": 0.0001991909702825681, "loss": 4.3626, "step": 163500 }, { "epoch": 0.30520248472704625, "grad_norm": 1.1510084867477417, "learning_rate": 0.00019919043689972782, "loss": 4.6443, "step": 163550 }, { "epoch": 0.30529579028642473, "grad_norm": 1.1867023706436157, "learning_rate": 0.0001991899033418339, "loss": 4.6603, "step": 163600 }, { "epoch": 0.30538909584580326, "grad_norm": 1.285799503326416, "learning_rate": 0.00019918936960888723, "loss": 4.7631, "step": 163650 }, { "epoch": 0.30548240140518174, "grad_norm": 1.011475920677185, "learning_rate": 0.00019918883570088877, "loss": 4.4944, "step": 163700 }, { "epoch": 0.3055757069645602, "grad_norm": 0.7682061791419983, "learning_rate": 0.0001991883016178395, "loss": 4.5854, "step": 163750 }, { "epoch": 0.3056690125239387, "grad_norm": 1.2138035297393799, "learning_rate": 0.0001991877673597403, "loss": 4.5669, "step": 163800 }, { "epoch": 0.3057623180833172, "grad_norm": 1.1199895143508911, "learning_rate": 0.00019918723292659216, "loss": 4.8251, "step": 163850 }, { "epoch": 0.3058556236426957, "grad_norm": 0.9371145367622375, "learning_rate": 0.00019918669831839601, "loss": 4.6153, "step": 163900 }, { "epoch": 0.30594892920207417, "grad_norm": 0.8127995729446411, "learning_rate": 0.00019918616353515277, "loss": 4.4168, "step": 163950 }, { "epoch": 0.30604223476145265, "grad_norm": 0.9031199216842651, "learning_rate": 0.00019918562857686344, "loss": 4.707, "step": 164000 }, { "epoch": 0.3061355403208312, "grad_norm": 1.1019059419631958, "learning_rate": 0.00019918509344352887, "loss": 4.577, "step": 164050 }, { "epoch": 0.30622884588020965, "grad_norm": 0.9607084393501282, "learning_rate": 0.0001991845581351501, "loss": 4.6266, "step": 164100 }, { "epoch": 0.30632215143958813, "grad_norm": 0.6045005917549133, "learning_rate": 0.00019918402265172805, "loss": 4.743, "step": 164150 }, { "epoch": 0.30641545699896666, "grad_norm": 0.9607996940612793, "learning_rate": 0.0001991834869932636, "loss": 4.637, "step": 164200 }, { "epoch": 0.30650876255834514, "grad_norm": 1.1145765781402588, "learning_rate": 0.0001991829511597578, "loss": 4.6641, "step": 164250 }, { "epoch": 0.3066020681177236, "grad_norm": 1.0428991317749023, "learning_rate": 0.0001991824151512115, "loss": 4.6888, "step": 164300 }, { "epoch": 0.3066953736771021, "grad_norm": 0.8830058574676514, "learning_rate": 0.00019918187896762575, "loss": 4.6689, "step": 164350 }, { "epoch": 0.3067886792364806, "grad_norm": 0.9978854656219482, "learning_rate": 0.0001991813426090014, "loss": 4.7586, "step": 164400 }, { "epoch": 0.3068819847958591, "grad_norm": 0.9433168768882751, "learning_rate": 0.00019918080607533943, "loss": 4.813, "step": 164450 }, { "epoch": 0.3069752903552376, "grad_norm": 0.64937824010849, "learning_rate": 0.00019918026936664078, "loss": 4.3087, "step": 164500 }, { "epoch": 0.3070685959146161, "grad_norm": 0.9438527822494507, "learning_rate": 0.00019917973248290643, "loss": 4.4295, "step": 164550 }, { "epoch": 0.3071619014739946, "grad_norm": 0.8068161606788635, "learning_rate": 0.00019917919542413728, "loss": 4.4977, "step": 164600 }, { "epoch": 0.30725520703337306, "grad_norm": 0.896666407585144, "learning_rate": 0.00019917865819033432, "loss": 4.6681, "step": 164650 }, { "epoch": 0.30734851259275153, "grad_norm": 1.1718946695327759, "learning_rate": 0.00019917812078149843, "loss": 4.565, "step": 164700 }, { "epoch": 0.30744181815213006, "grad_norm": 1.3140202760696411, "learning_rate": 0.00019917758319763066, "loss": 4.4831, "step": 164750 }, { "epoch": 0.30753512371150854, "grad_norm": 1.0430924892425537, "learning_rate": 0.0001991770454387319, "loss": 4.4573, "step": 164800 }, { "epoch": 0.307628429270887, "grad_norm": 0.9410809278488159, "learning_rate": 0.0001991765075048031, "loss": 4.5014, "step": 164850 }, { "epoch": 0.30772173483026555, "grad_norm": 0.7173283696174622, "learning_rate": 0.00019917596939584523, "loss": 4.7214, "step": 164900 }, { "epoch": 0.307815040389644, "grad_norm": 1.0757492780685425, "learning_rate": 0.00019917543111185921, "loss": 4.5144, "step": 164950 }, { "epoch": 0.3079083459490225, "grad_norm": 0.9228636026382446, "learning_rate": 0.000199174892652846, "loss": 4.6586, "step": 165000 }, { "epoch": 0.3079083459490225, "eval_loss": 4.838285446166992, "eval_runtime": 229.3082, "eval_samples_per_second": 11.373, "eval_steps_per_second": 11.373, "eval_tts_loss": 7.4116918211060066, "step": 165000 }, { "epoch": 0.308001651508401, "grad_norm": 1.0502113103866577, "learning_rate": 0.00019917435401880655, "loss": 4.6569, "step": 165050 }, { "epoch": 0.3080949570677795, "grad_norm": 0.9943327903747559, "learning_rate": 0.0001991738152097418, "loss": 4.5486, "step": 165100 }, { "epoch": 0.308188262627158, "grad_norm": 1.2838720083236694, "learning_rate": 0.00019917327622565276, "loss": 4.6833, "step": 165150 }, { "epoch": 0.30828156818653646, "grad_norm": 0.9620805978775024, "learning_rate": 0.00019917273706654033, "loss": 4.6248, "step": 165200 }, { "epoch": 0.308374873745915, "grad_norm": 1.1069529056549072, "learning_rate": 0.00019917219773240544, "loss": 4.4957, "step": 165250 }, { "epoch": 0.30846817930529347, "grad_norm": 0.6974848508834839, "learning_rate": 0.0001991716582232491, "loss": 4.7632, "step": 165300 }, { "epoch": 0.30856148486467194, "grad_norm": 1.0856902599334717, "learning_rate": 0.0001991711185390722, "loss": 4.6712, "step": 165350 }, { "epoch": 0.3086547904240504, "grad_norm": 1.2169914245605469, "learning_rate": 0.00019917057867987575, "loss": 4.5188, "step": 165400 }, { "epoch": 0.30874809598342895, "grad_norm": 0.9107069969177246, "learning_rate": 0.00019917003864566065, "loss": 4.7021, "step": 165450 }, { "epoch": 0.3088414015428074, "grad_norm": 0.910029947757721, "learning_rate": 0.00019916949843642794, "loss": 4.788, "step": 165500 }, { "epoch": 0.3089347071021859, "grad_norm": 1.238258957862854, "learning_rate": 0.00019916895805217844, "loss": 4.7447, "step": 165550 }, { "epoch": 0.30902801266156443, "grad_norm": 1.087679386138916, "learning_rate": 0.00019916841749291323, "loss": 4.6314, "step": 165600 }, { "epoch": 0.3091213182209429, "grad_norm": 0.9426969885826111, "learning_rate": 0.0001991678767586332, "loss": 4.4512, "step": 165650 }, { "epoch": 0.3092146237803214, "grad_norm": 0.944223940372467, "learning_rate": 0.00019916733584933932, "loss": 4.7139, "step": 165700 }, { "epoch": 0.30930792933969986, "grad_norm": 1.0137444734573364, "learning_rate": 0.00019916679476503254, "loss": 4.6459, "step": 165750 }, { "epoch": 0.3094012348990784, "grad_norm": 1.2591444253921509, "learning_rate": 0.00019916625350571384, "loss": 4.6593, "step": 165800 }, { "epoch": 0.30949454045845687, "grad_norm": 0.8761984705924988, "learning_rate": 0.00019916571207138412, "loss": 4.4277, "step": 165850 }, { "epoch": 0.30958784601783534, "grad_norm": 0.8550477027893066, "learning_rate": 0.00019916517046204434, "loss": 4.5806, "step": 165900 }, { "epoch": 0.3096811515772139, "grad_norm": 0.9725227355957031, "learning_rate": 0.00019916462867769553, "loss": 4.6497, "step": 165950 }, { "epoch": 0.30977445713659235, "grad_norm": 1.0279942750930786, "learning_rate": 0.00019916408671833858, "loss": 4.6852, "step": 166000 }, { "epoch": 0.3098677626959708, "grad_norm": 0.8512559533119202, "learning_rate": 0.0001991635445839745, "loss": 4.6959, "step": 166050 }, { "epoch": 0.3099610682553493, "grad_norm": 0.7771254181861877, "learning_rate": 0.00019916300227460417, "loss": 4.7392, "step": 166100 }, { "epoch": 0.31005437381472783, "grad_norm": 1.103044867515564, "learning_rate": 0.00019916245979022858, "loss": 4.3448, "step": 166150 }, { "epoch": 0.3101476793741063, "grad_norm": 1.1170908212661743, "learning_rate": 0.00019916191713084872, "loss": 4.4304, "step": 166200 }, { "epoch": 0.3102409849334848, "grad_norm": 1.0523208379745483, "learning_rate": 0.00019916137429646554, "loss": 4.7645, "step": 166250 }, { "epoch": 0.3103342904928633, "grad_norm": 1.361086368560791, "learning_rate": 0.00019916083128707995, "loss": 4.4946, "step": 166300 }, { "epoch": 0.3104275960522418, "grad_norm": 0.9509695172309875, "learning_rate": 0.00019916028810269297, "loss": 4.7445, "step": 166350 }, { "epoch": 0.31052090161162027, "grad_norm": 1.121620774269104, "learning_rate": 0.0001991597447433055, "loss": 4.8506, "step": 166400 }, { "epoch": 0.31061420717099875, "grad_norm": 1.0928542613983154, "learning_rate": 0.00019915920120891855, "loss": 4.6162, "step": 166450 }, { "epoch": 0.3107075127303773, "grad_norm": 1.339594841003418, "learning_rate": 0.00019915865749953304, "loss": 4.6252, "step": 166500 }, { "epoch": 0.31080081828975575, "grad_norm": 0.9323852062225342, "learning_rate": 0.00019915811361514996, "loss": 4.7945, "step": 166550 }, { "epoch": 0.31089412384913423, "grad_norm": 1.001569390296936, "learning_rate": 0.00019915756955577027, "loss": 4.5336, "step": 166600 }, { "epoch": 0.3109874294085127, "grad_norm": 0.9864224195480347, "learning_rate": 0.00019915702532139487, "loss": 4.5236, "step": 166650 }, { "epoch": 0.31108073496789124, "grad_norm": 0.9965605735778809, "learning_rate": 0.0001991564809120248, "loss": 4.6922, "step": 166700 }, { "epoch": 0.3111740405272697, "grad_norm": 1.0161752700805664, "learning_rate": 0.000199155936327661, "loss": 4.7956, "step": 166750 }, { "epoch": 0.3112673460866482, "grad_norm": 1.1669565439224243, "learning_rate": 0.00019915539156830443, "loss": 4.8049, "step": 166800 }, { "epoch": 0.3113606516460267, "grad_norm": 1.1081312894821167, "learning_rate": 0.000199154846633956, "loss": 4.6178, "step": 166850 }, { "epoch": 0.3114539572054052, "grad_norm": 0.9586905837059021, "learning_rate": 0.00019915430152461674, "loss": 4.8232, "step": 166900 }, { "epoch": 0.3115472627647837, "grad_norm": 1.1186403036117554, "learning_rate": 0.00019915375624028755, "loss": 4.6214, "step": 166950 }, { "epoch": 0.31164056832416215, "grad_norm": 1.1981947422027588, "learning_rate": 0.00019915321078096946, "loss": 4.7213, "step": 167000 }, { "epoch": 0.3117338738835407, "grad_norm": 0.7325296401977539, "learning_rate": 0.0001991526651466634, "loss": 4.6811, "step": 167050 }, { "epoch": 0.31182717944291916, "grad_norm": 1.0409541130065918, "learning_rate": 0.00019915211933737033, "loss": 4.5533, "step": 167100 }, { "epoch": 0.31192048500229763, "grad_norm": 0.8175338506698608, "learning_rate": 0.0001991515733530912, "loss": 4.5909, "step": 167150 }, { "epoch": 0.31201379056167616, "grad_norm": 1.0503971576690674, "learning_rate": 0.000199151027193827, "loss": 4.4858, "step": 167200 }, { "epoch": 0.31210709612105464, "grad_norm": 1.094265103340149, "learning_rate": 0.00019915048085957868, "loss": 4.5226, "step": 167250 }, { "epoch": 0.3122004016804331, "grad_norm": 0.8987294435501099, "learning_rate": 0.0001991499343503472, "loss": 4.7354, "step": 167300 }, { "epoch": 0.3122937072398116, "grad_norm": 1.0248451232910156, "learning_rate": 0.00019914938766613354, "loss": 4.5532, "step": 167350 }, { "epoch": 0.3123870127991901, "grad_norm": 0.8451243042945862, "learning_rate": 0.00019914884080693868, "loss": 4.5967, "step": 167400 }, { "epoch": 0.3124803183585686, "grad_norm": 0.8834465742111206, "learning_rate": 0.0001991482937727635, "loss": 4.2495, "step": 167450 }, { "epoch": 0.3125736239179471, "grad_norm": 0.8053299784660339, "learning_rate": 0.00019914774656360908, "loss": 4.9166, "step": 167500 }, { "epoch": 0.3126669294773256, "grad_norm": 1.0274240970611572, "learning_rate": 0.00019914719917947632, "loss": 4.7979, "step": 167550 }, { "epoch": 0.3127602350367041, "grad_norm": 0.8318864107131958, "learning_rate": 0.0001991466516203662, "loss": 4.5883, "step": 167600 }, { "epoch": 0.31285354059608256, "grad_norm": 1.0550457239151, "learning_rate": 0.00019914610388627969, "loss": 4.4031, "step": 167650 }, { "epoch": 0.31294684615546103, "grad_norm": 1.0894427299499512, "learning_rate": 0.00019914555597721773, "loss": 4.87, "step": 167700 }, { "epoch": 0.31304015171483957, "grad_norm": 0.8881887197494507, "learning_rate": 0.00019914500789318133, "loss": 4.8212, "step": 167750 }, { "epoch": 0.31313345727421804, "grad_norm": 0.9769049286842346, "learning_rate": 0.00019914445963417144, "loss": 4.5287, "step": 167800 }, { "epoch": 0.3132267628335965, "grad_norm": 1.0178450345993042, "learning_rate": 0.00019914391120018902, "loss": 4.5699, "step": 167850 }, { "epoch": 0.31332006839297505, "grad_norm": 0.9004485011100769, "learning_rate": 0.00019914336259123502, "loss": 4.6477, "step": 167900 }, { "epoch": 0.3134133739523535, "grad_norm": 1.098066806793213, "learning_rate": 0.00019914281380731047, "loss": 4.4141, "step": 167950 }, { "epoch": 0.313506679511732, "grad_norm": 0.9474552273750305, "learning_rate": 0.00019914226484841625, "loss": 4.4733, "step": 168000 }, { "epoch": 0.313506679511732, "eval_loss": 4.8274312019348145, "eval_runtime": 228.3749, "eval_samples_per_second": 11.42, "eval_steps_per_second": 11.42, "eval_tts_loss": 7.468050231518612, "step": 168000 }, { "epoch": 0.3135999850711105, "grad_norm": 1.2087820768356323, "learning_rate": 0.00019914171571455342, "loss": 4.6384, "step": 168050 }, { "epoch": 0.313693290630489, "grad_norm": 1.0665926933288574, "learning_rate": 0.00019914116640572287, "loss": 4.6831, "step": 168100 }, { "epoch": 0.3137865961898675, "grad_norm": 0.769088625907898, "learning_rate": 0.00019914061692192562, "loss": 4.7609, "step": 168150 }, { "epoch": 0.31387990174924596, "grad_norm": 1.0530633926391602, "learning_rate": 0.00019914006726316264, "loss": 4.9207, "step": 168200 }, { "epoch": 0.3139732073086245, "grad_norm": 0.705093502998352, "learning_rate": 0.00019913951742943486, "loss": 4.642, "step": 168250 }, { "epoch": 0.31406651286800297, "grad_norm": 0.9531348347663879, "learning_rate": 0.0001991389674207433, "loss": 4.9607, "step": 168300 }, { "epoch": 0.31415981842738144, "grad_norm": 0.8925808668136597, "learning_rate": 0.0001991384172370889, "loss": 4.4425, "step": 168350 }, { "epoch": 0.3142531239867599, "grad_norm": 0.8512563705444336, "learning_rate": 0.00019913786687847265, "loss": 4.5021, "step": 168400 }, { "epoch": 0.31434642954613845, "grad_norm": 1.233641266822815, "learning_rate": 0.00019913731634489548, "loss": 4.7805, "step": 168450 }, { "epoch": 0.3144397351055169, "grad_norm": 0.7689288854598999, "learning_rate": 0.0001991367656363584, "loss": 4.5333, "step": 168500 }, { "epoch": 0.3145330406648954, "grad_norm": 1.1268550157546997, "learning_rate": 0.00019913621475286237, "loss": 4.7591, "step": 168550 }, { "epoch": 0.31462634622427393, "grad_norm": 1.1939990520477295, "learning_rate": 0.00019913566369440837, "loss": 4.6831, "step": 168600 }, { "epoch": 0.3147196517836524, "grad_norm": 0.9455341100692749, "learning_rate": 0.00019913511246099735, "loss": 4.7392, "step": 168650 }, { "epoch": 0.3148129573430309, "grad_norm": 0.9590946435928345, "learning_rate": 0.00019913456105263033, "loss": 4.4734, "step": 168700 }, { "epoch": 0.31490626290240936, "grad_norm": 1.0091745853424072, "learning_rate": 0.00019913400946930824, "loss": 4.6294, "step": 168750 }, { "epoch": 0.3149995684617879, "grad_norm": 0.9442651867866516, "learning_rate": 0.00019913345771103207, "loss": 4.5777, "step": 168800 }, { "epoch": 0.31509287402116637, "grad_norm": 0.7429641485214233, "learning_rate": 0.00019913290577780278, "loss": 4.722, "step": 168850 }, { "epoch": 0.31518617958054485, "grad_norm": 1.2271195650100708, "learning_rate": 0.00019913235366962137, "loss": 4.8243, "step": 168900 }, { "epoch": 0.3152794851399234, "grad_norm": 1.0357447862625122, "learning_rate": 0.00019913180138648876, "loss": 4.5255, "step": 168950 }, { "epoch": 0.31537279069930185, "grad_norm": 1.144982933998108, "learning_rate": 0.00019913124892840603, "loss": 4.5941, "step": 169000 }, { "epoch": 0.31546609625868033, "grad_norm": 0.9120845794677734, "learning_rate": 0.00019913069629537403, "loss": 4.6719, "step": 169050 }, { "epoch": 0.3155594018180588, "grad_norm": 1.1090995073318481, "learning_rate": 0.00019913014348739384, "loss": 4.7637, "step": 169100 }, { "epoch": 0.31565270737743734, "grad_norm": 1.0764634609222412, "learning_rate": 0.00019912959050446633, "loss": 4.5816, "step": 169150 }, { "epoch": 0.3157460129368158, "grad_norm": 0.9311011433601379, "learning_rate": 0.00019912903734659257, "loss": 4.5876, "step": 169200 }, { "epoch": 0.3158393184961943, "grad_norm": 0.9119871854782104, "learning_rate": 0.00019912848401377353, "loss": 4.428, "step": 169250 }, { "epoch": 0.31593262405557276, "grad_norm": 0.9250431060791016, "learning_rate": 0.00019912793050601014, "loss": 4.528, "step": 169300 }, { "epoch": 0.3160259296149513, "grad_norm": 1.0978806018829346, "learning_rate": 0.00019912737682330336, "loss": 4.8015, "step": 169350 }, { "epoch": 0.31611923517432977, "grad_norm": 0.9182941913604736, "learning_rate": 0.00019912682296565422, "loss": 4.6908, "step": 169400 }, { "epoch": 0.31621254073370825, "grad_norm": 0.8654223084449768, "learning_rate": 0.0001991262689330637, "loss": 4.5184, "step": 169450 }, { "epoch": 0.3163058462930868, "grad_norm": 0.9357742071151733, "learning_rate": 0.00019912571472553272, "loss": 4.7359, "step": 169500 }, { "epoch": 0.31639915185246525, "grad_norm": 0.9018232226371765, "learning_rate": 0.00019912516034306232, "loss": 4.497, "step": 169550 }, { "epoch": 0.31649245741184373, "grad_norm": 1.170499324798584, "learning_rate": 0.00019912460578565345, "loss": 4.6351, "step": 169600 }, { "epoch": 0.3165857629712222, "grad_norm": 1.0474013090133667, "learning_rate": 0.0001991240510533071, "loss": 4.6865, "step": 169650 }, { "epoch": 0.31667906853060074, "grad_norm": 0.7851083874702454, "learning_rate": 0.00019912349614602425, "loss": 4.6992, "step": 169700 }, { "epoch": 0.3167723740899792, "grad_norm": 1.0943204164505005, "learning_rate": 0.00019912294106380586, "loss": 4.683, "step": 169750 }, { "epoch": 0.3168656796493577, "grad_norm": 1.037070870399475, "learning_rate": 0.00019912238580665288, "loss": 4.6626, "step": 169800 }, { "epoch": 0.3169589852087362, "grad_norm": 0.8870090842247009, "learning_rate": 0.00019912183037456638, "loss": 4.6193, "step": 169850 }, { "epoch": 0.3170522907681147, "grad_norm": 0.9513359069824219, "learning_rate": 0.00019912127476754728, "loss": 4.694, "step": 169900 }, { "epoch": 0.3171455963274932, "grad_norm": 1.1024489402770996, "learning_rate": 0.00019912071898559654, "loss": 4.6326, "step": 169950 }, { "epoch": 0.31723890188687165, "grad_norm": 1.0496639013290405, "learning_rate": 0.0001991201630287152, "loss": 4.4681, "step": 170000 }, { "epoch": 0.3173322074462502, "grad_norm": 0.8167015910148621, "learning_rate": 0.00019911960689690422, "loss": 4.6452, "step": 170050 }, { "epoch": 0.31742551300562866, "grad_norm": 0.811915397644043, "learning_rate": 0.00019911905059016456, "loss": 4.8109, "step": 170100 }, { "epoch": 0.31751881856500713, "grad_norm": 1.01187264919281, "learning_rate": 0.00019911849410849723, "loss": 4.6726, "step": 170150 }, { "epoch": 0.31761212412438566, "grad_norm": 1.0084149837493896, "learning_rate": 0.0001991179374519032, "loss": 4.5055, "step": 170200 }, { "epoch": 0.31770542968376414, "grad_norm": 0.9547125101089478, "learning_rate": 0.00019911738062038342, "loss": 4.5173, "step": 170250 }, { "epoch": 0.3177987352431426, "grad_norm": 0.9173873066902161, "learning_rate": 0.00019911682361393895, "loss": 4.8112, "step": 170300 }, { "epoch": 0.3178920408025211, "grad_norm": 1.2184867858886719, "learning_rate": 0.0001991162664325707, "loss": 4.4778, "step": 170350 }, { "epoch": 0.3179853463618996, "grad_norm": 1.0674309730529785, "learning_rate": 0.0001991157090762797, "loss": 4.5748, "step": 170400 }, { "epoch": 0.3180786519212781, "grad_norm": 0.7090270519256592, "learning_rate": 0.00019911515154506687, "loss": 4.4085, "step": 170450 }, { "epoch": 0.3181719574806566, "grad_norm": 0.9503317475318909, "learning_rate": 0.00019911459383893326, "loss": 4.5279, "step": 170500 }, { "epoch": 0.3182652630400351, "grad_norm": 0.6449097990989685, "learning_rate": 0.00019911403595787984, "loss": 4.5365, "step": 170550 }, { "epoch": 0.3183585685994136, "grad_norm": 1.0800449848175049, "learning_rate": 0.00019911347790190758, "loss": 4.5558, "step": 170600 }, { "epoch": 0.31845187415879206, "grad_norm": 0.8874872326850891, "learning_rate": 0.00019911291967101747, "loss": 4.6167, "step": 170650 }, { "epoch": 0.31854517971817053, "grad_norm": 0.8549138903617859, "learning_rate": 0.00019911236126521047, "loss": 4.6317, "step": 170700 }, { "epoch": 0.31863848527754907, "grad_norm": 1.0165430307388306, "learning_rate": 0.00019911180268448764, "loss": 4.6746, "step": 170750 }, { "epoch": 0.31873179083692754, "grad_norm": 1.049067497253418, "learning_rate": 0.00019911124392884988, "loss": 4.7359, "step": 170800 }, { "epoch": 0.318825096396306, "grad_norm": 1.071564793586731, "learning_rate": 0.00019911068499829825, "loss": 4.5707, "step": 170850 }, { "epoch": 0.31891840195568455, "grad_norm": 0.9487912654876709, "learning_rate": 0.00019911012589283366, "loss": 4.8717, "step": 170900 }, { "epoch": 0.319011707515063, "grad_norm": 1.0813241004943848, "learning_rate": 0.00019910956661245717, "loss": 4.6888, "step": 170950 }, { "epoch": 0.3191050130744415, "grad_norm": 1.02992844581604, "learning_rate": 0.0001991090071571697, "loss": 4.6039, "step": 171000 }, { "epoch": 0.3191050130744415, "eval_loss": 4.821337699890137, "eval_runtime": 229.1039, "eval_samples_per_second": 11.383, "eval_steps_per_second": 11.383, "eval_tts_loss": 7.466205600627692, "step": 171000 }, { "epoch": 0.31919831863382, "grad_norm": 1.2654902935028076, "learning_rate": 0.0001991084475269723, "loss": 4.5334, "step": 171050 }, { "epoch": 0.3192916241931985, "grad_norm": 0.9937511086463928, "learning_rate": 0.00019910788772186592, "loss": 4.6948, "step": 171100 }, { "epoch": 0.319384929752577, "grad_norm": 1.1610994338989258, "learning_rate": 0.00019910732774185155, "loss": 4.3105, "step": 171150 }, { "epoch": 0.31947823531195546, "grad_norm": 0.9847679734230042, "learning_rate": 0.00019910676758693022, "loss": 4.7063, "step": 171200 }, { "epoch": 0.319571540871334, "grad_norm": 1.15132474899292, "learning_rate": 0.00019910620725710284, "loss": 4.4479, "step": 171250 }, { "epoch": 0.31966484643071247, "grad_norm": 1.0646486282348633, "learning_rate": 0.00019910564675237048, "loss": 4.6407, "step": 171300 }, { "epoch": 0.31975815199009094, "grad_norm": 0.982887864112854, "learning_rate": 0.00019910508607273405, "loss": 4.4841, "step": 171350 }, { "epoch": 0.3198514575494694, "grad_norm": 1.1266651153564453, "learning_rate": 0.0001991045252181946, "loss": 4.8636, "step": 171400 }, { "epoch": 0.31994476310884795, "grad_norm": 0.9004781246185303, "learning_rate": 0.0001991039641887531, "loss": 4.7329, "step": 171450 }, { "epoch": 0.3200380686682264, "grad_norm": 0.904808759689331, "learning_rate": 0.00019910340298441055, "loss": 4.7403, "step": 171500 }, { "epoch": 0.3201313742276049, "grad_norm": 1.0027724504470825, "learning_rate": 0.0001991028416051679, "loss": 4.7172, "step": 171550 }, { "epoch": 0.32022467978698343, "grad_norm": 0.9115274548530579, "learning_rate": 0.00019910228005102623, "loss": 4.7143, "step": 171600 }, { "epoch": 0.3203179853463619, "grad_norm": 1.1834417581558228, "learning_rate": 0.0001991017183219864, "loss": 4.7278, "step": 171650 }, { "epoch": 0.3204112909057404, "grad_norm": 1.0323907136917114, "learning_rate": 0.00019910115641804952, "loss": 4.7366, "step": 171700 }, { "epoch": 0.32050459646511886, "grad_norm": 1.0845869779586792, "learning_rate": 0.00019910059433921653, "loss": 4.7036, "step": 171750 }, { "epoch": 0.3205979020244974, "grad_norm": 1.2340019941329956, "learning_rate": 0.00019910003208548844, "loss": 4.3835, "step": 171800 }, { "epoch": 0.32069120758387587, "grad_norm": 1.1320146322250366, "learning_rate": 0.0001990994696568662, "loss": 4.5122, "step": 171850 }, { "epoch": 0.32078451314325435, "grad_norm": 1.0517185926437378, "learning_rate": 0.00019909890705335087, "loss": 4.4093, "step": 171900 }, { "epoch": 0.3208778187026328, "grad_norm": 0.9432281255722046, "learning_rate": 0.00019909834427494342, "loss": 4.7337, "step": 171950 }, { "epoch": 0.32097112426201135, "grad_norm": 1.0861693620681763, "learning_rate": 0.00019909778132164475, "loss": 4.5552, "step": 172000 }, { "epoch": 0.32106442982138983, "grad_norm": 1.0468862056732178, "learning_rate": 0.00019909721819345603, "loss": 4.8132, "step": 172050 }, { "epoch": 0.3211577353807683, "grad_norm": 0.941402018070221, "learning_rate": 0.00019909665489037807, "loss": 4.7838, "step": 172100 }, { "epoch": 0.32125104094014684, "grad_norm": 0.92326420545578, "learning_rate": 0.000199096091412412, "loss": 4.476, "step": 172150 }, { "epoch": 0.3213443464995253, "grad_norm": 0.8617150187492371, "learning_rate": 0.00019909552775955876, "loss": 4.5015, "step": 172200 }, { "epoch": 0.3214376520589038, "grad_norm": 0.7417050004005432, "learning_rate": 0.00019909496393181932, "loss": 4.8644, "step": 172250 }, { "epoch": 0.32153095761828226, "grad_norm": 1.1346899271011353, "learning_rate": 0.00019909439992919474, "loss": 4.8367, "step": 172300 }, { "epoch": 0.3216242631776608, "grad_norm": 0.9893891215324402, "learning_rate": 0.00019909383575168597, "loss": 4.7965, "step": 172350 }, { "epoch": 0.3217175687370393, "grad_norm": 0.8581725358963013, "learning_rate": 0.000199093271399294, "loss": 4.6347, "step": 172400 }, { "epoch": 0.32181087429641775, "grad_norm": 0.9298039078712463, "learning_rate": 0.00019909270687201987, "loss": 4.7319, "step": 172450 }, { "epoch": 0.3219041798557963, "grad_norm": 1.0204685926437378, "learning_rate": 0.00019909214216986453, "loss": 4.6937, "step": 172500 }, { "epoch": 0.32199748541517476, "grad_norm": 1.0665239095687866, "learning_rate": 0.000199091577292829, "loss": 4.7268, "step": 172550 }, { "epoch": 0.32209079097455323, "grad_norm": 1.0132877826690674, "learning_rate": 0.0001990910122409143, "loss": 4.509, "step": 172600 }, { "epoch": 0.3221840965339317, "grad_norm": 1.0414512157440186, "learning_rate": 0.00019909044701412135, "loss": 4.4304, "step": 172650 }, { "epoch": 0.32227740209331024, "grad_norm": 0.763314962387085, "learning_rate": 0.0001990898816124512, "loss": 4.5164, "step": 172700 }, { "epoch": 0.3223707076526887, "grad_norm": 0.8184088468551636, "learning_rate": 0.00019908931603590486, "loss": 4.7437, "step": 172750 }, { "epoch": 0.3224640132120672, "grad_norm": 0.8922542333602905, "learning_rate": 0.00019908875028448332, "loss": 4.4946, "step": 172800 }, { "epoch": 0.3225573187714457, "grad_norm": 0.7964805364608765, "learning_rate": 0.00019908818435818756, "loss": 4.6092, "step": 172850 }, { "epoch": 0.3226506243308242, "grad_norm": 0.9622323513031006, "learning_rate": 0.0001990876182570186, "loss": 4.7243, "step": 172900 }, { "epoch": 0.3227439298902027, "grad_norm": 1.1066397428512573, "learning_rate": 0.0001990870519809774, "loss": 4.6053, "step": 172950 }, { "epoch": 0.32283723544958115, "grad_norm": 1.3279519081115723, "learning_rate": 0.00019908648553006503, "loss": 4.5265, "step": 173000 }, { "epoch": 0.3229305410089597, "grad_norm": 0.8593128323554993, "learning_rate": 0.00019908591890428244, "loss": 4.8224, "step": 173050 }, { "epoch": 0.32302384656833816, "grad_norm": 0.9954770803451538, "learning_rate": 0.00019908535210363062, "loss": 4.7258, "step": 173100 }, { "epoch": 0.32311715212771663, "grad_norm": 0.6582159996032715, "learning_rate": 0.0001990847851281106, "loss": 4.6627, "step": 173150 }, { "epoch": 0.32321045768709517, "grad_norm": 1.021619439125061, "learning_rate": 0.00019908421797772336, "loss": 4.7628, "step": 173200 }, { "epoch": 0.32330376324647364, "grad_norm": 1.0340412855148315, "learning_rate": 0.0001990836506524699, "loss": 4.4405, "step": 173250 }, { "epoch": 0.3233970688058521, "grad_norm": 1.0030783414840698, "learning_rate": 0.00019908308315235126, "loss": 4.7434, "step": 173300 }, { "epoch": 0.3234903743652306, "grad_norm": 1.1510215997695923, "learning_rate": 0.0001990825154773684, "loss": 4.6975, "step": 173350 }, { "epoch": 0.3235836799246091, "grad_norm": 0.7062892317771912, "learning_rate": 0.00019908194762752234, "loss": 4.6726, "step": 173400 }, { "epoch": 0.3236769854839876, "grad_norm": 1.1423907279968262, "learning_rate": 0.00019908137960281407, "loss": 4.524, "step": 173450 }, { "epoch": 0.3237702910433661, "grad_norm": 0.9802839159965515, "learning_rate": 0.00019908081140324456, "loss": 4.5331, "step": 173500 }, { "epoch": 0.3238635966027446, "grad_norm": 0.7058094143867493, "learning_rate": 0.00019908024302881493, "loss": 4.4531, "step": 173550 }, { "epoch": 0.3239569021621231, "grad_norm": 0.9642864465713501, "learning_rate": 0.00019907967447952603, "loss": 4.5526, "step": 173600 }, { "epoch": 0.32405020772150156, "grad_norm": 0.8382852077484131, "learning_rate": 0.000199079105755379, "loss": 4.6682, "step": 173650 }, { "epoch": 0.32414351328088004, "grad_norm": 1.2792258262634277, "learning_rate": 0.00019907853685637475, "loss": 4.6946, "step": 173700 }, { "epoch": 0.32423681884025857, "grad_norm": 1.1190528869628906, "learning_rate": 0.00019907796778251432, "loss": 4.5769, "step": 173750 }, { "epoch": 0.32433012439963704, "grad_norm": 1.1447540521621704, "learning_rate": 0.0001990773985337987, "loss": 4.8288, "step": 173800 }, { "epoch": 0.3244234299590155, "grad_norm": 1.0323656797409058, "learning_rate": 0.00019907682911022892, "loss": 4.6316, "step": 173850 }, { "epoch": 0.32451673551839405, "grad_norm": 0.9703738689422607, "learning_rate": 0.00019907625951180594, "loss": 4.6256, "step": 173900 }, { "epoch": 0.3246100410777725, "grad_norm": 1.1295841932296753, "learning_rate": 0.0001990756897385308, "loss": 4.542, "step": 173950 }, { "epoch": 0.324703346637151, "grad_norm": 1.204323172569275, "learning_rate": 0.00019907511979040454, "loss": 4.655, "step": 174000 }, { "epoch": 0.324703346637151, "eval_loss": 4.816260814666748, "eval_runtime": 230.4264, "eval_samples_per_second": 11.318, "eval_steps_per_second": 11.318, "eval_tts_loss": 7.435452688951045, "step": 174000 }, { "epoch": 0.3247966521965295, "grad_norm": 1.0280636548995972, "learning_rate": 0.00019907454966742812, "loss": 4.6904, "step": 174050 }, { "epoch": 0.324889957755908, "grad_norm": 1.1565958261489868, "learning_rate": 0.00019907397936960253, "loss": 4.3498, "step": 174100 }, { "epoch": 0.3249832633152865, "grad_norm": 1.0304701328277588, "learning_rate": 0.0001990734088969288, "loss": 4.5535, "step": 174150 }, { "epoch": 0.32507656887466496, "grad_norm": 0.8157714605331421, "learning_rate": 0.00019907283824940794, "loss": 4.5031, "step": 174200 }, { "epoch": 0.3251698744340435, "grad_norm": 0.6631931662559509, "learning_rate": 0.00019907226742704097, "loss": 4.8029, "step": 174250 }, { "epoch": 0.32526317999342197, "grad_norm": 1.1516517400741577, "learning_rate": 0.00019907169642982886, "loss": 4.7112, "step": 174300 }, { "epoch": 0.32535648555280045, "grad_norm": 0.7090115547180176, "learning_rate": 0.00019907112525777265, "loss": 4.8284, "step": 174350 }, { "epoch": 0.3254497911121789, "grad_norm": 1.1411478519439697, "learning_rate": 0.00019907055391087334, "loss": 4.7277, "step": 174400 }, { "epoch": 0.32554309667155745, "grad_norm": 0.763681948184967, "learning_rate": 0.00019906998238913192, "loss": 4.8038, "step": 174450 }, { "epoch": 0.32563640223093593, "grad_norm": 1.0577315092086792, "learning_rate": 0.0001990694106925494, "loss": 4.6134, "step": 174500 }, { "epoch": 0.3257297077903144, "grad_norm": 0.9963567852973938, "learning_rate": 0.00019906883882112682, "loss": 4.472, "step": 174550 }, { "epoch": 0.3258230133496929, "grad_norm": 0.8183549046516418, "learning_rate": 0.00019906826677486516, "loss": 4.4465, "step": 174600 }, { "epoch": 0.3259163189090714, "grad_norm": 0.908412516117096, "learning_rate": 0.00019906769455376545, "loss": 4.5436, "step": 174650 }, { "epoch": 0.3260096244684499, "grad_norm": 1.11734139919281, "learning_rate": 0.0001990671221578287, "loss": 4.8297, "step": 174700 }, { "epoch": 0.32610293002782836, "grad_norm": 1.123195767402649, "learning_rate": 0.0001990665495870559, "loss": 4.5935, "step": 174750 }, { "epoch": 0.3261962355872069, "grad_norm": 0.9942184090614319, "learning_rate": 0.00019906597684144807, "loss": 4.6195, "step": 174800 }, { "epoch": 0.32628954114658537, "grad_norm": 1.0256868600845337, "learning_rate": 0.00019906540392100624, "loss": 4.5741, "step": 174850 }, { "epoch": 0.32638284670596385, "grad_norm": 1.351872205734253, "learning_rate": 0.00019906483082573136, "loss": 4.5077, "step": 174900 }, { "epoch": 0.3264761522653423, "grad_norm": 0.7622151374816895, "learning_rate": 0.00019906425755562452, "loss": 4.5419, "step": 174950 }, { "epoch": 0.32656945782472085, "grad_norm": 0.8031182885169983, "learning_rate": 0.00019906368411068672, "loss": 4.5477, "step": 175000 }, { "epoch": 0.32666276338409933, "grad_norm": 1.1167811155319214, "learning_rate": 0.0001990631104909189, "loss": 4.7045, "step": 175050 }, { "epoch": 0.3267560689434778, "grad_norm": 1.2244101762771606, "learning_rate": 0.00019906253669632214, "loss": 4.7501, "step": 175100 }, { "epoch": 0.32684937450285634, "grad_norm": 0.8345790505409241, "learning_rate": 0.00019906196272689744, "loss": 4.5942, "step": 175150 }, { "epoch": 0.3269426800622348, "grad_norm": 0.927778422832489, "learning_rate": 0.00019906138858264577, "loss": 4.7552, "step": 175200 }, { "epoch": 0.3270359856216133, "grad_norm": 1.1129297018051147, "learning_rate": 0.00019906081426356823, "loss": 4.5184, "step": 175250 }, { "epoch": 0.32712929118099177, "grad_norm": 0.8951523900032043, "learning_rate": 0.00019906023976966577, "loss": 4.7909, "step": 175300 }, { "epoch": 0.3272225967403703, "grad_norm": 1.09775710105896, "learning_rate": 0.0001990596651009394, "loss": 4.61, "step": 175350 }, { "epoch": 0.3273159022997488, "grad_norm": 1.0176907777786255, "learning_rate": 0.00019905909025739015, "loss": 4.7255, "step": 175400 }, { "epoch": 0.32740920785912725, "grad_norm": 1.1389837265014648, "learning_rate": 0.00019905851523901907, "loss": 4.4602, "step": 175450 }, { "epoch": 0.3275025134185058, "grad_norm": 0.8852382302284241, "learning_rate": 0.00019905794004582708, "loss": 4.629, "step": 175500 }, { "epoch": 0.32759581897788426, "grad_norm": 1.0725276470184326, "learning_rate": 0.00019905736467781533, "loss": 4.5922, "step": 175550 }, { "epoch": 0.32768912453726273, "grad_norm": 0.8179895281791687, "learning_rate": 0.00019905678913498468, "loss": 4.6167, "step": 175600 }, { "epoch": 0.3277824300966412, "grad_norm": 0.7662099599838257, "learning_rate": 0.0001990562134173363, "loss": 4.46, "step": 175650 }, { "epoch": 0.32787573565601974, "grad_norm": 1.104725956916809, "learning_rate": 0.00019905563752487112, "loss": 4.6232, "step": 175700 }, { "epoch": 0.3279690412153982, "grad_norm": 1.0175747871398926, "learning_rate": 0.0001990550614575901, "loss": 4.5933, "step": 175750 }, { "epoch": 0.3280623467747767, "grad_norm": 1.1158710718154907, "learning_rate": 0.00019905448521549438, "loss": 4.6522, "step": 175800 }, { "epoch": 0.3281556523341552, "grad_norm": 1.6744157075881958, "learning_rate": 0.00019905390879858495, "loss": 4.7029, "step": 175850 }, { "epoch": 0.3282489578935337, "grad_norm": 0.9802636504173279, "learning_rate": 0.00019905333220686275, "loss": 4.5862, "step": 175900 }, { "epoch": 0.3283422634529122, "grad_norm": 0.7022011280059814, "learning_rate": 0.00019905275544032888, "loss": 4.5594, "step": 175950 }, { "epoch": 0.32843556901229065, "grad_norm": 0.8897641897201538, "learning_rate": 0.00019905217849898433, "loss": 4.6417, "step": 176000 }, { "epoch": 0.3285288745716692, "grad_norm": 1.1504569053649902, "learning_rate": 0.00019905160138283007, "loss": 4.6595, "step": 176050 }, { "epoch": 0.32862218013104766, "grad_norm": 1.0295069217681885, "learning_rate": 0.0001990510240918672, "loss": 4.5644, "step": 176100 }, { "epoch": 0.32871548569042613, "grad_norm": 0.9090021848678589, "learning_rate": 0.0001990504466260967, "loss": 4.4895, "step": 176150 }, { "epoch": 0.32880879124980467, "grad_norm": 0.9149088263511658, "learning_rate": 0.00019904986898551955, "loss": 4.6901, "step": 176200 }, { "epoch": 0.32890209680918314, "grad_norm": 0.8863251209259033, "learning_rate": 0.00019904929117013685, "loss": 4.6697, "step": 176250 }, { "epoch": 0.3289954023685616, "grad_norm": 1.07883882522583, "learning_rate": 0.00019904871317994957, "loss": 4.598, "step": 176300 }, { "epoch": 0.3290887079279401, "grad_norm": 0.9367828369140625, "learning_rate": 0.00019904813501495875, "loss": 4.6257, "step": 176350 }, { "epoch": 0.3291820134873186, "grad_norm": 1.0381193161010742, "learning_rate": 0.00019904755667516538, "loss": 4.6984, "step": 176400 }, { "epoch": 0.3292753190466971, "grad_norm": 0.847889244556427, "learning_rate": 0.00019904697816057053, "loss": 4.4849, "step": 176450 }, { "epoch": 0.3293686246060756, "grad_norm": 1.0124871730804443, "learning_rate": 0.00019904639947117514, "loss": 4.7115, "step": 176500 }, { "epoch": 0.3294619301654541, "grad_norm": 0.8684128522872925, "learning_rate": 0.0001990458206069803, "loss": 4.5261, "step": 176550 }, { "epoch": 0.3295552357248326, "grad_norm": 0.7277292609214783, "learning_rate": 0.00019904524156798704, "loss": 4.4937, "step": 176600 }, { "epoch": 0.32964854128421106, "grad_norm": 1.0759245157241821, "learning_rate": 0.00019904466235419634, "loss": 4.6008, "step": 176650 }, { "epoch": 0.32974184684358954, "grad_norm": 0.9541880488395691, "learning_rate": 0.00019904408296560924, "loss": 4.542, "step": 176700 }, { "epoch": 0.32983515240296807, "grad_norm": 0.8675011992454529, "learning_rate": 0.00019904350340222676, "loss": 4.5248, "step": 176750 }, { "epoch": 0.32992845796234654, "grad_norm": 0.9355060458183289, "learning_rate": 0.00019904292366404995, "loss": 4.5585, "step": 176800 }, { "epoch": 0.330021763521725, "grad_norm": 1.1286976337432861, "learning_rate": 0.00019904234375107976, "loss": 4.6191, "step": 176850 }, { "epoch": 0.33011506908110355, "grad_norm": 0.9177277684211731, "learning_rate": 0.00019904176366331727, "loss": 4.698, "step": 176900 }, { "epoch": 0.330208374640482, "grad_norm": 0.9858579039573669, "learning_rate": 0.00019904118340076355, "loss": 4.6723, "step": 176950 }, { "epoch": 0.3303016801998605, "grad_norm": 1.172903299331665, "learning_rate": 0.00019904060296341947, "loss": 4.5899, "step": 177000 }, { "epoch": 0.3303016801998605, "eval_loss": 4.8154120445251465, "eval_runtime": 228.9954, "eval_samples_per_second": 11.389, "eval_steps_per_second": 11.389, "eval_tts_loss": 7.542438022716634, "step": 177000 }, { "epoch": 0.330394985759239, "grad_norm": 1.2343531847000122, "learning_rate": 0.00019904002235128623, "loss": 4.5554, "step": 177050 }, { "epoch": 0.3304882913186175, "grad_norm": 1.2554324865341187, "learning_rate": 0.00019903944156436477, "loss": 4.6312, "step": 177100 }, { "epoch": 0.330581596877996, "grad_norm": 1.062771201133728, "learning_rate": 0.0001990388606026561, "loss": 4.7273, "step": 177150 }, { "epoch": 0.33067490243737446, "grad_norm": 0.921607494354248, "learning_rate": 0.0001990382794661613, "loss": 4.5602, "step": 177200 }, { "epoch": 0.330768207996753, "grad_norm": 1.0047394037246704, "learning_rate": 0.00019903769815488132, "loss": 4.4348, "step": 177250 }, { "epoch": 0.33086151355613147, "grad_norm": 0.6674817204475403, "learning_rate": 0.00019903711666881724, "loss": 4.6014, "step": 177300 }, { "epoch": 0.33095481911550995, "grad_norm": 1.1274831295013428, "learning_rate": 0.0001990365350079701, "loss": 4.6538, "step": 177350 }, { "epoch": 0.3310481246748884, "grad_norm": 0.9511878490447998, "learning_rate": 0.00019903595317234088, "loss": 4.8159, "step": 177400 }, { "epoch": 0.33114143023426695, "grad_norm": 1.1247509717941284, "learning_rate": 0.0001990353711619306, "loss": 4.6459, "step": 177450 }, { "epoch": 0.33123473579364543, "grad_norm": 1.111100673675537, "learning_rate": 0.00019903478897674036, "loss": 4.6574, "step": 177500 }, { "epoch": 0.3313280413530239, "grad_norm": 0.9724700450897217, "learning_rate": 0.00019903420661677114, "loss": 4.6917, "step": 177550 }, { "epoch": 0.3314213469124024, "grad_norm": 1.2161903381347656, "learning_rate": 0.00019903362408202395, "loss": 4.6725, "step": 177600 }, { "epoch": 0.3315146524717809, "grad_norm": 0.881798267364502, "learning_rate": 0.00019903304137249984, "loss": 4.4201, "step": 177650 }, { "epoch": 0.3316079580311594, "grad_norm": 1.228102684020996, "learning_rate": 0.00019903245848819988, "loss": 4.6403, "step": 177700 }, { "epoch": 0.33170126359053786, "grad_norm": 0.9526916146278381, "learning_rate": 0.00019903187542912504, "loss": 4.5728, "step": 177750 }, { "epoch": 0.3317945691499164, "grad_norm": 0.8187074661254883, "learning_rate": 0.00019903129219527632, "loss": 4.6696, "step": 177800 }, { "epoch": 0.33188787470929487, "grad_norm": 0.9834941029548645, "learning_rate": 0.00019903070878665485, "loss": 4.4247, "step": 177850 }, { "epoch": 0.33198118026867335, "grad_norm": 1.224478006362915, "learning_rate": 0.00019903012520326155, "loss": 4.6612, "step": 177900 }, { "epoch": 0.3320744858280518, "grad_norm": 0.7598810791969299, "learning_rate": 0.00019902954144509753, "loss": 4.563, "step": 177950 }, { "epoch": 0.33216779138743036, "grad_norm": 1.0683156251907349, "learning_rate": 0.0001990289575121638, "loss": 4.6277, "step": 178000 }, { "epoch": 0.33226109694680883, "grad_norm": 1.1003293991088867, "learning_rate": 0.0001990283734044614, "loss": 4.7756, "step": 178050 }, { "epoch": 0.3323544025061873, "grad_norm": 0.9530490040779114, "learning_rate": 0.00019902778912199128, "loss": 4.5517, "step": 178100 }, { "epoch": 0.33244770806556584, "grad_norm": 0.7573721408843994, "learning_rate": 0.0001990272046647546, "loss": 4.4958, "step": 178150 }, { "epoch": 0.3325410136249443, "grad_norm": 1.020322561264038, "learning_rate": 0.00019902662003275232, "loss": 4.5234, "step": 178200 }, { "epoch": 0.3326343191843228, "grad_norm": 0.9242296814918518, "learning_rate": 0.00019902603522598547, "loss": 4.6549, "step": 178250 }, { "epoch": 0.33272762474370127, "grad_norm": 1.179311752319336, "learning_rate": 0.00019902545024445508, "loss": 4.5256, "step": 178300 }, { "epoch": 0.3328209303030798, "grad_norm": 1.0598257780075073, "learning_rate": 0.00019902486508816222, "loss": 4.7532, "step": 178350 }, { "epoch": 0.3329142358624583, "grad_norm": 0.7825391292572021, "learning_rate": 0.0001990242797571079, "loss": 4.6937, "step": 178400 }, { "epoch": 0.33300754142183675, "grad_norm": 1.1020742654800415, "learning_rate": 0.00019902369425129311, "loss": 4.6548, "step": 178450 }, { "epoch": 0.3331008469812153, "grad_norm": 0.9558370113372803, "learning_rate": 0.00019902310857071895, "loss": 4.8421, "step": 178500 }, { "epoch": 0.33319415254059376, "grad_norm": 0.8280293941497803, "learning_rate": 0.00019902252271538644, "loss": 4.6858, "step": 178550 }, { "epoch": 0.33328745809997223, "grad_norm": 0.8872970342636108, "learning_rate": 0.00019902193668529656, "loss": 4.4993, "step": 178600 }, { "epoch": 0.3333807636593507, "grad_norm": 1.1399357318878174, "learning_rate": 0.00019902135048045042, "loss": 4.7029, "step": 178650 }, { "epoch": 0.33347406921872924, "grad_norm": 1.1143829822540283, "learning_rate": 0.00019902076410084903, "loss": 4.6988, "step": 178700 }, { "epoch": 0.3335673747781077, "grad_norm": 1.012130856513977, "learning_rate": 0.00019902017754649338, "loss": 4.6413, "step": 178750 }, { "epoch": 0.3336606803374862, "grad_norm": 1.0381497144699097, "learning_rate": 0.00019901959081738456, "loss": 4.5631, "step": 178800 }, { "epoch": 0.3337539858968647, "grad_norm": 0.788299024105072, "learning_rate": 0.00019901900391352357, "loss": 4.5039, "step": 178850 }, { "epoch": 0.3338472914562432, "grad_norm": 0.8483543992042542, "learning_rate": 0.0001990184168349115, "loss": 4.7003, "step": 178900 }, { "epoch": 0.3339405970156217, "grad_norm": 0.7698935866355896, "learning_rate": 0.0001990178295815493, "loss": 4.5215, "step": 178950 }, { "epoch": 0.33403390257500015, "grad_norm": 1.0911551713943481, "learning_rate": 0.00019901724215343807, "loss": 4.5445, "step": 179000 }, { "epoch": 0.3341272081343787, "grad_norm": 0.824788510799408, "learning_rate": 0.00019901665455057882, "loss": 4.5568, "step": 179050 }, { "epoch": 0.33422051369375716, "grad_norm": 0.88285893201828, "learning_rate": 0.00019901606677297262, "loss": 4.5505, "step": 179100 }, { "epoch": 0.33431381925313564, "grad_norm": 0.9610635638237, "learning_rate": 0.00019901547882062047, "loss": 4.6942, "step": 179150 }, { "epoch": 0.33440712481251417, "grad_norm": 0.9467892050743103, "learning_rate": 0.0001990148906935234, "loss": 4.5793, "step": 179200 }, { "epoch": 0.33450043037189264, "grad_norm": 1.0193818807601929, "learning_rate": 0.00019901430239168254, "loss": 4.5554, "step": 179250 }, { "epoch": 0.3345937359312711, "grad_norm": 1.2325133085250854, "learning_rate": 0.0001990137139150988, "loss": 4.6663, "step": 179300 }, { "epoch": 0.3346870414906496, "grad_norm": 0.6962212324142456, "learning_rate": 0.00019901312526377329, "loss": 4.5595, "step": 179350 }, { "epoch": 0.3347803470500281, "grad_norm": 1.0818673372268677, "learning_rate": 0.00019901253643770704, "loss": 4.5388, "step": 179400 }, { "epoch": 0.3348736526094066, "grad_norm": 0.9971811175346375, "learning_rate": 0.0001990119474369011, "loss": 4.4762, "step": 179450 }, { "epoch": 0.3349669581687851, "grad_norm": 0.9982391595840454, "learning_rate": 0.00019901135826135645, "loss": 4.5818, "step": 179500 }, { "epoch": 0.3350602637281636, "grad_norm": 2.8972480297088623, "learning_rate": 0.0001990107689110742, "loss": 4.5753, "step": 179550 }, { "epoch": 0.3351535692875421, "grad_norm": 0.804068922996521, "learning_rate": 0.00019901017938605536, "loss": 4.8127, "step": 179600 }, { "epoch": 0.33524687484692056, "grad_norm": 0.8897950053215027, "learning_rate": 0.000199009589686301, "loss": 4.6659, "step": 179650 }, { "epoch": 0.33534018040629904, "grad_norm": 1.096867561340332, "learning_rate": 0.00019900899981181212, "loss": 4.6793, "step": 179700 }, { "epoch": 0.33543348596567757, "grad_norm": 0.9849563241004944, "learning_rate": 0.00019900840976258977, "loss": 4.6498, "step": 179750 }, { "epoch": 0.33552679152505605, "grad_norm": 1.0917383432388306, "learning_rate": 0.000199007819538635, "loss": 4.4898, "step": 179800 }, { "epoch": 0.3356200970844345, "grad_norm": 1.0085644721984863, "learning_rate": 0.00019900722913994884, "loss": 4.7175, "step": 179850 }, { "epoch": 0.33571340264381305, "grad_norm": 0.8432663679122925, "learning_rate": 0.00019900663856653233, "loss": 4.5698, "step": 179900 }, { "epoch": 0.33580670820319153, "grad_norm": 1.037923812866211, "learning_rate": 0.00019900604781838656, "loss": 4.6207, "step": 179950 }, { "epoch": 0.33590001376257, "grad_norm": 0.9545549154281616, "learning_rate": 0.0001990054568955125, "loss": 4.6257, "step": 180000 }, { "epoch": 0.33590001376257, "eval_loss": 4.818175315856934, "eval_runtime": 228.2553, "eval_samples_per_second": 11.426, "eval_steps_per_second": 11.426, "eval_tts_loss": 7.496765146657618, "step": 180000 }, { "epoch": 0.3359933193219485, "grad_norm": 0.8834198713302612, "learning_rate": 0.00019900486579791128, "loss": 4.5444, "step": 180050 }, { "epoch": 0.336086624881327, "grad_norm": 1.05471932888031, "learning_rate": 0.00019900427452558385, "loss": 4.5951, "step": 180100 }, { "epoch": 0.3361799304407055, "grad_norm": 1.0553511381149292, "learning_rate": 0.0001990036830785313, "loss": 4.6423, "step": 180150 }, { "epoch": 0.33627323600008396, "grad_norm": 0.8488613963127136, "learning_rate": 0.00019900309145675472, "loss": 4.8316, "step": 180200 }, { "epoch": 0.33636654155946244, "grad_norm": 0.8496327996253967, "learning_rate": 0.00019900249966025503, "loss": 4.5013, "step": 180250 }, { "epoch": 0.33645984711884097, "grad_norm": 1.085153579711914, "learning_rate": 0.0001990019076890334, "loss": 4.5181, "step": 180300 }, { "epoch": 0.33655315267821945, "grad_norm": 1.0941439867019653, "learning_rate": 0.0001990013155430908, "loss": 4.4706, "step": 180350 }, { "epoch": 0.3366464582375979, "grad_norm": 0.9088996052742004, "learning_rate": 0.00019900072322242827, "loss": 4.8162, "step": 180400 }, { "epoch": 0.33673976379697645, "grad_norm": 0.8877679705619812, "learning_rate": 0.00019900013072704693, "loss": 4.4462, "step": 180450 }, { "epoch": 0.33683306935635493, "grad_norm": 1.0370676517486572, "learning_rate": 0.0001989995380569478, "loss": 4.6237, "step": 180500 }, { "epoch": 0.3369263749157334, "grad_norm": 0.7951025366783142, "learning_rate": 0.00019899894521213183, "loss": 4.5673, "step": 180550 }, { "epoch": 0.3370196804751119, "grad_norm": 1.1994744539260864, "learning_rate": 0.00019899835219260018, "loss": 4.4731, "step": 180600 }, { "epoch": 0.3371129860344904, "grad_norm": 1.1491518020629883, "learning_rate": 0.00019899775899835388, "loss": 4.5261, "step": 180650 }, { "epoch": 0.3372062915938689, "grad_norm": 0.9701890349388123, "learning_rate": 0.0001989971656293939, "loss": 4.8223, "step": 180700 }, { "epoch": 0.33729959715324737, "grad_norm": 0.7716856002807617, "learning_rate": 0.00019899657208572138, "loss": 4.7053, "step": 180750 }, { "epoch": 0.3373929027126259, "grad_norm": 1.2134188413619995, "learning_rate": 0.00019899597836733733, "loss": 4.8142, "step": 180800 }, { "epoch": 0.3374862082720044, "grad_norm": 1.0805442333221436, "learning_rate": 0.00019899538447424277, "loss": 4.5745, "step": 180850 }, { "epoch": 0.33757951383138285, "grad_norm": 0.9581602215766907, "learning_rate": 0.0001989947904064388, "loss": 4.7698, "step": 180900 }, { "epoch": 0.3376728193907613, "grad_norm": 1.0755600929260254, "learning_rate": 0.00019899419616392644, "loss": 4.6043, "step": 180950 }, { "epoch": 0.33776612495013986, "grad_norm": 0.9674501419067383, "learning_rate": 0.00019899360174670673, "loss": 4.683, "step": 181000 }, { "epoch": 0.33785943050951833, "grad_norm": 0.8484556078910828, "learning_rate": 0.00019899300715478077, "loss": 4.6563, "step": 181050 }, { "epoch": 0.3379527360688968, "grad_norm": 0.9506059885025024, "learning_rate": 0.00019899241238814953, "loss": 4.5925, "step": 181100 }, { "epoch": 0.33804604162827534, "grad_norm": 1.1092076301574707, "learning_rate": 0.0001989918174468141, "loss": 4.689, "step": 181150 }, { "epoch": 0.3381393471876538, "grad_norm": 0.8176827430725098, "learning_rate": 0.00019899122233077554, "loss": 4.5785, "step": 181200 }, { "epoch": 0.3382326527470323, "grad_norm": 0.9832103252410889, "learning_rate": 0.00019899062704003488, "loss": 4.4976, "step": 181250 }, { "epoch": 0.33832595830641077, "grad_norm": 0.6890990734100342, "learning_rate": 0.00019899003157459322, "loss": 4.5854, "step": 181300 }, { "epoch": 0.3384192638657893, "grad_norm": 1.0396530628204346, "learning_rate": 0.00019898943593445154, "loss": 4.6799, "step": 181350 }, { "epoch": 0.3385125694251678, "grad_norm": 1.0980576276779175, "learning_rate": 0.00019898884011961094, "loss": 4.5739, "step": 181400 }, { "epoch": 0.33860587498454625, "grad_norm": 0.8679540753364563, "learning_rate": 0.00019898824413007243, "loss": 4.5388, "step": 181450 }, { "epoch": 0.3386991805439248, "grad_norm": 0.857035219669342, "learning_rate": 0.0001989876479658371, "loss": 4.4769, "step": 181500 }, { "epoch": 0.33879248610330326, "grad_norm": 0.9443413019180298, "learning_rate": 0.00019898705162690598, "loss": 4.4387, "step": 181550 }, { "epoch": 0.33888579166268173, "grad_norm": 1.0914325714111328, "learning_rate": 0.00019898645511328016, "loss": 4.8009, "step": 181600 }, { "epoch": 0.3389790972220602, "grad_norm": 0.9410198926925659, "learning_rate": 0.00019898585842496064, "loss": 4.7744, "step": 181650 }, { "epoch": 0.33907240278143874, "grad_norm": 0.9771658778190613, "learning_rate": 0.0001989852615619485, "loss": 4.505, "step": 181700 }, { "epoch": 0.3391657083408172, "grad_norm": 0.8809220194816589, "learning_rate": 0.0001989846645242448, "loss": 4.6773, "step": 181750 }, { "epoch": 0.3392590139001957, "grad_norm": 0.9671509265899658, "learning_rate": 0.00019898406731185057, "loss": 4.573, "step": 181800 }, { "epoch": 0.3393523194595742, "grad_norm": 1.0768080949783325, "learning_rate": 0.0001989834699247669, "loss": 4.8166, "step": 181850 }, { "epoch": 0.3394456250189527, "grad_norm": 0.941792905330658, "learning_rate": 0.00019898287236299478, "loss": 4.7192, "step": 181900 }, { "epoch": 0.3395389305783312, "grad_norm": 0.8942563533782959, "learning_rate": 0.00019898227462653533, "loss": 4.5277, "step": 181950 }, { "epoch": 0.33963223613770965, "grad_norm": 0.9910199642181396, "learning_rate": 0.00019898167671538957, "loss": 4.6392, "step": 182000 }, { "epoch": 0.3397255416970882, "grad_norm": 1.0308103561401367, "learning_rate": 0.0001989810786295586, "loss": 4.6123, "step": 182050 }, { "epoch": 0.33981884725646666, "grad_norm": 0.8738309144973755, "learning_rate": 0.00019898048036904342, "loss": 4.5485, "step": 182100 }, { "epoch": 0.33991215281584514, "grad_norm": 0.8555976152420044, "learning_rate": 0.00019897988193384513, "loss": 4.6508, "step": 182150 }, { "epoch": 0.34000545837522367, "grad_norm": 0.9000989198684692, "learning_rate": 0.00019897928332396477, "loss": 4.7006, "step": 182200 }, { "epoch": 0.34009876393460214, "grad_norm": 0.8779635429382324, "learning_rate": 0.00019897868453940337, "loss": 4.7278, "step": 182250 }, { "epoch": 0.3401920694939806, "grad_norm": 1.2050081491470337, "learning_rate": 0.00019897808558016198, "loss": 4.7024, "step": 182300 }, { "epoch": 0.3402853750533591, "grad_norm": 0.8261715173721313, "learning_rate": 0.0001989774864462417, "loss": 4.6134, "step": 182350 }, { "epoch": 0.3403786806127376, "grad_norm": 1.1884026527404785, "learning_rate": 0.0001989768871376436, "loss": 4.5208, "step": 182400 }, { "epoch": 0.3404719861721161, "grad_norm": 1.0583335161209106, "learning_rate": 0.00019897628765436868, "loss": 4.628, "step": 182450 }, { "epoch": 0.3405652917314946, "grad_norm": 0.8147032856941223, "learning_rate": 0.00019897568799641805, "loss": 4.5084, "step": 182500 }, { "epoch": 0.3406585972908731, "grad_norm": 0.9693218469619751, "learning_rate": 0.00019897508816379273, "loss": 4.7667, "step": 182550 }, { "epoch": 0.3407519028502516, "grad_norm": 0.9098610877990723, "learning_rate": 0.0001989744881564938, "loss": 4.8388, "step": 182600 }, { "epoch": 0.34084520840963006, "grad_norm": 1.03932785987854, "learning_rate": 0.00019897388797452233, "loss": 4.6463, "step": 182650 }, { "epoch": 0.34093851396900854, "grad_norm": 0.9611286520957947, "learning_rate": 0.00019897328761787932, "loss": 4.7354, "step": 182700 }, { "epoch": 0.34103181952838707, "grad_norm": 1.1163007020950317, "learning_rate": 0.00019897268708656592, "loss": 4.6021, "step": 182750 }, { "epoch": 0.34112512508776555, "grad_norm": 0.9765385985374451, "learning_rate": 0.0001989720863805831, "loss": 4.5286, "step": 182800 }, { "epoch": 0.341218430647144, "grad_norm": 0.8786355257034302, "learning_rate": 0.00019897148549993202, "loss": 4.6033, "step": 182850 }, { "epoch": 0.3413117362065225, "grad_norm": 1.1458057165145874, "learning_rate": 0.00019897088444461363, "loss": 4.7663, "step": 182900 }, { "epoch": 0.34140504176590103, "grad_norm": 0.8316869139671326, "learning_rate": 0.00019897028321462905, "loss": 4.5552, "step": 182950 }, { "epoch": 0.3414983473252795, "grad_norm": 0.9072234630584717, "learning_rate": 0.00019896968180997934, "loss": 4.6009, "step": 183000 }, { "epoch": 0.3414983473252795, "eval_loss": 4.803169250488281, "eval_runtime": 229.034, "eval_samples_per_second": 11.387, "eval_steps_per_second": 11.387, "eval_tts_loss": 7.459224440081708, "step": 183000 }, { "epoch": 0.341591652884658, "grad_norm": 1.083984613418579, "learning_rate": 0.00019896908023066557, "loss": 4.469, "step": 183050 }, { "epoch": 0.3416849584440365, "grad_norm": 0.8765878677368164, "learning_rate": 0.00019896847847668877, "loss": 4.4339, "step": 183100 }, { "epoch": 0.341778264003415, "grad_norm": 0.9148985743522644, "learning_rate": 0.00019896787654805005, "loss": 4.6406, "step": 183150 }, { "epoch": 0.34187156956279346, "grad_norm": 0.8526818156242371, "learning_rate": 0.00019896727444475038, "loss": 4.7993, "step": 183200 }, { "epoch": 0.34196487512217194, "grad_norm": 0.9056457877159119, "learning_rate": 0.00019896667216679094, "loss": 4.4859, "step": 183250 }, { "epoch": 0.34205818068155047, "grad_norm": 1.0208319425582886, "learning_rate": 0.00019896606971417274, "loss": 4.7113, "step": 183300 }, { "epoch": 0.34215148624092895, "grad_norm": 0.9874845743179321, "learning_rate": 0.0001989654670868968, "loss": 4.5711, "step": 183350 }, { "epoch": 0.3422447918003074, "grad_norm": 0.988519549369812, "learning_rate": 0.00019896486428496425, "loss": 4.6615, "step": 183400 }, { "epoch": 0.34233809735968596, "grad_norm": 1.0255403518676758, "learning_rate": 0.0001989642613083761, "loss": 4.7423, "step": 183450 }, { "epoch": 0.34243140291906443, "grad_norm": 0.8926201462745667, "learning_rate": 0.0001989636581571335, "loss": 4.5053, "step": 183500 }, { "epoch": 0.3425247084784429, "grad_norm": 1.1894657611846924, "learning_rate": 0.0001989630548312374, "loss": 4.6994, "step": 183550 }, { "epoch": 0.3426180140378214, "grad_norm": 1.066410779953003, "learning_rate": 0.00019896245133068895, "loss": 4.6081, "step": 183600 }, { "epoch": 0.3427113195971999, "grad_norm": 1.3439453840255737, "learning_rate": 0.00019896184765548917, "loss": 4.6694, "step": 183650 }, { "epoch": 0.3428046251565784, "grad_norm": 1.0843331813812256, "learning_rate": 0.00019896124380563916, "loss": 4.5804, "step": 183700 }, { "epoch": 0.34289793071595687, "grad_norm": 1.108058214187622, "learning_rate": 0.00019896063978113996, "loss": 4.5375, "step": 183750 }, { "epoch": 0.3429912362753354, "grad_norm": 1.012802004814148, "learning_rate": 0.00019896003558199263, "loss": 4.5888, "step": 183800 }, { "epoch": 0.3430845418347139, "grad_norm": 0.9825184941291809, "learning_rate": 0.0001989594312081983, "loss": 4.6243, "step": 183850 }, { "epoch": 0.34317784739409235, "grad_norm": 0.7795212268829346, "learning_rate": 0.0001989588266597579, "loss": 4.805, "step": 183900 }, { "epoch": 0.3432711529534708, "grad_norm": 1.6016143560409546, "learning_rate": 0.00019895822193667265, "loss": 4.6542, "step": 183950 }, { "epoch": 0.34336445851284936, "grad_norm": 1.0046799182891846, "learning_rate": 0.00019895761703894355, "loss": 4.5883, "step": 184000 }, { "epoch": 0.34345776407222783, "grad_norm": 0.7315368056297302, "learning_rate": 0.00019895701196657163, "loss": 4.6325, "step": 184050 }, { "epoch": 0.3435510696316063, "grad_norm": 0.8666321635246277, "learning_rate": 0.00019895640671955804, "loss": 4.5639, "step": 184100 }, { "epoch": 0.34364437519098484, "grad_norm": 1.0640748739242554, "learning_rate": 0.00019895580129790377, "loss": 4.5574, "step": 184150 }, { "epoch": 0.3437376807503633, "grad_norm": 0.8577877879142761, "learning_rate": 0.00019895519570160996, "loss": 4.6898, "step": 184200 }, { "epoch": 0.3438309863097418, "grad_norm": 0.7682492733001709, "learning_rate": 0.0001989545899306776, "loss": 4.7193, "step": 184250 }, { "epoch": 0.34392429186912027, "grad_norm": 1.053924322128296, "learning_rate": 0.00019895398398510778, "loss": 4.8604, "step": 184300 }, { "epoch": 0.3440175974284988, "grad_norm": 0.9362823367118835, "learning_rate": 0.00019895337786490162, "loss": 4.3165, "step": 184350 }, { "epoch": 0.3441109029878773, "grad_norm": 0.9843419194221497, "learning_rate": 0.0001989527715700602, "loss": 4.3989, "step": 184400 }, { "epoch": 0.34420420854725575, "grad_norm": 1.27715265750885, "learning_rate": 0.0001989521651005845, "loss": 4.6078, "step": 184450 }, { "epoch": 0.3442975141066343, "grad_norm": 0.9157127141952515, "learning_rate": 0.00019895155845647567, "loss": 4.4988, "step": 184500 }, { "epoch": 0.34439081966601276, "grad_norm": 0.9360002875328064, "learning_rate": 0.00019895095163773472, "loss": 4.6132, "step": 184550 }, { "epoch": 0.34448412522539124, "grad_norm": 1.044089674949646, "learning_rate": 0.00019895034464436276, "loss": 4.351, "step": 184600 }, { "epoch": 0.3445774307847697, "grad_norm": 1.0638103485107422, "learning_rate": 0.00019894973747636084, "loss": 4.7351, "step": 184650 }, { "epoch": 0.34467073634414824, "grad_norm": 1.2173103094100952, "learning_rate": 0.00019894913013373007, "loss": 4.4605, "step": 184700 }, { "epoch": 0.3447640419035267, "grad_norm": 0.8456254601478577, "learning_rate": 0.00019894852261647148, "loss": 4.8245, "step": 184750 }, { "epoch": 0.3448573474629052, "grad_norm": 1.3228527307510376, "learning_rate": 0.00019894791492458617, "loss": 4.6516, "step": 184800 }, { "epoch": 0.3449506530222837, "grad_norm": 0.8793355226516724, "learning_rate": 0.0001989473070580752, "loss": 4.7263, "step": 184850 }, { "epoch": 0.3450439585816622, "grad_norm": 0.733553409576416, "learning_rate": 0.00019894669901693966, "loss": 4.5974, "step": 184900 }, { "epoch": 0.3451372641410407, "grad_norm": 0.9788616895675659, "learning_rate": 0.00019894609080118057, "loss": 4.4838, "step": 184950 }, { "epoch": 0.34523056970041915, "grad_norm": 1.099247932434082, "learning_rate": 0.00019894548241079907, "loss": 4.5225, "step": 185000 }, { "epoch": 0.3453238752597977, "grad_norm": 0.9463627934455872, "learning_rate": 0.00019894487384579618, "loss": 4.8215, "step": 185050 }, { "epoch": 0.34541718081917616, "grad_norm": 0.9453304409980774, "learning_rate": 0.000198944265106173, "loss": 4.4883, "step": 185100 }, { "epoch": 0.34551048637855464, "grad_norm": 1.1768165826797485, "learning_rate": 0.0001989436561919306, "loss": 4.8437, "step": 185150 }, { "epoch": 0.34560379193793317, "grad_norm": 0.7917547821998596, "learning_rate": 0.00019894304710307007, "loss": 4.3636, "step": 185200 }, { "epoch": 0.34569709749731165, "grad_norm": 0.802765429019928, "learning_rate": 0.00019894243783959246, "loss": 4.6204, "step": 185250 }, { "epoch": 0.3457904030566901, "grad_norm": 0.9512039422988892, "learning_rate": 0.00019894182840149885, "loss": 4.4626, "step": 185300 }, { "epoch": 0.3458837086160686, "grad_norm": 0.9868209362030029, "learning_rate": 0.00019894121878879034, "loss": 4.8379, "step": 185350 }, { "epoch": 0.34597701417544713, "grad_norm": 1.1373767852783203, "learning_rate": 0.000198940609001468, "loss": 4.4935, "step": 185400 }, { "epoch": 0.3460703197348256, "grad_norm": 0.8839678168296814, "learning_rate": 0.00019893999903953288, "loss": 4.7252, "step": 185450 }, { "epoch": 0.3461636252942041, "grad_norm": 0.8720890283584595, "learning_rate": 0.00019893938890298607, "loss": 4.7323, "step": 185500 }, { "epoch": 0.34625693085358256, "grad_norm": 0.9795693159103394, "learning_rate": 0.00019893877859182864, "loss": 4.8515, "step": 185550 }, { "epoch": 0.3463502364129611, "grad_norm": 1.0126317739486694, "learning_rate": 0.00019893816810606166, "loss": 4.7654, "step": 185600 }, { "epoch": 0.34644354197233956, "grad_norm": 1.0032151937484741, "learning_rate": 0.00019893755744568626, "loss": 4.638, "step": 185650 }, { "epoch": 0.34653684753171804, "grad_norm": 0.7030245661735535, "learning_rate": 0.00019893694661070349, "loss": 4.7943, "step": 185700 }, { "epoch": 0.34663015309109657, "grad_norm": 0.9134154915809631, "learning_rate": 0.00019893633560111437, "loss": 4.7419, "step": 185750 }, { "epoch": 0.34672345865047505, "grad_norm": 0.6795088052749634, "learning_rate": 0.00019893572441692005, "loss": 4.593, "step": 185800 }, { "epoch": 0.3468167642098535, "grad_norm": 0.9457302093505859, "learning_rate": 0.00019893511305812159, "loss": 4.6045, "step": 185850 }, { "epoch": 0.346910069769232, "grad_norm": 1.1078051328659058, "learning_rate": 0.00019893450152472003, "loss": 4.7646, "step": 185900 }, { "epoch": 0.34700337532861053, "grad_norm": 0.8041667342185974, "learning_rate": 0.00019893388981671653, "loss": 4.6, "step": 185950 }, { "epoch": 0.347096680887989, "grad_norm": 1.0429846048355103, "learning_rate": 0.00019893327793411207, "loss": 4.6922, "step": 186000 }, { "epoch": 0.347096680887989, "eval_loss": 4.805553913116455, "eval_runtime": 227.9003, "eval_samples_per_second": 11.444, "eval_steps_per_second": 11.444, "eval_tts_loss": 7.4744320140360205, "step": 186000 }, { "epoch": 0.3471899864473675, "grad_norm": 1.188132643699646, "learning_rate": 0.0001989326658769078, "loss": 4.6133, "step": 186050 }, { "epoch": 0.347283292006746, "grad_norm": 1.105686902999878, "learning_rate": 0.00019893205364510482, "loss": 4.5588, "step": 186100 }, { "epoch": 0.3473765975661245, "grad_norm": 1.0260311365127563, "learning_rate": 0.00019893144123870414, "loss": 4.5501, "step": 186150 }, { "epoch": 0.34746990312550297, "grad_norm": 1.099698781967163, "learning_rate": 0.00019893082865770687, "loss": 4.6763, "step": 186200 }, { "epoch": 0.34756320868488144, "grad_norm": 1.0089246034622192, "learning_rate": 0.0001989302159021141, "loss": 4.6909, "step": 186250 }, { "epoch": 0.34765651424426, "grad_norm": 1.0701543092727661, "learning_rate": 0.00019892960297192692, "loss": 4.6315, "step": 186300 }, { "epoch": 0.34774981980363845, "grad_norm": 1.1907925605773926, "learning_rate": 0.00019892898986714638, "loss": 4.4744, "step": 186350 }, { "epoch": 0.3478431253630169, "grad_norm": 1.0035996437072754, "learning_rate": 0.00019892837658777362, "loss": 4.6013, "step": 186400 }, { "epoch": 0.34793643092239546, "grad_norm": 1.0162055492401123, "learning_rate": 0.00019892776313380965, "loss": 4.4274, "step": 186450 }, { "epoch": 0.34802973648177393, "grad_norm": 1.1313321590423584, "learning_rate": 0.0001989271495052556, "loss": 4.7547, "step": 186500 }, { "epoch": 0.3481230420411524, "grad_norm": 0.8484470248222351, "learning_rate": 0.0001989265357021125, "loss": 4.8249, "step": 186550 }, { "epoch": 0.3482163476005309, "grad_norm": 0.9785260558128357, "learning_rate": 0.00019892592172438152, "loss": 4.7419, "step": 186600 }, { "epoch": 0.3483096531599094, "grad_norm": 0.9270712733268738, "learning_rate": 0.00019892530757206368, "loss": 4.7485, "step": 186650 }, { "epoch": 0.3484029587192879, "grad_norm": 1.1253376007080078, "learning_rate": 0.00019892469324516004, "loss": 4.6492, "step": 186700 }, { "epoch": 0.34849626427866637, "grad_norm": 0.8123875856399536, "learning_rate": 0.00019892407874367175, "loss": 4.5601, "step": 186750 }, { "epoch": 0.3485895698380449, "grad_norm": 0.9842862486839294, "learning_rate": 0.00019892346406759988, "loss": 4.6747, "step": 186800 }, { "epoch": 0.3486828753974234, "grad_norm": 1.0450785160064697, "learning_rate": 0.0001989228492169455, "loss": 4.6732, "step": 186850 }, { "epoch": 0.34877618095680185, "grad_norm": 1.2406466007232666, "learning_rate": 0.00019892223419170968, "loss": 4.5914, "step": 186900 }, { "epoch": 0.3488694865161803, "grad_norm": 1.202197790145874, "learning_rate": 0.0001989216189918935, "loss": 4.4839, "step": 186950 }, { "epoch": 0.34896279207555886, "grad_norm": 0.9749503135681152, "learning_rate": 0.00019892100361749813, "loss": 4.5923, "step": 187000 }, { "epoch": 0.34905609763493733, "grad_norm": 0.7598263621330261, "learning_rate": 0.00019892038806852455, "loss": 4.582, "step": 187050 }, { "epoch": 0.3491494031943158, "grad_norm": 1.0366371870040894, "learning_rate": 0.00019891977234497393, "loss": 4.6446, "step": 187100 }, { "epoch": 0.34924270875369434, "grad_norm": 0.82780522108078, "learning_rate": 0.0001989191564468473, "loss": 4.5483, "step": 187150 }, { "epoch": 0.3493360143130728, "grad_norm": 1.048930287361145, "learning_rate": 0.00019891854037414573, "loss": 4.6858, "step": 187200 }, { "epoch": 0.3494293198724513, "grad_norm": 1.0502203702926636, "learning_rate": 0.00019891792412687038, "loss": 4.821, "step": 187250 }, { "epoch": 0.34952262543182977, "grad_norm": 1.079811930656433, "learning_rate": 0.00019891730770502228, "loss": 4.6184, "step": 187300 }, { "epoch": 0.3496159309912083, "grad_norm": 0.9850397109985352, "learning_rate": 0.00019891669110860255, "loss": 4.7636, "step": 187350 }, { "epoch": 0.3497092365505868, "grad_norm": 1.421713948249817, "learning_rate": 0.00019891607433761226, "loss": 4.8562, "step": 187400 }, { "epoch": 0.34980254210996525, "grad_norm": 0.9206252098083496, "learning_rate": 0.00019891545739205249, "loss": 4.47, "step": 187450 }, { "epoch": 0.3498958476693438, "grad_norm": 0.9205539226531982, "learning_rate": 0.00019891484027192435, "loss": 4.5468, "step": 187500 }, { "epoch": 0.34998915322872226, "grad_norm": 1.1044520139694214, "learning_rate": 0.00019891422297722893, "loss": 4.7065, "step": 187550 }, { "epoch": 0.35008245878810074, "grad_norm": 1.2163364887237549, "learning_rate": 0.0001989136055079673, "loss": 4.5895, "step": 187600 }, { "epoch": 0.3501757643474792, "grad_norm": 1.0337022542953491, "learning_rate": 0.00019891298786414057, "loss": 4.752, "step": 187650 }, { "epoch": 0.35026906990685774, "grad_norm": 0.9887984395027161, "learning_rate": 0.0001989123700457498, "loss": 4.498, "step": 187700 }, { "epoch": 0.3503623754662362, "grad_norm": 0.8707548975944519, "learning_rate": 0.00019891175205279614, "loss": 4.4742, "step": 187750 }, { "epoch": 0.3504556810256147, "grad_norm": 0.9432954788208008, "learning_rate": 0.00019891113388528062, "loss": 4.7022, "step": 187800 }, { "epoch": 0.3505489865849932, "grad_norm": 1.063997745513916, "learning_rate": 0.00019891051554320438, "loss": 4.4942, "step": 187850 }, { "epoch": 0.3506422921443717, "grad_norm": 1.1391524076461792, "learning_rate": 0.00019890989702656844, "loss": 4.6385, "step": 187900 }, { "epoch": 0.3507355977037502, "grad_norm": 1.218582034111023, "learning_rate": 0.00019890927833537397, "loss": 4.665, "step": 187950 }, { "epoch": 0.35082890326312866, "grad_norm": 0.8306301832199097, "learning_rate": 0.00019890865946962198, "loss": 4.7438, "step": 188000 }, { "epoch": 0.3509222088225072, "grad_norm": 0.9843810796737671, "learning_rate": 0.00019890804042931362, "loss": 4.5901, "step": 188050 }, { "epoch": 0.35101551438188566, "grad_norm": 0.9525366425514221, "learning_rate": 0.00019890742121445002, "loss": 4.6783, "step": 188100 }, { "epoch": 0.35110881994126414, "grad_norm": 1.0917189121246338, "learning_rate": 0.00019890680182503218, "loss": 4.6924, "step": 188150 }, { "epoch": 0.3512021255006426, "grad_norm": 1.1915470361709595, "learning_rate": 0.00019890618226106125, "loss": 4.4894, "step": 188200 }, { "epoch": 0.35129543106002115, "grad_norm": 1.0157119035720825, "learning_rate": 0.00019890556252253832, "loss": 4.5125, "step": 188250 }, { "epoch": 0.3513887366193996, "grad_norm": 1.2147808074951172, "learning_rate": 0.0001989049426094645, "loss": 4.474, "step": 188300 }, { "epoch": 0.3514820421787781, "grad_norm": 0.9446842074394226, "learning_rate": 0.00019890432252184082, "loss": 4.6241, "step": 188350 }, { "epoch": 0.35157534773815663, "grad_norm": 0.6960589289665222, "learning_rate": 0.00019890370225966844, "loss": 4.5706, "step": 188400 }, { "epoch": 0.3516686532975351, "grad_norm": 0.8260095119476318, "learning_rate": 0.00019890308182294838, "loss": 4.7308, "step": 188450 }, { "epoch": 0.3517619588569136, "grad_norm": 0.9620236158370972, "learning_rate": 0.00019890246121168184, "loss": 4.5599, "step": 188500 }, { "epoch": 0.35185526441629206, "grad_norm": 0.8720660209655762, "learning_rate": 0.00019890184042586985, "loss": 4.5758, "step": 188550 }, { "epoch": 0.3519485699756706, "grad_norm": 1.1599175930023193, "learning_rate": 0.00019890121946551347, "loss": 4.7073, "step": 188600 }, { "epoch": 0.35204187553504906, "grad_norm": 1.200167179107666, "learning_rate": 0.00019890059833061388, "loss": 4.6389, "step": 188650 }, { "epoch": 0.35213518109442754, "grad_norm": 0.7152448296546936, "learning_rate": 0.0001988999770211721, "loss": 4.6249, "step": 188700 }, { "epoch": 0.35222848665380607, "grad_norm": 0.6849294900894165, "learning_rate": 0.0001988993555371893, "loss": 4.6701, "step": 188750 }, { "epoch": 0.35232179221318455, "grad_norm": 1.2049897909164429, "learning_rate": 0.00019889873387866653, "loss": 4.6354, "step": 188800 }, { "epoch": 0.352415097772563, "grad_norm": 0.9554798603057861, "learning_rate": 0.0001988981120456049, "loss": 4.4661, "step": 188850 }, { "epoch": 0.3525084033319415, "grad_norm": 1.052039623260498, "learning_rate": 0.0001988974900380055, "loss": 4.6176, "step": 188900 }, { "epoch": 0.35260170889132003, "grad_norm": 1.091506838798523, "learning_rate": 0.00019889686785586942, "loss": 4.801, "step": 188950 }, { "epoch": 0.3526950144506985, "grad_norm": 1.3394882678985596, "learning_rate": 0.00019889624549919777, "loss": 4.722, "step": 189000 }, { "epoch": 0.3526950144506985, "eval_loss": 4.806208610534668, "eval_runtime": 230.3394, "eval_samples_per_second": 11.322, "eval_steps_per_second": 11.322, "eval_tts_loss": 7.53537485083748, "step": 189000 }, { "epoch": 0.352788320010077, "grad_norm": 0.8463157415390015, "learning_rate": 0.00019889562296799166, "loss": 4.6327, "step": 189050 }, { "epoch": 0.3528816255694555, "grad_norm": 1.0070785284042358, "learning_rate": 0.00019889500026225217, "loss": 4.6743, "step": 189100 }, { "epoch": 0.352974931128834, "grad_norm": 0.8200750946998596, "learning_rate": 0.0001988943773819804, "loss": 4.3624, "step": 189150 }, { "epoch": 0.35306823668821247, "grad_norm": 0.9505131840705872, "learning_rate": 0.00019889375432717747, "loss": 4.8476, "step": 189200 }, { "epoch": 0.35316154224759094, "grad_norm": 0.9525319337844849, "learning_rate": 0.00019889313109784445, "loss": 4.511, "step": 189250 }, { "epoch": 0.3532548478069695, "grad_norm": 0.9823152422904968, "learning_rate": 0.00019889250769398247, "loss": 4.6123, "step": 189300 }, { "epoch": 0.35334815336634795, "grad_norm": 1.0408767461776733, "learning_rate": 0.0001988918841155926, "loss": 4.6926, "step": 189350 }, { "epoch": 0.3534414589257264, "grad_norm": 1.0379136800765991, "learning_rate": 0.00019889126036267595, "loss": 4.6514, "step": 189400 }, { "epoch": 0.35353476448510496, "grad_norm": 0.7835113406181335, "learning_rate": 0.00019889063643523366, "loss": 4.5634, "step": 189450 }, { "epoch": 0.35362807004448343, "grad_norm": 1.5030099153518677, "learning_rate": 0.00019889001233326677, "loss": 4.5985, "step": 189500 }, { "epoch": 0.3537213756038619, "grad_norm": 0.9151412844657898, "learning_rate": 0.0001988893880567764, "loss": 4.7988, "step": 189550 }, { "epoch": 0.3538146811632404, "grad_norm": 0.9782933592796326, "learning_rate": 0.00019888876360576367, "loss": 4.6997, "step": 189600 }, { "epoch": 0.3539079867226189, "grad_norm": 0.9720525145530701, "learning_rate": 0.00019888813898022968, "loss": 4.8163, "step": 189650 }, { "epoch": 0.3540012922819974, "grad_norm": 1.0412356853485107, "learning_rate": 0.00019888751418017554, "loss": 4.7658, "step": 189700 }, { "epoch": 0.35409459784137587, "grad_norm": 0.5875607132911682, "learning_rate": 0.0001988868892056023, "loss": 4.8729, "step": 189750 }, { "epoch": 0.3541879034007544, "grad_norm": 1.0475389957427979, "learning_rate": 0.00019888626405651112, "loss": 4.517, "step": 189800 }, { "epoch": 0.3542812089601329, "grad_norm": 1.3192886114120483, "learning_rate": 0.0001988856387329031, "loss": 4.5192, "step": 189850 }, { "epoch": 0.35437451451951135, "grad_norm": 1.179880142211914, "learning_rate": 0.00019888501323477928, "loss": 4.4085, "step": 189900 }, { "epoch": 0.35446782007888983, "grad_norm": 1.334714651107788, "learning_rate": 0.00019888438756214086, "loss": 4.7391, "step": 189950 }, { "epoch": 0.35456112563826836, "grad_norm": 1.2655901908874512, "learning_rate": 0.00019888376171498887, "loss": 4.756, "step": 190000 }, { "epoch": 0.35465443119764684, "grad_norm": 1.0103163719177246, "learning_rate": 0.00019888313569332443, "loss": 4.6622, "step": 190050 }, { "epoch": 0.3547477367570253, "grad_norm": 0.8582555055618286, "learning_rate": 0.00019888250949714865, "loss": 4.4321, "step": 190100 }, { "epoch": 0.35484104231640384, "grad_norm": 1.1723741292953491, "learning_rate": 0.00019888188312646267, "loss": 4.5144, "step": 190150 }, { "epoch": 0.3549343478757823, "grad_norm": 1.05409574508667, "learning_rate": 0.00019888125658126758, "loss": 4.618, "step": 190200 }, { "epoch": 0.3550276534351608, "grad_norm": 1.0219589471817017, "learning_rate": 0.00019888062986156443, "loss": 4.632, "step": 190250 }, { "epoch": 0.35512095899453927, "grad_norm": 1.1344385147094727, "learning_rate": 0.00019888000296735436, "loss": 4.4802, "step": 190300 }, { "epoch": 0.3552142645539178, "grad_norm": 1.0866692066192627, "learning_rate": 0.00019887937589863853, "loss": 4.6449, "step": 190350 }, { "epoch": 0.3553075701132963, "grad_norm": 1.0079729557037354, "learning_rate": 0.00019887874865541795, "loss": 4.6405, "step": 190400 }, { "epoch": 0.35540087567267475, "grad_norm": 1.0848729610443115, "learning_rate": 0.00019887812123769384, "loss": 4.6839, "step": 190450 }, { "epoch": 0.3554941812320533, "grad_norm": 0.8781519532203674, "learning_rate": 0.00019887749364546717, "loss": 4.5597, "step": 190500 }, { "epoch": 0.35558748679143176, "grad_norm": 0.9141809940338135, "learning_rate": 0.00019887686587873916, "loss": 4.577, "step": 190550 }, { "epoch": 0.35568079235081024, "grad_norm": 0.9193921685218811, "learning_rate": 0.00019887623793751088, "loss": 4.6156, "step": 190600 }, { "epoch": 0.3557740979101887, "grad_norm": 0.9867975115776062, "learning_rate": 0.00019887560982178343, "loss": 4.4942, "step": 190650 }, { "epoch": 0.35586740346956725, "grad_norm": 0.9040959477424622, "learning_rate": 0.00019887498153155794, "loss": 4.8029, "step": 190700 }, { "epoch": 0.3559607090289457, "grad_norm": 0.9523190855979919, "learning_rate": 0.00019887435306683549, "loss": 4.5111, "step": 190750 }, { "epoch": 0.3560540145883242, "grad_norm": 1.135631799697876, "learning_rate": 0.00019887372442761722, "loss": 4.7086, "step": 190800 }, { "epoch": 0.3561473201477027, "grad_norm": 1.1821600198745728, "learning_rate": 0.0001988730956139042, "loss": 4.524, "step": 190850 }, { "epoch": 0.3562406257070812, "grad_norm": 1.0236471891403198, "learning_rate": 0.00019887246662569758, "loss": 4.7486, "step": 190900 }, { "epoch": 0.3563339312664597, "grad_norm": 1.091172695159912, "learning_rate": 0.00019887183746299846, "loss": 4.572, "step": 190950 }, { "epoch": 0.35642723682583816, "grad_norm": 1.0206681489944458, "learning_rate": 0.00019887120812580792, "loss": 4.6126, "step": 191000 }, { "epoch": 0.3565205423852167, "grad_norm": 1.0186724662780762, "learning_rate": 0.00019887057861412712, "loss": 4.5049, "step": 191050 }, { "epoch": 0.35661384794459516, "grad_norm": 0.817267656326294, "learning_rate": 0.0001988699489279571, "loss": 4.6261, "step": 191100 }, { "epoch": 0.35670715350397364, "grad_norm": 1.2809371948242188, "learning_rate": 0.0001988693190672991, "loss": 4.7523, "step": 191150 }, { "epoch": 0.3568004590633521, "grad_norm": 0.6853193044662476, "learning_rate": 0.00019886868903215408, "loss": 4.4515, "step": 191200 }, { "epoch": 0.35689376462273065, "grad_norm": 1.1701523065567017, "learning_rate": 0.00019886805882252324, "loss": 4.6187, "step": 191250 }, { "epoch": 0.3569870701821091, "grad_norm": 0.8031530976295471, "learning_rate": 0.0001988674284384077, "loss": 4.5504, "step": 191300 }, { "epoch": 0.3570803757414876, "grad_norm": 0.855839192867279, "learning_rate": 0.0001988667978798085, "loss": 4.3099, "step": 191350 }, { "epoch": 0.35717368130086613, "grad_norm": 1.233931541442871, "learning_rate": 0.00019886616714672681, "loss": 4.5372, "step": 191400 }, { "epoch": 0.3572669868602446, "grad_norm": 1.1093002557754517, "learning_rate": 0.00019886553623916372, "loss": 4.8146, "step": 191450 }, { "epoch": 0.3573602924196231, "grad_norm": 1.2051666975021362, "learning_rate": 0.00019886490515712038, "loss": 4.7033, "step": 191500 }, { "epoch": 0.35745359797900156, "grad_norm": 1.0855170488357544, "learning_rate": 0.00019886427390059788, "loss": 4.5491, "step": 191550 }, { "epoch": 0.3575469035383801, "grad_norm": 1.0766313076019287, "learning_rate": 0.00019886364246959732, "loss": 4.5545, "step": 191600 }, { "epoch": 0.35764020909775857, "grad_norm": 0.8042473793029785, "learning_rate": 0.00019886301086411982, "loss": 4.8249, "step": 191650 }, { "epoch": 0.35773351465713704, "grad_norm": 0.9476457834243774, "learning_rate": 0.00019886237908416652, "loss": 4.8088, "step": 191700 }, { "epoch": 0.3578268202165156, "grad_norm": 1.0122379064559937, "learning_rate": 0.0001988617471297385, "loss": 4.4633, "step": 191750 }, { "epoch": 0.35792012577589405, "grad_norm": 1.0943959951400757, "learning_rate": 0.00019886111500083692, "loss": 4.4537, "step": 191800 }, { "epoch": 0.3580134313352725, "grad_norm": 0.9375251531600952, "learning_rate": 0.00019886048269746281, "loss": 4.7861, "step": 191850 }, { "epoch": 0.358106736894651, "grad_norm": 1.0100510120391846, "learning_rate": 0.00019885985021961738, "loss": 4.6012, "step": 191900 }, { "epoch": 0.35820004245402953, "grad_norm": 1.1179404258728027, "learning_rate": 0.00019885921756730173, "loss": 4.771, "step": 191950 }, { "epoch": 0.358293348013408, "grad_norm": 1.1079944372177124, "learning_rate": 0.00019885858474051694, "loss": 4.5788, "step": 192000 }, { "epoch": 0.358293348013408, "eval_loss": 4.794641971588135, "eval_runtime": 230.5927, "eval_samples_per_second": 11.31, "eval_steps_per_second": 11.31, "eval_tts_loss": 7.481033778209468, "step": 192000 }, { "epoch": 0.3583866535727865, "grad_norm": 1.0501129627227783, "learning_rate": 0.00019885795173926412, "loss": 4.589, "step": 192050 }, { "epoch": 0.358479959132165, "grad_norm": 0.9563352465629578, "learning_rate": 0.00019885731856354442, "loss": 4.7751, "step": 192100 }, { "epoch": 0.3585732646915435, "grad_norm": 1.539056658744812, "learning_rate": 0.00019885668521335898, "loss": 4.6909, "step": 192150 }, { "epoch": 0.35866657025092197, "grad_norm": 1.1265029907226562, "learning_rate": 0.00019885605168870884, "loss": 4.4262, "step": 192200 }, { "epoch": 0.35875987581030044, "grad_norm": 0.973284900188446, "learning_rate": 0.00019885541798959517, "loss": 4.8141, "step": 192250 }, { "epoch": 0.358853181369679, "grad_norm": 0.6788082718849182, "learning_rate": 0.0001988547841160191, "loss": 4.6917, "step": 192300 }, { "epoch": 0.35894648692905745, "grad_norm": 0.7461669445037842, "learning_rate": 0.0001988541500679817, "loss": 4.7468, "step": 192350 }, { "epoch": 0.3590397924884359, "grad_norm": 1.1224901676177979, "learning_rate": 0.00019885351584548414, "loss": 4.5873, "step": 192400 }, { "epoch": 0.35913309804781446, "grad_norm": 1.1169931888580322, "learning_rate": 0.00019885288144852754, "loss": 4.7951, "step": 192450 }, { "epoch": 0.35922640360719293, "grad_norm": 0.9802398085594177, "learning_rate": 0.00019885224687711298, "loss": 4.5911, "step": 192500 }, { "epoch": 0.3593197091665714, "grad_norm": 0.9591051340103149, "learning_rate": 0.00019885161213124158, "loss": 4.4823, "step": 192550 }, { "epoch": 0.3594130147259499, "grad_norm": 1.2416292428970337, "learning_rate": 0.0001988509772109145, "loss": 4.7929, "step": 192600 }, { "epoch": 0.3595063202853284, "grad_norm": 1.196144461631775, "learning_rate": 0.00019885034211613282, "loss": 4.6477, "step": 192650 }, { "epoch": 0.3595996258447069, "grad_norm": 1.0120468139648438, "learning_rate": 0.0001988497068468977, "loss": 4.7484, "step": 192700 }, { "epoch": 0.35969293140408537, "grad_norm": 0.7862088680267334, "learning_rate": 0.00019884907140321024, "loss": 4.5306, "step": 192750 }, { "epoch": 0.3597862369634639, "grad_norm": 1.0786339044570923, "learning_rate": 0.00019884843578507155, "loss": 4.861, "step": 192800 }, { "epoch": 0.3598795425228424, "grad_norm": 1.0364389419555664, "learning_rate": 0.00019884779999248278, "loss": 4.6999, "step": 192850 }, { "epoch": 0.35997284808222085, "grad_norm": 0.887571394443512, "learning_rate": 0.00019884716402544502, "loss": 4.6757, "step": 192900 }, { "epoch": 0.36006615364159933, "grad_norm": 0.6657192707061768, "learning_rate": 0.00019884652788395942, "loss": 4.5924, "step": 192950 }, { "epoch": 0.36015945920097786, "grad_norm": 1.246086597442627, "learning_rate": 0.00019884589156802707, "loss": 4.5792, "step": 193000 }, { "epoch": 0.36025276476035634, "grad_norm": 0.8808768391609192, "learning_rate": 0.00019884525507764912, "loss": 4.6044, "step": 193050 }, { "epoch": 0.3603460703197348, "grad_norm": 0.8309186697006226, "learning_rate": 0.00019884461841282672, "loss": 4.6616, "step": 193100 }, { "epoch": 0.36043937587911334, "grad_norm": 0.9923182129859924, "learning_rate": 0.0001988439815735609, "loss": 4.6831, "step": 193150 }, { "epoch": 0.3605326814384918, "grad_norm": 0.8415591716766357, "learning_rate": 0.0001988433445598529, "loss": 4.648, "step": 193200 }, { "epoch": 0.3606259869978703, "grad_norm": 1.2085046768188477, "learning_rate": 0.00019884270737170377, "loss": 4.429, "step": 193250 }, { "epoch": 0.36071929255724877, "grad_norm": 0.8806056976318359, "learning_rate": 0.00019884207000911465, "loss": 4.4001, "step": 193300 }, { "epoch": 0.3608125981166273, "grad_norm": 1.0138684511184692, "learning_rate": 0.00019884143247208665, "loss": 4.6307, "step": 193350 }, { "epoch": 0.3609059036760058, "grad_norm": 0.8585569858551025, "learning_rate": 0.00019884079476062094, "loss": 4.5959, "step": 193400 }, { "epoch": 0.36099920923538426, "grad_norm": 0.9443112015724182, "learning_rate": 0.00019884015687471863, "loss": 4.7276, "step": 193450 }, { "epoch": 0.3610925147947628, "grad_norm": 1.2051678895950317, "learning_rate": 0.00019883951881438078, "loss": 4.795, "step": 193500 }, { "epoch": 0.36118582035414126, "grad_norm": 0.7949090600013733, "learning_rate": 0.00019883888057960864, "loss": 4.7397, "step": 193550 }, { "epoch": 0.36127912591351974, "grad_norm": 0.9525768756866455, "learning_rate": 0.00019883824217040323, "loss": 4.6333, "step": 193600 }, { "epoch": 0.3613724314728982, "grad_norm": 1.2094985246658325, "learning_rate": 0.0001988376035867657, "loss": 4.5322, "step": 193650 }, { "epoch": 0.36146573703227675, "grad_norm": 1.0645558834075928, "learning_rate": 0.00019883696482869722, "loss": 4.7803, "step": 193700 }, { "epoch": 0.3615590425916552, "grad_norm": 1.0554648637771606, "learning_rate": 0.00019883632589619887, "loss": 4.5194, "step": 193750 }, { "epoch": 0.3616523481510337, "grad_norm": 0.8595163822174072, "learning_rate": 0.0001988356867892718, "loss": 4.6625, "step": 193800 }, { "epoch": 0.3617456537104122, "grad_norm": 0.9319025874137878, "learning_rate": 0.00019883504750791713, "loss": 4.7127, "step": 193850 }, { "epoch": 0.3618389592697907, "grad_norm": 1.2545781135559082, "learning_rate": 0.000198834408052136, "loss": 4.409, "step": 193900 }, { "epoch": 0.3619322648291692, "grad_norm": 1.0958348512649536, "learning_rate": 0.00019883376842192953, "loss": 4.629, "step": 193950 }, { "epoch": 0.36202557038854766, "grad_norm": 1.1290863752365112, "learning_rate": 0.00019883312861729887, "loss": 4.7757, "step": 194000 }, { "epoch": 0.3621188759479262, "grad_norm": 1.168502926826477, "learning_rate": 0.0001988324886382451, "loss": 4.5306, "step": 194050 }, { "epoch": 0.36221218150730466, "grad_norm": 1.0988664627075195, "learning_rate": 0.00019883184848476938, "loss": 4.5708, "step": 194100 }, { "epoch": 0.36230548706668314, "grad_norm": 1.10366952419281, "learning_rate": 0.00019883120815687285, "loss": 4.7049, "step": 194150 }, { "epoch": 0.3623987926260616, "grad_norm": 1.0056383609771729, "learning_rate": 0.0001988305676545566, "loss": 4.8057, "step": 194200 }, { "epoch": 0.36249209818544015, "grad_norm": 0.930894672870636, "learning_rate": 0.00019882992697782182, "loss": 4.5286, "step": 194250 }, { "epoch": 0.3625854037448186, "grad_norm": 1.0776501893997192, "learning_rate": 0.0001988292861266696, "loss": 4.56, "step": 194300 }, { "epoch": 0.3626787093041971, "grad_norm": 1.0516318082809448, "learning_rate": 0.00019882864510110105, "loss": 4.5943, "step": 194350 }, { "epoch": 0.36277201486357563, "grad_norm": 1.0238537788391113, "learning_rate": 0.00019882800390111735, "loss": 4.6839, "step": 194400 }, { "epoch": 0.3628653204229541, "grad_norm": 1.1146060228347778, "learning_rate": 0.00019882736252671963, "loss": 4.6616, "step": 194450 }, { "epoch": 0.3629586259823326, "grad_norm": 1.2036668062210083, "learning_rate": 0.00019882672097790898, "loss": 4.5828, "step": 194500 }, { "epoch": 0.36305193154171106, "grad_norm": 0.9518182873725891, "learning_rate": 0.00019882607925468656, "loss": 4.5861, "step": 194550 }, { "epoch": 0.3631452371010896, "grad_norm": 0.8470454812049866, "learning_rate": 0.0001988254373570535, "loss": 4.7907, "step": 194600 }, { "epoch": 0.36323854266046807, "grad_norm": 0.9194369316101074, "learning_rate": 0.00019882479528501095, "loss": 4.4232, "step": 194650 }, { "epoch": 0.36333184821984654, "grad_norm": 1.164222002029419, "learning_rate": 0.00019882415303856002, "loss": 4.4877, "step": 194700 }, { "epoch": 0.3634251537792251, "grad_norm": 0.9167250394821167, "learning_rate": 0.00019882351061770184, "loss": 4.5728, "step": 194750 }, { "epoch": 0.36351845933860355, "grad_norm": 0.8055945634841919, "learning_rate": 0.00019882286802243754, "loss": 4.6115, "step": 194800 }, { "epoch": 0.363611764897982, "grad_norm": 0.8844305872917175, "learning_rate": 0.0001988222252527683, "loss": 4.6615, "step": 194850 }, { "epoch": 0.3637050704573605, "grad_norm": 1.018107533454895, "learning_rate": 0.0001988215823086952, "loss": 4.6986, "step": 194900 }, { "epoch": 0.36379837601673903, "grad_norm": 0.80892413854599, "learning_rate": 0.00019882093919021938, "loss": 4.5929, "step": 194950 }, { "epoch": 0.3638916815761175, "grad_norm": 0.9718637466430664, "learning_rate": 0.00019882029589734203, "loss": 4.5956, "step": 195000 }, { "epoch": 0.3638916815761175, "eval_loss": 4.797578811645508, "eval_runtime": 230.7126, "eval_samples_per_second": 11.304, "eval_steps_per_second": 11.304, "eval_tts_loss": 7.486753235031148, "step": 195000 }, { "epoch": 0.363984987135496, "grad_norm": 0.9084148406982422, "learning_rate": 0.0001988196524300642, "loss": 4.5751, "step": 195050 }, { "epoch": 0.3640782926948745, "grad_norm": 1.0336567163467407, "learning_rate": 0.0001988190087883871, "loss": 4.7999, "step": 195100 }, { "epoch": 0.364171598254253, "grad_norm": 0.9693698287010193, "learning_rate": 0.00019881836497231183, "loss": 4.3249, "step": 195150 }, { "epoch": 0.36426490381363147, "grad_norm": 1.1222256422042847, "learning_rate": 0.00019881772098183956, "loss": 4.587, "step": 195200 }, { "epoch": 0.36435820937300994, "grad_norm": 1.0114097595214844, "learning_rate": 0.00019881707681697135, "loss": 4.6811, "step": 195250 }, { "epoch": 0.3644515149323885, "grad_norm": 0.9249664545059204, "learning_rate": 0.00019881643247770842, "loss": 4.8543, "step": 195300 }, { "epoch": 0.36454482049176695, "grad_norm": 1.0353797674179077, "learning_rate": 0.00019881578796405186, "loss": 4.6228, "step": 195350 }, { "epoch": 0.36463812605114543, "grad_norm": 1.0884183645248413, "learning_rate": 0.00019881514327600282, "loss": 4.7143, "step": 195400 }, { "epoch": 0.36473143161052396, "grad_norm": 1.049566388130188, "learning_rate": 0.00019881449841356247, "loss": 4.6686, "step": 195450 }, { "epoch": 0.36482473716990244, "grad_norm": 0.990344226360321, "learning_rate": 0.00019881385337673188, "loss": 4.585, "step": 195500 }, { "epoch": 0.3649180427292809, "grad_norm": 1.0016299486160278, "learning_rate": 0.00019881320816551225, "loss": 4.5803, "step": 195550 }, { "epoch": 0.3650113482886594, "grad_norm": 0.9439250826835632, "learning_rate": 0.0001988125627799047, "loss": 4.5019, "step": 195600 }, { "epoch": 0.3651046538480379, "grad_norm": 1.0324641466140747, "learning_rate": 0.00019881191721991034, "loss": 4.708, "step": 195650 }, { "epoch": 0.3651979594074164, "grad_norm": 1.1070998907089233, "learning_rate": 0.0001988112714855303, "loss": 4.6045, "step": 195700 }, { "epoch": 0.36529126496679487, "grad_norm": 0.7762126922607422, "learning_rate": 0.0001988106255767658, "loss": 4.9177, "step": 195750 }, { "epoch": 0.3653845705261734, "grad_norm": 1.03263258934021, "learning_rate": 0.00019880997949361793, "loss": 4.7215, "step": 195800 }, { "epoch": 0.3654778760855519, "grad_norm": 0.9962678551673889, "learning_rate": 0.00019880933323608783, "loss": 4.49, "step": 195850 }, { "epoch": 0.36557118164493035, "grad_norm": 0.8162546157836914, "learning_rate": 0.00019880868680417663, "loss": 4.7263, "step": 195900 }, { "epoch": 0.36566448720430883, "grad_norm": 0.7211110591888428, "learning_rate": 0.0001988080401978855, "loss": 4.7463, "step": 195950 }, { "epoch": 0.36575779276368736, "grad_norm": 0.9785488843917847, "learning_rate": 0.00019880739341721556, "loss": 4.7713, "step": 196000 }, { "epoch": 0.36585109832306584, "grad_norm": 0.8973463773727417, "learning_rate": 0.00019880674646216795, "loss": 4.7429, "step": 196050 }, { "epoch": 0.3659444038824443, "grad_norm": 1.1833984851837158, "learning_rate": 0.00019880609933274382, "loss": 4.5631, "step": 196100 }, { "epoch": 0.36603770944182284, "grad_norm": 1.098944067955017, "learning_rate": 0.0001988054520289443, "loss": 4.4881, "step": 196150 }, { "epoch": 0.3661310150012013, "grad_norm": 0.6843777894973755, "learning_rate": 0.00019880480455077059, "loss": 4.525, "step": 196200 }, { "epoch": 0.3662243205605798, "grad_norm": 0.9798048138618469, "learning_rate": 0.00019880415689822373, "loss": 4.5468, "step": 196250 }, { "epoch": 0.3663176261199583, "grad_norm": 1.0058033466339111, "learning_rate": 0.00019880350907130494, "loss": 4.5075, "step": 196300 }, { "epoch": 0.3664109316793368, "grad_norm": 0.8590142130851746, "learning_rate": 0.00019880286107001535, "loss": 4.5171, "step": 196350 }, { "epoch": 0.3665042372387153, "grad_norm": 0.8589456081390381, "learning_rate": 0.00019880221289435607, "loss": 4.672, "step": 196400 }, { "epoch": 0.36659754279809376, "grad_norm": 1.3543787002563477, "learning_rate": 0.00019880156454432827, "loss": 4.523, "step": 196450 }, { "epoch": 0.36669084835747223, "grad_norm": 0.9485663771629333, "learning_rate": 0.00019880091601993312, "loss": 4.8311, "step": 196500 }, { "epoch": 0.36678415391685076, "grad_norm": 0.8935502767562866, "learning_rate": 0.00019880026732117174, "loss": 4.604, "step": 196550 }, { "epoch": 0.36687745947622924, "grad_norm": 0.9361104369163513, "learning_rate": 0.00019879961844804524, "loss": 4.6418, "step": 196600 }, { "epoch": 0.3669707650356077, "grad_norm": 0.6516358852386475, "learning_rate": 0.00019879896940055481, "loss": 4.7734, "step": 196650 }, { "epoch": 0.36706407059498625, "grad_norm": 1.2583208084106445, "learning_rate": 0.0001987983201787016, "loss": 4.6789, "step": 196700 }, { "epoch": 0.3671573761543647, "grad_norm": 1.1848986148834229, "learning_rate": 0.00019879767078248673, "loss": 4.6755, "step": 196750 }, { "epoch": 0.3672506817137432, "grad_norm": 0.871610164642334, "learning_rate": 0.00019879702121191134, "loss": 4.5271, "step": 196800 }, { "epoch": 0.3673439872731217, "grad_norm": 1.1172187328338623, "learning_rate": 0.00019879637146697662, "loss": 4.6236, "step": 196850 }, { "epoch": 0.3674372928325002, "grad_norm": 0.910994827747345, "learning_rate": 0.00019879572154768367, "loss": 4.5114, "step": 196900 }, { "epoch": 0.3675305983918787, "grad_norm": 0.9143189787864685, "learning_rate": 0.00019879507145403365, "loss": 4.4976, "step": 196950 }, { "epoch": 0.36762390395125716, "grad_norm": 1.1715046167373657, "learning_rate": 0.0001987944211860277, "loss": 4.6836, "step": 197000 }, { "epoch": 0.3677172095106357, "grad_norm": 1.333890438079834, "learning_rate": 0.00019879377074366703, "loss": 4.7598, "step": 197050 }, { "epoch": 0.36781051507001417, "grad_norm": 1.0626753568649292, "learning_rate": 0.00019879312012695272, "loss": 4.7344, "step": 197100 }, { "epoch": 0.36790382062939264, "grad_norm": 0.8466847538948059, "learning_rate": 0.0001987924693358859, "loss": 4.6338, "step": 197150 }, { "epoch": 0.3679971261887711, "grad_norm": 1.0398815870285034, "learning_rate": 0.00019879181837046777, "loss": 4.5708, "step": 197200 }, { "epoch": 0.36809043174814965, "grad_norm": 1.2056119441986084, "learning_rate": 0.00019879116723069948, "loss": 4.8655, "step": 197250 }, { "epoch": 0.3681837373075281, "grad_norm": 1.0409513711929321, "learning_rate": 0.00019879051591658215, "loss": 4.6327, "step": 197300 }, { "epoch": 0.3682770428669066, "grad_norm": 1.0986347198486328, "learning_rate": 0.00019878986442811695, "loss": 4.6958, "step": 197350 }, { "epoch": 0.36837034842628513, "grad_norm": 1.093003749847412, "learning_rate": 0.00019878921276530502, "loss": 4.6851, "step": 197400 }, { "epoch": 0.3684636539856636, "grad_norm": 1.156713843345642, "learning_rate": 0.0001987885609281475, "loss": 4.7068, "step": 197450 }, { "epoch": 0.3685569595450421, "grad_norm": 1.0253217220306396, "learning_rate": 0.00019878790891664555, "loss": 4.5718, "step": 197500 }, { "epoch": 0.36865026510442056, "grad_norm": 0.8153467774391174, "learning_rate": 0.00019878725673080038, "loss": 4.6435, "step": 197550 }, { "epoch": 0.3687435706637991, "grad_norm": 0.8387562036514282, "learning_rate": 0.00019878660437061302, "loss": 4.733, "step": 197600 }, { "epoch": 0.36883687622317757, "grad_norm": 1.011487603187561, "learning_rate": 0.0001987859518360847, "loss": 4.772, "step": 197650 }, { "epoch": 0.36893018178255604, "grad_norm": 1.247372031211853, "learning_rate": 0.00019878529912721658, "loss": 4.6005, "step": 197700 }, { "epoch": 0.3690234873419346, "grad_norm": 0.8897690176963806, "learning_rate": 0.00019878464624400978, "loss": 4.6325, "step": 197750 }, { "epoch": 0.36911679290131305, "grad_norm": 1.1363788843154907, "learning_rate": 0.00019878399318646543, "loss": 4.7405, "step": 197800 }, { "epoch": 0.3692100984606915, "grad_norm": 0.8495743870735168, "learning_rate": 0.00019878333995458473, "loss": 4.4341, "step": 197850 }, { "epoch": 0.36930340402007, "grad_norm": 0.9996325969696045, "learning_rate": 0.00019878268654836882, "loss": 4.6559, "step": 197900 }, { "epoch": 0.36939670957944853, "grad_norm": 0.7758403420448303, "learning_rate": 0.00019878203296781885, "loss": 5.0731, "step": 197950 }, { "epoch": 0.369490015138827, "grad_norm": 0.9682093262672424, "learning_rate": 0.000198781379212936, "loss": 4.4734, "step": 198000 }, { "epoch": 0.369490015138827, "eval_loss": 4.794241905212402, "eval_runtime": 232.0461, "eval_samples_per_second": 11.239, "eval_steps_per_second": 11.239, "eval_tts_loss": 7.463459329881014, "step": 198000 }, { "epoch": 0.3695833206982055, "grad_norm": 1.051621675491333, "learning_rate": 0.00019878072528372136, "loss": 4.6863, "step": 198050 }, { "epoch": 0.369676626257584, "grad_norm": 1.0777297019958496, "learning_rate": 0.0001987800711801761, "loss": 4.5781, "step": 198100 }, { "epoch": 0.3697699318169625, "grad_norm": 0.8551693558692932, "learning_rate": 0.00019877941690230145, "loss": 4.6632, "step": 198150 }, { "epoch": 0.36986323737634097, "grad_norm": 0.8395803570747375, "learning_rate": 0.00019877876245009849, "loss": 4.6038, "step": 198200 }, { "epoch": 0.36995654293571945, "grad_norm": 0.860999584197998, "learning_rate": 0.0001987781078235684, "loss": 4.7329, "step": 198250 }, { "epoch": 0.370049848495098, "grad_norm": 1.099967360496521, "learning_rate": 0.0001987774530227123, "loss": 4.7567, "step": 198300 }, { "epoch": 0.37014315405447645, "grad_norm": 0.6936938166618347, "learning_rate": 0.00019877679804753141, "loss": 4.673, "step": 198350 }, { "epoch": 0.37023645961385493, "grad_norm": 0.959591805934906, "learning_rate": 0.00019877614289802686, "loss": 4.6583, "step": 198400 }, { "epoch": 0.37032976517323346, "grad_norm": 0.7232271432876587, "learning_rate": 0.00019877548757419976, "loss": 4.5433, "step": 198450 }, { "epoch": 0.37042307073261194, "grad_norm": 1.075395941734314, "learning_rate": 0.00019877483207605132, "loss": 4.5049, "step": 198500 }, { "epoch": 0.3705163762919904, "grad_norm": 1.0577384233474731, "learning_rate": 0.0001987741764035827, "loss": 4.642, "step": 198550 }, { "epoch": 0.3706096818513689, "grad_norm": 1.5923864841461182, "learning_rate": 0.000198773520556795, "loss": 4.7332, "step": 198600 }, { "epoch": 0.3707029874107474, "grad_norm": 0.9335559010505676, "learning_rate": 0.00019877286453568943, "loss": 4.4329, "step": 198650 }, { "epoch": 0.3707962929701259, "grad_norm": 1.0967572927474976, "learning_rate": 0.00019877220834026718, "loss": 4.4424, "step": 198700 }, { "epoch": 0.37088959852950437, "grad_norm": 1.1251230239868164, "learning_rate": 0.00019877155197052933, "loss": 4.5098, "step": 198750 }, { "epoch": 0.3709829040888829, "grad_norm": 1.2461490631103516, "learning_rate": 0.00019877089542647705, "loss": 4.636, "step": 198800 }, { "epoch": 0.3710762096482614, "grad_norm": 0.9544318914413452, "learning_rate": 0.00019877023870811154, "loss": 4.6384, "step": 198850 }, { "epoch": 0.37116951520763986, "grad_norm": 1.099353313446045, "learning_rate": 0.00019876958181543392, "loss": 4.7218, "step": 198900 }, { "epoch": 0.37126282076701833, "grad_norm": 0.7116580009460449, "learning_rate": 0.0001987689247484454, "loss": 4.5868, "step": 198950 }, { "epoch": 0.37135612632639686, "grad_norm": 0.9305589199066162, "learning_rate": 0.00019876826750714707, "loss": 4.5696, "step": 199000 }, { "epoch": 0.37144943188577534, "grad_norm": 1.0949960947036743, "learning_rate": 0.00019876761009154014, "loss": 4.4849, "step": 199050 }, { "epoch": 0.3715427374451538, "grad_norm": 0.9973856210708618, "learning_rate": 0.00019876695250162576, "loss": 4.6221, "step": 199100 }, { "epoch": 0.3716360430045323, "grad_norm": 0.7113000154495239, "learning_rate": 0.00019876629473740507, "loss": 4.6113, "step": 199150 }, { "epoch": 0.3717293485639108, "grad_norm": 0.8312941789627075, "learning_rate": 0.00019876563679887928, "loss": 4.6111, "step": 199200 }, { "epoch": 0.3718226541232893, "grad_norm": 1.2139105796813965, "learning_rate": 0.00019876497868604948, "loss": 4.5554, "step": 199250 }, { "epoch": 0.3719159596826678, "grad_norm": 1.0658022165298462, "learning_rate": 0.0001987643203989169, "loss": 4.6281, "step": 199300 }, { "epoch": 0.3720092652420463, "grad_norm": 0.9115583300590515, "learning_rate": 0.00019876366193748269, "loss": 4.8648, "step": 199350 }, { "epoch": 0.3721025708014248, "grad_norm": 1.0521410703659058, "learning_rate": 0.00019876300330174798, "loss": 4.5097, "step": 199400 }, { "epoch": 0.37219587636080326, "grad_norm": 0.9952971935272217, "learning_rate": 0.00019876234449171391, "loss": 4.7379, "step": 199450 }, { "epoch": 0.37228918192018173, "grad_norm": 0.8925333023071289, "learning_rate": 0.00019876168550738172, "loss": 4.5746, "step": 199500 }, { "epoch": 0.37238248747956026, "grad_norm": 1.0621514320373535, "learning_rate": 0.00019876102634875254, "loss": 4.7674, "step": 199550 }, { "epoch": 0.37247579303893874, "grad_norm": 1.0929023027420044, "learning_rate": 0.00019876036701582748, "loss": 4.7718, "step": 199600 }, { "epoch": 0.3725690985983172, "grad_norm": 0.9549880027770996, "learning_rate": 0.00019875970750860778, "loss": 4.7339, "step": 199650 }, { "epoch": 0.37266240415769575, "grad_norm": 0.9763375520706177, "learning_rate": 0.0001987590478270946, "loss": 4.6824, "step": 199700 }, { "epoch": 0.3727557097170742, "grad_norm": 0.9068183898925781, "learning_rate": 0.00019875838797128902, "loss": 4.7193, "step": 199750 }, { "epoch": 0.3728490152764527, "grad_norm": 0.9717890024185181, "learning_rate": 0.0001987577279411923, "loss": 4.5489, "step": 199800 }, { "epoch": 0.3729423208358312, "grad_norm": 1.2245581150054932, "learning_rate": 0.00019875706773680554, "loss": 4.705, "step": 199850 }, { "epoch": 0.3730356263952097, "grad_norm": 0.7146151065826416, "learning_rate": 0.00019875640735812997, "loss": 4.5574, "step": 199900 }, { "epoch": 0.3731289319545882, "grad_norm": 0.8383366465568542, "learning_rate": 0.00019875574680516667, "loss": 4.5159, "step": 199950 }, { "epoch": 0.37322223751396666, "grad_norm": 1.15836501121521, "learning_rate": 0.0001987550860779169, "loss": 4.8317, "step": 200000 }, { "epoch": 0.3733155430733452, "grad_norm": 0.8754977583885193, "learning_rate": 0.00019875442517638177, "loss": 4.7887, "step": 200050 }, { "epoch": 0.37340884863272367, "grad_norm": 0.8341622352600098, "learning_rate": 0.00019875376410056244, "loss": 4.5602, "step": 200100 }, { "epoch": 0.37350215419210214, "grad_norm": 0.9436699151992798, "learning_rate": 0.00019875310285046008, "loss": 4.5878, "step": 200150 }, { "epoch": 0.3735954597514806, "grad_norm": 1.203100562095642, "learning_rate": 0.0001987524414260759, "loss": 4.5758, "step": 200200 }, { "epoch": 0.37368876531085915, "grad_norm": 0.9803832173347473, "learning_rate": 0.00019875177982741106, "loss": 4.5436, "step": 200250 }, { "epoch": 0.3737820708702376, "grad_norm": 0.9167838096618652, "learning_rate": 0.00019875111805446664, "loss": 4.7414, "step": 200300 }, { "epoch": 0.3738753764296161, "grad_norm": 1.015909194946289, "learning_rate": 0.0001987504561072439, "loss": 4.7295, "step": 200350 }, { "epoch": 0.37396868198899463, "grad_norm": 1.1135481595993042, "learning_rate": 0.00019874979398574398, "loss": 4.4958, "step": 200400 }, { "epoch": 0.3740619875483731, "grad_norm": 0.9369817972183228, "learning_rate": 0.00019874913168996808, "loss": 4.5166, "step": 200450 }, { "epoch": 0.3741552931077516, "grad_norm": 1.2627172470092773, "learning_rate": 0.0001987484692199173, "loss": 4.661, "step": 200500 }, { "epoch": 0.37424859866713006, "grad_norm": 1.0819026231765747, "learning_rate": 0.00019874780657559287, "loss": 4.4839, "step": 200550 }, { "epoch": 0.3743419042265086, "grad_norm": 1.0878807306289673, "learning_rate": 0.0001987471437569959, "loss": 4.5501, "step": 200600 }, { "epoch": 0.37443520978588707, "grad_norm": 1.0913496017456055, "learning_rate": 0.00019874648076412762, "loss": 4.9156, "step": 200650 }, { "epoch": 0.37452851534526554, "grad_norm": 0.9594210386276245, "learning_rate": 0.00019874581759698917, "loss": 4.5145, "step": 200700 }, { "epoch": 0.3746218209046441, "grad_norm": 0.8627813458442688, "learning_rate": 0.00019874515425558173, "loss": 4.5963, "step": 200750 }, { "epoch": 0.37471512646402255, "grad_norm": 0.9256874918937683, "learning_rate": 0.0001987444907399065, "loss": 4.671, "step": 200800 }, { "epoch": 0.37480843202340103, "grad_norm": 1.5067219734191895, "learning_rate": 0.0001987438270499646, "loss": 4.672, "step": 200850 }, { "epoch": 0.3749017375827795, "grad_norm": 1.0874707698822021, "learning_rate": 0.00019874316318575718, "loss": 4.6663, "step": 200900 }, { "epoch": 0.37499504314215804, "grad_norm": 1.2468639612197876, "learning_rate": 0.00019874249914728548, "loss": 4.6853, "step": 200950 }, { "epoch": 0.3750883487015365, "grad_norm": 0.7165033221244812, "learning_rate": 0.00019874183493455064, "loss": 4.6766, "step": 201000 }, { "epoch": 0.3750883487015365, "eval_loss": 4.802398681640625, "eval_runtime": 230.2804, "eval_samples_per_second": 11.325, "eval_steps_per_second": 11.325, "eval_tts_loss": 7.475072805292653, "step": 201000 }, { "epoch": 0.375181654260915, "grad_norm": 1.1223444938659668, "learning_rate": 0.00019874117054755386, "loss": 4.6664, "step": 201050 }, { "epoch": 0.3752749598202935, "grad_norm": 1.1241016387939453, "learning_rate": 0.00019874050598629625, "loss": 4.5235, "step": 201100 }, { "epoch": 0.375368265379672, "grad_norm": 1.3051515817642212, "learning_rate": 0.00019873984125077905, "loss": 4.4631, "step": 201150 }, { "epoch": 0.37546157093905047, "grad_norm": 1.0892348289489746, "learning_rate": 0.0001987391763410034, "loss": 4.4371, "step": 201200 }, { "epoch": 0.37555487649842895, "grad_norm": 0.9838119745254517, "learning_rate": 0.00019873851125697044, "loss": 4.723, "step": 201250 }, { "epoch": 0.3756481820578075, "grad_norm": 1.3931784629821777, "learning_rate": 0.00019873784599868145, "loss": 4.7364, "step": 201300 }, { "epoch": 0.37574148761718595, "grad_norm": 0.9986215829849243, "learning_rate": 0.00019873718056613747, "loss": 4.4411, "step": 201350 }, { "epoch": 0.37583479317656443, "grad_norm": 1.0121062994003296, "learning_rate": 0.0001987365149593398, "loss": 4.3812, "step": 201400 }, { "epoch": 0.37592809873594296, "grad_norm": 1.1041994094848633, "learning_rate": 0.0001987358491782895, "loss": 4.4639, "step": 201450 }, { "epoch": 0.37602140429532144, "grad_norm": 0.7851685285568237, "learning_rate": 0.00019873518322298782, "loss": 4.4774, "step": 201500 }, { "epoch": 0.3761147098546999, "grad_norm": 1.1037462949752808, "learning_rate": 0.0001987345170934359, "loss": 4.6946, "step": 201550 }, { "epoch": 0.3762080154140784, "grad_norm": 0.9130716323852539, "learning_rate": 0.00019873385078963496, "loss": 4.6075, "step": 201600 }, { "epoch": 0.3763013209734569, "grad_norm": 1.2021543979644775, "learning_rate": 0.00019873318431158613, "loss": 4.5221, "step": 201650 }, { "epoch": 0.3763946265328354, "grad_norm": 0.9670273065567017, "learning_rate": 0.0001987325176592906, "loss": 4.8346, "step": 201700 }, { "epoch": 0.3764879320922139, "grad_norm": 0.8241245746612549, "learning_rate": 0.00019873185083274956, "loss": 4.487, "step": 201750 }, { "epoch": 0.37658123765159235, "grad_norm": 1.071000337600708, "learning_rate": 0.00019873118383196417, "loss": 4.5168, "step": 201800 }, { "epoch": 0.3766745432109709, "grad_norm": 0.627001166343689, "learning_rate": 0.0001987305166569356, "loss": 4.8239, "step": 201850 }, { "epoch": 0.37676784877034936, "grad_norm": 0.781301736831665, "learning_rate": 0.00019872984930766505, "loss": 4.4841, "step": 201900 }, { "epoch": 0.37686115432972783, "grad_norm": 0.7564511895179749, "learning_rate": 0.0001987291817841537, "loss": 4.4971, "step": 201950 }, { "epoch": 0.37695445988910636, "grad_norm": 0.9705767035484314, "learning_rate": 0.0001987285140864027, "loss": 4.4673, "step": 202000 }, { "epoch": 0.37704776544848484, "grad_norm": 0.9428151845932007, "learning_rate": 0.00019872784621441326, "loss": 4.8026, "step": 202050 }, { "epoch": 0.3771410710078633, "grad_norm": 1.1641000509262085, "learning_rate": 0.00019872717816818653, "loss": 4.7319, "step": 202100 }, { "epoch": 0.3772343765672418, "grad_norm": 1.1046082973480225, "learning_rate": 0.0001987265099477237, "loss": 4.8214, "step": 202150 }, { "epoch": 0.3773276821266203, "grad_norm": 1.2519506216049194, "learning_rate": 0.00019872584155302593, "loss": 4.6536, "step": 202200 }, { "epoch": 0.3774209876859988, "grad_norm": 1.1127455234527588, "learning_rate": 0.0001987251729840945, "loss": 4.5374, "step": 202250 }, { "epoch": 0.3775142932453773, "grad_norm": 1.0181680917739868, "learning_rate": 0.00019872450424093043, "loss": 4.4758, "step": 202300 }, { "epoch": 0.3776075988047558, "grad_norm": 0.7517116069793701, "learning_rate": 0.000198723835323535, "loss": 4.5575, "step": 202350 }, { "epoch": 0.3777009043641343, "grad_norm": 0.9754074215888977, "learning_rate": 0.00019872316623190938, "loss": 4.5489, "step": 202400 }, { "epoch": 0.37779420992351276, "grad_norm": 0.9626929759979248, "learning_rate": 0.00019872249696605473, "loss": 4.7142, "step": 202450 }, { "epoch": 0.37788751548289123, "grad_norm": 0.8055295944213867, "learning_rate": 0.0001987218275259723, "loss": 4.591, "step": 202500 }, { "epoch": 0.37798082104226977, "grad_norm": 0.821767270565033, "learning_rate": 0.00019872115791166314, "loss": 4.5601, "step": 202550 }, { "epoch": 0.37807412660164824, "grad_norm": 1.3287692070007324, "learning_rate": 0.00019872048812312855, "loss": 4.6671, "step": 202600 }, { "epoch": 0.3781674321610267, "grad_norm": 0.9818217158317566, "learning_rate": 0.00019871981816036965, "loss": 4.638, "step": 202650 }, { "epoch": 0.37826073772040525, "grad_norm": 1.010910987854004, "learning_rate": 0.00019871914802338763, "loss": 4.742, "step": 202700 }, { "epoch": 0.3783540432797837, "grad_norm": 1.0474064350128174, "learning_rate": 0.00019871847771218374, "loss": 4.7676, "step": 202750 }, { "epoch": 0.3784473488391622, "grad_norm": 0.8638759255409241, "learning_rate": 0.00019871780722675904, "loss": 4.7774, "step": 202800 }, { "epoch": 0.3785406543985407, "grad_norm": 1.0039918422698975, "learning_rate": 0.00019871713656711482, "loss": 4.5449, "step": 202850 }, { "epoch": 0.3786339599579192, "grad_norm": 1.126681923866272, "learning_rate": 0.00019871646573325222, "loss": 4.3922, "step": 202900 }, { "epoch": 0.3787272655172977, "grad_norm": 1.3228673934936523, "learning_rate": 0.00019871579472517241, "loss": 4.6475, "step": 202950 }, { "epoch": 0.37882057107667616, "grad_norm": 0.9036591649055481, "learning_rate": 0.00019871512354287664, "loss": 4.7047, "step": 203000 }, { "epoch": 0.3789138766360547, "grad_norm": 0.9258872866630554, "learning_rate": 0.00019871445218636597, "loss": 4.6349, "step": 203050 }, { "epoch": 0.37900718219543317, "grad_norm": 2.290501356124878, "learning_rate": 0.0001987137806556417, "loss": 4.4937, "step": 203100 }, { "epoch": 0.37910048775481164, "grad_norm": 0.8478452563285828, "learning_rate": 0.000198713108950705, "loss": 4.5709, "step": 203150 }, { "epoch": 0.3791937933141901, "grad_norm": 1.1010254621505737, "learning_rate": 0.000198712437071557, "loss": 4.7398, "step": 203200 }, { "epoch": 0.37928709887356865, "grad_norm": 1.4688420295715332, "learning_rate": 0.00019871176501819892, "loss": 4.6441, "step": 203250 }, { "epoch": 0.3793804044329471, "grad_norm": 1.065325379371643, "learning_rate": 0.00019871109279063196, "loss": 4.4881, "step": 203300 }, { "epoch": 0.3794737099923256, "grad_norm": 0.981877863407135, "learning_rate": 0.00019871042038885727, "loss": 4.5731, "step": 203350 }, { "epoch": 0.37956701555170413, "grad_norm": 1.0577549934387207, "learning_rate": 0.0001987097478128761, "loss": 4.7339, "step": 203400 }, { "epoch": 0.3796603211110826, "grad_norm": 0.9602108597755432, "learning_rate": 0.00019870907506268953, "loss": 4.5568, "step": 203450 }, { "epoch": 0.3797536266704611, "grad_norm": 1.1359353065490723, "learning_rate": 0.00019870840213829884, "loss": 4.6005, "step": 203500 }, { "epoch": 0.37984693222983956, "grad_norm": 0.9004436731338501, "learning_rate": 0.0001987077290397052, "loss": 4.7455, "step": 203550 }, { "epoch": 0.3799402377892181, "grad_norm": 1.0926775932312012, "learning_rate": 0.0001987070557669098, "loss": 4.4615, "step": 203600 }, { "epoch": 0.38003354334859657, "grad_norm": 0.8713423609733582, "learning_rate": 0.00019870638231991376, "loss": 4.4985, "step": 203650 }, { "epoch": 0.38012684890797505, "grad_norm": 0.6343261003494263, "learning_rate": 0.00019870570869871837, "loss": 4.437, "step": 203700 }, { "epoch": 0.3802201544673536, "grad_norm": 0.7554951310157776, "learning_rate": 0.00019870503490332475, "loss": 4.3759, "step": 203750 }, { "epoch": 0.38031346002673205, "grad_norm": 0.9022654891014099, "learning_rate": 0.0001987043609337341, "loss": 4.5305, "step": 203800 }, { "epoch": 0.38040676558611053, "grad_norm": 1.4872987270355225, "learning_rate": 0.00019870368678994766, "loss": 4.659, "step": 203850 }, { "epoch": 0.380500071145489, "grad_norm": 0.8781129121780396, "learning_rate": 0.00019870301247196655, "loss": 4.5687, "step": 203900 }, { "epoch": 0.38059337670486754, "grad_norm": 0.9872531294822693, "learning_rate": 0.000198702337979792, "loss": 4.4754, "step": 203950 }, { "epoch": 0.380686682264246, "grad_norm": 1.1369433403015137, "learning_rate": 0.0001987016633134252, "loss": 4.6713, "step": 204000 }, { "epoch": 0.380686682264246, "eval_loss": 4.801127910614014, "eval_runtime": 232.9265, "eval_samples_per_second": 11.197, "eval_steps_per_second": 11.197, "eval_tts_loss": 7.454781288711034, "step": 204000 }, { "epoch": 0.3807799878236245, "grad_norm": 1.0344752073287964, "learning_rate": 0.0001987009884728673, "loss": 4.5242, "step": 204050 }, { "epoch": 0.380873293383003, "grad_norm": 0.8404310345649719, "learning_rate": 0.00019870031345811955, "loss": 4.6894, "step": 204100 }, { "epoch": 0.3809665989423815, "grad_norm": 1.081472635269165, "learning_rate": 0.00019869963826918313, "loss": 4.5426, "step": 204150 }, { "epoch": 0.38105990450175997, "grad_norm": 0.9155741333961487, "learning_rate": 0.00019869896290605917, "loss": 4.781, "step": 204200 }, { "epoch": 0.38115321006113845, "grad_norm": 1.0043416023254395, "learning_rate": 0.00019869828736874894, "loss": 4.6185, "step": 204250 }, { "epoch": 0.381246515620517, "grad_norm": 0.6685605049133301, "learning_rate": 0.00019869761165725356, "loss": 4.7304, "step": 204300 }, { "epoch": 0.38133982117989546, "grad_norm": 0.9836552143096924, "learning_rate": 0.00019869693577157432, "loss": 4.6497, "step": 204350 }, { "epoch": 0.38143312673927393, "grad_norm": 1.1131340265274048, "learning_rate": 0.0001986962597117123, "loss": 4.5424, "step": 204400 }, { "epoch": 0.3815264322986524, "grad_norm": 0.8396437168121338, "learning_rate": 0.00019869558347766878, "loss": 4.6054, "step": 204450 }, { "epoch": 0.38161973785803094, "grad_norm": 0.8572611808776855, "learning_rate": 0.0001986949070694449, "loss": 4.5767, "step": 204500 }, { "epoch": 0.3817130434174094, "grad_norm": 0.9373936653137207, "learning_rate": 0.0001986942304870419, "loss": 4.5014, "step": 204550 }, { "epoch": 0.3818063489767879, "grad_norm": 1.0423073768615723, "learning_rate": 0.00019869355373046093, "loss": 4.3117, "step": 204600 }, { "epoch": 0.3818996545361664, "grad_norm": 0.8610841631889343, "learning_rate": 0.0001986928767997032, "loss": 4.2993, "step": 204650 }, { "epoch": 0.3819929600955449, "grad_norm": 1.0650521516799927, "learning_rate": 0.00019869219969476993, "loss": 4.7043, "step": 204700 }, { "epoch": 0.3820862656549234, "grad_norm": 1.2617119550704956, "learning_rate": 0.0001986915224156623, "loss": 4.7328, "step": 204750 }, { "epoch": 0.38217957121430185, "grad_norm": 1.0750634670257568, "learning_rate": 0.00019869084496238146, "loss": 4.6412, "step": 204800 }, { "epoch": 0.3822728767736804, "grad_norm": 1.0128196477890015, "learning_rate": 0.00019869016733492867, "loss": 4.7131, "step": 204850 }, { "epoch": 0.38236618233305886, "grad_norm": 0.9253135919570923, "learning_rate": 0.00019868948953330508, "loss": 4.5573, "step": 204900 }, { "epoch": 0.38245948789243733, "grad_norm": 1.0896155834197998, "learning_rate": 0.00019868881155751193, "loss": 4.5266, "step": 204950 }, { "epoch": 0.38255279345181586, "grad_norm": 1.0464351177215576, "learning_rate": 0.00019868813340755037, "loss": 4.5613, "step": 205000 }, { "epoch": 0.38264609901119434, "grad_norm": 0.8733029365539551, "learning_rate": 0.00019868745508342165, "loss": 4.4978, "step": 205050 }, { "epoch": 0.3827394045705728, "grad_norm": 1.1396583318710327, "learning_rate": 0.0001986867765851269, "loss": 4.5443, "step": 205100 }, { "epoch": 0.3828327101299513, "grad_norm": 1.0709539651870728, "learning_rate": 0.00019868609791266737, "loss": 4.7764, "step": 205150 }, { "epoch": 0.3829260156893298, "grad_norm": 0.9949414134025574, "learning_rate": 0.00019868541906604425, "loss": 4.5682, "step": 205200 }, { "epoch": 0.3830193212487083, "grad_norm": 0.9429978132247925, "learning_rate": 0.0001986847400452587, "loss": 4.5577, "step": 205250 }, { "epoch": 0.3831126268080868, "grad_norm": 1.0749703645706177, "learning_rate": 0.00019868406085031198, "loss": 4.6041, "step": 205300 }, { "epoch": 0.3832059323674653, "grad_norm": 0.7519810795783997, "learning_rate": 0.00019868338148120525, "loss": 4.5119, "step": 205350 }, { "epoch": 0.3832992379268438, "grad_norm": 1.1606097221374512, "learning_rate": 0.00019868270193793973, "loss": 4.6795, "step": 205400 }, { "epoch": 0.38339254348622226, "grad_norm": 0.6909898519515991, "learning_rate": 0.00019868202222051657, "loss": 4.7247, "step": 205450 }, { "epoch": 0.38348584904560074, "grad_norm": 1.001676321029663, "learning_rate": 0.00019868134232893704, "loss": 4.6081, "step": 205500 }, { "epoch": 0.38357915460497927, "grad_norm": 1.1560254096984863, "learning_rate": 0.00019868066226320227, "loss": 4.6179, "step": 205550 }, { "epoch": 0.38367246016435774, "grad_norm": 0.7835885882377625, "learning_rate": 0.00019867998202331352, "loss": 4.5546, "step": 205600 }, { "epoch": 0.3837657657237362, "grad_norm": 0.5636594295501709, "learning_rate": 0.00019867930160927195, "loss": 4.508, "step": 205650 }, { "epoch": 0.38385907128311475, "grad_norm": 1.0891164541244507, "learning_rate": 0.00019867862102107878, "loss": 4.6448, "step": 205700 }, { "epoch": 0.3839523768424932, "grad_norm": 0.8994442820549011, "learning_rate": 0.0001986779402587352, "loss": 4.4305, "step": 205750 }, { "epoch": 0.3840456824018717, "grad_norm": 1.0475820302963257, "learning_rate": 0.00019867725932224244, "loss": 4.5979, "step": 205800 }, { "epoch": 0.3841389879612502, "grad_norm": 1.0817757844924927, "learning_rate": 0.00019867657821160165, "loss": 4.4495, "step": 205850 }, { "epoch": 0.3842322935206287, "grad_norm": 1.2010945081710815, "learning_rate": 0.0001986758969268141, "loss": 4.5591, "step": 205900 }, { "epoch": 0.3843255990800072, "grad_norm": 0.9652662873268127, "learning_rate": 0.00019867521546788091, "loss": 4.5538, "step": 205950 }, { "epoch": 0.38441890463938566, "grad_norm": 1.1052769422531128, "learning_rate": 0.00019867453383480338, "loss": 4.6479, "step": 206000 }, { "epoch": 0.3845122101987642, "grad_norm": 1.1473944187164307, "learning_rate": 0.00019867385202758262, "loss": 4.5248, "step": 206050 }, { "epoch": 0.38460551575814267, "grad_norm": 1.0132917165756226, "learning_rate": 0.00019867317004621988, "loss": 4.6242, "step": 206100 }, { "epoch": 0.38469882131752114, "grad_norm": 1.299413800239563, "learning_rate": 0.00019867248789071638, "loss": 4.6021, "step": 206150 }, { "epoch": 0.3847921268768996, "grad_norm": 1.0853779315948486, "learning_rate": 0.0001986718055610733, "loss": 4.6003, "step": 206200 }, { "epoch": 0.38488543243627815, "grad_norm": 0.9251170754432678, "learning_rate": 0.0001986711230572918, "loss": 4.5685, "step": 206250 }, { "epoch": 0.38497873799565663, "grad_norm": 1.1247254610061646, "learning_rate": 0.00019867044037937315, "loss": 4.5821, "step": 206300 }, { "epoch": 0.3850720435550351, "grad_norm": 1.1961967945098877, "learning_rate": 0.00019866975752731855, "loss": 4.7458, "step": 206350 }, { "epoch": 0.38516534911441364, "grad_norm": 0.7294309735298157, "learning_rate": 0.00019866907450112917, "loss": 4.5956, "step": 206400 }, { "epoch": 0.3852586546737921, "grad_norm": 0.7703094482421875, "learning_rate": 0.00019866839130080623, "loss": 4.6983, "step": 206450 }, { "epoch": 0.3853519602331706, "grad_norm": 0.9793773293495178, "learning_rate": 0.00019866770792635098, "loss": 4.4386, "step": 206500 }, { "epoch": 0.38544526579254906, "grad_norm": 0.9847137928009033, "learning_rate": 0.00019866702437776454, "loss": 4.6341, "step": 206550 }, { "epoch": 0.3855385713519276, "grad_norm": 1.0143849849700928, "learning_rate": 0.00019866634065504818, "loss": 4.816, "step": 206600 }, { "epoch": 0.38563187691130607, "grad_norm": 0.8812471628189087, "learning_rate": 0.0001986656567582031, "loss": 4.4307, "step": 206650 }, { "epoch": 0.38572518247068455, "grad_norm": 0.9908750057220459, "learning_rate": 0.0001986649726872305, "loss": 4.7599, "step": 206700 }, { "epoch": 0.3858184880300631, "grad_norm": 1.2358583211898804, "learning_rate": 0.00019866428844213157, "loss": 4.7192, "step": 206750 }, { "epoch": 0.38591179358944155, "grad_norm": 0.8084837794303894, "learning_rate": 0.00019866360402290753, "loss": 4.6309, "step": 206800 }, { "epoch": 0.38600509914882003, "grad_norm": 0.917667806148529, "learning_rate": 0.0001986629194295596, "loss": 4.5326, "step": 206850 }, { "epoch": 0.3860984047081985, "grad_norm": 1.1653844118118286, "learning_rate": 0.00019866223466208897, "loss": 4.5962, "step": 206900 }, { "epoch": 0.38619171026757704, "grad_norm": 0.9410626888275146, "learning_rate": 0.00019866154972049687, "loss": 4.6704, "step": 206950 }, { "epoch": 0.3862850158269555, "grad_norm": 0.8716585040092468, "learning_rate": 0.00019866086460478447, "loss": 4.6156, "step": 207000 }, { "epoch": 0.3862850158269555, "eval_loss": 4.8045244216918945, "eval_runtime": 231.8143, "eval_samples_per_second": 11.25, "eval_steps_per_second": 11.25, "eval_tts_loss": 7.503764759250659, "step": 207000 }, { "epoch": 0.386378321386334, "grad_norm": 0.9506188631057739, "learning_rate": 0.00019866017931495302, "loss": 4.6642, "step": 207050 }, { "epoch": 0.38647162694571247, "grad_norm": 1.3123427629470825, "learning_rate": 0.00019865949385100372, "loss": 4.6589, "step": 207100 }, { "epoch": 0.386564932505091, "grad_norm": 1.0116405487060547, "learning_rate": 0.00019865880821293776, "loss": 4.6218, "step": 207150 }, { "epoch": 0.3866582380644695, "grad_norm": 1.0257128477096558, "learning_rate": 0.00019865812240075635, "loss": 4.8113, "step": 207200 }, { "epoch": 0.38675154362384795, "grad_norm": 1.1300727128982544, "learning_rate": 0.00019865743641446075, "loss": 4.6969, "step": 207250 }, { "epoch": 0.3868448491832265, "grad_norm": 0.8774489164352417, "learning_rate": 0.00019865675025405213, "loss": 4.6738, "step": 207300 }, { "epoch": 0.38693815474260496, "grad_norm": 1.0406553745269775, "learning_rate": 0.00019865606391953166, "loss": 4.7365, "step": 207350 }, { "epoch": 0.38703146030198343, "grad_norm": 0.9450544118881226, "learning_rate": 0.00019865537741090063, "loss": 4.454, "step": 207400 }, { "epoch": 0.3871247658613619, "grad_norm": 0.9951105117797852, "learning_rate": 0.0001986546907281602, "loss": 4.538, "step": 207450 }, { "epoch": 0.38721807142074044, "grad_norm": 0.7700023651123047, "learning_rate": 0.00019865400387131162, "loss": 4.4626, "step": 207500 }, { "epoch": 0.3873113769801189, "grad_norm": 0.8401572108268738, "learning_rate": 0.00019865331684035609, "loss": 4.7251, "step": 207550 }, { "epoch": 0.3874046825394974, "grad_norm": 1.3303920030593872, "learning_rate": 0.0001986526296352948, "loss": 4.6399, "step": 207600 }, { "epoch": 0.3874979880988759, "grad_norm": 0.9098189473152161, "learning_rate": 0.00019865194225612896, "loss": 4.6386, "step": 207650 }, { "epoch": 0.3875912936582544, "grad_norm": 1.0674896240234375, "learning_rate": 0.00019865125470285984, "loss": 4.652, "step": 207700 }, { "epoch": 0.3876845992176329, "grad_norm": 1.1178187131881714, "learning_rate": 0.0001986505669754886, "loss": 4.5139, "step": 207750 }, { "epoch": 0.38777790477701135, "grad_norm": 0.8213180303573608, "learning_rate": 0.00019864987907401645, "loss": 4.734, "step": 207800 }, { "epoch": 0.3878712103363899, "grad_norm": 1.1217759847640991, "learning_rate": 0.00019864919099844461, "loss": 4.5176, "step": 207850 }, { "epoch": 0.38796451589576836, "grad_norm": 0.978659451007843, "learning_rate": 0.00019864850274877433, "loss": 4.6919, "step": 207900 }, { "epoch": 0.38805782145514683, "grad_norm": 1.1488454341888428, "learning_rate": 0.0001986478143250068, "loss": 4.6732, "step": 207950 }, { "epoch": 0.38815112701452537, "grad_norm": 0.9637132287025452, "learning_rate": 0.00019864712572714323, "loss": 4.6186, "step": 208000 }, { "epoch": 0.38824443257390384, "grad_norm": 1.1298731565475464, "learning_rate": 0.00019864643695518487, "loss": 4.3077, "step": 208050 }, { "epoch": 0.3883377381332823, "grad_norm": 0.889434814453125, "learning_rate": 0.00019864574800913286, "loss": 4.4539, "step": 208100 }, { "epoch": 0.3884310436926608, "grad_norm": 1.0395722389221191, "learning_rate": 0.00019864505888898847, "loss": 4.6617, "step": 208150 }, { "epoch": 0.3885243492520393, "grad_norm": 0.8869847059249878, "learning_rate": 0.00019864436959475293, "loss": 4.5674, "step": 208200 }, { "epoch": 0.3886176548114178, "grad_norm": 1.071192979812622, "learning_rate": 0.0001986436801264274, "loss": 4.7536, "step": 208250 }, { "epoch": 0.3887109603707963, "grad_norm": 1.0335016250610352, "learning_rate": 0.00019864299048401315, "loss": 4.5433, "step": 208300 }, { "epoch": 0.3888042659301748, "grad_norm": 1.0146387815475464, "learning_rate": 0.00019864230066751138, "loss": 4.6243, "step": 208350 }, { "epoch": 0.3888975714895533, "grad_norm": 1.2553081512451172, "learning_rate": 0.00019864161067692333, "loss": 4.3292, "step": 208400 }, { "epoch": 0.38899087704893176, "grad_norm": 0.8601014018058777, "learning_rate": 0.00019864092051225017, "loss": 4.6767, "step": 208450 }, { "epoch": 0.38908418260831024, "grad_norm": 0.8037707209587097, "learning_rate": 0.00019864023017349315, "loss": 4.5461, "step": 208500 }, { "epoch": 0.38917748816768877, "grad_norm": 0.9061649441719055, "learning_rate": 0.00019863953966065348, "loss": 4.8048, "step": 208550 }, { "epoch": 0.38927079372706724, "grad_norm": 1.0509523153305054, "learning_rate": 0.00019863884897373236, "loss": 4.458, "step": 208600 }, { "epoch": 0.3893640992864457, "grad_norm": 1.026106834411621, "learning_rate": 0.00019863815811273106, "loss": 4.7856, "step": 208650 }, { "epoch": 0.38945740484582425, "grad_norm": 0.9003967642784119, "learning_rate": 0.00019863746707765074, "loss": 4.7185, "step": 208700 }, { "epoch": 0.3895507104052027, "grad_norm": 1.1053556203842163, "learning_rate": 0.00019863677586849266, "loss": 4.5309, "step": 208750 }, { "epoch": 0.3896440159645812, "grad_norm": 0.8523988127708435, "learning_rate": 0.00019863608448525798, "loss": 4.7219, "step": 208800 }, { "epoch": 0.3897373215239597, "grad_norm": 0.9140484929084778, "learning_rate": 0.00019863539292794803, "loss": 4.5942, "step": 208850 }, { "epoch": 0.3898306270833382, "grad_norm": 0.9049879312515259, "learning_rate": 0.00019863470119656394, "loss": 4.7273, "step": 208900 }, { "epoch": 0.3899239326427167, "grad_norm": 0.8477234840393066, "learning_rate": 0.00019863400929110697, "loss": 4.4955, "step": 208950 }, { "epoch": 0.39001723820209516, "grad_norm": 1.2483469247817993, "learning_rate": 0.00019863331721157833, "loss": 4.8071, "step": 209000 }, { "epoch": 0.3901105437614737, "grad_norm": 1.118116021156311, "learning_rate": 0.0001986326249579792, "loss": 4.6028, "step": 209050 }, { "epoch": 0.39020384932085217, "grad_norm": 1.069517731666565, "learning_rate": 0.0001986319325303109, "loss": 4.761, "step": 209100 }, { "epoch": 0.39029715488023065, "grad_norm": 1.1618387699127197, "learning_rate": 0.00019863123992857454, "loss": 4.5113, "step": 209150 }, { "epoch": 0.3903904604396091, "grad_norm": 0.905892014503479, "learning_rate": 0.00019863054715277143, "loss": 4.6528, "step": 209200 }, { "epoch": 0.39048376599898765, "grad_norm": 1.0331082344055176, "learning_rate": 0.00019862985420290277, "loss": 4.8151, "step": 209250 }, { "epoch": 0.39057707155836613, "grad_norm": 0.9063888788223267, "learning_rate": 0.00019862916107896971, "loss": 4.2769, "step": 209300 }, { "epoch": 0.3906703771177446, "grad_norm": 0.8382515907287598, "learning_rate": 0.00019862846778097358, "loss": 4.3706, "step": 209350 }, { "epoch": 0.39076368267712314, "grad_norm": 0.9924123883247375, "learning_rate": 0.00019862777430891556, "loss": 4.7743, "step": 209400 }, { "epoch": 0.3908569882365016, "grad_norm": 1.145404577255249, "learning_rate": 0.00019862708066279688, "loss": 4.5281, "step": 209450 }, { "epoch": 0.3909502937958801, "grad_norm": 1.0358670949935913, "learning_rate": 0.00019862638684261876, "loss": 4.6629, "step": 209500 }, { "epoch": 0.39104359935525856, "grad_norm": 1.2141669988632202, "learning_rate": 0.0001986256928483824, "loss": 4.5794, "step": 209550 }, { "epoch": 0.3911369049146371, "grad_norm": 0.9178851246833801, "learning_rate": 0.00019862499868008904, "loss": 4.6268, "step": 209600 }, { "epoch": 0.39123021047401557, "grad_norm": 1.0358548164367676, "learning_rate": 0.00019862430433773993, "loss": 4.7368, "step": 209650 }, { "epoch": 0.39132351603339405, "grad_norm": 0.8665915131568909, "learning_rate": 0.00019862360982133626, "loss": 4.3792, "step": 209700 }, { "epoch": 0.3914168215927725, "grad_norm": 1.0212570428848267, "learning_rate": 0.0001986229151308793, "loss": 4.8161, "step": 209750 }, { "epoch": 0.39151012715215106, "grad_norm": 1.306823968887329, "learning_rate": 0.00019862222026637022, "loss": 4.4615, "step": 209800 }, { "epoch": 0.39160343271152953, "grad_norm": 1.1909260749816895, "learning_rate": 0.00019862152522781028, "loss": 4.6705, "step": 209850 }, { "epoch": 0.391696738270908, "grad_norm": 0.8337374925613403, "learning_rate": 0.00019862083001520071, "loss": 4.5671, "step": 209900 }, { "epoch": 0.39179004383028654, "grad_norm": 1.068738341331482, "learning_rate": 0.0001986201346285427, "loss": 4.5786, "step": 209950 }, { "epoch": 0.391883349389665, "grad_norm": 0.9537367820739746, "learning_rate": 0.00019861943906783757, "loss": 4.6939, "step": 210000 }, { "epoch": 0.391883349389665, "eval_loss": 4.793107509613037, "eval_runtime": 232.1494, "eval_samples_per_second": 11.234, "eval_steps_per_second": 11.234, "eval_tts_loss": 7.506199054619971, "step": 210000 }, { "epoch": 0.3919766549490435, "grad_norm": 1.085239052772522, "learning_rate": 0.00019861874333308642, "loss": 4.599, "step": 210050 }, { "epoch": 0.39206996050842197, "grad_norm": 0.9580297470092773, "learning_rate": 0.00019861804742429056, "loss": 4.6775, "step": 210100 }, { "epoch": 0.3921632660678005, "grad_norm": 1.1435822248458862, "learning_rate": 0.00019861735134145118, "loss": 4.6902, "step": 210150 }, { "epoch": 0.392256571627179, "grad_norm": 1.1688177585601807, "learning_rate": 0.00019861665508456958, "loss": 4.63, "step": 210200 }, { "epoch": 0.39234987718655745, "grad_norm": 0.7775763273239136, "learning_rate": 0.0001986159586536469, "loss": 4.726, "step": 210250 }, { "epoch": 0.392443182745936, "grad_norm": 1.0630296468734741, "learning_rate": 0.0001986152620486844, "loss": 4.6786, "step": 210300 }, { "epoch": 0.39253648830531446, "grad_norm": 1.1185650825500488, "learning_rate": 0.0001986145652696833, "loss": 4.7136, "step": 210350 }, { "epoch": 0.39262979386469293, "grad_norm": 1.1582688093185425, "learning_rate": 0.00019861386831664487, "loss": 4.5912, "step": 210400 }, { "epoch": 0.3927230994240714, "grad_norm": 1.0869005918502808, "learning_rate": 0.00019861317118957032, "loss": 4.5373, "step": 210450 }, { "epoch": 0.39281640498344994, "grad_norm": 1.0537189245224, "learning_rate": 0.00019861247388846086, "loss": 4.7754, "step": 210500 }, { "epoch": 0.3929097105428284, "grad_norm": 1.4728331565856934, "learning_rate": 0.00019861177641331774, "loss": 4.7525, "step": 210550 }, { "epoch": 0.3930030161022069, "grad_norm": 1.0283353328704834, "learning_rate": 0.0001986110787641422, "loss": 4.4389, "step": 210600 }, { "epoch": 0.3930963216615854, "grad_norm": 1.0805425643920898, "learning_rate": 0.00019861038094093542, "loss": 4.6202, "step": 210650 }, { "epoch": 0.3931896272209639, "grad_norm": 0.8940476179122925, "learning_rate": 0.0001986096829436987, "loss": 4.6031, "step": 210700 }, { "epoch": 0.3932829327803424, "grad_norm": 0.9329911470413208, "learning_rate": 0.00019860898477243324, "loss": 4.4502, "step": 210750 }, { "epoch": 0.39337623833972085, "grad_norm": 0.9388894438743591, "learning_rate": 0.00019860828642714027, "loss": 4.675, "step": 210800 }, { "epoch": 0.3934695438990994, "grad_norm": 0.6746310591697693, "learning_rate": 0.000198607587907821, "loss": 4.8536, "step": 210850 }, { "epoch": 0.39356284945847786, "grad_norm": 0.9739928245544434, "learning_rate": 0.00019860688921447673, "loss": 4.6031, "step": 210900 }, { "epoch": 0.39365615501785634, "grad_norm": 1.030884027481079, "learning_rate": 0.00019860619034710864, "loss": 4.6616, "step": 210950 }, { "epoch": 0.39374946057723487, "grad_norm": 1.0043467283248901, "learning_rate": 0.00019860549130571795, "loss": 4.631, "step": 211000 }, { "epoch": 0.39384276613661334, "grad_norm": 1.2081234455108643, "learning_rate": 0.00019860479209030593, "loss": 4.5703, "step": 211050 }, { "epoch": 0.3939360716959918, "grad_norm": 0.9312835335731506, "learning_rate": 0.0001986040927008738, "loss": 4.6645, "step": 211100 }, { "epoch": 0.3940293772553703, "grad_norm": 0.9531163573265076, "learning_rate": 0.0001986033931374228, "loss": 4.6137, "step": 211150 }, { "epoch": 0.3941226828147488, "grad_norm": 1.0482721328735352, "learning_rate": 0.00019860269339995416, "loss": 4.3711, "step": 211200 }, { "epoch": 0.3942159883741273, "grad_norm": 0.7411187887191772, "learning_rate": 0.0001986019934884691, "loss": 4.4535, "step": 211250 }, { "epoch": 0.3943092939335058, "grad_norm": 1.1720820665359497, "learning_rate": 0.0001986012934029689, "loss": 4.6375, "step": 211300 }, { "epoch": 0.3944025994928843, "grad_norm": 0.9703693389892578, "learning_rate": 0.00019860059314345474, "loss": 4.4733, "step": 211350 }, { "epoch": 0.3944959050522628, "grad_norm": 1.1301404237747192, "learning_rate": 0.0001985998927099279, "loss": 4.6341, "step": 211400 }, { "epoch": 0.39458921061164126, "grad_norm": 0.7381196618080139, "learning_rate": 0.0001985991921023896, "loss": 4.675, "step": 211450 }, { "epoch": 0.39468251617101974, "grad_norm": 1.1832466125488281, "learning_rate": 0.00019859849132084108, "loss": 4.6554, "step": 211500 }, { "epoch": 0.39477582173039827, "grad_norm": 1.0502104759216309, "learning_rate": 0.00019859779036528354, "loss": 4.663, "step": 211550 }, { "epoch": 0.39486912728977674, "grad_norm": 1.127864122390747, "learning_rate": 0.0001985970892357183, "loss": 4.6673, "step": 211600 }, { "epoch": 0.3949624328491552, "grad_norm": 1.035516619682312, "learning_rate": 0.0001985963879321465, "loss": 4.623, "step": 211650 }, { "epoch": 0.39505573840853375, "grad_norm": 0.8197730779647827, "learning_rate": 0.00019859568645456946, "loss": 4.3858, "step": 211700 }, { "epoch": 0.39514904396791223, "grad_norm": 0.8772891163825989, "learning_rate": 0.00019859498480298836, "loss": 4.6063, "step": 211750 }, { "epoch": 0.3952423495272907, "grad_norm": 0.8859339952468872, "learning_rate": 0.00019859428297740444, "loss": 4.5472, "step": 211800 }, { "epoch": 0.3953356550866692, "grad_norm": 1.1464184522628784, "learning_rate": 0.000198593580977819, "loss": 4.6492, "step": 211850 }, { "epoch": 0.3954289606460477, "grad_norm": 1.1008684635162354, "learning_rate": 0.00019859287880423317, "loss": 4.5615, "step": 211900 }, { "epoch": 0.3955222662054262, "grad_norm": 1.2038742303848267, "learning_rate": 0.00019859217645664835, "loss": 4.6712, "step": 211950 }, { "epoch": 0.39561557176480466, "grad_norm": 1.1197429895401, "learning_rate": 0.00019859147393506564, "loss": 4.4475, "step": 212000 }, { "epoch": 0.3957088773241832, "grad_norm": 0.9153450727462769, "learning_rate": 0.00019859077123948632, "loss": 4.4927, "step": 212050 }, { "epoch": 0.39580218288356167, "grad_norm": 0.9062692523002625, "learning_rate": 0.00019859006836991166, "loss": 4.513, "step": 212100 }, { "epoch": 0.39589548844294015, "grad_norm": 1.01179838180542, "learning_rate": 0.00019858936532634282, "loss": 4.4879, "step": 212150 }, { "epoch": 0.3959887940023186, "grad_norm": 1.3619974851608276, "learning_rate": 0.00019858866210878116, "loss": 4.5294, "step": 212200 }, { "epoch": 0.39608209956169715, "grad_norm": 1.009998083114624, "learning_rate": 0.00019858795871722783, "loss": 4.6197, "step": 212250 }, { "epoch": 0.39617540512107563, "grad_norm": 1.0456039905548096, "learning_rate": 0.00019858725515168414, "loss": 4.3916, "step": 212300 }, { "epoch": 0.3962687106804541, "grad_norm": 1.0402274131774902, "learning_rate": 0.00019858655141215124, "loss": 4.6354, "step": 212350 }, { "epoch": 0.39636201623983264, "grad_norm": 1.4560786485671997, "learning_rate": 0.00019858584749863042, "loss": 4.6766, "step": 212400 }, { "epoch": 0.3964553217992111, "grad_norm": 0.6697715520858765, "learning_rate": 0.00019858514341112297, "loss": 4.5072, "step": 212450 }, { "epoch": 0.3965486273585896, "grad_norm": 0.9932947158813477, "learning_rate": 0.00019858443914963004, "loss": 4.7205, "step": 212500 }, { "epoch": 0.39664193291796807, "grad_norm": 1.2164796590805054, "learning_rate": 0.00019858373471415297, "loss": 4.5836, "step": 212550 }, { "epoch": 0.3967352384773466, "grad_norm": 1.0917052030563354, "learning_rate": 0.00019858303010469291, "loss": 4.5275, "step": 212600 }, { "epoch": 0.3968285440367251, "grad_norm": 0.9048320055007935, "learning_rate": 0.00019858232532125118, "loss": 4.709, "step": 212650 }, { "epoch": 0.39692184959610355, "grad_norm": 0.568315327167511, "learning_rate": 0.000198581620363829, "loss": 4.5221, "step": 212700 }, { "epoch": 0.397015155155482, "grad_norm": 1.132904052734375, "learning_rate": 0.00019858091523242756, "loss": 4.6825, "step": 212750 }, { "epoch": 0.39710846071486056, "grad_norm": 0.9325287342071533, "learning_rate": 0.0001985802099270482, "loss": 4.6598, "step": 212800 }, { "epoch": 0.39720176627423903, "grad_norm": 1.0554152727127075, "learning_rate": 0.00019857950444769208, "loss": 4.7556, "step": 212850 }, { "epoch": 0.3972950718336175, "grad_norm": 0.9634024500846863, "learning_rate": 0.0001985787987943605, "loss": 4.4973, "step": 212900 }, { "epoch": 0.39738837739299604, "grad_norm": 1.161877989768982, "learning_rate": 0.0001985780929670547, "loss": 4.6044, "step": 212950 }, { "epoch": 0.3974816829523745, "grad_norm": 0.954832136631012, "learning_rate": 0.00019857738696577588, "loss": 4.4247, "step": 213000 }, { "epoch": 0.3974816829523745, "eval_loss": 4.804925918579102, "eval_runtime": 229.7393, "eval_samples_per_second": 11.352, "eval_steps_per_second": 11.352, "eval_tts_loss": 7.437013648446089, "step": 213000 }, { "epoch": 0.397574988511753, "grad_norm": 0.9206513166427612, "learning_rate": 0.00019857668079052533, "loss": 4.6913, "step": 213050 }, { "epoch": 0.39766829407113147, "grad_norm": 1.1719000339508057, "learning_rate": 0.00019857597444130427, "loss": 4.6099, "step": 213100 }, { "epoch": 0.39776159963051, "grad_norm": 0.8833860158920288, "learning_rate": 0.00019857526791811398, "loss": 4.4701, "step": 213150 }, { "epoch": 0.3978549051898885, "grad_norm": 0.8113496899604797, "learning_rate": 0.00019857456122095567, "loss": 4.5576, "step": 213200 }, { "epoch": 0.39794821074926695, "grad_norm": 0.7514514327049255, "learning_rate": 0.0001985738543498306, "loss": 4.6348, "step": 213250 }, { "epoch": 0.3980415163086455, "grad_norm": 0.8051292896270752, "learning_rate": 0.00019857314730474006, "loss": 4.6339, "step": 213300 }, { "epoch": 0.39813482186802396, "grad_norm": 0.8697991967201233, "learning_rate": 0.00019857244008568523, "loss": 4.5064, "step": 213350 }, { "epoch": 0.39822812742740243, "grad_norm": 1.3794461488723755, "learning_rate": 0.0001985717326926674, "loss": 4.6744, "step": 213400 }, { "epoch": 0.3983214329867809, "grad_norm": 1.0612692832946777, "learning_rate": 0.0001985710251256878, "loss": 4.8218, "step": 213450 }, { "epoch": 0.39841473854615944, "grad_norm": 0.7990536689758301, "learning_rate": 0.00019857031738474768, "loss": 4.8063, "step": 213500 }, { "epoch": 0.3985080441055379, "grad_norm": 0.9997884035110474, "learning_rate": 0.0001985696094698483, "loss": 4.6703, "step": 213550 }, { "epoch": 0.3986013496649164, "grad_norm": 1.0094455480575562, "learning_rate": 0.00019856890138099088, "loss": 4.5359, "step": 213600 }, { "epoch": 0.3986946552242949, "grad_norm": 0.9724659323692322, "learning_rate": 0.00019856819311817674, "loss": 4.7684, "step": 213650 }, { "epoch": 0.3987879607836734, "grad_norm": 0.9077069759368896, "learning_rate": 0.00019856748468140704, "loss": 4.4383, "step": 213700 }, { "epoch": 0.3988812663430519, "grad_norm": 0.8644628524780273, "learning_rate": 0.00019856677607068307, "loss": 4.6682, "step": 213750 }, { "epoch": 0.39897457190243035, "grad_norm": 1.038778305053711, "learning_rate": 0.00019856606728600612, "loss": 4.6176, "step": 213800 }, { "epoch": 0.3990678774618089, "grad_norm": 0.8550258278846741, "learning_rate": 0.0001985653583273774, "loss": 4.3842, "step": 213850 }, { "epoch": 0.39916118302118736, "grad_norm": 0.929908275604248, "learning_rate": 0.00019856464919479813, "loss": 4.7997, "step": 213900 }, { "epoch": 0.39925448858056584, "grad_norm": 0.948686957359314, "learning_rate": 0.00019856393988826963, "loss": 4.608, "step": 213950 }, { "epoch": 0.39934779413994437, "grad_norm": 1.081762433052063, "learning_rate": 0.00019856323040779312, "loss": 4.5192, "step": 214000 }, { "epoch": 0.39944109969932284, "grad_norm": 1.0186339616775513, "learning_rate": 0.00019856252075336984, "loss": 4.5424, "step": 214050 }, { "epoch": 0.3995344052587013, "grad_norm": 1.243036150932312, "learning_rate": 0.00019856181092500104, "loss": 4.6291, "step": 214100 }, { "epoch": 0.3996277108180798, "grad_norm": 0.9121257066726685, "learning_rate": 0.000198561100922688, "loss": 4.4488, "step": 214150 }, { "epoch": 0.3997210163774583, "grad_norm": 1.058773159980774, "learning_rate": 0.00019856039074643195, "loss": 4.5231, "step": 214200 }, { "epoch": 0.3998143219368368, "grad_norm": 1.0914162397384644, "learning_rate": 0.00019855968039623418, "loss": 4.5719, "step": 214250 }, { "epoch": 0.3999076274962153, "grad_norm": 1.0850412845611572, "learning_rate": 0.0001985589698720959, "loss": 4.6784, "step": 214300 }, { "epoch": 0.4000009330555938, "grad_norm": 0.8191171884536743, "learning_rate": 0.00019855825917401837, "loss": 4.6503, "step": 214350 }, { "epoch": 0.4000942386149723, "grad_norm": 1.133804440498352, "learning_rate": 0.00019855754830200284, "loss": 4.6447, "step": 214400 }, { "epoch": 0.40018754417435076, "grad_norm": 0.927496612071991, "learning_rate": 0.0001985568372560506, "loss": 4.6549, "step": 214450 }, { "epoch": 0.40028084973372924, "grad_norm": 0.9108291864395142, "learning_rate": 0.00019855612603616288, "loss": 4.6962, "step": 214500 }, { "epoch": 0.40037415529310777, "grad_norm": 1.0643376111984253, "learning_rate": 0.00019855541464234094, "loss": 4.6508, "step": 214550 }, { "epoch": 0.40046746085248625, "grad_norm": 1.0396971702575684, "learning_rate": 0.00019855470307458606, "loss": 4.6046, "step": 214600 }, { "epoch": 0.4005607664118647, "grad_norm": 0.9104790091514587, "learning_rate": 0.00019855399133289944, "loss": 4.3958, "step": 214650 }, { "epoch": 0.40065407197124325, "grad_norm": 0.8371784687042236, "learning_rate": 0.00019855327941728238, "loss": 4.6615, "step": 214700 }, { "epoch": 0.40074737753062173, "grad_norm": 0.8036884665489197, "learning_rate": 0.0001985525673277361, "loss": 4.4121, "step": 214750 }, { "epoch": 0.4008406830900002, "grad_norm": 1.0669821500778198, "learning_rate": 0.0001985518550642619, "loss": 4.4635, "step": 214800 }, { "epoch": 0.4009339886493787, "grad_norm": 1.0389690399169922, "learning_rate": 0.00019855114262686102, "loss": 4.6575, "step": 214850 }, { "epoch": 0.4010272942087572, "grad_norm": 0.661183774471283, "learning_rate": 0.0001985504300155347, "loss": 4.6654, "step": 214900 }, { "epoch": 0.4011205997681357, "grad_norm": 1.4455314874649048, "learning_rate": 0.00019854971723028424, "loss": 4.7863, "step": 214950 }, { "epoch": 0.40121390532751416, "grad_norm": 0.9810767769813538, "learning_rate": 0.00019854900427111085, "loss": 4.4652, "step": 215000 }, { "epoch": 0.4013072108868927, "grad_norm": 0.9444175362586975, "learning_rate": 0.00019854829113801582, "loss": 4.6636, "step": 215050 }, { "epoch": 0.40140051644627117, "grad_norm": 0.8920708894729614, "learning_rate": 0.00019854757783100037, "loss": 4.5821, "step": 215100 }, { "epoch": 0.40149382200564965, "grad_norm": 0.9815887808799744, "learning_rate": 0.0001985468643500658, "loss": 4.6339, "step": 215150 }, { "epoch": 0.4015871275650281, "grad_norm": 0.9865546226501465, "learning_rate": 0.00019854615069521333, "loss": 4.5329, "step": 215200 }, { "epoch": 0.40168043312440666, "grad_norm": 1.2863136529922485, "learning_rate": 0.00019854543686644428, "loss": 4.7172, "step": 215250 }, { "epoch": 0.40177373868378513, "grad_norm": 1.207463264465332, "learning_rate": 0.00019854472286375986, "loss": 4.4284, "step": 215300 }, { "epoch": 0.4018670442431636, "grad_norm": 0.9688467383384705, "learning_rate": 0.00019854400868716133, "loss": 4.6533, "step": 215350 }, { "epoch": 0.4019603498025421, "grad_norm": 0.767192542552948, "learning_rate": 0.00019854329433665, "loss": 4.4779, "step": 215400 }, { "epoch": 0.4020536553619206, "grad_norm": 1.0222392082214355, "learning_rate": 0.00019854257981222705, "loss": 4.6248, "step": 215450 }, { "epoch": 0.4021469609212991, "grad_norm": 1.1537096500396729, "learning_rate": 0.00019854186511389382, "loss": 4.5435, "step": 215500 }, { "epoch": 0.40224026648067757, "grad_norm": 0.7932273745536804, "learning_rate": 0.00019854115024165153, "loss": 4.7153, "step": 215550 }, { "epoch": 0.4023335720400561, "grad_norm": 0.8604770302772522, "learning_rate": 0.00019854043519550144, "loss": 4.7003, "step": 215600 }, { "epoch": 0.4024268775994346, "grad_norm": 1.160075306892395, "learning_rate": 0.00019853971997544481, "loss": 4.4709, "step": 215650 }, { "epoch": 0.40252018315881305, "grad_norm": 0.9051239490509033, "learning_rate": 0.00019853900458148294, "loss": 4.7949, "step": 215700 }, { "epoch": 0.4026134887181915, "grad_norm": 0.8977974653244019, "learning_rate": 0.00019853828901361705, "loss": 4.5747, "step": 215750 }, { "epoch": 0.40270679427757006, "grad_norm": 1.0881613492965698, "learning_rate": 0.0001985375732718484, "loss": 4.6564, "step": 215800 }, { "epoch": 0.40280009983694853, "grad_norm": 1.3626446723937988, "learning_rate": 0.00019853685735617832, "loss": 4.613, "step": 215850 }, { "epoch": 0.402893405396327, "grad_norm": 1.0976074934005737, "learning_rate": 0.00019853614126660797, "loss": 4.6058, "step": 215900 }, { "epoch": 0.40298671095570554, "grad_norm": 1.0088049173355103, "learning_rate": 0.0001985354250031387, "loss": 4.5451, "step": 215950 }, { "epoch": 0.403080016515084, "grad_norm": 0.7824239730834961, "learning_rate": 0.00019853470856577175, "loss": 4.5725, "step": 216000 }, { "epoch": 0.403080016515084, "eval_loss": 4.79715633392334, "eval_runtime": 235.0521, "eval_samples_per_second": 11.095, "eval_steps_per_second": 11.095, "eval_tts_loss": 7.512439045195176, "step": 216000 }, { "epoch": 0.4031733220744625, "grad_norm": 0.9689962863922119, "learning_rate": 0.00019853399195450834, "loss": 4.4423, "step": 216050 }, { "epoch": 0.40326662763384097, "grad_norm": 1.1092582941055298, "learning_rate": 0.00019853327516934982, "loss": 4.5845, "step": 216100 }, { "epoch": 0.4033599331932195, "grad_norm": 1.217396855354309, "learning_rate": 0.0001985325582102974, "loss": 4.634, "step": 216150 }, { "epoch": 0.403453238752598, "grad_norm": 1.0966297388076782, "learning_rate": 0.00019853184107735232, "loss": 4.6351, "step": 216200 }, { "epoch": 0.40354654431197645, "grad_norm": 0.943762481212616, "learning_rate": 0.0001985311237705159, "loss": 4.7421, "step": 216250 }, { "epoch": 0.403639849871355, "grad_norm": 1.03568434715271, "learning_rate": 0.00019853040628978937, "loss": 4.6778, "step": 216300 }, { "epoch": 0.40373315543073346, "grad_norm": 1.1071677207946777, "learning_rate": 0.000198529688635174, "loss": 4.6034, "step": 216350 }, { "epoch": 0.40382646099011194, "grad_norm": 0.9357593059539795, "learning_rate": 0.0001985289708066711, "loss": 4.7582, "step": 216400 }, { "epoch": 0.4039197665494904, "grad_norm": 1.0563290119171143, "learning_rate": 0.00019852825280428186, "loss": 4.552, "step": 216450 }, { "epoch": 0.40401307210886894, "grad_norm": 1.0622097253799438, "learning_rate": 0.00019852753462800763, "loss": 4.5838, "step": 216500 }, { "epoch": 0.4041063776682474, "grad_norm": 1.0469093322753906, "learning_rate": 0.00019852681627784962, "loss": 4.5972, "step": 216550 }, { "epoch": 0.4041996832276259, "grad_norm": 0.950249969959259, "learning_rate": 0.00019852609775380912, "loss": 4.5714, "step": 216600 }, { "epoch": 0.4042929887870044, "grad_norm": 0.9721581935882568, "learning_rate": 0.00019852537905588741, "loss": 4.5639, "step": 216650 }, { "epoch": 0.4043862943463829, "grad_norm": 1.0957340002059937, "learning_rate": 0.00019852466018408573, "loss": 4.4685, "step": 216700 }, { "epoch": 0.4044795999057614, "grad_norm": 1.099837064743042, "learning_rate": 0.00019852394113840536, "loss": 4.6103, "step": 216750 }, { "epoch": 0.40457290546513985, "grad_norm": 1.1709736585617065, "learning_rate": 0.00019852322191884757, "loss": 4.6009, "step": 216800 }, { "epoch": 0.4046662110245184, "grad_norm": 0.804123044013977, "learning_rate": 0.0001985225025254136, "loss": 4.614, "step": 216850 }, { "epoch": 0.40475951658389686, "grad_norm": 0.7690798044204712, "learning_rate": 0.0001985217829581048, "loss": 4.5003, "step": 216900 }, { "epoch": 0.40485282214327534, "grad_norm": 1.1280100345611572, "learning_rate": 0.00019852106321692234, "loss": 4.3368, "step": 216950 }, { "epoch": 0.40494612770265387, "grad_norm": 0.9494564533233643, "learning_rate": 0.00019852034330186755, "loss": 4.5604, "step": 217000 }, { "epoch": 0.40503943326203234, "grad_norm": 0.8637983798980713, "learning_rate": 0.00019851962321294173, "loss": 4.6788, "step": 217050 }, { "epoch": 0.4051327388214108, "grad_norm": 1.0316267013549805, "learning_rate": 0.00019851890295014606, "loss": 4.6512, "step": 217100 }, { "epoch": 0.4052260443807893, "grad_norm": 1.0898293256759644, "learning_rate": 0.00019851818251348188, "loss": 4.5092, "step": 217150 }, { "epoch": 0.40531934994016783, "grad_norm": 0.9936012029647827, "learning_rate": 0.00019851746190295043, "loss": 4.5646, "step": 217200 }, { "epoch": 0.4054126554995463, "grad_norm": 1.9127103090286255, "learning_rate": 0.00019851674111855304, "loss": 4.6748, "step": 217250 }, { "epoch": 0.4055059610589248, "grad_norm": 1.153311014175415, "learning_rate": 0.0001985160201602909, "loss": 4.6777, "step": 217300 }, { "epoch": 0.4055992666183033, "grad_norm": 0.802513599395752, "learning_rate": 0.0001985152990281653, "loss": 4.7041, "step": 217350 }, { "epoch": 0.4056925721776818, "grad_norm": 0.7697170972824097, "learning_rate": 0.00019851457772217756, "loss": 4.6375, "step": 217400 }, { "epoch": 0.40578587773706026, "grad_norm": 1.0160504579544067, "learning_rate": 0.00019851385624232895, "loss": 4.5718, "step": 217450 }, { "epoch": 0.40587918329643874, "grad_norm": 1.2011131048202515, "learning_rate": 0.0001985131345886207, "loss": 4.5841, "step": 217500 }, { "epoch": 0.40597248885581727, "grad_norm": 0.7780730724334717, "learning_rate": 0.00019851241276105407, "loss": 4.6344, "step": 217550 }, { "epoch": 0.40606579441519575, "grad_norm": 1.0549157857894897, "learning_rate": 0.00019851169075963036, "loss": 4.7876, "step": 217600 }, { "epoch": 0.4061590999745742, "grad_norm": 0.9472026228904724, "learning_rate": 0.00019851096858435088, "loss": 4.6313, "step": 217650 }, { "epoch": 0.40625240553395275, "grad_norm": 0.9771903157234192, "learning_rate": 0.0001985102462352169, "loss": 4.5077, "step": 217700 }, { "epoch": 0.40634571109333123, "grad_norm": 1.2292876243591309, "learning_rate": 0.00019850952371222962, "loss": 4.5595, "step": 217750 }, { "epoch": 0.4064390166527097, "grad_norm": 1.2494375705718994, "learning_rate": 0.00019850880101539037, "loss": 4.713, "step": 217800 }, { "epoch": 0.4065323222120882, "grad_norm": 1.2384469509124756, "learning_rate": 0.00019850807814470045, "loss": 4.5567, "step": 217850 }, { "epoch": 0.4066256277714667, "grad_norm": 1.1389862298965454, "learning_rate": 0.00019850735510016108, "loss": 4.5898, "step": 217900 }, { "epoch": 0.4067189333308452, "grad_norm": 1.0417871475219727, "learning_rate": 0.00019850663188177356, "loss": 4.6815, "step": 217950 }, { "epoch": 0.40681223889022367, "grad_norm": 1.019178867340088, "learning_rate": 0.00019850590848953918, "loss": 4.6889, "step": 218000 }, { "epoch": 0.40690554444960214, "grad_norm": 0.7551344633102417, "learning_rate": 0.0001985051849234592, "loss": 4.6242, "step": 218050 }, { "epoch": 0.4069988500089807, "grad_norm": 1.3654025793075562, "learning_rate": 0.0001985044611835349, "loss": 4.7712, "step": 218100 }, { "epoch": 0.40709215556835915, "grad_norm": 0.6821979284286499, "learning_rate": 0.00019850373726976756, "loss": 4.6176, "step": 218150 }, { "epoch": 0.4071854611277376, "grad_norm": 1.0702706575393677, "learning_rate": 0.00019850301318215846, "loss": 4.4522, "step": 218200 }, { "epoch": 0.40727876668711616, "grad_norm": 1.0518271923065186, "learning_rate": 0.00019850228892070884, "loss": 4.7112, "step": 218250 }, { "epoch": 0.40737207224649463, "grad_norm": 1.0100867748260498, "learning_rate": 0.00019850156448542007, "loss": 4.7564, "step": 218300 }, { "epoch": 0.4074653778058731, "grad_norm": 1.5200986862182617, "learning_rate": 0.00019850083987629332, "loss": 4.5178, "step": 218350 }, { "epoch": 0.4075586833652516, "grad_norm": 1.1372010707855225, "learning_rate": 0.00019850011509332994, "loss": 4.5793, "step": 218400 }, { "epoch": 0.4076519889246301, "grad_norm": 1.1734660863876343, "learning_rate": 0.0001984993901365312, "loss": 4.7042, "step": 218450 }, { "epoch": 0.4077452944840086, "grad_norm": 0.9464728236198425, "learning_rate": 0.00019849866500589833, "loss": 4.5238, "step": 218500 }, { "epoch": 0.40783860004338707, "grad_norm": 1.0089821815490723, "learning_rate": 0.0001984979397014327, "loss": 4.4707, "step": 218550 }, { "epoch": 0.4079319056027656, "grad_norm": 0.7066876292228699, "learning_rate": 0.0001984972142231355, "loss": 4.556, "step": 218600 }, { "epoch": 0.4080252111621441, "grad_norm": 1.1190834045410156, "learning_rate": 0.00019849648857100808, "loss": 4.4077, "step": 218650 }, { "epoch": 0.40811851672152255, "grad_norm": 0.7868063449859619, "learning_rate": 0.00019849576274505164, "loss": 4.6604, "step": 218700 }, { "epoch": 0.408211822280901, "grad_norm": 1.2350362539291382, "learning_rate": 0.00019849503674526756, "loss": 4.5138, "step": 218750 }, { "epoch": 0.40830512784027956, "grad_norm": 1.3599216938018799, "learning_rate": 0.00019849431057165703, "loss": 4.6163, "step": 218800 }, { "epoch": 0.40839843339965803, "grad_norm": 0.9188289642333984, "learning_rate": 0.0001984935842242214, "loss": 4.6306, "step": 218850 }, { "epoch": 0.4084917389590365, "grad_norm": 0.9845192432403564, "learning_rate": 0.00019849285770296189, "loss": 4.656, "step": 218900 }, { "epoch": 0.40858504451841504, "grad_norm": 1.1785905361175537, "learning_rate": 0.00019849213100787987, "loss": 4.6391, "step": 218950 }, { "epoch": 0.4086783500777935, "grad_norm": 1.0496912002563477, "learning_rate": 0.00019849140413897654, "loss": 4.4645, "step": 219000 }, { "epoch": 0.4086783500777935, "eval_loss": 4.787511825561523, "eval_runtime": 234.8152, "eval_samples_per_second": 11.107, "eval_steps_per_second": 11.107, "eval_tts_loss": 7.520304385868353, "step": 219000 }, { "epoch": 0.408771655637172, "grad_norm": 1.1836047172546387, "learning_rate": 0.0001984906770962532, "loss": 4.5759, "step": 219050 }, { "epoch": 0.40886496119655047, "grad_norm": 0.866492748260498, "learning_rate": 0.00019848994987971117, "loss": 4.7295, "step": 219100 }, { "epoch": 0.408958266755929, "grad_norm": 0.7800747752189636, "learning_rate": 0.00019848922248935167, "loss": 4.6962, "step": 219150 }, { "epoch": 0.4090515723153075, "grad_norm": 1.0174862146377563, "learning_rate": 0.00019848849492517604, "loss": 4.4633, "step": 219200 }, { "epoch": 0.40914487787468595, "grad_norm": 1.1767679452896118, "learning_rate": 0.00019848776718718556, "loss": 4.6195, "step": 219250 }, { "epoch": 0.4092381834340645, "grad_norm": 1.1328825950622559, "learning_rate": 0.00019848703927538152, "loss": 4.518, "step": 219300 }, { "epoch": 0.40933148899344296, "grad_norm": 0.7945483326911926, "learning_rate": 0.00019848631118976514, "loss": 4.7111, "step": 219350 }, { "epoch": 0.40942479455282144, "grad_norm": 1.0717135667800903, "learning_rate": 0.00019848558293033777, "loss": 4.6445, "step": 219400 }, { "epoch": 0.4095181001121999, "grad_norm": 0.9013932347297668, "learning_rate": 0.00019848485449710067, "loss": 4.511, "step": 219450 }, { "epoch": 0.40961140567157844, "grad_norm": 1.1466339826583862, "learning_rate": 0.00019848412589005513, "loss": 4.4065, "step": 219500 }, { "epoch": 0.4097047112309569, "grad_norm": 1.0499449968338013, "learning_rate": 0.00019848339710920246, "loss": 4.489, "step": 219550 }, { "epoch": 0.4097980167903354, "grad_norm": 0.9530263543128967, "learning_rate": 0.0001984826681545439, "loss": 4.8684, "step": 219600 }, { "epoch": 0.4098913223497139, "grad_norm": 0.9534028768539429, "learning_rate": 0.00019848193902608077, "loss": 4.5792, "step": 219650 }, { "epoch": 0.4099846279090924, "grad_norm": 1.157947063446045, "learning_rate": 0.00019848120972381432, "loss": 4.5243, "step": 219700 }, { "epoch": 0.4100779334684709, "grad_norm": 1.022988200187683, "learning_rate": 0.00019848048024774588, "loss": 4.7144, "step": 219750 }, { "epoch": 0.41017123902784935, "grad_norm": 1.097163438796997, "learning_rate": 0.00019847975059787674, "loss": 4.8489, "step": 219800 }, { "epoch": 0.4102645445872279, "grad_norm": 1.2663757801055908, "learning_rate": 0.00019847902077420814, "loss": 4.629, "step": 219850 }, { "epoch": 0.41035785014660636, "grad_norm": 1.103257656097412, "learning_rate": 0.00019847829077674142, "loss": 4.6784, "step": 219900 }, { "epoch": 0.41045115570598484, "grad_norm": 1.0042039155960083, "learning_rate": 0.00019847756060547783, "loss": 4.6846, "step": 219950 }, { "epoch": 0.41054446126536337, "grad_norm": 0.9465295076370239, "learning_rate": 0.00019847683026041869, "loss": 4.3015, "step": 220000 }, { "epoch": 0.41063776682474185, "grad_norm": 0.9060609936714172, "learning_rate": 0.00019847609974156524, "loss": 4.3934, "step": 220050 }, { "epoch": 0.4107310723841203, "grad_norm": 0.7724514603614807, "learning_rate": 0.00019847536904891882, "loss": 4.617, "step": 220100 }, { "epoch": 0.4108243779434988, "grad_norm": 1.2579684257507324, "learning_rate": 0.00019847463818248068, "loss": 4.756, "step": 220150 }, { "epoch": 0.41091768350287733, "grad_norm": 0.7629663944244385, "learning_rate": 0.00019847390714225217, "loss": 4.5186, "step": 220200 }, { "epoch": 0.4110109890622558, "grad_norm": 0.8361650109291077, "learning_rate": 0.0001984731759282345, "loss": 4.6316, "step": 220250 }, { "epoch": 0.4111042946216343, "grad_norm": 1.2181036472320557, "learning_rate": 0.00019847244454042904, "loss": 4.7891, "step": 220300 }, { "epoch": 0.4111976001810128, "grad_norm": 0.9293192028999329, "learning_rate": 0.00019847171297883703, "loss": 4.4564, "step": 220350 }, { "epoch": 0.4112909057403913, "grad_norm": 0.9219282269477844, "learning_rate": 0.00019847098124345977, "loss": 4.5263, "step": 220400 }, { "epoch": 0.41138421129976976, "grad_norm": 1.1117901802062988, "learning_rate": 0.00019847024933429856, "loss": 4.6565, "step": 220450 }, { "epoch": 0.41147751685914824, "grad_norm": 1.5147013664245605, "learning_rate": 0.00019846951725135468, "loss": 4.8577, "step": 220500 }, { "epoch": 0.41157082241852677, "grad_norm": 1.095376968383789, "learning_rate": 0.00019846878499462942, "loss": 4.4866, "step": 220550 }, { "epoch": 0.41166412797790525, "grad_norm": 0.8250828385353088, "learning_rate": 0.00019846805256412407, "loss": 4.6568, "step": 220600 }, { "epoch": 0.4117574335372837, "grad_norm": 1.0293447971343994, "learning_rate": 0.00019846731995983995, "loss": 4.8399, "step": 220650 }, { "epoch": 0.4118507390966622, "grad_norm": 1.0483900308609009, "learning_rate": 0.00019846658718177836, "loss": 4.7965, "step": 220700 }, { "epoch": 0.41194404465604073, "grad_norm": 0.627288281917572, "learning_rate": 0.00019846585422994055, "loss": 4.1862, "step": 220750 }, { "epoch": 0.4120373502154192, "grad_norm": 1.12353515625, "learning_rate": 0.0001984651211043278, "loss": 4.7008, "step": 220800 }, { "epoch": 0.4121306557747977, "grad_norm": 1.2579227685928345, "learning_rate": 0.00019846438780494148, "loss": 4.5624, "step": 220850 }, { "epoch": 0.4122239613341762, "grad_norm": 1.1707371473312378, "learning_rate": 0.00019846365433178284, "loss": 4.6096, "step": 220900 }, { "epoch": 0.4123172668935547, "grad_norm": 0.7681998610496521, "learning_rate": 0.00019846292068485317, "loss": 4.6948, "step": 220950 }, { "epoch": 0.41241057245293317, "grad_norm": 0.5637853741645813, "learning_rate": 0.00019846218686415376, "loss": 4.5245, "step": 221000 }, { "epoch": 0.41250387801231164, "grad_norm": 0.9706264734268188, "learning_rate": 0.0001984614528696859, "loss": 4.6419, "step": 221050 }, { "epoch": 0.4125971835716902, "grad_norm": 1.179498553276062, "learning_rate": 0.00019846071870145094, "loss": 4.4572, "step": 221100 }, { "epoch": 0.41269048913106865, "grad_norm": 1.05915105342865, "learning_rate": 0.0001984599843594501, "loss": 4.7192, "step": 221150 }, { "epoch": 0.4127837946904471, "grad_norm": 1.2832460403442383, "learning_rate": 0.00019845924984368473, "loss": 4.632, "step": 221200 }, { "epoch": 0.41287710024982566, "grad_norm": 0.977168083190918, "learning_rate": 0.0001984585151541561, "loss": 4.6565, "step": 221250 }, { "epoch": 0.41297040580920413, "grad_norm": 0.9659879207611084, "learning_rate": 0.00019845778029086553, "loss": 4.5208, "step": 221300 }, { "epoch": 0.4130637113685826, "grad_norm": 1.013081669807434, "learning_rate": 0.0001984570452538143, "loss": 4.4713, "step": 221350 }, { "epoch": 0.4131570169279611, "grad_norm": 0.9163434505462646, "learning_rate": 0.00019845631004300368, "loss": 4.6295, "step": 221400 }, { "epoch": 0.4132503224873396, "grad_norm": 0.9959138631820679, "learning_rate": 0.00019845557465843504, "loss": 4.497, "step": 221450 }, { "epoch": 0.4133436280467181, "grad_norm": 1.0366419553756714, "learning_rate": 0.0001984548391001096, "loss": 4.6001, "step": 221500 }, { "epoch": 0.41343693360609657, "grad_norm": 0.8534669280052185, "learning_rate": 0.00019845410336802872, "loss": 4.6079, "step": 221550 }, { "epoch": 0.4135302391654751, "grad_norm": 1.0028185844421387, "learning_rate": 0.00019845336746219364, "loss": 4.5135, "step": 221600 }, { "epoch": 0.4136235447248536, "grad_norm": 0.957848846912384, "learning_rate": 0.0001984526313826057, "loss": 4.5515, "step": 221650 }, { "epoch": 0.41371685028423205, "grad_norm": 0.7522804141044617, "learning_rate": 0.00019845189512926622, "loss": 4.679, "step": 221700 }, { "epoch": 0.4138101558436105, "grad_norm": 0.7022336721420288, "learning_rate": 0.00019845115870217644, "loss": 4.642, "step": 221750 }, { "epoch": 0.41390346140298906, "grad_norm": 1.0571622848510742, "learning_rate": 0.0001984504221013377, "loss": 4.6447, "step": 221800 }, { "epoch": 0.41399676696236754, "grad_norm": 0.7391085624694824, "learning_rate": 0.00019844968532675125, "loss": 4.3801, "step": 221850 }, { "epoch": 0.414090072521746, "grad_norm": 0.9244349598884583, "learning_rate": 0.00019844894837841846, "loss": 4.5394, "step": 221900 }, { "epoch": 0.41418337808112454, "grad_norm": 0.9880529642105103, "learning_rate": 0.00019844821125634062, "loss": 4.4614, "step": 221950 }, { "epoch": 0.414276683640503, "grad_norm": 1.040346622467041, "learning_rate": 0.00019844747396051898, "loss": 4.574, "step": 222000 }, { "epoch": 0.414276683640503, "eval_loss": 4.79351282119751, "eval_runtime": 232.2857, "eval_samples_per_second": 11.228, "eval_steps_per_second": 11.228, "eval_tts_loss": 7.543720753433503, "step": 222000 }, { "epoch": 0.4143699891998815, "grad_norm": 0.9894281029701233, "learning_rate": 0.00019844673649095486, "loss": 4.5986, "step": 222050 }, { "epoch": 0.41446329475925997, "grad_norm": 0.9158909320831299, "learning_rate": 0.0001984459988476496, "loss": 4.7346, "step": 222100 }, { "epoch": 0.4145566003186385, "grad_norm": 0.9308950901031494, "learning_rate": 0.00019844526103060447, "loss": 4.5288, "step": 222150 }, { "epoch": 0.414649905878017, "grad_norm": 0.967197597026825, "learning_rate": 0.00019844452303982075, "loss": 4.6904, "step": 222200 }, { "epoch": 0.41474321143739545, "grad_norm": 0.8916376233100891, "learning_rate": 0.0001984437848752998, "loss": 4.4642, "step": 222250 }, { "epoch": 0.414836516996774, "grad_norm": 1.0411932468414307, "learning_rate": 0.00019844304653704285, "loss": 4.6579, "step": 222300 }, { "epoch": 0.41492982255615246, "grad_norm": 0.9220390915870667, "learning_rate": 0.0001984423080250513, "loss": 4.6703, "step": 222350 }, { "epoch": 0.41502312811553094, "grad_norm": 1.0786405801773071, "learning_rate": 0.00019844156933932635, "loss": 4.5595, "step": 222400 }, { "epoch": 0.4151164336749094, "grad_norm": 1.1047947406768799, "learning_rate": 0.0001984408304798694, "loss": 4.6576, "step": 222450 }, { "epoch": 0.41520973923428794, "grad_norm": 0.8863925933837891, "learning_rate": 0.00019844009144668167, "loss": 4.6185, "step": 222500 }, { "epoch": 0.4153030447936664, "grad_norm": 0.9184449911117554, "learning_rate": 0.0001984393522397645, "loss": 4.7059, "step": 222550 }, { "epoch": 0.4153963503530449, "grad_norm": 0.9710171818733215, "learning_rate": 0.00019843861285911922, "loss": 4.5444, "step": 222600 }, { "epoch": 0.41548965591242343, "grad_norm": 1.0862765312194824, "learning_rate": 0.00019843787330474708, "loss": 4.6252, "step": 222650 }, { "epoch": 0.4155829614718019, "grad_norm": 0.7898451685905457, "learning_rate": 0.00019843713357664943, "loss": 4.8069, "step": 222700 }, { "epoch": 0.4156762670311804, "grad_norm": 0.9102611541748047, "learning_rate": 0.00019843639367482757, "loss": 4.5326, "step": 222750 }, { "epoch": 0.41576957259055886, "grad_norm": 0.8616012334823608, "learning_rate": 0.00019843565359928275, "loss": 4.4656, "step": 222800 }, { "epoch": 0.4158628781499374, "grad_norm": 1.0981239080429077, "learning_rate": 0.00019843491335001638, "loss": 4.4092, "step": 222850 }, { "epoch": 0.41595618370931586, "grad_norm": 1.237775444984436, "learning_rate": 0.00019843417292702967, "loss": 4.5823, "step": 222900 }, { "epoch": 0.41604948926869434, "grad_norm": 1.199179768562317, "learning_rate": 0.000198433432330324, "loss": 4.7258, "step": 222950 }, { "epoch": 0.41614279482807287, "grad_norm": 1.1261119842529297, "learning_rate": 0.00019843269155990059, "loss": 4.5042, "step": 223000 }, { "epoch": 0.41623610038745135, "grad_norm": 1.0203489065170288, "learning_rate": 0.00019843195061576083, "loss": 4.49, "step": 223050 }, { "epoch": 0.4163294059468298, "grad_norm": 0.8318331837654114, "learning_rate": 0.000198431209497906, "loss": 4.5679, "step": 223100 }, { "epoch": 0.4164227115062083, "grad_norm": 1.0741044282913208, "learning_rate": 0.0001984304682063374, "loss": 4.6131, "step": 223150 }, { "epoch": 0.41651601706558683, "grad_norm": 1.0485063791275024, "learning_rate": 0.00019842972674105632, "loss": 4.4938, "step": 223200 }, { "epoch": 0.4166093226249653, "grad_norm": 0.9604942202568054, "learning_rate": 0.0001984289851020641, "loss": 4.536, "step": 223250 }, { "epoch": 0.4167026281843438, "grad_norm": 1.0192519426345825, "learning_rate": 0.0001984282432893621, "loss": 4.5778, "step": 223300 }, { "epoch": 0.41679593374372226, "grad_norm": 1.1776680946350098, "learning_rate": 0.0001984275013029515, "loss": 4.359, "step": 223350 }, { "epoch": 0.4168892393031008, "grad_norm": 0.999553382396698, "learning_rate": 0.00019842675914283372, "loss": 4.5106, "step": 223400 }, { "epoch": 0.41698254486247927, "grad_norm": 0.972808301448822, "learning_rate": 0.00019842601680901, "loss": 4.5058, "step": 223450 }, { "epoch": 0.41707585042185774, "grad_norm": 0.6613345146179199, "learning_rate": 0.0001984252743014817, "loss": 4.5789, "step": 223500 }, { "epoch": 0.4171691559812363, "grad_norm": 0.9558602571487427, "learning_rate": 0.00019842453162025007, "loss": 4.7024, "step": 223550 }, { "epoch": 0.41726246154061475, "grad_norm": 1.245826244354248, "learning_rate": 0.0001984237887653165, "loss": 4.5732, "step": 223600 }, { "epoch": 0.4173557670999932, "grad_norm": 1.244626760482788, "learning_rate": 0.00019842304573668224, "loss": 4.5236, "step": 223650 }, { "epoch": 0.4174490726593717, "grad_norm": 1.0705629587173462, "learning_rate": 0.00019842230253434864, "loss": 4.6263, "step": 223700 }, { "epoch": 0.41754237821875023, "grad_norm": 1.0318876504898071, "learning_rate": 0.00019842155915831698, "loss": 4.795, "step": 223750 }, { "epoch": 0.4176356837781287, "grad_norm": 1.3221536874771118, "learning_rate": 0.00019842081560858858, "loss": 4.7132, "step": 223800 }, { "epoch": 0.4177289893375072, "grad_norm": 0.8112679719924927, "learning_rate": 0.0001984200718851648, "loss": 4.4579, "step": 223850 }, { "epoch": 0.4178222948968857, "grad_norm": 1.2415705919265747, "learning_rate": 0.00019841932798804685, "loss": 4.7319, "step": 223900 }, { "epoch": 0.4179156004562642, "grad_norm": 1.0489301681518555, "learning_rate": 0.00019841858391723616, "loss": 4.7578, "step": 223950 }, { "epoch": 0.41800890601564267, "grad_norm": 1.1561607122421265, "learning_rate": 0.00019841783967273395, "loss": 4.5376, "step": 224000 }, { "epoch": 0.41810221157502114, "grad_norm": 1.2369205951690674, "learning_rate": 0.00019841709525454158, "loss": 4.506, "step": 224050 }, { "epoch": 0.4181955171343997, "grad_norm": 1.139877438545227, "learning_rate": 0.00019841635066266035, "loss": 4.7282, "step": 224100 }, { "epoch": 0.41828882269377815, "grad_norm": 0.8672619462013245, "learning_rate": 0.00019841560589709158, "loss": 4.7663, "step": 224150 }, { "epoch": 0.4183821282531566, "grad_norm": 1.0294967889785767, "learning_rate": 0.0001984148609578366, "loss": 4.7002, "step": 224200 }, { "epoch": 0.41847543381253516, "grad_norm": 1.0474852323532104, "learning_rate": 0.00019841411584489672, "loss": 4.4804, "step": 224250 }, { "epoch": 0.41856873937191363, "grad_norm": 1.1958084106445312, "learning_rate": 0.00019841337055827318, "loss": 4.4729, "step": 224300 }, { "epoch": 0.4186620449312921, "grad_norm": 0.6683170795440674, "learning_rate": 0.0001984126250979674, "loss": 4.5723, "step": 224350 }, { "epoch": 0.4187553504906706, "grad_norm": 1.3585337400436401, "learning_rate": 0.00019841187946398065, "loss": 4.6204, "step": 224400 }, { "epoch": 0.4188486560500491, "grad_norm": 0.8530429005622864, "learning_rate": 0.00019841113365631424, "loss": 4.7279, "step": 224450 }, { "epoch": 0.4189419616094276, "grad_norm": 0.943585216999054, "learning_rate": 0.00019841038767496948, "loss": 4.5772, "step": 224500 }, { "epoch": 0.41903526716880607, "grad_norm": 1.0253528356552124, "learning_rate": 0.00019840964151994774, "loss": 4.6783, "step": 224550 }, { "epoch": 0.4191285727281846, "grad_norm": 1.0521509647369385, "learning_rate": 0.00019840889519125027, "loss": 4.6578, "step": 224600 }, { "epoch": 0.4192218782875631, "grad_norm": 0.9954854249954224, "learning_rate": 0.00019840814868887842, "loss": 4.7111, "step": 224650 }, { "epoch": 0.41931518384694155, "grad_norm": 1.1058915853500366, "learning_rate": 0.00019840740201283353, "loss": 4.4783, "step": 224700 }, { "epoch": 0.41940848940632003, "grad_norm": 1.0141825675964355, "learning_rate": 0.00019840665516311688, "loss": 4.603, "step": 224750 }, { "epoch": 0.41950179496569856, "grad_norm": 0.7934043407440186, "learning_rate": 0.00019840590813972978, "loss": 4.6601, "step": 224800 }, { "epoch": 0.41959510052507704, "grad_norm": 1.1680604219436646, "learning_rate": 0.00019840516094267357, "loss": 4.5609, "step": 224850 }, { "epoch": 0.4196884060844555, "grad_norm": 0.995583713054657, "learning_rate": 0.00019840441357194956, "loss": 4.5134, "step": 224900 }, { "epoch": 0.41978171164383404, "grad_norm": 0.9852712750434875, "learning_rate": 0.00019840366602755908, "loss": 4.5728, "step": 224950 }, { "epoch": 0.4198750172032125, "grad_norm": 1.1983600854873657, "learning_rate": 0.00019840291830950345, "loss": 4.5515, "step": 225000 }, { "epoch": 0.4198750172032125, "eval_loss": 4.793398380279541, "eval_runtime": 231.8067, "eval_samples_per_second": 11.251, "eval_steps_per_second": 11.251, "eval_tts_loss": 7.515776119422042, "step": 225000 }, { "epoch": 0.419968322762591, "grad_norm": 1.0381780862808228, "learning_rate": 0.000198402170417784, "loss": 4.4676, "step": 225050 }, { "epoch": 0.42006162832196947, "grad_norm": 1.10462486743927, "learning_rate": 0.000198401422352402, "loss": 4.3191, "step": 225100 }, { "epoch": 0.420154933881348, "grad_norm": 1.2322332859039307, "learning_rate": 0.00019840067411335883, "loss": 4.7742, "step": 225150 }, { "epoch": 0.4202482394407265, "grad_norm": 0.9938576817512512, "learning_rate": 0.0001983999257006558, "loss": 4.5088, "step": 225200 }, { "epoch": 0.42034154500010495, "grad_norm": 0.9337325692176819, "learning_rate": 0.00019839917711429416, "loss": 4.6499, "step": 225250 }, { "epoch": 0.4204348505594835, "grad_norm": 0.9972164630889893, "learning_rate": 0.0001983984283542753, "loss": 4.5516, "step": 225300 }, { "epoch": 0.42052815611886196, "grad_norm": 1.3963088989257812, "learning_rate": 0.00019839767942060057, "loss": 4.5577, "step": 225350 }, { "epoch": 0.42062146167824044, "grad_norm": 1.1737512350082397, "learning_rate": 0.0001983969303132712, "loss": 4.4765, "step": 225400 }, { "epoch": 0.4207147672376189, "grad_norm": 0.8017054796218872, "learning_rate": 0.0001983961810322886, "loss": 4.524, "step": 225450 }, { "epoch": 0.42080807279699745, "grad_norm": 1.0323697328567505, "learning_rate": 0.00019839543157765404, "loss": 4.3256, "step": 225500 }, { "epoch": 0.4209013783563759, "grad_norm": 1.0249814987182617, "learning_rate": 0.00019839468194936887, "loss": 4.7338, "step": 225550 }, { "epoch": 0.4209946839157544, "grad_norm": 1.0047541856765747, "learning_rate": 0.00019839393214743435, "loss": 4.4109, "step": 225600 }, { "epoch": 0.42108798947513293, "grad_norm": 1.2357292175292969, "learning_rate": 0.0001983931821718519, "loss": 4.683, "step": 225650 }, { "epoch": 0.4211812950345114, "grad_norm": 1.0345031023025513, "learning_rate": 0.0001983924320226228, "loss": 4.5459, "step": 225700 }, { "epoch": 0.4212746005938899, "grad_norm": 0.9663260579109192, "learning_rate": 0.00019839168169974835, "loss": 4.7504, "step": 225750 }, { "epoch": 0.42136790615326836, "grad_norm": 1.1099520921707153, "learning_rate": 0.0001983909312032299, "loss": 4.435, "step": 225800 }, { "epoch": 0.4214612117126469, "grad_norm": 0.9571065902709961, "learning_rate": 0.00019839018053306875, "loss": 4.7323, "step": 225850 }, { "epoch": 0.42155451727202536, "grad_norm": 1.3123244047164917, "learning_rate": 0.00019838942968926625, "loss": 4.6093, "step": 225900 }, { "epoch": 0.42164782283140384, "grad_norm": 0.9905040860176086, "learning_rate": 0.00019838867867182373, "loss": 4.6339, "step": 225950 }, { "epoch": 0.4217411283907823, "grad_norm": 0.8658848404884338, "learning_rate": 0.0001983879274807425, "loss": 4.7805, "step": 226000 }, { "epoch": 0.42183443395016085, "grad_norm": 1.1252857446670532, "learning_rate": 0.0001983871761160239, "loss": 4.3846, "step": 226050 }, { "epoch": 0.4219277395095393, "grad_norm": 0.9474992752075195, "learning_rate": 0.00019838642457766924, "loss": 4.5989, "step": 226100 }, { "epoch": 0.4220210450689178, "grad_norm": 1.1194905042648315, "learning_rate": 0.00019838567286567985, "loss": 4.5753, "step": 226150 }, { "epoch": 0.42211435062829633, "grad_norm": 1.173120379447937, "learning_rate": 0.00019838492098005705, "loss": 4.6295, "step": 226200 }, { "epoch": 0.4222076561876748, "grad_norm": 0.8689015507698059, "learning_rate": 0.0001983841689208022, "loss": 4.6237, "step": 226250 }, { "epoch": 0.4223009617470533, "grad_norm": 1.1971979141235352, "learning_rate": 0.0001983834166879166, "loss": 4.5907, "step": 226300 }, { "epoch": 0.42239426730643176, "grad_norm": 1.1240501403808594, "learning_rate": 0.00019838266428140155, "loss": 4.6409, "step": 226350 }, { "epoch": 0.4224875728658103, "grad_norm": 0.8118560314178467, "learning_rate": 0.00019838191170125843, "loss": 4.4653, "step": 226400 }, { "epoch": 0.42258087842518877, "grad_norm": 0.9946422576904297, "learning_rate": 0.00019838115894748858, "loss": 4.542, "step": 226450 }, { "epoch": 0.42267418398456724, "grad_norm": 0.7898224592208862, "learning_rate": 0.00019838040602009323, "loss": 4.5506, "step": 226500 }, { "epoch": 0.4227674895439458, "grad_norm": 0.921465277671814, "learning_rate": 0.00019837965291907378, "loss": 4.6318, "step": 226550 }, { "epoch": 0.42286079510332425, "grad_norm": 1.0502792596817017, "learning_rate": 0.00019837889964443163, "loss": 4.5862, "step": 226600 }, { "epoch": 0.4229541006627027, "grad_norm": 0.8788375854492188, "learning_rate": 0.00019837814619616794, "loss": 4.6376, "step": 226650 }, { "epoch": 0.4230474062220812, "grad_norm": 0.8763489127159119, "learning_rate": 0.00019837739257428418, "loss": 4.8787, "step": 226700 }, { "epoch": 0.42314071178145973, "grad_norm": 0.9592618346214294, "learning_rate": 0.00019837663877878163, "loss": 4.5449, "step": 226750 }, { "epoch": 0.4232340173408382, "grad_norm": 0.5952169895172119, "learning_rate": 0.0001983758848096616, "loss": 4.506, "step": 226800 }, { "epoch": 0.4233273229002167, "grad_norm": 1.1658716201782227, "learning_rate": 0.00019837513066692548, "loss": 4.5342, "step": 226850 }, { "epoch": 0.4234206284595952, "grad_norm": 1.1232601404190063, "learning_rate": 0.00019837437635057453, "loss": 4.5859, "step": 226900 }, { "epoch": 0.4235139340189737, "grad_norm": 1.1959850788116455, "learning_rate": 0.00019837362186061013, "loss": 4.5792, "step": 226950 }, { "epoch": 0.42360723957835217, "grad_norm": 0.9566145539283752, "learning_rate": 0.0001983728671970336, "loss": 4.4619, "step": 227000 }, { "epoch": 0.42370054513773064, "grad_norm": 1.128745675086975, "learning_rate": 0.00019837211235984627, "loss": 4.6829, "step": 227050 }, { "epoch": 0.4237938506971092, "grad_norm": 1.116195559501648, "learning_rate": 0.00019837135734904947, "loss": 4.7385, "step": 227100 }, { "epoch": 0.42388715625648765, "grad_norm": 0.8514589667320251, "learning_rate": 0.00019837060216464452, "loss": 4.5505, "step": 227150 }, { "epoch": 0.4239804618158661, "grad_norm": 1.0022069215774536, "learning_rate": 0.00019836984680663276, "loss": 4.6128, "step": 227200 }, { "epoch": 0.42407376737524466, "grad_norm": 0.8854489326477051, "learning_rate": 0.00019836909127501554, "loss": 4.198, "step": 227250 }, { "epoch": 0.42416707293462313, "grad_norm": 0.9211372137069702, "learning_rate": 0.00019836833556979418, "loss": 4.5983, "step": 227300 }, { "epoch": 0.4242603784940016, "grad_norm": 0.9548203349113464, "learning_rate": 0.00019836757969097005, "loss": 4.5211, "step": 227350 }, { "epoch": 0.4243536840533801, "grad_norm": 0.9231182336807251, "learning_rate": 0.0001983668236385444, "loss": 4.5906, "step": 227400 }, { "epoch": 0.4244469896127586, "grad_norm": 1.0974338054656982, "learning_rate": 0.00019836606741251863, "loss": 4.8236, "step": 227450 }, { "epoch": 0.4245402951721371, "grad_norm": 1.2579002380371094, "learning_rate": 0.00019836531101289408, "loss": 4.7423, "step": 227500 }, { "epoch": 0.42463360073151557, "grad_norm": 0.9564178586006165, "learning_rate": 0.00019836455443967204, "loss": 4.5849, "step": 227550 }, { "epoch": 0.4247269062908941, "grad_norm": 0.7790428996086121, "learning_rate": 0.0001983637976928539, "loss": 4.4314, "step": 227600 }, { "epoch": 0.4248202118502726, "grad_norm": 1.0943279266357422, "learning_rate": 0.0001983630407724409, "loss": 4.4287, "step": 227650 }, { "epoch": 0.42491351740965105, "grad_norm": 0.8651321530342102, "learning_rate": 0.0001983622836784345, "loss": 4.6651, "step": 227700 }, { "epoch": 0.42500682296902953, "grad_norm": 1.2312800884246826, "learning_rate": 0.00019836152641083596, "loss": 4.6271, "step": 227750 }, { "epoch": 0.42510012852840806, "grad_norm": 0.8795189261436462, "learning_rate": 0.00019836076896964663, "loss": 4.7112, "step": 227800 }, { "epoch": 0.42519343408778654, "grad_norm": 0.861788809299469, "learning_rate": 0.00019836001135486786, "loss": 4.5759, "step": 227850 }, { "epoch": 0.425286739647165, "grad_norm": 0.9528782963752747, "learning_rate": 0.00019835925356650097, "loss": 4.6581, "step": 227900 }, { "epoch": 0.42538004520654354, "grad_norm": 0.9587149024009705, "learning_rate": 0.00019835849560454732, "loss": 4.3163, "step": 227950 }, { "epoch": 0.425473350765922, "grad_norm": 1.0508127212524414, "learning_rate": 0.00019835773746900822, "loss": 4.4864, "step": 228000 }, { "epoch": 0.425473350765922, "eval_loss": 4.791882514953613, "eval_runtime": 232.0509, "eval_samples_per_second": 11.239, "eval_steps_per_second": 11.239, "eval_tts_loss": 7.540685358295857, "step": 228000 }, { "epoch": 0.4255666563253005, "grad_norm": 1.0788441896438599, "learning_rate": 0.00019835697915988502, "loss": 4.6176, "step": 228050 }, { "epoch": 0.425659961884679, "grad_norm": 0.8570576310157776, "learning_rate": 0.00019835622067717904, "loss": 4.8553, "step": 228100 }, { "epoch": 0.4257532674440575, "grad_norm": 0.7701833844184875, "learning_rate": 0.00019835546202089165, "loss": 4.5462, "step": 228150 }, { "epoch": 0.425846573003436, "grad_norm": 0.9959319233894348, "learning_rate": 0.0001983547031910242, "loss": 4.4326, "step": 228200 }, { "epoch": 0.42593987856281446, "grad_norm": 0.9818710088729858, "learning_rate": 0.00019835394418757797, "loss": 4.7489, "step": 228250 }, { "epoch": 0.426033184122193, "grad_norm": 0.932654082775116, "learning_rate": 0.00019835318501055437, "loss": 4.571, "step": 228300 }, { "epoch": 0.42612648968157146, "grad_norm": 1.0089269876480103, "learning_rate": 0.00019835242565995466, "loss": 4.6125, "step": 228350 }, { "epoch": 0.42621979524094994, "grad_norm": 0.9701918363571167, "learning_rate": 0.00019835166613578026, "loss": 4.4908, "step": 228400 }, { "epoch": 0.4263131008003284, "grad_norm": 0.9656922221183777, "learning_rate": 0.00019835090643803246, "loss": 4.7232, "step": 228450 }, { "epoch": 0.42640640635970695, "grad_norm": 0.9137131571769714, "learning_rate": 0.0001983501465667126, "loss": 4.6645, "step": 228500 }, { "epoch": 0.4264997119190854, "grad_norm": 0.9897300601005554, "learning_rate": 0.00019834938652182204, "loss": 4.6834, "step": 228550 }, { "epoch": 0.4265930174784639, "grad_norm": 0.9887792468070984, "learning_rate": 0.00019834862630336214, "loss": 4.5679, "step": 228600 }, { "epoch": 0.42668632303784243, "grad_norm": 0.6781644821166992, "learning_rate": 0.0001983478659113342, "loss": 4.5704, "step": 228650 }, { "epoch": 0.4267796285972209, "grad_norm": 0.9480149745941162, "learning_rate": 0.00019834710534573962, "loss": 4.5627, "step": 228700 }, { "epoch": 0.4268729341565994, "grad_norm": 0.8901644945144653, "learning_rate": 0.00019834634460657964, "loss": 4.6213, "step": 228750 }, { "epoch": 0.42696623971597786, "grad_norm": 0.7598466873168945, "learning_rate": 0.00019834558369385573, "loss": 4.4153, "step": 228800 }, { "epoch": 0.4270595452753564, "grad_norm": 1.0075832605361938, "learning_rate": 0.00019834482260756912, "loss": 4.6078, "step": 228850 }, { "epoch": 0.42715285083473487, "grad_norm": 0.9788805842399597, "learning_rate": 0.00019834406134772124, "loss": 4.5693, "step": 228900 }, { "epoch": 0.42724615639411334, "grad_norm": 1.0992543697357178, "learning_rate": 0.00019834329991431338, "loss": 4.4368, "step": 228950 }, { "epoch": 0.4273394619534918, "grad_norm": 1.0395466089248657, "learning_rate": 0.0001983425383073469, "loss": 4.6703, "step": 229000 }, { "epoch": 0.42743276751287035, "grad_norm": 0.8829817771911621, "learning_rate": 0.00019834177652682314, "loss": 4.6088, "step": 229050 }, { "epoch": 0.4275260730722488, "grad_norm": 1.060113787651062, "learning_rate": 0.00019834101457274342, "loss": 4.5978, "step": 229100 }, { "epoch": 0.4276193786316273, "grad_norm": 1.1380964517593384, "learning_rate": 0.00019834025244510917, "loss": 4.5446, "step": 229150 }, { "epoch": 0.42771268419100583, "grad_norm": 1.2462329864501953, "learning_rate": 0.00019833949014392164, "loss": 4.3495, "step": 229200 }, { "epoch": 0.4278059897503843, "grad_norm": 0.9149359464645386, "learning_rate": 0.00019833872766918223, "loss": 4.5514, "step": 229250 }, { "epoch": 0.4278992953097628, "grad_norm": 0.770584762096405, "learning_rate": 0.00019833796502089225, "loss": 4.6721, "step": 229300 }, { "epoch": 0.42799260086914126, "grad_norm": 0.7645120024681091, "learning_rate": 0.00019833720219905308, "loss": 4.5611, "step": 229350 }, { "epoch": 0.4280859064285198, "grad_norm": 0.9542115330696106, "learning_rate": 0.00019833643920366605, "loss": 4.5287, "step": 229400 }, { "epoch": 0.42817921198789827, "grad_norm": 1.430662989616394, "learning_rate": 0.0001983356760347325, "loss": 4.4284, "step": 229450 }, { "epoch": 0.42827251754727674, "grad_norm": 0.8985748291015625, "learning_rate": 0.0001983349126922538, "loss": 4.5965, "step": 229500 }, { "epoch": 0.4283658231066553, "grad_norm": 0.901803195476532, "learning_rate": 0.00019833414917623128, "loss": 4.7505, "step": 229550 }, { "epoch": 0.42845912866603375, "grad_norm": 1.0440247058868408, "learning_rate": 0.0001983333854866663, "loss": 4.5123, "step": 229600 }, { "epoch": 0.4285524342254122, "grad_norm": 0.8303043842315674, "learning_rate": 0.00019833262162356015, "loss": 4.4886, "step": 229650 }, { "epoch": 0.4286457397847907, "grad_norm": 1.0328497886657715, "learning_rate": 0.00019833185758691427, "loss": 4.594, "step": 229700 }, { "epoch": 0.42873904534416923, "grad_norm": 1.0960317850112915, "learning_rate": 0.00019833109337672995, "loss": 4.5405, "step": 229750 }, { "epoch": 0.4288323509035477, "grad_norm": 0.9300960898399353, "learning_rate": 0.00019833032899300854, "loss": 4.7002, "step": 229800 }, { "epoch": 0.4289256564629262, "grad_norm": 1.045456051826477, "learning_rate": 0.0001983295644357514, "loss": 4.5931, "step": 229850 }, { "epoch": 0.4290189620223047, "grad_norm": 1.1042137145996094, "learning_rate": 0.0001983287997049599, "loss": 4.7307, "step": 229900 }, { "epoch": 0.4291122675816832, "grad_norm": 1.2343262434005737, "learning_rate": 0.00019832803480063538, "loss": 4.621, "step": 229950 }, { "epoch": 0.42920557314106167, "grad_norm": 1.0615925788879395, "learning_rate": 0.00019832726972277915, "loss": 4.6954, "step": 230000 }, { "epoch": 0.42929887870044015, "grad_norm": 1.0219675302505493, "learning_rate": 0.00019832650447139258, "loss": 4.6932, "step": 230050 }, { "epoch": 0.4293921842598187, "grad_norm": 0.7742038369178772, "learning_rate": 0.00019832573904647707, "loss": 4.649, "step": 230100 }, { "epoch": 0.42948548981919715, "grad_norm": 0.9617169499397278, "learning_rate": 0.00019832497344803393, "loss": 4.5471, "step": 230150 }, { "epoch": 0.42957879537857563, "grad_norm": 1.044552206993103, "learning_rate": 0.0001983242076760645, "loss": 4.4572, "step": 230200 }, { "epoch": 0.42967210093795416, "grad_norm": 0.9525365829467773, "learning_rate": 0.00019832344173057012, "loss": 4.6309, "step": 230250 }, { "epoch": 0.42976540649733264, "grad_norm": 0.9384802579879761, "learning_rate": 0.0001983226756115522, "loss": 4.6633, "step": 230300 }, { "epoch": 0.4298587120567111, "grad_norm": 0.7917965054512024, "learning_rate": 0.00019832190931901202, "loss": 4.6365, "step": 230350 }, { "epoch": 0.4299520176160896, "grad_norm": 0.9427030086517334, "learning_rate": 0.000198321142852951, "loss": 4.6762, "step": 230400 }, { "epoch": 0.4300453231754681, "grad_norm": 1.0474097728729248, "learning_rate": 0.00019832037621337047, "loss": 4.574, "step": 230450 }, { "epoch": 0.4301386287348466, "grad_norm": 1.0150837898254395, "learning_rate": 0.00019831960940027178, "loss": 4.6684, "step": 230500 }, { "epoch": 0.43023193429422507, "grad_norm": 1.0305523872375488, "learning_rate": 0.00019831884241365623, "loss": 4.4752, "step": 230550 }, { "epoch": 0.4303252398536036, "grad_norm": 1.2383668422698975, "learning_rate": 0.00019831807525352526, "loss": 4.4957, "step": 230600 }, { "epoch": 0.4304185454129821, "grad_norm": 1.0035946369171143, "learning_rate": 0.00019831730791988018, "loss": 4.5095, "step": 230650 }, { "epoch": 0.43051185097236055, "grad_norm": 0.8765376210212708, "learning_rate": 0.00019831654041272236, "loss": 4.3764, "step": 230700 }, { "epoch": 0.43060515653173903, "grad_norm": 1.094586730003357, "learning_rate": 0.00019831577273205314, "loss": 4.5836, "step": 230750 }, { "epoch": 0.43069846209111756, "grad_norm": 1.0345427989959717, "learning_rate": 0.00019831500487787386, "loss": 4.5329, "step": 230800 }, { "epoch": 0.43079176765049604, "grad_norm": 0.7867444157600403, "learning_rate": 0.00019831423685018592, "loss": 4.45, "step": 230850 }, { "epoch": 0.4308850732098745, "grad_norm": 1.01674222946167, "learning_rate": 0.00019831346864899066, "loss": 4.3076, "step": 230900 }, { "epoch": 0.43097837876925305, "grad_norm": 0.9102169871330261, "learning_rate": 0.0001983127002742894, "loss": 4.7141, "step": 230950 }, { "epoch": 0.4310716843286315, "grad_norm": 0.9253519773483276, "learning_rate": 0.00019831193172608353, "loss": 4.6601, "step": 231000 }, { "epoch": 0.4310716843286315, "eval_loss": 4.782818794250488, "eval_runtime": 230.7143, "eval_samples_per_second": 11.304, "eval_steps_per_second": 11.304, "eval_tts_loss": 7.499476979419509, "step": 231000 }, { "epoch": 0.43116498988801, "grad_norm": 0.9714641571044922, "learning_rate": 0.0001983111630043744, "loss": 4.826, "step": 231050 }, { "epoch": 0.4312582954473885, "grad_norm": 1.0827927589416504, "learning_rate": 0.0001983103941091634, "loss": 4.5804, "step": 231100 }, { "epoch": 0.431351601006767, "grad_norm": 0.9766815304756165, "learning_rate": 0.0001983096250404518, "loss": 4.4809, "step": 231150 }, { "epoch": 0.4314449065661455, "grad_norm": 1.1361207962036133, "learning_rate": 0.00019830885579824104, "loss": 4.7216, "step": 231200 }, { "epoch": 0.43153821212552396, "grad_norm": 0.9344854950904846, "learning_rate": 0.00019830808638253243, "loss": 4.4767, "step": 231250 }, { "epoch": 0.4316315176849025, "grad_norm": 0.6080878376960754, "learning_rate": 0.00019830731679332738, "loss": 4.582, "step": 231300 }, { "epoch": 0.43172482324428096, "grad_norm": 1.0502957105636597, "learning_rate": 0.00019830654703062716, "loss": 4.5399, "step": 231350 }, { "epoch": 0.43181812880365944, "grad_norm": 1.0605568885803223, "learning_rate": 0.0001983057770944332, "loss": 4.5047, "step": 231400 }, { "epoch": 0.4319114343630379, "grad_norm": 1.162824273109436, "learning_rate": 0.00019830500698474687, "loss": 4.6726, "step": 231450 }, { "epoch": 0.43200473992241645, "grad_norm": 0.957460343837738, "learning_rate": 0.0001983042367015695, "loss": 4.6275, "step": 231500 }, { "epoch": 0.4320980454817949, "grad_norm": 0.8100277185440063, "learning_rate": 0.0001983034662449024, "loss": 4.5962, "step": 231550 }, { "epoch": 0.4321913510411734, "grad_norm": 1.1255825757980347, "learning_rate": 0.00019830269561474704, "loss": 4.6635, "step": 231600 }, { "epoch": 0.4322846566005519, "grad_norm": 1.0411120653152466, "learning_rate": 0.0001983019248111047, "loss": 4.4638, "step": 231650 }, { "epoch": 0.4323779621599304, "grad_norm": 0.6740716099739075, "learning_rate": 0.00019830115383397674, "loss": 4.5441, "step": 231700 }, { "epoch": 0.4324712677193089, "grad_norm": 1.0795879364013672, "learning_rate": 0.00019830038268336454, "loss": 4.8265, "step": 231750 }, { "epoch": 0.43256457327868736, "grad_norm": 1.218779444694519, "learning_rate": 0.00019829961135926948, "loss": 4.7943, "step": 231800 }, { "epoch": 0.4326578788380659, "grad_norm": 0.9649580717086792, "learning_rate": 0.00019829883986169285, "loss": 4.5642, "step": 231850 }, { "epoch": 0.43275118439744437, "grad_norm": 0.9463967084884644, "learning_rate": 0.00019829806819063613, "loss": 4.5809, "step": 231900 }, { "epoch": 0.43284448995682284, "grad_norm": 1.0788993835449219, "learning_rate": 0.0001982972963461006, "loss": 4.5609, "step": 231950 }, { "epoch": 0.4329377955162013, "grad_norm": 1.184398889541626, "learning_rate": 0.0001982965243280876, "loss": 4.6776, "step": 232000 }, { "epoch": 0.43303110107557985, "grad_norm": 0.9360151886940002, "learning_rate": 0.00019829575213659858, "loss": 4.5851, "step": 232050 }, { "epoch": 0.4331244066349583, "grad_norm": 1.1861097812652588, "learning_rate": 0.00019829497977163482, "loss": 4.5354, "step": 232100 }, { "epoch": 0.4332177121943368, "grad_norm": 0.9626287221908569, "learning_rate": 0.00019829420723319773, "loss": 4.8662, "step": 232150 }, { "epoch": 0.43331101775371533, "grad_norm": 0.9649963974952698, "learning_rate": 0.00019829343452128865, "loss": 4.5901, "step": 232200 }, { "epoch": 0.4334043233130938, "grad_norm": 1.0452684164047241, "learning_rate": 0.000198292661635909, "loss": 4.7717, "step": 232250 }, { "epoch": 0.4334976288724723, "grad_norm": 0.963649332523346, "learning_rate": 0.00019829188857706004, "loss": 4.6523, "step": 232300 }, { "epoch": 0.43359093443185076, "grad_norm": 1.0967086553573608, "learning_rate": 0.00019829111534474324, "loss": 4.7513, "step": 232350 }, { "epoch": 0.4336842399912293, "grad_norm": 1.1611350774765015, "learning_rate": 0.00019829034193895987, "loss": 4.6203, "step": 232400 }, { "epoch": 0.43377754555060777, "grad_norm": 1.3620021343231201, "learning_rate": 0.00019828956835971136, "loss": 4.5346, "step": 232450 }, { "epoch": 0.43387085110998624, "grad_norm": 1.0658037662506104, "learning_rate": 0.00019828879460699904, "loss": 4.6917, "step": 232500 }, { "epoch": 0.4339641566693648, "grad_norm": 1.0672107934951782, "learning_rate": 0.00019828802068082432, "loss": 4.7152, "step": 232550 }, { "epoch": 0.43405746222874325, "grad_norm": 0.8459929823875427, "learning_rate": 0.0001982872465811885, "loss": 4.4931, "step": 232600 }, { "epoch": 0.4341507677881217, "grad_norm": 0.9519999027252197, "learning_rate": 0.00019828647230809302, "loss": 4.6345, "step": 232650 }, { "epoch": 0.4342440733475002, "grad_norm": 0.7836452126502991, "learning_rate": 0.00019828569786153918, "loss": 4.6803, "step": 232700 }, { "epoch": 0.43433737890687873, "grad_norm": 1.1806153059005737, "learning_rate": 0.00019828492324152844, "loss": 4.5936, "step": 232750 }, { "epoch": 0.4344306844662572, "grad_norm": 1.0644776821136475, "learning_rate": 0.00019828414844806204, "loss": 4.5906, "step": 232800 }, { "epoch": 0.4345239900256357, "grad_norm": 1.392525315284729, "learning_rate": 0.00019828337348114144, "loss": 4.6483, "step": 232850 }, { "epoch": 0.4346172955850142, "grad_norm": 1.0627968311309814, "learning_rate": 0.00019828259834076797, "loss": 4.6276, "step": 232900 }, { "epoch": 0.4347106011443927, "grad_norm": 0.71669602394104, "learning_rate": 0.000198281823026943, "loss": 4.5737, "step": 232950 }, { "epoch": 0.43480390670377117, "grad_norm": 1.0099847316741943, "learning_rate": 0.00019828104753966792, "loss": 4.491, "step": 233000 }, { "epoch": 0.43489721226314965, "grad_norm": 1.1775974035263062, "learning_rate": 0.0001982802718789441, "loss": 4.4005, "step": 233050 }, { "epoch": 0.4349905178225282, "grad_norm": 1.0557302236557007, "learning_rate": 0.00019827949604477285, "loss": 4.5978, "step": 233100 }, { "epoch": 0.43508382338190665, "grad_norm": 1.1460533142089844, "learning_rate": 0.0001982787200371556, "loss": 4.6675, "step": 233150 }, { "epoch": 0.43517712894128513, "grad_norm": 1.1376761198043823, "learning_rate": 0.00019827794385609372, "loss": 4.5659, "step": 233200 }, { "epoch": 0.43527043450066366, "grad_norm": 0.9409826993942261, "learning_rate": 0.00019827716750158854, "loss": 4.6083, "step": 233250 }, { "epoch": 0.43536374006004214, "grad_norm": 0.7400208711624146, "learning_rate": 0.00019827639097364145, "loss": 4.7026, "step": 233300 }, { "epoch": 0.4354570456194206, "grad_norm": 1.1714075803756714, "learning_rate": 0.00019827561427225384, "loss": 4.6554, "step": 233350 }, { "epoch": 0.4355503511787991, "grad_norm": 0.8486500382423401, "learning_rate": 0.00019827483739742705, "loss": 4.6074, "step": 233400 }, { "epoch": 0.4356436567381776, "grad_norm": 1.0664108991622925, "learning_rate": 0.00019827406034916248, "loss": 4.5721, "step": 233450 }, { "epoch": 0.4357369622975561, "grad_norm": 0.9367774128913879, "learning_rate": 0.00019827328312746145, "loss": 4.5607, "step": 233500 }, { "epoch": 0.4358302678569346, "grad_norm": 1.0547155141830444, "learning_rate": 0.00019827250573232542, "loss": 4.5985, "step": 233550 }, { "epoch": 0.4359235734163131, "grad_norm": 0.9410718679428101, "learning_rate": 0.00019827172816375566, "loss": 4.5643, "step": 233600 }, { "epoch": 0.4360168789756916, "grad_norm": 0.9093536734580994, "learning_rate": 0.0001982709504217536, "loss": 4.6474, "step": 233650 }, { "epoch": 0.43611018453507006, "grad_norm": 0.9456886053085327, "learning_rate": 0.00019827017250632063, "loss": 4.7075, "step": 233700 }, { "epoch": 0.43620349009444853, "grad_norm": 1.0371719598770142, "learning_rate": 0.00019826939441745805, "loss": 4.5081, "step": 233750 }, { "epoch": 0.43629679565382706, "grad_norm": 0.7762008905410767, "learning_rate": 0.00019826861615516734, "loss": 4.5744, "step": 233800 }, { "epoch": 0.43639010121320554, "grad_norm": 0.8495267629623413, "learning_rate": 0.0001982678377194498, "loss": 4.6972, "step": 233850 }, { "epoch": 0.436483406772584, "grad_norm": 0.9476446509361267, "learning_rate": 0.00019826705911030678, "loss": 4.6825, "step": 233900 }, { "epoch": 0.43657671233196255, "grad_norm": 0.92109614610672, "learning_rate": 0.0001982662803277397, "loss": 4.5082, "step": 233950 }, { "epoch": 0.436670017891341, "grad_norm": 1.5745224952697754, "learning_rate": 0.00019826550137174996, "loss": 4.702, "step": 234000 }, { "epoch": 0.436670017891341, "eval_loss": 4.786718368530273, "eval_runtime": 234.4074, "eval_samples_per_second": 11.126, "eval_steps_per_second": 11.126, "eval_tts_loss": 7.521374580757464, "step": 234000 }, { "epoch": 0.4367633234507195, "grad_norm": 0.6546002626419067, "learning_rate": 0.00019826472224233888, "loss": 4.6363, "step": 234050 }, { "epoch": 0.436856629010098, "grad_norm": 1.0382810831069946, "learning_rate": 0.00019826394293950784, "loss": 4.6492, "step": 234100 }, { "epoch": 0.4369499345694765, "grad_norm": 1.125950574874878, "learning_rate": 0.00019826316346325826, "loss": 4.6651, "step": 234150 }, { "epoch": 0.437043240128855, "grad_norm": 0.9821774959564209, "learning_rate": 0.00019826238381359147, "loss": 4.4427, "step": 234200 }, { "epoch": 0.43713654568823346, "grad_norm": 0.9345731735229492, "learning_rate": 0.00019826160399050887, "loss": 4.6285, "step": 234250 }, { "epoch": 0.43722985124761193, "grad_norm": 1.1006258726119995, "learning_rate": 0.0001982608239940118, "loss": 4.6635, "step": 234300 }, { "epoch": 0.43732315680699047, "grad_norm": 0.8882812261581421, "learning_rate": 0.0001982600438241017, "loss": 4.5461, "step": 234350 }, { "epoch": 0.43741646236636894, "grad_norm": 0.8373991847038269, "learning_rate": 0.0001982592634807799, "loss": 4.6949, "step": 234400 }, { "epoch": 0.4375097679257474, "grad_norm": 0.9993740916252136, "learning_rate": 0.00019825848296404778, "loss": 4.6201, "step": 234450 }, { "epoch": 0.43760307348512595, "grad_norm": 1.093626856803894, "learning_rate": 0.00019825770227390675, "loss": 4.6239, "step": 234500 }, { "epoch": 0.4376963790445044, "grad_norm": 1.159334421157837, "learning_rate": 0.00019825692141035815, "loss": 4.4909, "step": 234550 }, { "epoch": 0.4377896846038829, "grad_norm": 0.8569579124450684, "learning_rate": 0.0001982561403734034, "loss": 4.54, "step": 234600 }, { "epoch": 0.4378829901632614, "grad_norm": 1.0872637033462524, "learning_rate": 0.00019825535916304378, "loss": 4.5892, "step": 234650 }, { "epoch": 0.4379762957226399, "grad_norm": 1.0390820503234863, "learning_rate": 0.0001982545777792808, "loss": 4.4236, "step": 234700 }, { "epoch": 0.4380696012820184, "grad_norm": 0.8615942597389221, "learning_rate": 0.00019825379622211576, "loss": 4.7319, "step": 234750 }, { "epoch": 0.43816290684139686, "grad_norm": 0.9036928415298462, "learning_rate": 0.00019825301449155006, "loss": 4.6797, "step": 234800 }, { "epoch": 0.4382562124007754, "grad_norm": 0.8488964438438416, "learning_rate": 0.00019825223258758507, "loss": 4.6368, "step": 234850 }, { "epoch": 0.43834951796015387, "grad_norm": 0.7917913198471069, "learning_rate": 0.0001982514505102222, "loss": 4.5903, "step": 234900 }, { "epoch": 0.43844282351953234, "grad_norm": 1.0873136520385742, "learning_rate": 0.0001982506682594628, "loss": 4.5301, "step": 234950 }, { "epoch": 0.4385361290789108, "grad_norm": 1.2058392763137817, "learning_rate": 0.00019824988583530824, "loss": 4.3379, "step": 235000 }, { "epoch": 0.43862943463828935, "grad_norm": 1.043069839477539, "learning_rate": 0.00019824910323775993, "loss": 4.9557, "step": 235050 }, { "epoch": 0.4387227401976678, "grad_norm": 0.8327590823173523, "learning_rate": 0.00019824832046681925, "loss": 4.5355, "step": 235100 }, { "epoch": 0.4388160457570463, "grad_norm": 1.312573790550232, "learning_rate": 0.00019824753752248754, "loss": 4.7489, "step": 235150 }, { "epoch": 0.43890935131642483, "grad_norm": 1.0750068426132202, "learning_rate": 0.00019824675440476624, "loss": 4.61, "step": 235200 }, { "epoch": 0.4390026568758033, "grad_norm": 1.0304391384124756, "learning_rate": 0.0001982459711136567, "loss": 4.5803, "step": 235250 }, { "epoch": 0.4390959624351818, "grad_norm": 1.0730409622192383, "learning_rate": 0.0001982451876491603, "loss": 4.661, "step": 235300 }, { "epoch": 0.43918926799456026, "grad_norm": 1.1401572227478027, "learning_rate": 0.00019824440401127846, "loss": 4.5335, "step": 235350 }, { "epoch": 0.4392825735539388, "grad_norm": 1.187408685684204, "learning_rate": 0.00019824362020001251, "loss": 4.6488, "step": 235400 }, { "epoch": 0.43937587911331727, "grad_norm": 0.830162763595581, "learning_rate": 0.00019824283621536386, "loss": 4.716, "step": 235450 }, { "epoch": 0.43946918467269575, "grad_norm": 1.0726609230041504, "learning_rate": 0.0001982420520573339, "loss": 4.5043, "step": 235500 }, { "epoch": 0.4395624902320743, "grad_norm": 0.8797732591629028, "learning_rate": 0.000198241267725924, "loss": 4.5403, "step": 235550 }, { "epoch": 0.43965579579145275, "grad_norm": 1.1645294427871704, "learning_rate": 0.00019824048322113552, "loss": 4.5505, "step": 235600 }, { "epoch": 0.43974910135083123, "grad_norm": 1.0088536739349365, "learning_rate": 0.0001982396985429699, "loss": 4.5699, "step": 235650 }, { "epoch": 0.4398424069102097, "grad_norm": 1.086678385734558, "learning_rate": 0.00019823891369142847, "loss": 4.5694, "step": 235700 }, { "epoch": 0.43993571246958824, "grad_norm": 1.2575753927230835, "learning_rate": 0.00019823812866651271, "loss": 4.6272, "step": 235750 }, { "epoch": 0.4400290180289667, "grad_norm": 0.8871008157730103, "learning_rate": 0.00019823734346822387, "loss": 4.5924, "step": 235800 }, { "epoch": 0.4401223235883452, "grad_norm": 0.8986257314682007, "learning_rate": 0.00019823655809656345, "loss": 4.7974, "step": 235850 }, { "epoch": 0.4402156291477237, "grad_norm": 0.5982351899147034, "learning_rate": 0.00019823577255153276, "loss": 4.4457, "step": 235900 }, { "epoch": 0.4403089347071022, "grad_norm": 1.1470563411712646, "learning_rate": 0.00019823498683313324, "loss": 4.5922, "step": 235950 }, { "epoch": 0.44040224026648067, "grad_norm": 0.9602048397064209, "learning_rate": 0.00019823420094136623, "loss": 4.5407, "step": 236000 }, { "epoch": 0.44049554582585915, "grad_norm": 1.228560209274292, "learning_rate": 0.00019823341487623314, "loss": 4.6289, "step": 236050 }, { "epoch": 0.4405888513852377, "grad_norm": 1.1307828426361084, "learning_rate": 0.00019823262863773538, "loss": 4.5017, "step": 236100 }, { "epoch": 0.44068215694461615, "grad_norm": 0.942937970161438, "learning_rate": 0.0001982318422258743, "loss": 4.6442, "step": 236150 }, { "epoch": 0.44077546250399463, "grad_norm": 0.9702286720275879, "learning_rate": 0.00019823105564065129, "loss": 4.5807, "step": 236200 }, { "epoch": 0.44086876806337316, "grad_norm": 0.9227506518363953, "learning_rate": 0.00019823026888206777, "loss": 4.5915, "step": 236250 }, { "epoch": 0.44096207362275164, "grad_norm": 0.7443217635154724, "learning_rate": 0.00019822948195012512, "loss": 4.5009, "step": 236300 }, { "epoch": 0.4410553791821301, "grad_norm": 0.9936513304710388, "learning_rate": 0.0001982286948448247, "loss": 4.7613, "step": 236350 }, { "epoch": 0.4411486847415086, "grad_norm": 0.7316447496414185, "learning_rate": 0.00019822790756616793, "loss": 4.4194, "step": 236400 }, { "epoch": 0.4412419903008871, "grad_norm": 0.6408354640007019, "learning_rate": 0.00019822712011415615, "loss": 4.5336, "step": 236450 }, { "epoch": 0.4413352958602656, "grad_norm": 1.0418235063552856, "learning_rate": 0.00019822633248879084, "loss": 4.4547, "step": 236500 }, { "epoch": 0.4414286014196441, "grad_norm": 1.010443925857544, "learning_rate": 0.0001982255446900733, "loss": 4.5134, "step": 236550 }, { "epoch": 0.4415219069790226, "grad_norm": 1.3255016803741455, "learning_rate": 0.00019822475671800498, "loss": 4.7933, "step": 236600 }, { "epoch": 0.4416152125384011, "grad_norm": 0.9770745038986206, "learning_rate": 0.00019822396857258724, "loss": 4.7556, "step": 236650 }, { "epoch": 0.44170851809777956, "grad_norm": 1.0298558473587036, "learning_rate": 0.00019822318025382147, "loss": 4.6711, "step": 236700 }, { "epoch": 0.44180182365715803, "grad_norm": 0.7105265259742737, "learning_rate": 0.00019822239176170908, "loss": 4.5719, "step": 236750 }, { "epoch": 0.44189512921653656, "grad_norm": 0.9798986315727234, "learning_rate": 0.00019822160309625143, "loss": 4.5355, "step": 236800 }, { "epoch": 0.44198843477591504, "grad_norm": 1.1659282445907593, "learning_rate": 0.00019822081425744997, "loss": 4.6384, "step": 236850 }, { "epoch": 0.4420817403352935, "grad_norm": 0.9095149636268616, "learning_rate": 0.00019822002524530602, "loss": 4.5297, "step": 236900 }, { "epoch": 0.442175045894672, "grad_norm": 0.8372174501419067, "learning_rate": 0.00019821923605982104, "loss": 4.6121, "step": 236950 }, { "epoch": 0.4422683514540505, "grad_norm": 0.9444534182548523, "learning_rate": 0.00019821844670099637, "loss": 4.5425, "step": 237000 }, { "epoch": 0.4422683514540505, "eval_loss": 4.778234958648682, "eval_runtime": 232.886, "eval_samples_per_second": 11.199, "eval_steps_per_second": 11.199, "eval_tts_loss": 7.544113754277672, "step": 237000 }, { "epoch": 0.442361657013429, "grad_norm": 1.01829195022583, "learning_rate": 0.0001982176571688334, "loss": 4.5883, "step": 237050 }, { "epoch": 0.4424549625728075, "grad_norm": 0.8927636742591858, "learning_rate": 0.0001982168674633336, "loss": 4.6002, "step": 237100 }, { "epoch": 0.442548268132186, "grad_norm": 0.9533956050872803, "learning_rate": 0.00019821607758449826, "loss": 4.4706, "step": 237150 }, { "epoch": 0.4426415736915645, "grad_norm": 1.0447924137115479, "learning_rate": 0.00019821528753232888, "loss": 4.7974, "step": 237200 }, { "epoch": 0.44273487925094296, "grad_norm": 0.9149044156074524, "learning_rate": 0.00019821449730682676, "loss": 4.4807, "step": 237250 }, { "epoch": 0.44282818481032143, "grad_norm": 0.6954190135002136, "learning_rate": 0.00019821370690799333, "loss": 4.3259, "step": 237300 }, { "epoch": 0.44292149036969997, "grad_norm": 0.8902153372764587, "learning_rate": 0.00019821291633583001, "loss": 4.6436, "step": 237350 }, { "epoch": 0.44301479592907844, "grad_norm": 0.9041743874549866, "learning_rate": 0.00019821212559033814, "loss": 4.5148, "step": 237400 }, { "epoch": 0.4431081014884569, "grad_norm": 1.206248164176941, "learning_rate": 0.0001982113346715192, "loss": 4.5886, "step": 237450 }, { "epoch": 0.44320140704783545, "grad_norm": 0.9443100690841675, "learning_rate": 0.0001982105435793745, "loss": 4.5412, "step": 237500 }, { "epoch": 0.4432947126072139, "grad_norm": 1.071516990661621, "learning_rate": 0.00019820975231390548, "loss": 4.5386, "step": 237550 }, { "epoch": 0.4433880181665924, "grad_norm": 1.0057834386825562, "learning_rate": 0.0001982089608751135, "loss": 4.6202, "step": 237600 }, { "epoch": 0.4434813237259709, "grad_norm": 1.0348365306854248, "learning_rate": 0.00019820816926300006, "loss": 4.4221, "step": 237650 }, { "epoch": 0.4435746292853494, "grad_norm": 0.8951256275177002, "learning_rate": 0.00019820737747756642, "loss": 4.5601, "step": 237700 }, { "epoch": 0.4436679348447279, "grad_norm": 0.783212423324585, "learning_rate": 0.00019820658551881405, "loss": 4.5649, "step": 237750 }, { "epoch": 0.44376124040410636, "grad_norm": 1.0642147064208984, "learning_rate": 0.0001982057933867443, "loss": 4.8408, "step": 237800 }, { "epoch": 0.4438545459634849, "grad_norm": 0.9804556965827942, "learning_rate": 0.00019820500108135865, "loss": 4.7134, "step": 237850 }, { "epoch": 0.44394785152286337, "grad_norm": 0.9035722017288208, "learning_rate": 0.0001982042086026585, "loss": 4.6311, "step": 237900 }, { "epoch": 0.44404115708224184, "grad_norm": 1.0001559257507324, "learning_rate": 0.00019820341595064514, "loss": 4.804, "step": 237950 }, { "epoch": 0.4441344626416203, "grad_norm": 1.2685824632644653, "learning_rate": 0.00019820262312532, "loss": 4.6524, "step": 238000 }, { "epoch": 0.44422776820099885, "grad_norm": 1.238340973854065, "learning_rate": 0.00019820183012668458, "loss": 4.6042, "step": 238050 }, { "epoch": 0.4443210737603773, "grad_norm": 1.1636897325515747, "learning_rate": 0.00019820103695474014, "loss": 4.562, "step": 238100 }, { "epoch": 0.4444143793197558, "grad_norm": 1.0242795944213867, "learning_rate": 0.0001982002436094882, "loss": 4.3683, "step": 238150 }, { "epoch": 0.44450768487913433, "grad_norm": 0.8869125843048096, "learning_rate": 0.0001981994500909301, "loss": 4.6679, "step": 238200 }, { "epoch": 0.4446009904385128, "grad_norm": 0.8452954292297363, "learning_rate": 0.00019819865639906725, "loss": 4.6062, "step": 238250 }, { "epoch": 0.4446942959978913, "grad_norm": 1.0370817184448242, "learning_rate": 0.00019819786253390102, "loss": 4.5246, "step": 238300 }, { "epoch": 0.44478760155726976, "grad_norm": 1.0226502418518066, "learning_rate": 0.0001981970684954329, "loss": 4.7068, "step": 238350 }, { "epoch": 0.4448809071166483, "grad_norm": 0.9036027789115906, "learning_rate": 0.00019819627428366418, "loss": 4.5923, "step": 238400 }, { "epoch": 0.44497421267602677, "grad_norm": 1.0699036121368408, "learning_rate": 0.00019819547989859633, "loss": 4.8253, "step": 238450 }, { "epoch": 0.44506751823540525, "grad_norm": 0.9457048177719116, "learning_rate": 0.00019819468534023073, "loss": 4.3733, "step": 238500 }, { "epoch": 0.4451608237947838, "grad_norm": 1.337355375289917, "learning_rate": 0.0001981938906085688, "loss": 4.4989, "step": 238550 }, { "epoch": 0.44525412935416225, "grad_norm": 1.0064990520477295, "learning_rate": 0.00019819309570361194, "loss": 4.7772, "step": 238600 }, { "epoch": 0.44534743491354073, "grad_norm": 0.9159983396530151, "learning_rate": 0.00019819230062536152, "loss": 4.6947, "step": 238650 }, { "epoch": 0.4454407404729192, "grad_norm": 1.0956370830535889, "learning_rate": 0.00019819150537381898, "loss": 4.6732, "step": 238700 }, { "epoch": 0.44553404603229774, "grad_norm": 0.8136811852455139, "learning_rate": 0.0001981907099489857, "loss": 4.5511, "step": 238750 }, { "epoch": 0.4456273515916762, "grad_norm": 0.7191414833068848, "learning_rate": 0.00019818991435086313, "loss": 4.8481, "step": 238800 }, { "epoch": 0.4457206571510547, "grad_norm": 1.0575743913650513, "learning_rate": 0.00019818911857945258, "loss": 4.4962, "step": 238850 }, { "epoch": 0.4458139627104332, "grad_norm": 0.9950775504112244, "learning_rate": 0.00019818832263475552, "loss": 4.44, "step": 238900 }, { "epoch": 0.4459072682698117, "grad_norm": 1.380645990371704, "learning_rate": 0.00019818752651677338, "loss": 4.5819, "step": 238950 }, { "epoch": 0.44600057382919017, "grad_norm": 1.2283092737197876, "learning_rate": 0.00019818673022550755, "loss": 4.6796, "step": 239000 }, { "epoch": 0.44609387938856865, "grad_norm": 1.224413275718689, "learning_rate": 0.00019818593376095937, "loss": 4.6762, "step": 239050 }, { "epoch": 0.4461871849479472, "grad_norm": 1.0246546268463135, "learning_rate": 0.00019818513712313032, "loss": 4.6132, "step": 239100 }, { "epoch": 0.44628049050732566, "grad_norm": 1.1078890562057495, "learning_rate": 0.00019818434031202177, "loss": 4.5609, "step": 239150 }, { "epoch": 0.44637379606670413, "grad_norm": 1.020003080368042, "learning_rate": 0.00019818354332763513, "loss": 4.5842, "step": 239200 }, { "epoch": 0.44646710162608266, "grad_norm": 0.9246664643287659, "learning_rate": 0.00019818274616997182, "loss": 4.7182, "step": 239250 }, { "epoch": 0.44656040718546114, "grad_norm": 0.9110824465751648, "learning_rate": 0.00019818194883903324, "loss": 4.4936, "step": 239300 }, { "epoch": 0.4466537127448396, "grad_norm": 1.1827949285507202, "learning_rate": 0.0001981811513348208, "loss": 4.3821, "step": 239350 }, { "epoch": 0.4467470183042181, "grad_norm": 0.9574491381645203, "learning_rate": 0.0001981803536573359, "loss": 4.8013, "step": 239400 }, { "epoch": 0.4468403238635966, "grad_norm": 1.1559712886810303, "learning_rate": 0.00019817955580657995, "loss": 4.4273, "step": 239450 }, { "epoch": 0.4469336294229751, "grad_norm": 0.8281153440475464, "learning_rate": 0.00019817875778255434, "loss": 4.6266, "step": 239500 }, { "epoch": 0.4470269349823536, "grad_norm": 1.210089087486267, "learning_rate": 0.00019817795958526052, "loss": 4.5594, "step": 239550 }, { "epoch": 0.44712024054173205, "grad_norm": 1.2433289289474487, "learning_rate": 0.00019817716121469986, "loss": 4.5299, "step": 239600 }, { "epoch": 0.4472135461011106, "grad_norm": 1.1075997352600098, "learning_rate": 0.00019817636267087382, "loss": 4.3679, "step": 239650 }, { "epoch": 0.44730685166048906, "grad_norm": 1.0246319770812988, "learning_rate": 0.00019817556395378374, "loss": 4.8387, "step": 239700 }, { "epoch": 0.44740015721986753, "grad_norm": 1.2445541620254517, "learning_rate": 0.00019817476506343108, "loss": 4.6812, "step": 239750 }, { "epoch": 0.44749346277924607, "grad_norm": 0.791220486164093, "learning_rate": 0.00019817396599981725, "loss": 4.6479, "step": 239800 }, { "epoch": 0.44758676833862454, "grad_norm": 1.1053506135940552, "learning_rate": 0.0001981731667629436, "loss": 4.5058, "step": 239850 }, { "epoch": 0.447680073898003, "grad_norm": 0.7307910919189453, "learning_rate": 0.00019817236735281162, "loss": 4.4622, "step": 239900 }, { "epoch": 0.4477733794573815, "grad_norm": 0.864814281463623, "learning_rate": 0.00019817156776942265, "loss": 4.5375, "step": 239950 }, { "epoch": 0.44786668501676, "grad_norm": 1.1574825048446655, "learning_rate": 0.0001981707680127782, "loss": 4.661, "step": 240000 }, { "epoch": 0.44786668501676, "eval_loss": 4.771146774291992, "eval_runtime": 230.5723, "eval_samples_per_second": 11.311, "eval_steps_per_second": 11.311, "eval_tts_loss": 7.4719085190530885, "step": 240000 }, { "epoch": 0.4479599905761385, "grad_norm": 1.3045135736465454, "learning_rate": 0.00019816996808287956, "loss": 4.6628, "step": 240050 }, { "epoch": 0.448053296135517, "grad_norm": 1.4040054082870483, "learning_rate": 0.0001981691679797282, "loss": 4.763, "step": 240100 }, { "epoch": 0.4481466016948955, "grad_norm": 0.8123361468315125, "learning_rate": 0.00019816836770332556, "loss": 4.6492, "step": 240150 }, { "epoch": 0.448239907254274, "grad_norm": 0.8047076463699341, "learning_rate": 0.000198167567253673, "loss": 4.7971, "step": 240200 }, { "epoch": 0.44833321281365246, "grad_norm": 0.7521131634712219, "learning_rate": 0.000198166766630772, "loss": 4.7822, "step": 240250 }, { "epoch": 0.44842651837303094, "grad_norm": 0.808711051940918, "learning_rate": 0.0001981659658346239, "loss": 4.5455, "step": 240300 }, { "epoch": 0.44851982393240947, "grad_norm": 0.6877180933952332, "learning_rate": 0.00019816516486523015, "loss": 4.5178, "step": 240350 }, { "epoch": 0.44861312949178794, "grad_norm": 1.0691016912460327, "learning_rate": 0.00019816436372259216, "loss": 4.6918, "step": 240400 }, { "epoch": 0.4487064350511664, "grad_norm": 1.0807589292526245, "learning_rate": 0.0001981635624067113, "loss": 4.7728, "step": 240450 }, { "epoch": 0.44879974061054495, "grad_norm": 0.9326990842819214, "learning_rate": 0.00019816276091758907, "loss": 4.5443, "step": 240500 }, { "epoch": 0.4488930461699234, "grad_norm": 0.9165334105491638, "learning_rate": 0.00019816195925522682, "loss": 4.6545, "step": 240550 }, { "epoch": 0.4489863517293019, "grad_norm": 0.8920701146125793, "learning_rate": 0.000198161157419626, "loss": 4.7592, "step": 240600 }, { "epoch": 0.4490796572886804, "grad_norm": 1.1457407474517822, "learning_rate": 0.000198160355410788, "loss": 4.5365, "step": 240650 }, { "epoch": 0.4491729628480589, "grad_norm": 0.9040488004684448, "learning_rate": 0.00019815955322871425, "loss": 4.6286, "step": 240700 }, { "epoch": 0.4492662684074374, "grad_norm": 0.8481893539428711, "learning_rate": 0.00019815875087340613, "loss": 4.5136, "step": 240750 }, { "epoch": 0.44935957396681586, "grad_norm": 1.0525832176208496, "learning_rate": 0.00019815794834486511, "loss": 4.791, "step": 240800 }, { "epoch": 0.4494528795261944, "grad_norm": 1.0377291440963745, "learning_rate": 0.00019815714564309258, "loss": 4.6304, "step": 240850 }, { "epoch": 0.44954618508557287, "grad_norm": 0.7769307494163513, "learning_rate": 0.00019815634276808996, "loss": 4.4798, "step": 240900 }, { "epoch": 0.44963949064495135, "grad_norm": 0.9710248112678528, "learning_rate": 0.00019815553971985867, "loss": 4.4885, "step": 240950 }, { "epoch": 0.4497327962043298, "grad_norm": 1.0528430938720703, "learning_rate": 0.00019815473649840014, "loss": 4.6653, "step": 241000 }, { "epoch": 0.44982610176370835, "grad_norm": 0.9350594282150269, "learning_rate": 0.00019815393310371575, "loss": 4.6053, "step": 241050 }, { "epoch": 0.44991940732308683, "grad_norm": 0.9134218692779541, "learning_rate": 0.00019815312953580692, "loss": 4.6681, "step": 241100 }, { "epoch": 0.4500127128824653, "grad_norm": 1.070358395576477, "learning_rate": 0.00019815232579467514, "loss": 4.4734, "step": 241150 }, { "epoch": 0.45010601844184384, "grad_norm": 1.081626296043396, "learning_rate": 0.00019815152188032173, "loss": 4.5605, "step": 241200 }, { "epoch": 0.4501993240012223, "grad_norm": 1.0206654071807861, "learning_rate": 0.00019815071779274816, "loss": 4.7068, "step": 241250 }, { "epoch": 0.4502926295606008, "grad_norm": 0.9931300282478333, "learning_rate": 0.00019814991353195587, "loss": 4.5779, "step": 241300 }, { "epoch": 0.45038593511997926, "grad_norm": 1.139255166053772, "learning_rate": 0.00019814910909794623, "loss": 4.5736, "step": 241350 }, { "epoch": 0.4504792406793578, "grad_norm": 1.1006131172180176, "learning_rate": 0.00019814830449072068, "loss": 4.7273, "step": 241400 }, { "epoch": 0.45057254623873627, "grad_norm": 0.8194779753684998, "learning_rate": 0.00019814749971028067, "loss": 4.7137, "step": 241450 }, { "epoch": 0.45066585179811475, "grad_norm": 1.067018747329712, "learning_rate": 0.00019814669475662754, "loss": 4.5206, "step": 241500 }, { "epoch": 0.4507591573574933, "grad_norm": 1.050444483757019, "learning_rate": 0.0001981458896297628, "loss": 4.6427, "step": 241550 }, { "epoch": 0.45085246291687175, "grad_norm": 1.1245814561843872, "learning_rate": 0.00019814508432968784, "loss": 4.4676, "step": 241600 }, { "epoch": 0.45094576847625023, "grad_norm": 1.1817681789398193, "learning_rate": 0.00019814427885640407, "loss": 4.8481, "step": 241650 }, { "epoch": 0.4510390740356287, "grad_norm": 1.0630196332931519, "learning_rate": 0.00019814347320991292, "loss": 4.4368, "step": 241700 }, { "epoch": 0.45113237959500724, "grad_norm": 1.3640244007110596, "learning_rate": 0.00019814266739021577, "loss": 4.5863, "step": 241750 }, { "epoch": 0.4512256851543857, "grad_norm": 0.9691682457923889, "learning_rate": 0.00019814186139731409, "loss": 4.6831, "step": 241800 }, { "epoch": 0.4513189907137642, "grad_norm": 0.8320736289024353, "learning_rate": 0.00019814105523120935, "loss": 4.4026, "step": 241850 }, { "epoch": 0.4514122962731427, "grad_norm": 0.7617359757423401, "learning_rate": 0.00019814024889190285, "loss": 4.591, "step": 241900 }, { "epoch": 0.4515056018325212, "grad_norm": 0.9816135168075562, "learning_rate": 0.00019813944237939612, "loss": 4.7765, "step": 241950 }, { "epoch": 0.4515989073918997, "grad_norm": 0.8725814819335938, "learning_rate": 0.00019813863569369052, "loss": 4.5988, "step": 242000 }, { "epoch": 0.45169221295127815, "grad_norm": 0.9102336764335632, "learning_rate": 0.00019813782883478754, "loss": 4.6437, "step": 242050 }, { "epoch": 0.4517855185106567, "grad_norm": 1.0546931028366089, "learning_rate": 0.0001981370218026885, "loss": 4.5362, "step": 242100 }, { "epoch": 0.45187882407003516, "grad_norm": 1.28167724609375, "learning_rate": 0.00019813621459739494, "loss": 4.7866, "step": 242150 }, { "epoch": 0.45197212962941363, "grad_norm": 1.2853349447250366, "learning_rate": 0.00019813540721890817, "loss": 4.6789, "step": 242200 }, { "epoch": 0.4520654351887921, "grad_norm": 1.1249480247497559, "learning_rate": 0.0001981345996672297, "loss": 4.4926, "step": 242250 }, { "epoch": 0.45215874074817064, "grad_norm": 0.5457555651664734, "learning_rate": 0.00019813379194236094, "loss": 4.4383, "step": 242300 }, { "epoch": 0.4522520463075491, "grad_norm": 1.1314606666564941, "learning_rate": 0.00019813298404430332, "loss": 4.7619, "step": 242350 }, { "epoch": 0.4523453518669276, "grad_norm": 1.0931801795959473, "learning_rate": 0.0001981321759730582, "loss": 4.7695, "step": 242400 }, { "epoch": 0.4524386574263061, "grad_norm": 1.2083085775375366, "learning_rate": 0.00019813136772862708, "loss": 4.5025, "step": 242450 }, { "epoch": 0.4525319629856846, "grad_norm": 1.1209338903427124, "learning_rate": 0.00019813055931101138, "loss": 4.649, "step": 242500 }, { "epoch": 0.4526252685450631, "grad_norm": 1.0771358013153076, "learning_rate": 0.0001981297507202125, "loss": 4.5248, "step": 242550 }, { "epoch": 0.45271857410444155, "grad_norm": 1.0569473505020142, "learning_rate": 0.00019812894195623187, "loss": 4.5961, "step": 242600 }, { "epoch": 0.4528118796638201, "grad_norm": 0.7088388204574585, "learning_rate": 0.0001981281330190709, "loss": 4.6979, "step": 242650 }, { "epoch": 0.45290518522319856, "grad_norm": 0.8838179707527161, "learning_rate": 0.0001981273239087311, "loss": 4.7185, "step": 242700 }, { "epoch": 0.45299849078257703, "grad_norm": 1.145406723022461, "learning_rate": 0.0001981265146252138, "loss": 4.6553, "step": 242750 }, { "epoch": 0.45309179634195557, "grad_norm": 0.8475630879402161, "learning_rate": 0.00019812570516852048, "loss": 4.6364, "step": 242800 }, { "epoch": 0.45318510190133404, "grad_norm": 0.8554997444152832, "learning_rate": 0.00019812489553865255, "loss": 4.3752, "step": 242850 }, { "epoch": 0.4532784074607125, "grad_norm": 1.0469423532485962, "learning_rate": 0.00019812408573561144, "loss": 4.6012, "step": 242900 }, { "epoch": 0.453371713020091, "grad_norm": 0.9449438452720642, "learning_rate": 0.00019812327575939858, "loss": 4.4068, "step": 242950 }, { "epoch": 0.4534650185794695, "grad_norm": 0.8156939744949341, "learning_rate": 0.00019812246561001543, "loss": 4.4701, "step": 243000 }, { "epoch": 0.4534650185794695, "eval_loss": 4.782342910766602, "eval_runtime": 233.441, "eval_samples_per_second": 11.172, "eval_steps_per_second": 11.172, "eval_tts_loss": 7.457396795253263, "step": 243000 }, { "epoch": 0.453558324138848, "grad_norm": 1.1916180849075317, "learning_rate": 0.00019812165528746337, "loss": 4.5937, "step": 243050 }, { "epoch": 0.4536516296982265, "grad_norm": 1.1467726230621338, "learning_rate": 0.00019812084479174384, "loss": 4.737, "step": 243100 }, { "epoch": 0.453744935257605, "grad_norm": 1.2688440084457397, "learning_rate": 0.0001981200341228583, "loss": 4.7588, "step": 243150 }, { "epoch": 0.4538382408169835, "grad_norm": 0.977979302406311, "learning_rate": 0.0001981192232808082, "loss": 4.5473, "step": 243200 }, { "epoch": 0.45393154637636196, "grad_norm": 0.8517860174179077, "learning_rate": 0.00019811841226559487, "loss": 4.6273, "step": 243250 }, { "epoch": 0.45402485193574044, "grad_norm": 1.3185665607452393, "learning_rate": 0.00019811760107721983, "loss": 4.6099, "step": 243300 }, { "epoch": 0.45411815749511897, "grad_norm": 1.0361919403076172, "learning_rate": 0.00019811678971568453, "loss": 4.5083, "step": 243350 }, { "epoch": 0.45421146305449744, "grad_norm": 1.0736933946609497, "learning_rate": 0.00019811597818099032, "loss": 4.5282, "step": 243400 }, { "epoch": 0.4543047686138759, "grad_norm": 0.9656780362129211, "learning_rate": 0.00019811516647313867, "loss": 4.7517, "step": 243450 }, { "epoch": 0.45439807417325445, "grad_norm": 1.222460389137268, "learning_rate": 0.00019811435459213104, "loss": 4.5887, "step": 243500 }, { "epoch": 0.4544913797326329, "grad_norm": 1.2716476917266846, "learning_rate": 0.00019811354253796882, "loss": 4.6937, "step": 243550 }, { "epoch": 0.4545846852920114, "grad_norm": 0.9141228199005127, "learning_rate": 0.00019811273031065347, "loss": 4.5894, "step": 243600 }, { "epoch": 0.4546779908513899, "grad_norm": 1.0707182884216309, "learning_rate": 0.00019811191791018642, "loss": 4.6077, "step": 243650 }, { "epoch": 0.4547712964107684, "grad_norm": 1.1825801134109497, "learning_rate": 0.00019811110533656909, "loss": 4.5034, "step": 243700 }, { "epoch": 0.4548646019701469, "grad_norm": 1.3206427097320557, "learning_rate": 0.0001981102925898029, "loss": 4.7394, "step": 243750 }, { "epoch": 0.45495790752952536, "grad_norm": 0.9127716422080994, "learning_rate": 0.00019810947966988934, "loss": 4.5219, "step": 243800 }, { "epoch": 0.4550512130889039, "grad_norm": 0.9963350296020508, "learning_rate": 0.0001981086665768298, "loss": 4.5856, "step": 243850 }, { "epoch": 0.45514451864828237, "grad_norm": 0.8234110474586487, "learning_rate": 0.0001981078533106257, "loss": 4.7194, "step": 243900 }, { "epoch": 0.45523782420766085, "grad_norm": 0.8868279457092285, "learning_rate": 0.00019810703987127855, "loss": 4.5398, "step": 243950 }, { "epoch": 0.4553311297670393, "grad_norm": 0.8374065160751343, "learning_rate": 0.0001981062262587897, "loss": 4.5695, "step": 244000 }, { "epoch": 0.45542443532641785, "grad_norm": 0.884613573551178, "learning_rate": 0.00019810541247316062, "loss": 4.615, "step": 244050 }, { "epoch": 0.45551774088579633, "grad_norm": 0.8596709966659546, "learning_rate": 0.00019810459851439278, "loss": 4.4475, "step": 244100 }, { "epoch": 0.4556110464451748, "grad_norm": 0.7834071516990662, "learning_rate": 0.00019810378438248758, "loss": 4.4734, "step": 244150 }, { "epoch": 0.45570435200455334, "grad_norm": 0.9869714975357056, "learning_rate": 0.00019810297007744645, "loss": 4.6483, "step": 244200 }, { "epoch": 0.4557976575639318, "grad_norm": 1.230589509010315, "learning_rate": 0.00019810215559927085, "loss": 4.5132, "step": 244250 }, { "epoch": 0.4558909631233103, "grad_norm": 1.0174620151519775, "learning_rate": 0.0001981013409479622, "loss": 4.6885, "step": 244300 }, { "epoch": 0.45598426868268876, "grad_norm": 1.0130627155303955, "learning_rate": 0.00019810052612352194, "loss": 4.4784, "step": 244350 }, { "epoch": 0.4560775742420673, "grad_norm": 0.9201186895370483, "learning_rate": 0.0001980997111259515, "loss": 4.5096, "step": 244400 }, { "epoch": 0.45617087980144577, "grad_norm": 1.1127043962478638, "learning_rate": 0.00019809889595525239, "loss": 4.5587, "step": 244450 }, { "epoch": 0.45626418536082425, "grad_norm": 1.2005751132965088, "learning_rate": 0.0001980980806114259, "loss": 4.6419, "step": 244500 }, { "epoch": 0.4563574909202028, "grad_norm": 1.1009554862976074, "learning_rate": 0.00019809726509447363, "loss": 4.5501, "step": 244550 }, { "epoch": 0.45645079647958126, "grad_norm": 1.0324193239212036, "learning_rate": 0.00019809644940439692, "loss": 4.3942, "step": 244600 }, { "epoch": 0.45654410203895973, "grad_norm": 1.0688228607177734, "learning_rate": 0.00019809563354119725, "loss": 4.5763, "step": 244650 }, { "epoch": 0.4566374075983382, "grad_norm": 0.9501363039016724, "learning_rate": 0.00019809481750487603, "loss": 4.3471, "step": 244700 }, { "epoch": 0.45673071315771674, "grad_norm": 1.0971622467041016, "learning_rate": 0.00019809400129543474, "loss": 4.7443, "step": 244750 }, { "epoch": 0.4568240187170952, "grad_norm": 1.5378892421722412, "learning_rate": 0.0001980931849128748, "loss": 4.4834, "step": 244800 }, { "epoch": 0.4569173242764737, "grad_norm": 1.0331060886383057, "learning_rate": 0.0001980923683571976, "loss": 4.5969, "step": 244850 }, { "epoch": 0.4570106298358522, "grad_norm": 0.9876800775527954, "learning_rate": 0.00019809155162840467, "loss": 4.6108, "step": 244900 }, { "epoch": 0.4571039353952307, "grad_norm": 1.113776683807373, "learning_rate": 0.00019809073472649738, "loss": 4.7422, "step": 244950 }, { "epoch": 0.4571972409546092, "grad_norm": 0.8026116490364075, "learning_rate": 0.00019808991765147727, "loss": 4.6976, "step": 245000 }, { "epoch": 0.45729054651398765, "grad_norm": 1.0806735754013062, "learning_rate": 0.00019808910040334564, "loss": 4.5532, "step": 245050 }, { "epoch": 0.4573838520733662, "grad_norm": 1.0394002199172974, "learning_rate": 0.00019808828298210408, "loss": 4.6943, "step": 245100 }, { "epoch": 0.45747715763274466, "grad_norm": 1.022925853729248, "learning_rate": 0.0001980874653877539, "loss": 4.5151, "step": 245150 }, { "epoch": 0.45757046319212313, "grad_norm": 0.7271643280982971, "learning_rate": 0.00019808664762029662, "loss": 4.4889, "step": 245200 }, { "epoch": 0.4576637687515016, "grad_norm": 0.8295145630836487, "learning_rate": 0.00019808582967973366, "loss": 4.5364, "step": 245250 }, { "epoch": 0.45775707431088014, "grad_norm": 1.0905122756958008, "learning_rate": 0.0001980850115660665, "loss": 4.7712, "step": 245300 }, { "epoch": 0.4578503798702586, "grad_norm": 1.2582670450210571, "learning_rate": 0.0001980841932792965, "loss": 4.557, "step": 245350 }, { "epoch": 0.4579436854296371, "grad_norm": 0.8376700282096863, "learning_rate": 0.0001980833748194252, "loss": 4.5676, "step": 245400 }, { "epoch": 0.4580369909890156, "grad_norm": 0.8749433159828186, "learning_rate": 0.00019808255618645397, "loss": 4.729, "step": 245450 }, { "epoch": 0.4581302965483941, "grad_norm": 1.354799747467041, "learning_rate": 0.0001980817373803843, "loss": 4.5384, "step": 245500 }, { "epoch": 0.4582236021077726, "grad_norm": 1.1467299461364746, "learning_rate": 0.00019808091840121765, "loss": 4.6499, "step": 245550 }, { "epoch": 0.45831690766715105, "grad_norm": 1.0075682401657104, "learning_rate": 0.0001980800992489554, "loss": 4.643, "step": 245600 }, { "epoch": 0.4584102132265296, "grad_norm": 1.1275964975357056, "learning_rate": 0.00019807927992359904, "loss": 4.7408, "step": 245650 }, { "epoch": 0.45850351878590806, "grad_norm": 0.9080273509025574, "learning_rate": 0.00019807846042515004, "loss": 4.8073, "step": 245700 }, { "epoch": 0.45859682434528654, "grad_norm": 0.7634953856468201, "learning_rate": 0.00019807764075360977, "loss": 4.6437, "step": 245750 }, { "epoch": 0.45869012990466507, "grad_norm": 0.9849189519882202, "learning_rate": 0.00019807682090897976, "loss": 4.6905, "step": 245800 }, { "epoch": 0.45878343546404354, "grad_norm": 0.7696005702018738, "learning_rate": 0.0001980760008912614, "loss": 4.5023, "step": 245850 }, { "epoch": 0.458876741023422, "grad_norm": 1.1144013404846191, "learning_rate": 0.00019807518070045615, "loss": 4.7656, "step": 245900 }, { "epoch": 0.4589700465828005, "grad_norm": 1.2693634033203125, "learning_rate": 0.00019807436033656545, "loss": 4.572, "step": 245950 }, { "epoch": 0.459063352142179, "grad_norm": 1.1602495908737183, "learning_rate": 0.0001980735397995908, "loss": 4.5726, "step": 246000 }, { "epoch": 0.459063352142179, "eval_loss": 4.774869441986084, "eval_runtime": 233.5371, "eval_samples_per_second": 11.167, "eval_steps_per_second": 11.167, "eval_tts_loss": 7.48305198906898, "step": 246000 }, { "epoch": 0.4591566577015575, "grad_norm": 0.9617483615875244, "learning_rate": 0.0001980727190895336, "loss": 4.4622, "step": 246050 }, { "epoch": 0.459249963260936, "grad_norm": 0.9876244068145752, "learning_rate": 0.00019807189820639527, "loss": 4.7104, "step": 246100 }, { "epoch": 0.4593432688203145, "grad_norm": 1.061923861503601, "learning_rate": 0.00019807107715017733, "loss": 4.6161, "step": 246150 }, { "epoch": 0.459436574379693, "grad_norm": 1.3784229755401611, "learning_rate": 0.00019807025592088117, "loss": 4.5802, "step": 246200 }, { "epoch": 0.45952987993907146, "grad_norm": 1.1820831298828125, "learning_rate": 0.00019806943451850827, "loss": 4.5958, "step": 246250 }, { "epoch": 0.45962318549844994, "grad_norm": 0.9838506579399109, "learning_rate": 0.0001980686129430601, "loss": 4.7938, "step": 246300 }, { "epoch": 0.45971649105782847, "grad_norm": 1.1807385683059692, "learning_rate": 0.00019806779119453807, "loss": 4.5621, "step": 246350 }, { "epoch": 0.45980979661720695, "grad_norm": 1.1486557722091675, "learning_rate": 0.00019806696927294364, "loss": 4.5431, "step": 246400 }, { "epoch": 0.4599031021765854, "grad_norm": 1.1213480234146118, "learning_rate": 0.00019806614717827827, "loss": 4.5494, "step": 246450 }, { "epoch": 0.45999640773596395, "grad_norm": 1.0412524938583374, "learning_rate": 0.00019806532491054338, "loss": 4.6432, "step": 246500 }, { "epoch": 0.46008971329534243, "grad_norm": 1.125463843345642, "learning_rate": 0.00019806450246974048, "loss": 4.6128, "step": 246550 }, { "epoch": 0.4601830188547209, "grad_norm": 1.0406776666641235, "learning_rate": 0.00019806367985587099, "loss": 4.7215, "step": 246600 }, { "epoch": 0.4602763244140994, "grad_norm": 1.2460758686065674, "learning_rate": 0.00019806285706893632, "loss": 4.6226, "step": 246650 }, { "epoch": 0.4603696299734779, "grad_norm": 0.9566603302955627, "learning_rate": 0.000198062034108938, "loss": 4.4905, "step": 246700 }, { "epoch": 0.4604629355328564, "grad_norm": 1.013724684715271, "learning_rate": 0.00019806121097587745, "loss": 4.507, "step": 246750 }, { "epoch": 0.46055624109223486, "grad_norm": 0.9310964941978455, "learning_rate": 0.00019806038766975606, "loss": 4.3787, "step": 246800 }, { "epoch": 0.4606495466516134, "grad_norm": 1.0492833852767944, "learning_rate": 0.0001980595641905754, "loss": 4.5084, "step": 246850 }, { "epoch": 0.46074285221099187, "grad_norm": 0.5750857591629028, "learning_rate": 0.00019805874053833685, "loss": 4.5357, "step": 246900 }, { "epoch": 0.46083615777037035, "grad_norm": 1.0940406322479248, "learning_rate": 0.00019805791671304187, "loss": 4.4754, "step": 246950 }, { "epoch": 0.4609294633297488, "grad_norm": 1.030585527420044, "learning_rate": 0.0001980570927146919, "loss": 4.6358, "step": 247000 }, { "epoch": 0.46102276888912735, "grad_norm": 1.0150128602981567, "learning_rate": 0.00019805626854328844, "loss": 4.7402, "step": 247050 }, { "epoch": 0.46111607444850583, "grad_norm": 2.006843090057373, "learning_rate": 0.00019805544419883292, "loss": 4.5641, "step": 247100 }, { "epoch": 0.4612093800078843, "grad_norm": 1.1281328201293945, "learning_rate": 0.00019805461968132679, "loss": 4.5954, "step": 247150 }, { "epoch": 0.46130268556726284, "grad_norm": 1.0709025859832764, "learning_rate": 0.00019805379499077152, "loss": 4.7013, "step": 247200 }, { "epoch": 0.4613959911266413, "grad_norm": 1.0696786642074585, "learning_rate": 0.00019805297012716856, "loss": 4.6369, "step": 247250 }, { "epoch": 0.4614892966860198, "grad_norm": 1.0185669660568237, "learning_rate": 0.00019805214509051932, "loss": 4.685, "step": 247300 }, { "epoch": 0.46158260224539827, "grad_norm": 0.921750009059906, "learning_rate": 0.00019805131988082534, "loss": 4.6481, "step": 247350 }, { "epoch": 0.4616759078047768, "grad_norm": 0.625173032283783, "learning_rate": 0.000198050494498088, "loss": 4.4107, "step": 247400 }, { "epoch": 0.4617692133641553, "grad_norm": 0.8325273990631104, "learning_rate": 0.0001980496689423088, "loss": 4.7302, "step": 247450 }, { "epoch": 0.46186251892353375, "grad_norm": 1.2000508308410645, "learning_rate": 0.00019804884321348917, "loss": 4.7503, "step": 247500 }, { "epoch": 0.4619558244829123, "grad_norm": 1.06386137008667, "learning_rate": 0.00019804801731163063, "loss": 4.5668, "step": 247550 }, { "epoch": 0.46204913004229076, "grad_norm": 1.1253514289855957, "learning_rate": 0.00019804719123673456, "loss": 4.4186, "step": 247600 }, { "epoch": 0.46214243560166923, "grad_norm": 0.9880508780479431, "learning_rate": 0.00019804636498880248, "loss": 4.6186, "step": 247650 }, { "epoch": 0.4622357411610477, "grad_norm": 1.0324673652648926, "learning_rate": 0.00019804553856783577, "loss": 4.4925, "step": 247700 }, { "epoch": 0.46232904672042624, "grad_norm": 0.7357877492904663, "learning_rate": 0.00019804471197383596, "loss": 4.6977, "step": 247750 }, { "epoch": 0.4624223522798047, "grad_norm": 0.9549800753593445, "learning_rate": 0.0001980438852068045, "loss": 4.4168, "step": 247800 }, { "epoch": 0.4625156578391832, "grad_norm": 0.9433833956718445, "learning_rate": 0.0001980430582667428, "loss": 4.6086, "step": 247850 }, { "epoch": 0.46260896339856167, "grad_norm": 1.2438510656356812, "learning_rate": 0.00019804223115365237, "loss": 4.6693, "step": 247900 }, { "epoch": 0.4627022689579402, "grad_norm": 0.8715135455131531, "learning_rate": 0.00019804140386753467, "loss": 4.6429, "step": 247950 }, { "epoch": 0.4627955745173187, "grad_norm": 0.9428792595863342, "learning_rate": 0.00019804057640839112, "loss": 4.5094, "step": 248000 }, { "epoch": 0.46288888007669715, "grad_norm": 0.8112114071846008, "learning_rate": 0.0001980397487762232, "loss": 4.538, "step": 248050 }, { "epoch": 0.4629821856360757, "grad_norm": 0.899664044380188, "learning_rate": 0.0001980389209710324, "loss": 4.4226, "step": 248100 }, { "epoch": 0.46307549119545416, "grad_norm": 1.0924978256225586, "learning_rate": 0.00019803809299282014, "loss": 4.6605, "step": 248150 }, { "epoch": 0.46316879675483263, "grad_norm": 0.9904999136924744, "learning_rate": 0.0001980372648415879, "loss": 4.5332, "step": 248200 }, { "epoch": 0.4632621023142111, "grad_norm": 1.1240946054458618, "learning_rate": 0.00019803643651733713, "loss": 4.6454, "step": 248250 }, { "epoch": 0.46335540787358964, "grad_norm": 1.0554358959197998, "learning_rate": 0.0001980356080200693, "loss": 4.757, "step": 248300 }, { "epoch": 0.4634487134329681, "grad_norm": 1.0673205852508545, "learning_rate": 0.00019803477934978586, "loss": 4.3607, "step": 248350 }, { "epoch": 0.4635420189923466, "grad_norm": 1.0549492835998535, "learning_rate": 0.00019803395050648832, "loss": 4.4935, "step": 248400 }, { "epoch": 0.4636353245517251, "grad_norm": 1.18800687789917, "learning_rate": 0.00019803312149017807, "loss": 4.7759, "step": 248450 }, { "epoch": 0.4637286301111036, "grad_norm": 1.0293747186660767, "learning_rate": 0.00019803229230085662, "loss": 4.5521, "step": 248500 }, { "epoch": 0.4638219356704821, "grad_norm": 1.0756644010543823, "learning_rate": 0.00019803146293852545, "loss": 4.5545, "step": 248550 }, { "epoch": 0.46391524122986055, "grad_norm": 0.8822924494743347, "learning_rate": 0.00019803063340318597, "loss": 4.4807, "step": 248600 }, { "epoch": 0.4640085467892391, "grad_norm": 1.109789490699768, "learning_rate": 0.00019802980369483967, "loss": 4.5922, "step": 248650 }, { "epoch": 0.46410185234861756, "grad_norm": 1.0648267269134521, "learning_rate": 0.00019802897381348804, "loss": 4.5294, "step": 248700 }, { "epoch": 0.46419515790799604, "grad_norm": 0.9832372069358826, "learning_rate": 0.00019802814375913247, "loss": 4.6105, "step": 248750 }, { "epoch": 0.46428846346737457, "grad_norm": 1.0333669185638428, "learning_rate": 0.00019802731353177452, "loss": 4.6382, "step": 248800 }, { "epoch": 0.46438176902675304, "grad_norm": 1.1665089130401611, "learning_rate": 0.00019802648313141562, "loss": 4.5055, "step": 248850 }, { "epoch": 0.4644750745861315, "grad_norm": 0.7576875686645508, "learning_rate": 0.00019802565255805718, "loss": 4.555, "step": 248900 }, { "epoch": 0.46456838014551, "grad_norm": 0.8509981036186218, "learning_rate": 0.00019802482181170074, "loss": 4.6874, "step": 248950 }, { "epoch": 0.4646616857048885, "grad_norm": 1.0222673416137695, "learning_rate": 0.00019802399089234775, "loss": 4.7886, "step": 249000 }, { "epoch": 0.4646616857048885, "eval_loss": 4.778683662414551, "eval_runtime": 230.8207, "eval_samples_per_second": 11.299, "eval_steps_per_second": 11.299, "eval_tts_loss": 7.546051602269254, "step": 249000 }, { "epoch": 0.464754991264267, "grad_norm": 0.9813963770866394, "learning_rate": 0.00019802315979999965, "loss": 4.4973, "step": 249050 }, { "epoch": 0.4648482968236455, "grad_norm": 1.1679600477218628, "learning_rate": 0.0001980223285346579, "loss": 4.716, "step": 249100 }, { "epoch": 0.464941602383024, "grad_norm": 1.014891505241394, "learning_rate": 0.000198021497096324, "loss": 4.641, "step": 249150 }, { "epoch": 0.4650349079424025, "grad_norm": 1.1043691635131836, "learning_rate": 0.00019802066548499941, "loss": 4.712, "step": 249200 }, { "epoch": 0.46512821350178096, "grad_norm": 1.0716991424560547, "learning_rate": 0.0001980198337006856, "loss": 4.6739, "step": 249250 }, { "epoch": 0.46522151906115944, "grad_norm": 1.2139348983764648, "learning_rate": 0.00019801900174338402, "loss": 4.6826, "step": 249300 }, { "epoch": 0.46531482462053797, "grad_norm": 0.8604678511619568, "learning_rate": 0.00019801816961309617, "loss": 4.6827, "step": 249350 }, { "epoch": 0.46540813017991645, "grad_norm": 1.0838696956634521, "learning_rate": 0.00019801733730982348, "loss": 4.6474, "step": 249400 }, { "epoch": 0.4655014357392949, "grad_norm": 1.5749475955963135, "learning_rate": 0.00019801650483356742, "loss": 4.4896, "step": 249450 }, { "epoch": 0.46559474129867345, "grad_norm": 0.7585879564285278, "learning_rate": 0.0001980156721843295, "loss": 4.3649, "step": 249500 }, { "epoch": 0.46568804685805193, "grad_norm": 1.1551792621612549, "learning_rate": 0.00019801483936211117, "loss": 4.6776, "step": 249550 }, { "epoch": 0.4657813524174304, "grad_norm": 0.8028654456138611, "learning_rate": 0.00019801400636691385, "loss": 4.4243, "step": 249600 }, { "epoch": 0.4658746579768089, "grad_norm": 1.0627497434616089, "learning_rate": 0.0001980131731987391, "loss": 4.3195, "step": 249650 }, { "epoch": 0.4659679635361874, "grad_norm": 1.1818534135818481, "learning_rate": 0.00019801233985758835, "loss": 4.5887, "step": 249700 }, { "epoch": 0.4660612690955659, "grad_norm": 1.0179072618484497, "learning_rate": 0.000198011506343463, "loss": 4.5652, "step": 249750 }, { "epoch": 0.46615457465494436, "grad_norm": 1.0190846920013428, "learning_rate": 0.00019801067265636465, "loss": 4.7446, "step": 249800 }, { "epoch": 0.4662478802143229, "grad_norm": 0.9870844483375549, "learning_rate": 0.0001980098387962947, "loss": 4.6793, "step": 249850 }, { "epoch": 0.46634118577370137, "grad_norm": 0.8866018652915955, "learning_rate": 0.00019800900476325462, "loss": 4.7872, "step": 249900 }, { "epoch": 0.46643449133307985, "grad_norm": 1.2715367078781128, "learning_rate": 0.0001980081705572459, "loss": 4.7906, "step": 249950 }, { "epoch": 0.4665277968924583, "grad_norm": 1.1326353549957275, "learning_rate": 0.00019800733617827, "loss": 4.7138, "step": 250000 }, { "epoch": 0.46662110245183686, "grad_norm": 1.0294839143753052, "learning_rate": 0.0001980065016263284, "loss": 4.5253, "step": 250050 }, { "epoch": 0.46671440801121533, "grad_norm": 1.1062980890274048, "learning_rate": 0.00019800566690142258, "loss": 4.6491, "step": 250100 }, { "epoch": 0.4668077135705938, "grad_norm": 0.8462019562721252, "learning_rate": 0.00019800483200355394, "loss": 4.6245, "step": 250150 }, { "epoch": 0.46690101912997234, "grad_norm": 0.8871804475784302, "learning_rate": 0.0001980039969327241, "loss": 4.7166, "step": 250200 }, { "epoch": 0.4669943246893508, "grad_norm": 1.0749971866607666, "learning_rate": 0.0001980031616889344, "loss": 4.4303, "step": 250250 }, { "epoch": 0.4670876302487293, "grad_norm": 1.1696592569351196, "learning_rate": 0.00019800232627218637, "loss": 4.6637, "step": 250300 }, { "epoch": 0.46718093580810777, "grad_norm": 1.017404317855835, "learning_rate": 0.0001980014906824815, "loss": 4.6735, "step": 250350 }, { "epoch": 0.4672742413674863, "grad_norm": 0.9707728624343872, "learning_rate": 0.00019800065491982122, "loss": 4.5803, "step": 250400 }, { "epoch": 0.4673675469268648, "grad_norm": 1.125417709350586, "learning_rate": 0.00019799981898420703, "loss": 4.6582, "step": 250450 }, { "epoch": 0.46746085248624325, "grad_norm": 0.9616281986236572, "learning_rate": 0.00019799898287564042, "loss": 4.545, "step": 250500 }, { "epoch": 0.4675541580456217, "grad_norm": 0.8503236770629883, "learning_rate": 0.00019799814659412284, "loss": 4.6893, "step": 250550 }, { "epoch": 0.46764746360500026, "grad_norm": 1.1284072399139404, "learning_rate": 0.00019799731013965576, "loss": 4.6788, "step": 250600 }, { "epoch": 0.46774076916437873, "grad_norm": 0.9626160264015198, "learning_rate": 0.00019799647351224068, "loss": 4.577, "step": 250650 }, { "epoch": 0.4678340747237572, "grad_norm": 0.9178403615951538, "learning_rate": 0.00019799563671187907, "loss": 4.3979, "step": 250700 }, { "epoch": 0.46792738028313574, "grad_norm": 0.9638570547103882, "learning_rate": 0.0001979947997385724, "loss": 4.4919, "step": 250750 }, { "epoch": 0.4680206858425142, "grad_norm": 0.9107949733734131, "learning_rate": 0.00019799396259232216, "loss": 4.5359, "step": 250800 }, { "epoch": 0.4681139914018927, "grad_norm": 1.2097649574279785, "learning_rate": 0.00019799312527312983, "loss": 4.5872, "step": 250850 }, { "epoch": 0.46820729696127117, "grad_norm": 0.9304313063621521, "learning_rate": 0.00019799228778099687, "loss": 4.6143, "step": 250900 }, { "epoch": 0.4683006025206497, "grad_norm": 0.9003384113311768, "learning_rate": 0.00019799145011592474, "loss": 4.6902, "step": 250950 }, { "epoch": 0.4683939080800282, "grad_norm": 0.825849711894989, "learning_rate": 0.00019799061227791494, "loss": 4.5176, "step": 251000 }, { "epoch": 0.46848721363940665, "grad_norm": 0.8515427708625793, "learning_rate": 0.00019798977426696896, "loss": 4.5787, "step": 251050 }, { "epoch": 0.4685805191987852, "grad_norm": 0.9630306959152222, "learning_rate": 0.0001979889360830883, "loss": 4.4875, "step": 251100 }, { "epoch": 0.46867382475816366, "grad_norm": 0.694108784198761, "learning_rate": 0.00019798809772627438, "loss": 4.2851, "step": 251150 }, { "epoch": 0.46876713031754214, "grad_norm": 0.8134151697158813, "learning_rate": 0.00019798725919652872, "loss": 4.6372, "step": 251200 }, { "epoch": 0.4688604358769206, "grad_norm": 0.8471952080726624, "learning_rate": 0.00019798642049385278, "loss": 4.3836, "step": 251250 }, { "epoch": 0.46895374143629914, "grad_norm": 1.0113226175308228, "learning_rate": 0.00019798558161824805, "loss": 4.3661, "step": 251300 }, { "epoch": 0.4690470469956776, "grad_norm": 0.840195894241333, "learning_rate": 0.000197984742569716, "loss": 4.5986, "step": 251350 }, { "epoch": 0.4691403525550561, "grad_norm": 0.9539893865585327, "learning_rate": 0.00019798390334825817, "loss": 4.6197, "step": 251400 }, { "epoch": 0.4692336581144346, "grad_norm": 0.8172125816345215, "learning_rate": 0.00019798306395387596, "loss": 4.6553, "step": 251450 }, { "epoch": 0.4693269636738131, "grad_norm": 1.068747878074646, "learning_rate": 0.00019798222438657085, "loss": 4.5775, "step": 251500 }, { "epoch": 0.4694202692331916, "grad_norm": 1.1480756998062134, "learning_rate": 0.00019798138464634439, "loss": 4.4341, "step": 251550 }, { "epoch": 0.46951357479257005, "grad_norm": 1.1255130767822266, "learning_rate": 0.00019798054473319803, "loss": 4.4635, "step": 251600 }, { "epoch": 0.4696068803519486, "grad_norm": 0.9743191599845886, "learning_rate": 0.00019797970464713321, "loss": 4.8709, "step": 251650 }, { "epoch": 0.46970018591132706, "grad_norm": 1.054312825202942, "learning_rate": 0.0001979788643881515, "loss": 4.7223, "step": 251700 }, { "epoch": 0.46979349147070554, "grad_norm": 0.6565726399421692, "learning_rate": 0.00019797802395625432, "loss": 4.6015, "step": 251750 }, { "epoch": 0.46988679703008407, "grad_norm": 1.1399985551834106, "learning_rate": 0.00019797718335144316, "loss": 4.7225, "step": 251800 }, { "epoch": 0.46998010258946255, "grad_norm": 0.9679462909698486, "learning_rate": 0.00019797634257371952, "loss": 4.6508, "step": 251850 }, { "epoch": 0.470073408148841, "grad_norm": 1.1389963626861572, "learning_rate": 0.00019797550162308487, "loss": 4.5277, "step": 251900 }, { "epoch": 0.4701667137082195, "grad_norm": 0.8405935764312744, "learning_rate": 0.00019797466049954068, "loss": 4.4807, "step": 251950 }, { "epoch": 0.47026001926759803, "grad_norm": 0.9860919713973999, "learning_rate": 0.00019797381920308847, "loss": 4.5945, "step": 252000 }, { "epoch": 0.47026001926759803, "eval_loss": 4.771847724914551, "eval_runtime": 228.5901, "eval_samples_per_second": 11.409, "eval_steps_per_second": 11.409, "eval_tts_loss": 7.527745360290014, "step": 252000 }, { "epoch": 0.4703533248269765, "grad_norm": 0.8719726204872131, "learning_rate": 0.0001979729777337297, "loss": 4.5289, "step": 252050 }, { "epoch": 0.470446630386355, "grad_norm": 1.0118962526321411, "learning_rate": 0.00019797213609146588, "loss": 4.6344, "step": 252100 }, { "epoch": 0.4705399359457335, "grad_norm": 1.014058232307434, "learning_rate": 0.00019797129427629847, "loss": 4.6503, "step": 252150 }, { "epoch": 0.470633241505112, "grad_norm": 0.9134781360626221, "learning_rate": 0.00019797045228822898, "loss": 4.538, "step": 252200 }, { "epoch": 0.47072654706449046, "grad_norm": 1.0648183822631836, "learning_rate": 0.00019796961012725884, "loss": 4.5453, "step": 252250 }, { "epoch": 0.47081985262386894, "grad_norm": 1.1719633340835571, "learning_rate": 0.0001979687677933896, "loss": 4.7073, "step": 252300 }, { "epoch": 0.47091315818324747, "grad_norm": 1.206669569015503, "learning_rate": 0.00019796792528662272, "loss": 4.708, "step": 252350 }, { "epoch": 0.47100646374262595, "grad_norm": 1.1847809553146362, "learning_rate": 0.00019796708260695972, "loss": 4.5306, "step": 252400 }, { "epoch": 0.4710997693020044, "grad_norm": 0.8223580718040466, "learning_rate": 0.00019796623975440204, "loss": 4.5695, "step": 252450 }, { "epoch": 0.47119307486138295, "grad_norm": 0.9964529871940613, "learning_rate": 0.00019796539672895116, "loss": 4.6491, "step": 252500 }, { "epoch": 0.47128638042076143, "grad_norm": 0.9831516146659851, "learning_rate": 0.0001979645535306086, "loss": 4.6033, "step": 252550 }, { "epoch": 0.4713796859801399, "grad_norm": 0.7860525250434875, "learning_rate": 0.00019796371015937586, "loss": 4.4849, "step": 252600 }, { "epoch": 0.4714729915395184, "grad_norm": 1.078437328338623, "learning_rate": 0.00019796286661525437, "loss": 4.7393, "step": 252650 }, { "epoch": 0.4715662970988969, "grad_norm": 0.8938062191009521, "learning_rate": 0.0001979620228982457, "loss": 4.4925, "step": 252700 }, { "epoch": 0.4716596026582754, "grad_norm": 1.3449671268463135, "learning_rate": 0.00019796117900835127, "loss": 4.5993, "step": 252750 }, { "epoch": 0.47175290821765387, "grad_norm": 1.0457773208618164, "learning_rate": 0.00019796033494557262, "loss": 4.507, "step": 252800 }, { "epoch": 0.4718462137770324, "grad_norm": 1.118150234222412, "learning_rate": 0.0001979594907099112, "loss": 4.5657, "step": 252850 }, { "epoch": 0.4719395193364109, "grad_norm": 1.0260703563690186, "learning_rate": 0.00019795864630136852, "loss": 4.5898, "step": 252900 }, { "epoch": 0.47203282489578935, "grad_norm": 0.9269238114356995, "learning_rate": 0.00019795780171994606, "loss": 4.808, "step": 252950 }, { "epoch": 0.4721261304551678, "grad_norm": 1.0303397178649902, "learning_rate": 0.00019795695696564534, "loss": 4.4025, "step": 253000 }, { "epoch": 0.47221943601454636, "grad_norm": 0.9815906882286072, "learning_rate": 0.00019795611203846778, "loss": 4.6513, "step": 253050 }, { "epoch": 0.47231274157392483, "grad_norm": 1.0139660835266113, "learning_rate": 0.00019795526693841496, "loss": 4.5168, "step": 253100 }, { "epoch": 0.4724060471333033, "grad_norm": 1.0596623420715332, "learning_rate": 0.0001979544216654883, "loss": 4.6385, "step": 253150 }, { "epoch": 0.4724993526926818, "grad_norm": 1.116756796836853, "learning_rate": 0.00019795357621968935, "loss": 4.6691, "step": 253200 }, { "epoch": 0.4725926582520603, "grad_norm": 0.955316960811615, "learning_rate": 0.00019795273060101956, "loss": 4.6902, "step": 253250 }, { "epoch": 0.4726859638114388, "grad_norm": 0.9806995987892151, "learning_rate": 0.00019795188480948043, "loss": 4.6256, "step": 253300 }, { "epoch": 0.47277926937081727, "grad_norm": 1.410809874534607, "learning_rate": 0.00019795103884507345, "loss": 4.5668, "step": 253350 }, { "epoch": 0.4728725749301958, "grad_norm": 0.9873063564300537, "learning_rate": 0.00019795019270780014, "loss": 4.5323, "step": 253400 }, { "epoch": 0.4729658804895743, "grad_norm": 0.8761279582977295, "learning_rate": 0.00019794934639766196, "loss": 4.5232, "step": 253450 }, { "epoch": 0.47305918604895275, "grad_norm": 0.7007517218589783, "learning_rate": 0.00019794849991466046, "loss": 4.4708, "step": 253500 }, { "epoch": 0.4731524916083312, "grad_norm": 0.8728896379470825, "learning_rate": 0.00019794765325879704, "loss": 4.5846, "step": 253550 }, { "epoch": 0.47324579716770976, "grad_norm": 1.0133166313171387, "learning_rate": 0.00019794680643007325, "loss": 4.5587, "step": 253600 }, { "epoch": 0.47333910272708823, "grad_norm": 1.3427821397781372, "learning_rate": 0.0001979459594284906, "loss": 4.663, "step": 253650 }, { "epoch": 0.4734324082864667, "grad_norm": 1.2627538442611694, "learning_rate": 0.00019794511225405058, "loss": 4.6541, "step": 253700 }, { "epoch": 0.47352571384584524, "grad_norm": 1.0716121196746826, "learning_rate": 0.00019794426490675463, "loss": 4.5518, "step": 253750 }, { "epoch": 0.4736190194052237, "grad_norm": 1.0432318449020386, "learning_rate": 0.00019794341738660432, "loss": 4.5539, "step": 253800 }, { "epoch": 0.4737123249646022, "grad_norm": 1.083561658859253, "learning_rate": 0.00019794256969360113, "loss": 4.7249, "step": 253850 }, { "epoch": 0.47380563052398067, "grad_norm": 1.0139247179031372, "learning_rate": 0.00019794172182774648, "loss": 4.5797, "step": 253900 }, { "epoch": 0.4738989360833592, "grad_norm": 1.054151177406311, "learning_rate": 0.00019794087378904195, "loss": 4.6604, "step": 253950 }, { "epoch": 0.4739922416427377, "grad_norm": 1.1216671466827393, "learning_rate": 0.00019794002557748904, "loss": 4.6178, "step": 254000 }, { "epoch": 0.47408554720211615, "grad_norm": 1.1557576656341553, "learning_rate": 0.0001979391771930892, "loss": 4.5867, "step": 254050 }, { "epoch": 0.4741788527614947, "grad_norm": 1.1198077201843262, "learning_rate": 0.00019793832863584392, "loss": 4.6404, "step": 254100 }, { "epoch": 0.47427215832087316, "grad_norm": 0.9180806279182434, "learning_rate": 0.00019793747990575473, "loss": 4.4542, "step": 254150 }, { "epoch": 0.47436546388025164, "grad_norm": 0.9826111793518066, "learning_rate": 0.00019793663100282314, "loss": 4.6157, "step": 254200 }, { "epoch": 0.4744587694396301, "grad_norm": 1.0048807859420776, "learning_rate": 0.00019793578192705064, "loss": 4.4314, "step": 254250 }, { "epoch": 0.47455207499900864, "grad_norm": 1.0159192085266113, "learning_rate": 0.0001979349326784387, "loss": 4.6976, "step": 254300 }, { "epoch": 0.4746453805583871, "grad_norm": 1.0906916856765747, "learning_rate": 0.00019793408325698882, "loss": 4.7204, "step": 254350 }, { "epoch": 0.4747386861177656, "grad_norm": 1.0381954908370972, "learning_rate": 0.00019793323366270254, "loss": 4.4653, "step": 254400 }, { "epoch": 0.4748319916771441, "grad_norm": 1.0073879957199097, "learning_rate": 0.0001979323838955813, "loss": 4.5649, "step": 254450 }, { "epoch": 0.4749252972365226, "grad_norm": 1.224362850189209, "learning_rate": 0.00019793153395562663, "loss": 4.7343, "step": 254500 }, { "epoch": 0.4750186027959011, "grad_norm": 0.9153668880462646, "learning_rate": 0.00019793068384284007, "loss": 4.6776, "step": 254550 }, { "epoch": 0.47511190835527956, "grad_norm": 1.0687617063522339, "learning_rate": 0.0001979298335572231, "loss": 4.4508, "step": 254600 }, { "epoch": 0.4752052139146581, "grad_norm": 1.1399562358856201, "learning_rate": 0.00019792898309877716, "loss": 4.5694, "step": 254650 }, { "epoch": 0.47529851947403656, "grad_norm": 1.0620476007461548, "learning_rate": 0.00019792813246750384, "loss": 4.4951, "step": 254700 }, { "epoch": 0.47539182503341504, "grad_norm": 0.987824559211731, "learning_rate": 0.00019792728166340457, "loss": 4.5894, "step": 254750 }, { "epoch": 0.47548513059279357, "grad_norm": 0.9981347322463989, "learning_rate": 0.00019792643068648088, "loss": 4.5573, "step": 254800 }, { "epoch": 0.47557843615217205, "grad_norm": 1.3554757833480835, "learning_rate": 0.00019792557953673428, "loss": 4.4722, "step": 254850 }, { "epoch": 0.4756717417115505, "grad_norm": 1.2664356231689453, "learning_rate": 0.00019792472821416625, "loss": 4.467, "step": 254900 }, { "epoch": 0.475765047270929, "grad_norm": 1.125602126121521, "learning_rate": 0.00019792387671877833, "loss": 4.5172, "step": 254950 }, { "epoch": 0.47585835283030753, "grad_norm": 0.9335258603096008, "learning_rate": 0.00019792302505057196, "loss": 4.5266, "step": 255000 }, { "epoch": 0.47585835283030753, "eval_loss": 4.763278484344482, "eval_runtime": 230.157, "eval_samples_per_second": 11.331, "eval_steps_per_second": 11.331, "eval_tts_loss": 7.542246853862746, "step": 255000 }, { "epoch": 0.475951658389686, "grad_norm": 1.0923206806182861, "learning_rate": 0.00019792217320954868, "loss": 4.7171, "step": 255050 }, { "epoch": 0.4760449639490645, "grad_norm": 1.037209391593933, "learning_rate": 0.00019792132119571002, "loss": 4.4465, "step": 255100 }, { "epoch": 0.476138269508443, "grad_norm": 1.2115193605422974, "learning_rate": 0.00019792046900905748, "loss": 4.4359, "step": 255150 }, { "epoch": 0.4762315750678215, "grad_norm": 1.0448733568191528, "learning_rate": 0.0001979196166495925, "loss": 4.5225, "step": 255200 }, { "epoch": 0.47632488062719996, "grad_norm": 1.1041306257247925, "learning_rate": 0.00019791876411731665, "loss": 4.7055, "step": 255250 }, { "epoch": 0.47641818618657844, "grad_norm": 0.9525231122970581, "learning_rate": 0.0001979179114122314, "loss": 4.6905, "step": 255300 }, { "epoch": 0.47651149174595697, "grad_norm": 1.0401257276535034, "learning_rate": 0.00019791705853433827, "loss": 4.5725, "step": 255350 }, { "epoch": 0.47660479730533545, "grad_norm": 0.9512485265731812, "learning_rate": 0.00019791620548363875, "loss": 4.6771, "step": 255400 }, { "epoch": 0.4766981028647139, "grad_norm": 0.9402346611022949, "learning_rate": 0.00019791535226013437, "loss": 4.6207, "step": 255450 }, { "epoch": 0.47679140842409246, "grad_norm": 1.0512655973434448, "learning_rate": 0.0001979144988638266, "loss": 4.5787, "step": 255500 }, { "epoch": 0.47688471398347093, "grad_norm": 0.9695249795913696, "learning_rate": 0.000197913645294717, "loss": 4.5116, "step": 255550 }, { "epoch": 0.4769780195428494, "grad_norm": 0.9149754643440247, "learning_rate": 0.000197912791552807, "loss": 4.6068, "step": 255600 }, { "epoch": 0.4770713251022279, "grad_norm": 1.0737544298171997, "learning_rate": 0.00019791193763809818, "loss": 4.6188, "step": 255650 }, { "epoch": 0.4771646306616064, "grad_norm": 0.9186537861824036, "learning_rate": 0.00019791108355059203, "loss": 4.446, "step": 255700 }, { "epoch": 0.4772579362209849, "grad_norm": 1.1221696138381958, "learning_rate": 0.00019791022929029005, "loss": 4.5936, "step": 255750 }, { "epoch": 0.47735124178036337, "grad_norm": 1.1481096744537354, "learning_rate": 0.00019790937485719373, "loss": 4.5989, "step": 255800 }, { "epoch": 0.47744454733974184, "grad_norm": 0.8442092537879944, "learning_rate": 0.00019790852025130455, "loss": 4.5942, "step": 255850 }, { "epoch": 0.4775378528991204, "grad_norm": 0.8604154586791992, "learning_rate": 0.00019790766547262412, "loss": 4.5971, "step": 255900 }, { "epoch": 0.47763115845849885, "grad_norm": 0.8299677968025208, "learning_rate": 0.00019790681052115385, "loss": 4.4849, "step": 255950 }, { "epoch": 0.4777244640178773, "grad_norm": 0.9092508554458618, "learning_rate": 0.0001979059553968953, "loss": 4.5899, "step": 256000 }, { "epoch": 0.47781776957725586, "grad_norm": 0.8719881176948547, "learning_rate": 0.00019790510009984997, "loss": 4.6535, "step": 256050 }, { "epoch": 0.47791107513663433, "grad_norm": 1.0991895198822021, "learning_rate": 0.00019790424463001936, "loss": 4.6057, "step": 256100 }, { "epoch": 0.4780043806960128, "grad_norm": 0.9137606620788574, "learning_rate": 0.00019790338898740498, "loss": 4.6306, "step": 256150 }, { "epoch": 0.4780976862553913, "grad_norm": 0.8069062829017639, "learning_rate": 0.00019790253317200835, "loss": 4.6064, "step": 256200 }, { "epoch": 0.4781909918147698, "grad_norm": 1.0056474208831787, "learning_rate": 0.00019790167718383098, "loss": 4.4567, "step": 256250 }, { "epoch": 0.4782842973741483, "grad_norm": 1.1140251159667969, "learning_rate": 0.00019790082102287437, "loss": 4.6005, "step": 256300 }, { "epoch": 0.47837760293352677, "grad_norm": 1.049212098121643, "learning_rate": 0.00019789996468914004, "loss": 4.5948, "step": 256350 }, { "epoch": 0.4784709084929053, "grad_norm": 1.2043838500976562, "learning_rate": 0.00019789910818262952, "loss": 4.7654, "step": 256400 }, { "epoch": 0.4785642140522838, "grad_norm": 1.0834003686904907, "learning_rate": 0.00019789825150334426, "loss": 4.4575, "step": 256450 }, { "epoch": 0.47865751961166225, "grad_norm": 1.3048619031906128, "learning_rate": 0.00019789739465128582, "loss": 4.628, "step": 256500 }, { "epoch": 0.47875082517104073, "grad_norm": 0.926943302154541, "learning_rate": 0.00019789653762645576, "loss": 4.6248, "step": 256550 }, { "epoch": 0.47884413073041926, "grad_norm": 1.0053600072860718, "learning_rate": 0.00019789568042885547, "loss": 4.4205, "step": 256600 }, { "epoch": 0.47893743628979774, "grad_norm": 0.9861886501312256, "learning_rate": 0.00019789482305848655, "loss": 4.7163, "step": 256650 }, { "epoch": 0.4790307418491762, "grad_norm": 1.014960527420044, "learning_rate": 0.0001978939655153505, "loss": 4.679, "step": 256700 }, { "epoch": 0.47912404740855474, "grad_norm": 0.6750307679176331, "learning_rate": 0.0001978931077994488, "loss": 4.5416, "step": 256750 }, { "epoch": 0.4792173529679332, "grad_norm": 0.891221284866333, "learning_rate": 0.000197892249910783, "loss": 4.2551, "step": 256800 }, { "epoch": 0.4793106585273117, "grad_norm": 0.7130653262138367, "learning_rate": 0.00019789139184935462, "loss": 4.6392, "step": 256850 }, { "epoch": 0.47940396408669017, "grad_norm": 0.7216616272926331, "learning_rate": 0.00019789053361516517, "loss": 4.5822, "step": 256900 }, { "epoch": 0.4794972696460687, "grad_norm": 1.1904911994934082, "learning_rate": 0.00019788967520821614, "loss": 4.7572, "step": 256950 }, { "epoch": 0.4795905752054472, "grad_norm": 0.8761834502220154, "learning_rate": 0.00019788881662850902, "loss": 4.6696, "step": 257000 }, { "epoch": 0.47968388076482565, "grad_norm": 0.9479681849479675, "learning_rate": 0.00019788795787604538, "loss": 4.6753, "step": 257050 }, { "epoch": 0.4797771863242042, "grad_norm": 1.1873890161514282, "learning_rate": 0.00019788709895082675, "loss": 4.4175, "step": 257100 }, { "epoch": 0.47987049188358266, "grad_norm": 1.023085594177246, "learning_rate": 0.0001978862398528546, "loss": 4.5694, "step": 257150 }, { "epoch": 0.47996379744296114, "grad_norm": 0.9726206064224243, "learning_rate": 0.00019788538058213045, "loss": 4.5753, "step": 257200 }, { "epoch": 0.4800571030023396, "grad_norm": 0.7751583456993103, "learning_rate": 0.00019788452113865583, "loss": 4.4491, "step": 257250 }, { "epoch": 0.48015040856171814, "grad_norm": 1.2681344747543335, "learning_rate": 0.00019788366152243225, "loss": 4.5964, "step": 257300 }, { "epoch": 0.4802437141210966, "grad_norm": 1.1450700759887695, "learning_rate": 0.00019788280173346125, "loss": 4.6184, "step": 257350 }, { "epoch": 0.4803370196804751, "grad_norm": 0.9162212610244751, "learning_rate": 0.00019788194177174431, "loss": 4.4201, "step": 257400 }, { "epoch": 0.48043032523985363, "grad_norm": 1.0206190347671509, "learning_rate": 0.00019788108163728296, "loss": 4.6341, "step": 257450 }, { "epoch": 0.4805236307992321, "grad_norm": 0.9956310391426086, "learning_rate": 0.00019788022133007872, "loss": 4.5461, "step": 257500 }, { "epoch": 0.4806169363586106, "grad_norm": 0.968657374382019, "learning_rate": 0.00019787936085013315, "loss": 4.3365, "step": 257550 }, { "epoch": 0.48071024191798906, "grad_norm": 0.9663876891136169, "learning_rate": 0.0001978785001974477, "loss": 4.6494, "step": 257600 }, { "epoch": 0.4808035474773676, "grad_norm": 0.8891162276268005, "learning_rate": 0.00019787763937202392, "loss": 4.6504, "step": 257650 }, { "epoch": 0.48089685303674606, "grad_norm": 1.1420527696609497, "learning_rate": 0.00019787677837386334, "loss": 4.5885, "step": 257700 }, { "epoch": 0.48099015859612454, "grad_norm": 1.198561429977417, "learning_rate": 0.00019787591720296747, "loss": 4.6936, "step": 257750 }, { "epoch": 0.48108346415550307, "grad_norm": 0.9371161460876465, "learning_rate": 0.0001978750558593378, "loss": 4.5657, "step": 257800 }, { "epoch": 0.48117676971488155, "grad_norm": 0.8604893684387207, "learning_rate": 0.0001978741943429759, "loss": 4.4049, "step": 257850 }, { "epoch": 0.48127007527426, "grad_norm": 0.8791077733039856, "learning_rate": 0.00019787333265388325, "loss": 4.3852, "step": 257900 }, { "epoch": 0.4813633808336385, "grad_norm": 1.054795265197754, "learning_rate": 0.00019787247079206143, "loss": 4.6211, "step": 257950 }, { "epoch": 0.48145668639301703, "grad_norm": 1.1401352882385254, "learning_rate": 0.0001978716087575119, "loss": 4.6098, "step": 258000 }, { "epoch": 0.48145668639301703, "eval_loss": 4.777442932128906, "eval_runtime": 230.5421, "eval_samples_per_second": 11.312, "eval_steps_per_second": 11.312, "eval_tts_loss": 7.488510146787093, "step": 258000 }, { "epoch": 0.4815499919523955, "grad_norm": 1.9049067497253418, "learning_rate": 0.00019787074655023618, "loss": 4.5569, "step": 258050 }, { "epoch": 0.481643297511774, "grad_norm": 1.1054402589797974, "learning_rate": 0.00019786988417023585, "loss": 4.6882, "step": 258100 }, { "epoch": 0.4817366030711525, "grad_norm": 0.8335074186325073, "learning_rate": 0.00019786902161751234, "loss": 4.6304, "step": 258150 }, { "epoch": 0.481829908630531, "grad_norm": 0.9190850257873535, "learning_rate": 0.00019786815889206728, "loss": 4.5275, "step": 258200 }, { "epoch": 0.48192321418990947, "grad_norm": 0.9075625538825989, "learning_rate": 0.0001978672959939021, "loss": 4.5176, "step": 258250 }, { "epoch": 0.48201651974928794, "grad_norm": 0.9331411719322205, "learning_rate": 0.00019786643292301842, "loss": 4.5701, "step": 258300 }, { "epoch": 0.4821098253086665, "grad_norm": 0.732326865196228, "learning_rate": 0.00019786556967941768, "loss": 4.6525, "step": 258350 }, { "epoch": 0.48220313086804495, "grad_norm": 1.284225583076477, "learning_rate": 0.0001978647062631014, "loss": 4.8359, "step": 258400 }, { "epoch": 0.4822964364274234, "grad_norm": 1.2217979431152344, "learning_rate": 0.00019786384267407117, "loss": 4.6221, "step": 258450 }, { "epoch": 0.4823897419868019, "grad_norm": 0.6436966061592102, "learning_rate": 0.00019786297891232846, "loss": 4.5166, "step": 258500 }, { "epoch": 0.48248304754618043, "grad_norm": 0.9446617364883423, "learning_rate": 0.0001978621149778748, "loss": 4.689, "step": 258550 }, { "epoch": 0.4825763531055589, "grad_norm": 0.9324028491973877, "learning_rate": 0.00019786125087071173, "loss": 4.5694, "step": 258600 }, { "epoch": 0.4826696586649374, "grad_norm": 0.9746628999710083, "learning_rate": 0.00019786038659084078, "loss": 4.6517, "step": 258650 }, { "epoch": 0.4827629642243159, "grad_norm": 0.9371702671051025, "learning_rate": 0.0001978595221382635, "loss": 4.6591, "step": 258700 }, { "epoch": 0.4828562697836944, "grad_norm": 0.8363770246505737, "learning_rate": 0.00019785865751298135, "loss": 4.4523, "step": 258750 }, { "epoch": 0.48294957534307287, "grad_norm": 0.8663115501403809, "learning_rate": 0.00019785779271499587, "loss": 4.5986, "step": 258800 }, { "epoch": 0.48304288090245134, "grad_norm": 0.9521824717521667, "learning_rate": 0.00019785692774430862, "loss": 4.6115, "step": 258850 }, { "epoch": 0.4831361864618299, "grad_norm": 1.0983976125717163, "learning_rate": 0.0001978560626009211, "loss": 4.5424, "step": 258900 }, { "epoch": 0.48322949202120835, "grad_norm": 0.9267615675926208, "learning_rate": 0.0001978551972848349, "loss": 4.5658, "step": 258950 }, { "epoch": 0.4833227975805868, "grad_norm": 1.1117475032806396, "learning_rate": 0.00019785433179605143, "loss": 4.6279, "step": 259000 }, { "epoch": 0.48341610313996536, "grad_norm": 0.9783563613891602, "learning_rate": 0.00019785346613457235, "loss": 4.5828, "step": 259050 }, { "epoch": 0.48350940869934383, "grad_norm": 1.0470584630966187, "learning_rate": 0.00019785260030039906, "loss": 4.5992, "step": 259100 }, { "epoch": 0.4836027142587223, "grad_norm": 1.1836624145507812, "learning_rate": 0.00019785173429353318, "loss": 4.6148, "step": 259150 }, { "epoch": 0.4836960198181008, "grad_norm": 1.0793054103851318, "learning_rate": 0.00019785086811397617, "loss": 4.4822, "step": 259200 }, { "epoch": 0.4837893253774793, "grad_norm": 0.7679390907287598, "learning_rate": 0.0001978500017617296, "loss": 4.6483, "step": 259250 }, { "epoch": 0.4838826309368578, "grad_norm": 0.987201988697052, "learning_rate": 0.00019784913523679504, "loss": 4.5346, "step": 259300 }, { "epoch": 0.48397593649623627, "grad_norm": 1.2286121845245361, "learning_rate": 0.00019784826853917395, "loss": 4.751, "step": 259350 }, { "epoch": 0.4840692420556148, "grad_norm": 1.232087254524231, "learning_rate": 0.0001978474016688679, "loss": 4.6215, "step": 259400 }, { "epoch": 0.4841625476149933, "grad_norm": 0.9811508059501648, "learning_rate": 0.00019784653462587835, "loss": 4.3616, "step": 259450 }, { "epoch": 0.48425585317437175, "grad_norm": 1.0040799379348755, "learning_rate": 0.00019784566741020695, "loss": 4.4287, "step": 259500 }, { "epoch": 0.48434915873375023, "grad_norm": 0.8686191439628601, "learning_rate": 0.0001978448000218551, "loss": 4.468, "step": 259550 }, { "epoch": 0.48444246429312876, "grad_norm": 0.8420849442481995, "learning_rate": 0.0001978439324608244, "loss": 4.6569, "step": 259600 }, { "epoch": 0.48453576985250724, "grad_norm": 0.7843796610832214, "learning_rate": 0.00019784306472711642, "loss": 4.4424, "step": 259650 }, { "epoch": 0.4846290754118857, "grad_norm": 1.2961535453796387, "learning_rate": 0.00019784219682073264, "loss": 4.6343, "step": 259700 }, { "epoch": 0.48472238097126424, "grad_norm": 1.09566330909729, "learning_rate": 0.0001978413287416746, "loss": 4.4992, "step": 259750 }, { "epoch": 0.4848156865306427, "grad_norm": 0.9257076978683472, "learning_rate": 0.0001978404604899438, "loss": 4.6446, "step": 259800 }, { "epoch": 0.4849089920900212, "grad_norm": 0.9009526371955872, "learning_rate": 0.0001978395920655418, "loss": 4.4916, "step": 259850 }, { "epoch": 0.48500229764939967, "grad_norm": 1.0999066829681396, "learning_rate": 0.00019783872346847018, "loss": 4.4681, "step": 259900 }, { "epoch": 0.4850956032087782, "grad_norm": 0.8918809294700623, "learning_rate": 0.00019783785469873038, "loss": 4.5746, "step": 259950 }, { "epoch": 0.4851889087681567, "grad_norm": 1.1592248678207397, "learning_rate": 0.000197836985756324, "loss": 4.6462, "step": 260000 }, { "epoch": 0.48528221432753516, "grad_norm": 1.0789110660552979, "learning_rate": 0.00019783611664125257, "loss": 4.6372, "step": 260050 }, { "epoch": 0.4853755198869137, "grad_norm": 0.7961848974227905, "learning_rate": 0.0001978352473535176, "loss": 4.7098, "step": 260100 }, { "epoch": 0.48546882544629216, "grad_norm": 1.1101049184799194, "learning_rate": 0.0001978343778931206, "loss": 4.4104, "step": 260150 }, { "epoch": 0.48556213100567064, "grad_norm": 1.1163572072982788, "learning_rate": 0.00019783350826006317, "loss": 4.6022, "step": 260200 }, { "epoch": 0.4856554365650491, "grad_norm": 0.8605142831802368, "learning_rate": 0.00019783263845434682, "loss": 4.6625, "step": 260250 }, { "epoch": 0.48574874212442765, "grad_norm": 0.9262938499450684, "learning_rate": 0.00019783176847597308, "loss": 4.6614, "step": 260300 }, { "epoch": 0.4858420476838061, "grad_norm": 1.254042387008667, "learning_rate": 0.00019783089832494345, "loss": 4.5324, "step": 260350 }, { "epoch": 0.4859353532431846, "grad_norm": 1.2252981662750244, "learning_rate": 0.00019783002800125954, "loss": 4.541, "step": 260400 }, { "epoch": 0.48602865880256313, "grad_norm": 0.9264373779296875, "learning_rate": 0.00019782915750492282, "loss": 4.6012, "step": 260450 }, { "epoch": 0.4861219643619416, "grad_norm": 1.1347676515579224, "learning_rate": 0.00019782828683593485, "loss": 4.7729, "step": 260500 }, { "epoch": 0.4862152699213201, "grad_norm": 0.581415593624115, "learning_rate": 0.00019782741599429717, "loss": 4.5686, "step": 260550 }, { "epoch": 0.48630857548069856, "grad_norm": 1.1789880990982056, "learning_rate": 0.00019782654498001133, "loss": 4.5014, "step": 260600 }, { "epoch": 0.4864018810400771, "grad_norm": 1.1918727159500122, "learning_rate": 0.0001978256737930788, "loss": 4.5867, "step": 260650 }, { "epoch": 0.48649518659945556, "grad_norm": 1.2388694286346436, "learning_rate": 0.00019782480243350122, "loss": 4.4532, "step": 260700 }, { "epoch": 0.48658849215883404, "grad_norm": 0.9649792313575745, "learning_rate": 0.00019782393090128007, "loss": 4.6906, "step": 260750 }, { "epoch": 0.48668179771821257, "grad_norm": 0.8217645883560181, "learning_rate": 0.00019782305919641689, "loss": 4.4997, "step": 260800 }, { "epoch": 0.48677510327759105, "grad_norm": 1.2379895448684692, "learning_rate": 0.0001978221873189132, "loss": 4.7031, "step": 260850 }, { "epoch": 0.4868684088369695, "grad_norm": 0.9519298672676086, "learning_rate": 0.00019782131526877058, "loss": 4.7055, "step": 260900 }, { "epoch": 0.486961714396348, "grad_norm": 0.8161171674728394, "learning_rate": 0.00019782044304599055, "loss": 4.826, "step": 260950 }, { "epoch": 0.48705501995572653, "grad_norm": 1.0965524911880493, "learning_rate": 0.00019781957065057463, "loss": 4.6886, "step": 261000 }, { "epoch": 0.48705501995572653, "eval_loss": 4.7727813720703125, "eval_runtime": 228.7535, "eval_samples_per_second": 11.401, "eval_steps_per_second": 11.401, "eval_tts_loss": 7.503641583035455, "step": 261000 }, { "epoch": 0.487148325515105, "grad_norm": 1.025140643119812, "learning_rate": 0.00019781869808252443, "loss": 4.613, "step": 261050 }, { "epoch": 0.4872416310744835, "grad_norm": 1.0564911365509033, "learning_rate": 0.0001978178253418414, "loss": 4.5312, "step": 261100 }, { "epoch": 0.487334936633862, "grad_norm": 0.8013015985488892, "learning_rate": 0.00019781695242852712, "loss": 4.5406, "step": 261150 }, { "epoch": 0.4874282421932405, "grad_norm": 0.806037425994873, "learning_rate": 0.00019781607934258314, "loss": 4.6685, "step": 261200 }, { "epoch": 0.48752154775261897, "grad_norm": 1.0473806858062744, "learning_rate": 0.000197815206084011, "loss": 4.5322, "step": 261250 }, { "epoch": 0.48761485331199744, "grad_norm": 1.0605626106262207, "learning_rate": 0.0001978143326528122, "loss": 4.4764, "step": 261300 }, { "epoch": 0.487708158871376, "grad_norm": 0.7920166254043579, "learning_rate": 0.00019781345904898835, "loss": 4.6327, "step": 261350 }, { "epoch": 0.48780146443075445, "grad_norm": 0.9014812111854553, "learning_rate": 0.00019781258527254097, "loss": 4.6423, "step": 261400 }, { "epoch": 0.4878947699901329, "grad_norm": 0.8125473856925964, "learning_rate": 0.00019781171132347155, "loss": 4.6761, "step": 261450 }, { "epoch": 0.4879880755495114, "grad_norm": 1.018195390701294, "learning_rate": 0.00019781083720178167, "loss": 4.6347, "step": 261500 }, { "epoch": 0.48808138110888993, "grad_norm": 1.0549075603485107, "learning_rate": 0.0001978099629074729, "loss": 4.633, "step": 261550 }, { "epoch": 0.4881746866682684, "grad_norm": 1.154897689819336, "learning_rate": 0.00019780908844054673, "loss": 4.6492, "step": 261600 }, { "epoch": 0.4882679922276469, "grad_norm": 0.9253442883491516, "learning_rate": 0.00019780821380100473, "loss": 4.6132, "step": 261650 }, { "epoch": 0.4883612977870254, "grad_norm": 1.029456615447998, "learning_rate": 0.00019780733898884848, "loss": 4.4266, "step": 261700 }, { "epoch": 0.4884546033464039, "grad_norm": 1.1559395790100098, "learning_rate": 0.00019780646400407946, "loss": 4.4551, "step": 261750 }, { "epoch": 0.48854790890578237, "grad_norm": 1.0872691869735718, "learning_rate": 0.00019780558884669922, "loss": 4.4269, "step": 261800 }, { "epoch": 0.48864121446516084, "grad_norm": 0.7114251852035522, "learning_rate": 0.00019780471351670933, "loss": 4.405, "step": 261850 }, { "epoch": 0.4887345200245394, "grad_norm": 0.846502423286438, "learning_rate": 0.00019780383801411134, "loss": 4.4984, "step": 261900 }, { "epoch": 0.48882782558391785, "grad_norm": 1.0822960138320923, "learning_rate": 0.0001978029623389068, "loss": 4.5524, "step": 261950 }, { "epoch": 0.48892113114329633, "grad_norm": 1.0587904453277588, "learning_rate": 0.0001978020864910972, "loss": 4.5385, "step": 262000 }, { "epoch": 0.48901443670267486, "grad_norm": 0.9366457462310791, "learning_rate": 0.00019780121047068417, "loss": 4.7893, "step": 262050 }, { "epoch": 0.48910774226205334, "grad_norm": 1.1019724607467651, "learning_rate": 0.00019780033427766922, "loss": 4.5838, "step": 262100 }, { "epoch": 0.4892010478214318, "grad_norm": 1.3364999294281006, "learning_rate": 0.00019779945791205386, "loss": 4.5512, "step": 262150 }, { "epoch": 0.4892943533808103, "grad_norm": 0.7997466921806335, "learning_rate": 0.00019779858137383967, "loss": 4.4808, "step": 262200 }, { "epoch": 0.4893876589401888, "grad_norm": 1.033316731452942, "learning_rate": 0.0001977977046630282, "loss": 4.5601, "step": 262250 }, { "epoch": 0.4894809644995673, "grad_norm": 1.1399928331375122, "learning_rate": 0.00019779682777962097, "loss": 4.5472, "step": 262300 }, { "epoch": 0.48957427005894577, "grad_norm": 0.9686580896377563, "learning_rate": 0.00019779595072361955, "loss": 4.8388, "step": 262350 }, { "epoch": 0.4896675756183243, "grad_norm": 0.9062378406524658, "learning_rate": 0.00019779507349502548, "loss": 4.5663, "step": 262400 }, { "epoch": 0.4897608811777028, "grad_norm": 1.001134991645813, "learning_rate": 0.00019779419609384035, "loss": 4.6099, "step": 262450 }, { "epoch": 0.48985418673708125, "grad_norm": 1.1371427774429321, "learning_rate": 0.00019779331852006564, "loss": 4.5837, "step": 262500 }, { "epoch": 0.48994749229645973, "grad_norm": 0.9281883835792542, "learning_rate": 0.00019779244077370293, "loss": 4.7212, "step": 262550 }, { "epoch": 0.49004079785583826, "grad_norm": 1.3296451568603516, "learning_rate": 0.0001977915628547538, "loss": 4.8147, "step": 262600 }, { "epoch": 0.49013410341521674, "grad_norm": 1.0357704162597656, "learning_rate": 0.00019779068476321975, "loss": 4.7545, "step": 262650 }, { "epoch": 0.4902274089745952, "grad_norm": 1.1672788858413696, "learning_rate": 0.00019778980649910234, "loss": 4.6424, "step": 262700 }, { "epoch": 0.49032071453397374, "grad_norm": 1.162309169769287, "learning_rate": 0.0001977889280624031, "loss": 4.7544, "step": 262750 }, { "epoch": 0.4904140200933522, "grad_norm": 1.0663206577301025, "learning_rate": 0.00019778804945312365, "loss": 4.3885, "step": 262800 }, { "epoch": 0.4905073256527307, "grad_norm": 0.9657242298126221, "learning_rate": 0.0001977871706712655, "loss": 4.7163, "step": 262850 }, { "epoch": 0.4906006312121092, "grad_norm": 1.080690860748291, "learning_rate": 0.0001977862917168302, "loss": 4.5714, "step": 262900 }, { "epoch": 0.4906939367714877, "grad_norm": 1.0465108156204224, "learning_rate": 0.00019778541258981926, "loss": 4.511, "step": 262950 }, { "epoch": 0.4907872423308662, "grad_norm": 1.0422312021255493, "learning_rate": 0.00019778453329023432, "loss": 4.4401, "step": 263000 }, { "epoch": 0.49088054789024466, "grad_norm": 1.068161129951477, "learning_rate": 0.00019778365381807687, "loss": 4.5966, "step": 263050 }, { "epoch": 0.4909738534496232, "grad_norm": 1.1453481912612915, "learning_rate": 0.00019778277417334845, "loss": 4.5744, "step": 263100 }, { "epoch": 0.49106715900900166, "grad_norm": 0.9773776531219482, "learning_rate": 0.00019778189435605068, "loss": 4.7388, "step": 263150 }, { "epoch": 0.49116046456838014, "grad_norm": 0.8863257169723511, "learning_rate": 0.00019778101436618506, "loss": 4.4572, "step": 263200 }, { "epoch": 0.4912537701277586, "grad_norm": 0.8632976412773132, "learning_rate": 0.00019778013420375314, "loss": 4.7325, "step": 263250 }, { "epoch": 0.49134707568713715, "grad_norm": 1.222376823425293, "learning_rate": 0.00019777925386875648, "loss": 4.4925, "step": 263300 }, { "epoch": 0.4914403812465156, "grad_norm": 1.2068899869918823, "learning_rate": 0.00019777837336119668, "loss": 4.6069, "step": 263350 }, { "epoch": 0.4915336868058941, "grad_norm": 1.1874123811721802, "learning_rate": 0.00019777749268107525, "loss": 4.5623, "step": 263400 }, { "epoch": 0.49162699236527263, "grad_norm": 0.8214643001556396, "learning_rate": 0.00019777661182839373, "loss": 4.5084, "step": 263450 }, { "epoch": 0.4917202979246511, "grad_norm": 1.1660773754119873, "learning_rate": 0.0001977757308031537, "loss": 4.5678, "step": 263500 }, { "epoch": 0.4918136034840296, "grad_norm": 0.8827805519104004, "learning_rate": 0.00019777484960535667, "loss": 4.7326, "step": 263550 }, { "epoch": 0.49190690904340806, "grad_norm": 0.9647439122200012, "learning_rate": 0.0001977739682350043, "loss": 4.7222, "step": 263600 }, { "epoch": 0.4920002146027866, "grad_norm": 0.7739233374595642, "learning_rate": 0.00019777308669209807, "loss": 4.4787, "step": 263650 }, { "epoch": 0.49209352016216507, "grad_norm": 0.7975419759750366, "learning_rate": 0.0001977722049766395, "loss": 4.6783, "step": 263700 }, { "epoch": 0.49218682572154354, "grad_norm": 1.198824167251587, "learning_rate": 0.00019777132308863022, "loss": 4.5501, "step": 263750 }, { "epoch": 0.4922801312809221, "grad_norm": 1.0451983213424683, "learning_rate": 0.00019777044102807175, "loss": 4.63, "step": 263800 }, { "epoch": 0.49237343684030055, "grad_norm": 1.064078450202942, "learning_rate": 0.00019776955879496569, "loss": 4.5926, "step": 263850 }, { "epoch": 0.492466742399679, "grad_norm": 0.6753122806549072, "learning_rate": 0.00019776867638931353, "loss": 4.5474, "step": 263900 }, { "epoch": 0.4925600479590575, "grad_norm": 1.0703619718551636, "learning_rate": 0.00019776779381111686, "loss": 4.6989, "step": 263950 }, { "epoch": 0.49265335351843603, "grad_norm": 0.9324034452438354, "learning_rate": 0.00019776691106037722, "loss": 4.6068, "step": 264000 }, { "epoch": 0.49265335351843603, "eval_loss": 4.763203144073486, "eval_runtime": 228.6077, "eval_samples_per_second": 11.408, "eval_steps_per_second": 11.408, "eval_tts_loss": 7.533669790340176, "step": 264000 }, { "epoch": 0.4927466590778145, "grad_norm": 1.1262904405593872, "learning_rate": 0.00019776602813709624, "loss": 4.6063, "step": 264050 }, { "epoch": 0.492839964637193, "grad_norm": 0.7986791133880615, "learning_rate": 0.00019776514504127538, "loss": 4.5345, "step": 264100 }, { "epoch": 0.49293327019657146, "grad_norm": 1.026489496231079, "learning_rate": 0.00019776426177291624, "loss": 4.638, "step": 264150 }, { "epoch": 0.49302657575595, "grad_norm": 1.1013755798339844, "learning_rate": 0.00019776337833202039, "loss": 4.5174, "step": 264200 }, { "epoch": 0.49311988131532847, "grad_norm": 0.9843734502792358, "learning_rate": 0.0001977624947185894, "loss": 4.3527, "step": 264250 }, { "epoch": 0.49321318687470694, "grad_norm": 1.009853720664978, "learning_rate": 0.00019776161093262474, "loss": 4.5395, "step": 264300 }, { "epoch": 0.4933064924340855, "grad_norm": 1.039201021194458, "learning_rate": 0.00019776072697412812, "loss": 4.5555, "step": 264350 }, { "epoch": 0.49339979799346395, "grad_norm": 1.022912859916687, "learning_rate": 0.000197759842843101, "loss": 4.6428, "step": 264400 }, { "epoch": 0.4934931035528424, "grad_norm": 1.0572845935821533, "learning_rate": 0.0001977589585395449, "loss": 4.4889, "step": 264450 }, { "epoch": 0.4935864091122209, "grad_norm": 1.2854913473129272, "learning_rate": 0.0001977580740634615, "loss": 4.695, "step": 264500 }, { "epoch": 0.49367971467159943, "grad_norm": 1.0805784463882446, "learning_rate": 0.00019775718941485228, "loss": 4.6472, "step": 264550 }, { "epoch": 0.4937730202309779, "grad_norm": 1.0864139795303345, "learning_rate": 0.00019775630459371884, "loss": 4.5395, "step": 264600 }, { "epoch": 0.4938663257903564, "grad_norm": 0.8742742538452148, "learning_rate": 0.0001977554196000627, "loss": 4.5496, "step": 264650 }, { "epoch": 0.4939596313497349, "grad_norm": 1.1790704727172852, "learning_rate": 0.00019775453443388546, "loss": 4.5932, "step": 264700 }, { "epoch": 0.4940529369091134, "grad_norm": 0.8600195050239563, "learning_rate": 0.00019775364909518865, "loss": 4.3761, "step": 264750 }, { "epoch": 0.49414624246849187, "grad_norm": 1.3297063112258911, "learning_rate": 0.00019775276358397384, "loss": 4.7526, "step": 264800 }, { "epoch": 0.49423954802787035, "grad_norm": 0.9508358240127563, "learning_rate": 0.00019775187790024263, "loss": 4.4664, "step": 264850 }, { "epoch": 0.4943328535872489, "grad_norm": 1.2345694303512573, "learning_rate": 0.00019775099204399657, "loss": 4.6711, "step": 264900 }, { "epoch": 0.49442615914662735, "grad_norm": 1.387581467628479, "learning_rate": 0.00019775010601523718, "loss": 4.5366, "step": 264950 }, { "epoch": 0.49451946470600583, "grad_norm": 0.9651834964752197, "learning_rate": 0.00019774921981396607, "loss": 4.5564, "step": 265000 }, { "epoch": 0.49461277026538436, "grad_norm": 1.261367917060852, "learning_rate": 0.00019774833344018476, "loss": 4.4319, "step": 265050 }, { "epoch": 0.49470607582476284, "grad_norm": 0.7602758407592773, "learning_rate": 0.00019774744689389487, "loss": 4.737, "step": 265100 }, { "epoch": 0.4947993813841413, "grad_norm": 1.285857081413269, "learning_rate": 0.0001977465601750979, "loss": 4.6235, "step": 265150 }, { "epoch": 0.4948926869435198, "grad_norm": 1.1164087057113647, "learning_rate": 0.00019774567328379548, "loss": 4.4363, "step": 265200 }, { "epoch": 0.4949859925028983, "grad_norm": 1.1866155862808228, "learning_rate": 0.00019774478621998912, "loss": 4.5297, "step": 265250 }, { "epoch": 0.4950792980622768, "grad_norm": 1.0542068481445312, "learning_rate": 0.00019774389898368045, "loss": 4.6326, "step": 265300 }, { "epoch": 0.49517260362165527, "grad_norm": 1.035804033279419, "learning_rate": 0.00019774301157487095, "loss": 4.7021, "step": 265350 }, { "epoch": 0.4952659091810338, "grad_norm": 1.0542982816696167, "learning_rate": 0.00019774212399356226, "loss": 4.7791, "step": 265400 }, { "epoch": 0.4953592147404123, "grad_norm": 1.058598279953003, "learning_rate": 0.00019774123623975592, "loss": 4.4212, "step": 265450 }, { "epoch": 0.49545252029979076, "grad_norm": 1.0170423984527588, "learning_rate": 0.0001977403483134535, "loss": 4.3866, "step": 265500 }, { "epoch": 0.49554582585916923, "grad_norm": 0.9709385633468628, "learning_rate": 0.00019773946021465656, "loss": 4.5444, "step": 265550 }, { "epoch": 0.49563913141854776, "grad_norm": 1.150532603263855, "learning_rate": 0.00019773857194336667, "loss": 4.6281, "step": 265600 }, { "epoch": 0.49573243697792624, "grad_norm": 0.9875307679176331, "learning_rate": 0.0001977376834995854, "loss": 4.6653, "step": 265650 }, { "epoch": 0.4958257425373047, "grad_norm": 1.2764935493469238, "learning_rate": 0.00019773679488331429, "loss": 4.4393, "step": 265700 }, { "epoch": 0.49591904809668325, "grad_norm": 0.9200681447982788, "learning_rate": 0.00019773590609455498, "loss": 4.7367, "step": 265750 }, { "epoch": 0.4960123536560617, "grad_norm": 0.8322024941444397, "learning_rate": 0.00019773501713330895, "loss": 4.5703, "step": 265800 }, { "epoch": 0.4961056592154402, "grad_norm": 0.9457756876945496, "learning_rate": 0.00019773412799957784, "loss": 4.6463, "step": 265850 }, { "epoch": 0.4961989647748187, "grad_norm": 1.083602786064148, "learning_rate": 0.00019773323869336318, "loss": 4.728, "step": 265900 }, { "epoch": 0.4962922703341972, "grad_norm": 0.7881966829299927, "learning_rate": 0.00019773234921466654, "loss": 4.4035, "step": 265950 }, { "epoch": 0.4963855758935757, "grad_norm": 0.9676995873451233, "learning_rate": 0.00019773145956348954, "loss": 4.542, "step": 266000 }, { "epoch": 0.49647888145295416, "grad_norm": 1.1931356191635132, "learning_rate": 0.00019773056973983367, "loss": 4.7262, "step": 266050 }, { "epoch": 0.4965721870123327, "grad_norm": 1.042542815208435, "learning_rate": 0.00019772967974370054, "loss": 4.7454, "step": 266100 }, { "epoch": 0.49666549257171116, "grad_norm": 0.9951144456863403, "learning_rate": 0.00019772878957509177, "loss": 4.4392, "step": 266150 }, { "epoch": 0.49675879813108964, "grad_norm": 1.0997178554534912, "learning_rate": 0.00019772789923400882, "loss": 4.454, "step": 266200 }, { "epoch": 0.4968521036904681, "grad_norm": 1.0528903007507324, "learning_rate": 0.00019772700872045337, "loss": 4.6066, "step": 266250 }, { "epoch": 0.49694540924984665, "grad_norm": 1.052966833114624, "learning_rate": 0.00019772611803442692, "loss": 4.5549, "step": 266300 }, { "epoch": 0.4970387148092251, "grad_norm": 0.7441802620887756, "learning_rate": 0.00019772522717593108, "loss": 4.6553, "step": 266350 }, { "epoch": 0.4971320203686036, "grad_norm": 1.0470805168151855, "learning_rate": 0.0001977243361449674, "loss": 4.5749, "step": 266400 }, { "epoch": 0.49722532592798213, "grad_norm": 0.9941937327384949, "learning_rate": 0.00019772344494153746, "loss": 4.6922, "step": 266450 }, { "epoch": 0.4973186314873606, "grad_norm": 0.9114199280738831, "learning_rate": 0.00019772255356564282, "loss": 4.532, "step": 266500 }, { "epoch": 0.4974119370467391, "grad_norm": 1.1066757440567017, "learning_rate": 0.00019772166201728508, "loss": 4.7109, "step": 266550 }, { "epoch": 0.49750524260611756, "grad_norm": 0.7902519106864929, "learning_rate": 0.0001977207702964658, "loss": 4.7002, "step": 266600 }, { "epoch": 0.4975985481654961, "grad_norm": 0.84111088514328, "learning_rate": 0.00019771987840318657, "loss": 4.6185, "step": 266650 }, { "epoch": 0.49769185372487457, "grad_norm": 0.8162730932235718, "learning_rate": 0.00019771898633744894, "loss": 4.6529, "step": 266700 }, { "epoch": 0.49778515928425304, "grad_norm": 1.0409742593765259, "learning_rate": 0.0001977180940992545, "loss": 4.5277, "step": 266750 }, { "epoch": 0.4978784648436315, "grad_norm": 0.9645498394966125, "learning_rate": 0.00019771720168860481, "loss": 4.6303, "step": 266800 }, { "epoch": 0.49797177040301005, "grad_norm": 1.0645626783370972, "learning_rate": 0.00019771630910550146, "loss": 4.7958, "step": 266850 }, { "epoch": 0.4980650759623885, "grad_norm": 1.721364974975586, "learning_rate": 0.000197715416349946, "loss": 4.5781, "step": 266900 }, { "epoch": 0.498158381521767, "grad_norm": 0.9426038265228271, "learning_rate": 0.00019771452342194005, "loss": 4.5271, "step": 266950 }, { "epoch": 0.49825168708114553, "grad_norm": 1.0938518047332764, "learning_rate": 0.00019771363032148516, "loss": 4.5928, "step": 267000 }, { "epoch": 0.49825168708114553, "eval_loss": 4.766524314880371, "eval_runtime": 229.0764, "eval_samples_per_second": 11.385, "eval_steps_per_second": 11.385, "eval_tts_loss": 7.488897705084981, "step": 267000 }, { "epoch": 0.498344992640524, "grad_norm": 1.268874168395996, "learning_rate": 0.00019771273704858287, "loss": 4.4298, "step": 267050 }, { "epoch": 0.4984382981999025, "grad_norm": 1.0305707454681396, "learning_rate": 0.0001977118436032348, "loss": 4.657, "step": 267100 }, { "epoch": 0.49853160375928096, "grad_norm": 0.9082435369491577, "learning_rate": 0.00019771094998544255, "loss": 4.6786, "step": 267150 }, { "epoch": 0.4986249093186595, "grad_norm": 0.9867738485336304, "learning_rate": 0.00019771005619520767, "loss": 4.6024, "step": 267200 }, { "epoch": 0.49871821487803797, "grad_norm": 1.0489767789840698, "learning_rate": 0.00019770916223253168, "loss": 4.5919, "step": 267250 }, { "epoch": 0.49881152043741644, "grad_norm": 1.0940841436386108, "learning_rate": 0.00019770826809741628, "loss": 4.6786, "step": 267300 }, { "epoch": 0.498904825996795, "grad_norm": 1.0874834060668945, "learning_rate": 0.00019770737378986292, "loss": 4.8278, "step": 267350 }, { "epoch": 0.49899813155617345, "grad_norm": 1.1011561155319214, "learning_rate": 0.00019770647930987328, "loss": 4.3219, "step": 267400 }, { "epoch": 0.49909143711555193, "grad_norm": 1.0629221200942993, "learning_rate": 0.00019770558465744887, "loss": 4.6627, "step": 267450 }, { "epoch": 0.4991847426749304, "grad_norm": 0.8989550471305847, "learning_rate": 0.0001977046898325913, "loss": 4.7171, "step": 267500 }, { "epoch": 0.49927804823430894, "grad_norm": 1.0148009061813354, "learning_rate": 0.00019770379483530215, "loss": 4.6872, "step": 267550 }, { "epoch": 0.4993713537936874, "grad_norm": 1.0268181562423706, "learning_rate": 0.00019770289966558298, "loss": 4.5703, "step": 267600 }, { "epoch": 0.4994646593530659, "grad_norm": 0.9970580339431763, "learning_rate": 0.0001977020043234354, "loss": 4.4933, "step": 267650 }, { "epoch": 0.4995579649124444, "grad_norm": 1.5218355655670166, "learning_rate": 0.00019770110880886095, "loss": 4.6482, "step": 267700 }, { "epoch": 0.4996512704718229, "grad_norm": 0.9240729212760925, "learning_rate": 0.00019770021312186126, "loss": 4.5817, "step": 267750 }, { "epoch": 0.49974457603120137, "grad_norm": 1.1267787218093872, "learning_rate": 0.00019769931726243786, "loss": 4.6775, "step": 267800 }, { "epoch": 0.49983788159057985, "grad_norm": 0.8717399835586548, "learning_rate": 0.0001976984212305924, "loss": 4.7974, "step": 267850 }, { "epoch": 0.4999311871499584, "grad_norm": 0.7997406125068665, "learning_rate": 0.00019769752502632638, "loss": 4.7391, "step": 267900 }, { "epoch": 0.5000244927093368, "grad_norm": 1.135779619216919, "learning_rate": 0.00019769662864964143, "loss": 4.5182, "step": 267950 }, { "epoch": 0.5001177982687154, "grad_norm": 1.0822643041610718, "learning_rate": 0.0001976957321005391, "loss": 4.5152, "step": 268000 }, { "epoch": 0.5002111038280939, "grad_norm": 0.8304023146629333, "learning_rate": 0.00019769483537902103, "loss": 4.5748, "step": 268050 }, { "epoch": 0.5003044093874723, "grad_norm": 0.9894828200340271, "learning_rate": 0.00019769393848508875, "loss": 4.5601, "step": 268100 }, { "epoch": 0.5003977149468508, "grad_norm": 1.1318585872650146, "learning_rate": 0.00019769304141874385, "loss": 4.4261, "step": 268150 }, { "epoch": 0.5004910205062293, "grad_norm": 1.3329347372055054, "learning_rate": 0.00019769214417998793, "loss": 4.7159, "step": 268200 }, { "epoch": 0.5005843260656078, "grad_norm": 1.1386692523956299, "learning_rate": 0.00019769124676882255, "loss": 4.374, "step": 268250 }, { "epoch": 0.5006776316249862, "grad_norm": 1.0569969415664673, "learning_rate": 0.0001976903491852493, "loss": 4.7144, "step": 268300 }, { "epoch": 0.5007709371843648, "grad_norm": 0.9701844453811646, "learning_rate": 0.0001976894514292698, "loss": 4.4343, "step": 268350 }, { "epoch": 0.5008642427437433, "grad_norm": 0.9479658007621765, "learning_rate": 0.00019768855350088562, "loss": 4.3689, "step": 268400 }, { "epoch": 0.5009575483031218, "grad_norm": 1.123121976852417, "learning_rate": 0.0001976876554000983, "loss": 4.3821, "step": 268450 }, { "epoch": 0.5010508538625003, "grad_norm": 1.1315399408340454, "learning_rate": 0.00019768675712690948, "loss": 4.7261, "step": 268500 }, { "epoch": 0.5011441594218787, "grad_norm": 0.9239130616188049, "learning_rate": 0.0001976858586813207, "loss": 4.7715, "step": 268550 }, { "epoch": 0.5012374649812572, "grad_norm": 1.1073530912399292, "learning_rate": 0.00019768496006333357, "loss": 4.5403, "step": 268600 }, { "epoch": 0.5013307705406357, "grad_norm": 0.8722808957099915, "learning_rate": 0.00019768406127294969, "loss": 4.6618, "step": 268650 }, { "epoch": 0.5014240761000143, "grad_norm": 1.0455543994903564, "learning_rate": 0.0001976831623101706, "loss": 4.6246, "step": 268700 }, { "epoch": 0.5015173816593927, "grad_norm": 1.1023845672607422, "learning_rate": 0.00019768226317499796, "loss": 4.6571, "step": 268750 }, { "epoch": 0.5016106872187712, "grad_norm": 1.113635778427124, "learning_rate": 0.0001976813638674333, "loss": 4.6877, "step": 268800 }, { "epoch": 0.5017039927781497, "grad_norm": 1.1741982698440552, "learning_rate": 0.0001976804643874782, "loss": 4.5927, "step": 268850 }, { "epoch": 0.5017972983375282, "grad_norm": 1.0700513124465942, "learning_rate": 0.00019767956473513428, "loss": 4.5712, "step": 268900 }, { "epoch": 0.5018906038969067, "grad_norm": 0.9588819146156311, "learning_rate": 0.00019767866491040312, "loss": 4.6768, "step": 268950 }, { "epoch": 0.5019839094562851, "grad_norm": 1.2364482879638672, "learning_rate": 0.0001976777649132863, "loss": 4.643, "step": 269000 }, { "epoch": 0.5020772150156637, "grad_norm": 1.096693754196167, "learning_rate": 0.0001976768647437854, "loss": 4.4986, "step": 269050 }, { "epoch": 0.5021705205750422, "grad_norm": 0.6519501209259033, "learning_rate": 0.00019767596440190203, "loss": 4.5439, "step": 269100 }, { "epoch": 0.5022638261344207, "grad_norm": 1.2315655946731567, "learning_rate": 0.00019767506388763777, "loss": 4.7023, "step": 269150 }, { "epoch": 0.5023571316937991, "grad_norm": 1.1877769231796265, "learning_rate": 0.00019767416320099423, "loss": 4.4986, "step": 269200 }, { "epoch": 0.5024504372531776, "grad_norm": 1.1195615530014038, "learning_rate": 0.00019767326234197295, "loss": 4.6128, "step": 269250 }, { "epoch": 0.5025437428125561, "grad_norm": 0.8910592794418335, "learning_rate": 0.00019767236131057553, "loss": 4.4873, "step": 269300 }, { "epoch": 0.5026370483719346, "grad_norm": 0.8761081099510193, "learning_rate": 0.0001976714601068036, "loss": 4.8037, "step": 269350 }, { "epoch": 0.5027303539313132, "grad_norm": 1.1296546459197998, "learning_rate": 0.00019767055873065877, "loss": 4.6226, "step": 269400 }, { "epoch": 0.5028236594906916, "grad_norm": 0.8964772820472717, "learning_rate": 0.00019766965718214253, "loss": 4.4358, "step": 269450 }, { "epoch": 0.5029169650500701, "grad_norm": 1.181470274925232, "learning_rate": 0.00019766875546125656, "loss": 4.7548, "step": 269500 }, { "epoch": 0.5030102706094486, "grad_norm": 1.027578353881836, "learning_rate": 0.00019766785356800242, "loss": 4.6212, "step": 269550 }, { "epoch": 0.5031035761688271, "grad_norm": 0.7062067985534668, "learning_rate": 0.00019766695150238167, "loss": 4.5479, "step": 269600 }, { "epoch": 0.5031968817282055, "grad_norm": 0.781272828578949, "learning_rate": 0.000197666049264396, "loss": 4.4475, "step": 269650 }, { "epoch": 0.503290187287584, "grad_norm": 0.8852269053459167, "learning_rate": 0.00019766514685404688, "loss": 4.7096, "step": 269700 }, { "epoch": 0.5033834928469626, "grad_norm": 0.8530430197715759, "learning_rate": 0.000197664244271336, "loss": 4.3999, "step": 269750 }, { "epoch": 0.5034767984063411, "grad_norm": 0.670219898223877, "learning_rate": 0.0001976633415162649, "loss": 4.7789, "step": 269800 }, { "epoch": 0.5035701039657196, "grad_norm": 0.8426247239112854, "learning_rate": 0.00019766243858883518, "loss": 4.5608, "step": 269850 }, { "epoch": 0.503663409525098, "grad_norm": 1.1550434827804565, "learning_rate": 0.00019766153548904844, "loss": 4.605, "step": 269900 }, { "epoch": 0.5037567150844765, "grad_norm": 0.9927839636802673, "learning_rate": 0.00019766063221690627, "loss": 4.41, "step": 269950 }, { "epoch": 0.503850020643855, "grad_norm": 0.9224207401275635, "learning_rate": 0.00019765972877241027, "loss": 4.5662, "step": 270000 }, { "epoch": 0.503850020643855, "eval_loss": 4.762325286865234, "eval_runtime": 230.697, "eval_samples_per_second": 11.305, "eval_steps_per_second": 11.305, "eval_tts_loss": 7.51998570571265, "step": 270000 }, { "epoch": 0.5039433262032335, "grad_norm": 1.099448561668396, "learning_rate": 0.00019765882515556204, "loss": 4.7066, "step": 270050 }, { "epoch": 0.504036631762612, "grad_norm": 0.9349024891853333, "learning_rate": 0.00019765792136636315, "loss": 4.5372, "step": 270100 }, { "epoch": 0.5041299373219905, "grad_norm": 0.8908520936965942, "learning_rate": 0.00019765701740481523, "loss": 4.4929, "step": 270150 }, { "epoch": 0.504223242881369, "grad_norm": 0.9843934774398804, "learning_rate": 0.00019765611327091987, "loss": 4.6941, "step": 270200 }, { "epoch": 0.5043165484407475, "grad_norm": 1.1821966171264648, "learning_rate": 0.00019765520896467863, "loss": 4.3483, "step": 270250 }, { "epoch": 0.504409854000126, "grad_norm": 1.0089640617370605, "learning_rate": 0.00019765430448609315, "loss": 4.6964, "step": 270300 }, { "epoch": 0.5045031595595044, "grad_norm": 0.9127652645111084, "learning_rate": 0.00019765339983516498, "loss": 4.4906, "step": 270350 }, { "epoch": 0.5045964651188829, "grad_norm": 1.262275218963623, "learning_rate": 0.00019765249501189574, "loss": 4.8157, "step": 270400 }, { "epoch": 0.5046897706782615, "grad_norm": 0.7381452322006226, "learning_rate": 0.00019765159001628707, "loss": 4.4541, "step": 270450 }, { "epoch": 0.50478307623764, "grad_norm": 1.1626837253570557, "learning_rate": 0.00019765068484834048, "loss": 4.405, "step": 270500 }, { "epoch": 0.5048763817970184, "grad_norm": 1.3286311626434326, "learning_rate": 0.00019764977950805762, "loss": 4.6207, "step": 270550 }, { "epoch": 0.5049696873563969, "grad_norm": 0.9503090977668762, "learning_rate": 0.00019764887399544012, "loss": 4.4284, "step": 270600 }, { "epoch": 0.5050629929157754, "grad_norm": 1.025394082069397, "learning_rate": 0.00019764796831048948, "loss": 4.5853, "step": 270650 }, { "epoch": 0.5051562984751539, "grad_norm": 1.0864533185958862, "learning_rate": 0.00019764706245320741, "loss": 4.6884, "step": 270700 }, { "epoch": 0.5052496040345323, "grad_norm": 1.156358242034912, "learning_rate": 0.00019764615642359543, "loss": 4.5813, "step": 270750 }, { "epoch": 0.5053429095939109, "grad_norm": 1.1821043491363525, "learning_rate": 0.00019764525022165517, "loss": 4.6636, "step": 270800 }, { "epoch": 0.5054362151532894, "grad_norm": 1.121275782585144, "learning_rate": 0.00019764434384738825, "loss": 4.669, "step": 270850 }, { "epoch": 0.5055295207126679, "grad_norm": 0.9881389141082764, "learning_rate": 0.0001976434373007962, "loss": 4.6018, "step": 270900 }, { "epoch": 0.5056228262720464, "grad_norm": 1.1502811908721924, "learning_rate": 0.0001976425305818807, "loss": 4.6788, "step": 270950 }, { "epoch": 0.5057161318314248, "grad_norm": 0.8529852032661438, "learning_rate": 0.00019764162369064327, "loss": 4.6025, "step": 271000 }, { "epoch": 0.5058094373908033, "grad_norm": 0.8883637189865112, "learning_rate": 0.0001976407166270856, "loss": 4.5644, "step": 271050 }, { "epoch": 0.5059027429501818, "grad_norm": 1.2354636192321777, "learning_rate": 0.00019763980939120922, "loss": 4.4785, "step": 271100 }, { "epoch": 0.5059960485095604, "grad_norm": 0.7551540732383728, "learning_rate": 0.0001976389019830158, "loss": 4.4654, "step": 271150 }, { "epoch": 0.5060893540689388, "grad_norm": 1.2407974004745483, "learning_rate": 0.00019763799440250684, "loss": 4.5676, "step": 271200 }, { "epoch": 0.5061826596283173, "grad_norm": 1.2217525243759155, "learning_rate": 0.00019763708664968403, "loss": 4.6668, "step": 271250 }, { "epoch": 0.5062759651876958, "grad_norm": 1.1022509336471558, "learning_rate": 0.00019763617872454892, "loss": 4.5131, "step": 271300 }, { "epoch": 0.5063692707470743, "grad_norm": 0.9949933290481567, "learning_rate": 0.00019763527062710315, "loss": 4.6129, "step": 271350 }, { "epoch": 0.5064625763064527, "grad_norm": 0.9555957913398743, "learning_rate": 0.0001976343623573483, "loss": 4.4298, "step": 271400 }, { "epoch": 0.5065558818658312, "grad_norm": 0.9663435220718384, "learning_rate": 0.00019763345391528598, "loss": 4.5994, "step": 271450 }, { "epoch": 0.5066491874252098, "grad_norm": 1.056382417678833, "learning_rate": 0.0001976325453009178, "loss": 4.5895, "step": 271500 }, { "epoch": 0.5067424929845883, "grad_norm": 1.0316522121429443, "learning_rate": 0.00019763163651424537, "loss": 4.67, "step": 271550 }, { "epoch": 0.5068357985439668, "grad_norm": 1.1362007856369019, "learning_rate": 0.00019763072755527028, "loss": 4.7587, "step": 271600 }, { "epoch": 0.5069291041033452, "grad_norm": 1.2041569948196411, "learning_rate": 0.0001976298184239941, "loss": 4.6065, "step": 271650 }, { "epoch": 0.5070224096627237, "grad_norm": 0.9852795600891113, "learning_rate": 0.0001976289091204185, "loss": 4.4963, "step": 271700 }, { "epoch": 0.5071157152221022, "grad_norm": 0.8907833695411682, "learning_rate": 0.00019762799964454504, "loss": 4.6828, "step": 271750 }, { "epoch": 0.5072090207814807, "grad_norm": 1.104365348815918, "learning_rate": 0.00019762708999637534, "loss": 4.8107, "step": 271800 }, { "epoch": 0.5073023263408593, "grad_norm": 1.104337215423584, "learning_rate": 0.00019762618017591098, "loss": 4.536, "step": 271850 }, { "epoch": 0.5073956319002377, "grad_norm": 1.1071338653564453, "learning_rate": 0.00019762527018315365, "loss": 4.7241, "step": 271900 }, { "epoch": 0.5074889374596162, "grad_norm": 1.119663119316101, "learning_rate": 0.00019762436001810485, "loss": 4.5535, "step": 271950 }, { "epoch": 0.5075822430189947, "grad_norm": 0.6204298734664917, "learning_rate": 0.00019762344968076624, "loss": 4.4117, "step": 272000 }, { "epoch": 0.5076755485783732, "grad_norm": 1.1064519882202148, "learning_rate": 0.00019762253917113943, "loss": 4.7273, "step": 272050 }, { "epoch": 0.5077688541377516, "grad_norm": 0.985309362411499, "learning_rate": 0.000197621628489226, "loss": 4.5156, "step": 272100 }, { "epoch": 0.5078621596971301, "grad_norm": 1.3585158586502075, "learning_rate": 0.0001976207176350276, "loss": 4.5401, "step": 272150 }, { "epoch": 0.5079554652565087, "grad_norm": 1.3111861944198608, "learning_rate": 0.00019761980660854578, "loss": 4.6386, "step": 272200 }, { "epoch": 0.5080487708158872, "grad_norm": 1.2555516958236694, "learning_rate": 0.00019761889540978217, "loss": 4.6786, "step": 272250 }, { "epoch": 0.5081420763752656, "grad_norm": 1.0087488889694214, "learning_rate": 0.00019761798403873844, "loss": 4.5918, "step": 272300 }, { "epoch": 0.5082353819346441, "grad_norm": 1.2685153484344482, "learning_rate": 0.0001976170724954161, "loss": 4.7424, "step": 272350 }, { "epoch": 0.5083286874940226, "grad_norm": 0.9734815955162048, "learning_rate": 0.00019761616077981683, "loss": 4.81, "step": 272400 }, { "epoch": 0.5084219930534011, "grad_norm": 1.008641242980957, "learning_rate": 0.00019761524889194218, "loss": 4.3828, "step": 272450 }, { "epoch": 0.5085152986127796, "grad_norm": 0.7137326598167419, "learning_rate": 0.0001976143368317938, "loss": 4.5799, "step": 272500 }, { "epoch": 0.5086086041721581, "grad_norm": 1.0536925792694092, "learning_rate": 0.00019761342459937331, "loss": 4.4982, "step": 272550 }, { "epoch": 0.5087019097315366, "grad_norm": 1.1726641654968262, "learning_rate": 0.00019761251219468226, "loss": 4.5372, "step": 272600 }, { "epoch": 0.5087952152909151, "grad_norm": 0.9436787962913513, "learning_rate": 0.00019761159961772235, "loss": 4.7957, "step": 272650 }, { "epoch": 0.5088885208502936, "grad_norm": 1.3651800155639648, "learning_rate": 0.00019761068686849513, "loss": 4.5775, "step": 272700 }, { "epoch": 0.508981826409672, "grad_norm": 1.1570991277694702, "learning_rate": 0.0001976097739470022, "loss": 4.6765, "step": 272750 }, { "epoch": 0.5090751319690505, "grad_norm": 1.0292539596557617, "learning_rate": 0.0001976088608532452, "loss": 4.6702, "step": 272800 }, { "epoch": 0.509168437528429, "grad_norm": 1.0812225341796875, "learning_rate": 0.00019760794758722576, "loss": 4.486, "step": 272850 }, { "epoch": 0.5092617430878076, "grad_norm": 0.784271240234375, "learning_rate": 0.00019760703414894545, "loss": 4.4899, "step": 272900 }, { "epoch": 0.5093550486471861, "grad_norm": 1.1616487503051758, "learning_rate": 0.0001976061205384059, "loss": 4.909, "step": 272950 }, { "epoch": 0.5094483542065645, "grad_norm": 0.8870630860328674, "learning_rate": 0.00019760520675560873, "loss": 4.613, "step": 273000 }, { "epoch": 0.5094483542065645, "eval_loss": 4.762674808502197, "eval_runtime": 230.6095, "eval_samples_per_second": 11.309, "eval_steps_per_second": 11.309, "eval_tts_loss": 7.561236815871328, "step": 273000 }, { "epoch": 0.509541659765943, "grad_norm": 1.079444169998169, "learning_rate": 0.00019760429280055552, "loss": 4.5656, "step": 273050 }, { "epoch": 0.5096349653253215, "grad_norm": 1.0470081567764282, "learning_rate": 0.00019760337867324795, "loss": 4.4551, "step": 273100 }, { "epoch": 0.5097282708847, "grad_norm": 1.2054589986801147, "learning_rate": 0.00019760246437368757, "loss": 4.5344, "step": 273150 }, { "epoch": 0.5098215764440784, "grad_norm": 1.3342701196670532, "learning_rate": 0.000197601549901876, "loss": 4.4564, "step": 273200 }, { "epoch": 0.509914882003457, "grad_norm": 0.9556697010993958, "learning_rate": 0.00019760063525781486, "loss": 4.6027, "step": 273250 }, { "epoch": 0.5100081875628355, "grad_norm": 1.4421682357788086, "learning_rate": 0.00019759972044150579, "loss": 4.7149, "step": 273300 }, { "epoch": 0.510101493122214, "grad_norm": 1.053886890411377, "learning_rate": 0.0001975988054529504, "loss": 4.4437, "step": 273350 }, { "epoch": 0.5101947986815925, "grad_norm": 0.9334540963172913, "learning_rate": 0.00019759789029215026, "loss": 4.4206, "step": 273400 }, { "epoch": 0.5102881042409709, "grad_norm": 0.9920104146003723, "learning_rate": 0.00019759697495910702, "loss": 4.4509, "step": 273450 }, { "epoch": 0.5103814098003494, "grad_norm": 1.2183308601379395, "learning_rate": 0.00019759605945382234, "loss": 4.5994, "step": 273500 }, { "epoch": 0.5104747153597279, "grad_norm": 1.0757365226745605, "learning_rate": 0.00019759514377629773, "loss": 4.5886, "step": 273550 }, { "epoch": 0.5105680209191064, "grad_norm": 0.8739063143730164, "learning_rate": 0.0001975942279265349, "loss": 4.6799, "step": 273600 }, { "epoch": 0.510661326478485, "grad_norm": 1.4280765056610107, "learning_rate": 0.0001975933119045354, "loss": 4.6141, "step": 273650 }, { "epoch": 0.5107546320378634, "grad_norm": 0.8664491176605225, "learning_rate": 0.0001975923957103009, "loss": 4.5229, "step": 273700 }, { "epoch": 0.5108479375972419, "grad_norm": 1.3105825185775757, "learning_rate": 0.000197591479343833, "loss": 4.61, "step": 273750 }, { "epoch": 0.5109412431566204, "grad_norm": 1.0648022890090942, "learning_rate": 0.0001975905628051333, "loss": 4.5341, "step": 273800 }, { "epoch": 0.5110345487159988, "grad_norm": 1.1797337532043457, "learning_rate": 0.00019758964609420342, "loss": 4.6844, "step": 273850 }, { "epoch": 0.5111278542753773, "grad_norm": 0.919456958770752, "learning_rate": 0.000197588729211045, "loss": 4.6024, "step": 273900 }, { "epoch": 0.5112211598347558, "grad_norm": 1.0891369581222534, "learning_rate": 0.00019758781215565964, "loss": 4.6621, "step": 273950 }, { "epoch": 0.5113144653941344, "grad_norm": 1.0899498462677002, "learning_rate": 0.00019758689492804896, "loss": 4.6085, "step": 274000 }, { "epoch": 0.5114077709535129, "grad_norm": 0.9327117204666138, "learning_rate": 0.00019758597752821458, "loss": 4.4207, "step": 274050 }, { "epoch": 0.5115010765128913, "grad_norm": 1.0181550979614258, "learning_rate": 0.0001975850599561581, "loss": 4.5633, "step": 274100 }, { "epoch": 0.5115943820722698, "grad_norm": 1.0915321111679077, "learning_rate": 0.00019758414221188118, "loss": 4.7161, "step": 274150 }, { "epoch": 0.5116876876316483, "grad_norm": 1.1579314470291138, "learning_rate": 0.00019758322429538541, "loss": 4.6698, "step": 274200 }, { "epoch": 0.5117809931910268, "grad_norm": 1.0257996320724487, "learning_rate": 0.00019758230620667245, "loss": 4.762, "step": 274250 }, { "epoch": 0.5118742987504052, "grad_norm": 1.128806710243225, "learning_rate": 0.00019758138794574387, "loss": 4.7106, "step": 274300 }, { "epoch": 0.5119676043097838, "grad_norm": 1.203101396560669, "learning_rate": 0.00019758046951260129, "loss": 4.5531, "step": 274350 }, { "epoch": 0.5120609098691623, "grad_norm": 1.0775619745254517, "learning_rate": 0.0001975795509072464, "loss": 4.6696, "step": 274400 }, { "epoch": 0.5121542154285408, "grad_norm": 0.9243199825286865, "learning_rate": 0.00019757863212968072, "loss": 4.3668, "step": 274450 }, { "epoch": 0.5122475209879193, "grad_norm": 0.9755812883377075, "learning_rate": 0.00019757771317990597, "loss": 4.3343, "step": 274500 }, { "epoch": 0.5123408265472977, "grad_norm": 0.8233321309089661, "learning_rate": 0.0001975767940579237, "loss": 4.6018, "step": 274550 }, { "epoch": 0.5124341321066762, "grad_norm": 0.9258056879043579, "learning_rate": 0.00019757587476373556, "loss": 4.528, "step": 274600 }, { "epoch": 0.5125274376660547, "grad_norm": 0.988038957118988, "learning_rate": 0.00019757495529734316, "loss": 4.5291, "step": 274650 }, { "epoch": 0.5126207432254333, "grad_norm": 1.2431509494781494, "learning_rate": 0.00019757403565874815, "loss": 4.6244, "step": 274700 }, { "epoch": 0.5127140487848117, "grad_norm": 0.9292747974395752, "learning_rate": 0.00019757311584795212, "loss": 4.4623, "step": 274750 }, { "epoch": 0.5128073543441902, "grad_norm": 1.3109321594238281, "learning_rate": 0.0001975721958649567, "loss": 4.6401, "step": 274800 }, { "epoch": 0.5129006599035687, "grad_norm": 0.715238094329834, "learning_rate": 0.00019757127570976354, "loss": 4.8008, "step": 274850 }, { "epoch": 0.5129939654629472, "grad_norm": 1.2275428771972656, "learning_rate": 0.00019757035538237427, "loss": 4.5437, "step": 274900 }, { "epoch": 0.5130872710223257, "grad_norm": 0.7300710678100586, "learning_rate": 0.00019756943488279048, "loss": 4.3459, "step": 274950 }, { "epoch": 0.5131805765817041, "grad_norm": 1.039902925491333, "learning_rate": 0.00019756851421101377, "loss": 4.6869, "step": 275000 }, { "epoch": 0.5132738821410827, "grad_norm": 1.2344797849655151, "learning_rate": 0.00019756759336704582, "loss": 4.6589, "step": 275050 }, { "epoch": 0.5133671877004612, "grad_norm": 1.012023687362671, "learning_rate": 0.00019756667235088825, "loss": 4.6262, "step": 275100 }, { "epoch": 0.5134604932598397, "grad_norm": 0.8712413907051086, "learning_rate": 0.00019756575116254265, "loss": 4.4664, "step": 275150 }, { "epoch": 0.5135537988192181, "grad_norm": 1.4855979681015015, "learning_rate": 0.00019756482980201067, "loss": 4.4876, "step": 275200 }, { "epoch": 0.5136471043785966, "grad_norm": 0.8439715504646301, "learning_rate": 0.00019756390826929394, "loss": 4.5116, "step": 275250 }, { "epoch": 0.5137404099379751, "grad_norm": 1.170244574546814, "learning_rate": 0.00019756298656439407, "loss": 4.5469, "step": 275300 }, { "epoch": 0.5138337154973536, "grad_norm": 0.8794098496437073, "learning_rate": 0.00019756206468731272, "loss": 4.6529, "step": 275350 }, { "epoch": 0.5139270210567322, "grad_norm": 0.6970331072807312, "learning_rate": 0.00019756114263805148, "loss": 4.4499, "step": 275400 }, { "epoch": 0.5140203266161106, "grad_norm": 1.5795390605926514, "learning_rate": 0.00019756022041661196, "loss": 4.6471, "step": 275450 }, { "epoch": 0.5141136321754891, "grad_norm": 1.1724786758422852, "learning_rate": 0.00019755929802299583, "loss": 4.5844, "step": 275500 }, { "epoch": 0.5142069377348676, "grad_norm": 1.2464286088943481, "learning_rate": 0.00019755837545720474, "loss": 4.4931, "step": 275550 }, { "epoch": 0.5143002432942461, "grad_norm": 0.9109342694282532, "learning_rate": 0.00019755745271924028, "loss": 4.6377, "step": 275600 }, { "epoch": 0.5143935488536245, "grad_norm": 0.9874877333641052, "learning_rate": 0.00019755652980910405, "loss": 4.8261, "step": 275650 }, { "epoch": 0.514486854413003, "grad_norm": 1.1274828910827637, "learning_rate": 0.0001975556067267977, "loss": 4.6338, "step": 275700 }, { "epoch": 0.5145801599723816, "grad_norm": 1.0080583095550537, "learning_rate": 0.00019755468347232289, "loss": 4.4742, "step": 275750 }, { "epoch": 0.5146734655317601, "grad_norm": 0.6097507476806641, "learning_rate": 0.00019755376004568126, "loss": 4.5828, "step": 275800 }, { "epoch": 0.5147667710911386, "grad_norm": 1.115775465965271, "learning_rate": 0.00019755283644687434, "loss": 4.454, "step": 275850 }, { "epoch": 0.514860076650517, "grad_norm": 1.1590081453323364, "learning_rate": 0.00019755191267590387, "loss": 4.6587, "step": 275900 }, { "epoch": 0.5149533822098955, "grad_norm": 1.1706831455230713, "learning_rate": 0.00019755098873277144, "loss": 4.5509, "step": 275950 }, { "epoch": 0.515046687769274, "grad_norm": 1.1684876680374146, "learning_rate": 0.0001975500646174787, "loss": 4.4862, "step": 276000 }, { "epoch": 0.515046687769274, "eval_loss": 4.763826370239258, "eval_runtime": 231.8604, "eval_samples_per_second": 11.248, "eval_steps_per_second": 11.248, "eval_tts_loss": 7.552985722040529, "step": 276000 }, { "epoch": 0.5151399933286525, "grad_norm": 0.8337443470954895, "learning_rate": 0.00019754914033002724, "loss": 4.4825, "step": 276050 }, { "epoch": 0.515233298888031, "grad_norm": 0.8753382563591003, "learning_rate": 0.00019754821587041868, "loss": 4.5664, "step": 276100 }, { "epoch": 0.5153266044474095, "grad_norm": 0.86767578125, "learning_rate": 0.00019754729123865473, "loss": 4.2846, "step": 276150 }, { "epoch": 0.515419910006788, "grad_norm": 0.8605402708053589, "learning_rate": 0.00019754636643473693, "loss": 4.5609, "step": 276200 }, { "epoch": 0.5155132155661665, "grad_norm": 1.1940838098526, "learning_rate": 0.000197545441458667, "loss": 4.7445, "step": 276250 }, { "epoch": 0.515606521125545, "grad_norm": 0.7858842611312866, "learning_rate": 0.00019754451631044654, "loss": 4.6351, "step": 276300 }, { "epoch": 0.5156998266849234, "grad_norm": 1.128357172012329, "learning_rate": 0.0001975435909900771, "loss": 4.4674, "step": 276350 }, { "epoch": 0.5157931322443019, "grad_norm": 1.2002841234207153, "learning_rate": 0.00019754266549756045, "loss": 4.7851, "step": 276400 }, { "epoch": 0.5158864378036805, "grad_norm": 1.0065298080444336, "learning_rate": 0.00019754173983289812, "loss": 4.557, "step": 276450 }, { "epoch": 0.515979743363059, "grad_norm": 0.8537634015083313, "learning_rate": 0.0001975408139960918, "loss": 4.6682, "step": 276500 }, { "epoch": 0.5160730489224374, "grad_norm": 1.0204187631607056, "learning_rate": 0.00019753988798714312, "loss": 4.7636, "step": 276550 }, { "epoch": 0.5161663544818159, "grad_norm": 1.0077636241912842, "learning_rate": 0.0001975389618060537, "loss": 4.4058, "step": 276600 }, { "epoch": 0.5162596600411944, "grad_norm": 1.3681881427764893, "learning_rate": 0.00019753803545282517, "loss": 4.5586, "step": 276650 }, { "epoch": 0.5163529656005729, "grad_norm": 1.233417272567749, "learning_rate": 0.00019753710892745914, "loss": 4.7313, "step": 276700 }, { "epoch": 0.5164462711599513, "grad_norm": 1.226003885269165, "learning_rate": 0.00019753618222995728, "loss": 4.5046, "step": 276750 }, { "epoch": 0.5165395767193299, "grad_norm": 1.183275580406189, "learning_rate": 0.00019753525536032128, "loss": 4.6232, "step": 276800 }, { "epoch": 0.5166328822787084, "grad_norm": 1.11515212059021, "learning_rate": 0.00019753432831855265, "loss": 4.6647, "step": 276850 }, { "epoch": 0.5167261878380869, "grad_norm": 0.8872556686401367, "learning_rate": 0.0001975334011046531, "loss": 4.5093, "step": 276900 }, { "epoch": 0.5168194933974654, "grad_norm": 0.9870370626449585, "learning_rate": 0.0001975324737186243, "loss": 4.6153, "step": 276950 }, { "epoch": 0.5169127989568438, "grad_norm": 1.020976185798645, "learning_rate": 0.0001975315461604678, "loss": 4.4416, "step": 277000 }, { "epoch": 0.5170061045162223, "grad_norm": 0.7729629278182983, "learning_rate": 0.00019753061843018533, "loss": 4.5748, "step": 277050 }, { "epoch": 0.5170994100756008, "grad_norm": 1.0765986442565918, "learning_rate": 0.00019752969052777843, "loss": 4.6085, "step": 277100 }, { "epoch": 0.5171927156349794, "grad_norm": 1.0300875902175903, "learning_rate": 0.0001975287624532488, "loss": 4.5239, "step": 277150 }, { "epoch": 0.5172860211943578, "grad_norm": 0.9864205718040466, "learning_rate": 0.00019752783420659808, "loss": 4.6854, "step": 277200 }, { "epoch": 0.5173793267537363, "grad_norm": 0.8580088019371033, "learning_rate": 0.0001975269057878279, "loss": 4.55, "step": 277250 }, { "epoch": 0.5174726323131148, "grad_norm": 0.8540656566619873, "learning_rate": 0.00019752597719693987, "loss": 4.488, "step": 277300 }, { "epoch": 0.5175659378724933, "grad_norm": 0.6463267207145691, "learning_rate": 0.00019752504843393567, "loss": 4.5579, "step": 277350 }, { "epoch": 0.5176592434318718, "grad_norm": 1.2098805904388428, "learning_rate": 0.00019752411949881688, "loss": 4.7437, "step": 277400 }, { "epoch": 0.5177525489912502, "grad_norm": 0.8357241749763489, "learning_rate": 0.00019752319039158521, "loss": 4.5648, "step": 277450 }, { "epoch": 0.5178458545506288, "grad_norm": 0.9252105355262756, "learning_rate": 0.00019752226111224226, "loss": 4.6191, "step": 277500 }, { "epoch": 0.5179391601100073, "grad_norm": 0.9856184124946594, "learning_rate": 0.0001975213316607897, "loss": 4.6529, "step": 277550 }, { "epoch": 0.5180324656693858, "grad_norm": 1.0845974683761597, "learning_rate": 0.00019752040203722913, "loss": 4.4621, "step": 277600 }, { "epoch": 0.5181257712287642, "grad_norm": 0.7452815771102905, "learning_rate": 0.0001975194722415622, "loss": 4.5138, "step": 277650 }, { "epoch": 0.5182190767881427, "grad_norm": 1.0747730731964111, "learning_rate": 0.00019751854227379058, "loss": 4.678, "step": 277700 }, { "epoch": 0.5183123823475212, "grad_norm": 1.055133581161499, "learning_rate": 0.0001975176121339159, "loss": 4.653, "step": 277750 }, { "epoch": 0.5184056879068997, "grad_norm": 0.7984835505485535, "learning_rate": 0.0001975166818219398, "loss": 4.4666, "step": 277800 }, { "epoch": 0.5184989934662783, "grad_norm": 1.2567082643508911, "learning_rate": 0.00019751575133786385, "loss": 4.6693, "step": 277850 }, { "epoch": 0.5185922990256567, "grad_norm": 0.9567890167236328, "learning_rate": 0.00019751482068168982, "loss": 4.5737, "step": 277900 }, { "epoch": 0.5186856045850352, "grad_norm": 0.942671537399292, "learning_rate": 0.00019751388985341927, "loss": 4.7707, "step": 277950 }, { "epoch": 0.5187789101444137, "grad_norm": 1.07314932346344, "learning_rate": 0.00019751295885305384, "loss": 4.7499, "step": 278000 }, { "epoch": 0.5188722157037922, "grad_norm": 0.8657567501068115, "learning_rate": 0.00019751202768059526, "loss": 4.5627, "step": 278050 }, { "epoch": 0.5189655212631706, "grad_norm": 0.9359160661697388, "learning_rate": 0.00019751109633604506, "loss": 4.5867, "step": 278100 }, { "epoch": 0.5190588268225491, "grad_norm": 1.0121476650238037, "learning_rate": 0.00019751016481940492, "loss": 4.7104, "step": 278150 }, { "epoch": 0.5191521323819277, "grad_norm": 1.4921369552612305, "learning_rate": 0.00019750923313067653, "loss": 4.5998, "step": 278200 }, { "epoch": 0.5192454379413062, "grad_norm": 1.2651481628417969, "learning_rate": 0.00019750830126986148, "loss": 4.3789, "step": 278250 }, { "epoch": 0.5193387435006847, "grad_norm": 0.9953886866569519, "learning_rate": 0.00019750736923696146, "loss": 4.6677, "step": 278300 }, { "epoch": 0.5194320490600631, "grad_norm": 0.7058857083320618, "learning_rate": 0.00019750643703197806, "loss": 4.5534, "step": 278350 }, { "epoch": 0.5195253546194416, "grad_norm": 1.0814391374588013, "learning_rate": 0.00019750550465491297, "loss": 4.485, "step": 278400 }, { "epoch": 0.5196186601788201, "grad_norm": 0.9021806716918945, "learning_rate": 0.00019750457210576783, "loss": 4.485, "step": 278450 }, { "epoch": 0.5197119657381986, "grad_norm": 0.9229526519775391, "learning_rate": 0.00019750363938454425, "loss": 4.4707, "step": 278500 }, { "epoch": 0.5198052712975771, "grad_norm": 1.1465047597885132, "learning_rate": 0.00019750270649124393, "loss": 4.4243, "step": 278550 }, { "epoch": 0.5198985768569556, "grad_norm": 0.8875910639762878, "learning_rate": 0.00019750177342586848, "loss": 4.5245, "step": 278600 }, { "epoch": 0.5199918824163341, "grad_norm": 1.1508044004440308, "learning_rate": 0.00019750084018841955, "loss": 4.7269, "step": 278650 }, { "epoch": 0.5200851879757126, "grad_norm": 0.9982178211212158, "learning_rate": 0.00019749990677889878, "loss": 4.806, "step": 278700 }, { "epoch": 0.520178493535091, "grad_norm": 0.7332342267036438, "learning_rate": 0.00019749897319730787, "loss": 4.5537, "step": 278750 }, { "epoch": 0.5202717990944695, "grad_norm": 1.0460069179534912, "learning_rate": 0.0001974980394436484, "loss": 4.4889, "step": 278800 }, { "epoch": 0.520365104653848, "grad_norm": 1.1912788152694702, "learning_rate": 0.00019749710551792205, "loss": 4.7645, "step": 278850 }, { "epoch": 0.5204584102132265, "grad_norm": 1.0818390846252441, "learning_rate": 0.00019749617142013047, "loss": 4.7387, "step": 278900 }, { "epoch": 0.5205517157726051, "grad_norm": 1.156233549118042, "learning_rate": 0.00019749523715027528, "loss": 4.6504, "step": 278950 }, { "epoch": 0.5206450213319835, "grad_norm": 0.9997411966323853, "learning_rate": 0.00019749430270835818, "loss": 4.4489, "step": 279000 }, { "epoch": 0.5206450213319835, "eval_loss": 4.75526237487793, "eval_runtime": 229.0585, "eval_samples_per_second": 11.386, "eval_steps_per_second": 11.386, "eval_tts_loss": 7.540098349903738, "step": 279000 }, { "epoch": 0.520738326891362, "grad_norm": 0.8896755576133728, "learning_rate": 0.00019749336809438077, "loss": 4.7217, "step": 279050 }, { "epoch": 0.5208316324507405, "grad_norm": 0.9033337831497192, "learning_rate": 0.00019749243330834473, "loss": 4.6086, "step": 279100 }, { "epoch": 0.520924938010119, "grad_norm": 0.7716773152351379, "learning_rate": 0.0001974914983502517, "loss": 4.7269, "step": 279150 }, { "epoch": 0.5210182435694974, "grad_norm": 1.14053213596344, "learning_rate": 0.00019749056322010332, "loss": 4.6691, "step": 279200 }, { "epoch": 0.5211115491288759, "grad_norm": 1.0877877473831177, "learning_rate": 0.00019748962791790126, "loss": 4.4902, "step": 279250 }, { "epoch": 0.5212048546882545, "grad_norm": 0.9544693231582642, "learning_rate": 0.00019748869244364716, "loss": 4.5944, "step": 279300 }, { "epoch": 0.521298160247633, "grad_norm": 1.047554850578308, "learning_rate": 0.0001974877567973427, "loss": 4.7411, "step": 279350 }, { "epoch": 0.5213914658070115, "grad_norm": 0.8173419237136841, "learning_rate": 0.00019748682097898946, "loss": 4.4421, "step": 279400 }, { "epoch": 0.5214847713663899, "grad_norm": 1.2062064409255981, "learning_rate": 0.00019748588498858912, "loss": 4.6812, "step": 279450 }, { "epoch": 0.5215780769257684, "grad_norm": 1.0467079877853394, "learning_rate": 0.00019748494882614338, "loss": 4.5961, "step": 279500 }, { "epoch": 0.5216713824851469, "grad_norm": 0.8883413076400757, "learning_rate": 0.00019748401249165384, "loss": 4.6766, "step": 279550 }, { "epoch": 0.5217646880445254, "grad_norm": 0.8880034685134888, "learning_rate": 0.0001974830759851222, "loss": 4.2964, "step": 279600 }, { "epoch": 0.521857993603904, "grad_norm": 1.1952457427978516, "learning_rate": 0.00019748213930655006, "loss": 4.663, "step": 279650 }, { "epoch": 0.5219512991632824, "grad_norm": 0.9282484650611877, "learning_rate": 0.00019748120245593911, "loss": 4.6977, "step": 279700 }, { "epoch": 0.5220446047226609, "grad_norm": 1.0370795726776123, "learning_rate": 0.000197480265433291, "loss": 4.5207, "step": 279750 }, { "epoch": 0.5221379102820394, "grad_norm": 1.0368770360946655, "learning_rate": 0.00019747932823860733, "loss": 4.4436, "step": 279800 }, { "epoch": 0.5222312158414179, "grad_norm": 1.122261881828308, "learning_rate": 0.00019747839087188985, "loss": 4.6599, "step": 279850 }, { "epoch": 0.5223245214007963, "grad_norm": 1.1524388790130615, "learning_rate": 0.00019747745333314012, "loss": 4.4843, "step": 279900 }, { "epoch": 0.5224178269601748, "grad_norm": 1.1980907917022705, "learning_rate": 0.00019747651562235984, "loss": 4.6278, "step": 279950 }, { "epoch": 0.5225111325195534, "grad_norm": 1.0899757146835327, "learning_rate": 0.0001974755777395507, "loss": 4.5952, "step": 280000 }, { "epoch": 0.5226044380789319, "grad_norm": 1.0352386236190796, "learning_rate": 0.0001974746396847143, "loss": 4.5709, "step": 280050 }, { "epoch": 0.5226977436383103, "grad_norm": 1.149746298789978, "learning_rate": 0.00019747370145785226, "loss": 4.4802, "step": 280100 }, { "epoch": 0.5227910491976888, "grad_norm": 1.0010826587677002, "learning_rate": 0.00019747276305896632, "loss": 4.7044, "step": 280150 }, { "epoch": 0.5228843547570673, "grad_norm": 1.0683485269546509, "learning_rate": 0.00019747182448805816, "loss": 4.5503, "step": 280200 }, { "epoch": 0.5229776603164458, "grad_norm": 1.128632664680481, "learning_rate": 0.00019747088574512933, "loss": 4.5131, "step": 280250 }, { "epoch": 0.5230709658758242, "grad_norm": 1.312676191329956, "learning_rate": 0.00019746994683018154, "loss": 4.6438, "step": 280300 }, { "epoch": 0.5231642714352028, "grad_norm": 1.1339898109436035, "learning_rate": 0.00019746900774321644, "loss": 4.4215, "step": 280350 }, { "epoch": 0.5232575769945813, "grad_norm": 1.3833768367767334, "learning_rate": 0.00019746806848423568, "loss": 4.4671, "step": 280400 }, { "epoch": 0.5233508825539598, "grad_norm": 1.1775107383728027, "learning_rate": 0.00019746712905324097, "loss": 4.6874, "step": 280450 }, { "epoch": 0.5234441881133383, "grad_norm": 0.881015419960022, "learning_rate": 0.00019746618945023388, "loss": 4.6857, "step": 280500 }, { "epoch": 0.5235374936727167, "grad_norm": 1.1742804050445557, "learning_rate": 0.00019746524967521615, "loss": 4.5829, "step": 280550 }, { "epoch": 0.5236307992320952, "grad_norm": 1.0308231115341187, "learning_rate": 0.00019746430972818942, "loss": 4.4821, "step": 280600 }, { "epoch": 0.5237241047914737, "grad_norm": 0.9045799374580383, "learning_rate": 0.00019746336960915528, "loss": 4.6749, "step": 280650 }, { "epoch": 0.5238174103508523, "grad_norm": 1.095934510231018, "learning_rate": 0.00019746242931811548, "loss": 4.6171, "step": 280700 }, { "epoch": 0.5239107159102308, "grad_norm": 1.1799335479736328, "learning_rate": 0.00019746148885507162, "loss": 4.3884, "step": 280750 }, { "epoch": 0.5240040214696092, "grad_norm": 0.9894862771034241, "learning_rate": 0.0001974605482200254, "loss": 4.65, "step": 280800 }, { "epoch": 0.5240973270289877, "grad_norm": 0.993606448173523, "learning_rate": 0.00019745960741297843, "loss": 4.3739, "step": 280850 }, { "epoch": 0.5241906325883662, "grad_norm": 1.1276586055755615, "learning_rate": 0.00019745866643393245, "loss": 4.9291, "step": 280900 }, { "epoch": 0.5242839381477447, "grad_norm": 1.249640941619873, "learning_rate": 0.00019745772528288903, "loss": 4.5495, "step": 280950 }, { "epoch": 0.5243772437071231, "grad_norm": 0.9969247579574585, "learning_rate": 0.00019745678395984988, "loss": 4.7876, "step": 281000 }, { "epoch": 0.5244705492665017, "grad_norm": 0.9947354793548584, "learning_rate": 0.00019745584246481666, "loss": 4.6367, "step": 281050 }, { "epoch": 0.5245638548258802, "grad_norm": 1.1620121002197266, "learning_rate": 0.00019745490079779105, "loss": 4.6612, "step": 281100 }, { "epoch": 0.5246571603852587, "grad_norm": 0.8015241026878357, "learning_rate": 0.00019745395895877466, "loss": 4.5458, "step": 281150 }, { "epoch": 0.5247504659446371, "grad_norm": 1.141109585762024, "learning_rate": 0.00019745301694776916, "loss": 4.5927, "step": 281200 }, { "epoch": 0.5248437715040156, "grad_norm": 0.8895716071128845, "learning_rate": 0.0001974520747647763, "loss": 4.6357, "step": 281250 }, { "epoch": 0.5249370770633941, "grad_norm": 1.0158172845840454, "learning_rate": 0.00019745113240979763, "loss": 4.738, "step": 281300 }, { "epoch": 0.5250303826227726, "grad_norm": 1.0733691453933716, "learning_rate": 0.00019745018988283483, "loss": 4.5997, "step": 281350 }, { "epoch": 0.5251236881821512, "grad_norm": 1.1334929466247559, "learning_rate": 0.00019744924718388964, "loss": 4.6377, "step": 281400 }, { "epoch": 0.5252169937415296, "grad_norm": 1.1703345775604248, "learning_rate": 0.00019744830431296366, "loss": 4.6239, "step": 281450 }, { "epoch": 0.5253102993009081, "grad_norm": 1.0354031324386597, "learning_rate": 0.00019744736127005856, "loss": 4.5597, "step": 281500 }, { "epoch": 0.5254036048602866, "grad_norm": 1.0485466718673706, "learning_rate": 0.00019744641805517603, "loss": 4.5579, "step": 281550 }, { "epoch": 0.5254969104196651, "grad_norm": 1.360655665397644, "learning_rate": 0.0001974454746683177, "loss": 4.4888, "step": 281600 }, { "epoch": 0.5255902159790435, "grad_norm": 1.0404666662216187, "learning_rate": 0.0001974445311094853, "loss": 4.5823, "step": 281650 }, { "epoch": 0.525683521538422, "grad_norm": 1.0680186748504639, "learning_rate": 0.0001974435873786804, "loss": 4.5325, "step": 281700 }, { "epoch": 0.5257768270978006, "grad_norm": 1.1032402515411377, "learning_rate": 0.0001974426434759047, "loss": 4.5647, "step": 281750 }, { "epoch": 0.5258701326571791, "grad_norm": 0.9431357383728027, "learning_rate": 0.0001974416994011599, "loss": 4.5412, "step": 281800 }, { "epoch": 0.5259634382165576, "grad_norm": 1.0254426002502441, "learning_rate": 0.00019744075515444766, "loss": 4.4711, "step": 281850 }, { "epoch": 0.526056743775936, "grad_norm": 0.9435511231422424, "learning_rate": 0.00019743981073576966, "loss": 4.5718, "step": 281900 }, { "epoch": 0.5261500493353145, "grad_norm": 1.1502196788787842, "learning_rate": 0.00019743886614512747, "loss": 4.5823, "step": 281950 }, { "epoch": 0.526243354894693, "grad_norm": 1.0509811639785767, "learning_rate": 0.00019743792138252286, "loss": 4.7095, "step": 282000 }, { "epoch": 0.526243354894693, "eval_loss": 4.762018203735352, "eval_runtime": 231.817, "eval_samples_per_second": 11.25, "eval_steps_per_second": 11.25, "eval_tts_loss": 7.497885950622463, "step": 282000 }, { "epoch": 0.5263366604540715, "grad_norm": 0.8279403448104858, "learning_rate": 0.00019743697644795745, "loss": 4.4441, "step": 282050 }, { "epoch": 0.52642996601345, "grad_norm": 1.0076203346252441, "learning_rate": 0.00019743603134143295, "loss": 4.5973, "step": 282100 }, { "epoch": 0.5265232715728285, "grad_norm": 0.8520145416259766, "learning_rate": 0.000197435086062951, "loss": 4.7364, "step": 282150 }, { "epoch": 0.526616577132207, "grad_norm": 1.1191004514694214, "learning_rate": 0.00019743414061251324, "loss": 4.529, "step": 282200 }, { "epoch": 0.5267098826915855, "grad_norm": 1.0632494688034058, "learning_rate": 0.00019743319499012139, "loss": 4.4006, "step": 282250 }, { "epoch": 0.526803188250964, "grad_norm": 0.871149480342865, "learning_rate": 0.00019743224919577707, "loss": 4.6199, "step": 282300 }, { "epoch": 0.5268964938103424, "grad_norm": 1.0625333786010742, "learning_rate": 0.00019743130322948197, "loss": 4.5028, "step": 282350 }, { "epoch": 0.5269897993697209, "grad_norm": 1.019814133644104, "learning_rate": 0.00019743035709123778, "loss": 4.6573, "step": 282400 }, { "epoch": 0.5270831049290995, "grad_norm": 0.9306842088699341, "learning_rate": 0.00019742941078104616, "loss": 4.7412, "step": 282450 }, { "epoch": 0.527176410488478, "grad_norm": 1.135619878768921, "learning_rate": 0.0001974284642989088, "loss": 4.6812, "step": 282500 }, { "epoch": 0.5272697160478564, "grad_norm": 1.3619680404663086, "learning_rate": 0.0001974275176448273, "loss": 4.5811, "step": 282550 }, { "epoch": 0.5273630216072349, "grad_norm": 0.803629994392395, "learning_rate": 0.00019742657081880336, "loss": 4.776, "step": 282600 }, { "epoch": 0.5274563271666134, "grad_norm": 0.9954638481140137, "learning_rate": 0.0001974256238208387, "loss": 4.7264, "step": 282650 }, { "epoch": 0.5275496327259919, "grad_norm": 1.0035852193832397, "learning_rate": 0.00019742467665093493, "loss": 4.6593, "step": 282700 }, { "epoch": 0.5276429382853703, "grad_norm": 1.4352021217346191, "learning_rate": 0.00019742372930909379, "loss": 4.5499, "step": 282750 }, { "epoch": 0.5277362438447489, "grad_norm": 1.2543692588806152, "learning_rate": 0.00019742278179531687, "loss": 4.6667, "step": 282800 }, { "epoch": 0.5278295494041274, "grad_norm": 0.8242788314819336, "learning_rate": 0.0001974218341096059, "loss": 4.4721, "step": 282850 }, { "epoch": 0.5279228549635059, "grad_norm": 0.9672014117240906, "learning_rate": 0.00019742088625196251, "loss": 4.6143, "step": 282900 }, { "epoch": 0.5280161605228844, "grad_norm": 0.9670060873031616, "learning_rate": 0.00019741993822238843, "loss": 4.4835, "step": 282950 }, { "epoch": 0.5281094660822628, "grad_norm": 1.4916218519210815, "learning_rate": 0.0001974189900208853, "loss": 4.6265, "step": 283000 }, { "epoch": 0.5282027716416413, "grad_norm": 0.8706250786781311, "learning_rate": 0.00019741804164745476, "loss": 4.6028, "step": 283050 }, { "epoch": 0.5282960772010198, "grad_norm": 0.9748098254203796, "learning_rate": 0.00019741709310209854, "loss": 4.5401, "step": 283100 }, { "epoch": 0.5283893827603984, "grad_norm": 1.1996750831604004, "learning_rate": 0.0001974161443848183, "loss": 4.6417, "step": 283150 }, { "epoch": 0.5284826883197768, "grad_norm": 0.9673647880554199, "learning_rate": 0.0001974151954956157, "loss": 4.559, "step": 283200 }, { "epoch": 0.5285759938791553, "grad_norm": 0.6728837490081787, "learning_rate": 0.00019741424643449242, "loss": 4.497, "step": 283250 }, { "epoch": 0.5286692994385338, "grad_norm": 1.0573639869689941, "learning_rate": 0.00019741329720145015, "loss": 4.6833, "step": 283300 }, { "epoch": 0.5287626049979123, "grad_norm": 1.2526932954788208, "learning_rate": 0.00019741234779649053, "loss": 4.7223, "step": 283350 }, { "epoch": 0.5288559105572908, "grad_norm": 1.370250940322876, "learning_rate": 0.00019741139821961525, "loss": 4.5744, "step": 283400 }, { "epoch": 0.5289492161166692, "grad_norm": 1.1258753538131714, "learning_rate": 0.00019741044847082598, "loss": 4.734, "step": 283450 }, { "epoch": 0.5290425216760478, "grad_norm": 0.8937044143676758, "learning_rate": 0.00019740949855012445, "loss": 4.3539, "step": 283500 }, { "epoch": 0.5291358272354263, "grad_norm": 1.048108696937561, "learning_rate": 0.00019740854845751228, "loss": 4.4583, "step": 283550 }, { "epoch": 0.5292291327948048, "grad_norm": 0.9477057456970215, "learning_rate": 0.00019740759819299118, "loss": 4.553, "step": 283600 }, { "epoch": 0.5293224383541832, "grad_norm": 1.0268933773040771, "learning_rate": 0.00019740664775656276, "loss": 4.5117, "step": 283650 }, { "epoch": 0.5294157439135617, "grad_norm": 1.1293946504592896, "learning_rate": 0.00019740569714822876, "loss": 4.7847, "step": 283700 }, { "epoch": 0.5295090494729402, "grad_norm": 1.2568365335464478, "learning_rate": 0.00019740474636799089, "loss": 4.5597, "step": 283750 }, { "epoch": 0.5296023550323187, "grad_norm": 1.0086328983306885, "learning_rate": 0.00019740379541585075, "loss": 4.5726, "step": 283800 }, { "epoch": 0.5296956605916973, "grad_norm": 0.9730863571166992, "learning_rate": 0.00019740284429181003, "loss": 4.7144, "step": 283850 }, { "epoch": 0.5297889661510757, "grad_norm": 1.4535603523254395, "learning_rate": 0.00019740189299587044, "loss": 4.7222, "step": 283900 }, { "epoch": 0.5298822717104542, "grad_norm": 0.6582089066505432, "learning_rate": 0.00019740094152803365, "loss": 4.7405, "step": 283950 }, { "epoch": 0.5299755772698327, "grad_norm": 1.1787588596343994, "learning_rate": 0.00019739998988830136, "loss": 4.5965, "step": 284000 }, { "epoch": 0.5300688828292112, "grad_norm": 0.9802767038345337, "learning_rate": 0.00019739903807667518, "loss": 4.3648, "step": 284050 }, { "epoch": 0.5301621883885896, "grad_norm": 0.9081062078475952, "learning_rate": 0.00019739808609315688, "loss": 4.4767, "step": 284100 }, { "epoch": 0.5302554939479681, "grad_norm": 1.005858063697815, "learning_rate": 0.00019739713393774805, "loss": 4.5397, "step": 284150 }, { "epoch": 0.5303487995073466, "grad_norm": 0.6593231558799744, "learning_rate": 0.00019739618161045047, "loss": 4.5068, "step": 284200 }, { "epoch": 0.5304421050667252, "grad_norm": 0.8046698570251465, "learning_rate": 0.00019739522911126574, "loss": 4.6746, "step": 284250 }, { "epoch": 0.5305354106261037, "grad_norm": 1.3987061977386475, "learning_rate": 0.00019739427644019556, "loss": 4.469, "step": 284300 }, { "epoch": 0.5306287161854821, "grad_norm": 1.4002022743225098, "learning_rate": 0.00019739332359724164, "loss": 4.5972, "step": 284350 }, { "epoch": 0.5307220217448606, "grad_norm": 1.0923137664794922, "learning_rate": 0.00019739237058240562, "loss": 4.6223, "step": 284400 }, { "epoch": 0.5308153273042391, "grad_norm": 0.9077520370483398, "learning_rate": 0.00019739141739568921, "loss": 4.6044, "step": 284450 }, { "epoch": 0.5309086328636176, "grad_norm": 1.186133861541748, "learning_rate": 0.0001973904640370941, "loss": 4.8774, "step": 284500 }, { "epoch": 0.531001938422996, "grad_norm": 1.0076979398727417, "learning_rate": 0.0001973895105066219, "loss": 4.4713, "step": 284550 }, { "epoch": 0.5310952439823746, "grad_norm": 0.9802193641662598, "learning_rate": 0.00019738855680427443, "loss": 4.4804, "step": 284600 }, { "epoch": 0.5311885495417531, "grad_norm": 1.033267855644226, "learning_rate": 0.00019738760293005325, "loss": 4.6651, "step": 284650 }, { "epoch": 0.5312818551011316, "grad_norm": 0.9412552714347839, "learning_rate": 0.0001973866488839601, "loss": 4.8155, "step": 284700 }, { "epoch": 0.53137516066051, "grad_norm": 1.0584665536880493, "learning_rate": 0.00019738569466599664, "loss": 4.6357, "step": 284750 }, { "epoch": 0.5314684662198885, "grad_norm": 1.006569266319275, "learning_rate": 0.00019738474027616455, "loss": 4.4633, "step": 284800 }, { "epoch": 0.531561771779267, "grad_norm": 1.1155530214309692, "learning_rate": 0.00019738378571446555, "loss": 4.6122, "step": 284850 }, { "epoch": 0.5316550773386455, "grad_norm": 0.9938474297523499, "learning_rate": 0.0001973828309809013, "loss": 4.5603, "step": 284900 }, { "epoch": 0.5317483828980241, "grad_norm": 1.090856909751892, "learning_rate": 0.0001973818760754735, "loss": 4.4701, "step": 284950 }, { "epoch": 0.5318416884574025, "grad_norm": 1.146928310394287, "learning_rate": 0.0001973809209981838, "loss": 4.6746, "step": 285000 }, { "epoch": 0.5318416884574025, "eval_loss": 4.75473165512085, "eval_runtime": 230.3186, "eval_samples_per_second": 11.323, "eval_steps_per_second": 11.323, "eval_tts_loss": 7.53693535750549, "step": 285000 }, { "epoch": 0.531934994016781, "grad_norm": 0.9732300639152527, "learning_rate": 0.00019737996574903393, "loss": 4.5959, "step": 285050 }, { "epoch": 0.5320282995761595, "grad_norm": 1.0271157026290894, "learning_rate": 0.00019737901032802554, "loss": 4.5514, "step": 285100 }, { "epoch": 0.532121605135538, "grad_norm": 1.106013298034668, "learning_rate": 0.00019737805473516034, "loss": 4.7324, "step": 285150 }, { "epoch": 0.5322149106949164, "grad_norm": 0.9477126598358154, "learning_rate": 0.00019737709897044, "loss": 4.6364, "step": 285200 }, { "epoch": 0.5323082162542949, "grad_norm": 1.0403019189834595, "learning_rate": 0.00019737614303386622, "loss": 4.802, "step": 285250 }, { "epoch": 0.5324015218136735, "grad_norm": 1.1093212366104126, "learning_rate": 0.0001973751869254407, "loss": 4.4366, "step": 285300 }, { "epoch": 0.532494827373052, "grad_norm": 0.9458768367767334, "learning_rate": 0.00019737423064516507, "loss": 4.5341, "step": 285350 }, { "epoch": 0.5325881329324305, "grad_norm": 0.8419497609138489, "learning_rate": 0.00019737327419304106, "loss": 4.3237, "step": 285400 }, { "epoch": 0.5326814384918089, "grad_norm": 0.7255857586860657, "learning_rate": 0.0001973723175690704, "loss": 4.5856, "step": 285450 }, { "epoch": 0.5327747440511874, "grad_norm": 1.0696274042129517, "learning_rate": 0.0001973713607732547, "loss": 4.5296, "step": 285500 }, { "epoch": 0.5328680496105659, "grad_norm": 1.033064603805542, "learning_rate": 0.0001973704038055957, "loss": 4.64, "step": 285550 }, { "epoch": 0.5329613551699444, "grad_norm": 1.201291799545288, "learning_rate": 0.00019736944666609503, "loss": 4.6272, "step": 285600 }, { "epoch": 0.533054660729323, "grad_norm": 1.2754902839660645, "learning_rate": 0.00019736848935475446, "loss": 4.7426, "step": 285650 }, { "epoch": 0.5331479662887014, "grad_norm": 1.2380539178848267, "learning_rate": 0.0001973675318715756, "loss": 4.5593, "step": 285700 }, { "epoch": 0.5332412718480799, "grad_norm": 0.910490095615387, "learning_rate": 0.00019736657421656019, "loss": 4.4716, "step": 285750 }, { "epoch": 0.5333345774074584, "grad_norm": 0.6188794374465942, "learning_rate": 0.0001973656163897099, "loss": 4.5351, "step": 285800 }, { "epoch": 0.5334278829668369, "grad_norm": 1.1422240734100342, "learning_rate": 0.00019736465839102646, "loss": 4.5875, "step": 285850 }, { "epoch": 0.5335211885262153, "grad_norm": 1.2172739505767822, "learning_rate": 0.00019736370022051153, "loss": 4.5143, "step": 285900 }, { "epoch": 0.5336144940855938, "grad_norm": 1.0027328729629517, "learning_rate": 0.00019736274187816678, "loss": 4.6219, "step": 285950 }, { "epoch": 0.5337077996449724, "grad_norm": 0.7851713299751282, "learning_rate": 0.00019736178336399394, "loss": 4.5339, "step": 286000 }, { "epoch": 0.5338011052043509, "grad_norm": 0.8879555463790894, "learning_rate": 0.00019736082467799467, "loss": 4.5193, "step": 286050 }, { "epoch": 0.5338944107637293, "grad_norm": 0.8688876032829285, "learning_rate": 0.0001973598658201707, "loss": 4.5729, "step": 286100 }, { "epoch": 0.5339877163231078, "grad_norm": 1.100091814994812, "learning_rate": 0.00019735890679052368, "loss": 4.7068, "step": 286150 }, { "epoch": 0.5340810218824863, "grad_norm": 1.1109236478805542, "learning_rate": 0.0001973579475890553, "loss": 4.6203, "step": 286200 }, { "epoch": 0.5341743274418648, "grad_norm": 0.8393378853797913, "learning_rate": 0.0001973569882157673, "loss": 4.6772, "step": 286250 }, { "epoch": 0.5342676330012432, "grad_norm": 1.0183576345443726, "learning_rate": 0.00019735602867066134, "loss": 4.716, "step": 286300 }, { "epoch": 0.5343609385606218, "grad_norm": 1.0126652717590332, "learning_rate": 0.00019735506895373912, "loss": 4.6462, "step": 286350 }, { "epoch": 0.5344542441200003, "grad_norm": 0.8533194661140442, "learning_rate": 0.00019735410906500232, "loss": 4.542, "step": 286400 }, { "epoch": 0.5345475496793788, "grad_norm": 0.7949962019920349, "learning_rate": 0.00019735314900445268, "loss": 4.5383, "step": 286450 }, { "epoch": 0.5346408552387573, "grad_norm": 1.039992094039917, "learning_rate": 0.00019735218877209182, "loss": 4.4467, "step": 286500 }, { "epoch": 0.5347341607981357, "grad_norm": 0.6137852668762207, "learning_rate": 0.0001973512283679215, "loss": 4.7501, "step": 286550 }, { "epoch": 0.5348274663575142, "grad_norm": 1.16130793094635, "learning_rate": 0.00019735026779194338, "loss": 4.5853, "step": 286600 }, { "epoch": 0.5349207719168927, "grad_norm": 1.0255588293075562, "learning_rate": 0.0001973493070441592, "loss": 4.5439, "step": 286650 }, { "epoch": 0.5350140774762713, "grad_norm": 0.9397153258323669, "learning_rate": 0.00019734834612457057, "loss": 4.4539, "step": 286700 }, { "epoch": 0.5351073830356498, "grad_norm": 0.7592172622680664, "learning_rate": 0.00019734738503317928, "loss": 4.4875, "step": 286750 }, { "epoch": 0.5352006885950282, "grad_norm": 1.4141420125961304, "learning_rate": 0.00019734642376998695, "loss": 4.4221, "step": 286800 }, { "epoch": 0.5352939941544067, "grad_norm": 0.8302573561668396, "learning_rate": 0.00019734546233499532, "loss": 4.3809, "step": 286850 }, { "epoch": 0.5353872997137852, "grad_norm": 1.0930051803588867, "learning_rate": 0.00019734450072820612, "loss": 4.6485, "step": 286900 }, { "epoch": 0.5354806052731637, "grad_norm": 1.100099802017212, "learning_rate": 0.00019734353894962096, "loss": 4.5825, "step": 286950 }, { "epoch": 0.5355739108325421, "grad_norm": 1.1723564863204956, "learning_rate": 0.00019734257699924158, "loss": 4.3963, "step": 287000 }, { "epoch": 0.5356672163919207, "grad_norm": 1.0333951711654663, "learning_rate": 0.0001973416148770697, "loss": 4.4069, "step": 287050 }, { "epoch": 0.5357605219512992, "grad_norm": 1.0501785278320312, "learning_rate": 0.00019734065258310702, "loss": 4.5181, "step": 287100 }, { "epoch": 0.5358538275106777, "grad_norm": 1.1242258548736572, "learning_rate": 0.00019733969011735518, "loss": 4.3266, "step": 287150 }, { "epoch": 0.5359471330700561, "grad_norm": 1.1503238677978516, "learning_rate": 0.00019733872747981592, "loss": 4.5229, "step": 287200 }, { "epoch": 0.5360404386294346, "grad_norm": 0.9693351984024048, "learning_rate": 0.00019733776467049095, "loss": 4.5938, "step": 287250 }, { "epoch": 0.5361337441888131, "grad_norm": 0.7347463965415955, "learning_rate": 0.00019733680168938192, "loss": 4.4832, "step": 287300 }, { "epoch": 0.5362270497481916, "grad_norm": 1.387575626373291, "learning_rate": 0.00019733583853649058, "loss": 4.5632, "step": 287350 }, { "epoch": 0.5363203553075702, "grad_norm": 1.0806927680969238, "learning_rate": 0.00019733487521181862, "loss": 4.7334, "step": 287400 }, { "epoch": 0.5364136608669486, "grad_norm": 1.3513052463531494, "learning_rate": 0.00019733391171536773, "loss": 4.571, "step": 287450 }, { "epoch": 0.5365069664263271, "grad_norm": 1.1874183416366577, "learning_rate": 0.00019733294804713959, "loss": 4.5931, "step": 287500 }, { "epoch": 0.5366002719857056, "grad_norm": 1.212813138961792, "learning_rate": 0.00019733198420713595, "loss": 4.5991, "step": 287550 }, { "epoch": 0.5366935775450841, "grad_norm": 0.9041852355003357, "learning_rate": 0.00019733102019535846, "loss": 4.5846, "step": 287600 }, { "epoch": 0.5367868831044625, "grad_norm": 1.1560322046279907, "learning_rate": 0.00019733005601180886, "loss": 4.5619, "step": 287650 }, { "epoch": 0.536880188663841, "grad_norm": 0.8640111684799194, "learning_rate": 0.00019732909165648882, "loss": 4.7589, "step": 287700 }, { "epoch": 0.5369734942232196, "grad_norm": 1.0471079349517822, "learning_rate": 0.00019732812712940008, "loss": 4.6379, "step": 287750 }, { "epoch": 0.5370667997825981, "grad_norm": 1.2303370237350464, "learning_rate": 0.0001973271624305443, "loss": 4.5282, "step": 287800 }, { "epoch": 0.5371601053419766, "grad_norm": 0.6238991022109985, "learning_rate": 0.00019732619755992323, "loss": 4.533, "step": 287850 }, { "epoch": 0.537253410901355, "grad_norm": 1.0186882019042969, "learning_rate": 0.00019732523251753855, "loss": 4.4196, "step": 287900 }, { "epoch": 0.5373467164607335, "grad_norm": 0.9189663529396057, "learning_rate": 0.00019732426730339193, "loss": 4.8847, "step": 287950 }, { "epoch": 0.537440022020112, "grad_norm": 1.0216429233551025, "learning_rate": 0.0001973233019174851, "loss": 4.6982, "step": 288000 }, { "epoch": 0.537440022020112, "eval_loss": 4.758412837982178, "eval_runtime": 228.433, "eval_samples_per_second": 11.417, "eval_steps_per_second": 11.417, "eval_tts_loss": 7.586343273851861, "step": 288000 }, { "epoch": 0.5375333275794905, "grad_norm": 0.9328665137290955, "learning_rate": 0.0001973223363598198, "loss": 4.4907, "step": 288050 }, { "epoch": 0.537626633138869, "grad_norm": 0.9139004349708557, "learning_rate": 0.0001973213706303977, "loss": 4.4406, "step": 288100 }, { "epoch": 0.5377199386982475, "grad_norm": 0.7561057806015015, "learning_rate": 0.00019732040472922048, "loss": 4.6152, "step": 288150 }, { "epoch": 0.537813244257626, "grad_norm": 1.2258198261260986, "learning_rate": 0.00019731943865628987, "loss": 4.6745, "step": 288200 }, { "epoch": 0.5379065498170045, "grad_norm": 1.0092732906341553, "learning_rate": 0.0001973184724116076, "loss": 4.5054, "step": 288250 }, { "epoch": 0.537999855376383, "grad_norm": 1.1553289890289307, "learning_rate": 0.0001973175059951753, "loss": 4.6214, "step": 288300 }, { "epoch": 0.5380931609357614, "grad_norm": 0.8745859861373901, "learning_rate": 0.00019731653940699477, "loss": 4.4469, "step": 288350 }, { "epoch": 0.5381864664951399, "grad_norm": 1.3128819465637207, "learning_rate": 0.00019731557264706767, "loss": 4.8051, "step": 288400 }, { "epoch": 0.5382797720545185, "grad_norm": 0.9993950724601746, "learning_rate": 0.00019731460571539568, "loss": 4.6709, "step": 288450 }, { "epoch": 0.538373077613897, "grad_norm": 0.7750226259231567, "learning_rate": 0.00019731363861198056, "loss": 4.744, "step": 288500 }, { "epoch": 0.5384663831732754, "grad_norm": 1.1615644693374634, "learning_rate": 0.000197312671336824, "loss": 4.6706, "step": 288550 }, { "epoch": 0.5385596887326539, "grad_norm": 1.0617265701293945, "learning_rate": 0.00019731170388992766, "loss": 4.4617, "step": 288600 }, { "epoch": 0.5386529942920324, "grad_norm": 1.0319828987121582, "learning_rate": 0.0001973107362712933, "loss": 4.6425, "step": 288650 }, { "epoch": 0.5387462998514109, "grad_norm": 0.7510588765144348, "learning_rate": 0.00019730976848092265, "loss": 4.4306, "step": 288700 }, { "epoch": 0.5388396054107893, "grad_norm": 0.9257532358169556, "learning_rate": 0.00019730880051881732, "loss": 4.6251, "step": 288750 }, { "epoch": 0.5389329109701679, "grad_norm": 1.1619385480880737, "learning_rate": 0.0001973078323849791, "loss": 4.7905, "step": 288800 }, { "epoch": 0.5390262165295464, "grad_norm": 1.0947250127792358, "learning_rate": 0.0001973068640794097, "loss": 4.4057, "step": 288850 }, { "epoch": 0.5391195220889249, "grad_norm": 1.0521148443222046, "learning_rate": 0.00019730589560211078, "loss": 4.635, "step": 288900 }, { "epoch": 0.5392128276483034, "grad_norm": 0.9681885838508606, "learning_rate": 0.0001973049269530841, "loss": 4.4493, "step": 288950 }, { "epoch": 0.5393061332076818, "grad_norm": 0.8884767293930054, "learning_rate": 0.0001973039581323313, "loss": 4.5077, "step": 289000 }, { "epoch": 0.5393994387670603, "grad_norm": 1.1336555480957031, "learning_rate": 0.0001973029891398542, "loss": 4.536, "step": 289050 }, { "epoch": 0.5394927443264388, "grad_norm": 1.0752379894256592, "learning_rate": 0.0001973020199756544, "loss": 4.6676, "step": 289100 }, { "epoch": 0.5395860498858174, "grad_norm": 0.9591339230537415, "learning_rate": 0.00019730105063973369, "loss": 4.5266, "step": 289150 }, { "epoch": 0.5396793554451959, "grad_norm": 1.1089969873428345, "learning_rate": 0.00019730008113209372, "loss": 4.4951, "step": 289200 }, { "epoch": 0.5397726610045743, "grad_norm": 0.8516735434532166, "learning_rate": 0.00019729911145273624, "loss": 4.4916, "step": 289250 }, { "epoch": 0.5398659665639528, "grad_norm": 1.008996605873108, "learning_rate": 0.00019729814160166292, "loss": 4.6332, "step": 289300 }, { "epoch": 0.5399592721233313, "grad_norm": 0.9984986782073975, "learning_rate": 0.00019729717157887552, "loss": 4.6355, "step": 289350 }, { "epoch": 0.5400525776827098, "grad_norm": 0.8811632990837097, "learning_rate": 0.00019729620138437572, "loss": 4.3783, "step": 289400 }, { "epoch": 0.5401458832420882, "grad_norm": 1.0923969745635986, "learning_rate": 0.00019729523101816526, "loss": 4.771, "step": 289450 }, { "epoch": 0.5402391888014668, "grad_norm": 0.8841258883476257, "learning_rate": 0.00019729426048024586, "loss": 4.5809, "step": 289500 }, { "epoch": 0.5403324943608453, "grad_norm": 1.0029546022415161, "learning_rate": 0.00019729328977061917, "loss": 4.4276, "step": 289550 }, { "epoch": 0.5404257999202238, "grad_norm": 0.8529361486434937, "learning_rate": 0.00019729231888928697, "loss": 4.5002, "step": 289600 }, { "epoch": 0.5405191054796022, "grad_norm": 1.010185956954956, "learning_rate": 0.00019729134783625095, "loss": 4.6486, "step": 289650 }, { "epoch": 0.5406124110389807, "grad_norm": 0.8875432014465332, "learning_rate": 0.00019729037661151279, "loss": 4.5652, "step": 289700 }, { "epoch": 0.5407057165983592, "grad_norm": 1.2176717519760132, "learning_rate": 0.00019728940521507424, "loss": 4.4401, "step": 289750 }, { "epoch": 0.5407990221577377, "grad_norm": 0.6937578916549683, "learning_rate": 0.00019728843364693702, "loss": 4.5127, "step": 289800 }, { "epoch": 0.5408923277171162, "grad_norm": 1.0732295513153076, "learning_rate": 0.00019728746190710283, "loss": 4.3988, "step": 289850 }, { "epoch": 0.5409856332764947, "grad_norm": 1.096102237701416, "learning_rate": 0.00019728648999557338, "loss": 4.6227, "step": 289900 }, { "epoch": 0.5410789388358732, "grad_norm": 0.8393336534500122, "learning_rate": 0.00019728551791235038, "loss": 4.6566, "step": 289950 }, { "epoch": 0.5411722443952517, "grad_norm": 1.1003714799880981, "learning_rate": 0.0001972845456574356, "loss": 4.6147, "step": 290000 }, { "epoch": 0.5412655499546302, "grad_norm": 1.0207267999649048, "learning_rate": 0.0001972835732308307, "loss": 4.6324, "step": 290050 }, { "epoch": 0.5413588555140086, "grad_norm": 0.9939537644386292, "learning_rate": 0.0001972826006325374, "loss": 4.4746, "step": 290100 }, { "epoch": 0.5414521610733871, "grad_norm": 1.1141278743743896, "learning_rate": 0.0001972816278625574, "loss": 4.4754, "step": 290150 }, { "epoch": 0.5415454666327656, "grad_norm": 0.9295952916145325, "learning_rate": 0.00019728065492089248, "loss": 4.6563, "step": 290200 }, { "epoch": 0.5416387721921442, "grad_norm": 1.3614526987075806, "learning_rate": 0.00019727968180754433, "loss": 4.6514, "step": 290250 }, { "epoch": 0.5417320777515227, "grad_norm": 1.0252643823623657, "learning_rate": 0.00019727870852251464, "loss": 4.5044, "step": 290300 }, { "epoch": 0.5418253833109011, "grad_norm": 0.8520663380622864, "learning_rate": 0.00019727773506580517, "loss": 4.6316, "step": 290350 }, { "epoch": 0.5419186888702796, "grad_norm": 1.0140016078948975, "learning_rate": 0.00019727676143741757, "loss": 4.5094, "step": 290400 }, { "epoch": 0.5420119944296581, "grad_norm": 1.1733412742614746, "learning_rate": 0.00019727578763735365, "loss": 4.4411, "step": 290450 }, { "epoch": 0.5421052999890366, "grad_norm": 1.280668020248413, "learning_rate": 0.00019727481366561508, "loss": 4.5537, "step": 290500 }, { "epoch": 0.542198605548415, "grad_norm": 1.0060126781463623, "learning_rate": 0.00019727383952220353, "loss": 4.5231, "step": 290550 }, { "epoch": 0.5422919111077936, "grad_norm": 0.9130061268806458, "learning_rate": 0.0001972728652071208, "loss": 4.5402, "step": 290600 }, { "epoch": 0.5423852166671721, "grad_norm": 1.172722339630127, "learning_rate": 0.00019727189072036859, "loss": 4.6175, "step": 290650 }, { "epoch": 0.5424785222265506, "grad_norm": 0.979167640209198, "learning_rate": 0.0001972709160619486, "loss": 4.5181, "step": 290700 }, { "epoch": 0.542571827785929, "grad_norm": 0.9845288991928101, "learning_rate": 0.00019726994123186252, "loss": 4.5178, "step": 290750 }, { "epoch": 0.5426651333453075, "grad_norm": 0.8215778470039368, "learning_rate": 0.00019726896623011215, "loss": 4.6082, "step": 290800 }, { "epoch": 0.542758438904686, "grad_norm": 0.7376789450645447, "learning_rate": 0.00019726799105669918, "loss": 4.5592, "step": 290850 }, { "epoch": 0.5428517444640645, "grad_norm": 1.229047417640686, "learning_rate": 0.0001972670157116253, "loss": 4.6113, "step": 290900 }, { "epoch": 0.5429450500234431, "grad_norm": 1.0673273801803589, "learning_rate": 0.00019726604019489226, "loss": 4.5896, "step": 290950 }, { "epoch": 0.5430383555828215, "grad_norm": 1.2181791067123413, "learning_rate": 0.00019726506450650175, "loss": 4.6041, "step": 291000 }, { "epoch": 0.5430383555828215, "eval_loss": 4.755862236022949, "eval_runtime": 229.1406, "eval_samples_per_second": 11.382, "eval_steps_per_second": 11.382, "eval_tts_loss": 7.533074107711015, "step": 291000 }, { "epoch": 0.5431316611422, "grad_norm": 1.0295772552490234, "learning_rate": 0.00019726408864645555, "loss": 4.5021, "step": 291050 }, { "epoch": 0.5432249667015785, "grad_norm": 1.02863609790802, "learning_rate": 0.00019726311261475532, "loss": 4.7297, "step": 291100 }, { "epoch": 0.543318272260957, "grad_norm": 1.064806580543518, "learning_rate": 0.0001972621364114028, "loss": 4.7794, "step": 291150 }, { "epoch": 0.5434115778203354, "grad_norm": 1.2632663249969482, "learning_rate": 0.00019726116003639975, "loss": 4.5501, "step": 291200 }, { "epoch": 0.5435048833797139, "grad_norm": 0.9783966541290283, "learning_rate": 0.00019726018348974788, "loss": 4.6422, "step": 291250 }, { "epoch": 0.5435981889390925, "grad_norm": 1.000856637954712, "learning_rate": 0.0001972592067714489, "loss": 4.5299, "step": 291300 }, { "epoch": 0.543691494498471, "grad_norm": 1.0604976415634155, "learning_rate": 0.0001972582298815045, "loss": 4.2955, "step": 291350 }, { "epoch": 0.5437848000578495, "grad_norm": 1.129164695739746, "learning_rate": 0.00019725725281991645, "loss": 4.6644, "step": 291400 }, { "epoch": 0.5438781056172279, "grad_norm": 0.9613283276557922, "learning_rate": 0.00019725627558668646, "loss": 4.7118, "step": 291450 }, { "epoch": 0.5439714111766064, "grad_norm": 0.8535525798797607, "learning_rate": 0.00019725529818181626, "loss": 4.5516, "step": 291500 }, { "epoch": 0.5440647167359849, "grad_norm": 1.3331067562103271, "learning_rate": 0.00019725432060530755, "loss": 4.5355, "step": 291550 }, { "epoch": 0.5441580222953634, "grad_norm": 1.053837776184082, "learning_rate": 0.0001972533428571621, "loss": 4.4569, "step": 291600 }, { "epoch": 0.544251327854742, "grad_norm": 1.1131651401519775, "learning_rate": 0.0001972523649373816, "loss": 4.8562, "step": 291650 }, { "epoch": 0.5443446334141204, "grad_norm": 1.1066210269927979, "learning_rate": 0.00019725138684596782, "loss": 4.59, "step": 291700 }, { "epoch": 0.5444379389734989, "grad_norm": 0.9143478870391846, "learning_rate": 0.00019725040858292243, "loss": 4.5997, "step": 291750 }, { "epoch": 0.5445312445328774, "grad_norm": 0.9488169550895691, "learning_rate": 0.00019724943014824716, "loss": 4.4625, "step": 291800 }, { "epoch": 0.5446245500922559, "grad_norm": 1.1619845628738403, "learning_rate": 0.0001972484515419438, "loss": 4.5641, "step": 291850 }, { "epoch": 0.5447178556516343, "grad_norm": 0.7851349115371704, "learning_rate": 0.000197247472764014, "loss": 4.7522, "step": 291900 }, { "epoch": 0.5448111612110128, "grad_norm": 1.1872843503952026, "learning_rate": 0.00019724649381445955, "loss": 4.6147, "step": 291950 }, { "epoch": 0.5449044667703914, "grad_norm": 1.1382789611816406, "learning_rate": 0.0001972455146932821, "loss": 4.6354, "step": 292000 }, { "epoch": 0.5449977723297699, "grad_norm": 1.0083820819854736, "learning_rate": 0.00019724453540048348, "loss": 4.6133, "step": 292050 }, { "epoch": 0.5450910778891483, "grad_norm": 1.1564075946807861, "learning_rate": 0.00019724355593606535, "loss": 4.5128, "step": 292100 }, { "epoch": 0.5451843834485268, "grad_norm": 1.0278780460357666, "learning_rate": 0.00019724257630002946, "loss": 4.7728, "step": 292150 }, { "epoch": 0.5452776890079053, "grad_norm": 1.1431865692138672, "learning_rate": 0.00019724159649237754, "loss": 4.4999, "step": 292200 }, { "epoch": 0.5453709945672838, "grad_norm": 1.0673056840896606, "learning_rate": 0.0001972406165131113, "loss": 4.5965, "step": 292250 }, { "epoch": 0.5454643001266622, "grad_norm": 1.0915521383285522, "learning_rate": 0.00019723963636223245, "loss": 4.4474, "step": 292300 }, { "epoch": 0.5455576056860408, "grad_norm": 1.1890692710876465, "learning_rate": 0.0001972386560397428, "loss": 4.6021, "step": 292350 }, { "epoch": 0.5456509112454193, "grad_norm": 1.1801838874816895, "learning_rate": 0.000197237675545644, "loss": 4.4898, "step": 292400 }, { "epoch": 0.5457442168047978, "grad_norm": 1.9533765316009521, "learning_rate": 0.00019723669487993784, "loss": 4.5624, "step": 292450 }, { "epoch": 0.5458375223641763, "grad_norm": 1.0451918840408325, "learning_rate": 0.00019723571404262603, "loss": 4.6414, "step": 292500 }, { "epoch": 0.5459308279235547, "grad_norm": 0.9235672354698181, "learning_rate": 0.00019723473303371022, "loss": 4.4504, "step": 292550 }, { "epoch": 0.5460241334829332, "grad_norm": 0.9666757583618164, "learning_rate": 0.00019723375185319228, "loss": 4.5351, "step": 292600 }, { "epoch": 0.5461174390423117, "grad_norm": 0.9865828156471252, "learning_rate": 0.00019723277050107387, "loss": 4.6113, "step": 292650 }, { "epoch": 0.5462107446016903, "grad_norm": 0.8017159700393677, "learning_rate": 0.00019723178897735672, "loss": 4.6902, "step": 292700 }, { "epoch": 0.5463040501610688, "grad_norm": 0.959333598613739, "learning_rate": 0.00019723080728204257, "loss": 4.5285, "step": 292750 }, { "epoch": 0.5463973557204472, "grad_norm": 1.0091277360916138, "learning_rate": 0.00019722982541513318, "loss": 4.5637, "step": 292800 }, { "epoch": 0.5464906612798257, "grad_norm": 0.985924243927002, "learning_rate": 0.00019722884337663023, "loss": 4.686, "step": 292850 }, { "epoch": 0.5465839668392042, "grad_norm": 1.0899864435195923, "learning_rate": 0.00019722786116653548, "loss": 4.3514, "step": 292900 }, { "epoch": 0.5466772723985827, "grad_norm": 1.169295072555542, "learning_rate": 0.00019722687878485066, "loss": 4.5164, "step": 292950 }, { "epoch": 0.5467705779579611, "grad_norm": 1.144992709159851, "learning_rate": 0.00019722589623157751, "loss": 4.6521, "step": 293000 }, { "epoch": 0.5468638835173397, "grad_norm": 1.1579240560531616, "learning_rate": 0.00019722491350671777, "loss": 4.6411, "step": 293050 }, { "epoch": 0.5469571890767182, "grad_norm": 0.9811488389968872, "learning_rate": 0.00019722393061027315, "loss": 4.8124, "step": 293100 }, { "epoch": 0.5470504946360967, "grad_norm": 1.1563857793807983, "learning_rate": 0.0001972229475422454, "loss": 4.5458, "step": 293150 }, { "epoch": 0.5471438001954751, "grad_norm": 1.1877684593200684, "learning_rate": 0.00019722196430263628, "loss": 4.5323, "step": 293200 }, { "epoch": 0.5472371057548536, "grad_norm": 1.0926393270492554, "learning_rate": 0.00019722098089144748, "loss": 4.5031, "step": 293250 }, { "epoch": 0.5473304113142321, "grad_norm": 0.8998591303825378, "learning_rate": 0.00019721999730868076, "loss": 4.7053, "step": 293300 }, { "epoch": 0.5474237168736106, "grad_norm": 1.27166748046875, "learning_rate": 0.00019721901355433782, "loss": 4.553, "step": 293350 }, { "epoch": 0.5475170224329892, "grad_norm": 1.5627135038375854, "learning_rate": 0.00019721802962842047, "loss": 4.6294, "step": 293400 }, { "epoch": 0.5476103279923676, "grad_norm": 1.1435675621032715, "learning_rate": 0.0001972170455309304, "loss": 4.4011, "step": 293450 }, { "epoch": 0.5477036335517461, "grad_norm": 1.1832562685012817, "learning_rate": 0.00019721606126186933, "loss": 4.4907, "step": 293500 }, { "epoch": 0.5477969391111246, "grad_norm": 0.8468302488327026, "learning_rate": 0.00019721507682123903, "loss": 4.6218, "step": 293550 }, { "epoch": 0.5478902446705031, "grad_norm": 1.006789207458496, "learning_rate": 0.0001972140922090412, "loss": 4.5976, "step": 293600 }, { "epoch": 0.5479835502298815, "grad_norm": 0.8985641002655029, "learning_rate": 0.00019721310742527764, "loss": 4.5135, "step": 293650 }, { "epoch": 0.54807685578926, "grad_norm": 0.9235044121742249, "learning_rate": 0.00019721212246995006, "loss": 4.5865, "step": 293700 }, { "epoch": 0.5481701613486386, "grad_norm": 1.1763442754745483, "learning_rate": 0.00019721113734306014, "loss": 4.7422, "step": 293750 }, { "epoch": 0.5482634669080171, "grad_norm": 1.071958065032959, "learning_rate": 0.0001972101520446097, "loss": 4.7605, "step": 293800 }, { "epoch": 0.5483567724673956, "grad_norm": 0.9237344861030579, "learning_rate": 0.00019720916657460041, "loss": 4.6234, "step": 293850 }, { "epoch": 0.548450078026774, "grad_norm": 0.7987807393074036, "learning_rate": 0.00019720818093303408, "loss": 4.6693, "step": 293900 }, { "epoch": 0.5485433835861525, "grad_norm": 1.0376191139221191, "learning_rate": 0.0001972071951199124, "loss": 4.715, "step": 293950 }, { "epoch": 0.548636689145531, "grad_norm": 0.8929983377456665, "learning_rate": 0.00019720620913523716, "loss": 4.5809, "step": 294000 }, { "epoch": 0.548636689145531, "eval_loss": 4.764010429382324, "eval_runtime": 229.1673, "eval_samples_per_second": 11.38, "eval_steps_per_second": 11.38, "eval_tts_loss": 7.565109917429773, "step": 294000 }, { "epoch": 0.5487299947049095, "grad_norm": 0.9813073873519897, "learning_rate": 0.00019720522297901003, "loss": 4.6842, "step": 294050 }, { "epoch": 0.548823300264288, "grad_norm": 0.7194148302078247, "learning_rate": 0.00019720423665123277, "loss": 4.5848, "step": 294100 }, { "epoch": 0.5489166058236665, "grad_norm": 1.0934531688690186, "learning_rate": 0.00019720325015190718, "loss": 4.619, "step": 294150 }, { "epoch": 0.549009911383045, "grad_norm": 0.726915717124939, "learning_rate": 0.00019720226348103492, "loss": 4.5181, "step": 294200 }, { "epoch": 0.5491032169424235, "grad_norm": 0.8358513712882996, "learning_rate": 0.00019720127663861781, "loss": 4.3869, "step": 294250 }, { "epoch": 0.549196522501802, "grad_norm": 0.9418166875839233, "learning_rate": 0.0001972002896246575, "loss": 4.4952, "step": 294300 }, { "epoch": 0.5492898280611804, "grad_norm": 0.9657112956047058, "learning_rate": 0.0001971993024391558, "loss": 4.5174, "step": 294350 }, { "epoch": 0.5493831336205589, "grad_norm": 0.9693840146064758, "learning_rate": 0.00019719831508211447, "loss": 4.5318, "step": 294400 }, { "epoch": 0.5494764391799375, "grad_norm": 0.6684638261795044, "learning_rate": 0.00019719732755353516, "loss": 4.728, "step": 294450 }, { "epoch": 0.549569744739316, "grad_norm": 1.061577558517456, "learning_rate": 0.0001971963398534197, "loss": 4.7139, "step": 294500 }, { "epoch": 0.5496630502986944, "grad_norm": 1.12180495262146, "learning_rate": 0.0001971953519817698, "loss": 4.6283, "step": 294550 }, { "epoch": 0.5497563558580729, "grad_norm": 1.1466017961502075, "learning_rate": 0.0001971943639385872, "loss": 4.477, "step": 294600 }, { "epoch": 0.5498496614174514, "grad_norm": 1.0311684608459473, "learning_rate": 0.00019719337572387367, "loss": 4.5822, "step": 294650 }, { "epoch": 0.5499429669768299, "grad_norm": 0.9233488440513611, "learning_rate": 0.00019719238733763091, "loss": 4.5704, "step": 294700 }, { "epoch": 0.5500362725362083, "grad_norm": 1.1542483568191528, "learning_rate": 0.00019719139877986069, "loss": 4.62, "step": 294750 }, { "epoch": 0.5501295780955869, "grad_norm": 1.068114995956421, "learning_rate": 0.00019719041005056477, "loss": 4.6344, "step": 294800 }, { "epoch": 0.5502228836549654, "grad_norm": 1.212415099143982, "learning_rate": 0.00019718942114974484, "loss": 4.3569, "step": 294850 }, { "epoch": 0.5503161892143439, "grad_norm": 0.9979035258293152, "learning_rate": 0.00019718843207740272, "loss": 4.5337, "step": 294900 }, { "epoch": 0.5504094947737224, "grad_norm": 1.1693506240844727, "learning_rate": 0.0001971874428335401, "loss": 4.5714, "step": 294950 }, { "epoch": 0.5505028003331008, "grad_norm": 0.8749262690544128, "learning_rate": 0.00019718645341815875, "loss": 4.5207, "step": 295000 }, { "epoch": 0.5505961058924793, "grad_norm": 1.2598340511322021, "learning_rate": 0.00019718546383126043, "loss": 4.6344, "step": 295050 }, { "epoch": 0.5506894114518578, "grad_norm": 1.0515248775482178, "learning_rate": 0.00019718447407284683, "loss": 4.6527, "step": 295100 }, { "epoch": 0.5507827170112363, "grad_norm": 0.9423948526382446, "learning_rate": 0.00019718348414291974, "loss": 4.7807, "step": 295150 }, { "epoch": 0.5508760225706149, "grad_norm": 1.0501548051834106, "learning_rate": 0.0001971824940414809, "loss": 4.5323, "step": 295200 }, { "epoch": 0.5509693281299933, "grad_norm": 1.265120029449463, "learning_rate": 0.0001971815037685321, "loss": 4.7063, "step": 295250 }, { "epoch": 0.5510626336893718, "grad_norm": 1.0201817750930786, "learning_rate": 0.00019718051332407498, "loss": 4.6494, "step": 295300 }, { "epoch": 0.5511559392487503, "grad_norm": 0.9585539698600769, "learning_rate": 0.0001971795227081114, "loss": 4.5697, "step": 295350 }, { "epoch": 0.5512492448081288, "grad_norm": 0.8855648636817932, "learning_rate": 0.00019717853192064303, "loss": 4.69, "step": 295400 }, { "epoch": 0.5513425503675072, "grad_norm": 0.7688419818878174, "learning_rate": 0.00019717754096167166, "loss": 4.5284, "step": 295450 }, { "epoch": 0.5514358559268857, "grad_norm": 1.1366264820098877, "learning_rate": 0.00019717654983119907, "loss": 4.7289, "step": 295500 }, { "epoch": 0.5515291614862643, "grad_norm": 2.7926149368286133, "learning_rate": 0.00019717555852922695, "loss": 4.735, "step": 295550 }, { "epoch": 0.5516224670456428, "grad_norm": 1.3501007556915283, "learning_rate": 0.00019717456705575703, "loss": 4.4595, "step": 295600 }, { "epoch": 0.5517157726050212, "grad_norm": 0.9407560229301453, "learning_rate": 0.00019717357541079112, "loss": 4.703, "step": 295650 }, { "epoch": 0.5518090781643997, "grad_norm": 1.0523128509521484, "learning_rate": 0.00019717258359433094, "loss": 4.5778, "step": 295700 }, { "epoch": 0.5519023837237782, "grad_norm": 1.0216292142868042, "learning_rate": 0.00019717159160637828, "loss": 4.5917, "step": 295750 }, { "epoch": 0.5519956892831567, "grad_norm": 1.2850682735443115, "learning_rate": 0.00019717059944693484, "loss": 4.5741, "step": 295800 }, { "epoch": 0.5520889948425352, "grad_norm": 0.8473016023635864, "learning_rate": 0.00019716960711600235, "loss": 4.6217, "step": 295850 }, { "epoch": 0.5521823004019137, "grad_norm": 1.0704396963119507, "learning_rate": 0.00019716861461358266, "loss": 4.5319, "step": 295900 }, { "epoch": 0.5522756059612922, "grad_norm": 1.325152039527893, "learning_rate": 0.00019716762193967743, "loss": 4.6891, "step": 295950 }, { "epoch": 0.5523689115206707, "grad_norm": 0.8382663130760193, "learning_rate": 0.00019716662909428844, "loss": 4.4307, "step": 296000 }, { "epoch": 0.5524622170800492, "grad_norm": 0.8413145542144775, "learning_rate": 0.00019716563607741748, "loss": 4.7154, "step": 296050 }, { "epoch": 0.5525555226394276, "grad_norm": 0.9249318838119507, "learning_rate": 0.00019716464288906625, "loss": 4.6292, "step": 296100 }, { "epoch": 0.5526488281988061, "grad_norm": 0.9210410714149475, "learning_rate": 0.0001971636495292365, "loss": 4.6499, "step": 296150 }, { "epoch": 0.5527421337581846, "grad_norm": 0.881159245967865, "learning_rate": 0.00019716265599793004, "loss": 4.7111, "step": 296200 }, { "epoch": 0.5528354393175632, "grad_norm": 0.7526847720146179, "learning_rate": 0.00019716166229514855, "loss": 4.6092, "step": 296250 }, { "epoch": 0.5529287448769417, "grad_norm": 1.2674932479858398, "learning_rate": 0.00019716066842089387, "loss": 4.5312, "step": 296300 }, { "epoch": 0.5530220504363201, "grad_norm": 0.6895270347595215, "learning_rate": 0.00019715967437516766, "loss": 4.6097, "step": 296350 }, { "epoch": 0.5531153559956986, "grad_norm": 1.333682656288147, "learning_rate": 0.00019715868015797176, "loss": 4.6198, "step": 296400 }, { "epoch": 0.5532086615550771, "grad_norm": 1.0232683420181274, "learning_rate": 0.00019715768576930787, "loss": 4.6465, "step": 296450 }, { "epoch": 0.5533019671144556, "grad_norm": 0.875064492225647, "learning_rate": 0.00019715669120917774, "loss": 4.6424, "step": 296500 }, { "epoch": 0.553395272673834, "grad_norm": 1.2495737075805664, "learning_rate": 0.00019715569647758317, "loss": 4.694, "step": 296550 }, { "epoch": 0.5534885782332126, "grad_norm": 0.9244838953018188, "learning_rate": 0.0001971547015745259, "loss": 4.6244, "step": 296600 }, { "epoch": 0.5535818837925911, "grad_norm": 1.1226541996002197, "learning_rate": 0.00019715370650000765, "loss": 4.369, "step": 296650 }, { "epoch": 0.5536751893519696, "grad_norm": 1.1125637292861938, "learning_rate": 0.0001971527112540302, "loss": 4.8984, "step": 296700 }, { "epoch": 0.553768494911348, "grad_norm": 0.9522160887718201, "learning_rate": 0.00019715171583659536, "loss": 4.6531, "step": 296750 }, { "epoch": 0.5538618004707265, "grad_norm": 1.3746414184570312, "learning_rate": 0.0001971507202477048, "loss": 4.321, "step": 296800 }, { "epoch": 0.553955106030105, "grad_norm": 0.8997330069541931, "learning_rate": 0.0001971497244873603, "loss": 4.5425, "step": 296850 }, { "epoch": 0.5540484115894835, "grad_norm": 1.069753646850586, "learning_rate": 0.00019714872855556363, "loss": 4.4188, "step": 296900 }, { "epoch": 0.5541417171488621, "grad_norm": 1.023139476776123, "learning_rate": 0.00019714773245231657, "loss": 4.4299, "step": 296950 }, { "epoch": 0.5542350227082405, "grad_norm": 0.9713852405548096, "learning_rate": 0.00019714673617762086, "loss": 4.4206, "step": 297000 }, { "epoch": 0.5542350227082405, "eval_loss": 4.7526021003723145, "eval_runtime": 231.4288, "eval_samples_per_second": 11.269, "eval_steps_per_second": 11.269, "eval_tts_loss": 7.537269182848898, "step": 297000 }, { "epoch": 0.554328328267619, "grad_norm": 0.9608700275421143, "learning_rate": 0.00019714573973147822, "loss": 4.6437, "step": 297050 }, { "epoch": 0.5544216338269975, "grad_norm": 1.1964012384414673, "learning_rate": 0.00019714474311389047, "loss": 4.6537, "step": 297100 }, { "epoch": 0.554514939386376, "grad_norm": 1.1816233396530151, "learning_rate": 0.00019714374632485933, "loss": 4.4968, "step": 297150 }, { "epoch": 0.5546082449457544, "grad_norm": 0.9269901514053345, "learning_rate": 0.00019714274936438657, "loss": 4.6129, "step": 297200 }, { "epoch": 0.5547015505051329, "grad_norm": 1.0753899812698364, "learning_rate": 0.00019714175223247397, "loss": 4.6587, "step": 297250 }, { "epoch": 0.5547948560645115, "grad_norm": 1.0030367374420166, "learning_rate": 0.00019714075492912326, "loss": 4.6356, "step": 297300 }, { "epoch": 0.55488816162389, "grad_norm": 0.9227311015129089, "learning_rate": 0.0001971397574543362, "loss": 4.4709, "step": 297350 }, { "epoch": 0.5549814671832685, "grad_norm": 1.0956863164901733, "learning_rate": 0.00019713875980811456, "loss": 4.4858, "step": 297400 }, { "epoch": 0.5550747727426469, "grad_norm": 0.9432074427604675, "learning_rate": 0.0001971377619904601, "loss": 4.4785, "step": 297450 }, { "epoch": 0.5551680783020254, "grad_norm": 1.325101375579834, "learning_rate": 0.0001971367640013746, "loss": 4.4846, "step": 297500 }, { "epoch": 0.5552613838614039, "grad_norm": 1.0917229652404785, "learning_rate": 0.00019713576584085978, "loss": 4.465, "step": 297550 }, { "epoch": 0.5553546894207824, "grad_norm": 1.0985119342803955, "learning_rate": 0.00019713476750891744, "loss": 4.7151, "step": 297600 }, { "epoch": 0.555447994980161, "grad_norm": 0.9184651970863342, "learning_rate": 0.00019713376900554931, "loss": 4.5393, "step": 297650 }, { "epoch": 0.5555413005395394, "grad_norm": 1.0063183307647705, "learning_rate": 0.0001971327703307572, "loss": 4.5729, "step": 297700 }, { "epoch": 0.5556346060989179, "grad_norm": 0.9919562935829163, "learning_rate": 0.00019713177148454283, "loss": 4.7591, "step": 297750 }, { "epoch": 0.5557279116582964, "grad_norm": 0.8779546022415161, "learning_rate": 0.00019713077246690794, "loss": 4.4183, "step": 297800 }, { "epoch": 0.5558212172176749, "grad_norm": 0.8050033450126648, "learning_rate": 0.00019712977327785436, "loss": 4.463, "step": 297850 }, { "epoch": 0.5559145227770533, "grad_norm": 1.0061942338943481, "learning_rate": 0.00019712877391738382, "loss": 4.8949, "step": 297900 }, { "epoch": 0.5560078283364318, "grad_norm": 0.9918883442878723, "learning_rate": 0.0001971277743854981, "loss": 4.5452, "step": 297950 }, { "epoch": 0.5561011338958104, "grad_norm": 0.9152721762657166, "learning_rate": 0.00019712677468219892, "loss": 4.603, "step": 298000 }, { "epoch": 0.5561944394551889, "grad_norm": 0.9235228896141052, "learning_rate": 0.0001971257748074881, "loss": 4.4085, "step": 298050 }, { "epoch": 0.5562877450145673, "grad_norm": 1.0980701446533203, "learning_rate": 0.00019712477476136733, "loss": 4.4507, "step": 298100 }, { "epoch": 0.5563810505739458, "grad_norm": 0.8840431571006775, "learning_rate": 0.00019712377454383847, "loss": 4.7334, "step": 298150 }, { "epoch": 0.5564743561333243, "grad_norm": 0.9748885035514832, "learning_rate": 0.0001971227741549032, "loss": 4.8757, "step": 298200 }, { "epoch": 0.5565676616927028, "grad_norm": 1.0682768821716309, "learning_rate": 0.00019712177359456335, "loss": 4.6206, "step": 298250 }, { "epoch": 0.5566609672520813, "grad_norm": 1.0033636093139648, "learning_rate": 0.00019712077286282065, "loss": 4.4473, "step": 298300 }, { "epoch": 0.5567542728114598, "grad_norm": 0.8800263404846191, "learning_rate": 0.00019711977195967687, "loss": 4.5585, "step": 298350 }, { "epoch": 0.5568475783708383, "grad_norm": 1.004462718963623, "learning_rate": 0.00019711877088513378, "loss": 4.703, "step": 298400 }, { "epoch": 0.5569408839302168, "grad_norm": 0.99757981300354, "learning_rate": 0.00019711776963919316, "loss": 4.751, "step": 298450 }, { "epoch": 0.5570341894895953, "grad_norm": 0.8325973749160767, "learning_rate": 0.00019711676822185674, "loss": 4.7005, "step": 298500 }, { "epoch": 0.5571274950489737, "grad_norm": 0.8714603781700134, "learning_rate": 0.00019711576663312635, "loss": 4.7723, "step": 298550 }, { "epoch": 0.5572208006083522, "grad_norm": 1.3608721494674683, "learning_rate": 0.00019711476487300372, "loss": 4.6039, "step": 298600 }, { "epoch": 0.5573141061677307, "grad_norm": 1.2337629795074463, "learning_rate": 0.00019711376294149056, "loss": 4.4813, "step": 298650 }, { "epoch": 0.5574074117271093, "grad_norm": 1.042840600013733, "learning_rate": 0.00019711276083858877, "loss": 4.5122, "step": 298700 }, { "epoch": 0.5575007172864878, "grad_norm": 1.5062463283538818, "learning_rate": 0.0001971117585643, "loss": 4.6468, "step": 298750 }, { "epoch": 0.5575940228458662, "grad_norm": 1.0929310321807861, "learning_rate": 0.00019711075611862606, "loss": 4.6295, "step": 298800 }, { "epoch": 0.5576873284052447, "grad_norm": 0.9297996163368225, "learning_rate": 0.00019710975350156873, "loss": 4.448, "step": 298850 }, { "epoch": 0.5577806339646232, "grad_norm": 1.0459295511245728, "learning_rate": 0.00019710875071312977, "loss": 4.4881, "step": 298900 }, { "epoch": 0.5578739395240017, "grad_norm": 0.9883279800415039, "learning_rate": 0.00019710774775331098, "loss": 4.611, "step": 298950 }, { "epoch": 0.5579672450833801, "grad_norm": 0.9868349432945251, "learning_rate": 0.00019710674462211408, "loss": 4.525, "step": 299000 }, { "epoch": 0.5580605506427587, "grad_norm": 0.8064708113670349, "learning_rate": 0.00019710574131954087, "loss": 4.493, "step": 299050 }, { "epoch": 0.5581538562021372, "grad_norm": 1.1615211963653564, "learning_rate": 0.0001971047378455931, "loss": 4.6057, "step": 299100 }, { "epoch": 0.5582471617615157, "grad_norm": 1.1374075412750244, "learning_rate": 0.00019710373420027256, "loss": 4.5673, "step": 299150 }, { "epoch": 0.5583404673208942, "grad_norm": 0.9291498064994812, "learning_rate": 0.00019710273038358103, "loss": 4.4507, "step": 299200 }, { "epoch": 0.5584337728802726, "grad_norm": 1.0071158409118652, "learning_rate": 0.00019710172639552024, "loss": 4.7346, "step": 299250 }, { "epoch": 0.5585270784396511, "grad_norm": 0.8895450830459595, "learning_rate": 0.000197100722236092, "loss": 4.4564, "step": 299300 }, { "epoch": 0.5586203839990296, "grad_norm": 0.9139736890792847, "learning_rate": 0.00019709971790529807, "loss": 4.6733, "step": 299350 }, { "epoch": 0.5587136895584082, "grad_norm": 1.2478666305541992, "learning_rate": 0.0001970987134031402, "loss": 4.7858, "step": 299400 }, { "epoch": 0.5588069951177866, "grad_norm": 0.9026058912277222, "learning_rate": 0.00019709770872962024, "loss": 4.6203, "step": 299450 }, { "epoch": 0.5589003006771651, "grad_norm": 1.2664495706558228, "learning_rate": 0.00019709670388473987, "loss": 4.6035, "step": 299500 }, { "epoch": 0.5589936062365436, "grad_norm": 1.0902659893035889, "learning_rate": 0.00019709569886850093, "loss": 4.7008, "step": 299550 }, { "epoch": 0.5590869117959221, "grad_norm": 0.7903751134872437, "learning_rate": 0.00019709469368090515, "loss": 4.5173, "step": 299600 }, { "epoch": 0.5591802173553005, "grad_norm": 0.8331606984138489, "learning_rate": 0.0001970936883219543, "loss": 4.7115, "step": 299650 }, { "epoch": 0.559273522914679, "grad_norm": 1.0867011547088623, "learning_rate": 0.00019709268279165018, "loss": 4.6903, "step": 299700 }, { "epoch": 0.5593668284740576, "grad_norm": 1.0519412755966187, "learning_rate": 0.00019709167708999457, "loss": 4.4897, "step": 299750 }, { "epoch": 0.5594601340334361, "grad_norm": 1.4251899719238281, "learning_rate": 0.00019709067121698924, "loss": 4.6184, "step": 299800 }, { "epoch": 0.5595534395928146, "grad_norm": 1.2636020183563232, "learning_rate": 0.00019708966517263595, "loss": 4.4974, "step": 299850 }, { "epoch": 0.559646745152193, "grad_norm": 0.7390130758285522, "learning_rate": 0.00019708865895693648, "loss": 4.6559, "step": 299900 }, { "epoch": 0.5597400507115715, "grad_norm": 0.8000625967979431, "learning_rate": 0.00019708765256989263, "loss": 4.5117, "step": 299950 }, { "epoch": 0.55983335627095, "grad_norm": 1.0556968450546265, "learning_rate": 0.00019708664601150616, "loss": 4.4716, "step": 300000 }, { "epoch": 0.55983335627095, "eval_loss": 4.755502700805664, "eval_runtime": 232.7346, "eval_samples_per_second": 11.206, "eval_steps_per_second": 11.206, "eval_tts_loss": 7.561606850610771, "step": 300000 }, { "epoch": 0.5599266618303285, "grad_norm": 1.1514503955841064, "learning_rate": 0.0001970856392817788, "loss": 4.5287, "step": 300050 }, { "epoch": 0.560019967389707, "grad_norm": 1.0816736221313477, "learning_rate": 0.00019708463238071242, "loss": 4.6709, "step": 300100 }, { "epoch": 0.5601132729490855, "grad_norm": 0.9680812954902649, "learning_rate": 0.0001970836253083087, "loss": 4.5883, "step": 300150 }, { "epoch": 0.560206578508464, "grad_norm": 1.0057451725006104, "learning_rate": 0.0001970826180645695, "loss": 4.4862, "step": 300200 }, { "epoch": 0.5602998840678425, "grad_norm": 0.8072318434715271, "learning_rate": 0.00019708161064949656, "loss": 4.5203, "step": 300250 }, { "epoch": 0.560393189627221, "grad_norm": 0.8877402544021606, "learning_rate": 0.00019708060306309163, "loss": 4.6058, "step": 300300 }, { "epoch": 0.5604864951865994, "grad_norm": 0.8873353600502014, "learning_rate": 0.00019707959530535654, "loss": 4.8, "step": 300350 }, { "epoch": 0.5605798007459779, "grad_norm": 0.993877112865448, "learning_rate": 0.00019707858737629303, "loss": 4.4375, "step": 300400 }, { "epoch": 0.5606731063053564, "grad_norm": 0.8739638924598694, "learning_rate": 0.00019707757927590292, "loss": 4.4138, "step": 300450 }, { "epoch": 0.560766411864735, "grad_norm": 1.0241881608963013, "learning_rate": 0.00019707657100418794, "loss": 4.4477, "step": 300500 }, { "epoch": 0.5608597174241134, "grad_norm": 0.8494142293930054, "learning_rate": 0.0001970755625611499, "loss": 4.5795, "step": 300550 }, { "epoch": 0.5609530229834919, "grad_norm": 1.0680198669433594, "learning_rate": 0.00019707455394679058, "loss": 4.562, "step": 300600 }, { "epoch": 0.5610463285428704, "grad_norm": 0.9774929881095886, "learning_rate": 0.00019707354516111173, "loss": 4.6641, "step": 300650 }, { "epoch": 0.5611396341022489, "grad_norm": 1.0798308849334717, "learning_rate": 0.0001970725362041152, "loss": 4.433, "step": 300700 }, { "epoch": 0.5612329396616274, "grad_norm": 1.1354361772537231, "learning_rate": 0.0001970715270758027, "loss": 4.6163, "step": 300750 }, { "epoch": 0.5613262452210058, "grad_norm": 1.0032528638839722, "learning_rate": 0.00019707051777617603, "loss": 4.5807, "step": 300800 }, { "epoch": 0.5614195507803844, "grad_norm": 0.8227291107177734, "learning_rate": 0.00019706950830523695, "loss": 4.6253, "step": 300850 }, { "epoch": 0.5615128563397629, "grad_norm": 1.0578374862670898, "learning_rate": 0.00019706849866298732, "loss": 4.483, "step": 300900 }, { "epoch": 0.5616061618991414, "grad_norm": 0.8574528694152832, "learning_rate": 0.00019706748884942883, "loss": 4.7146, "step": 300950 }, { "epoch": 0.5616994674585198, "grad_norm": 1.0092363357543945, "learning_rate": 0.00019706647886456332, "loss": 4.4874, "step": 301000 }, { "epoch": 0.5617927730178983, "grad_norm": 1.14267098903656, "learning_rate": 0.00019706546870839256, "loss": 4.7584, "step": 301050 }, { "epoch": 0.5618860785772768, "grad_norm": 0.9092347621917725, "learning_rate": 0.0001970644583809183, "loss": 4.4427, "step": 301100 }, { "epoch": 0.5619793841366553, "grad_norm": 0.960862934589386, "learning_rate": 0.00019706344788214238, "loss": 4.6718, "step": 301150 }, { "epoch": 0.5620726896960339, "grad_norm": 0.8959313631057739, "learning_rate": 0.00019706243721206654, "loss": 4.5922, "step": 301200 }, { "epoch": 0.5621659952554123, "grad_norm": 0.8489114046096802, "learning_rate": 0.00019706142637069255, "loss": 4.7201, "step": 301250 }, { "epoch": 0.5622593008147908, "grad_norm": 1.1516261100769043, "learning_rate": 0.00019706041535802227, "loss": 4.5853, "step": 301300 }, { "epoch": 0.5623526063741693, "grad_norm": 1.163648009300232, "learning_rate": 0.00019705940417405742, "loss": 4.5929, "step": 301350 }, { "epoch": 0.5624459119335478, "grad_norm": 0.963965654373169, "learning_rate": 0.0001970583928187998, "loss": 4.5252, "step": 301400 }, { "epoch": 0.5625392174929262, "grad_norm": 1.6231391429901123, "learning_rate": 0.00019705738129225118, "loss": 4.5953, "step": 301450 }, { "epoch": 0.5626325230523047, "grad_norm": 0.8425366878509521, "learning_rate": 0.00019705636959441335, "loss": 4.7018, "step": 301500 }, { "epoch": 0.5627258286116833, "grad_norm": 1.016891598701477, "learning_rate": 0.00019705535772528813, "loss": 4.5982, "step": 301550 }, { "epoch": 0.5628191341710618, "grad_norm": 1.1888262033462524, "learning_rate": 0.0001970543456848773, "loss": 4.6413, "step": 301600 }, { "epoch": 0.5629124397304403, "grad_norm": 1.064733624458313, "learning_rate": 0.00019705333347318257, "loss": 4.5598, "step": 301650 }, { "epoch": 0.5630057452898187, "grad_norm": 0.8338524103164673, "learning_rate": 0.00019705232109020583, "loss": 4.4923, "step": 301700 }, { "epoch": 0.5630990508491972, "grad_norm": 1.349608302116394, "learning_rate": 0.0001970513085359488, "loss": 4.5175, "step": 301750 }, { "epoch": 0.5631923564085757, "grad_norm": 0.7603903412818909, "learning_rate": 0.00019705029581041327, "loss": 4.5393, "step": 301800 }, { "epoch": 0.5632856619679542, "grad_norm": 1.117794394493103, "learning_rate": 0.00019704928291360108, "loss": 4.4936, "step": 301850 }, { "epoch": 0.5633789675273327, "grad_norm": 1.124933123588562, "learning_rate": 0.00019704826984551398, "loss": 4.5207, "step": 301900 }, { "epoch": 0.5634722730867112, "grad_norm": 1.0115653276443481, "learning_rate": 0.00019704725660615372, "loss": 4.7858, "step": 301950 }, { "epoch": 0.5635655786460897, "grad_norm": 0.9481130838394165, "learning_rate": 0.0001970462431955222, "loss": 4.6487, "step": 302000 }, { "epoch": 0.5636588842054682, "grad_norm": 1.1163173913955688, "learning_rate": 0.00019704522961362105, "loss": 4.6169, "step": 302050 }, { "epoch": 0.5637521897648466, "grad_norm": 1.0677237510681152, "learning_rate": 0.0001970442158604522, "loss": 4.8105, "step": 302100 }, { "epoch": 0.5638454953242251, "grad_norm": 1.0545969009399414, "learning_rate": 0.00019704320193601738, "loss": 4.4465, "step": 302150 }, { "epoch": 0.5639388008836036, "grad_norm": 1.149702548980713, "learning_rate": 0.0001970421878403184, "loss": 4.4358, "step": 302200 }, { "epoch": 0.5640321064429822, "grad_norm": 1.150503158569336, "learning_rate": 0.000197041173573357, "loss": 4.7382, "step": 302250 }, { "epoch": 0.5641254120023607, "grad_norm": 0.8477134704589844, "learning_rate": 0.000197040159135135, "loss": 4.6298, "step": 302300 }, { "epoch": 0.5642187175617391, "grad_norm": 1.2105333805084229, "learning_rate": 0.00019703914452565423, "loss": 4.4616, "step": 302350 }, { "epoch": 0.5643120231211176, "grad_norm": 0.7254282832145691, "learning_rate": 0.0001970381297449164, "loss": 4.5311, "step": 302400 }, { "epoch": 0.5644053286804961, "grad_norm": 0.9602898359298706, "learning_rate": 0.00019703711479292338, "loss": 4.6499, "step": 302450 }, { "epoch": 0.5644986342398746, "grad_norm": 1.1957898139953613, "learning_rate": 0.0001970360996696769, "loss": 4.6785, "step": 302500 }, { "epoch": 0.564591939799253, "grad_norm": 1.1815950870513916, "learning_rate": 0.0001970350843751788, "loss": 4.526, "step": 302550 }, { "epoch": 0.5646852453586316, "grad_norm": 0.9010161757469177, "learning_rate": 0.00019703406890943085, "loss": 4.6529, "step": 302600 }, { "epoch": 0.5647785509180101, "grad_norm": 0.8498725891113281, "learning_rate": 0.00019703305327243485, "loss": 4.4077, "step": 302650 }, { "epoch": 0.5648718564773886, "grad_norm": 0.838350236415863, "learning_rate": 0.00019703203746419254, "loss": 4.5618, "step": 302700 }, { "epoch": 0.564965162036767, "grad_norm": 1.0802860260009766, "learning_rate": 0.0001970310214847058, "loss": 4.6637, "step": 302750 }, { "epoch": 0.5650584675961455, "grad_norm": 1.061564564704895, "learning_rate": 0.00019703000533397635, "loss": 4.5875, "step": 302800 }, { "epoch": 0.565151773155524, "grad_norm": 1.123292326927185, "learning_rate": 0.00019702898901200602, "loss": 4.7532, "step": 302850 }, { "epoch": 0.5652450787149025, "grad_norm": 0.901685357093811, "learning_rate": 0.00019702797251879662, "loss": 4.6985, "step": 302900 }, { "epoch": 0.5653383842742811, "grad_norm": 0.8660471439361572, "learning_rate": 0.0001970269558543499, "loss": 4.7795, "step": 302950 }, { "epoch": 0.5654316898336595, "grad_norm": 1.0963807106018066, "learning_rate": 0.00019702593901866768, "loss": 4.675, "step": 303000 }, { "epoch": 0.5654316898336595, "eval_loss": 4.748389720916748, "eval_runtime": 230.9537, "eval_samples_per_second": 11.292, "eval_steps_per_second": 11.292, "eval_tts_loss": 7.572273884011444, "step": 303000 }, { "epoch": 0.565524995393038, "grad_norm": 1.2750355005264282, "learning_rate": 0.00019702492201175177, "loss": 4.7116, "step": 303050 }, { "epoch": 0.5656183009524165, "grad_norm": 0.9990673661231995, "learning_rate": 0.0001970239048336039, "loss": 4.6189, "step": 303100 }, { "epoch": 0.565711606511795, "grad_norm": 1.0602651834487915, "learning_rate": 0.00019702288748422596, "loss": 4.8896, "step": 303150 }, { "epoch": 0.5658049120711734, "grad_norm": 0.9759448170661926, "learning_rate": 0.00019702186996361967, "loss": 4.6535, "step": 303200 }, { "epoch": 0.5658982176305519, "grad_norm": 0.7759549021720886, "learning_rate": 0.00019702085227178687, "loss": 4.63, "step": 303250 }, { "epoch": 0.5659915231899305, "grad_norm": 1.1305766105651855, "learning_rate": 0.0001970198344087293, "loss": 4.5686, "step": 303300 }, { "epoch": 0.566084828749309, "grad_norm": 1.027217149734497, "learning_rate": 0.0001970188163744488, "loss": 4.5261, "step": 303350 }, { "epoch": 0.5661781343086875, "grad_norm": 0.970268726348877, "learning_rate": 0.0001970177981689472, "loss": 4.6596, "step": 303400 }, { "epoch": 0.5662714398680659, "grad_norm": 1.0845476388931274, "learning_rate": 0.00019701677979222623, "loss": 4.645, "step": 303450 }, { "epoch": 0.5663647454274444, "grad_norm": 1.0397199392318726, "learning_rate": 0.00019701576124428772, "loss": 4.6715, "step": 303500 }, { "epoch": 0.5664580509868229, "grad_norm": 1.0684022903442383, "learning_rate": 0.00019701474252513347, "loss": 4.5604, "step": 303550 }, { "epoch": 0.5665513565462014, "grad_norm": 0.8587548732757568, "learning_rate": 0.00019701372363476523, "loss": 4.4554, "step": 303600 }, { "epoch": 0.56664466210558, "grad_norm": 1.0176758766174316, "learning_rate": 0.0001970127045731849, "loss": 4.6871, "step": 303650 }, { "epoch": 0.5667379676649584, "grad_norm": 0.9839355945587158, "learning_rate": 0.00019701168534039416, "loss": 4.4283, "step": 303700 }, { "epoch": 0.5668312732243369, "grad_norm": 1.0790977478027344, "learning_rate": 0.00019701066593639488, "loss": 4.5313, "step": 303750 }, { "epoch": 0.5669245787837154, "grad_norm": 1.2821701765060425, "learning_rate": 0.00019700964636118885, "loss": 4.6008, "step": 303800 }, { "epoch": 0.5670178843430939, "grad_norm": 0.9878072738647461, "learning_rate": 0.00019700862661477787, "loss": 4.3788, "step": 303850 }, { "epoch": 0.5671111899024723, "grad_norm": 1.4116833209991455, "learning_rate": 0.00019700760669716375, "loss": 4.5283, "step": 303900 }, { "epoch": 0.5672044954618508, "grad_norm": 0.9004996418952942, "learning_rate": 0.00019700658660834824, "loss": 4.7421, "step": 303950 }, { "epoch": 0.5672978010212294, "grad_norm": 1.0024306774139404, "learning_rate": 0.00019700556634833318, "loss": 4.7196, "step": 304000 }, { "epoch": 0.5673911065806079, "grad_norm": 1.0337103605270386, "learning_rate": 0.00019700454591712037, "loss": 4.5365, "step": 304050 }, { "epoch": 0.5674844121399863, "grad_norm": 1.1965677738189697, "learning_rate": 0.0001970035253147116, "loss": 4.4689, "step": 304100 }, { "epoch": 0.5675777176993648, "grad_norm": 0.9932730197906494, "learning_rate": 0.00019700250454110868, "loss": 4.7301, "step": 304150 }, { "epoch": 0.5676710232587433, "grad_norm": 0.9931434392929077, "learning_rate": 0.00019700148359631342, "loss": 4.4457, "step": 304200 }, { "epoch": 0.5677643288181218, "grad_norm": 0.8123148083686829, "learning_rate": 0.00019700046248032757, "loss": 4.4538, "step": 304250 }, { "epoch": 0.5678576343775003, "grad_norm": 1.0292284488677979, "learning_rate": 0.000196999441193153, "loss": 4.7729, "step": 304300 }, { "epoch": 0.5679509399368788, "grad_norm": 1.213063359260559, "learning_rate": 0.0001969984197347915, "loss": 4.584, "step": 304350 }, { "epoch": 0.5680442454962573, "grad_norm": 0.8655990362167358, "learning_rate": 0.0001969973981052448, "loss": 4.497, "step": 304400 }, { "epoch": 0.5681375510556358, "grad_norm": 1.1007928848266602, "learning_rate": 0.00019699637630451485, "loss": 4.4047, "step": 304450 }, { "epoch": 0.5682308566150143, "grad_norm": 1.1282696723937988, "learning_rate": 0.0001969953543326033, "loss": 4.575, "step": 304500 }, { "epoch": 0.5683241621743927, "grad_norm": 1.417205810546875, "learning_rate": 0.00019699433218951201, "loss": 4.6682, "step": 304550 }, { "epoch": 0.5684174677337712, "grad_norm": 1.053480863571167, "learning_rate": 0.00019699330987524283, "loss": 4.4982, "step": 304600 }, { "epoch": 0.5685107732931497, "grad_norm": 0.7026126980781555, "learning_rate": 0.00019699228738979748, "loss": 4.6107, "step": 304650 }, { "epoch": 0.5686040788525283, "grad_norm": 0.6384779214859009, "learning_rate": 0.00019699126473317783, "loss": 4.4579, "step": 304700 }, { "epoch": 0.5686973844119068, "grad_norm": 1.152971625328064, "learning_rate": 0.0001969902419053857, "loss": 4.6327, "step": 304750 }, { "epoch": 0.5687906899712852, "grad_norm": 1.2194347381591797, "learning_rate": 0.0001969892189064228, "loss": 4.7345, "step": 304800 }, { "epoch": 0.5688839955306637, "grad_norm": 0.9324080944061279, "learning_rate": 0.00019698819573629102, "loss": 4.6878, "step": 304850 }, { "epoch": 0.5689773010900422, "grad_norm": 1.0412653684616089, "learning_rate": 0.00019698717239499214, "loss": 4.6517, "step": 304900 }, { "epoch": 0.5690706066494207, "grad_norm": 1.0900930166244507, "learning_rate": 0.00019698614888252798, "loss": 4.4744, "step": 304950 }, { "epoch": 0.5691639122087991, "grad_norm": 1.0790338516235352, "learning_rate": 0.0001969851251989003, "loss": 4.5466, "step": 305000 }, { "epoch": 0.5692572177681777, "grad_norm": 1.126724123954773, "learning_rate": 0.00019698410134411096, "loss": 4.5577, "step": 305050 }, { "epoch": 0.5693505233275562, "grad_norm": 1.2070733308792114, "learning_rate": 0.00019698307731816174, "loss": 4.6863, "step": 305100 }, { "epoch": 0.5694438288869347, "grad_norm": 0.6844452023506165, "learning_rate": 0.00019698205312105444, "loss": 4.4842, "step": 305150 }, { "epoch": 0.5695371344463132, "grad_norm": 1.2279843091964722, "learning_rate": 0.0001969810287527909, "loss": 4.537, "step": 305200 }, { "epoch": 0.5696304400056916, "grad_norm": 0.9124521017074585, "learning_rate": 0.00019698000421337292, "loss": 4.3272, "step": 305250 }, { "epoch": 0.5697237455650701, "grad_norm": 1.001929521560669, "learning_rate": 0.00019697897950280227, "loss": 4.6252, "step": 305300 }, { "epoch": 0.5698170511244486, "grad_norm": 1.156294584274292, "learning_rate": 0.00019697795462108081, "loss": 4.6983, "step": 305350 }, { "epoch": 0.5699103566838272, "grad_norm": 1.0211153030395508, "learning_rate": 0.00019697692956821032, "loss": 4.4763, "step": 305400 }, { "epoch": 0.5700036622432056, "grad_norm": 1.119654893875122, "learning_rate": 0.0001969759043441926, "loss": 4.6318, "step": 305450 }, { "epoch": 0.5700969678025841, "grad_norm": 1.1528464555740356, "learning_rate": 0.00019697487894902946, "loss": 4.5226, "step": 305500 }, { "epoch": 0.5701902733619626, "grad_norm": 0.6716154217720032, "learning_rate": 0.00019697385338272275, "loss": 4.6249, "step": 305550 }, { "epoch": 0.5702835789213411, "grad_norm": 0.9842194318771362, "learning_rate": 0.00019697282764527424, "loss": 4.5753, "step": 305600 }, { "epoch": 0.5703768844807195, "grad_norm": 1.140363097190857, "learning_rate": 0.00019697180173668574, "loss": 4.5553, "step": 305650 }, { "epoch": 0.570470190040098, "grad_norm": 1.0138086080551147, "learning_rate": 0.0001969707756569591, "loss": 4.3587, "step": 305700 }, { "epoch": 0.5705634955994766, "grad_norm": 2.0012075901031494, "learning_rate": 0.00019696974940609606, "loss": 4.5402, "step": 305750 }, { "epoch": 0.5706568011588551, "grad_norm": 1.1730917692184448, "learning_rate": 0.0001969687229840985, "loss": 4.687, "step": 305800 }, { "epoch": 0.5707501067182336, "grad_norm": 1.1034109592437744, "learning_rate": 0.0001969676963909682, "loss": 4.6527, "step": 305850 }, { "epoch": 0.570843412277612, "grad_norm": 0.922483503818512, "learning_rate": 0.00019696666962670703, "loss": 4.4858, "step": 305900 }, { "epoch": 0.5709367178369905, "grad_norm": 1.0203330516815186, "learning_rate": 0.0001969656426913167, "loss": 4.7398, "step": 305950 }, { "epoch": 0.571030023396369, "grad_norm": 0.7679086923599243, "learning_rate": 0.00019696461558479906, "loss": 4.5363, "step": 306000 }, { "epoch": 0.571030023396369, "eval_loss": 4.761473178863525, "eval_runtime": 231.9504, "eval_samples_per_second": 11.244, "eval_steps_per_second": 11.244, "eval_tts_loss": 7.575939671752083, "step": 306000 }, { "epoch": 0.5711233289557475, "grad_norm": 1.031692385673523, "learning_rate": 0.00019696358830715598, "loss": 4.6841, "step": 306050 }, { "epoch": 0.5712166345151259, "grad_norm": 0.9715311527252197, "learning_rate": 0.0001969625608583892, "loss": 4.7471, "step": 306100 }, { "epoch": 0.5713099400745045, "grad_norm": 0.7275256514549255, "learning_rate": 0.00019696153323850056, "loss": 4.5227, "step": 306150 }, { "epoch": 0.571403245633883, "grad_norm": 0.9675799012184143, "learning_rate": 0.00019696050544749187, "loss": 4.3886, "step": 306200 }, { "epoch": 0.5714965511932615, "grad_norm": 0.7808157205581665, "learning_rate": 0.00019695947748536496, "loss": 4.3582, "step": 306250 }, { "epoch": 0.57158985675264, "grad_norm": 1.1058979034423828, "learning_rate": 0.00019695844935212163, "loss": 4.6779, "step": 306300 }, { "epoch": 0.5716831623120184, "grad_norm": 1.121512532234192, "learning_rate": 0.0001969574210477637, "loss": 4.8738, "step": 306350 }, { "epoch": 0.5717764678713969, "grad_norm": 0.8467114567756653, "learning_rate": 0.00019695639257229298, "loss": 4.4983, "step": 306400 }, { "epoch": 0.5718697734307754, "grad_norm": 0.9501798748970032, "learning_rate": 0.00019695536392571128, "loss": 4.6996, "step": 306450 }, { "epoch": 0.571963078990154, "grad_norm": 1.0179754495620728, "learning_rate": 0.00019695433510802043, "loss": 4.4817, "step": 306500 }, { "epoch": 0.5720563845495324, "grad_norm": 1.097285509109497, "learning_rate": 0.0001969533061192222, "loss": 4.5062, "step": 306550 }, { "epoch": 0.5721496901089109, "grad_norm": 1.1028809547424316, "learning_rate": 0.0001969522769593185, "loss": 4.4032, "step": 306600 }, { "epoch": 0.5722429956682894, "grad_norm": 0.931887149810791, "learning_rate": 0.00019695124762831106, "loss": 4.5055, "step": 306650 }, { "epoch": 0.5723363012276679, "grad_norm": 1.1663527488708496, "learning_rate": 0.00019695021812620177, "loss": 4.5626, "step": 306700 }, { "epoch": 0.5724296067870464, "grad_norm": 1.0238298177719116, "learning_rate": 0.00019694918845299236, "loss": 4.5973, "step": 306750 }, { "epoch": 0.5725229123464248, "grad_norm": 1.0825369358062744, "learning_rate": 0.0001969481586086847, "loss": 4.5194, "step": 306800 }, { "epoch": 0.5726162179058034, "grad_norm": 1.1342600584030151, "learning_rate": 0.0001969471285932806, "loss": 4.768, "step": 306850 }, { "epoch": 0.5727095234651819, "grad_norm": 1.0217125415802002, "learning_rate": 0.0001969460984067819, "loss": 4.7457, "step": 306900 }, { "epoch": 0.5728028290245604, "grad_norm": 1.3335061073303223, "learning_rate": 0.00019694506804919035, "loss": 4.6568, "step": 306950 }, { "epoch": 0.5728961345839388, "grad_norm": 0.8185517191886902, "learning_rate": 0.00019694403752050784, "loss": 4.6078, "step": 307000 }, { "epoch": 0.5729894401433173, "grad_norm": 1.478460669517517, "learning_rate": 0.00019694300682073616, "loss": 4.6365, "step": 307050 }, { "epoch": 0.5730827457026958, "grad_norm": 1.2282335758209229, "learning_rate": 0.00019694197594987713, "loss": 4.5315, "step": 307100 }, { "epoch": 0.5731760512620743, "grad_norm": 0.8165243864059448, "learning_rate": 0.00019694094490793255, "loss": 4.5083, "step": 307150 }, { "epoch": 0.5732693568214529, "grad_norm": 1.2821803092956543, "learning_rate": 0.0001969399136949043, "loss": 4.6808, "step": 307200 }, { "epoch": 0.5733626623808313, "grad_norm": 0.8003031015396118, "learning_rate": 0.00019693888231079413, "loss": 4.4266, "step": 307250 }, { "epoch": 0.5734559679402098, "grad_norm": 1.1053504943847656, "learning_rate": 0.00019693785075560388, "loss": 4.5509, "step": 307300 }, { "epoch": 0.5735492734995883, "grad_norm": 1.2611793279647827, "learning_rate": 0.00019693681902933541, "loss": 4.513, "step": 307350 }, { "epoch": 0.5736425790589668, "grad_norm": 0.9037719964981079, "learning_rate": 0.0001969357871319905, "loss": 4.367, "step": 307400 }, { "epoch": 0.5737358846183452, "grad_norm": 0.9977772235870361, "learning_rate": 0.00019693475506357098, "loss": 4.6679, "step": 307450 }, { "epoch": 0.5738291901777237, "grad_norm": 0.8833919167518616, "learning_rate": 0.00019693372282407866, "loss": 4.6624, "step": 307500 }, { "epoch": 0.5739224957371023, "grad_norm": 1.1849719285964966, "learning_rate": 0.0001969326904135154, "loss": 4.6164, "step": 307550 }, { "epoch": 0.5740158012964808, "grad_norm": 1.1160576343536377, "learning_rate": 0.00019693165783188298, "loss": 4.5444, "step": 307600 }, { "epoch": 0.5741091068558593, "grad_norm": 0.7999420166015625, "learning_rate": 0.00019693062507918324, "loss": 4.4761, "step": 307650 }, { "epoch": 0.5742024124152377, "grad_norm": 1.1961227655410767, "learning_rate": 0.000196929592155418, "loss": 4.5654, "step": 307700 }, { "epoch": 0.5742957179746162, "grad_norm": 1.1302611827850342, "learning_rate": 0.0001969285590605891, "loss": 4.532, "step": 307750 }, { "epoch": 0.5743890235339947, "grad_norm": 1.0442041158676147, "learning_rate": 0.00019692752579469834, "loss": 4.4396, "step": 307800 }, { "epoch": 0.5744823290933732, "grad_norm": 0.9419299364089966, "learning_rate": 0.00019692649235774755, "loss": 4.6001, "step": 307850 }, { "epoch": 0.5745756346527517, "grad_norm": 1.114232063293457, "learning_rate": 0.00019692545874973858, "loss": 4.5381, "step": 307900 }, { "epoch": 0.5746689402121302, "grad_norm": 1.1545286178588867, "learning_rate": 0.0001969244249706732, "loss": 4.5677, "step": 307950 }, { "epoch": 0.5747622457715087, "grad_norm": 1.051513433456421, "learning_rate": 0.00019692339102055328, "loss": 4.5663, "step": 308000 }, { "epoch": 0.5748555513308872, "grad_norm": 1.200150489807129, "learning_rate": 0.0001969223568993806, "loss": 4.5208, "step": 308050 }, { "epoch": 0.5749488568902656, "grad_norm": 0.8030568361282349, "learning_rate": 0.00019692132260715702, "loss": 4.344, "step": 308100 }, { "epoch": 0.5750421624496441, "grad_norm": 1.1953867673873901, "learning_rate": 0.00019692028814388438, "loss": 4.4139, "step": 308150 }, { "epoch": 0.5751354680090226, "grad_norm": 0.8794069886207581, "learning_rate": 0.00019691925350956449, "loss": 4.553, "step": 308200 }, { "epoch": 0.5752287735684012, "grad_norm": 1.015227198600769, "learning_rate": 0.00019691821870419917, "loss": 4.7194, "step": 308250 }, { "epoch": 0.5753220791277797, "grad_norm": 1.3813424110412598, "learning_rate": 0.00019691718372779024, "loss": 4.4443, "step": 308300 }, { "epoch": 0.5754153846871581, "grad_norm": 0.9208297729492188, "learning_rate": 0.00019691614858033954, "loss": 4.6475, "step": 308350 }, { "epoch": 0.5755086902465366, "grad_norm": 1.0939719676971436, "learning_rate": 0.00019691511326184888, "loss": 4.5593, "step": 308400 }, { "epoch": 0.5756019958059151, "grad_norm": 1.1301640272140503, "learning_rate": 0.0001969140777723201, "loss": 4.4633, "step": 308450 }, { "epoch": 0.5756953013652936, "grad_norm": 1.2512931823730469, "learning_rate": 0.000196913042111755, "loss": 4.6935, "step": 308500 }, { "epoch": 0.575788606924672, "grad_norm": 1.0407778024673462, "learning_rate": 0.0001969120062801555, "loss": 4.5598, "step": 308550 }, { "epoch": 0.5758819124840506, "grad_norm": 1.2233508825302124, "learning_rate": 0.0001969109702775233, "loss": 4.635, "step": 308600 }, { "epoch": 0.5759752180434291, "grad_norm": 1.032896637916565, "learning_rate": 0.0001969099341038603, "loss": 4.7609, "step": 308650 }, { "epoch": 0.5760685236028076, "grad_norm": 0.5946270227432251, "learning_rate": 0.00019690889775916834, "loss": 4.6175, "step": 308700 }, { "epoch": 0.5761618291621861, "grad_norm": 0.6330205798149109, "learning_rate": 0.00019690786124344921, "loss": 4.5726, "step": 308750 }, { "epoch": 0.5762551347215645, "grad_norm": 1.0800895690917969, "learning_rate": 0.00019690682455670476, "loss": 4.4216, "step": 308800 }, { "epoch": 0.576348440280943, "grad_norm": 0.8748204708099365, "learning_rate": 0.00019690578769893682, "loss": 4.668, "step": 308850 }, { "epoch": 0.5764417458403215, "grad_norm": 0.9928598403930664, "learning_rate": 0.00019690475067014721, "loss": 4.5196, "step": 308900 }, { "epoch": 0.5765350513997001, "grad_norm": 1.1503949165344238, "learning_rate": 0.00019690371347033775, "loss": 4.5349, "step": 308950 }, { "epoch": 0.5766283569590785, "grad_norm": 0.8418891429901123, "learning_rate": 0.00019690267609951035, "loss": 4.5646, "step": 309000 }, { "epoch": 0.5766283569590785, "eval_loss": 4.752696990966797, "eval_runtime": 228.2633, "eval_samples_per_second": 11.425, "eval_steps_per_second": 11.425, "eval_tts_loss": 7.530654103705157, "step": 309000 }, { "epoch": 0.576721662518457, "grad_norm": 0.988182544708252, "learning_rate": 0.0001969016385576667, "loss": 4.5647, "step": 309050 }, { "epoch": 0.5768149680778355, "grad_norm": 1.031363606452942, "learning_rate": 0.00019690060084480874, "loss": 4.3385, "step": 309100 }, { "epoch": 0.576908273637214, "grad_norm": 0.8367418646812439, "learning_rate": 0.00019689956296093828, "loss": 4.6036, "step": 309150 }, { "epoch": 0.5770015791965925, "grad_norm": 1.2426990270614624, "learning_rate": 0.00019689852490605713, "loss": 4.3699, "step": 309200 }, { "epoch": 0.5770948847559709, "grad_norm": 1.0288184881210327, "learning_rate": 0.0001968974866801671, "loss": 4.5097, "step": 309250 }, { "epoch": 0.5771881903153495, "grad_norm": 1.1623977422714233, "learning_rate": 0.00019689644828327013, "loss": 4.5612, "step": 309300 }, { "epoch": 0.577281495874728, "grad_norm": 1.1653733253479004, "learning_rate": 0.00019689540971536796, "loss": 4.5268, "step": 309350 }, { "epoch": 0.5773748014341065, "grad_norm": 1.2442301511764526, "learning_rate": 0.0001968943709764624, "loss": 4.5712, "step": 309400 }, { "epoch": 0.5774681069934849, "grad_norm": 1.1720430850982666, "learning_rate": 0.00019689333206655536, "loss": 4.6003, "step": 309450 }, { "epoch": 0.5775614125528634, "grad_norm": 1.1067763566970825, "learning_rate": 0.00019689229298564863, "loss": 4.699, "step": 309500 }, { "epoch": 0.5776547181122419, "grad_norm": 0.7484050989151001, "learning_rate": 0.000196891253733744, "loss": 4.6018, "step": 309550 }, { "epoch": 0.5777480236716204, "grad_norm": 1.0562546253204346, "learning_rate": 0.00019689021431084343, "loss": 4.714, "step": 309600 }, { "epoch": 0.577841329230999, "grad_norm": 1.1429967880249023, "learning_rate": 0.00019688917471694865, "loss": 4.6596, "step": 309650 }, { "epoch": 0.5779346347903774, "grad_norm": 1.0727895498275757, "learning_rate": 0.0001968881349520615, "loss": 4.6577, "step": 309700 }, { "epoch": 0.5780279403497559, "grad_norm": 0.8976140022277832, "learning_rate": 0.00019688709501618389, "loss": 4.6248, "step": 309750 }, { "epoch": 0.5781212459091344, "grad_norm": 0.8005908727645874, "learning_rate": 0.00019688605490931758, "loss": 4.6077, "step": 309800 }, { "epoch": 0.5782145514685129, "grad_norm": 1.1622130870819092, "learning_rate": 0.00019688501463146444, "loss": 4.6376, "step": 309850 }, { "epoch": 0.5783078570278913, "grad_norm": 0.9488914608955383, "learning_rate": 0.00019688397418262627, "loss": 4.384, "step": 309900 }, { "epoch": 0.5784011625872698, "grad_norm": 1.159583330154419, "learning_rate": 0.00019688293356280495, "loss": 4.7019, "step": 309950 }, { "epoch": 0.5784944681466484, "grad_norm": 0.9173402190208435, "learning_rate": 0.00019688189277200228, "loss": 4.5651, "step": 310000 }, { "epoch": 0.5785877737060269, "grad_norm": 1.0882700681686401, "learning_rate": 0.00019688085181022012, "loss": 4.3936, "step": 310050 }, { "epoch": 0.5786810792654054, "grad_norm": 1.401253581047058, "learning_rate": 0.00019687981067746037, "loss": 4.5729, "step": 310100 }, { "epoch": 0.5787743848247838, "grad_norm": 1.2113391160964966, "learning_rate": 0.00019687876937372472, "loss": 4.4746, "step": 310150 }, { "epoch": 0.5788676903841623, "grad_norm": 0.9598744511604309, "learning_rate": 0.00019687772789901513, "loss": 4.6706, "step": 310200 }, { "epoch": 0.5789609959435408, "grad_norm": 1.0837336778640747, "learning_rate": 0.00019687668625333337, "loss": 4.5265, "step": 310250 }, { "epoch": 0.5790543015029193, "grad_norm": 1.3954722881317139, "learning_rate": 0.00019687564443668132, "loss": 4.3927, "step": 310300 }, { "epoch": 0.5791476070622978, "grad_norm": 0.8956559896469116, "learning_rate": 0.0001968746024490608, "loss": 4.6382, "step": 310350 }, { "epoch": 0.5792409126216763, "grad_norm": 1.6203763484954834, "learning_rate": 0.00019687356029047365, "loss": 4.6029, "step": 310400 }, { "epoch": 0.5793342181810548, "grad_norm": 0.983693540096283, "learning_rate": 0.00019687251796092175, "loss": 4.5395, "step": 310450 }, { "epoch": 0.5794275237404333, "grad_norm": 1.05988347530365, "learning_rate": 0.00019687147546040685, "loss": 4.3837, "step": 310500 }, { "epoch": 0.5795208292998117, "grad_norm": 1.071925163269043, "learning_rate": 0.00019687043278893084, "loss": 4.553, "step": 310550 }, { "epoch": 0.5796141348591902, "grad_norm": 1.0153021812438965, "learning_rate": 0.0001968693899464956, "loss": 4.3998, "step": 310600 }, { "epoch": 0.5797074404185687, "grad_norm": 1.4140474796295166, "learning_rate": 0.00019686834693310293, "loss": 4.5908, "step": 310650 }, { "epoch": 0.5798007459779473, "grad_norm": 1.0499241352081299, "learning_rate": 0.00019686730374875465, "loss": 4.6639, "step": 310700 }, { "epoch": 0.5798940515373258, "grad_norm": 0.9955544471740723, "learning_rate": 0.00019686626039345262, "loss": 4.8298, "step": 310750 }, { "epoch": 0.5799873570967042, "grad_norm": 1.2786093950271606, "learning_rate": 0.00019686521686719868, "loss": 4.4873, "step": 310800 }, { "epoch": 0.5800806626560827, "grad_norm": 0.96354740858078, "learning_rate": 0.0001968641731699947, "loss": 4.5876, "step": 310850 }, { "epoch": 0.5801739682154612, "grad_norm": 1.0890089273452759, "learning_rate": 0.00019686312930184246, "loss": 4.6488, "step": 310900 }, { "epoch": 0.5802672737748397, "grad_norm": 1.1017330884933472, "learning_rate": 0.00019686208526274385, "loss": 4.6347, "step": 310950 }, { "epoch": 0.5803605793342181, "grad_norm": 1.126384973526001, "learning_rate": 0.00019686104105270073, "loss": 4.5384, "step": 311000 }, { "epoch": 0.5804538848935967, "grad_norm": 0.9543178081512451, "learning_rate": 0.0001968599966717149, "loss": 4.4462, "step": 311050 }, { "epoch": 0.5805471904529752, "grad_norm": 1.119261622428894, "learning_rate": 0.00019685895211978824, "loss": 4.7619, "step": 311100 }, { "epoch": 0.5806404960123537, "grad_norm": 0.9374808073043823, "learning_rate": 0.00019685790739692256, "loss": 4.5444, "step": 311150 }, { "epoch": 0.5807338015717322, "grad_norm": 0.8393676280975342, "learning_rate": 0.0001968568625031197, "loss": 4.7381, "step": 311200 }, { "epoch": 0.5808271071311106, "grad_norm": 0.9694864153862, "learning_rate": 0.00019685581743838154, "loss": 4.584, "step": 311250 }, { "epoch": 0.5809204126904891, "grad_norm": 0.9850889444351196, "learning_rate": 0.00019685477220270987, "loss": 4.578, "step": 311300 }, { "epoch": 0.5810137182498676, "grad_norm": 1.149882197380066, "learning_rate": 0.0001968537267961066, "loss": 4.6047, "step": 311350 }, { "epoch": 0.5811070238092461, "grad_norm": 0.8860505819320679, "learning_rate": 0.00019685268121857354, "loss": 4.6453, "step": 311400 }, { "epoch": 0.5812003293686246, "grad_norm": 0.9555177092552185, "learning_rate": 0.00019685163547011257, "loss": 4.5644, "step": 311450 }, { "epoch": 0.5812936349280031, "grad_norm": 0.7401108145713806, "learning_rate": 0.00019685058955072547, "loss": 4.5516, "step": 311500 }, { "epoch": 0.5813869404873816, "grad_norm": 0.8870165944099426, "learning_rate": 0.00019684954346041412, "loss": 4.4763, "step": 311550 }, { "epoch": 0.5814802460467601, "grad_norm": 1.1874957084655762, "learning_rate": 0.00019684849719918038, "loss": 4.3779, "step": 311600 }, { "epoch": 0.5815735516061386, "grad_norm": 0.9011445045471191, "learning_rate": 0.00019684745076702608, "loss": 4.7085, "step": 311650 }, { "epoch": 0.581666857165517, "grad_norm": 1.0121692419052124, "learning_rate": 0.00019684640416395304, "loss": 4.6841, "step": 311700 }, { "epoch": 0.5817601627248955, "grad_norm": 1.0623122453689575, "learning_rate": 0.00019684535738996318, "loss": 4.337, "step": 311750 }, { "epoch": 0.5818534682842741, "grad_norm": 1.0233830213546753, "learning_rate": 0.00019684431044505828, "loss": 4.4701, "step": 311800 }, { "epoch": 0.5819467738436526, "grad_norm": 0.7241932153701782, "learning_rate": 0.0001968432633292402, "loss": 4.3979, "step": 311850 }, { "epoch": 0.582040079403031, "grad_norm": 1.0669363737106323, "learning_rate": 0.00019684221604251084, "loss": 4.6099, "step": 311900 }, { "epoch": 0.5821333849624095, "grad_norm": 1.0737048387527466, "learning_rate": 0.00019684116858487198, "loss": 4.6644, "step": 311950 }, { "epoch": 0.582226690521788, "grad_norm": 1.3133255243301392, "learning_rate": 0.0001968401209563255, "loss": 4.4478, "step": 312000 }, { "epoch": 0.582226690521788, "eval_loss": 4.749062538146973, "eval_runtime": 231.04, "eval_samples_per_second": 11.288, "eval_steps_per_second": 11.288, "eval_tts_loss": 7.521649108948243, "step": 312000 }, { "epoch": 0.5823199960811665, "grad_norm": 1.216781497001648, "learning_rate": 0.00019683907315687328, "loss": 4.5382, "step": 312050 }, { "epoch": 0.582413301640545, "grad_norm": 1.1261478662490845, "learning_rate": 0.00019683802518651709, "loss": 4.5286, "step": 312100 }, { "epoch": 0.5825066071999235, "grad_norm": 0.9089665412902832, "learning_rate": 0.00019683697704525884, "loss": 4.5094, "step": 312150 }, { "epoch": 0.582599912759302, "grad_norm": 0.9918016791343689, "learning_rate": 0.00019683592873310035, "loss": 4.7113, "step": 312200 }, { "epoch": 0.5826932183186805, "grad_norm": 1.0284326076507568, "learning_rate": 0.00019683488025004352, "loss": 4.6837, "step": 312250 }, { "epoch": 0.582786523878059, "grad_norm": 0.9192066192626953, "learning_rate": 0.00019683383159609013, "loss": 4.6697, "step": 312300 }, { "epoch": 0.5828798294374374, "grad_norm": 0.8472697138786316, "learning_rate": 0.0001968327827712421, "loss": 4.627, "step": 312350 }, { "epoch": 0.5829731349968159, "grad_norm": 1.0066317319869995, "learning_rate": 0.0001968317337755012, "loss": 4.6281, "step": 312400 }, { "epoch": 0.5830664405561944, "grad_norm": 0.9507394433021545, "learning_rate": 0.00019683068460886935, "loss": 4.6588, "step": 312450 }, { "epoch": 0.583159746115573, "grad_norm": 1.2607237100601196, "learning_rate": 0.0001968296352713484, "loss": 4.4912, "step": 312500 }, { "epoch": 0.5832530516749515, "grad_norm": 1.0062910318374634, "learning_rate": 0.00019682858576294015, "loss": 4.6857, "step": 312550 }, { "epoch": 0.5833463572343299, "grad_norm": 0.8089048862457275, "learning_rate": 0.0001968275360836465, "loss": 4.6043, "step": 312600 }, { "epoch": 0.5834396627937084, "grad_norm": 0.5660948753356934, "learning_rate": 0.0001968264862334693, "loss": 4.4691, "step": 312650 }, { "epoch": 0.5835329683530869, "grad_norm": 1.0028718709945679, "learning_rate": 0.00019682543621241035, "loss": 4.6358, "step": 312700 }, { "epoch": 0.5836262739124654, "grad_norm": 1.0709588527679443, "learning_rate": 0.0001968243860204716, "loss": 4.5166, "step": 312750 }, { "epoch": 0.5837195794718438, "grad_norm": 0.80045086145401, "learning_rate": 0.0001968233356576548, "loss": 4.611, "step": 312800 }, { "epoch": 0.5838128850312224, "grad_norm": 1.1337125301361084, "learning_rate": 0.00019682228512396188, "loss": 4.5636, "step": 312850 }, { "epoch": 0.5839061905906009, "grad_norm": 1.2380632162094116, "learning_rate": 0.00019682123441939464, "loss": 4.5597, "step": 312900 }, { "epoch": 0.5839994961499794, "grad_norm": 1.3010809421539307, "learning_rate": 0.00019682018354395497, "loss": 4.5734, "step": 312950 }, { "epoch": 0.5840928017093578, "grad_norm": 1.0648142099380493, "learning_rate": 0.0001968191324976447, "loss": 4.5486, "step": 313000 }, { "epoch": 0.5841861072687363, "grad_norm": 0.9465823173522949, "learning_rate": 0.00019681808128046567, "loss": 4.5559, "step": 313050 }, { "epoch": 0.5842794128281148, "grad_norm": 1.0052610635757446, "learning_rate": 0.00019681702989241982, "loss": 4.7299, "step": 313100 }, { "epoch": 0.5843727183874933, "grad_norm": 1.0187366008758545, "learning_rate": 0.00019681597833350894, "loss": 4.6535, "step": 313150 }, { "epoch": 0.5844660239468719, "grad_norm": 0.8939917087554932, "learning_rate": 0.00019681492660373487, "loss": 4.6603, "step": 313200 }, { "epoch": 0.5845593295062503, "grad_norm": 1.0251716375350952, "learning_rate": 0.00019681387470309948, "loss": 4.7555, "step": 313250 }, { "epoch": 0.5846526350656288, "grad_norm": 0.9094771146774292, "learning_rate": 0.00019681282263160467, "loss": 4.6101, "step": 313300 }, { "epoch": 0.5847459406250073, "grad_norm": 0.9584060907363892, "learning_rate": 0.00019681177038925223, "loss": 4.7022, "step": 313350 }, { "epoch": 0.5848392461843858, "grad_norm": 0.9828512668609619, "learning_rate": 0.00019681071797604408, "loss": 4.6973, "step": 313400 }, { "epoch": 0.5849325517437642, "grad_norm": 1.157536506652832, "learning_rate": 0.000196809665391982, "loss": 4.3445, "step": 313450 }, { "epoch": 0.5850258573031427, "grad_norm": 1.133435606956482, "learning_rate": 0.00019680861263706792, "loss": 4.5914, "step": 313500 }, { "epoch": 0.5851191628625213, "grad_norm": 0.9810517430305481, "learning_rate": 0.00019680755971130371, "loss": 4.4814, "step": 313550 }, { "epoch": 0.5852124684218998, "grad_norm": 0.7324022650718689, "learning_rate": 0.00019680650661469114, "loss": 4.5606, "step": 313600 }, { "epoch": 0.5853057739812783, "grad_norm": 1.1952625513076782, "learning_rate": 0.00019680545334723216, "loss": 4.6152, "step": 313650 }, { "epoch": 0.5853990795406567, "grad_norm": 1.237855315208435, "learning_rate": 0.00019680439990892853, "loss": 4.4655, "step": 313700 }, { "epoch": 0.5854923851000352, "grad_norm": 1.2508081197738647, "learning_rate": 0.0001968033462997822, "loss": 4.5237, "step": 313750 }, { "epoch": 0.5855856906594137, "grad_norm": 0.8722935318946838, "learning_rate": 0.00019680229251979503, "loss": 4.6543, "step": 313800 }, { "epoch": 0.5856789962187922, "grad_norm": 1.2657021284103394, "learning_rate": 0.0001968012385689688, "loss": 4.6198, "step": 313850 }, { "epoch": 0.5857723017781707, "grad_norm": 1.203904151916504, "learning_rate": 0.00019680018444730542, "loss": 4.5212, "step": 313900 }, { "epoch": 0.5858656073375492, "grad_norm": 1.0725904703140259, "learning_rate": 0.0001967991301548068, "loss": 4.5757, "step": 313950 }, { "epoch": 0.5859589128969277, "grad_norm": 1.0794929265975952, "learning_rate": 0.00019679807569147472, "loss": 4.3468, "step": 314000 }, { "epoch": 0.5860522184563062, "grad_norm": 1.049497127532959, "learning_rate": 0.00019679702105731103, "loss": 4.7022, "step": 314050 }, { "epoch": 0.5861455240156846, "grad_norm": 1.3520044088363647, "learning_rate": 0.00019679596625231768, "loss": 4.418, "step": 314100 }, { "epoch": 0.5862388295750631, "grad_norm": 1.1205198764801025, "learning_rate": 0.00019679491127649646, "loss": 4.5213, "step": 314150 }, { "epoch": 0.5863321351344416, "grad_norm": 0.9296178221702576, "learning_rate": 0.00019679385612984924, "loss": 4.578, "step": 314200 }, { "epoch": 0.5864254406938202, "grad_norm": 1.0487147569656372, "learning_rate": 0.0001967928008123779, "loss": 4.5632, "step": 314250 }, { "epoch": 0.5865187462531987, "grad_norm": 1.1871947050094604, "learning_rate": 0.0001967917453240843, "loss": 4.6728, "step": 314300 }, { "epoch": 0.5866120518125771, "grad_norm": 0.8288167715072632, "learning_rate": 0.0001967906896649703, "loss": 4.598, "step": 314350 }, { "epoch": 0.5867053573719556, "grad_norm": 1.000169277191162, "learning_rate": 0.0001967896338350378, "loss": 4.49, "step": 314400 }, { "epoch": 0.5867986629313341, "grad_norm": 1.1729748249053955, "learning_rate": 0.00019678857783428857, "loss": 4.6127, "step": 314450 }, { "epoch": 0.5868919684907126, "grad_norm": 1.2053636312484741, "learning_rate": 0.00019678752166272457, "loss": 4.5601, "step": 314500 }, { "epoch": 0.586985274050091, "grad_norm": 0.915949821472168, "learning_rate": 0.0001967864653203476, "loss": 4.4926, "step": 314550 }, { "epoch": 0.5870785796094696, "grad_norm": 1.1071586608886719, "learning_rate": 0.00019678540880715956, "loss": 4.5221, "step": 314600 }, { "epoch": 0.5871718851688481, "grad_norm": 0.8765965700149536, "learning_rate": 0.0001967843521231623, "loss": 4.4275, "step": 314650 }, { "epoch": 0.5872651907282266, "grad_norm": 1.1023027896881104, "learning_rate": 0.0001967832952683577, "loss": 4.5292, "step": 314700 }, { "epoch": 0.5873584962876051, "grad_norm": 1.0290693044662476, "learning_rate": 0.0001967822382427476, "loss": 4.6013, "step": 314750 }, { "epoch": 0.5874518018469835, "grad_norm": 1.0654975175857544, "learning_rate": 0.0001967811810463339, "loss": 4.6032, "step": 314800 }, { "epoch": 0.587545107406362, "grad_norm": 1.0474263429641724, "learning_rate": 0.00019678012367911843, "loss": 4.4041, "step": 314850 }, { "epoch": 0.5876384129657405, "grad_norm": 1.2105765342712402, "learning_rate": 0.00019677906614110307, "loss": 4.4584, "step": 314900 }, { "epoch": 0.5877317185251191, "grad_norm": 1.050726056098938, "learning_rate": 0.00019677800843228966, "loss": 4.5704, "step": 314950 }, { "epoch": 0.5878250240844975, "grad_norm": 0.7779209017753601, "learning_rate": 0.00019677695055268015, "loss": 4.6361, "step": 315000 }, { "epoch": 0.5878250240844975, "eval_loss": 4.74893856048584, "eval_runtime": 233.0687, "eval_samples_per_second": 11.19, "eval_steps_per_second": 11.19, "eval_tts_loss": 7.577803379387532, "step": 315000 }, { "epoch": 0.587918329643876, "grad_norm": 1.0922472476959229, "learning_rate": 0.0001967758925022763, "loss": 4.5707, "step": 315050 }, { "epoch": 0.5880116352032545, "grad_norm": 1.1798323392868042, "learning_rate": 0.00019677483428108005, "loss": 4.5508, "step": 315100 }, { "epoch": 0.588104940762633, "grad_norm": 1.1150181293487549, "learning_rate": 0.00019677377588909326, "loss": 4.6886, "step": 315150 }, { "epoch": 0.5881982463220115, "grad_norm": 1.1528747081756592, "learning_rate": 0.00019677271732631776, "loss": 4.5604, "step": 315200 }, { "epoch": 0.5882915518813899, "grad_norm": 1.16209876537323, "learning_rate": 0.00019677165859275544, "loss": 4.6568, "step": 315250 }, { "epoch": 0.5883848574407685, "grad_norm": 1.0389351844787598, "learning_rate": 0.00019677059968840817, "loss": 4.5282, "step": 315300 }, { "epoch": 0.588478163000147, "grad_norm": 1.0864149332046509, "learning_rate": 0.00019676954061327787, "loss": 4.539, "step": 315350 }, { "epoch": 0.5885714685595255, "grad_norm": 0.9973792433738708, "learning_rate": 0.0001967684813673663, "loss": 4.5436, "step": 315400 }, { "epoch": 0.5886647741189039, "grad_norm": 1.0296919345855713, "learning_rate": 0.0001967674219506754, "loss": 4.7594, "step": 315450 }, { "epoch": 0.5887580796782824, "grad_norm": 1.2115896940231323, "learning_rate": 0.00019676636236320703, "loss": 4.676, "step": 315500 }, { "epoch": 0.5888513852376609, "grad_norm": 1.299649715423584, "learning_rate": 0.00019676530260496304, "loss": 4.6847, "step": 315550 }, { "epoch": 0.5889446907970394, "grad_norm": 1.1729029417037964, "learning_rate": 0.00019676424267594535, "loss": 4.3708, "step": 315600 }, { "epoch": 0.589037996356418, "grad_norm": 1.0936784744262695, "learning_rate": 0.00019676318257615577, "loss": 4.6959, "step": 315650 }, { "epoch": 0.5891313019157964, "grad_norm": 0.9710637331008911, "learning_rate": 0.00019676212230559622, "loss": 4.5585, "step": 315700 }, { "epoch": 0.5892246074751749, "grad_norm": 1.055420160293579, "learning_rate": 0.00019676106186426856, "loss": 4.7101, "step": 315750 }, { "epoch": 0.5893179130345534, "grad_norm": 1.2843619585037231, "learning_rate": 0.0001967600012521746, "loss": 4.4441, "step": 315800 }, { "epoch": 0.5894112185939319, "grad_norm": 1.2306737899780273, "learning_rate": 0.0001967589404693163, "loss": 4.4583, "step": 315850 }, { "epoch": 0.5895045241533103, "grad_norm": 1.2592270374298096, "learning_rate": 0.0001967578795156955, "loss": 4.6707, "step": 315900 }, { "epoch": 0.5895978297126888, "grad_norm": 1.2055423259735107, "learning_rate": 0.00019675681839131408, "loss": 4.4714, "step": 315950 }, { "epoch": 0.5896911352720674, "grad_norm": 0.9702626466751099, "learning_rate": 0.00019675575709617388, "loss": 4.5387, "step": 316000 }, { "epoch": 0.5897844408314459, "grad_norm": 0.9379885196685791, "learning_rate": 0.00019675469563027676, "loss": 4.5659, "step": 316050 }, { "epoch": 0.5898777463908244, "grad_norm": 0.9110227227210999, "learning_rate": 0.00019675363399362467, "loss": 4.5908, "step": 316100 }, { "epoch": 0.5899710519502028, "grad_norm": 1.2341092824935913, "learning_rate": 0.00019675257218621943, "loss": 4.4516, "step": 316150 }, { "epoch": 0.5900643575095813, "grad_norm": 1.1738377809524536, "learning_rate": 0.00019675151020806292, "loss": 4.4439, "step": 316200 }, { "epoch": 0.5901576630689598, "grad_norm": 0.6892131567001343, "learning_rate": 0.00019675044805915704, "loss": 4.4533, "step": 316250 }, { "epoch": 0.5902509686283383, "grad_norm": 0.9237385988235474, "learning_rate": 0.00019674938573950362, "loss": 4.479, "step": 316300 }, { "epoch": 0.5903442741877168, "grad_norm": 1.0071806907653809, "learning_rate": 0.00019674832324910457, "loss": 4.6307, "step": 316350 }, { "epoch": 0.5904375797470953, "grad_norm": 0.7943397164344788, "learning_rate": 0.00019674726058796174, "loss": 4.3602, "step": 316400 }, { "epoch": 0.5905308853064738, "grad_norm": 0.9103261828422546, "learning_rate": 0.00019674619775607703, "loss": 4.6764, "step": 316450 }, { "epoch": 0.5906241908658523, "grad_norm": 0.9532055854797363, "learning_rate": 0.00019674513475345228, "loss": 4.5013, "step": 316500 }, { "epoch": 0.5907174964252307, "grad_norm": 0.7558811902999878, "learning_rate": 0.0001967440715800894, "loss": 4.5183, "step": 316550 }, { "epoch": 0.5908108019846092, "grad_norm": 1.2909374237060547, "learning_rate": 0.00019674300823599027, "loss": 4.6535, "step": 316600 }, { "epoch": 0.5909041075439877, "grad_norm": 1.1105871200561523, "learning_rate": 0.00019674194472115678, "loss": 4.384, "step": 316650 }, { "epoch": 0.5909974131033662, "grad_norm": 1.0830601453781128, "learning_rate": 0.00019674088103559072, "loss": 4.38, "step": 316700 }, { "epoch": 0.5910907186627448, "grad_norm": 0.8381434679031372, "learning_rate": 0.00019673981717929407, "loss": 4.4828, "step": 316750 }, { "epoch": 0.5911840242221232, "grad_norm": 0.9999323487281799, "learning_rate": 0.00019673875315226865, "loss": 4.532, "step": 316800 }, { "epoch": 0.5912773297815017, "grad_norm": 0.8143675923347473, "learning_rate": 0.00019673768895451637, "loss": 4.5682, "step": 316850 }, { "epoch": 0.5913706353408802, "grad_norm": 1.1459548473358154, "learning_rate": 0.00019673662458603907, "loss": 4.4766, "step": 316900 }, { "epoch": 0.5914639409002587, "grad_norm": 0.8771162033081055, "learning_rate": 0.00019673556004683862, "loss": 4.6515, "step": 316950 }, { "epoch": 0.5915572464596371, "grad_norm": 1.024323582649231, "learning_rate": 0.00019673449533691697, "loss": 4.692, "step": 317000 }, { "epoch": 0.5916505520190156, "grad_norm": 1.246565818786621, "learning_rate": 0.00019673343045627595, "loss": 4.6446, "step": 317050 }, { "epoch": 0.5917438575783942, "grad_norm": 0.9760048389434814, "learning_rate": 0.00019673236540491745, "loss": 4.7288, "step": 317100 }, { "epoch": 0.5918371631377727, "grad_norm": 1.1647191047668457, "learning_rate": 0.00019673130018284333, "loss": 4.4353, "step": 317150 }, { "epoch": 0.5919304686971512, "grad_norm": 1.019235372543335, "learning_rate": 0.00019673023479005548, "loss": 4.612, "step": 317200 }, { "epoch": 0.5920237742565296, "grad_norm": 0.9864053726196289, "learning_rate": 0.00019672916922655582, "loss": 4.7742, "step": 317250 }, { "epoch": 0.5921170798159081, "grad_norm": 1.1842503547668457, "learning_rate": 0.00019672810349234617, "loss": 4.5876, "step": 317300 }, { "epoch": 0.5922103853752866, "grad_norm": 1.0289043188095093, "learning_rate": 0.00019672703758742842, "loss": 4.706, "step": 317350 }, { "epoch": 0.5923036909346651, "grad_norm": 0.9660289287567139, "learning_rate": 0.00019672597151180451, "loss": 4.4939, "step": 317400 }, { "epoch": 0.5923969964940436, "grad_norm": 1.1275559663772583, "learning_rate": 0.00019672490526547628, "loss": 4.517, "step": 317450 }, { "epoch": 0.5924903020534221, "grad_norm": 1.21570885181427, "learning_rate": 0.00019672383884844556, "loss": 4.8104, "step": 317500 }, { "epoch": 0.5925836076128006, "grad_norm": 0.8104692697525024, "learning_rate": 0.0001967227722607143, "loss": 4.3482, "step": 317550 }, { "epoch": 0.5926769131721791, "grad_norm": 1.08836030960083, "learning_rate": 0.0001967217055022844, "loss": 4.5051, "step": 317600 }, { "epoch": 0.5927702187315576, "grad_norm": 1.0493236780166626, "learning_rate": 0.0001967206385731577, "loss": 4.6316, "step": 317650 }, { "epoch": 0.592863524290936, "grad_norm": 0.74932861328125, "learning_rate": 0.00019671957147333609, "loss": 4.4358, "step": 317700 }, { "epoch": 0.5929568298503145, "grad_norm": 1.2372153997421265, "learning_rate": 0.00019671850420282142, "loss": 4.5856, "step": 317750 }, { "epoch": 0.5930501354096931, "grad_norm": 1.0992337465286255, "learning_rate": 0.00019671743676161562, "loss": 4.5747, "step": 317800 }, { "epoch": 0.5931434409690716, "grad_norm": 1.0623835325241089, "learning_rate": 0.0001967163691497206, "loss": 4.4083, "step": 317850 }, { "epoch": 0.59323674652845, "grad_norm": 0.9603080749511719, "learning_rate": 0.00019671530136713817, "loss": 4.6688, "step": 317900 }, { "epoch": 0.5933300520878285, "grad_norm": 0.849819540977478, "learning_rate": 0.00019671423341387028, "loss": 4.5017, "step": 317950 }, { "epoch": 0.593423357647207, "grad_norm": 0.8867913484573364, "learning_rate": 0.00019671316528991874, "loss": 4.4862, "step": 318000 }, { "epoch": 0.593423357647207, "eval_loss": 4.749143123626709, "eval_runtime": 230.0262, "eval_samples_per_second": 11.338, "eval_steps_per_second": 11.338, "eval_tts_loss": 7.517176300315815, "step": 318000 }, { "epoch": 0.5935166632065855, "grad_norm": 0.8609456419944763, "learning_rate": 0.0001967120969952855, "loss": 4.5275, "step": 318050 }, { "epoch": 0.593609968765964, "grad_norm": 1.0815144777297974, "learning_rate": 0.0001967110285299725, "loss": 4.7102, "step": 318100 }, { "epoch": 0.5937032743253425, "grad_norm": 0.7996091842651367, "learning_rate": 0.00019670995989398146, "loss": 4.6476, "step": 318150 }, { "epoch": 0.593796579884721, "grad_norm": 1.0403865575790405, "learning_rate": 0.00019670889108731437, "loss": 4.4335, "step": 318200 }, { "epoch": 0.5938898854440995, "grad_norm": 1.0239715576171875, "learning_rate": 0.00019670782210997313, "loss": 4.6901, "step": 318250 }, { "epoch": 0.593983191003478, "grad_norm": 1.121938943862915, "learning_rate": 0.00019670675296195957, "loss": 4.4311, "step": 318300 }, { "epoch": 0.5940764965628564, "grad_norm": 1.120389699935913, "learning_rate": 0.00019670568364327565, "loss": 4.6799, "step": 318350 }, { "epoch": 0.5941698021222349, "grad_norm": 1.1791412830352783, "learning_rate": 0.00019670461415392319, "loss": 4.7471, "step": 318400 }, { "epoch": 0.5942631076816134, "grad_norm": 1.0191036462783813, "learning_rate": 0.0001967035444939041, "loss": 4.6189, "step": 318450 }, { "epoch": 0.594356413240992, "grad_norm": 1.2128996849060059, "learning_rate": 0.00019670247466322026, "loss": 4.5339, "step": 318500 }, { "epoch": 0.5944497188003705, "grad_norm": 0.9911199808120728, "learning_rate": 0.00019670140466187356, "loss": 4.576, "step": 318550 }, { "epoch": 0.5945430243597489, "grad_norm": 1.3055840730667114, "learning_rate": 0.0001967003344898659, "loss": 4.5811, "step": 318600 }, { "epoch": 0.5946363299191274, "grad_norm": 0.8344447016716003, "learning_rate": 0.0001966992641471992, "loss": 4.823, "step": 318650 }, { "epoch": 0.5947296354785059, "grad_norm": 1.0247474908828735, "learning_rate": 0.0001966981936338753, "loss": 4.6469, "step": 318700 }, { "epoch": 0.5948229410378844, "grad_norm": 0.9239888787269592, "learning_rate": 0.00019669712294989608, "loss": 4.5128, "step": 318750 }, { "epoch": 0.5949162465972628, "grad_norm": 0.8752967715263367, "learning_rate": 0.00019669605209526346, "loss": 4.7802, "step": 318800 }, { "epoch": 0.5950095521566414, "grad_norm": 1.1598962545394897, "learning_rate": 0.00019669498106997932, "loss": 4.5692, "step": 318850 }, { "epoch": 0.5951028577160199, "grad_norm": 1.3275585174560547, "learning_rate": 0.00019669390987404558, "loss": 4.6229, "step": 318900 }, { "epoch": 0.5951961632753984, "grad_norm": 1.271799921989441, "learning_rate": 0.00019669283850746409, "loss": 4.5638, "step": 318950 }, { "epoch": 0.5952894688347768, "grad_norm": 1.048512578010559, "learning_rate": 0.00019669176697023676, "loss": 4.6437, "step": 319000 }, { "epoch": 0.5953827743941553, "grad_norm": 0.7832193970680237, "learning_rate": 0.00019669069526236546, "loss": 4.3553, "step": 319050 }, { "epoch": 0.5954760799535338, "grad_norm": 0.8954023718833923, "learning_rate": 0.00019668962338385208, "loss": 4.5636, "step": 319100 }, { "epoch": 0.5955693855129123, "grad_norm": 1.0028667449951172, "learning_rate": 0.00019668855133469856, "loss": 4.5156, "step": 319150 }, { "epoch": 0.5956626910722909, "grad_norm": 1.0486000776290894, "learning_rate": 0.00019668747911490673, "loss": 4.242, "step": 319200 }, { "epoch": 0.5957559966316693, "grad_norm": 0.7862451076507568, "learning_rate": 0.00019668640672447854, "loss": 4.3009, "step": 319250 }, { "epoch": 0.5958493021910478, "grad_norm": 0.964422345161438, "learning_rate": 0.00019668533416341584, "loss": 4.5527, "step": 319300 }, { "epoch": 0.5959426077504263, "grad_norm": 0.8378289341926575, "learning_rate": 0.00019668426143172054, "loss": 4.5578, "step": 319350 }, { "epoch": 0.5960359133098048, "grad_norm": 0.9096817374229431, "learning_rate": 0.00019668318852939453, "loss": 4.6013, "step": 319400 }, { "epoch": 0.5961292188691832, "grad_norm": 1.119133472442627, "learning_rate": 0.0001966821154564397, "loss": 4.6927, "step": 319450 }, { "epoch": 0.5962225244285617, "grad_norm": 1.016858458518982, "learning_rate": 0.00019668104221285795, "loss": 4.6666, "step": 319500 }, { "epoch": 0.5963158299879403, "grad_norm": 0.9536043405532837, "learning_rate": 0.00019667996879865117, "loss": 4.735, "step": 319550 }, { "epoch": 0.5964091355473188, "grad_norm": 1.2586673498153687, "learning_rate": 0.00019667889521382127, "loss": 4.7085, "step": 319600 }, { "epoch": 0.5965024411066973, "grad_norm": 1.001456379890442, "learning_rate": 0.00019667782145837013, "loss": 4.5489, "step": 319650 }, { "epoch": 0.5965957466660757, "grad_norm": 1.457856297492981, "learning_rate": 0.00019667674753229964, "loss": 4.5942, "step": 319700 }, { "epoch": 0.5966890522254542, "grad_norm": 1.1986068487167358, "learning_rate": 0.00019667567343561172, "loss": 4.6471, "step": 319750 }, { "epoch": 0.5967823577848327, "grad_norm": 0.9569765329360962, "learning_rate": 0.00019667459916830818, "loss": 4.6439, "step": 319800 }, { "epoch": 0.5968756633442112, "grad_norm": 1.1844605207443237, "learning_rate": 0.00019667352473039107, "loss": 4.6101, "step": 319850 }, { "epoch": 0.5969689689035897, "grad_norm": 1.2512829303741455, "learning_rate": 0.00019667245012186215, "loss": 4.5151, "step": 319900 }, { "epoch": 0.5970622744629682, "grad_norm": 1.1351796388626099, "learning_rate": 0.00019667137534272336, "loss": 4.2293, "step": 319950 }, { "epoch": 0.5971555800223467, "grad_norm": 1.0945556163787842, "learning_rate": 0.00019667030039297663, "loss": 4.6776, "step": 320000 }, { "epoch": 0.5972488855817252, "grad_norm": 0.9205471277236938, "learning_rate": 0.0001966692252726238, "loss": 4.3958, "step": 320050 }, { "epoch": 0.5973421911411037, "grad_norm": 1.8201923370361328, "learning_rate": 0.00019666814998166684, "loss": 4.5612, "step": 320100 }, { "epoch": 0.5974354967004821, "grad_norm": 0.9468337297439575, "learning_rate": 0.00019666707452010757, "loss": 4.7242, "step": 320150 }, { "epoch": 0.5975288022598606, "grad_norm": 1.0803794860839844, "learning_rate": 0.00019666599888794792, "loss": 4.5611, "step": 320200 }, { "epoch": 0.5976221078192392, "grad_norm": 1.1365025043487549, "learning_rate": 0.0001966649230851898, "loss": 4.4926, "step": 320250 }, { "epoch": 0.5977154133786177, "grad_norm": 0.9078410267829895, "learning_rate": 0.00019666384711183513, "loss": 4.7069, "step": 320300 }, { "epoch": 0.5978087189379961, "grad_norm": 0.7499719858169556, "learning_rate": 0.00019666277096788573, "loss": 4.4693, "step": 320350 }, { "epoch": 0.5979020244973746, "grad_norm": 1.1963690519332886, "learning_rate": 0.00019666169465334357, "loss": 4.6755, "step": 320400 }, { "epoch": 0.5979953300567531, "grad_norm": 1.081960678100586, "learning_rate": 0.0001966606181682105, "loss": 4.718, "step": 320450 }, { "epoch": 0.5980886356161316, "grad_norm": 1.1670255661010742, "learning_rate": 0.00019665954151248848, "loss": 4.4774, "step": 320500 }, { "epoch": 0.59818194117551, "grad_norm": 0.851716935634613, "learning_rate": 0.00019665846468617937, "loss": 4.6525, "step": 320550 }, { "epoch": 0.5982752467348886, "grad_norm": 1.07505464553833, "learning_rate": 0.00019665738768928506, "loss": 4.5536, "step": 320600 }, { "epoch": 0.5983685522942671, "grad_norm": 1.0349235534667969, "learning_rate": 0.0001966563105218075, "loss": 4.3931, "step": 320650 }, { "epoch": 0.5984618578536456, "grad_norm": 1.1968237161636353, "learning_rate": 0.00019665523318374858, "loss": 4.6618, "step": 320700 }, { "epoch": 0.5985551634130241, "grad_norm": 1.3349511623382568, "learning_rate": 0.0001966541556751101, "loss": 4.5394, "step": 320750 }, { "epoch": 0.5986484689724025, "grad_norm": 1.3234177827835083, "learning_rate": 0.0001966530779958941, "loss": 4.5405, "step": 320800 }, { "epoch": 0.598741774531781, "grad_norm": 1.0499647855758667, "learning_rate": 0.0001966520001461024, "loss": 4.4868, "step": 320850 }, { "epoch": 0.5988350800911595, "grad_norm": 1.1239477396011353, "learning_rate": 0.00019665092212573694, "loss": 4.6189, "step": 320900 }, { "epoch": 0.5989283856505381, "grad_norm": 1.0901520252227783, "learning_rate": 0.0001966498439347996, "loss": 4.8085, "step": 320950 }, { "epoch": 0.5990216912099166, "grad_norm": 1.0736844539642334, "learning_rate": 0.0001966487655732923, "loss": 4.6435, "step": 321000 }, { "epoch": 0.5990216912099166, "eval_loss": 4.745899200439453, "eval_runtime": 229.1766, "eval_samples_per_second": 11.38, "eval_steps_per_second": 11.38, "eval_tts_loss": 7.537432325952486, "step": 321000 }, { "epoch": 0.599114996769295, "grad_norm": 0.7315561771392822, "learning_rate": 0.00019664768704121693, "loss": 4.6113, "step": 321050 }, { "epoch": 0.5992083023286735, "grad_norm": 1.0384379625320435, "learning_rate": 0.0001966466083385754, "loss": 4.7575, "step": 321100 }, { "epoch": 0.599301607888052, "grad_norm": 1.1658464670181274, "learning_rate": 0.00019664552946536962, "loss": 4.5812, "step": 321150 }, { "epoch": 0.5993949134474305, "grad_norm": 1.0108740329742432, "learning_rate": 0.00019664445042160146, "loss": 4.5712, "step": 321200 }, { "epoch": 0.5994882190068089, "grad_norm": 0.908758819103241, "learning_rate": 0.00019664337120727288, "loss": 4.6208, "step": 321250 }, { "epoch": 0.5995815245661875, "grad_norm": 0.9615160822868347, "learning_rate": 0.00019664229182238575, "loss": 4.5738, "step": 321300 }, { "epoch": 0.599674830125566, "grad_norm": 1.0263842344284058, "learning_rate": 0.00019664121226694196, "loss": 4.4814, "step": 321350 }, { "epoch": 0.5997681356849445, "grad_norm": 0.9682742953300476, "learning_rate": 0.00019664013254094343, "loss": 4.62, "step": 321400 }, { "epoch": 0.599861441244323, "grad_norm": 1.1773290634155273, "learning_rate": 0.0001966390526443921, "loss": 4.4712, "step": 321450 }, { "epoch": 0.5999547468037014, "grad_norm": 1.0511239767074585, "learning_rate": 0.00019663797257728983, "loss": 4.6363, "step": 321500 }, { "epoch": 0.6000480523630799, "grad_norm": 1.1144319772720337, "learning_rate": 0.00019663689233963855, "loss": 4.5899, "step": 321550 }, { "epoch": 0.6001413579224584, "grad_norm": 0.9538633227348328, "learning_rate": 0.00019663581193144015, "loss": 4.5972, "step": 321600 }, { "epoch": 0.600234663481837, "grad_norm": 1.1423015594482422, "learning_rate": 0.00019663473135269653, "loss": 4.5939, "step": 321650 }, { "epoch": 0.6003279690412154, "grad_norm": 1.0630602836608887, "learning_rate": 0.00019663365060340965, "loss": 4.6106, "step": 321700 }, { "epoch": 0.6004212746005939, "grad_norm": 1.1156061887741089, "learning_rate": 0.00019663256968358134, "loss": 4.4516, "step": 321750 }, { "epoch": 0.6005145801599724, "grad_norm": 1.0120784044265747, "learning_rate": 0.00019663148859321359, "loss": 4.8363, "step": 321800 }, { "epoch": 0.6006078857193509, "grad_norm": 0.9239031076431274, "learning_rate": 0.00019663040733230823, "loss": 4.3221, "step": 321850 }, { "epoch": 0.6007011912787293, "grad_norm": 0.9571710824966431, "learning_rate": 0.00019662932590086722, "loss": 4.6294, "step": 321900 }, { "epoch": 0.6007944968381078, "grad_norm": 0.7756795883178711, "learning_rate": 0.00019662824429889246, "loss": 4.4496, "step": 321950 }, { "epoch": 0.6008878023974864, "grad_norm": 1.1667654514312744, "learning_rate": 0.00019662716252638585, "loss": 4.6833, "step": 322000 }, { "epoch": 0.6009811079568649, "grad_norm": 1.0731756687164307, "learning_rate": 0.00019662608058334928, "loss": 4.5627, "step": 322050 }, { "epoch": 0.6010744135162434, "grad_norm": 1.1679133176803589, "learning_rate": 0.00019662499846978467, "loss": 4.5755, "step": 322100 }, { "epoch": 0.6011677190756218, "grad_norm": 0.9971390962600708, "learning_rate": 0.00019662391618569398, "loss": 4.6366, "step": 322150 }, { "epoch": 0.6012610246350003, "grad_norm": 1.1999056339263916, "learning_rate": 0.00019662283373107905, "loss": 4.4628, "step": 322200 }, { "epoch": 0.6013543301943788, "grad_norm": 0.9386271834373474, "learning_rate": 0.00019662175110594183, "loss": 4.5262, "step": 322250 }, { "epoch": 0.6014476357537573, "grad_norm": 1.5020827054977417, "learning_rate": 0.00019662066831028423, "loss": 4.5555, "step": 322300 }, { "epoch": 0.6015409413131357, "grad_norm": 1.0585623979568481, "learning_rate": 0.00019661958534410813, "loss": 4.4205, "step": 322350 }, { "epoch": 0.6016342468725143, "grad_norm": 0.9941714406013489, "learning_rate": 0.00019661850220741549, "loss": 4.6991, "step": 322400 }, { "epoch": 0.6017275524318928, "grad_norm": 1.1517648696899414, "learning_rate": 0.00019661741890020817, "loss": 4.6615, "step": 322450 }, { "epoch": 0.6018208579912713, "grad_norm": 1.0368093252182007, "learning_rate": 0.00019661633542248811, "loss": 4.5439, "step": 322500 }, { "epoch": 0.6019141635506498, "grad_norm": 0.943163275718689, "learning_rate": 0.00019661525177425723, "loss": 4.5757, "step": 322550 }, { "epoch": 0.6020074691100282, "grad_norm": 1.1650466918945312, "learning_rate": 0.00019661416795551743, "loss": 4.4431, "step": 322600 }, { "epoch": 0.6021007746694067, "grad_norm": 0.9699743986129761, "learning_rate": 0.00019661308396627062, "loss": 4.5648, "step": 322650 }, { "epoch": 0.6021940802287852, "grad_norm": 1.0106393098831177, "learning_rate": 0.0001966119998065187, "loss": 4.5206, "step": 322700 }, { "epoch": 0.6022873857881638, "grad_norm": 0.9035084247589111, "learning_rate": 0.0001966109154762636, "loss": 4.4606, "step": 322750 }, { "epoch": 0.6023806913475422, "grad_norm": 1.0321792364120483, "learning_rate": 0.00019660983097550728, "loss": 4.4261, "step": 322800 }, { "epoch": 0.6024739969069207, "grad_norm": 1.080522060394287, "learning_rate": 0.00019660874630425155, "loss": 4.595, "step": 322850 }, { "epoch": 0.6025673024662992, "grad_norm": 1.063269853591919, "learning_rate": 0.0001966076614624984, "loss": 4.5491, "step": 322900 }, { "epoch": 0.6026606080256777, "grad_norm": 0.8457520008087158, "learning_rate": 0.00019660657645024972, "loss": 4.5702, "step": 322950 }, { "epoch": 0.6027539135850561, "grad_norm": 0.9796984195709229, "learning_rate": 0.00019660549126750747, "loss": 4.4528, "step": 323000 }, { "epoch": 0.6028472191444346, "grad_norm": 0.9538129568099976, "learning_rate": 0.00019660440591427348, "loss": 4.3992, "step": 323050 }, { "epoch": 0.6029405247038132, "grad_norm": 0.8515247106552124, "learning_rate": 0.00019660332039054973, "loss": 4.5346, "step": 323100 }, { "epoch": 0.6030338302631917, "grad_norm": 1.1083250045776367, "learning_rate": 0.00019660223469633814, "loss": 4.6768, "step": 323150 }, { "epoch": 0.6031271358225702, "grad_norm": 0.8830024600028992, "learning_rate": 0.00019660114883164056, "loss": 4.4938, "step": 323200 }, { "epoch": 0.6032204413819486, "grad_norm": 1.167148232460022, "learning_rate": 0.00019660006279645899, "loss": 4.4916, "step": 323250 }, { "epoch": 0.6033137469413271, "grad_norm": 1.0859688520431519, "learning_rate": 0.00019659897659079527, "loss": 4.6358, "step": 323300 }, { "epoch": 0.6034070525007056, "grad_norm": 0.9899167418479919, "learning_rate": 0.00019659789021465139, "loss": 4.4966, "step": 323350 }, { "epoch": 0.6035003580600841, "grad_norm": 1.0375924110412598, "learning_rate": 0.0001965968036680292, "loss": 4.4735, "step": 323400 }, { "epoch": 0.6035936636194627, "grad_norm": 1.1162281036376953, "learning_rate": 0.00019659571695093065, "loss": 4.7, "step": 323450 }, { "epoch": 0.6036869691788411, "grad_norm": 0.8468580842018127, "learning_rate": 0.00019659463006335768, "loss": 4.444, "step": 323500 }, { "epoch": 0.6037802747382196, "grad_norm": 1.0031423568725586, "learning_rate": 0.00019659354300531216, "loss": 4.4444, "step": 323550 }, { "epoch": 0.6038735802975981, "grad_norm": 0.908850371837616, "learning_rate": 0.00019659245577679607, "loss": 4.5278, "step": 323600 }, { "epoch": 0.6039668858569766, "grad_norm": 0.9931618571281433, "learning_rate": 0.00019659136837781125, "loss": 4.6133, "step": 323650 }, { "epoch": 0.604060191416355, "grad_norm": 1.185490369796753, "learning_rate": 0.00019659028080835968, "loss": 4.7496, "step": 323700 }, { "epoch": 0.6041534969757335, "grad_norm": 1.058326244354248, "learning_rate": 0.00019658919306844323, "loss": 4.6327, "step": 323750 }, { "epoch": 0.6042468025351121, "grad_norm": 1.1122000217437744, "learning_rate": 0.0001965881051580639, "loss": 4.7566, "step": 323800 }, { "epoch": 0.6043401080944906, "grad_norm": 0.9507600665092468, "learning_rate": 0.00019658701707722353, "loss": 4.4221, "step": 323850 }, { "epoch": 0.604433413653869, "grad_norm": 1.1114460229873657, "learning_rate": 0.00019658592882592407, "loss": 4.7225, "step": 323900 }, { "epoch": 0.6045267192132475, "grad_norm": 0.9252007603645325, "learning_rate": 0.00019658484040416743, "loss": 4.6596, "step": 323950 }, { "epoch": 0.604620024772626, "grad_norm": 1.2695643901824951, "learning_rate": 0.00019658375181195556, "loss": 4.771, "step": 324000 }, { "epoch": 0.604620024772626, "eval_loss": 4.752098560333252, "eval_runtime": 228.5606, "eval_samples_per_second": 11.411, "eval_steps_per_second": 11.411, "eval_tts_loss": 7.5316662976555415, "step": 324000 }, { "epoch": 0.6047133303320045, "grad_norm": 1.2202486991882324, "learning_rate": 0.00019658266304929035, "loss": 4.5332, "step": 324050 }, { "epoch": 0.604806635891383, "grad_norm": 0.8959473967552185, "learning_rate": 0.00019658157411617374, "loss": 4.7924, "step": 324100 }, { "epoch": 0.6048999414507615, "grad_norm": 0.8497489094734192, "learning_rate": 0.00019658048501260765, "loss": 4.5222, "step": 324150 }, { "epoch": 0.60499324701014, "grad_norm": 1.0305777788162231, "learning_rate": 0.000196579395738594, "loss": 4.5465, "step": 324200 }, { "epoch": 0.6050865525695185, "grad_norm": 1.0758594274520874, "learning_rate": 0.00019657830629413468, "loss": 4.6718, "step": 324250 }, { "epoch": 0.605179858128897, "grad_norm": 1.2500865459442139, "learning_rate": 0.0001965772166792317, "loss": 4.6455, "step": 324300 }, { "epoch": 0.6052731636882754, "grad_norm": 1.0861148834228516, "learning_rate": 0.0001965761268938869, "loss": 4.6696, "step": 324350 }, { "epoch": 0.6053664692476539, "grad_norm": 1.0526787042617798, "learning_rate": 0.00019657503693810223, "loss": 4.682, "step": 324400 }, { "epoch": 0.6054597748070324, "grad_norm": 0.9431971311569214, "learning_rate": 0.0001965739468118796, "loss": 4.4591, "step": 324450 }, { "epoch": 0.605553080366411, "grad_norm": 1.114660382270813, "learning_rate": 0.00019657285651522095, "loss": 4.5528, "step": 324500 }, { "epoch": 0.6056463859257895, "grad_norm": 1.0945026874542236, "learning_rate": 0.00019657176604812818, "loss": 4.6842, "step": 324550 }, { "epoch": 0.6057396914851679, "grad_norm": 0.9303328990936279, "learning_rate": 0.00019657067541060328, "loss": 4.7461, "step": 324600 }, { "epoch": 0.6058329970445464, "grad_norm": 0.9786701202392578, "learning_rate": 0.00019656958460264812, "loss": 4.7805, "step": 324650 }, { "epoch": 0.6059263026039249, "grad_norm": 0.8835315704345703, "learning_rate": 0.0001965684936242646, "loss": 4.7447, "step": 324700 }, { "epoch": 0.6060196081633034, "grad_norm": 1.2973759174346924, "learning_rate": 0.00019656740247545474, "loss": 4.8283, "step": 324750 }, { "epoch": 0.6061129137226818, "grad_norm": 1.4678105115890503, "learning_rate": 0.00019656631115622035, "loss": 4.6572, "step": 324800 }, { "epoch": 0.6062062192820604, "grad_norm": 0.8270549178123474, "learning_rate": 0.00019656521966656342, "loss": 4.336, "step": 324850 }, { "epoch": 0.6062995248414389, "grad_norm": 1.040366530418396, "learning_rate": 0.0001965641280064859, "loss": 4.4321, "step": 324900 }, { "epoch": 0.6063928304008174, "grad_norm": 1.086034893989563, "learning_rate": 0.00019656303617598966, "loss": 4.5608, "step": 324950 }, { "epoch": 0.6064861359601958, "grad_norm": 1.1229069232940674, "learning_rate": 0.00019656194417507665, "loss": 4.6664, "step": 325000 }, { "epoch": 0.6065794415195743, "grad_norm": 0.9855028390884399, "learning_rate": 0.00019656085200374883, "loss": 4.7031, "step": 325050 }, { "epoch": 0.6066727470789528, "grad_norm": 0.996859073638916, "learning_rate": 0.00019655975966200806, "loss": 4.4463, "step": 325100 }, { "epoch": 0.6067660526383313, "grad_norm": 1.2329950332641602, "learning_rate": 0.00019655866714985633, "loss": 4.4839, "step": 325150 }, { "epoch": 0.6068593581977099, "grad_norm": 0.7367371916770935, "learning_rate": 0.00019655757446729552, "loss": 4.488, "step": 325200 }, { "epoch": 0.6069526637570883, "grad_norm": 1.0623233318328857, "learning_rate": 0.00019655648161432757, "loss": 4.4816, "step": 325250 }, { "epoch": 0.6070459693164668, "grad_norm": 1.2126883268356323, "learning_rate": 0.00019655538859095445, "loss": 4.7547, "step": 325300 }, { "epoch": 0.6071392748758453, "grad_norm": 0.8081970810890198, "learning_rate": 0.00019655429539717803, "loss": 4.5895, "step": 325350 }, { "epoch": 0.6072325804352238, "grad_norm": 0.9101871848106384, "learning_rate": 0.00019655320203300033, "loss": 4.5339, "step": 325400 }, { "epoch": 0.6073258859946022, "grad_norm": 0.9889973402023315, "learning_rate": 0.00019655210849842315, "loss": 4.5158, "step": 325450 }, { "epoch": 0.6074191915539807, "grad_norm": 1.2756493091583252, "learning_rate": 0.00019655101479344853, "loss": 4.6334, "step": 325500 }, { "epoch": 0.6075124971133593, "grad_norm": 1.1303919553756714, "learning_rate": 0.00019654992091807833, "loss": 4.4023, "step": 325550 }, { "epoch": 0.6076058026727378, "grad_norm": 0.8666788935661316, "learning_rate": 0.00019654882687231452, "loss": 4.5264, "step": 325600 }, { "epoch": 0.6076991082321163, "grad_norm": 0.773320734500885, "learning_rate": 0.00019654773265615904, "loss": 4.8125, "step": 325650 }, { "epoch": 0.6077924137914947, "grad_norm": 0.9333234429359436, "learning_rate": 0.00019654663826961378, "loss": 4.6164, "step": 325700 }, { "epoch": 0.6078857193508732, "grad_norm": 1.298832654953003, "learning_rate": 0.00019654554371268068, "loss": 4.6582, "step": 325750 }, { "epoch": 0.6079790249102517, "grad_norm": 0.8918960690498352, "learning_rate": 0.0001965444489853617, "loss": 4.7427, "step": 325800 }, { "epoch": 0.6080723304696302, "grad_norm": 1.246414303779602, "learning_rate": 0.00019654335408765872, "loss": 4.6355, "step": 325850 }, { "epoch": 0.6081656360290087, "grad_norm": 1.3998520374298096, "learning_rate": 0.00019654225901957374, "loss": 4.6412, "step": 325900 }, { "epoch": 0.6082589415883872, "grad_norm": 1.056199312210083, "learning_rate": 0.00019654116378110866, "loss": 4.4848, "step": 325950 }, { "epoch": 0.6083522471477657, "grad_norm": 1.1381001472473145, "learning_rate": 0.0001965400683722654, "loss": 4.6771, "step": 326000 }, { "epoch": 0.6084455527071442, "grad_norm": 0.9033389687538147, "learning_rate": 0.0001965389727930459, "loss": 4.3929, "step": 326050 }, { "epoch": 0.6085388582665227, "grad_norm": 1.209061622619629, "learning_rate": 0.00019653787704345212, "loss": 4.5338, "step": 326100 }, { "epoch": 0.6086321638259011, "grad_norm": 0.8938238620758057, "learning_rate": 0.00019653678112348597, "loss": 4.5691, "step": 326150 }, { "epoch": 0.6087254693852796, "grad_norm": 1.1634777784347534, "learning_rate": 0.0001965356850331494, "loss": 4.5017, "step": 326200 }, { "epoch": 0.6088187749446582, "grad_norm": 0.9691982865333557, "learning_rate": 0.00019653458877244432, "loss": 4.6735, "step": 326250 }, { "epoch": 0.6089120805040367, "grad_norm": 1.0083643198013306, "learning_rate": 0.00019653349234137265, "loss": 4.5339, "step": 326300 }, { "epoch": 0.6090053860634151, "grad_norm": 1.2493237257003784, "learning_rate": 0.0001965323957399364, "loss": 4.6472, "step": 326350 }, { "epoch": 0.6090986916227936, "grad_norm": 0.7249242067337036, "learning_rate": 0.00019653129896813742, "loss": 4.7061, "step": 326400 }, { "epoch": 0.6091919971821721, "grad_norm": 1.0830210447311401, "learning_rate": 0.0001965302020259777, "loss": 4.4675, "step": 326450 }, { "epoch": 0.6092853027415506, "grad_norm": 0.8388657569885254, "learning_rate": 0.00019652910491345913, "loss": 4.4492, "step": 326500 }, { "epoch": 0.609378608300929, "grad_norm": 0.9378383159637451, "learning_rate": 0.0001965280076305837, "loss": 4.6353, "step": 326550 }, { "epoch": 0.6094719138603076, "grad_norm": 0.8669517636299133, "learning_rate": 0.0001965269101773533, "loss": 4.5646, "step": 326600 }, { "epoch": 0.6095652194196861, "grad_norm": 0.952555239200592, "learning_rate": 0.00019652581255376992, "loss": 4.6136, "step": 326650 }, { "epoch": 0.6096585249790646, "grad_norm": 1.1579066514968872, "learning_rate": 0.00019652471475983544, "loss": 4.4012, "step": 326700 }, { "epoch": 0.6097518305384431, "grad_norm": 1.158402442932129, "learning_rate": 0.00019652361679555185, "loss": 4.4434, "step": 326750 }, { "epoch": 0.6098451360978215, "grad_norm": 1.1343615055084229, "learning_rate": 0.00019652251866092104, "loss": 4.5192, "step": 326800 }, { "epoch": 0.6099384416572, "grad_norm": 1.0151411294937134, "learning_rate": 0.00019652142035594497, "loss": 4.5552, "step": 326850 }, { "epoch": 0.6100317472165785, "grad_norm": 1.1065433025360107, "learning_rate": 0.00019652032188062556, "loss": 4.5854, "step": 326900 }, { "epoch": 0.6101250527759571, "grad_norm": 1.094696283340454, "learning_rate": 0.0001965192232349648, "loss": 4.7376, "step": 326950 }, { "epoch": 0.6102183583353356, "grad_norm": 1.0514317750930786, "learning_rate": 0.00019651812441896457, "loss": 4.5541, "step": 327000 }, { "epoch": 0.6102183583353356, "eval_loss": 4.746123790740967, "eval_runtime": 229.2631, "eval_samples_per_second": 11.376, "eval_steps_per_second": 11.376, "eval_tts_loss": 7.612345226578975, "step": 327000 }, { "epoch": 0.610311663894714, "grad_norm": 1.2415906190872192, "learning_rate": 0.00019651702543262685, "loss": 4.4649, "step": 327050 }, { "epoch": 0.6104049694540925, "grad_norm": 1.0724464654922485, "learning_rate": 0.00019651592627595353, "loss": 4.8505, "step": 327100 }, { "epoch": 0.610498275013471, "grad_norm": 0.9932399392127991, "learning_rate": 0.00019651482694894664, "loss": 4.4503, "step": 327150 }, { "epoch": 0.6105915805728495, "grad_norm": 0.9129202961921692, "learning_rate": 0.000196513727451608, "loss": 4.5826, "step": 327200 }, { "epoch": 0.6106848861322279, "grad_norm": 1.6551772356033325, "learning_rate": 0.00019651262778393967, "loss": 4.7348, "step": 327250 }, { "epoch": 0.6107781916916065, "grad_norm": 0.7570912837982178, "learning_rate": 0.00019651152794594346, "loss": 4.504, "step": 327300 }, { "epoch": 0.610871497250985, "grad_norm": 0.8969416618347168, "learning_rate": 0.00019651042793762146, "loss": 4.6502, "step": 327350 }, { "epoch": 0.6109648028103635, "grad_norm": 1.240637183189392, "learning_rate": 0.0001965093277589755, "loss": 4.5202, "step": 327400 }, { "epoch": 0.611058108369742, "grad_norm": 1.1303819417953491, "learning_rate": 0.0001965082274100076, "loss": 4.5728, "step": 327450 }, { "epoch": 0.6111514139291204, "grad_norm": 1.054078221321106, "learning_rate": 0.0001965071268907196, "loss": 4.5804, "step": 327500 }, { "epoch": 0.6112447194884989, "grad_norm": 0.9103609919548035, "learning_rate": 0.00019650602620111354, "loss": 4.4957, "step": 327550 }, { "epoch": 0.6113380250478774, "grad_norm": 0.5890389084815979, "learning_rate": 0.0001965049253411913, "loss": 4.5091, "step": 327600 }, { "epoch": 0.6114313306072559, "grad_norm": 1.1171053647994995, "learning_rate": 0.00019650382431095483, "loss": 4.5723, "step": 327650 }, { "epoch": 0.6115246361666344, "grad_norm": 1.0724706649780273, "learning_rate": 0.00019650272311040614, "loss": 4.7041, "step": 327700 }, { "epoch": 0.6116179417260129, "grad_norm": 0.8792927861213684, "learning_rate": 0.0001965016217395471, "loss": 4.7052, "step": 327750 }, { "epoch": 0.6117112472853914, "grad_norm": 0.9894546270370483, "learning_rate": 0.00019650052019837968, "loss": 4.7049, "step": 327800 }, { "epoch": 0.6118045528447699, "grad_norm": 1.0347297191619873, "learning_rate": 0.00019649941848690582, "loss": 4.4355, "step": 327850 }, { "epoch": 0.6118978584041483, "grad_norm": 1.1192511320114136, "learning_rate": 0.00019649831660512748, "loss": 4.5305, "step": 327900 }, { "epoch": 0.6119911639635268, "grad_norm": 1.2721318006515503, "learning_rate": 0.00019649721455304657, "loss": 4.3744, "step": 327950 }, { "epoch": 0.6120844695229053, "grad_norm": 1.1381678581237793, "learning_rate": 0.00019649611233066506, "loss": 4.7289, "step": 328000 }, { "epoch": 0.6121777750822839, "grad_norm": 1.2317228317260742, "learning_rate": 0.00019649500993798488, "loss": 4.5512, "step": 328050 }, { "epoch": 0.6122710806416624, "grad_norm": 0.9952607154846191, "learning_rate": 0.000196493907375008, "loss": 4.5351, "step": 328100 }, { "epoch": 0.6123643862010408, "grad_norm": 0.5989564061164856, "learning_rate": 0.0001964928046417364, "loss": 4.3921, "step": 328150 }, { "epoch": 0.6124576917604193, "grad_norm": 0.7932347059249878, "learning_rate": 0.0001964917017381719, "loss": 4.4653, "step": 328200 }, { "epoch": 0.6125509973197978, "grad_norm": 1.1928980350494385, "learning_rate": 0.00019649059866431656, "loss": 4.5038, "step": 328250 }, { "epoch": 0.6126443028791763, "grad_norm": 1.196129560470581, "learning_rate": 0.00019648949542017228, "loss": 4.6942, "step": 328300 }, { "epoch": 0.6127376084385547, "grad_norm": 1.213106632232666, "learning_rate": 0.00019648839200574103, "loss": 4.5907, "step": 328350 }, { "epoch": 0.6128309139979333, "grad_norm": 1.1690261363983154, "learning_rate": 0.00019648728842102474, "loss": 4.6448, "step": 328400 }, { "epoch": 0.6129242195573118, "grad_norm": 1.2848875522613525, "learning_rate": 0.00019648618466602534, "loss": 4.5353, "step": 328450 }, { "epoch": 0.6130175251166903, "grad_norm": 1.101271152496338, "learning_rate": 0.00019648508074074486, "loss": 4.8021, "step": 328500 }, { "epoch": 0.6131108306760688, "grad_norm": 1.112776756286621, "learning_rate": 0.00019648397664518514, "loss": 4.4725, "step": 328550 }, { "epoch": 0.6132041362354472, "grad_norm": 1.0443400144577026, "learning_rate": 0.0001964828723793482, "loss": 4.6165, "step": 328600 }, { "epoch": 0.6132974417948257, "grad_norm": 1.1482632160186768, "learning_rate": 0.00019648176794323597, "loss": 4.4479, "step": 328650 }, { "epoch": 0.6133907473542042, "grad_norm": 1.1257174015045166, "learning_rate": 0.0001964806633368504, "loss": 4.8037, "step": 328700 }, { "epoch": 0.6134840529135828, "grad_norm": 0.9680014252662659, "learning_rate": 0.0001964795585601934, "loss": 4.6661, "step": 328750 }, { "epoch": 0.6135773584729612, "grad_norm": 1.0088255405426025, "learning_rate": 0.00019647845361326697, "loss": 4.3305, "step": 328800 }, { "epoch": 0.6136706640323397, "grad_norm": 1.115094542503357, "learning_rate": 0.00019647734849607306, "loss": 4.6662, "step": 328850 }, { "epoch": 0.6137639695917182, "grad_norm": 1.0315027236938477, "learning_rate": 0.0001964762432086136, "loss": 4.544, "step": 328900 }, { "epoch": 0.6138572751510967, "grad_norm": 0.9989319443702698, "learning_rate": 0.00019647513775089055, "loss": 4.6008, "step": 328950 }, { "epoch": 0.6139505807104751, "grad_norm": 0.8979176878929138, "learning_rate": 0.00019647403212290585, "loss": 4.5104, "step": 329000 }, { "epoch": 0.6140438862698536, "grad_norm": 0.7627320289611816, "learning_rate": 0.00019647292632466145, "loss": 4.666, "step": 329050 }, { "epoch": 0.6141371918292322, "grad_norm": 1.0842419862747192, "learning_rate": 0.00019647182035615934, "loss": 4.5201, "step": 329100 }, { "epoch": 0.6142304973886107, "grad_norm": 1.143351674079895, "learning_rate": 0.00019647071421740142, "loss": 4.5823, "step": 329150 }, { "epoch": 0.6143238029479892, "grad_norm": 1.2588927745819092, "learning_rate": 0.00019646960790838965, "loss": 4.5743, "step": 329200 }, { "epoch": 0.6144171085073676, "grad_norm": 1.2234686613082886, "learning_rate": 0.000196468501429126, "loss": 4.6187, "step": 329250 }, { "epoch": 0.6145104140667461, "grad_norm": 0.8479459881782532, "learning_rate": 0.00019646739477961246, "loss": 4.6488, "step": 329300 }, { "epoch": 0.6146037196261246, "grad_norm": 0.9231317639350891, "learning_rate": 0.00019646628795985088, "loss": 4.6329, "step": 329350 }, { "epoch": 0.6146970251855031, "grad_norm": 0.9470799565315247, "learning_rate": 0.00019646518096984333, "loss": 4.4673, "step": 329400 }, { "epoch": 0.6147903307448817, "grad_norm": 0.9557851552963257, "learning_rate": 0.0001964640738095917, "loss": 4.5174, "step": 329450 }, { "epoch": 0.6148836363042601, "grad_norm": 1.2796331644058228, "learning_rate": 0.00019646296647909793, "loss": 4.6437, "step": 329500 }, { "epoch": 0.6149769418636386, "grad_norm": 1.0128943920135498, "learning_rate": 0.000196461858978364, "loss": 4.546, "step": 329550 }, { "epoch": 0.6150702474230171, "grad_norm": 0.8020675182342529, "learning_rate": 0.00019646075130739187, "loss": 4.6278, "step": 329600 }, { "epoch": 0.6151635529823956, "grad_norm": 1.032940149307251, "learning_rate": 0.00019645964346618348, "loss": 4.4691, "step": 329650 }, { "epoch": 0.615256858541774, "grad_norm": 0.9859622120857239, "learning_rate": 0.00019645853545474082, "loss": 4.3977, "step": 329700 }, { "epoch": 0.6153501641011525, "grad_norm": 1.0631998777389526, "learning_rate": 0.00019645742727306578, "loss": 4.5291, "step": 329750 }, { "epoch": 0.6154434696605311, "grad_norm": 0.8335729837417603, "learning_rate": 0.00019645631892116039, "loss": 4.4934, "step": 329800 }, { "epoch": 0.6155367752199096, "grad_norm": 1.0670087337493896, "learning_rate": 0.00019645521039902655, "loss": 4.77, "step": 329850 }, { "epoch": 0.615630080779288, "grad_norm": 0.9757839441299438, "learning_rate": 0.00019645410170666621, "loss": 4.6015, "step": 329900 }, { "epoch": 0.6157233863386665, "grad_norm": 0.6826449036598206, "learning_rate": 0.00019645299284408137, "loss": 4.5966, "step": 329950 }, { "epoch": 0.615816691898045, "grad_norm": 1.2171297073364258, "learning_rate": 0.000196451883811274, "loss": 4.5911, "step": 330000 }, { "epoch": 0.615816691898045, "eval_loss": 4.750162601470947, "eval_runtime": 230.0209, "eval_samples_per_second": 11.338, "eval_steps_per_second": 11.338, "eval_tts_loss": 7.530302074275042, "step": 330000 }, { "epoch": 0.6159099974574235, "grad_norm": 1.0892624855041504, "learning_rate": 0.000196450774608246, "loss": 4.3776, "step": 330050 }, { "epoch": 0.616003303016802, "grad_norm": 1.1487836837768555, "learning_rate": 0.00019644966523499937, "loss": 4.5873, "step": 330100 }, { "epoch": 0.6160966085761805, "grad_norm": 1.07069993019104, "learning_rate": 0.00019644855569153605, "loss": 4.5787, "step": 330150 }, { "epoch": 0.616189914135559, "grad_norm": 1.1551017761230469, "learning_rate": 0.00019644744597785795, "loss": 4.6652, "step": 330200 }, { "epoch": 0.6162832196949375, "grad_norm": 1.1474007368087769, "learning_rate": 0.00019644633609396713, "loss": 4.407, "step": 330250 }, { "epoch": 0.616376525254316, "grad_norm": 0.6405335068702698, "learning_rate": 0.0001964452260398655, "loss": 4.5194, "step": 330300 }, { "epoch": 0.6164698308136944, "grad_norm": 0.8228738903999329, "learning_rate": 0.000196444115815555, "loss": 4.7268, "step": 330350 }, { "epoch": 0.6165631363730729, "grad_norm": 0.929740846157074, "learning_rate": 0.0001964430054210376, "loss": 4.3615, "step": 330400 }, { "epoch": 0.6166564419324514, "grad_norm": 1.1017268896102905, "learning_rate": 0.00019644189485631527, "loss": 4.5888, "step": 330450 }, { "epoch": 0.61674974749183, "grad_norm": 0.9419089555740356, "learning_rate": 0.00019644078412138995, "loss": 4.5174, "step": 330500 }, { "epoch": 0.6168430530512085, "grad_norm": 1.1504666805267334, "learning_rate": 0.00019643967321626363, "loss": 4.4526, "step": 330550 }, { "epoch": 0.6169363586105869, "grad_norm": 1.1194062232971191, "learning_rate": 0.00019643856214093824, "loss": 4.5075, "step": 330600 }, { "epoch": 0.6170296641699654, "grad_norm": 1.017431378364563, "learning_rate": 0.00019643745089541577, "loss": 4.5547, "step": 330650 }, { "epoch": 0.6171229697293439, "grad_norm": 0.9872207045555115, "learning_rate": 0.0001964363394796982, "loss": 4.5328, "step": 330700 }, { "epoch": 0.6172162752887224, "grad_norm": 0.8828074336051941, "learning_rate": 0.0001964352278937874, "loss": 4.6521, "step": 330750 }, { "epoch": 0.6173095808481008, "grad_norm": 0.8775408864021301, "learning_rate": 0.00019643411613768543, "loss": 4.7866, "step": 330800 }, { "epoch": 0.6174028864074794, "grad_norm": 0.7545701265335083, "learning_rate": 0.0001964330042113942, "loss": 4.5833, "step": 330850 }, { "epoch": 0.6174961919668579, "grad_norm": 1.1457773447036743, "learning_rate": 0.00019643189211491572, "loss": 4.5884, "step": 330900 }, { "epoch": 0.6175894975262364, "grad_norm": 1.1608442068099976, "learning_rate": 0.00019643077984825185, "loss": 4.586, "step": 330950 }, { "epoch": 0.6176828030856149, "grad_norm": 1.0994585752487183, "learning_rate": 0.00019642966741140468, "loss": 4.3464, "step": 331000 }, { "epoch": 0.6177761086449933, "grad_norm": 1.1020636558532715, "learning_rate": 0.0001964285548043761, "loss": 4.5699, "step": 331050 }, { "epoch": 0.6178694142043718, "grad_norm": 1.0513229370117188, "learning_rate": 0.00019642744202716806, "loss": 4.4984, "step": 331100 }, { "epoch": 0.6179627197637503, "grad_norm": 1.385101079940796, "learning_rate": 0.00019642632907978258, "loss": 4.5985, "step": 331150 }, { "epoch": 0.6180560253231289, "grad_norm": 1.1317663192749023, "learning_rate": 0.0001964252159622216, "loss": 4.3054, "step": 331200 }, { "epoch": 0.6181493308825073, "grad_norm": 0.9027521014213562, "learning_rate": 0.00019642410267448705, "loss": 4.6536, "step": 331250 }, { "epoch": 0.6182426364418858, "grad_norm": 1.294648289680481, "learning_rate": 0.00019642298921658096, "loss": 4.7736, "step": 331300 }, { "epoch": 0.6183359420012643, "grad_norm": 1.1973425149917603, "learning_rate": 0.00019642187558850522, "loss": 4.5538, "step": 331350 }, { "epoch": 0.6184292475606428, "grad_norm": 1.1227333545684814, "learning_rate": 0.0001964207617902619, "loss": 4.6983, "step": 331400 }, { "epoch": 0.6185225531200212, "grad_norm": 0.9092786312103271, "learning_rate": 0.00019641964782185283, "loss": 4.4916, "step": 331450 }, { "epoch": 0.6186158586793997, "grad_norm": 1.4974592924118042, "learning_rate": 0.0001964185336832801, "loss": 4.7164, "step": 331500 }, { "epoch": 0.6187091642387783, "grad_norm": 1.073347568511963, "learning_rate": 0.0001964174193745456, "loss": 4.5721, "step": 331550 }, { "epoch": 0.6188024697981568, "grad_norm": 1.0728085041046143, "learning_rate": 0.0001964163048956513, "loss": 4.6056, "step": 331600 }, { "epoch": 0.6188957753575353, "grad_norm": 0.6657750010490417, "learning_rate": 0.0001964151902465992, "loss": 4.397, "step": 331650 }, { "epoch": 0.6189890809169137, "grad_norm": 0.7082281112670898, "learning_rate": 0.00019641407542739128, "loss": 4.4559, "step": 331700 }, { "epoch": 0.6190823864762922, "grad_norm": 1.0563206672668457, "learning_rate": 0.00019641296043802946, "loss": 4.6829, "step": 331750 }, { "epoch": 0.6191756920356707, "grad_norm": 1.0246212482452393, "learning_rate": 0.00019641184527851573, "loss": 4.5568, "step": 331800 }, { "epoch": 0.6192689975950492, "grad_norm": 1.1611067056655884, "learning_rate": 0.00019641072994885207, "loss": 4.6479, "step": 331850 }, { "epoch": 0.6193623031544278, "grad_norm": 1.0870575904846191, "learning_rate": 0.00019640961444904043, "loss": 4.7873, "step": 331900 }, { "epoch": 0.6194556087138062, "grad_norm": 1.0575101375579834, "learning_rate": 0.00019640849877908275, "loss": 4.5286, "step": 331950 }, { "epoch": 0.6195489142731847, "grad_norm": 1.4265297651290894, "learning_rate": 0.00019640738293898107, "loss": 4.6938, "step": 332000 }, { "epoch": 0.6196422198325632, "grad_norm": 0.8864887356758118, "learning_rate": 0.00019640626692873735, "loss": 4.4216, "step": 332050 }, { "epoch": 0.6197355253919417, "grad_norm": 0.8195017576217651, "learning_rate": 0.00019640515074835347, "loss": 4.6461, "step": 332100 }, { "epoch": 0.6198288309513201, "grad_norm": 1.165611743927002, "learning_rate": 0.0001964040343978315, "loss": 4.6621, "step": 332150 }, { "epoch": 0.6199221365106986, "grad_norm": 1.1882660388946533, "learning_rate": 0.00019640291787717336, "loss": 4.5354, "step": 332200 }, { "epoch": 0.6200154420700772, "grad_norm": 1.2898271083831787, "learning_rate": 0.00019640180118638104, "loss": 4.777, "step": 332250 }, { "epoch": 0.6201087476294557, "grad_norm": 1.1433839797973633, "learning_rate": 0.0001964006843254565, "loss": 4.6701, "step": 332300 }, { "epoch": 0.6202020531888341, "grad_norm": 0.9746190309524536, "learning_rate": 0.0001963995672944017, "loss": 4.6845, "step": 332350 }, { "epoch": 0.6202953587482126, "grad_norm": 1.1168901920318604, "learning_rate": 0.00019639845009321864, "loss": 4.5668, "step": 332400 }, { "epoch": 0.6203886643075911, "grad_norm": 0.7818266749382019, "learning_rate": 0.00019639733272190925, "loss": 4.6588, "step": 332450 }, { "epoch": 0.6204819698669696, "grad_norm": 0.812650203704834, "learning_rate": 0.00019639621518047556, "loss": 4.6722, "step": 332500 }, { "epoch": 0.620575275426348, "grad_norm": 0.8418398499488831, "learning_rate": 0.00019639509746891952, "loss": 4.5468, "step": 332550 }, { "epoch": 0.6206685809857266, "grad_norm": 1.0831886529922485, "learning_rate": 0.00019639397958724308, "loss": 4.6973, "step": 332600 }, { "epoch": 0.6207618865451051, "grad_norm": 1.0995817184448242, "learning_rate": 0.00019639286153544822, "loss": 4.6884, "step": 332650 }, { "epoch": 0.6208551921044836, "grad_norm": 1.1868793964385986, "learning_rate": 0.00019639174331353694, "loss": 4.5945, "step": 332700 }, { "epoch": 0.6209484976638621, "grad_norm": 1.0501141548156738, "learning_rate": 0.0001963906249215112, "loss": 4.6657, "step": 332750 }, { "epoch": 0.6210418032232405, "grad_norm": 1.1675609350204468, "learning_rate": 0.0001963895063593729, "loss": 4.5511, "step": 332800 }, { "epoch": 0.621135108782619, "grad_norm": 0.9354573488235474, "learning_rate": 0.00019638838762712414, "loss": 4.6609, "step": 332850 }, { "epoch": 0.6212284143419975, "grad_norm": 0.8872344493865967, "learning_rate": 0.00019638726872476685, "loss": 4.6845, "step": 332900 }, { "epoch": 0.621321719901376, "grad_norm": 1.1700870990753174, "learning_rate": 0.00019638614965230295, "loss": 4.5461, "step": 332950 }, { "epoch": 0.6214150254607546, "grad_norm": 1.4714617729187012, "learning_rate": 0.00019638503040973447, "loss": 4.5, "step": 333000 }, { "epoch": 0.6214150254607546, "eval_loss": 4.740175247192383, "eval_runtime": 230.3212, "eval_samples_per_second": 11.323, "eval_steps_per_second": 11.323, "eval_tts_loss": 7.545454389351403, "step": 333000 }, { "epoch": 0.621508331020133, "grad_norm": 0.8408044576644897, "learning_rate": 0.00019638391099706338, "loss": 4.4279, "step": 333050 }, { "epoch": 0.6216016365795115, "grad_norm": 1.0031147003173828, "learning_rate": 0.00019638279141429162, "loss": 4.5333, "step": 333100 }, { "epoch": 0.62169494213889, "grad_norm": 0.9783959984779358, "learning_rate": 0.00019638167166142125, "loss": 4.4112, "step": 333150 }, { "epoch": 0.6217882476982685, "grad_norm": 1.0063459873199463, "learning_rate": 0.00019638055173845415, "loss": 4.7576, "step": 333200 }, { "epoch": 0.6218815532576469, "grad_norm": 0.936040461063385, "learning_rate": 0.00019637943164539233, "loss": 4.6556, "step": 333250 }, { "epoch": 0.6219748588170254, "grad_norm": 1.168675422668457, "learning_rate": 0.00019637831138223777, "loss": 4.4459, "step": 333300 }, { "epoch": 0.622068164376404, "grad_norm": 1.1717939376831055, "learning_rate": 0.00019637719094899245, "loss": 4.781, "step": 333350 }, { "epoch": 0.6221614699357825, "grad_norm": 0.8017193078994751, "learning_rate": 0.00019637607034565834, "loss": 4.73, "step": 333400 }, { "epoch": 0.622254775495161, "grad_norm": 1.0900770425796509, "learning_rate": 0.00019637494957223746, "loss": 4.7813, "step": 333450 }, { "epoch": 0.6223480810545394, "grad_norm": 1.1926219463348389, "learning_rate": 0.00019637382862873172, "loss": 4.6637, "step": 333500 }, { "epoch": 0.6224413866139179, "grad_norm": 1.0957951545715332, "learning_rate": 0.00019637270751514315, "loss": 4.6358, "step": 333550 }, { "epoch": 0.6225346921732964, "grad_norm": 1.0752534866333008, "learning_rate": 0.00019637158623147366, "loss": 4.6348, "step": 333600 }, { "epoch": 0.6226279977326749, "grad_norm": 1.108163595199585, "learning_rate": 0.00019637046477772533, "loss": 4.761, "step": 333650 }, { "epoch": 0.6227213032920534, "grad_norm": 0.9737891554832458, "learning_rate": 0.00019636934315390007, "loss": 4.4272, "step": 333700 }, { "epoch": 0.6228146088514319, "grad_norm": 1.2830264568328857, "learning_rate": 0.00019636822135999987, "loss": 4.6189, "step": 333750 }, { "epoch": 0.6229079144108104, "grad_norm": 1.093766212463379, "learning_rate": 0.00019636709939602673, "loss": 4.6984, "step": 333800 }, { "epoch": 0.6230012199701889, "grad_norm": 1.0020575523376465, "learning_rate": 0.0001963659772619826, "loss": 4.6488, "step": 333850 }, { "epoch": 0.6230945255295673, "grad_norm": 0.9393331408500671, "learning_rate": 0.00019636485495786948, "loss": 4.49, "step": 333900 }, { "epoch": 0.6231878310889458, "grad_norm": 1.1477243900299072, "learning_rate": 0.00019636373248368936, "loss": 4.6383, "step": 333950 }, { "epoch": 0.6232811366483243, "grad_norm": 1.0864436626434326, "learning_rate": 0.00019636260983944422, "loss": 4.6965, "step": 334000 }, { "epoch": 0.6233744422077029, "grad_norm": 0.757732629776001, "learning_rate": 0.00019636148702513597, "loss": 4.7961, "step": 334050 }, { "epoch": 0.6234677477670814, "grad_norm": 1.0098869800567627, "learning_rate": 0.00019636036404076669, "loss": 4.6886, "step": 334100 }, { "epoch": 0.6235610533264598, "grad_norm": 1.1908249855041504, "learning_rate": 0.00019635924088633834, "loss": 4.4147, "step": 334150 }, { "epoch": 0.6236543588858383, "grad_norm": 0.871414065361023, "learning_rate": 0.00019635811756185288, "loss": 4.7605, "step": 334200 }, { "epoch": 0.6237476644452168, "grad_norm": 1.0281577110290527, "learning_rate": 0.00019635699406731226, "loss": 4.5573, "step": 334250 }, { "epoch": 0.6238409700045953, "grad_norm": 0.8493218421936035, "learning_rate": 0.00019635587040271851, "loss": 4.3602, "step": 334300 }, { "epoch": 0.6239342755639737, "grad_norm": 0.9594669938087463, "learning_rate": 0.00019635474656807365, "loss": 4.5932, "step": 334350 }, { "epoch": 0.6240275811233523, "grad_norm": 0.9037676453590393, "learning_rate": 0.00019635362256337956, "loss": 4.5376, "step": 334400 }, { "epoch": 0.6241208866827308, "grad_norm": 1.2145931720733643, "learning_rate": 0.0001963524983886383, "loss": 4.4981, "step": 334450 }, { "epoch": 0.6242141922421093, "grad_norm": 1.0076943635940552, "learning_rate": 0.00019635137404385185, "loss": 4.558, "step": 334500 }, { "epoch": 0.6243074978014878, "grad_norm": 1.2487367391586304, "learning_rate": 0.00019635024952902215, "loss": 4.6688, "step": 334550 }, { "epoch": 0.6244008033608662, "grad_norm": 1.0044771432876587, "learning_rate": 0.00019634912484415123, "loss": 4.6066, "step": 334600 }, { "epoch": 0.6244941089202447, "grad_norm": 1.2532018423080444, "learning_rate": 0.00019634799998924108, "loss": 4.5018, "step": 334650 }, { "epoch": 0.6245874144796232, "grad_norm": 0.9828773140907288, "learning_rate": 0.0001963468749642936, "loss": 4.5975, "step": 334700 }, { "epoch": 0.6246807200390018, "grad_norm": 1.2060518264770508, "learning_rate": 0.0001963457497693109, "loss": 4.6228, "step": 334750 }, { "epoch": 0.6247740255983802, "grad_norm": 0.9976147413253784, "learning_rate": 0.0001963446244042949, "loss": 4.5998, "step": 334800 }, { "epoch": 0.6248673311577587, "grad_norm": 1.1077107191085815, "learning_rate": 0.00019634349886924753, "loss": 4.5616, "step": 334850 }, { "epoch": 0.6249606367171372, "grad_norm": 0.9811950325965881, "learning_rate": 0.00019634237316417088, "loss": 4.5548, "step": 334900 }, { "epoch": 0.6250539422765157, "grad_norm": 1.1001842021942139, "learning_rate": 0.0001963412472890669, "loss": 4.4049, "step": 334950 }, { "epoch": 0.6251472478358941, "grad_norm": 1.3210327625274658, "learning_rate": 0.00019634012124393756, "loss": 4.7167, "step": 335000 }, { "epoch": 0.6252405533952726, "grad_norm": 1.1540125608444214, "learning_rate": 0.00019633899502878487, "loss": 4.4382, "step": 335050 }, { "epoch": 0.6253338589546512, "grad_norm": 1.3277180194854736, "learning_rate": 0.00019633786864361078, "loss": 4.608, "step": 335100 }, { "epoch": 0.6254271645140297, "grad_norm": 1.3375109434127808, "learning_rate": 0.00019633674208841733, "loss": 4.3782, "step": 335150 }, { "epoch": 0.6255204700734082, "grad_norm": 1.0912272930145264, "learning_rate": 0.00019633561536320647, "loss": 4.6823, "step": 335200 }, { "epoch": 0.6256137756327866, "grad_norm": 1.2482918500900269, "learning_rate": 0.0001963344884679802, "loss": 4.6102, "step": 335250 }, { "epoch": 0.6257070811921651, "grad_norm": 1.2576767206192017, "learning_rate": 0.0001963333614027405, "loss": 4.6146, "step": 335300 }, { "epoch": 0.6258003867515436, "grad_norm": 0.9687120318412781, "learning_rate": 0.00019633223416748937, "loss": 4.5318, "step": 335350 }, { "epoch": 0.6258936923109221, "grad_norm": 0.7464566230773926, "learning_rate": 0.0001963311067622288, "loss": 4.4003, "step": 335400 }, { "epoch": 0.6259869978703007, "grad_norm": 1.0068252086639404, "learning_rate": 0.00019632997918696079, "loss": 4.6142, "step": 335450 }, { "epoch": 0.6260803034296791, "grad_norm": 1.048789143562317, "learning_rate": 0.00019632885144168732, "loss": 4.5993, "step": 335500 }, { "epoch": 0.6261736089890576, "grad_norm": 1.1501978635787964, "learning_rate": 0.00019632772352641035, "loss": 4.4245, "step": 335550 }, { "epoch": 0.6262669145484361, "grad_norm": 1.0876208543777466, "learning_rate": 0.0001963265954411319, "loss": 4.6859, "step": 335600 }, { "epoch": 0.6263602201078146, "grad_norm": 1.1919074058532715, "learning_rate": 0.00019632546718585396, "loss": 4.6262, "step": 335650 }, { "epoch": 0.626453525667193, "grad_norm": 1.3580552339553833, "learning_rate": 0.0001963243387605785, "loss": 4.5983, "step": 335700 }, { "epoch": 0.6265468312265715, "grad_norm": 0.9644747376441956, "learning_rate": 0.00019632321016530759, "loss": 4.2747, "step": 335750 }, { "epoch": 0.6266401367859501, "grad_norm": 0.8680166602134705, "learning_rate": 0.00019632208140004314, "loss": 4.5933, "step": 335800 }, { "epoch": 0.6267334423453286, "grad_norm": 0.9055318236351013, "learning_rate": 0.00019632095246478715, "loss": 4.3666, "step": 335850 }, { "epoch": 0.626826747904707, "grad_norm": 1.0014026165008545, "learning_rate": 0.00019631982335954163, "loss": 4.5995, "step": 335900 }, { "epoch": 0.6269200534640855, "grad_norm": 1.083059549331665, "learning_rate": 0.00019631869408430857, "loss": 4.635, "step": 335950 }, { "epoch": 0.627013359023464, "grad_norm": 0.9607086777687073, "learning_rate": 0.00019631756463908996, "loss": 4.7649, "step": 336000 }, { "epoch": 0.627013359023464, "eval_loss": 4.74292516708374, "eval_runtime": 230.2863, "eval_samples_per_second": 11.325, "eval_steps_per_second": 11.325, "eval_tts_loss": 7.567721385224522, "step": 336000 }, { "epoch": 0.6271066645828425, "grad_norm": 0.7822778224945068, "learning_rate": 0.0001963164350238878, "loss": 4.5734, "step": 336050 }, { "epoch": 0.627199970142221, "grad_norm": 1.253950595855713, "learning_rate": 0.00019631530523870407, "loss": 4.5294, "step": 336100 }, { "epoch": 0.6272932757015995, "grad_norm": 0.8558539748191833, "learning_rate": 0.00019631417528354078, "loss": 4.5185, "step": 336150 }, { "epoch": 0.627386581260978, "grad_norm": 0.9739894270896912, "learning_rate": 0.00019631304515839993, "loss": 4.6097, "step": 336200 }, { "epoch": 0.6274798868203565, "grad_norm": 1.14170503616333, "learning_rate": 0.00019631191486328348, "loss": 4.6264, "step": 336250 }, { "epoch": 0.627573192379735, "grad_norm": 1.0193452835083008, "learning_rate": 0.00019631078439819345, "loss": 4.699, "step": 336300 }, { "epoch": 0.6276664979391134, "grad_norm": 1.1254428625106812, "learning_rate": 0.00019630965376313185, "loss": 4.6496, "step": 336350 }, { "epoch": 0.6277598034984919, "grad_norm": 0.9944987893104553, "learning_rate": 0.00019630852295810067, "loss": 4.5546, "step": 336400 }, { "epoch": 0.6278531090578704, "grad_norm": 1.0571918487548828, "learning_rate": 0.00019630739198310185, "loss": 4.4923, "step": 336450 }, { "epoch": 0.627946414617249, "grad_norm": 1.2111427783966064, "learning_rate": 0.00019630626083813745, "loss": 4.5467, "step": 336500 }, { "epoch": 0.6280397201766275, "grad_norm": 0.6809943318367004, "learning_rate": 0.00019630512952320946, "loss": 4.57, "step": 336550 }, { "epoch": 0.6281330257360059, "grad_norm": 1.091772198677063, "learning_rate": 0.00019630399803831984, "loss": 4.6803, "step": 336600 }, { "epoch": 0.6282263312953844, "grad_norm": 1.1936126947402954, "learning_rate": 0.00019630286638347064, "loss": 4.4784, "step": 336650 }, { "epoch": 0.6283196368547629, "grad_norm": 1.1190840005874634, "learning_rate": 0.00019630173455866378, "loss": 4.7071, "step": 336700 }, { "epoch": 0.6284129424141414, "grad_norm": 1.2292462587356567, "learning_rate": 0.00019630060256390134, "loss": 4.6232, "step": 336750 }, { "epoch": 0.6285062479735198, "grad_norm": 1.1477453708648682, "learning_rate": 0.00019629947039918525, "loss": 4.7842, "step": 336800 }, { "epoch": 0.6285995535328984, "grad_norm": 1.0196152925491333, "learning_rate": 0.00019629833806451757, "loss": 4.6162, "step": 336850 }, { "epoch": 0.6286928590922769, "grad_norm": 1.0956239700317383, "learning_rate": 0.00019629720555990025, "loss": 4.4714, "step": 336900 }, { "epoch": 0.6287861646516554, "grad_norm": 0.8698761463165283, "learning_rate": 0.00019629607288533532, "loss": 4.5083, "step": 336950 }, { "epoch": 0.6288794702110339, "grad_norm": 1.1955455541610718, "learning_rate": 0.00019629494004082476, "loss": 4.585, "step": 337000 }, { "epoch": 0.6289727757704123, "grad_norm": 1.2584973573684692, "learning_rate": 0.00019629380702637053, "loss": 4.405, "step": 337050 }, { "epoch": 0.6290660813297908, "grad_norm": 0.8850629925727844, "learning_rate": 0.00019629267384197474, "loss": 4.5174, "step": 337100 }, { "epoch": 0.6291593868891693, "grad_norm": 0.861888587474823, "learning_rate": 0.00019629154048763929, "loss": 4.4397, "step": 337150 }, { "epoch": 0.6292526924485479, "grad_norm": 0.9065359830856323, "learning_rate": 0.00019629040696336622, "loss": 4.785, "step": 337200 }, { "epoch": 0.6293459980079263, "grad_norm": 0.9942500591278076, "learning_rate": 0.0001962892732691575, "loss": 4.5742, "step": 337250 }, { "epoch": 0.6294393035673048, "grad_norm": 1.2220995426177979, "learning_rate": 0.0001962881394050152, "loss": 4.6268, "step": 337300 }, { "epoch": 0.6295326091266833, "grad_norm": 1.0713739395141602, "learning_rate": 0.00019628700537094128, "loss": 4.3832, "step": 337350 }, { "epoch": 0.6296259146860618, "grad_norm": 0.8994389772415161, "learning_rate": 0.0001962858711669377, "loss": 4.7237, "step": 337400 }, { "epoch": 0.6297192202454402, "grad_norm": 1.032889723777771, "learning_rate": 0.0001962847367930065, "loss": 4.6565, "step": 337450 }, { "epoch": 0.6298125258048187, "grad_norm": 0.8554074764251709, "learning_rate": 0.0001962836022491497, "loss": 4.6186, "step": 337500 }, { "epoch": 0.6299058313641973, "grad_norm": 0.938651978969574, "learning_rate": 0.0001962824675353693, "loss": 4.5568, "step": 337550 }, { "epoch": 0.6299991369235758, "grad_norm": 0.7361330986022949, "learning_rate": 0.00019628133265166725, "loss": 4.5691, "step": 337600 }, { "epoch": 0.6300924424829543, "grad_norm": 0.9456680417060852, "learning_rate": 0.0001962801975980456, "loss": 4.6024, "step": 337650 }, { "epoch": 0.6301857480423327, "grad_norm": 1.1112492084503174, "learning_rate": 0.00019627906237450638, "loss": 4.5557, "step": 337700 }, { "epoch": 0.6302790536017112, "grad_norm": 1.0722355842590332, "learning_rate": 0.0001962779269810515, "loss": 4.5067, "step": 337750 }, { "epoch": 0.6303723591610897, "grad_norm": 1.1839230060577393, "learning_rate": 0.00019627679141768305, "loss": 4.7353, "step": 337800 }, { "epoch": 0.6304656647204682, "grad_norm": 1.101914644241333, "learning_rate": 0.00019627565568440297, "loss": 4.7507, "step": 337850 }, { "epoch": 0.6305589702798468, "grad_norm": 0.9712167382240295, "learning_rate": 0.00019627451978121335, "loss": 4.4696, "step": 337900 }, { "epoch": 0.6306522758392252, "grad_norm": 1.0436922311782837, "learning_rate": 0.0001962733837081161, "loss": 4.4775, "step": 337950 }, { "epoch": 0.6307455813986037, "grad_norm": 1.427233338356018, "learning_rate": 0.0001962722474651133, "loss": 4.6929, "step": 338000 }, { "epoch": 0.6308388869579822, "grad_norm": 0.9068809747695923, "learning_rate": 0.0001962711110522069, "loss": 4.6332, "step": 338050 }, { "epoch": 0.6309321925173607, "grad_norm": 0.6984463930130005, "learning_rate": 0.00019626997446939895, "loss": 4.6269, "step": 338100 }, { "epoch": 0.6310254980767391, "grad_norm": 1.048166275024414, "learning_rate": 0.0001962688377166914, "loss": 4.4084, "step": 338150 }, { "epoch": 0.6311188036361176, "grad_norm": 0.901068389415741, "learning_rate": 0.0001962677007940863, "loss": 4.6114, "step": 338200 }, { "epoch": 0.6312121091954962, "grad_norm": 1.1569806337356567, "learning_rate": 0.00019626656370158566, "loss": 4.7474, "step": 338250 }, { "epoch": 0.6313054147548747, "grad_norm": 1.0751103162765503, "learning_rate": 0.0001962654264391915, "loss": 4.5038, "step": 338300 }, { "epoch": 0.6313987203142531, "grad_norm": 0.8912768959999084, "learning_rate": 0.00019626428900690573, "loss": 4.4918, "step": 338350 }, { "epoch": 0.6314920258736316, "grad_norm": 1.0260379314422607, "learning_rate": 0.00019626315140473048, "loss": 4.6335, "step": 338400 }, { "epoch": 0.6315853314330101, "grad_norm": 1.0572807788848877, "learning_rate": 0.0001962620136326677, "loss": 4.6774, "step": 338450 }, { "epoch": 0.6316786369923886, "grad_norm": 0.8457328677177429, "learning_rate": 0.00019626087569071942, "loss": 4.5486, "step": 338500 }, { "epoch": 0.631771942551767, "grad_norm": 0.9951227903366089, "learning_rate": 0.0001962597375788876, "loss": 4.4332, "step": 338550 }, { "epoch": 0.6318652481111455, "grad_norm": 0.9773802757263184, "learning_rate": 0.0001962585992971743, "loss": 4.7251, "step": 338600 }, { "epoch": 0.6319585536705241, "grad_norm": 1.052430272102356, "learning_rate": 0.0001962574608455815, "loss": 4.4294, "step": 338650 }, { "epoch": 0.6320518592299026, "grad_norm": 1.225507140159607, "learning_rate": 0.0001962563222241112, "loss": 4.7384, "step": 338700 }, { "epoch": 0.6321451647892811, "grad_norm": 1.1235377788543701, "learning_rate": 0.00019625518343276545, "loss": 4.3392, "step": 338750 }, { "epoch": 0.6322384703486595, "grad_norm": 1.0696687698364258, "learning_rate": 0.00019625404447154625, "loss": 4.6364, "step": 338800 }, { "epoch": 0.632331775908038, "grad_norm": 0.9930275678634644, "learning_rate": 0.0001962529053404556, "loss": 4.6083, "step": 338850 }, { "epoch": 0.6324250814674165, "grad_norm": 1.2023861408233643, "learning_rate": 0.0001962517660394955, "loss": 4.6714, "step": 338900 }, { "epoch": 0.632518387026795, "grad_norm": 0.8565691709518433, "learning_rate": 0.00019625062656866793, "loss": 4.7474, "step": 338950 }, { "epoch": 0.6326116925861736, "grad_norm": 1.2467950582504272, "learning_rate": 0.00019624948692797498, "loss": 4.4604, "step": 339000 }, { "epoch": 0.6326116925861736, "eval_loss": 4.749098300933838, "eval_runtime": 229.0294, "eval_samples_per_second": 11.387, "eval_steps_per_second": 11.387, "eval_tts_loss": 7.533419189001919, "step": 339000 }, { "epoch": 0.632704998145552, "grad_norm": 0.9784042835235596, "learning_rate": 0.00019624834711741863, "loss": 4.6181, "step": 339050 }, { "epoch": 0.6327983037049305, "grad_norm": 1.1489437818527222, "learning_rate": 0.00019624720713700084, "loss": 4.5812, "step": 339100 }, { "epoch": 0.632891609264309, "grad_norm": 1.233275055885315, "learning_rate": 0.0001962460669867237, "loss": 4.7015, "step": 339150 }, { "epoch": 0.6329849148236875, "grad_norm": 0.9534322023391724, "learning_rate": 0.00019624492666658915, "loss": 4.589, "step": 339200 }, { "epoch": 0.6330782203830659, "grad_norm": 1.1035577058792114, "learning_rate": 0.0001962437861765993, "loss": 4.6331, "step": 339250 }, { "epoch": 0.6331715259424444, "grad_norm": 1.225083589553833, "learning_rate": 0.00019624264551675605, "loss": 4.6044, "step": 339300 }, { "epoch": 0.633264831501823, "grad_norm": 1.2086414098739624, "learning_rate": 0.00019624150468706148, "loss": 4.3887, "step": 339350 }, { "epoch": 0.6333581370612015, "grad_norm": 1.15715491771698, "learning_rate": 0.00019624036368751758, "loss": 4.7368, "step": 339400 }, { "epoch": 0.63345144262058, "grad_norm": 1.059969425201416, "learning_rate": 0.0001962392225181264, "loss": 4.6418, "step": 339450 }, { "epoch": 0.6335447481799584, "grad_norm": 1.1142557859420776, "learning_rate": 0.00019623808117888992, "loss": 4.6561, "step": 339500 }, { "epoch": 0.6336380537393369, "grad_norm": 1.1521546840667725, "learning_rate": 0.00019623693966981014, "loss": 4.6243, "step": 339550 }, { "epoch": 0.6337313592987154, "grad_norm": 1.0312284231185913, "learning_rate": 0.0001962357979908891, "loss": 4.698, "step": 339600 }, { "epoch": 0.6338246648580939, "grad_norm": 1.1738924980163574, "learning_rate": 0.00019623465614212882, "loss": 4.3374, "step": 339650 }, { "epoch": 0.6339179704174724, "grad_norm": 1.2590388059616089, "learning_rate": 0.0001962335141235313, "loss": 4.6054, "step": 339700 }, { "epoch": 0.6340112759768509, "grad_norm": 1.1100465059280396, "learning_rate": 0.00019623237193509857, "loss": 4.822, "step": 339750 }, { "epoch": 0.6341045815362294, "grad_norm": 1.1239413022994995, "learning_rate": 0.00019623122957683261, "loss": 4.5495, "step": 339800 }, { "epoch": 0.6341978870956079, "grad_norm": 1.021667242050171, "learning_rate": 0.00019623008704873547, "loss": 4.6736, "step": 339850 }, { "epoch": 0.6342911926549863, "grad_norm": 0.9145529270172119, "learning_rate": 0.00019622894435080917, "loss": 4.4413, "step": 339900 }, { "epoch": 0.6343844982143648, "grad_norm": 1.1026692390441895, "learning_rate": 0.00019622780148305572, "loss": 4.7521, "step": 339950 }, { "epoch": 0.6344778037737433, "grad_norm": 1.013266921043396, "learning_rate": 0.00019622665844547713, "loss": 4.6753, "step": 340000 }, { "epoch": 0.6345711093331219, "grad_norm": 0.95827317237854, "learning_rate": 0.00019622551523807542, "loss": 4.4974, "step": 340050 }, { "epoch": 0.6346644148925004, "grad_norm": 1.094130277633667, "learning_rate": 0.00019622437186085257, "loss": 4.7321, "step": 340100 }, { "epoch": 0.6347577204518788, "grad_norm": 0.9615687727928162, "learning_rate": 0.00019622322831381068, "loss": 4.4676, "step": 340150 }, { "epoch": 0.6348510260112573, "grad_norm": 0.982330858707428, "learning_rate": 0.0001962220845969517, "loss": 4.4142, "step": 340200 }, { "epoch": 0.6349443315706358, "grad_norm": 1.1653809547424316, "learning_rate": 0.0001962209407102777, "loss": 4.5996, "step": 340250 }, { "epoch": 0.6350376371300143, "grad_norm": 1.0871142148971558, "learning_rate": 0.00019621979665379065, "loss": 4.6704, "step": 340300 }, { "epoch": 0.6351309426893927, "grad_norm": 0.9514403939247131, "learning_rate": 0.00019621865242749258, "loss": 4.4426, "step": 340350 }, { "epoch": 0.6352242482487713, "grad_norm": 1.1947722434997559, "learning_rate": 0.0001962175080313855, "loss": 4.6943, "step": 340400 }, { "epoch": 0.6353175538081498, "grad_norm": 0.9871910810470581, "learning_rate": 0.00019621636346547149, "loss": 4.5358, "step": 340450 }, { "epoch": 0.6354108593675283, "grad_norm": 1.0049558877944946, "learning_rate": 0.00019621521872975254, "loss": 4.4537, "step": 340500 }, { "epoch": 0.6355041649269068, "grad_norm": 0.8265572190284729, "learning_rate": 0.00019621407382423064, "loss": 4.5113, "step": 340550 }, { "epoch": 0.6355974704862852, "grad_norm": 1.2362648248672485, "learning_rate": 0.0001962129287489078, "loss": 4.5154, "step": 340600 }, { "epoch": 0.6356907760456637, "grad_norm": 1.2209618091583252, "learning_rate": 0.00019621178350378608, "loss": 4.5803, "step": 340650 }, { "epoch": 0.6357840816050422, "grad_norm": 1.2041095495224, "learning_rate": 0.00019621063808886752, "loss": 4.8815, "step": 340700 }, { "epoch": 0.6358773871644208, "grad_norm": 1.169942855834961, "learning_rate": 0.00019620949250415408, "loss": 4.6944, "step": 340750 }, { "epoch": 0.6359706927237992, "grad_norm": 0.9905920624732971, "learning_rate": 0.00019620834674964784, "loss": 4.7523, "step": 340800 }, { "epoch": 0.6360639982831777, "grad_norm": 1.0956676006317139, "learning_rate": 0.0001962072008253508, "loss": 4.5541, "step": 340850 }, { "epoch": 0.6361573038425562, "grad_norm": 1.159255027770996, "learning_rate": 0.00019620605473126497, "loss": 4.5287, "step": 340900 }, { "epoch": 0.6362506094019347, "grad_norm": 1.0262867212295532, "learning_rate": 0.00019620490846739236, "loss": 4.6522, "step": 340950 }, { "epoch": 0.6363439149613132, "grad_norm": 0.8731298446655273, "learning_rate": 0.00019620376203373501, "loss": 4.4661, "step": 341000 }, { "epoch": 0.6364372205206916, "grad_norm": 1.0459544658660889, "learning_rate": 0.000196202615430295, "loss": 4.5979, "step": 341050 }, { "epoch": 0.6365305260800702, "grad_norm": 0.9693548679351807, "learning_rate": 0.00019620146865707426, "loss": 4.5553, "step": 341100 }, { "epoch": 0.6366238316394487, "grad_norm": 1.110594630241394, "learning_rate": 0.00019620032171407487, "loss": 4.402, "step": 341150 }, { "epoch": 0.6367171371988272, "grad_norm": 0.8167380094528198, "learning_rate": 0.00019619917460129882, "loss": 4.5829, "step": 341200 }, { "epoch": 0.6368104427582056, "grad_norm": 1.1120448112487793, "learning_rate": 0.00019619802731874817, "loss": 4.4741, "step": 341250 }, { "epoch": 0.6369037483175841, "grad_norm": 1.108361840248108, "learning_rate": 0.00019619687986642493, "loss": 4.4353, "step": 341300 }, { "epoch": 0.6369970538769626, "grad_norm": 0.9473559260368347, "learning_rate": 0.00019619573224433112, "loss": 4.6361, "step": 341350 }, { "epoch": 0.6370903594363411, "grad_norm": 1.1996275186538696, "learning_rate": 0.00019619458445246877, "loss": 4.6839, "step": 341400 }, { "epoch": 0.6371836649957197, "grad_norm": 0.9558074474334717, "learning_rate": 0.0001961934364908399, "loss": 4.5789, "step": 341450 }, { "epoch": 0.6372769705550981, "grad_norm": 0.9312458038330078, "learning_rate": 0.00019619228835944654, "loss": 4.4597, "step": 341500 }, { "epoch": 0.6373702761144766, "grad_norm": 0.9136353731155396, "learning_rate": 0.00019619114005829072, "loss": 4.5402, "step": 341550 }, { "epoch": 0.6374635816738551, "grad_norm": 1.1716053485870361, "learning_rate": 0.00019618999158737448, "loss": 4.6056, "step": 341600 }, { "epoch": 0.6375568872332336, "grad_norm": 0.9839272499084473, "learning_rate": 0.0001961888429466998, "loss": 4.636, "step": 341650 }, { "epoch": 0.637650192792612, "grad_norm": 1.1562862396240234, "learning_rate": 0.00019618769413626874, "loss": 4.5179, "step": 341700 }, { "epoch": 0.6377434983519905, "grad_norm": 1.0224076509475708, "learning_rate": 0.00019618654515608333, "loss": 4.3989, "step": 341750 }, { "epoch": 0.6378368039113691, "grad_norm": 1.2475993633270264, "learning_rate": 0.0001961853960061456, "loss": 4.4837, "step": 341800 }, { "epoch": 0.6379301094707476, "grad_norm": 1.106919765472412, "learning_rate": 0.00019618424668645757, "loss": 4.5131, "step": 341850 }, { "epoch": 0.638023415030126, "grad_norm": 0.9902898073196411, "learning_rate": 0.00019618309719702127, "loss": 4.5768, "step": 341900 }, { "epoch": 0.6381167205895045, "grad_norm": 1.1790491342544556, "learning_rate": 0.00019618194753783872, "loss": 4.5059, "step": 341950 }, { "epoch": 0.638210026148883, "grad_norm": 1.2637087106704712, "learning_rate": 0.00019618079770891196, "loss": 4.4652, "step": 342000 }, { "epoch": 0.638210026148883, "eval_loss": 4.740108489990234, "eval_runtime": 231.3959, "eval_samples_per_second": 11.271, "eval_steps_per_second": 11.271, "eval_tts_loss": 7.586605188102559, "step": 342000 }, { "epoch": 0.6383033317082615, "grad_norm": 1.101304531097412, "learning_rate": 0.00019617964771024302, "loss": 4.8453, "step": 342050 }, { "epoch": 0.63839663726764, "grad_norm": 1.254040002822876, "learning_rate": 0.00019617849754183394, "loss": 4.5925, "step": 342100 }, { "epoch": 0.6384899428270185, "grad_norm": 0.9550309777259827, "learning_rate": 0.0001961773472036867, "loss": 4.6263, "step": 342150 }, { "epoch": 0.638583248386397, "grad_norm": 0.9817168712615967, "learning_rate": 0.0001961761966958034, "loss": 4.4791, "step": 342200 }, { "epoch": 0.6386765539457755, "grad_norm": 0.9158251285552979, "learning_rate": 0.00019617504601818596, "loss": 4.4927, "step": 342250 }, { "epoch": 0.638769859505154, "grad_norm": 1.2883267402648926, "learning_rate": 0.00019617389517083655, "loss": 4.6163, "step": 342300 }, { "epoch": 0.6388631650645324, "grad_norm": 0.8760630488395691, "learning_rate": 0.00019617274415375716, "loss": 4.6279, "step": 342350 }, { "epoch": 0.6389564706239109, "grad_norm": 1.0041249990463257, "learning_rate": 0.00019617159296694974, "loss": 4.4317, "step": 342400 }, { "epoch": 0.6390497761832894, "grad_norm": 1.0369528532028198, "learning_rate": 0.00019617044161041643, "loss": 4.6907, "step": 342450 }, { "epoch": 0.639143081742668, "grad_norm": 0.9990165829658508, "learning_rate": 0.0001961692900841592, "loss": 4.5248, "step": 342500 }, { "epoch": 0.6392363873020465, "grad_norm": 1.1646652221679688, "learning_rate": 0.00019616813838818008, "loss": 4.5413, "step": 342550 }, { "epoch": 0.6393296928614249, "grad_norm": 0.7702269554138184, "learning_rate": 0.00019616698652248112, "loss": 4.5439, "step": 342600 }, { "epoch": 0.6394229984208034, "grad_norm": 1.0360941886901855, "learning_rate": 0.00019616583448706435, "loss": 4.6135, "step": 342650 }, { "epoch": 0.6395163039801819, "grad_norm": 1.2609333992004395, "learning_rate": 0.0001961646822819318, "loss": 4.5714, "step": 342700 }, { "epoch": 0.6396096095395604, "grad_norm": 1.1983611583709717, "learning_rate": 0.0001961635299070855, "loss": 4.722, "step": 342750 }, { "epoch": 0.6397029150989388, "grad_norm": 1.2513660192489624, "learning_rate": 0.0001961623773625275, "loss": 4.5354, "step": 342800 }, { "epoch": 0.6397962206583174, "grad_norm": 0.9544387459754944, "learning_rate": 0.00019616122464825983, "loss": 4.597, "step": 342850 }, { "epoch": 0.6398895262176959, "grad_norm": 0.9964279532432556, "learning_rate": 0.0001961600717642845, "loss": 4.7352, "step": 342900 }, { "epoch": 0.6399828317770744, "grad_norm": 0.8601515889167786, "learning_rate": 0.0001961589187106036, "loss": 4.1705, "step": 342950 }, { "epoch": 0.6400761373364529, "grad_norm": 1.0321756601333618, "learning_rate": 0.0001961577654872191, "loss": 4.6684, "step": 343000 }, { "epoch": 0.6401694428958313, "grad_norm": 1.2796767950057983, "learning_rate": 0.00019615661209413307, "loss": 4.6704, "step": 343050 }, { "epoch": 0.6402627484552098, "grad_norm": 1.3482799530029297, "learning_rate": 0.00019615545853134754, "loss": 4.3253, "step": 343100 }, { "epoch": 0.6403560540145883, "grad_norm": 1.0156035423278809, "learning_rate": 0.00019615430479886454, "loss": 4.5649, "step": 343150 }, { "epoch": 0.6404493595739669, "grad_norm": 0.8863063454627991, "learning_rate": 0.00019615315089668611, "loss": 4.6938, "step": 343200 }, { "epoch": 0.6405426651333453, "grad_norm": 0.9971023797988892, "learning_rate": 0.0001961519968248143, "loss": 4.5935, "step": 343250 }, { "epoch": 0.6406359706927238, "grad_norm": 0.84346604347229, "learning_rate": 0.00019615084258325113, "loss": 4.6598, "step": 343300 }, { "epoch": 0.6407292762521023, "grad_norm": 0.9728134274482727, "learning_rate": 0.00019614968817199864, "loss": 4.6571, "step": 343350 }, { "epoch": 0.6408225818114808, "grad_norm": 2.26699161529541, "learning_rate": 0.00019614853359105884, "loss": 4.4916, "step": 343400 }, { "epoch": 0.6409158873708592, "grad_norm": 1.1074600219726562, "learning_rate": 0.00019614737884043383, "loss": 4.5896, "step": 343450 }, { "epoch": 0.6410091929302377, "grad_norm": 0.9999982118606567, "learning_rate": 0.0001961462239201256, "loss": 4.5574, "step": 343500 }, { "epoch": 0.6411024984896163, "grad_norm": 1.1165833473205566, "learning_rate": 0.00019614506883013622, "loss": 4.5864, "step": 343550 }, { "epoch": 0.6411958040489948, "grad_norm": 0.9048035740852356, "learning_rate": 0.0001961439135704677, "loss": 4.6054, "step": 343600 }, { "epoch": 0.6412891096083733, "grad_norm": 1.244683861732483, "learning_rate": 0.00019614275814112207, "loss": 4.695, "step": 343650 }, { "epoch": 0.6413824151677517, "grad_norm": 0.6611224412918091, "learning_rate": 0.0001961416025421014, "loss": 4.4653, "step": 343700 }, { "epoch": 0.6414757207271302, "grad_norm": 0.6442915797233582, "learning_rate": 0.00019614044677340771, "loss": 4.5202, "step": 343750 }, { "epoch": 0.6415690262865087, "grad_norm": 1.1567631959915161, "learning_rate": 0.0001961392908350431, "loss": 4.6907, "step": 343800 }, { "epoch": 0.6416623318458872, "grad_norm": 0.9971233606338501, "learning_rate": 0.00019613813472700952, "loss": 4.5955, "step": 343850 }, { "epoch": 0.6417556374052656, "grad_norm": 1.0582727193832397, "learning_rate": 0.00019613697844930906, "loss": 4.5997, "step": 343900 }, { "epoch": 0.6418489429646442, "grad_norm": 1.181840419769287, "learning_rate": 0.0001961358220019437, "loss": 4.7453, "step": 343950 }, { "epoch": 0.6419422485240227, "grad_norm": 0.9383758902549744, "learning_rate": 0.00019613466538491559, "loss": 4.5304, "step": 344000 }, { "epoch": 0.6420355540834012, "grad_norm": 1.198830485343933, "learning_rate": 0.00019613350859822669, "loss": 4.4426, "step": 344050 }, { "epoch": 0.6421288596427797, "grad_norm": 1.2523542642593384, "learning_rate": 0.00019613235164187907, "loss": 4.4969, "step": 344100 }, { "epoch": 0.6422221652021581, "grad_norm": 0.9709693789482117, "learning_rate": 0.00019613119451587473, "loss": 4.5729, "step": 344150 }, { "epoch": 0.6423154707615366, "grad_norm": 1.276157021522522, "learning_rate": 0.00019613003722021577, "loss": 4.6016, "step": 344200 }, { "epoch": 0.6424087763209151, "grad_norm": 1.1887445449829102, "learning_rate": 0.0001961288797549042, "loss": 4.8722, "step": 344250 }, { "epoch": 0.6425020818802937, "grad_norm": 1.132991909980774, "learning_rate": 0.0001961277221199421, "loss": 4.5887, "step": 344300 }, { "epoch": 0.6425953874396721, "grad_norm": 0.9599609971046448, "learning_rate": 0.00019612656431533147, "loss": 4.5974, "step": 344350 }, { "epoch": 0.6426886929990506, "grad_norm": 1.0121450424194336, "learning_rate": 0.00019612540634107438, "loss": 4.6816, "step": 344400 }, { "epoch": 0.6427819985584291, "grad_norm": 0.8555306196212769, "learning_rate": 0.00019612424819717282, "loss": 4.4254, "step": 344450 }, { "epoch": 0.6428753041178076, "grad_norm": 1.1181089878082275, "learning_rate": 0.00019612308988362892, "loss": 4.4815, "step": 344500 }, { "epoch": 0.642968609677186, "grad_norm": 1.0256164073944092, "learning_rate": 0.00019612193140044465, "loss": 4.5175, "step": 344550 }, { "epoch": 0.6430619152365645, "grad_norm": 0.7289327383041382, "learning_rate": 0.00019612077274762212, "loss": 4.3594, "step": 344600 }, { "epoch": 0.6431552207959431, "grad_norm": 1.0339686870574951, "learning_rate": 0.00019611961392516328, "loss": 4.4817, "step": 344650 }, { "epoch": 0.6432485263553216, "grad_norm": 0.9205882549285889, "learning_rate": 0.00019611845493307028, "loss": 4.6025, "step": 344700 }, { "epoch": 0.6433418319147001, "grad_norm": 1.0576496124267578, "learning_rate": 0.00019611729577134513, "loss": 4.3981, "step": 344750 }, { "epoch": 0.6434351374740785, "grad_norm": 0.7700652480125427, "learning_rate": 0.00019611613643998982, "loss": 4.5055, "step": 344800 }, { "epoch": 0.643528443033457, "grad_norm": 0.9929338097572327, "learning_rate": 0.00019611497693900647, "loss": 4.6463, "step": 344850 }, { "epoch": 0.6436217485928355, "grad_norm": 0.9840719103813171, "learning_rate": 0.0001961138172683971, "loss": 4.6944, "step": 344900 }, { "epoch": 0.643715054152214, "grad_norm": 1.1600414514541626, "learning_rate": 0.00019611265742816374, "loss": 4.5144, "step": 344950 }, { "epoch": 0.6438083597115926, "grad_norm": 1.071171760559082, "learning_rate": 0.00019611149741830845, "loss": 4.6351, "step": 345000 }, { "epoch": 0.6438083597115926, "eval_loss": 4.739168643951416, "eval_runtime": 233.3772, "eval_samples_per_second": 11.175, "eval_steps_per_second": 11.175, "eval_tts_loss": 7.525498071853986, "step": 345000 }, { "epoch": 0.643901665270971, "grad_norm": 1.0771901607513428, "learning_rate": 0.00019611033723883329, "loss": 4.7873, "step": 345050 }, { "epoch": 0.6439949708303495, "grad_norm": 0.9674336314201355, "learning_rate": 0.0001961091768897403, "loss": 4.3937, "step": 345100 }, { "epoch": 0.644088276389728, "grad_norm": 0.6305899620056152, "learning_rate": 0.00019610801637103152, "loss": 4.3422, "step": 345150 }, { "epoch": 0.6441815819491065, "grad_norm": 1.0925474166870117, "learning_rate": 0.000196106855682709, "loss": 4.536, "step": 345200 }, { "epoch": 0.6442748875084849, "grad_norm": 0.9010369777679443, "learning_rate": 0.00019610569482477478, "loss": 4.6296, "step": 345250 }, { "epoch": 0.6443681930678634, "grad_norm": 1.1146564483642578, "learning_rate": 0.00019610453379723092, "loss": 4.5091, "step": 345300 }, { "epoch": 0.644461498627242, "grad_norm": 0.842292845249176, "learning_rate": 0.00019610337260007947, "loss": 4.6014, "step": 345350 }, { "epoch": 0.6445548041866205, "grad_norm": 1.0273958444595337, "learning_rate": 0.00019610221123332247, "loss": 4.571, "step": 345400 }, { "epoch": 0.644648109745999, "grad_norm": 1.2018096446990967, "learning_rate": 0.00019610104969696203, "loss": 4.6055, "step": 345450 }, { "epoch": 0.6447414153053774, "grad_norm": 1.0792499780654907, "learning_rate": 0.0001960998879910001, "loss": 4.4501, "step": 345500 }, { "epoch": 0.6448347208647559, "grad_norm": 1.2928937673568726, "learning_rate": 0.0001960987261154388, "loss": 4.7066, "step": 345550 }, { "epoch": 0.6449280264241344, "grad_norm": 0.9761016964912415, "learning_rate": 0.0001960975640702801, "loss": 4.448, "step": 345600 }, { "epoch": 0.6450213319835129, "grad_norm": 1.1374536752700806, "learning_rate": 0.00019609640185552616, "loss": 4.7384, "step": 345650 }, { "epoch": 0.6451146375428914, "grad_norm": 1.1401821374893188, "learning_rate": 0.00019609523947117897, "loss": 4.6449, "step": 345700 }, { "epoch": 0.6452079431022699, "grad_norm": 1.1021229028701782, "learning_rate": 0.0001960940769172406, "loss": 4.6003, "step": 345750 }, { "epoch": 0.6453012486616484, "grad_norm": 1.2322521209716797, "learning_rate": 0.0001960929141937131, "loss": 4.6746, "step": 345800 }, { "epoch": 0.6453945542210269, "grad_norm": 1.06720769405365, "learning_rate": 0.00019609175130059847, "loss": 4.6942, "step": 345850 }, { "epoch": 0.6454878597804053, "grad_norm": 1.0128451585769653, "learning_rate": 0.00019609058823789882, "loss": 4.5013, "step": 345900 }, { "epoch": 0.6455811653397838, "grad_norm": 0.9844326376914978, "learning_rate": 0.00019608942500561622, "loss": 4.6259, "step": 345950 }, { "epoch": 0.6456744708991623, "grad_norm": 1.1689870357513428, "learning_rate": 0.00019608826160375265, "loss": 4.673, "step": 346000 }, { "epoch": 0.6457677764585409, "grad_norm": 1.0861157178878784, "learning_rate": 0.00019608709803231028, "loss": 4.5371, "step": 346050 }, { "epoch": 0.6458610820179194, "grad_norm": 0.9207714200019836, "learning_rate": 0.00019608593429129104, "loss": 4.6154, "step": 346100 }, { "epoch": 0.6459543875772978, "grad_norm": 1.0358107089996338, "learning_rate": 0.000196084770380697, "loss": 4.5926, "step": 346150 }, { "epoch": 0.6460476931366763, "grad_norm": 1.0891337394714355, "learning_rate": 0.0001960836063005303, "loss": 4.3183, "step": 346200 }, { "epoch": 0.6461409986960548, "grad_norm": 0.8938190937042236, "learning_rate": 0.00019608244205079293, "loss": 4.5577, "step": 346250 }, { "epoch": 0.6462343042554333, "grad_norm": 1.108068823814392, "learning_rate": 0.00019608127763148691, "loss": 4.3522, "step": 346300 }, { "epoch": 0.6463276098148117, "grad_norm": 0.9931000471115112, "learning_rate": 0.00019608011304261437, "loss": 4.716, "step": 346350 }, { "epoch": 0.6464209153741903, "grad_norm": 0.6367947459220886, "learning_rate": 0.00019607894828417737, "loss": 4.5168, "step": 346400 }, { "epoch": 0.6465142209335688, "grad_norm": 1.301096796989441, "learning_rate": 0.0001960777833561779, "loss": 4.689, "step": 346450 }, { "epoch": 0.6466075264929473, "grad_norm": 1.2469911575317383, "learning_rate": 0.00019607661825861805, "loss": 4.5518, "step": 346500 }, { "epoch": 0.6467008320523258, "grad_norm": 0.7290911078453064, "learning_rate": 0.00019607545299149987, "loss": 4.5329, "step": 346550 }, { "epoch": 0.6467941376117042, "grad_norm": 0.720203697681427, "learning_rate": 0.00019607428755482543, "loss": 4.6408, "step": 346600 }, { "epoch": 0.6468874431710827, "grad_norm": 1.0603928565979004, "learning_rate": 0.00019607312194859676, "loss": 4.6633, "step": 346650 }, { "epoch": 0.6469807487304612, "grad_norm": 1.269835352897644, "learning_rate": 0.00019607195617281597, "loss": 4.6081, "step": 346700 }, { "epoch": 0.6470740542898398, "grad_norm": 0.9368632435798645, "learning_rate": 0.00019607079022748508, "loss": 4.7006, "step": 346750 }, { "epoch": 0.6471673598492182, "grad_norm": 1.0011993646621704, "learning_rate": 0.0001960696241126061, "loss": 4.5515, "step": 346800 }, { "epoch": 0.6472606654085967, "grad_norm": 0.9085133671760559, "learning_rate": 0.0001960684578281812, "loss": 4.5498, "step": 346850 }, { "epoch": 0.6473539709679752, "grad_norm": 1.0289874076843262, "learning_rate": 0.00019606729137421232, "loss": 4.6087, "step": 346900 }, { "epoch": 0.6474472765273537, "grad_norm": 1.0617784261703491, "learning_rate": 0.0001960661247507016, "loss": 4.5837, "step": 346950 }, { "epoch": 0.6475405820867322, "grad_norm": 1.0014612674713135, "learning_rate": 0.00019606495795765108, "loss": 4.3512, "step": 347000 }, { "epoch": 0.6476338876461106, "grad_norm": 0.9258717894554138, "learning_rate": 0.0001960637909950628, "loss": 4.6539, "step": 347050 }, { "epoch": 0.6477271932054892, "grad_norm": 0.7328335642814636, "learning_rate": 0.00019606262386293884, "loss": 4.7303, "step": 347100 }, { "epoch": 0.6478204987648677, "grad_norm": 0.9870677590370178, "learning_rate": 0.00019606145656128122, "loss": 4.4358, "step": 347150 }, { "epoch": 0.6479138043242462, "grad_norm": 1.216899037361145, "learning_rate": 0.00019606028909009208, "loss": 4.6176, "step": 347200 }, { "epoch": 0.6480071098836246, "grad_norm": 1.1379703283309937, "learning_rate": 0.0001960591214493734, "loss": 4.3831, "step": 347250 }, { "epoch": 0.6481004154430031, "grad_norm": 1.1051362752914429, "learning_rate": 0.0001960579536391273, "loss": 4.6166, "step": 347300 }, { "epoch": 0.6481937210023816, "grad_norm": 1.2240265607833862, "learning_rate": 0.00019605678565935578, "loss": 4.5038, "step": 347350 }, { "epoch": 0.6482870265617601, "grad_norm": 0.9971303343772888, "learning_rate": 0.00019605561751006095, "loss": 4.7579, "step": 347400 }, { "epoch": 0.6483803321211387, "grad_norm": 0.888080894947052, "learning_rate": 0.00019605444919124486, "loss": 4.5582, "step": 347450 }, { "epoch": 0.6484736376805171, "grad_norm": 1.0499393939971924, "learning_rate": 0.00019605328070290958, "loss": 4.6159, "step": 347500 }, { "epoch": 0.6485669432398956, "grad_norm": 0.9919920563697815, "learning_rate": 0.00019605211204505712, "loss": 4.521, "step": 347550 }, { "epoch": 0.6486602487992741, "grad_norm": 1.0972288846969604, "learning_rate": 0.00019605094321768963, "loss": 4.4626, "step": 347600 }, { "epoch": 0.6487535543586526, "grad_norm": 1.0431736707687378, "learning_rate": 0.00019604977422080913, "loss": 4.4371, "step": 347650 }, { "epoch": 0.648846859918031, "grad_norm": 1.1692547798156738, "learning_rate": 0.00019604860505441764, "loss": 4.6365, "step": 347700 }, { "epoch": 0.6489401654774095, "grad_norm": 1.0646686553955078, "learning_rate": 0.00019604743571851726, "loss": 4.6142, "step": 347750 }, { "epoch": 0.6490334710367881, "grad_norm": 0.9616683721542358, "learning_rate": 0.00019604626621311008, "loss": 4.6137, "step": 347800 }, { "epoch": 0.6491267765961666, "grad_norm": 1.0787080526351929, "learning_rate": 0.00019604509653819814, "loss": 4.7056, "step": 347850 }, { "epoch": 0.649220082155545, "grad_norm": 1.1852656602859497, "learning_rate": 0.0001960439266937835, "loss": 4.5779, "step": 347900 }, { "epoch": 0.6493133877149235, "grad_norm": 0.8655521869659424, "learning_rate": 0.00019604275667986824, "loss": 4.6498, "step": 347950 }, { "epoch": 0.649406693274302, "grad_norm": 1.107316493988037, "learning_rate": 0.0001960415864964544, "loss": 4.6731, "step": 348000 }, { "epoch": 0.649406693274302, "eval_loss": 4.729848861694336, "eval_runtime": 229.1395, "eval_samples_per_second": 11.382, "eval_steps_per_second": 11.382, "eval_tts_loss": 7.577490967087036, "step": 348000 }, { "epoch": 0.6494999988336805, "grad_norm": 0.722476065158844, "learning_rate": 0.00019604041614354408, "loss": 4.731, "step": 348050 }, { "epoch": 0.649593304393059, "grad_norm": 1.176112413406372, "learning_rate": 0.0001960392456211393, "loss": 4.5264, "step": 348100 }, { "epoch": 0.6496866099524375, "grad_norm": 1.0830116271972656, "learning_rate": 0.00019603807492924214, "loss": 4.3844, "step": 348150 }, { "epoch": 0.649779915511816, "grad_norm": 0.9252073764801025, "learning_rate": 0.0001960369040678547, "loss": 4.736, "step": 348200 }, { "epoch": 0.6498732210711945, "grad_norm": 1.025813102722168, "learning_rate": 0.000196035733036979, "loss": 4.7346, "step": 348250 }, { "epoch": 0.649966526630573, "grad_norm": 0.72474604845047, "learning_rate": 0.00019603456183661715, "loss": 4.6141, "step": 348300 }, { "epoch": 0.6500598321899514, "grad_norm": 1.258725881576538, "learning_rate": 0.0001960333904667712, "loss": 4.4785, "step": 348350 }, { "epoch": 0.6501531377493299, "grad_norm": 0.8761577606201172, "learning_rate": 0.0001960322189274432, "loss": 4.6884, "step": 348400 }, { "epoch": 0.6502464433087084, "grad_norm": 1.0042715072631836, "learning_rate": 0.00019603104721863522, "loss": 4.6647, "step": 348450 }, { "epoch": 0.650339748868087, "grad_norm": 1.3143467903137207, "learning_rate": 0.00019602987534034932, "loss": 4.4965, "step": 348500 }, { "epoch": 0.6504330544274655, "grad_norm": 1.3221840858459473, "learning_rate": 0.00019602870329258763, "loss": 4.5676, "step": 348550 }, { "epoch": 0.6505263599868439, "grad_norm": 1.199800729751587, "learning_rate": 0.0001960275310753522, "loss": 4.4183, "step": 348600 }, { "epoch": 0.6506196655462224, "grad_norm": 0.9423922300338745, "learning_rate": 0.000196026358688645, "loss": 4.4449, "step": 348650 }, { "epoch": 0.6507129711056009, "grad_norm": 1.0496889352798462, "learning_rate": 0.00019602518613246822, "loss": 4.3336, "step": 348700 }, { "epoch": 0.6508062766649794, "grad_norm": 1.4044764041900635, "learning_rate": 0.0001960240134068239, "loss": 4.5279, "step": 348750 }, { "epoch": 0.6508995822243578, "grad_norm": 1.3792707920074463, "learning_rate": 0.00019602284051171406, "loss": 4.5458, "step": 348800 }, { "epoch": 0.6509928877837364, "grad_norm": 1.1532506942749023, "learning_rate": 0.00019602166744714082, "loss": 4.6879, "step": 348850 }, { "epoch": 0.6510861933431149, "grad_norm": 1.7418402433395386, "learning_rate": 0.0001960204942131062, "loss": 4.6084, "step": 348900 }, { "epoch": 0.6511794989024934, "grad_norm": 1.2199933528900146, "learning_rate": 0.00019601932080961235, "loss": 4.6386, "step": 348950 }, { "epoch": 0.6512728044618719, "grad_norm": 1.0035910606384277, "learning_rate": 0.00019601814723666126, "loss": 4.4855, "step": 349000 }, { "epoch": 0.6513661100212503, "grad_norm": 0.8540064096450806, "learning_rate": 0.00019601697349425502, "loss": 4.5604, "step": 349050 }, { "epoch": 0.6514594155806288, "grad_norm": 1.0423887968063354, "learning_rate": 0.00019601579958239575, "loss": 4.4431, "step": 349100 }, { "epoch": 0.6515527211400073, "grad_norm": 0.993720531463623, "learning_rate": 0.00019601462550108547, "loss": 4.5087, "step": 349150 }, { "epoch": 0.6516460266993858, "grad_norm": 1.0810538530349731, "learning_rate": 0.00019601345125032629, "loss": 4.4436, "step": 349200 }, { "epoch": 0.6517393322587643, "grad_norm": 0.9943142533302307, "learning_rate": 0.00019601227683012023, "loss": 4.4959, "step": 349250 }, { "epoch": 0.6518326378181428, "grad_norm": 1.1125222444534302, "learning_rate": 0.00019601110224046943, "loss": 4.6668, "step": 349300 }, { "epoch": 0.6519259433775213, "grad_norm": 0.7313642501831055, "learning_rate": 0.00019600992748137593, "loss": 4.6501, "step": 349350 }, { "epoch": 0.6520192489368998, "grad_norm": 0.9498952627182007, "learning_rate": 0.00019600875255284178, "loss": 4.6597, "step": 349400 }, { "epoch": 0.6521125544962783, "grad_norm": 1.002515435218811, "learning_rate": 0.00019600757745486907, "loss": 4.6331, "step": 349450 }, { "epoch": 0.6522058600556567, "grad_norm": 0.8719493746757507, "learning_rate": 0.0001960064021874599, "loss": 4.4631, "step": 349500 }, { "epoch": 0.6522991656150352, "grad_norm": 1.3565610647201538, "learning_rate": 0.00019600522675061632, "loss": 4.6105, "step": 349550 }, { "epoch": 0.6523924711744138, "grad_norm": 1.4063760042190552, "learning_rate": 0.0001960040511443404, "loss": 4.4435, "step": 349600 }, { "epoch": 0.6524857767337923, "grad_norm": 1.1056792736053467, "learning_rate": 0.0001960028753686342, "loss": 4.8335, "step": 349650 }, { "epoch": 0.6525790822931707, "grad_norm": 0.9742892980575562, "learning_rate": 0.00019600169942349984, "loss": 4.3601, "step": 349700 }, { "epoch": 0.6526723878525492, "grad_norm": 1.1894937753677368, "learning_rate": 0.0001960005233089394, "loss": 4.7548, "step": 349750 }, { "epoch": 0.6527656934119277, "grad_norm": 1.2229342460632324, "learning_rate": 0.0001959993470249549, "loss": 4.4735, "step": 349800 }, { "epoch": 0.6528589989713062, "grad_norm": 1.0048850774765015, "learning_rate": 0.00019599817057154844, "loss": 4.4905, "step": 349850 }, { "epoch": 0.6529523045306846, "grad_norm": 1.1758800745010376, "learning_rate": 0.00019599699394872211, "loss": 4.4454, "step": 349900 }, { "epoch": 0.6530456100900632, "grad_norm": 0.9137585163116455, "learning_rate": 0.00019599581715647798, "loss": 4.4121, "step": 349950 }, { "epoch": 0.6531389156494417, "grad_norm": 1.040393590927124, "learning_rate": 0.0001959946401948181, "loss": 4.6267, "step": 350000 }, { "epoch": 0.6532322212088202, "grad_norm": 1.0021544694900513, "learning_rate": 0.00019599346306374457, "loss": 4.6441, "step": 350050 }, { "epoch": 0.6533255267681987, "grad_norm": 0.8930035829544067, "learning_rate": 0.0001959922857632595, "loss": 4.634, "step": 350100 }, { "epoch": 0.6534188323275771, "grad_norm": 1.142181634902954, "learning_rate": 0.00019599110829336493, "loss": 4.661, "step": 350150 }, { "epoch": 0.6535121378869556, "grad_norm": 1.200129747390747, "learning_rate": 0.00019598993065406294, "loss": 4.563, "step": 350200 }, { "epoch": 0.6536054434463341, "grad_norm": 1.0493800640106201, "learning_rate": 0.00019598875284535558, "loss": 4.5014, "step": 350250 }, { "epoch": 0.6536987490057127, "grad_norm": 0.9547832012176514, "learning_rate": 0.00019598757486724499, "loss": 4.645, "step": 350300 }, { "epoch": 0.6537920545650912, "grad_norm": 1.0982476472854614, "learning_rate": 0.00019598639671973323, "loss": 4.4349, "step": 350350 }, { "epoch": 0.6538853601244696, "grad_norm": 1.0742329359054565, "learning_rate": 0.00019598521840282236, "loss": 4.5784, "step": 350400 }, { "epoch": 0.6539786656838481, "grad_norm": 1.203382134437561, "learning_rate": 0.00019598403991651447, "loss": 4.7065, "step": 350450 }, { "epoch": 0.6540719712432266, "grad_norm": 1.0984567403793335, "learning_rate": 0.00019598286126081162, "loss": 4.6502, "step": 350500 }, { "epoch": 0.6541652768026051, "grad_norm": 1.1957650184631348, "learning_rate": 0.00019598168243571592, "loss": 4.5351, "step": 350550 }, { "epoch": 0.6542585823619835, "grad_norm": 1.264735460281372, "learning_rate": 0.00019598050344122944, "loss": 4.6503, "step": 350600 }, { "epoch": 0.6543518879213621, "grad_norm": 0.8449005484580994, "learning_rate": 0.00019597932427735425, "loss": 4.377, "step": 350650 }, { "epoch": 0.6544451934807406, "grad_norm": 1.2481213808059692, "learning_rate": 0.00019597814494409245, "loss": 4.55, "step": 350700 }, { "epoch": 0.6545384990401191, "grad_norm": 1.057641625404358, "learning_rate": 0.00019597696544144616, "loss": 4.41, "step": 350750 }, { "epoch": 0.6546318045994975, "grad_norm": 0.9384470582008362, "learning_rate": 0.00019597578576941734, "loss": 4.4782, "step": 350800 }, { "epoch": 0.654725110158876, "grad_norm": 0.7581617832183838, "learning_rate": 0.00019597460592800817, "loss": 4.6777, "step": 350850 }, { "epoch": 0.6548184157182545, "grad_norm": 1.164628028869629, "learning_rate": 0.0001959734259172207, "loss": 4.9107, "step": 350900 }, { "epoch": 0.654911721277633, "grad_norm": 1.0642462968826294, "learning_rate": 0.00019597224573705705, "loss": 4.7088, "step": 350950 }, { "epoch": 0.6550050268370116, "grad_norm": 1.156158447265625, "learning_rate": 0.00019597106538751926, "loss": 4.33, "step": 351000 }, { "epoch": 0.6550050268370116, "eval_loss": 4.736559867858887, "eval_runtime": 228.9675, "eval_samples_per_second": 11.39, "eval_steps_per_second": 11.39, "eval_tts_loss": 7.536694011900951, "step": 351000 }, { "epoch": 0.65509833239639, "grad_norm": 0.8227680325508118, "learning_rate": 0.00019596988486860942, "loss": 4.671, "step": 351050 }, { "epoch": 0.6551916379557685, "grad_norm": 1.0057629346847534, "learning_rate": 0.0001959687041803296, "loss": 4.557, "step": 351100 }, { "epoch": 0.655284943515147, "grad_norm": 1.1353752613067627, "learning_rate": 0.00019596752332268194, "loss": 4.4945, "step": 351150 }, { "epoch": 0.6553782490745255, "grad_norm": 1.2325284481048584, "learning_rate": 0.00019596634229566848, "loss": 4.5777, "step": 351200 }, { "epoch": 0.6554715546339039, "grad_norm": 1.2234023809432983, "learning_rate": 0.00019596516109929128, "loss": 4.6234, "step": 351250 }, { "epoch": 0.6555648601932824, "grad_norm": 1.1709771156311035, "learning_rate": 0.00019596397973355249, "loss": 4.5459, "step": 351300 }, { "epoch": 0.655658165752661, "grad_norm": 1.100290298461914, "learning_rate": 0.00019596279819845416, "loss": 4.4377, "step": 351350 }, { "epoch": 0.6557514713120395, "grad_norm": 1.1520344018936157, "learning_rate": 0.00019596161649399835, "loss": 4.3091, "step": 351400 }, { "epoch": 0.655844776871418, "grad_norm": 1.0503275394439697, "learning_rate": 0.00019596043462018718, "loss": 4.6749, "step": 351450 }, { "epoch": 0.6559380824307964, "grad_norm": 0.8855152726173401, "learning_rate": 0.00019595925257702277, "loss": 4.6931, "step": 351500 }, { "epoch": 0.6560313879901749, "grad_norm": 0.8062465786933899, "learning_rate": 0.00019595807036450712, "loss": 4.5645, "step": 351550 }, { "epoch": 0.6561246935495534, "grad_norm": 1.2063981294631958, "learning_rate": 0.00019595688798264237, "loss": 4.6111, "step": 351600 }, { "epoch": 0.6562179991089319, "grad_norm": 1.2455005645751953, "learning_rate": 0.00019595570543143057, "loss": 4.5999, "step": 351650 }, { "epoch": 0.6563113046683104, "grad_norm": 1.0943653583526611, "learning_rate": 0.00019595452271087388, "loss": 4.5354, "step": 351700 }, { "epoch": 0.6564046102276889, "grad_norm": 1.1074237823486328, "learning_rate": 0.00019595333982097432, "loss": 4.4646, "step": 351750 }, { "epoch": 0.6564979157870674, "grad_norm": 1.2968482971191406, "learning_rate": 0.000195952156761734, "loss": 4.5801, "step": 351800 }, { "epoch": 0.6565912213464459, "grad_norm": 1.4533486366271973, "learning_rate": 0.00019595097353315502, "loss": 4.5672, "step": 351850 }, { "epoch": 0.6566845269058244, "grad_norm": 1.104537844657898, "learning_rate": 0.00019594979013523944, "loss": 4.5806, "step": 351900 }, { "epoch": 0.6567778324652028, "grad_norm": 0.7298156023025513, "learning_rate": 0.00019594860656798937, "loss": 4.5813, "step": 351950 }, { "epoch": 0.6568711380245813, "grad_norm": 1.117599606513977, "learning_rate": 0.00019594742283140688, "loss": 4.6431, "step": 352000 }, { "epoch": 0.6569644435839599, "grad_norm": 1.0720120668411255, "learning_rate": 0.00019594623892549406, "loss": 4.558, "step": 352050 }, { "epoch": 0.6570577491433384, "grad_norm": 0.9862524271011353, "learning_rate": 0.00019594505485025305, "loss": 4.5758, "step": 352100 }, { "epoch": 0.6571510547027168, "grad_norm": 1.1951582431793213, "learning_rate": 0.00019594387060568587, "loss": 4.6974, "step": 352150 }, { "epoch": 0.6572443602620953, "grad_norm": 1.1000356674194336, "learning_rate": 0.00019594268619179466, "loss": 4.5496, "step": 352200 }, { "epoch": 0.6573376658214738, "grad_norm": 1.2195320129394531, "learning_rate": 0.00019594150160858148, "loss": 4.4902, "step": 352250 }, { "epoch": 0.6574309713808523, "grad_norm": 1.0285000801086426, "learning_rate": 0.00019594031685604842, "loss": 4.5667, "step": 352300 }, { "epoch": 0.6575242769402307, "grad_norm": 1.1513861417770386, "learning_rate": 0.0001959391319341976, "loss": 4.6403, "step": 352350 }, { "epoch": 0.6576175824996093, "grad_norm": 1.0986008644104004, "learning_rate": 0.0001959379468430311, "loss": 4.6092, "step": 352400 }, { "epoch": 0.6577108880589878, "grad_norm": 1.1743237972259521, "learning_rate": 0.00019593676158255097, "loss": 4.6803, "step": 352450 }, { "epoch": 0.6578041936183663, "grad_norm": 0.9898801445960999, "learning_rate": 0.00019593557615275935, "loss": 4.3212, "step": 352500 }, { "epoch": 0.6578974991777448, "grad_norm": 1.0967254638671875, "learning_rate": 0.00019593439055365833, "loss": 4.5072, "step": 352550 }, { "epoch": 0.6579908047371232, "grad_norm": 1.0041605234146118, "learning_rate": 0.00019593320478524998, "loss": 4.4197, "step": 352600 }, { "epoch": 0.6580841102965017, "grad_norm": 1.07600736618042, "learning_rate": 0.0001959320188475364, "loss": 4.6252, "step": 352650 }, { "epoch": 0.6581774158558802, "grad_norm": 0.9869751930236816, "learning_rate": 0.00019593083274051975, "loss": 4.3965, "step": 352700 }, { "epoch": 0.6582707214152588, "grad_norm": 0.9461277723312378, "learning_rate": 0.00019592964646420198, "loss": 4.4762, "step": 352750 }, { "epoch": 0.6583640269746373, "grad_norm": 1.2213248014450073, "learning_rate": 0.0001959284600185853, "loss": 4.5955, "step": 352800 }, { "epoch": 0.6584573325340157, "grad_norm": 1.1953409910202026, "learning_rate": 0.00019592727340367173, "loss": 4.5722, "step": 352850 }, { "epoch": 0.6585506380933942, "grad_norm": 0.9113812446594238, "learning_rate": 0.00019592608661946344, "loss": 4.5533, "step": 352900 }, { "epoch": 0.6586439436527727, "grad_norm": 1.179384708404541, "learning_rate": 0.00019592489966596249, "loss": 4.6624, "step": 352950 }, { "epoch": 0.6587372492121512, "grad_norm": 1.031244158744812, "learning_rate": 0.00019592371254317095, "loss": 4.4156, "step": 353000 }, { "epoch": 0.6588305547715296, "grad_norm": 1.3826984167099, "learning_rate": 0.00019592252525109092, "loss": 4.5798, "step": 353050 }, { "epoch": 0.6589238603309082, "grad_norm": 0.905318558216095, "learning_rate": 0.00019592133778972452, "loss": 4.6347, "step": 353100 }, { "epoch": 0.6590171658902867, "grad_norm": 0.880496621131897, "learning_rate": 0.00019592015015907385, "loss": 4.4655, "step": 353150 }, { "epoch": 0.6591104714496652, "grad_norm": 0.9531444907188416, "learning_rate": 0.000195918962359141, "loss": 4.6297, "step": 353200 }, { "epoch": 0.6592037770090436, "grad_norm": 1.0310364961624146, "learning_rate": 0.00019591777438992804, "loss": 4.6778, "step": 353250 }, { "epoch": 0.6592970825684221, "grad_norm": 1.0869261026382446, "learning_rate": 0.00019591658625143712, "loss": 4.6961, "step": 353300 }, { "epoch": 0.6593903881278006, "grad_norm": 0.9768805503845215, "learning_rate": 0.00019591539794367027, "loss": 4.772, "step": 353350 }, { "epoch": 0.6594836936871791, "grad_norm": 1.2685409784317017, "learning_rate": 0.0001959142094666296, "loss": 4.4007, "step": 353400 }, { "epoch": 0.6595769992465577, "grad_norm": 0.9491780400276184, "learning_rate": 0.0001959130208203173, "loss": 4.4167, "step": 353450 }, { "epoch": 0.6596703048059361, "grad_norm": 0.995601236820221, "learning_rate": 0.00019591183200473532, "loss": 4.5628, "step": 353500 }, { "epoch": 0.6597636103653146, "grad_norm": 0.6878821849822998, "learning_rate": 0.00019591064301988586, "loss": 4.5096, "step": 353550 }, { "epoch": 0.6598569159246931, "grad_norm": 1.0458654165267944, "learning_rate": 0.000195909453865771, "loss": 4.4981, "step": 353600 }, { "epoch": 0.6599502214840716, "grad_norm": 1.0209881067276, "learning_rate": 0.00019590826454239283, "loss": 4.3819, "step": 353650 }, { "epoch": 0.66004352704345, "grad_norm": 0.883777916431427, "learning_rate": 0.00019590707504975343, "loss": 4.5237, "step": 353700 }, { "epoch": 0.6601368326028285, "grad_norm": 0.9451062679290771, "learning_rate": 0.00019590588538785495, "loss": 4.5695, "step": 353750 }, { "epoch": 0.6602301381622071, "grad_norm": 1.0650455951690674, "learning_rate": 0.00019590469555669942, "loss": 4.5875, "step": 353800 }, { "epoch": 0.6603234437215856, "grad_norm": 0.835949182510376, "learning_rate": 0.000195903505556289, "loss": 4.3907, "step": 353850 }, { "epoch": 0.660416749280964, "grad_norm": 1.0406574010849, "learning_rate": 0.00019590231538662575, "loss": 4.499, "step": 353900 }, { "epoch": 0.6605100548403425, "grad_norm": 0.9681634306907654, "learning_rate": 0.0001959011250477118, "loss": 4.4877, "step": 353950 }, { "epoch": 0.660603360399721, "grad_norm": 1.1119762659072876, "learning_rate": 0.0001958999345395492, "loss": 4.596, "step": 354000 }, { "epoch": 0.660603360399721, "eval_loss": 4.737687587738037, "eval_runtime": 229.9813, "eval_samples_per_second": 11.34, "eval_steps_per_second": 11.34, "eval_tts_loss": 7.5381670642552745, "step": 354000 }, { "epoch": 0.6606966659590995, "grad_norm": 1.0130051374435425, "learning_rate": 0.00019589874386214017, "loss": 4.4346, "step": 354050 }, { "epoch": 0.660789971518478, "grad_norm": 1.2727903127670288, "learning_rate": 0.0001958975530154867, "loss": 4.5517, "step": 354100 }, { "epoch": 0.6608832770778565, "grad_norm": 1.147301435470581, "learning_rate": 0.0001958963619995909, "loss": 4.5535, "step": 354150 }, { "epoch": 0.660976582637235, "grad_norm": 0.9156633615493774, "learning_rate": 0.0001958951708144549, "loss": 4.5594, "step": 354200 }, { "epoch": 0.6610698881966135, "grad_norm": 0.8328635692596436, "learning_rate": 0.0001958939794600808, "loss": 4.5648, "step": 354250 }, { "epoch": 0.661163193755992, "grad_norm": 0.9792226552963257, "learning_rate": 0.0001958927879364707, "loss": 4.6311, "step": 354300 }, { "epoch": 0.6612564993153704, "grad_norm": 1.0784177780151367, "learning_rate": 0.0001958915962436267, "loss": 4.4926, "step": 354350 }, { "epoch": 0.6613498048747489, "grad_norm": 0.9137364029884338, "learning_rate": 0.00019589040438155092, "loss": 4.6259, "step": 354400 }, { "epoch": 0.6614431104341274, "grad_norm": 1.2003227472305298, "learning_rate": 0.00019588921235024541, "loss": 4.6057, "step": 354450 }, { "epoch": 0.661536415993506, "grad_norm": 0.9942118525505066, "learning_rate": 0.00019588802014971238, "loss": 4.5277, "step": 354500 }, { "epoch": 0.6616297215528845, "grad_norm": 1.1411582231521606, "learning_rate": 0.0001958868277799538, "loss": 4.6279, "step": 354550 }, { "epoch": 0.6617230271122629, "grad_norm": 0.9816586971282959, "learning_rate": 0.00019588563524097188, "loss": 4.4651, "step": 354600 }, { "epoch": 0.6618163326716414, "grad_norm": 1.1638094186782837, "learning_rate": 0.00019588444253276866, "loss": 4.6132, "step": 354650 }, { "epoch": 0.6619096382310199, "grad_norm": 1.100682020187378, "learning_rate": 0.0001958832496553463, "loss": 4.4892, "step": 354700 }, { "epoch": 0.6620029437903984, "grad_norm": 1.091890573501587, "learning_rate": 0.00019588205660870685, "loss": 4.5033, "step": 354750 }, { "epoch": 0.6620962493497768, "grad_norm": 1.1308807134628296, "learning_rate": 0.00019588086339285244, "loss": 4.4093, "step": 354800 }, { "epoch": 0.6621895549091553, "grad_norm": 0.8613793849945068, "learning_rate": 0.0001958796700077852, "loss": 4.6998, "step": 354850 }, { "epoch": 0.6622828604685339, "grad_norm": 0.9791272282600403, "learning_rate": 0.00019587847645350718, "loss": 4.4469, "step": 354900 }, { "epoch": 0.6623761660279124, "grad_norm": 0.8936867713928223, "learning_rate": 0.00019587728273002056, "loss": 4.5074, "step": 354950 }, { "epoch": 0.6624694715872909, "grad_norm": 1.1822404861450195, "learning_rate": 0.00019587608883732738, "loss": 4.4717, "step": 355000 }, { "epoch": 0.6625627771466693, "grad_norm": 0.9689282178878784, "learning_rate": 0.00019587489477542976, "loss": 4.56, "step": 355050 }, { "epoch": 0.6626560827060478, "grad_norm": 0.9512212872505188, "learning_rate": 0.00019587370054432985, "loss": 4.6973, "step": 355100 }, { "epoch": 0.6627493882654263, "grad_norm": 1.2343778610229492, "learning_rate": 0.00019587250614402972, "loss": 4.5545, "step": 355150 }, { "epoch": 0.6628426938248048, "grad_norm": 1.140806794166565, "learning_rate": 0.00019587131157453147, "loss": 4.5647, "step": 355200 }, { "epoch": 0.6629359993841833, "grad_norm": 0.8960227966308594, "learning_rate": 0.00019587011683583724, "loss": 4.6904, "step": 355250 }, { "epoch": 0.6630293049435618, "grad_norm": 1.0368685722351074, "learning_rate": 0.00019586892192794913, "loss": 4.5606, "step": 355300 }, { "epoch": 0.6631226105029403, "grad_norm": 1.2226470708847046, "learning_rate": 0.00019586772685086922, "loss": 4.4205, "step": 355350 }, { "epoch": 0.6632159160623188, "grad_norm": 1.3335462808609009, "learning_rate": 0.00019586653160459964, "loss": 4.3505, "step": 355400 }, { "epoch": 0.6633092216216973, "grad_norm": 0.8618627786636353, "learning_rate": 0.00019586533618914253, "loss": 4.5772, "step": 355450 }, { "epoch": 0.6634025271810757, "grad_norm": 0.8240525722503662, "learning_rate": 0.00019586414060449996, "loss": 4.4362, "step": 355500 }, { "epoch": 0.6634958327404542, "grad_norm": 1.1974875926971436, "learning_rate": 0.00019586294485067404, "loss": 4.514, "step": 355550 }, { "epoch": 0.6635891382998328, "grad_norm": 0.8910797238349915, "learning_rate": 0.0001958617489276669, "loss": 4.4302, "step": 355600 }, { "epoch": 0.6636824438592113, "grad_norm": 0.7819907069206238, "learning_rate": 0.0001958605528354806, "loss": 4.5097, "step": 355650 }, { "epoch": 0.6637757494185897, "grad_norm": 0.8999784588813782, "learning_rate": 0.00019585935657411732, "loss": 4.3274, "step": 355700 }, { "epoch": 0.6638690549779682, "grad_norm": 0.8418229222297668, "learning_rate": 0.00019585816014357915, "loss": 4.5235, "step": 355750 }, { "epoch": 0.6639623605373467, "grad_norm": 1.0023940801620483, "learning_rate": 0.0001958569635438682, "loss": 4.4684, "step": 355800 }, { "epoch": 0.6640556660967252, "grad_norm": 1.0244237184524536, "learning_rate": 0.00019585576677498656, "loss": 4.573, "step": 355850 }, { "epoch": 0.6641489716561036, "grad_norm": 1.1590559482574463, "learning_rate": 0.00019585456983693637, "loss": 4.4707, "step": 355900 }, { "epoch": 0.6642422772154822, "grad_norm": 0.977392852306366, "learning_rate": 0.0001958533727297197, "loss": 4.68, "step": 355950 }, { "epoch": 0.6643355827748607, "grad_norm": 1.1569302082061768, "learning_rate": 0.00019585217545333873, "loss": 4.5928, "step": 356000 }, { "epoch": 0.6644288883342392, "grad_norm": 1.1341270208358765, "learning_rate": 0.0001958509780077955, "loss": 4.697, "step": 356050 }, { "epoch": 0.6645221938936177, "grad_norm": 1.0654176473617554, "learning_rate": 0.0001958497803930922, "loss": 4.5901, "step": 356100 }, { "epoch": 0.6646154994529961, "grad_norm": 0.9715544581413269, "learning_rate": 0.0001958485826092309, "loss": 4.3667, "step": 356150 }, { "epoch": 0.6647088050123746, "grad_norm": 0.9358380436897278, "learning_rate": 0.00019584738465621367, "loss": 4.6559, "step": 356200 }, { "epoch": 0.6648021105717531, "grad_norm": 1.0738136768341064, "learning_rate": 0.00019584618653404272, "loss": 4.7552, "step": 356250 }, { "epoch": 0.6648954161311317, "grad_norm": 0.8731696009635925, "learning_rate": 0.0001958449882427201, "loss": 4.589, "step": 356300 }, { "epoch": 0.6649887216905102, "grad_norm": 1.1186859607696533, "learning_rate": 0.00019584378978224794, "loss": 4.4606, "step": 356350 }, { "epoch": 0.6650820272498886, "grad_norm": 1.0473381280899048, "learning_rate": 0.00019584259115262834, "loss": 4.5423, "step": 356400 }, { "epoch": 0.6651753328092671, "grad_norm": 0.7153509259223938, "learning_rate": 0.00019584139235386344, "loss": 4.5403, "step": 356450 }, { "epoch": 0.6652686383686456, "grad_norm": 0.6838157773017883, "learning_rate": 0.00019584019338595536, "loss": 4.5531, "step": 356500 }, { "epoch": 0.6653619439280241, "grad_norm": 1.0864678621292114, "learning_rate": 0.0001958389942489062, "loss": 4.5307, "step": 356550 }, { "epoch": 0.6654552494874025, "grad_norm": 1.2366557121276855, "learning_rate": 0.00019583779494271805, "loss": 4.4318, "step": 356600 }, { "epoch": 0.6655485550467811, "grad_norm": 1.153394103050232, "learning_rate": 0.00019583659546739307, "loss": 4.5182, "step": 356650 }, { "epoch": 0.6656418606061596, "grad_norm": 1.1458213329315186, "learning_rate": 0.0001958353958229334, "loss": 4.6159, "step": 356700 }, { "epoch": 0.6657351661655381, "grad_norm": 1.208235502243042, "learning_rate": 0.00019583419600934106, "loss": 4.614, "step": 356750 }, { "epoch": 0.6658284717249165, "grad_norm": 0.8747703433036804, "learning_rate": 0.00019583299602661826, "loss": 4.7372, "step": 356800 }, { "epoch": 0.665921777284295, "grad_norm": 1.0256723165512085, "learning_rate": 0.00019583179587476707, "loss": 4.4512, "step": 356850 }, { "epoch": 0.6660150828436735, "grad_norm": 1.1350177526474, "learning_rate": 0.00019583059555378963, "loss": 4.4028, "step": 356900 }, { "epoch": 0.666108388403052, "grad_norm": 0.8919464945793152, "learning_rate": 0.00019582939506368807, "loss": 4.5091, "step": 356950 }, { "epoch": 0.6662016939624306, "grad_norm": 1.2155777215957642, "learning_rate": 0.00019582819440446447, "loss": 4.6637, "step": 357000 }, { "epoch": 0.6662016939624306, "eval_loss": 4.740085124969482, "eval_runtime": 233.202, "eval_samples_per_second": 11.183, "eval_steps_per_second": 11.183, "eval_tts_loss": 7.604406293443021, "step": 357000 }, { "epoch": 0.666294999521809, "grad_norm": 1.0060449838638306, "learning_rate": 0.00019582699357612095, "loss": 4.5867, "step": 357050 }, { "epoch": 0.6663883050811875, "grad_norm": 1.1478132009506226, "learning_rate": 0.0001958257925786597, "loss": 4.4616, "step": 357100 }, { "epoch": 0.666481610640566, "grad_norm": 1.0752184391021729, "learning_rate": 0.00019582459141208273, "loss": 4.56, "step": 357150 }, { "epoch": 0.6665749161999445, "grad_norm": 1.305156946182251, "learning_rate": 0.00019582339007639225, "loss": 4.6758, "step": 357200 }, { "epoch": 0.6666682217593229, "grad_norm": 0.9395535588264465, "learning_rate": 0.00019582218857159035, "loss": 4.537, "step": 357250 }, { "epoch": 0.6667615273187014, "grad_norm": 1.31413996219635, "learning_rate": 0.00019582098689767912, "loss": 4.5146, "step": 357300 }, { "epoch": 0.66685483287808, "grad_norm": 1.1975082159042358, "learning_rate": 0.00019581978505466073, "loss": 4.5509, "step": 357350 }, { "epoch": 0.6669481384374585, "grad_norm": 0.9953302145004272, "learning_rate": 0.00019581858304253727, "loss": 4.4775, "step": 357400 }, { "epoch": 0.667041443996837, "grad_norm": 1.0209640264511108, "learning_rate": 0.00019581738086131087, "loss": 4.4213, "step": 357450 }, { "epoch": 0.6671347495562154, "grad_norm": 1.3481035232543945, "learning_rate": 0.00019581617851098366, "loss": 4.6205, "step": 357500 }, { "epoch": 0.6672280551155939, "grad_norm": 1.155086636543274, "learning_rate": 0.00019581497599155775, "loss": 4.5012, "step": 357550 }, { "epoch": 0.6673213606749724, "grad_norm": 1.0793018341064453, "learning_rate": 0.00019581377330303527, "loss": 4.6158, "step": 357600 }, { "epoch": 0.6674146662343509, "grad_norm": 1.0425056219100952, "learning_rate": 0.00019581257044541835, "loss": 4.5017, "step": 357650 }, { "epoch": 0.6675079717937294, "grad_norm": 0.9235154986381531, "learning_rate": 0.00019581136741870906, "loss": 4.5477, "step": 357700 }, { "epoch": 0.6676012773531079, "grad_norm": 0.9167792201042175, "learning_rate": 0.0001958101642229096, "loss": 4.687, "step": 357750 }, { "epoch": 0.6676945829124864, "grad_norm": 0.7527122497558594, "learning_rate": 0.00019580896085802206, "loss": 4.3443, "step": 357800 }, { "epoch": 0.6677878884718649, "grad_norm": 0.9313538670539856, "learning_rate": 0.00019580775732404854, "loss": 4.3117, "step": 357850 }, { "epoch": 0.6678811940312434, "grad_norm": 0.930475115776062, "learning_rate": 0.0001958065536209912, "loss": 4.4561, "step": 357900 }, { "epoch": 0.6679744995906218, "grad_norm": 0.9051177501678467, "learning_rate": 0.00019580534974885215, "loss": 4.6001, "step": 357950 }, { "epoch": 0.6680678051500003, "grad_norm": 0.922713041305542, "learning_rate": 0.0001958041457076335, "loss": 4.6231, "step": 358000 }, { "epoch": 0.6681611107093789, "grad_norm": 1.1627779006958008, "learning_rate": 0.00019580294149733742, "loss": 4.4918, "step": 358050 }, { "epoch": 0.6682544162687574, "grad_norm": 1.1160223484039307, "learning_rate": 0.00019580173711796598, "loss": 4.5517, "step": 358100 }, { "epoch": 0.6683477218281358, "grad_norm": 1.067106008529663, "learning_rate": 0.00019580053256952135, "loss": 4.5297, "step": 358150 }, { "epoch": 0.6684410273875143, "grad_norm": 0.7735969424247742, "learning_rate": 0.00019579932785200566, "loss": 4.4722, "step": 358200 }, { "epoch": 0.6685343329468928, "grad_norm": 1.1097967624664307, "learning_rate": 0.00019579812296542098, "loss": 4.5994, "step": 358250 }, { "epoch": 0.6686276385062713, "grad_norm": 1.2591450214385986, "learning_rate": 0.00019579691790976944, "loss": 4.6635, "step": 358300 }, { "epoch": 0.6687209440656497, "grad_norm": 1.1494500637054443, "learning_rate": 0.00019579571268505324, "loss": 4.6788, "step": 358350 }, { "epoch": 0.6688142496250283, "grad_norm": 0.966873824596405, "learning_rate": 0.00019579450729127445, "loss": 4.4838, "step": 358400 }, { "epoch": 0.6689075551844068, "grad_norm": 1.4860520362854004, "learning_rate": 0.0001957933017284352, "loss": 4.6094, "step": 358450 }, { "epoch": 0.6690008607437853, "grad_norm": 1.0449298620224, "learning_rate": 0.00019579209599653765, "loss": 4.6897, "step": 358500 }, { "epoch": 0.6690941663031638, "grad_norm": 1.0751429796218872, "learning_rate": 0.0001957908900955839, "loss": 4.5401, "step": 358550 }, { "epoch": 0.6691874718625422, "grad_norm": 0.7464136481285095, "learning_rate": 0.00019578968402557607, "loss": 4.5338, "step": 358600 }, { "epoch": 0.6692807774219207, "grad_norm": 0.916556715965271, "learning_rate": 0.0001957884777865163, "loss": 4.6826, "step": 358650 }, { "epoch": 0.6693740829812992, "grad_norm": 0.8825215101242065, "learning_rate": 0.00019578727137840673, "loss": 4.6662, "step": 358700 }, { "epoch": 0.6694673885406778, "grad_norm": 1.216315507888794, "learning_rate": 0.0001957860648012495, "loss": 4.6074, "step": 358750 }, { "epoch": 0.6695606941000563, "grad_norm": 1.1252713203430176, "learning_rate": 0.00019578485805504667, "loss": 4.624, "step": 358800 }, { "epoch": 0.6696539996594347, "grad_norm": 0.9946227073669434, "learning_rate": 0.0001957836511398005, "loss": 4.404, "step": 358850 }, { "epoch": 0.6697473052188132, "grad_norm": 0.8797948956489563, "learning_rate": 0.00019578244405551296, "loss": 4.5698, "step": 358900 }, { "epoch": 0.6698406107781917, "grad_norm": 0.8493410348892212, "learning_rate": 0.0001957812368021863, "loss": 4.5516, "step": 358950 }, { "epoch": 0.6699339163375702, "grad_norm": 0.9304854869842529, "learning_rate": 0.00019578002937982258, "loss": 4.5322, "step": 359000 }, { "epoch": 0.6700272218969486, "grad_norm": 1.0428178310394287, "learning_rate": 0.00019577882178842398, "loss": 4.3919, "step": 359050 }, { "epoch": 0.6701205274563272, "grad_norm": 0.9189246296882629, "learning_rate": 0.00019577761402799263, "loss": 4.6786, "step": 359100 }, { "epoch": 0.6702138330157057, "grad_norm": 1.20093834400177, "learning_rate": 0.00019577640609853063, "loss": 4.7049, "step": 359150 }, { "epoch": 0.6703071385750842, "grad_norm": 1.0256885290145874, "learning_rate": 0.00019577519800004009, "loss": 4.4463, "step": 359200 }, { "epoch": 0.6704004441344626, "grad_norm": 1.009544014930725, "learning_rate": 0.00019577398973252323, "loss": 4.4192, "step": 359250 }, { "epoch": 0.6704937496938411, "grad_norm": 0.9781863689422607, "learning_rate": 0.0001957727812959821, "loss": 4.7262, "step": 359300 }, { "epoch": 0.6705870552532196, "grad_norm": 0.7879834175109863, "learning_rate": 0.00019577157269041887, "loss": 4.548, "step": 359350 }, { "epoch": 0.6706803608125981, "grad_norm": 0.9353818893432617, "learning_rate": 0.00019577036391583568, "loss": 4.64, "step": 359400 }, { "epoch": 0.6707736663719767, "grad_norm": 1.063107967376709, "learning_rate": 0.00019576915497223466, "loss": 4.7691, "step": 359450 }, { "epoch": 0.6708669719313551, "grad_norm": 1.1930428743362427, "learning_rate": 0.00019576794585961788, "loss": 4.5924, "step": 359500 }, { "epoch": 0.6709602774907336, "grad_norm": 1.2417848110198975, "learning_rate": 0.00019576673657798756, "loss": 4.4907, "step": 359550 }, { "epoch": 0.6710535830501121, "grad_norm": 0.8512386679649353, "learning_rate": 0.0001957655271273458, "loss": 4.402, "step": 359600 }, { "epoch": 0.6711468886094906, "grad_norm": 1.0181959867477417, "learning_rate": 0.00019576431750769471, "loss": 4.7324, "step": 359650 }, { "epoch": 0.671240194168869, "grad_norm": 0.8717050552368164, "learning_rate": 0.0001957631077190365, "loss": 4.4207, "step": 359700 }, { "epoch": 0.6713334997282475, "grad_norm": 1.1153290271759033, "learning_rate": 0.00019576189776137324, "loss": 4.5153, "step": 359750 }, { "epoch": 0.6714268052876261, "grad_norm": 1.100089430809021, "learning_rate": 0.00019576068763470705, "loss": 4.6634, "step": 359800 }, { "epoch": 0.6715201108470046, "grad_norm": 1.1623677015304565, "learning_rate": 0.0001957594773390401, "loss": 4.5501, "step": 359850 }, { "epoch": 0.6716134164063831, "grad_norm": 1.1294299364089966, "learning_rate": 0.00019575826687437455, "loss": 4.5091, "step": 359900 }, { "epoch": 0.6717067219657615, "grad_norm": 1.3196748495101929, "learning_rate": 0.0001957570562407125, "loss": 4.5307, "step": 359950 }, { "epoch": 0.67180002752514, "grad_norm": 1.1398862600326538, "learning_rate": 0.00019575584543805608, "loss": 4.5979, "step": 360000 }, { "epoch": 0.67180002752514, "eval_loss": 4.742733478546143, "eval_runtime": 231.3397, "eval_samples_per_second": 11.273, "eval_steps_per_second": 11.273, "eval_tts_loss": 7.567788422349097, "step": 360000 }, { "epoch": 0.6718933330845185, "grad_norm": 0.9543121457099915, "learning_rate": 0.00019575463446640744, "loss": 4.5505, "step": 360050 }, { "epoch": 0.671986638643897, "grad_norm": 1.1611236333847046, "learning_rate": 0.00019575342332576873, "loss": 4.5798, "step": 360100 }, { "epoch": 0.6720799442032754, "grad_norm": 1.2661935091018677, "learning_rate": 0.0001957522120161421, "loss": 4.5332, "step": 360150 }, { "epoch": 0.672173249762654, "grad_norm": 1.149210810661316, "learning_rate": 0.0001957510005375296, "loss": 4.413, "step": 360200 }, { "epoch": 0.6722665553220325, "grad_norm": 0.6918655633926392, "learning_rate": 0.0001957497888899335, "loss": 4.4898, "step": 360250 }, { "epoch": 0.672359860881411, "grad_norm": 0.9333517551422119, "learning_rate": 0.00019574857707335583, "loss": 4.6921, "step": 360300 }, { "epoch": 0.6724531664407895, "grad_norm": 1.155081868171692, "learning_rate": 0.00019574736508779877, "loss": 4.6113, "step": 360350 }, { "epoch": 0.6725464720001679, "grad_norm": 0.6867272257804871, "learning_rate": 0.00019574615293326447, "loss": 4.5905, "step": 360400 }, { "epoch": 0.6726397775595464, "grad_norm": 1.0847917795181274, "learning_rate": 0.00019574494060975504, "loss": 4.6095, "step": 360450 }, { "epoch": 0.6727330831189249, "grad_norm": 0.8768734335899353, "learning_rate": 0.00019574372811727265, "loss": 4.6401, "step": 360500 }, { "epoch": 0.6728263886783035, "grad_norm": 1.27104651927948, "learning_rate": 0.00019574251545581946, "loss": 4.5056, "step": 360550 }, { "epoch": 0.6729196942376819, "grad_norm": 1.0164141654968262, "learning_rate": 0.00019574130262539754, "loss": 4.6382, "step": 360600 }, { "epoch": 0.6730129997970604, "grad_norm": 1.1899137496948242, "learning_rate": 0.00019574008962600908, "loss": 4.5382, "step": 360650 }, { "epoch": 0.6731063053564389, "grad_norm": 1.0625330209732056, "learning_rate": 0.0001957388764576562, "loss": 4.6297, "step": 360700 }, { "epoch": 0.6731996109158174, "grad_norm": 0.9368055462837219, "learning_rate": 0.00019573766312034103, "loss": 4.7591, "step": 360750 }, { "epoch": 0.6732929164751958, "grad_norm": 1.0346652269363403, "learning_rate": 0.00019573644961406577, "loss": 4.5228, "step": 360800 }, { "epoch": 0.6733862220345743, "grad_norm": 1.0345101356506348, "learning_rate": 0.00019573523593883249, "loss": 4.6289, "step": 360850 }, { "epoch": 0.6734795275939529, "grad_norm": 1.3405543565750122, "learning_rate": 0.00019573402209464334, "loss": 4.6347, "step": 360900 }, { "epoch": 0.6735728331533314, "grad_norm": 0.9958884716033936, "learning_rate": 0.00019573280808150053, "loss": 4.5511, "step": 360950 }, { "epoch": 0.6736661387127099, "grad_norm": 1.1798781156539917, "learning_rate": 0.00019573159389940618, "loss": 4.4071, "step": 361000 }, { "epoch": 0.6737594442720883, "grad_norm": 0.8254154324531555, "learning_rate": 0.00019573037954836236, "loss": 4.6153, "step": 361050 }, { "epoch": 0.6738527498314668, "grad_norm": 1.03472900390625, "learning_rate": 0.0001957291650283713, "loss": 4.4614, "step": 361100 }, { "epoch": 0.6739460553908453, "grad_norm": 0.7731339931488037, "learning_rate": 0.00019572795033943508, "loss": 4.5691, "step": 361150 }, { "epoch": 0.6740393609502238, "grad_norm": 0.7978653907775879, "learning_rate": 0.00019572673548155586, "loss": 4.6731, "step": 361200 }, { "epoch": 0.6741326665096024, "grad_norm": 0.8550422191619873, "learning_rate": 0.00019572552045473581, "loss": 4.4369, "step": 361250 }, { "epoch": 0.6742259720689808, "grad_norm": 1.2098850011825562, "learning_rate": 0.0001957243052589771, "loss": 4.614, "step": 361300 }, { "epoch": 0.6743192776283593, "grad_norm": 1.0120937824249268, "learning_rate": 0.00019572308989428182, "loss": 4.5158, "step": 361350 }, { "epoch": 0.6744125831877378, "grad_norm": 1.155301809310913, "learning_rate": 0.0001957218743606521, "loss": 4.5678, "step": 361400 }, { "epoch": 0.6745058887471163, "grad_norm": 1.0626083612442017, "learning_rate": 0.0001957206586580901, "loss": 4.4325, "step": 361450 }, { "epoch": 0.6745991943064947, "grad_norm": 0.9684640765190125, "learning_rate": 0.00019571944278659804, "loss": 4.61, "step": 361500 }, { "epoch": 0.6746924998658732, "grad_norm": 0.969910204410553, "learning_rate": 0.00019571822674617796, "loss": 4.4414, "step": 361550 }, { "epoch": 0.6747858054252518, "grad_norm": 1.1967800855636597, "learning_rate": 0.00019571701053683205, "loss": 4.6256, "step": 361600 }, { "epoch": 0.6748791109846303, "grad_norm": 0.913022518157959, "learning_rate": 0.00019571579415856246, "loss": 4.617, "step": 361650 }, { "epoch": 0.6749724165440087, "grad_norm": 0.9121726155281067, "learning_rate": 0.00019571457761137137, "loss": 4.5678, "step": 361700 }, { "epoch": 0.6750657221033872, "grad_norm": 1.1886693239212036, "learning_rate": 0.00019571336089526086, "loss": 4.6723, "step": 361750 }, { "epoch": 0.6751590276627657, "grad_norm": 1.0208622217178345, "learning_rate": 0.00019571214401023314, "loss": 4.5746, "step": 361800 }, { "epoch": 0.6752523332221442, "grad_norm": 1.072257399559021, "learning_rate": 0.0001957109269562903, "loss": 4.5361, "step": 361850 }, { "epoch": 0.6753456387815227, "grad_norm": 1.0889387130737305, "learning_rate": 0.00019570970973343453, "loss": 4.747, "step": 361900 }, { "epoch": 0.6754389443409012, "grad_norm": 1.2097222805023193, "learning_rate": 0.00019570849234166793, "loss": 4.5434, "step": 361950 }, { "epoch": 0.6755322499002797, "grad_norm": 1.045190453529358, "learning_rate": 0.0001957072747809927, "loss": 4.4708, "step": 362000 }, { "epoch": 0.6756255554596582, "grad_norm": 1.025077223777771, "learning_rate": 0.000195706057051411, "loss": 4.4576, "step": 362050 }, { "epoch": 0.6757188610190367, "grad_norm": 1.1330491304397583, "learning_rate": 0.0001957048391529249, "loss": 4.6768, "step": 362100 }, { "epoch": 0.6758121665784151, "grad_norm": 1.0041680335998535, "learning_rate": 0.00019570362108553663, "loss": 4.5468, "step": 362150 }, { "epoch": 0.6759054721377936, "grad_norm": 1.4580110311508179, "learning_rate": 0.00019570240284924834, "loss": 4.3164, "step": 362200 }, { "epoch": 0.6759987776971721, "grad_norm": 0.8612103462219238, "learning_rate": 0.00019570118444406208, "loss": 4.5728, "step": 362250 }, { "epoch": 0.6760920832565507, "grad_norm": 1.135723352432251, "learning_rate": 0.0001956999658699801, "loss": 4.5523, "step": 362300 }, { "epoch": 0.6761853888159292, "grad_norm": 0.9825050830841064, "learning_rate": 0.0001956987471270045, "loss": 4.5548, "step": 362350 }, { "epoch": 0.6762786943753076, "grad_norm": 1.275405764579773, "learning_rate": 0.00019569752821513748, "loss": 4.4826, "step": 362400 }, { "epoch": 0.6763719999346861, "grad_norm": 1.0499589443206787, "learning_rate": 0.00019569630913438116, "loss": 4.635, "step": 362450 }, { "epoch": 0.6764653054940646, "grad_norm": 0.9857321381568909, "learning_rate": 0.0001956950898847377, "loss": 4.5577, "step": 362500 }, { "epoch": 0.6765586110534431, "grad_norm": 1.0807452201843262, "learning_rate": 0.0001956938704662092, "loss": 4.393, "step": 362550 }, { "epoch": 0.6766519166128215, "grad_norm": 1.0702356100082397, "learning_rate": 0.00019569265087879788, "loss": 4.721, "step": 362600 }, { "epoch": 0.6767452221722001, "grad_norm": 0.8599978685379028, "learning_rate": 0.00019569143112250587, "loss": 4.5146, "step": 362650 }, { "epoch": 0.6768385277315786, "grad_norm": 0.9826505780220032, "learning_rate": 0.00019569021119733533, "loss": 4.5429, "step": 362700 }, { "epoch": 0.6769318332909571, "grad_norm": 1.0868836641311646, "learning_rate": 0.0001956889911032884, "loss": 4.527, "step": 362750 }, { "epoch": 0.6770251388503356, "grad_norm": 1.200915813446045, "learning_rate": 0.00019568777084036723, "loss": 4.5566, "step": 362800 }, { "epoch": 0.677118444409714, "grad_norm": 1.088509440422058, "learning_rate": 0.000195686550408574, "loss": 4.4763, "step": 362850 }, { "epoch": 0.6772117499690925, "grad_norm": 1.1293811798095703, "learning_rate": 0.00019568532980791082, "loss": 4.5489, "step": 362900 }, { "epoch": 0.677305055528471, "grad_norm": 1.165267825126648, "learning_rate": 0.0001956841090383799, "loss": 4.5753, "step": 362950 }, { "epoch": 0.6773983610878496, "grad_norm": 0.7049471735954285, "learning_rate": 0.00019568288809998334, "loss": 4.6423, "step": 363000 }, { "epoch": 0.6773983610878496, "eval_loss": 4.743714809417725, "eval_runtime": 231.9151, "eval_samples_per_second": 11.245, "eval_steps_per_second": 11.245, "eval_tts_loss": 7.5692039081952505, "step": 363000 }, { "epoch": 0.677491666647228, "grad_norm": 1.0355195999145508, "learning_rate": 0.00019568166699272333, "loss": 4.6849, "step": 363050 }, { "epoch": 0.6775849722066065, "grad_norm": 1.2881120443344116, "learning_rate": 0.00019568044571660197, "loss": 4.6834, "step": 363100 }, { "epoch": 0.677678277765985, "grad_norm": 1.2323970794677734, "learning_rate": 0.00019567922427162152, "loss": 4.5766, "step": 363150 }, { "epoch": 0.6777715833253635, "grad_norm": 1.278564691543579, "learning_rate": 0.00019567800265778403, "loss": 4.6025, "step": 363200 }, { "epoch": 0.677864888884742, "grad_norm": 0.8138118386268616, "learning_rate": 0.00019567678087509174, "loss": 4.5658, "step": 363250 }, { "epoch": 0.6779581944441204, "grad_norm": 0.691251814365387, "learning_rate": 0.00019567555892354674, "loss": 4.5252, "step": 363300 }, { "epoch": 0.678051500003499, "grad_norm": 0.8609877824783325, "learning_rate": 0.00019567433680315122, "loss": 4.5487, "step": 363350 }, { "epoch": 0.6781448055628775, "grad_norm": 0.8605664372444153, "learning_rate": 0.00019567311451390733, "loss": 4.4648, "step": 363400 }, { "epoch": 0.678238111122256, "grad_norm": 1.4465529918670654, "learning_rate": 0.00019567189205581723, "loss": 4.7005, "step": 363450 }, { "epoch": 0.6783314166816344, "grad_norm": 1.1591824293136597, "learning_rate": 0.00019567066942888307, "loss": 4.4901, "step": 363500 }, { "epoch": 0.6784247222410129, "grad_norm": 1.1295784711837769, "learning_rate": 0.000195669446633107, "loss": 4.583, "step": 363550 }, { "epoch": 0.6785180278003914, "grad_norm": 0.9203224778175354, "learning_rate": 0.0001956682236684912, "loss": 4.3986, "step": 363600 }, { "epoch": 0.6786113333597699, "grad_norm": 0.9815191030502319, "learning_rate": 0.00019566700053503784, "loss": 4.5553, "step": 363650 }, { "epoch": 0.6787046389191485, "grad_norm": 0.9840654730796814, "learning_rate": 0.00019566577723274903, "loss": 4.4235, "step": 363700 }, { "epoch": 0.6787979444785269, "grad_norm": 1.0703121423721313, "learning_rate": 0.00019566455376162698, "loss": 4.5737, "step": 363750 }, { "epoch": 0.6788912500379054, "grad_norm": 0.8637380599975586, "learning_rate": 0.00019566333012167379, "loss": 4.5224, "step": 363800 }, { "epoch": 0.6789845555972839, "grad_norm": 0.9189258813858032, "learning_rate": 0.00019566210631289172, "loss": 4.4868, "step": 363850 }, { "epoch": 0.6790778611566624, "grad_norm": 0.9535520076751709, "learning_rate": 0.00019566088233528278, "loss": 4.4389, "step": 363900 }, { "epoch": 0.6791711667160408, "grad_norm": 1.0472358465194702, "learning_rate": 0.00019565965818884925, "loss": 4.6446, "step": 363950 }, { "epoch": 0.6792644722754193, "grad_norm": 1.2085245847702026, "learning_rate": 0.00019565843387359325, "loss": 4.5661, "step": 364000 }, { "epoch": 0.6793577778347979, "grad_norm": 1.175116777420044, "learning_rate": 0.000195657209389517, "loss": 4.571, "step": 364050 }, { "epoch": 0.6794510833941764, "grad_norm": 1.7191215753555298, "learning_rate": 0.00019565598473662255, "loss": 4.5139, "step": 364100 }, { "epoch": 0.6795443889535548, "grad_norm": 0.8249337077140808, "learning_rate": 0.0001956547599149121, "loss": 4.6354, "step": 364150 }, { "epoch": 0.6796376945129333, "grad_norm": 0.8554646372795105, "learning_rate": 0.00019565353492438788, "loss": 4.7298, "step": 364200 }, { "epoch": 0.6797310000723118, "grad_norm": 0.9862709045410156, "learning_rate": 0.000195652309765052, "loss": 4.5764, "step": 364250 }, { "epoch": 0.6798243056316903, "grad_norm": 1.3434244394302368, "learning_rate": 0.0001956510844369066, "loss": 4.7812, "step": 364300 }, { "epoch": 0.6799176111910687, "grad_norm": 0.7632136344909668, "learning_rate": 0.00019564985893995385, "loss": 4.6408, "step": 364350 }, { "epoch": 0.6800109167504473, "grad_norm": 1.0645220279693604, "learning_rate": 0.00019564863327419598, "loss": 4.4689, "step": 364400 }, { "epoch": 0.6801042223098258, "grad_norm": 0.6220859289169312, "learning_rate": 0.00019564740743963506, "loss": 4.5658, "step": 364450 }, { "epoch": 0.6801975278692043, "grad_norm": 1.4747847318649292, "learning_rate": 0.00019564618143627332, "loss": 4.6266, "step": 364500 }, { "epoch": 0.6802908334285828, "grad_norm": 1.1730914115905762, "learning_rate": 0.00019564495526411287, "loss": 4.6482, "step": 364550 }, { "epoch": 0.6803841389879612, "grad_norm": 0.8440443873405457, "learning_rate": 0.00019564372892315592, "loss": 4.552, "step": 364600 }, { "epoch": 0.6804774445473397, "grad_norm": 1.2068709135055542, "learning_rate": 0.00019564250241340463, "loss": 4.759, "step": 364650 }, { "epoch": 0.6805707501067182, "grad_norm": 0.9385483264923096, "learning_rate": 0.00019564127573486114, "loss": 4.4719, "step": 364700 }, { "epoch": 0.6806640556660968, "grad_norm": 0.8869559168815613, "learning_rate": 0.00019564004888752762, "loss": 4.5069, "step": 364750 }, { "epoch": 0.6807573612254753, "grad_norm": 1.1940581798553467, "learning_rate": 0.00019563882187140626, "loss": 4.7019, "step": 364800 }, { "epoch": 0.6808506667848537, "grad_norm": 1.2081252336502075, "learning_rate": 0.00019563759468649918, "loss": 4.5911, "step": 364850 }, { "epoch": 0.6809439723442322, "grad_norm": 1.456714153289795, "learning_rate": 0.0001956363673328086, "loss": 4.4947, "step": 364900 }, { "epoch": 0.6810372779036107, "grad_norm": 1.0345370769500732, "learning_rate": 0.00019563513981033664, "loss": 4.5642, "step": 364950 }, { "epoch": 0.6811305834629892, "grad_norm": 0.933967113494873, "learning_rate": 0.00019563391211908552, "loss": 4.6249, "step": 365000 }, { "epoch": 0.6812238890223676, "grad_norm": 1.1223982572555542, "learning_rate": 0.00019563268425905735, "loss": 4.5568, "step": 365050 }, { "epoch": 0.6813171945817462, "grad_norm": 0.7929447889328003, "learning_rate": 0.00019563145623025432, "loss": 4.4747, "step": 365100 }, { "epoch": 0.6814105001411247, "grad_norm": 1.0209295749664307, "learning_rate": 0.0001956302280326786, "loss": 4.7439, "step": 365150 }, { "epoch": 0.6815038057005032, "grad_norm": 0.9150970578193665, "learning_rate": 0.00019562899966633234, "loss": 4.4238, "step": 365200 }, { "epoch": 0.6815971112598816, "grad_norm": 0.9653599262237549, "learning_rate": 0.00019562777113121772, "loss": 4.5515, "step": 365250 }, { "epoch": 0.6816904168192601, "grad_norm": 1.1098837852478027, "learning_rate": 0.00019562654242733692, "loss": 4.6102, "step": 365300 }, { "epoch": 0.6817837223786386, "grad_norm": 0.9841729998588562, "learning_rate": 0.0001956253135546921, "loss": 4.5646, "step": 365350 }, { "epoch": 0.6818770279380171, "grad_norm": 1.2507600784301758, "learning_rate": 0.00019562408451328543, "loss": 4.6874, "step": 365400 }, { "epoch": 0.6819703334973956, "grad_norm": 0.8660342693328857, "learning_rate": 0.00019562285530311908, "loss": 4.5343, "step": 365450 }, { "epoch": 0.6820636390567741, "grad_norm": 1.0164939165115356, "learning_rate": 0.0001956216259241952, "loss": 4.6378, "step": 365500 }, { "epoch": 0.6821569446161526, "grad_norm": 0.9830291867256165, "learning_rate": 0.000195620396376516, "loss": 4.5258, "step": 365550 }, { "epoch": 0.6822502501755311, "grad_norm": 0.8058454394340515, "learning_rate": 0.0001956191666600836, "loss": 4.5012, "step": 365600 }, { "epoch": 0.6823435557349096, "grad_norm": 0.7354638576507568, "learning_rate": 0.0001956179367749002, "loss": 4.5234, "step": 365650 }, { "epoch": 0.682436861294288, "grad_norm": 0.9891838431358337, "learning_rate": 0.00019561670672096798, "loss": 4.5401, "step": 365700 }, { "epoch": 0.6825301668536665, "grad_norm": 0.7514564394950867, "learning_rate": 0.0001956154764982891, "loss": 4.5046, "step": 365750 }, { "epoch": 0.682623472413045, "grad_norm": 1.202129602432251, "learning_rate": 0.0001956142461068657, "loss": 4.4916, "step": 365800 }, { "epoch": 0.6827167779724236, "grad_norm": 1.1306408643722534, "learning_rate": 0.0001956130155467, "loss": 4.5489, "step": 365850 }, { "epoch": 0.6828100835318021, "grad_norm": 0.8608388900756836, "learning_rate": 0.00019561178481779413, "loss": 4.4624, "step": 365900 }, { "epoch": 0.6829033890911805, "grad_norm": 1.213686227798462, "learning_rate": 0.00019561055392015031, "loss": 4.4172, "step": 365950 }, { "epoch": 0.682996694650559, "grad_norm": 1.3149348497390747, "learning_rate": 0.00019560932285377067, "loss": 4.6674, "step": 366000 }, { "epoch": 0.682996694650559, "eval_loss": 4.745150089263916, "eval_runtime": 230.3389, "eval_samples_per_second": 11.322, "eval_steps_per_second": 11.322, "eval_tts_loss": 7.550067961383097, "step": 366000 }, { "epoch": 0.6830900002099375, "grad_norm": 1.0034360885620117, "learning_rate": 0.00019560809161865743, "loss": 4.6237, "step": 366050 }, { "epoch": 0.683183305769316, "grad_norm": 1.0511972904205322, "learning_rate": 0.00019560686021481268, "loss": 4.4821, "step": 366100 }, { "epoch": 0.6832766113286944, "grad_norm": 1.0858912467956543, "learning_rate": 0.00019560562864223867, "loss": 4.4884, "step": 366150 }, { "epoch": 0.683369916888073, "grad_norm": 0.6770386099815369, "learning_rate": 0.00019560439690093756, "loss": 4.4349, "step": 366200 }, { "epoch": 0.6834632224474515, "grad_norm": 1.1042362451553345, "learning_rate": 0.00019560316499091148, "loss": 4.6419, "step": 366250 }, { "epoch": 0.68355652800683, "grad_norm": 1.1873412132263184, "learning_rate": 0.00019560193291216267, "loss": 4.4536, "step": 366300 }, { "epoch": 0.6836498335662085, "grad_norm": 1.1348243951797485, "learning_rate": 0.00019560070066469324, "loss": 4.5514, "step": 366350 }, { "epoch": 0.6837431391255869, "grad_norm": 0.8972763419151306, "learning_rate": 0.00019559946824850542, "loss": 4.639, "step": 366400 }, { "epoch": 0.6838364446849654, "grad_norm": 1.2635902166366577, "learning_rate": 0.00019559823566360135, "loss": 4.5739, "step": 366450 }, { "epoch": 0.6839297502443439, "grad_norm": 0.8604574799537659, "learning_rate": 0.00019559700290998323, "loss": 4.5483, "step": 366500 }, { "epoch": 0.6840230558037225, "grad_norm": 1.056391716003418, "learning_rate": 0.0001955957699876532, "loss": 4.4846, "step": 366550 }, { "epoch": 0.6841163613631009, "grad_norm": 1.2378082275390625, "learning_rate": 0.00019559453689661346, "loss": 4.6516, "step": 366600 }, { "epoch": 0.6842096669224794, "grad_norm": 1.1226683855056763, "learning_rate": 0.00019559330363686617, "loss": 4.4961, "step": 366650 }, { "epoch": 0.6843029724818579, "grad_norm": 0.9702514410018921, "learning_rate": 0.00019559207020841354, "loss": 4.8676, "step": 366700 }, { "epoch": 0.6843962780412364, "grad_norm": 0.9354146122932434, "learning_rate": 0.00019559083661125772, "loss": 4.5858, "step": 366750 }, { "epoch": 0.6844895836006148, "grad_norm": 0.993274986743927, "learning_rate": 0.0001955896028454009, "loss": 4.4758, "step": 366800 }, { "epoch": 0.6845828891599933, "grad_norm": 1.0623561143875122, "learning_rate": 0.00019558836891084522, "loss": 4.3529, "step": 366850 }, { "epoch": 0.6846761947193719, "grad_norm": 0.9752203226089478, "learning_rate": 0.00019558713480759293, "loss": 4.6037, "step": 366900 }, { "epoch": 0.6847695002787504, "grad_norm": 0.9326674938201904, "learning_rate": 0.00019558590053564613, "loss": 4.6296, "step": 366950 }, { "epoch": 0.6848628058381289, "grad_norm": 0.8091044425964355, "learning_rate": 0.00019558466609500706, "loss": 4.698, "step": 367000 }, { "epoch": 0.6849561113975073, "grad_norm": 0.9038903117179871, "learning_rate": 0.00019558343148567788, "loss": 4.3031, "step": 367050 }, { "epoch": 0.6850494169568858, "grad_norm": 1.1735448837280273, "learning_rate": 0.00019558219670766074, "loss": 4.7071, "step": 367100 }, { "epoch": 0.6851427225162643, "grad_norm": 1.3830543756484985, "learning_rate": 0.00019558096176095787, "loss": 4.5986, "step": 367150 }, { "epoch": 0.6852360280756428, "grad_norm": 1.1118428707122803, "learning_rate": 0.00019557972664557136, "loss": 4.458, "step": 367200 }, { "epoch": 0.6853293336350214, "grad_norm": 1.1908377408981323, "learning_rate": 0.0001955784913615035, "loss": 4.5695, "step": 367250 }, { "epoch": 0.6854226391943998, "grad_norm": 1.0074695348739624, "learning_rate": 0.0001955772559087564, "loss": 4.7084, "step": 367300 }, { "epoch": 0.6855159447537783, "grad_norm": 1.166852355003357, "learning_rate": 0.0001955760202873323, "loss": 4.6143, "step": 367350 }, { "epoch": 0.6856092503131568, "grad_norm": 0.8768144845962524, "learning_rate": 0.00019557478449723327, "loss": 4.587, "step": 367400 }, { "epoch": 0.6857025558725353, "grad_norm": 1.2766517400741577, "learning_rate": 0.00019557354853846163, "loss": 4.6175, "step": 367450 }, { "epoch": 0.6857958614319137, "grad_norm": 1.0810915231704712, "learning_rate": 0.00019557231241101945, "loss": 4.5643, "step": 367500 }, { "epoch": 0.6858891669912922, "grad_norm": 1.072299838066101, "learning_rate": 0.000195571076114909, "loss": 4.4177, "step": 367550 }, { "epoch": 0.6859824725506708, "grad_norm": 0.7774250507354736, "learning_rate": 0.0001955698396501324, "loss": 4.4628, "step": 367600 }, { "epoch": 0.6860757781100493, "grad_norm": 0.9877749085426331, "learning_rate": 0.0001955686030166918, "loss": 4.5595, "step": 367650 }, { "epoch": 0.6861690836694277, "grad_norm": 0.8914515376091003, "learning_rate": 0.00019556736621458948, "loss": 4.5708, "step": 367700 }, { "epoch": 0.6862623892288062, "grad_norm": 0.8482488989830017, "learning_rate": 0.00019556612924382754, "loss": 4.558, "step": 367750 }, { "epoch": 0.6863556947881847, "grad_norm": 1.1105358600616455, "learning_rate": 0.00019556489210440824, "loss": 4.5492, "step": 367800 }, { "epoch": 0.6864490003475632, "grad_norm": 1.0241068601608276, "learning_rate": 0.00019556365479633372, "loss": 4.3774, "step": 367850 }, { "epoch": 0.6865423059069417, "grad_norm": 1.1066597700119019, "learning_rate": 0.00019556241731960613, "loss": 4.7176, "step": 367900 }, { "epoch": 0.6866356114663202, "grad_norm": 1.0185424089431763, "learning_rate": 0.0001955611796742277, "loss": 4.3786, "step": 367950 }, { "epoch": 0.6867289170256987, "grad_norm": 1.2061375379562378, "learning_rate": 0.00019555994186020058, "loss": 4.6226, "step": 368000 }, { "epoch": 0.6868222225850772, "grad_norm": 1.1836224794387817, "learning_rate": 0.000195558703877527, "loss": 4.5512, "step": 368050 }, { "epoch": 0.6869155281444557, "grad_norm": 0.7828928232192993, "learning_rate": 0.00019555746572620916, "loss": 4.6906, "step": 368100 }, { "epoch": 0.6870088337038341, "grad_norm": 0.9676310420036316, "learning_rate": 0.00019555622740624918, "loss": 4.4422, "step": 368150 }, { "epoch": 0.6871021392632126, "grad_norm": 1.104750156402588, "learning_rate": 0.00019555498891764926, "loss": 4.6423, "step": 368200 }, { "epoch": 0.6871954448225911, "grad_norm": 1.1759475469589233, "learning_rate": 0.00019555375026041158, "loss": 4.4813, "step": 368250 }, { "epoch": 0.6872887503819697, "grad_norm": 1.128739356994629, "learning_rate": 0.00019555251143453837, "loss": 4.5668, "step": 368300 }, { "epoch": 0.6873820559413482, "grad_norm": 1.0008203983306885, "learning_rate": 0.0001955512724400318, "loss": 4.3894, "step": 368350 }, { "epoch": 0.6874753615007266, "grad_norm": 0.8730279803276062, "learning_rate": 0.00019555003327689403, "loss": 4.4255, "step": 368400 }, { "epoch": 0.6875686670601051, "grad_norm": 1.139626145362854, "learning_rate": 0.00019554879394512726, "loss": 4.5072, "step": 368450 }, { "epoch": 0.6876619726194836, "grad_norm": 0.7365878224372864, "learning_rate": 0.00019554755444473369, "loss": 4.4943, "step": 368500 }, { "epoch": 0.6877552781788621, "grad_norm": 1.0410829782485962, "learning_rate": 0.0001955463147757155, "loss": 4.5267, "step": 368550 }, { "epoch": 0.6878485837382405, "grad_norm": 1.2380183935165405, "learning_rate": 0.00019554507493807486, "loss": 4.3674, "step": 368600 }, { "epoch": 0.6879418892976191, "grad_norm": 1.0722304582595825, "learning_rate": 0.000195543834931814, "loss": 4.4804, "step": 368650 }, { "epoch": 0.6880351948569976, "grad_norm": 1.3019156455993652, "learning_rate": 0.00019554259475693507, "loss": 4.5368, "step": 368700 }, { "epoch": 0.6881285004163761, "grad_norm": 1.0796579122543335, "learning_rate": 0.00019554135441344025, "loss": 4.5918, "step": 368750 }, { "epoch": 0.6882218059757546, "grad_norm": 1.541914463043213, "learning_rate": 0.0001955401139013318, "loss": 4.4014, "step": 368800 }, { "epoch": 0.688315111535133, "grad_norm": 1.0862489938735962, "learning_rate": 0.00019553887322061182, "loss": 4.545, "step": 368850 }, { "epoch": 0.6884084170945115, "grad_norm": 1.548496961593628, "learning_rate": 0.00019553763237128256, "loss": 4.7347, "step": 368900 }, { "epoch": 0.68850172265389, "grad_norm": 1.0292510986328125, "learning_rate": 0.00019553639135334617, "loss": 4.5278, "step": 368950 }, { "epoch": 0.6885950282132686, "grad_norm": 0.8597036004066467, "learning_rate": 0.00019553515016680487, "loss": 4.4832, "step": 369000 }, { "epoch": 0.6885950282132686, "eval_loss": 4.74437952041626, "eval_runtime": 228.019, "eval_samples_per_second": 11.438, "eval_steps_per_second": 11.438, "eval_tts_loss": 7.562755081445505, "step": 369000 }, { "epoch": 0.688688333772647, "grad_norm": 0.8960505127906799, "learning_rate": 0.00019553390881166081, "loss": 4.4335, "step": 369050 }, { "epoch": 0.6887816393320255, "grad_norm": 0.9919257760047913, "learning_rate": 0.00019553266728791627, "loss": 4.4563, "step": 369100 }, { "epoch": 0.688874944891404, "grad_norm": 0.7080592513084412, "learning_rate": 0.00019553142559557336, "loss": 4.4533, "step": 369150 }, { "epoch": 0.6889682504507825, "grad_norm": 1.239512324333191, "learning_rate": 0.00019553018373463427, "loss": 4.4225, "step": 369200 }, { "epoch": 0.689061556010161, "grad_norm": 0.8886670470237732, "learning_rate": 0.00019552894170510123, "loss": 4.592, "step": 369250 }, { "epoch": 0.6891548615695394, "grad_norm": 0.9981276988983154, "learning_rate": 0.00019552769950697643, "loss": 4.4187, "step": 369300 }, { "epoch": 0.689248167128918, "grad_norm": 1.1080340147018433, "learning_rate": 0.00019552645714026202, "loss": 4.6639, "step": 369350 }, { "epoch": 0.6893414726882965, "grad_norm": 1.150093674659729, "learning_rate": 0.00019552521460496026, "loss": 4.5057, "step": 369400 }, { "epoch": 0.689434778247675, "grad_norm": 1.16274893283844, "learning_rate": 0.00019552397190107326, "loss": 4.3424, "step": 369450 }, { "epoch": 0.6895280838070534, "grad_norm": 1.1337528228759766, "learning_rate": 0.0001955227290286033, "loss": 4.4874, "step": 369500 }, { "epoch": 0.6896213893664319, "grad_norm": 0.9183812737464905, "learning_rate": 0.0001955214859875525, "loss": 4.5758, "step": 369550 }, { "epoch": 0.6897146949258104, "grad_norm": 0.9727858901023865, "learning_rate": 0.0001955202427779231, "loss": 4.4889, "step": 369600 }, { "epoch": 0.6898080004851889, "grad_norm": 0.9852572083473206, "learning_rate": 0.00019551899939971725, "loss": 4.5826, "step": 369650 }, { "epoch": 0.6899013060445675, "grad_norm": 0.9665806889533997, "learning_rate": 0.0001955177558529372, "loss": 4.4956, "step": 369700 }, { "epoch": 0.6899946116039459, "grad_norm": 1.0975180864334106, "learning_rate": 0.00019551651213758512, "loss": 4.6726, "step": 369750 }, { "epoch": 0.6900879171633244, "grad_norm": 0.9996809959411621, "learning_rate": 0.0001955152682536632, "loss": 4.6079, "step": 369800 }, { "epoch": 0.6901812227227029, "grad_norm": 1.0467857122421265, "learning_rate": 0.00019551402420117363, "loss": 4.6961, "step": 369850 }, { "epoch": 0.6902745282820814, "grad_norm": 0.9526649117469788, "learning_rate": 0.00019551277998011864, "loss": 4.4674, "step": 369900 }, { "epoch": 0.6903678338414598, "grad_norm": 1.1415717601776123, "learning_rate": 0.00019551153559050035, "loss": 4.5515, "step": 369950 }, { "epoch": 0.6904611394008383, "grad_norm": 0.8399020433425903, "learning_rate": 0.00019551029103232105, "loss": 4.5819, "step": 370000 }, { "epoch": 0.6905544449602169, "grad_norm": 0.8485755324363708, "learning_rate": 0.00019550904630558286, "loss": 4.5073, "step": 370050 }, { "epoch": 0.6906477505195954, "grad_norm": 1.2100976705551147, "learning_rate": 0.00019550780141028802, "loss": 4.7205, "step": 370100 }, { "epoch": 0.6907410560789738, "grad_norm": 1.1322890520095825, "learning_rate": 0.00019550655634643872, "loss": 4.2542, "step": 370150 }, { "epoch": 0.6908343616383523, "grad_norm": 1.1923885345458984, "learning_rate": 0.00019550531111403715, "loss": 4.6466, "step": 370200 }, { "epoch": 0.6909276671977308, "grad_norm": 0.8890280723571777, "learning_rate": 0.00019550406571308552, "loss": 4.7235, "step": 370250 }, { "epoch": 0.6910209727571093, "grad_norm": 0.961129367351532, "learning_rate": 0.000195502820143586, "loss": 4.5495, "step": 370300 }, { "epoch": 0.6911142783164878, "grad_norm": 1.1575649976730347, "learning_rate": 0.00019550157440554084, "loss": 4.6855, "step": 370350 }, { "epoch": 0.6912075838758663, "grad_norm": 0.886288046836853, "learning_rate": 0.0001955003284989522, "loss": 4.4955, "step": 370400 }, { "epoch": 0.6913008894352448, "grad_norm": 1.3565956354141235, "learning_rate": 0.00019549908242382224, "loss": 4.6876, "step": 370450 }, { "epoch": 0.6913941949946233, "grad_norm": 0.9993792772293091, "learning_rate": 0.00019549783618015324, "loss": 4.6915, "step": 370500 }, { "epoch": 0.6914875005540018, "grad_norm": 1.1421886682510376, "learning_rate": 0.00019549658976794737, "loss": 4.5883, "step": 370550 }, { "epoch": 0.6915808061133802, "grad_norm": 0.9610868096351624, "learning_rate": 0.00019549534318720677, "loss": 4.5152, "step": 370600 }, { "epoch": 0.6916741116727587, "grad_norm": 0.917777419090271, "learning_rate": 0.00019549409643793375, "loss": 4.6137, "step": 370650 }, { "epoch": 0.6917674172321372, "grad_norm": 1.0799082517623901, "learning_rate": 0.00019549284952013042, "loss": 4.5084, "step": 370700 }, { "epoch": 0.6918607227915158, "grad_norm": 0.9805840849876404, "learning_rate": 0.00019549160243379903, "loss": 4.5066, "step": 370750 }, { "epoch": 0.6919540283508943, "grad_norm": 1.0332810878753662, "learning_rate": 0.00019549035517894177, "loss": 4.5318, "step": 370800 }, { "epoch": 0.6920473339102727, "grad_norm": 1.0908498764038086, "learning_rate": 0.0001954891077555608, "loss": 4.3942, "step": 370850 }, { "epoch": 0.6921406394696512, "grad_norm": 1.0144426822662354, "learning_rate": 0.00019548786016365837, "loss": 4.5183, "step": 370900 }, { "epoch": 0.6922339450290297, "grad_norm": 0.7185133099555969, "learning_rate": 0.0001954866124032367, "loss": 4.6889, "step": 370950 }, { "epoch": 0.6923272505884082, "grad_norm": 0.8159246444702148, "learning_rate": 0.00019548536447429793, "loss": 4.5158, "step": 371000 }, { "epoch": 0.6924205561477866, "grad_norm": 1.108191967010498, "learning_rate": 0.0001954841163768443, "loss": 4.6321, "step": 371050 }, { "epoch": 0.6925138617071651, "grad_norm": 1.095153570175171, "learning_rate": 0.00019548286811087802, "loss": 4.4778, "step": 371100 }, { "epoch": 0.6926071672665437, "grad_norm": 1.071548581123352, "learning_rate": 0.00019548161967640124, "loss": 4.3622, "step": 371150 }, { "epoch": 0.6927004728259222, "grad_norm": 1.2594348192214966, "learning_rate": 0.00019548037107341624, "loss": 4.4866, "step": 371200 }, { "epoch": 0.6927937783853007, "grad_norm": 0.8258947134017944, "learning_rate": 0.00019547912230192518, "loss": 4.3015, "step": 371250 }, { "epoch": 0.6928870839446791, "grad_norm": 1.2690938711166382, "learning_rate": 0.00019547787336193025, "loss": 4.6411, "step": 371300 }, { "epoch": 0.6929803895040576, "grad_norm": 0.8823020458221436, "learning_rate": 0.0001954766242534337, "loss": 4.5026, "step": 371350 }, { "epoch": 0.6930736950634361, "grad_norm": 1.126285195350647, "learning_rate": 0.00019547537497643767, "loss": 4.5313, "step": 371400 }, { "epoch": 0.6931670006228146, "grad_norm": 1.3908414840698242, "learning_rate": 0.00019547412553094442, "loss": 4.4273, "step": 371450 }, { "epoch": 0.6932603061821931, "grad_norm": 1.1015129089355469, "learning_rate": 0.00019547287591695613, "loss": 4.4119, "step": 371500 }, { "epoch": 0.6933536117415716, "grad_norm": 1.1324613094329834, "learning_rate": 0.00019547162613447504, "loss": 4.673, "step": 371550 }, { "epoch": 0.6934469173009501, "grad_norm": 1.167299747467041, "learning_rate": 0.00019547037618350332, "loss": 4.6662, "step": 371600 }, { "epoch": 0.6935402228603286, "grad_norm": 1.0740166902542114, "learning_rate": 0.00019546912606404316, "loss": 4.475, "step": 371650 }, { "epoch": 0.693633528419707, "grad_norm": 1.5000436305999756, "learning_rate": 0.00019546787577609683, "loss": 4.6277, "step": 371700 }, { "epoch": 0.6937268339790855, "grad_norm": 1.1670953035354614, "learning_rate": 0.00019546662531966645, "loss": 4.4253, "step": 371750 }, { "epoch": 0.693820139538464, "grad_norm": 1.2359504699707031, "learning_rate": 0.0001954653746947543, "loss": 4.606, "step": 371800 }, { "epoch": 0.6939134450978426, "grad_norm": 0.9903602004051208, "learning_rate": 0.0001954641239013626, "loss": 4.5031, "step": 371850 }, { "epoch": 0.6940067506572211, "grad_norm": 1.4943287372589111, "learning_rate": 0.00019546287293949346, "loss": 4.4695, "step": 371900 }, { "epoch": 0.6941000562165995, "grad_norm": 1.1207112073898315, "learning_rate": 0.00019546162180914916, "loss": 4.6373, "step": 371950 }, { "epoch": 0.694193361775978, "grad_norm": 1.0234216451644897, "learning_rate": 0.0001954603705103319, "loss": 4.5827, "step": 372000 }, { "epoch": 0.694193361775978, "eval_loss": 4.737115859985352, "eval_runtime": 230.2817, "eval_samples_per_second": 11.325, "eval_steps_per_second": 11.325, "eval_tts_loss": 7.549153191890237, "step": 372000 }, { "epoch": 0.6942866673353565, "grad_norm": 1.2011725902557373, "learning_rate": 0.00019545911904304388, "loss": 4.5869, "step": 372050 }, { "epoch": 0.694379972894735, "grad_norm": 0.8918785452842712, "learning_rate": 0.00019545786740728734, "loss": 4.3698, "step": 372100 }, { "epoch": 0.6944732784541134, "grad_norm": 1.012628436088562, "learning_rate": 0.00019545661560306442, "loss": 4.5073, "step": 372150 }, { "epoch": 0.694566584013492, "grad_norm": 1.0200284719467163, "learning_rate": 0.0001954553636303774, "loss": 4.6447, "step": 372200 }, { "epoch": 0.6946598895728705, "grad_norm": 0.9943399429321289, "learning_rate": 0.00019545411148922846, "loss": 4.3431, "step": 372250 }, { "epoch": 0.694753195132249, "grad_norm": 0.9219287037849426, "learning_rate": 0.00019545285917961977, "loss": 4.5561, "step": 372300 }, { "epoch": 0.6948465006916275, "grad_norm": 1.167061448097229, "learning_rate": 0.0001954516067015536, "loss": 4.7088, "step": 372350 }, { "epoch": 0.6949398062510059, "grad_norm": 1.2277780771255493, "learning_rate": 0.00019545035405503213, "loss": 4.6485, "step": 372400 }, { "epoch": 0.6950331118103844, "grad_norm": 1.0556718111038208, "learning_rate": 0.00019544910124005762, "loss": 4.5689, "step": 372450 }, { "epoch": 0.6951264173697629, "grad_norm": 0.7697402834892273, "learning_rate": 0.0001954478482566322, "loss": 4.5025, "step": 372500 }, { "epoch": 0.6952197229291415, "grad_norm": 1.5871305465698242, "learning_rate": 0.00019544659510475814, "loss": 4.3397, "step": 372550 }, { "epoch": 0.69531302848852, "grad_norm": 0.971028745174408, "learning_rate": 0.00019544534178443762, "loss": 4.6454, "step": 372600 }, { "epoch": 0.6954063340478984, "grad_norm": 1.0348927974700928, "learning_rate": 0.00019544408829567288, "loss": 4.6747, "step": 372650 }, { "epoch": 0.6954996396072769, "grad_norm": 0.8070318698883057, "learning_rate": 0.0001954428346384661, "loss": 4.5578, "step": 372700 }, { "epoch": 0.6955929451666554, "grad_norm": 1.4318119287490845, "learning_rate": 0.00019544158081281953, "loss": 4.5287, "step": 372750 }, { "epoch": 0.6956862507260339, "grad_norm": 1.0857127904891968, "learning_rate": 0.00019544032681873536, "loss": 4.61, "step": 372800 }, { "epoch": 0.6957795562854123, "grad_norm": 1.012400507926941, "learning_rate": 0.0001954390726562158, "loss": 4.5802, "step": 372850 }, { "epoch": 0.6958728618447909, "grad_norm": 0.7758302688598633, "learning_rate": 0.0001954378183252631, "loss": 4.6574, "step": 372900 }, { "epoch": 0.6959661674041694, "grad_norm": 0.9272875785827637, "learning_rate": 0.0001954365638258794, "loss": 4.6471, "step": 372950 }, { "epoch": 0.6960594729635479, "grad_norm": 0.9731526970863342, "learning_rate": 0.00019543530915806697, "loss": 4.4795, "step": 373000 }, { "epoch": 0.6961527785229263, "grad_norm": 1.3055826425552368, "learning_rate": 0.00019543405432182804, "loss": 4.6253, "step": 373050 }, { "epoch": 0.6962460840823048, "grad_norm": 1.2052303552627563, "learning_rate": 0.00019543279931716475, "loss": 4.62, "step": 373100 }, { "epoch": 0.6963393896416833, "grad_norm": 1.5236045122146606, "learning_rate": 0.0001954315441440794, "loss": 4.3725, "step": 373150 }, { "epoch": 0.6964326952010618, "grad_norm": 0.7871856093406677, "learning_rate": 0.00019543028880257416, "loss": 4.5097, "step": 373200 }, { "epoch": 0.6965260007604404, "grad_norm": 0.9522252678871155, "learning_rate": 0.00019542903329265122, "loss": 4.6569, "step": 373250 }, { "epoch": 0.6966193063198188, "grad_norm": 1.0478781461715698, "learning_rate": 0.00019542777761431286, "loss": 4.5729, "step": 373300 }, { "epoch": 0.6967126118791973, "grad_norm": 1.0283840894699097, "learning_rate": 0.00019542652176756124, "loss": 4.5673, "step": 373350 }, { "epoch": 0.6968059174385758, "grad_norm": 0.9941511154174805, "learning_rate": 0.00019542526575239862, "loss": 4.3899, "step": 373400 }, { "epoch": 0.6968992229979543, "grad_norm": 0.7745673656463623, "learning_rate": 0.00019542400956882718, "loss": 4.413, "step": 373450 }, { "epoch": 0.6969925285573327, "grad_norm": 0.7503075003623962, "learning_rate": 0.00019542275321684916, "loss": 4.4243, "step": 373500 }, { "epoch": 0.6970858341167112, "grad_norm": 0.9689463973045349, "learning_rate": 0.0001954214966964668, "loss": 4.5236, "step": 373550 }, { "epoch": 0.6971791396760898, "grad_norm": 0.9288371801376343, "learning_rate": 0.00019542024000768224, "loss": 4.5392, "step": 373600 }, { "epoch": 0.6972724452354683, "grad_norm": 0.9681942462921143, "learning_rate": 0.00019541898315049777, "loss": 4.5209, "step": 373650 }, { "epoch": 0.6973657507948468, "grad_norm": 1.1082279682159424, "learning_rate": 0.00019541772612491557, "loss": 4.4681, "step": 373700 }, { "epoch": 0.6974590563542252, "grad_norm": 1.3885143995285034, "learning_rate": 0.0001954164689309379, "loss": 4.6048, "step": 373750 }, { "epoch": 0.6975523619136037, "grad_norm": 1.106627106666565, "learning_rate": 0.00019541521156856692, "loss": 4.6785, "step": 373800 }, { "epoch": 0.6976456674729822, "grad_norm": 1.0747969150543213, "learning_rate": 0.00019541395403780491, "loss": 4.8345, "step": 373850 }, { "epoch": 0.6977389730323607, "grad_norm": 1.0435670614242554, "learning_rate": 0.00019541269633865403, "loss": 4.4634, "step": 373900 }, { "epoch": 0.6978322785917392, "grad_norm": 1.3467811346054077, "learning_rate": 0.00019541143847111658, "loss": 4.4974, "step": 373950 }, { "epoch": 0.6979255841511177, "grad_norm": 1.0148619413375854, "learning_rate": 0.00019541018043519467, "loss": 4.4536, "step": 374000 }, { "epoch": 0.6980188897104962, "grad_norm": 0.9159517288208008, "learning_rate": 0.00019540892223089064, "loss": 4.4959, "step": 374050 }, { "epoch": 0.6981121952698747, "grad_norm": 0.9616969227790833, "learning_rate": 0.0001954076638582066, "loss": 4.5215, "step": 374100 }, { "epoch": 0.6982055008292531, "grad_norm": 1.0186048746109009, "learning_rate": 0.00019540640531714483, "loss": 4.7394, "step": 374150 }, { "epoch": 0.6982988063886316, "grad_norm": 1.0677536725997925, "learning_rate": 0.00019540514660770756, "loss": 4.4439, "step": 374200 }, { "epoch": 0.6983921119480101, "grad_norm": 1.3627240657806396, "learning_rate": 0.00019540388772989697, "loss": 4.6828, "step": 374250 }, { "epoch": 0.6984854175073887, "grad_norm": 0.9445304274559021, "learning_rate": 0.0001954026286837153, "loss": 4.6344, "step": 374300 }, { "epoch": 0.6985787230667672, "grad_norm": 0.7951604723930359, "learning_rate": 0.0001954013694691648, "loss": 4.6203, "step": 374350 }, { "epoch": 0.6986720286261456, "grad_norm": 1.1664619445800781, "learning_rate": 0.00019540011008624766, "loss": 4.6081, "step": 374400 }, { "epoch": 0.6987653341855241, "grad_norm": 1.1221282482147217, "learning_rate": 0.0001953988505349661, "loss": 4.7079, "step": 374450 }, { "epoch": 0.6988586397449026, "grad_norm": 1.0007444620132446, "learning_rate": 0.00019539759081532238, "loss": 4.497, "step": 374500 }, { "epoch": 0.6989519453042811, "grad_norm": 1.0179212093353271, "learning_rate": 0.0001953963309273187, "loss": 4.5788, "step": 374550 }, { "epoch": 0.6990452508636595, "grad_norm": 1.243072271347046, "learning_rate": 0.00019539507087095725, "loss": 4.5208, "step": 374600 }, { "epoch": 0.6991385564230381, "grad_norm": 1.132432222366333, "learning_rate": 0.0001953938106462403, "loss": 4.7902, "step": 374650 }, { "epoch": 0.6992318619824166, "grad_norm": 1.2821457386016846, "learning_rate": 0.00019539255025317006, "loss": 4.5234, "step": 374700 }, { "epoch": 0.6993251675417951, "grad_norm": 1.0446197986602783, "learning_rate": 0.00019539128969174878, "loss": 4.5727, "step": 374750 }, { "epoch": 0.6994184731011736, "grad_norm": 0.5463175177574158, "learning_rate": 0.00019539002896197863, "loss": 4.4624, "step": 374800 }, { "epoch": 0.699511778660552, "grad_norm": 0.9870461225509644, "learning_rate": 0.00019538876806386187, "loss": 4.5074, "step": 374850 }, { "epoch": 0.6996050842199305, "grad_norm": 1.0726468563079834, "learning_rate": 0.0001953875069974007, "loss": 4.4284, "step": 374900 }, { "epoch": 0.699698389779309, "grad_norm": 1.2654712200164795, "learning_rate": 0.0001953862457625974, "loss": 4.574, "step": 374950 }, { "epoch": 0.6997916953386876, "grad_norm": 0.9296988844871521, "learning_rate": 0.00019538498435945414, "loss": 4.5662, "step": 375000 }, { "epoch": 0.6997916953386876, "eval_loss": 4.739675521850586, "eval_runtime": 230.6566, "eval_samples_per_second": 11.307, "eval_steps_per_second": 11.307, "eval_tts_loss": 7.568869058622699, "step": 375000 }, { "epoch": 0.699885000898066, "grad_norm": 0.8359538912773132, "learning_rate": 0.00019538372278797315, "loss": 4.3433, "step": 375050 }, { "epoch": 0.6999783064574445, "grad_norm": 1.1886621713638306, "learning_rate": 0.00019538246104815668, "loss": 4.5397, "step": 375100 }, { "epoch": 0.700071612016823, "grad_norm": 0.8133793473243713, "learning_rate": 0.00019538119914000695, "loss": 4.3441, "step": 375150 }, { "epoch": 0.7001649175762015, "grad_norm": 0.9325920939445496, "learning_rate": 0.00019537993706352622, "loss": 4.5081, "step": 375200 }, { "epoch": 0.70025822313558, "grad_norm": 0.9943052530288696, "learning_rate": 0.00019537867481871665, "loss": 4.3981, "step": 375250 }, { "epoch": 0.7003515286949584, "grad_norm": 0.8147107362747192, "learning_rate": 0.0001953774124055805, "loss": 4.6357, "step": 375300 }, { "epoch": 0.700444834254337, "grad_norm": 1.0404388904571533, "learning_rate": 0.00019537614982412, "loss": 4.7682, "step": 375350 }, { "epoch": 0.7005381398137155, "grad_norm": 1.011792540550232, "learning_rate": 0.0001953748870743374, "loss": 4.3843, "step": 375400 }, { "epoch": 0.700631445373094, "grad_norm": 1.0004193782806396, "learning_rate": 0.0001953736241562349, "loss": 4.6198, "step": 375450 }, { "epoch": 0.7007247509324724, "grad_norm": 1.302375316619873, "learning_rate": 0.00019537236106981473, "loss": 4.7347, "step": 375500 }, { "epoch": 0.7008180564918509, "grad_norm": 1.1643980741500854, "learning_rate": 0.00019537109781507913, "loss": 4.6438, "step": 375550 }, { "epoch": 0.7009113620512294, "grad_norm": 0.9268198013305664, "learning_rate": 0.0001953698343920303, "loss": 4.5663, "step": 375600 }, { "epoch": 0.7010046676106079, "grad_norm": 0.8678167462348938, "learning_rate": 0.0001953685708006705, "loss": 4.6092, "step": 375650 }, { "epoch": 0.7010979731699865, "grad_norm": 0.8486493825912476, "learning_rate": 0.00019536730704100195, "loss": 4.6546, "step": 375700 }, { "epoch": 0.7011912787293649, "grad_norm": 1.1792452335357666, "learning_rate": 0.00019536604311302691, "loss": 4.6587, "step": 375750 }, { "epoch": 0.7012845842887434, "grad_norm": 1.2924436330795288, "learning_rate": 0.00019536477901674758, "loss": 4.7177, "step": 375800 }, { "epoch": 0.7013778898481219, "grad_norm": 1.1593018770217896, "learning_rate": 0.00019536351475216618, "loss": 4.3727, "step": 375850 }, { "epoch": 0.7014711954075004, "grad_norm": 1.359183430671692, "learning_rate": 0.00019536225031928494, "loss": 4.6466, "step": 375900 }, { "epoch": 0.7015645009668788, "grad_norm": 1.0704420804977417, "learning_rate": 0.00019536098571810615, "loss": 4.5263, "step": 375950 }, { "epoch": 0.7016578065262573, "grad_norm": 0.9053759574890137, "learning_rate": 0.00019535972094863197, "loss": 4.4246, "step": 376000 }, { "epoch": 0.7017511120856359, "grad_norm": 0.872847855091095, "learning_rate": 0.00019535845601086464, "loss": 4.8251, "step": 376050 }, { "epoch": 0.7018444176450144, "grad_norm": 1.0933473110198975, "learning_rate": 0.00019535719090480646, "loss": 4.6556, "step": 376100 }, { "epoch": 0.7019377232043928, "grad_norm": 1.1355681419372559, "learning_rate": 0.00019535592563045963, "loss": 4.3005, "step": 376150 }, { "epoch": 0.7020310287637713, "grad_norm": 1.106989860534668, "learning_rate": 0.00019535466018782632, "loss": 4.4891, "step": 376200 }, { "epoch": 0.7021243343231498, "grad_norm": 1.1264011859893799, "learning_rate": 0.00019535339457690882, "loss": 4.5817, "step": 376250 }, { "epoch": 0.7022176398825283, "grad_norm": 0.9267123937606812, "learning_rate": 0.00019535212879770938, "loss": 4.457, "step": 376300 }, { "epoch": 0.7023109454419068, "grad_norm": 1.2155795097351074, "learning_rate": 0.00019535086285023018, "loss": 4.5067, "step": 376350 }, { "epoch": 0.7024042510012852, "grad_norm": 1.0612845420837402, "learning_rate": 0.0001953495967344735, "loss": 4.5759, "step": 376400 }, { "epoch": 0.7024975565606638, "grad_norm": 1.1359730958938599, "learning_rate": 0.0001953483304504416, "loss": 4.5738, "step": 376450 }, { "epoch": 0.7025908621200423, "grad_norm": 1.0387494564056396, "learning_rate": 0.0001953470639981366, "loss": 4.7473, "step": 376500 }, { "epoch": 0.7026841676794208, "grad_norm": 1.2349492311477661, "learning_rate": 0.00019534579737756085, "loss": 4.6316, "step": 376550 }, { "epoch": 0.7027774732387992, "grad_norm": 1.0554500818252563, "learning_rate": 0.00019534453058871654, "loss": 4.4325, "step": 376600 }, { "epoch": 0.7028707787981777, "grad_norm": 0.9486688375473022, "learning_rate": 0.00019534326363160592, "loss": 4.5899, "step": 376650 }, { "epoch": 0.7029640843575562, "grad_norm": 0.8850482106208801, "learning_rate": 0.00019534199650623118, "loss": 4.7084, "step": 376700 }, { "epoch": 0.7030573899169347, "grad_norm": 0.7264657020568848, "learning_rate": 0.00019534072921259464, "loss": 4.4336, "step": 376750 }, { "epoch": 0.7031506954763133, "grad_norm": 1.0185405015945435, "learning_rate": 0.00019533946175069846, "loss": 4.4325, "step": 376800 }, { "epoch": 0.7032440010356917, "grad_norm": 1.1543070077896118, "learning_rate": 0.0001953381941205449, "loss": 4.6538, "step": 376850 }, { "epoch": 0.7033373065950702, "grad_norm": 1.1681410074234009, "learning_rate": 0.0001953369263221362, "loss": 4.5526, "step": 376900 }, { "epoch": 0.7034306121544487, "grad_norm": 0.8196843862533569, "learning_rate": 0.00019533565835547461, "loss": 4.601, "step": 376950 }, { "epoch": 0.7035239177138272, "grad_norm": 1.115649700164795, "learning_rate": 0.0001953343902205624, "loss": 4.4583, "step": 377000 }, { "epoch": 0.7036172232732056, "grad_norm": 1.201120376586914, "learning_rate": 0.00019533312191740172, "loss": 4.4975, "step": 377050 }, { "epoch": 0.7037105288325841, "grad_norm": 0.9064095616340637, "learning_rate": 0.00019533185344599484, "loss": 4.578, "step": 377100 }, { "epoch": 0.7038038343919627, "grad_norm": 1.2267018556594849, "learning_rate": 0.00019533058480634405, "loss": 4.5731, "step": 377150 }, { "epoch": 0.7038971399513412, "grad_norm": 1.2399940490722656, "learning_rate": 0.00019532931599845155, "loss": 4.6491, "step": 377200 }, { "epoch": 0.7039904455107197, "grad_norm": 0.991210401058197, "learning_rate": 0.00019532804702231956, "loss": 4.5972, "step": 377250 }, { "epoch": 0.7040837510700981, "grad_norm": 0.9711277484893799, "learning_rate": 0.00019532677787795035, "loss": 4.5564, "step": 377300 }, { "epoch": 0.7041770566294766, "grad_norm": 1.2676489353179932, "learning_rate": 0.00019532550856534613, "loss": 4.4359, "step": 377350 }, { "epoch": 0.7042703621888551, "grad_norm": 1.5608159303665161, "learning_rate": 0.0001953242390845092, "loss": 4.4103, "step": 377400 }, { "epoch": 0.7043636677482336, "grad_norm": 0.9402786493301392, "learning_rate": 0.00019532296943544172, "loss": 4.4642, "step": 377450 }, { "epoch": 0.7044569733076121, "grad_norm": 0.8223136067390442, "learning_rate": 0.000195321699618146, "loss": 4.4978, "step": 377500 }, { "epoch": 0.7045502788669906, "grad_norm": 0.744204044342041, "learning_rate": 0.00019532042963262423, "loss": 4.419, "step": 377550 }, { "epoch": 0.7046435844263691, "grad_norm": 1.0643330812454224, "learning_rate": 0.00019531915947887868, "loss": 4.5853, "step": 377600 }, { "epoch": 0.7047368899857476, "grad_norm": 1.1969504356384277, "learning_rate": 0.0001953178891569116, "loss": 4.6716, "step": 377650 }, { "epoch": 0.704830195545126, "grad_norm": 0.9131072759628296, "learning_rate": 0.0001953166186667252, "loss": 4.5512, "step": 377700 }, { "epoch": 0.7049235011045045, "grad_norm": 1.1578060388565063, "learning_rate": 0.00019531534800832175, "loss": 4.5252, "step": 377750 }, { "epoch": 0.705016806663883, "grad_norm": 1.3309463262557983, "learning_rate": 0.0001953140771817035, "loss": 4.8214, "step": 377800 }, { "epoch": 0.7051101122232616, "grad_norm": 1.3318638801574707, "learning_rate": 0.00019531280618687266, "loss": 4.3906, "step": 377850 }, { "epoch": 0.7052034177826401, "grad_norm": 0.9537114500999451, "learning_rate": 0.00019531153502383149, "loss": 4.6278, "step": 377900 }, { "epoch": 0.7052967233420185, "grad_norm": 1.287827968597412, "learning_rate": 0.0001953102636925822, "loss": 4.4588, "step": 377950 }, { "epoch": 0.705390028901397, "grad_norm": 1.0049210786819458, "learning_rate": 0.0001953089921931271, "loss": 4.3171, "step": 378000 }, { "epoch": 0.705390028901397, "eval_loss": 4.741976261138916, "eval_runtime": 229.2276, "eval_samples_per_second": 11.377, "eval_steps_per_second": 11.377, "eval_tts_loss": 7.590075247879626, "step": 378000 }, { "epoch": 0.7054833344607755, "grad_norm": 0.9821569919586182, "learning_rate": 0.0001953077205254684, "loss": 4.7729, "step": 378050 }, { "epoch": 0.705576640020154, "grad_norm": 1.08197820186615, "learning_rate": 0.00019530644868960833, "loss": 4.5847, "step": 378100 }, { "epoch": 0.7056699455795324, "grad_norm": 0.9573339223861694, "learning_rate": 0.00019530517668554914, "loss": 4.653, "step": 378150 }, { "epoch": 0.705763251138911, "grad_norm": 1.0623559951782227, "learning_rate": 0.0001953039045132931, "loss": 4.5052, "step": 378200 }, { "epoch": 0.7058565566982895, "grad_norm": 0.9007604718208313, "learning_rate": 0.00019530263217284243, "loss": 4.545, "step": 378250 }, { "epoch": 0.705949862257668, "grad_norm": 1.1949926614761353, "learning_rate": 0.00019530135966419938, "loss": 4.5462, "step": 378300 }, { "epoch": 0.7060431678170465, "grad_norm": 1.0200376510620117, "learning_rate": 0.00019530008698736618, "loss": 4.5997, "step": 378350 }, { "epoch": 0.7061364733764249, "grad_norm": 1.2400939464569092, "learning_rate": 0.00019529881414234515, "loss": 4.5039, "step": 378400 }, { "epoch": 0.7062297789358034, "grad_norm": 0.8792778253555298, "learning_rate": 0.00019529754112913844, "loss": 4.7137, "step": 378450 }, { "epoch": 0.7063230844951819, "grad_norm": 1.0362474918365479, "learning_rate": 0.00019529626794774835, "loss": 4.5742, "step": 378500 }, { "epoch": 0.7064163900545605, "grad_norm": 0.8532072901725769, "learning_rate": 0.0001952949945981771, "loss": 4.6874, "step": 378550 }, { "epoch": 0.706509695613939, "grad_norm": 0.7794941663742065, "learning_rate": 0.00019529372108042695, "loss": 4.3726, "step": 378600 }, { "epoch": 0.7066030011733174, "grad_norm": 1.080212116241455, "learning_rate": 0.00019529244739450018, "loss": 4.4039, "step": 378650 }, { "epoch": 0.7066963067326959, "grad_norm": 0.8216705918312073, "learning_rate": 0.000195291173540399, "loss": 4.7282, "step": 378700 }, { "epoch": 0.7067896122920744, "grad_norm": 1.154890537261963, "learning_rate": 0.00019528989951812564, "loss": 4.4009, "step": 378750 }, { "epoch": 0.7068829178514529, "grad_norm": 1.1282939910888672, "learning_rate": 0.0001952886253276824, "loss": 4.5085, "step": 378800 }, { "epoch": 0.7069762234108313, "grad_norm": 0.9745324850082397, "learning_rate": 0.0001952873509690715, "loss": 4.6955, "step": 378850 }, { "epoch": 0.7070695289702099, "grad_norm": 0.9145590662956238, "learning_rate": 0.0001952860764422952, "loss": 4.5594, "step": 378900 }, { "epoch": 0.7071628345295884, "grad_norm": 0.9380348324775696, "learning_rate": 0.00019528480174735573, "loss": 4.4592, "step": 378950 }, { "epoch": 0.7072561400889669, "grad_norm": 0.9864219427108765, "learning_rate": 0.00019528352688425533, "loss": 4.5247, "step": 379000 }, { "epoch": 0.7073494456483453, "grad_norm": 1.1214337348937988, "learning_rate": 0.0001952822518529963, "loss": 4.5793, "step": 379050 }, { "epoch": 0.7074427512077238, "grad_norm": 0.8948968648910522, "learning_rate": 0.00019528097665358086, "loss": 4.412, "step": 379100 }, { "epoch": 0.7075360567671023, "grad_norm": 0.9934725165367126, "learning_rate": 0.00019527970128601122, "loss": 4.6638, "step": 379150 }, { "epoch": 0.7076293623264808, "grad_norm": 0.9161415100097656, "learning_rate": 0.00019527842575028971, "loss": 4.6051, "step": 379200 }, { "epoch": 0.7077226678858594, "grad_norm": 1.0251588821411133, "learning_rate": 0.00019527715004641856, "loss": 4.3529, "step": 379250 }, { "epoch": 0.7078159734452378, "grad_norm": 1.0846620798110962, "learning_rate": 0.00019527587417439996, "loss": 4.5257, "step": 379300 }, { "epoch": 0.7079092790046163, "grad_norm": 1.136697769165039, "learning_rate": 0.00019527459813423623, "loss": 4.4867, "step": 379350 }, { "epoch": 0.7080025845639948, "grad_norm": 1.052685022354126, "learning_rate": 0.0001952733219259296, "loss": 4.6143, "step": 379400 }, { "epoch": 0.7080958901233733, "grad_norm": 0.9956856966018677, "learning_rate": 0.00019527204554948231, "loss": 4.5272, "step": 379450 }, { "epoch": 0.7081891956827517, "grad_norm": 1.0275317430496216, "learning_rate": 0.00019527076900489662, "loss": 4.7828, "step": 379500 }, { "epoch": 0.7082825012421302, "grad_norm": 0.8791360855102539, "learning_rate": 0.00019526949229217483, "loss": 4.5505, "step": 379550 }, { "epoch": 0.7083758068015088, "grad_norm": 0.9856266975402832, "learning_rate": 0.0001952682154113191, "loss": 4.7794, "step": 379600 }, { "epoch": 0.7084691123608873, "grad_norm": 0.8814522624015808, "learning_rate": 0.00019526693836233175, "loss": 4.5435, "step": 379650 }, { "epoch": 0.7085624179202658, "grad_norm": 0.9303852319717407, "learning_rate": 0.000195265661145215, "loss": 4.6448, "step": 379700 }, { "epoch": 0.7086557234796442, "grad_norm": 0.9985621571540833, "learning_rate": 0.00019526438375997115, "loss": 4.6232, "step": 379750 }, { "epoch": 0.7087490290390227, "grad_norm": 1.2800402641296387, "learning_rate": 0.00019526310620660242, "loss": 4.6542, "step": 379800 }, { "epoch": 0.7088423345984012, "grad_norm": 1.2168068885803223, "learning_rate": 0.00019526182848511104, "loss": 4.6047, "step": 379850 }, { "epoch": 0.7089356401577797, "grad_norm": 0.9573338627815247, "learning_rate": 0.0001952605505954993, "loss": 4.4232, "step": 379900 }, { "epoch": 0.7090289457171582, "grad_norm": 0.991240918636322, "learning_rate": 0.0001952592725377695, "loss": 4.5541, "step": 379950 }, { "epoch": 0.7091222512765367, "grad_norm": 1.210038185119629, "learning_rate": 0.0001952579943119238, "loss": 4.6207, "step": 380000 }, { "epoch": 0.7092155568359152, "grad_norm": 1.135339379310608, "learning_rate": 0.0001952567159179645, "loss": 4.5715, "step": 380050 }, { "epoch": 0.7093088623952937, "grad_norm": 1.2163465023040771, "learning_rate": 0.00019525543735589388, "loss": 4.5323, "step": 380100 }, { "epoch": 0.7094021679546721, "grad_norm": 1.2491923570632935, "learning_rate": 0.00019525415862571416, "loss": 4.6131, "step": 380150 }, { "epoch": 0.7094954735140506, "grad_norm": 0.6898648738861084, "learning_rate": 0.0001952528797274276, "loss": 4.6363, "step": 380200 }, { "epoch": 0.7095887790734291, "grad_norm": 0.7697023153305054, "learning_rate": 0.0001952516006610365, "loss": 4.4572, "step": 380250 }, { "epoch": 0.7096820846328077, "grad_norm": 0.7443915605545044, "learning_rate": 0.00019525032142654304, "loss": 4.5029, "step": 380300 }, { "epoch": 0.7097753901921862, "grad_norm": 1.4766536951065063, "learning_rate": 0.00019524904202394955, "loss": 4.5291, "step": 380350 }, { "epoch": 0.7098686957515646, "grad_norm": 0.9718853831291199, "learning_rate": 0.00019524776245325825, "loss": 4.6251, "step": 380400 }, { "epoch": 0.7099620013109431, "grad_norm": 1.1660414934158325, "learning_rate": 0.00019524648271447144, "loss": 4.5658, "step": 380450 }, { "epoch": 0.7100553068703216, "grad_norm": 1.0906556844711304, "learning_rate": 0.00019524520280759134, "loss": 4.4738, "step": 380500 }, { "epoch": 0.7101486124297001, "grad_norm": 1.1851836442947388, "learning_rate": 0.00019524392273262018, "loss": 4.5588, "step": 380550 }, { "epoch": 0.7102419179890785, "grad_norm": 1.3385717868804932, "learning_rate": 0.0001952426424895603, "loss": 4.4783, "step": 380600 }, { "epoch": 0.7103352235484571, "grad_norm": 1.1230698823928833, "learning_rate": 0.00019524136207841385, "loss": 4.7116, "step": 380650 }, { "epoch": 0.7104285291078356, "grad_norm": 1.1039788722991943, "learning_rate": 0.0001952400814991832, "loss": 4.5662, "step": 380700 }, { "epoch": 0.7105218346672141, "grad_norm": 0.9813905358314514, "learning_rate": 0.00019523880075187058, "loss": 4.612, "step": 380750 }, { "epoch": 0.7106151402265926, "grad_norm": 1.1137398481369019, "learning_rate": 0.0001952375198364782, "loss": 4.5109, "step": 380800 }, { "epoch": 0.710708445785971, "grad_norm": 1.0999574661254883, "learning_rate": 0.00019523623875300836, "loss": 4.4076, "step": 380850 }, { "epoch": 0.7108017513453495, "grad_norm": 1.1628655195236206, "learning_rate": 0.0001952349575014633, "loss": 4.6405, "step": 380900 }, { "epoch": 0.710895056904728, "grad_norm": 1.07998526096344, "learning_rate": 0.00019523367608184535, "loss": 4.588, "step": 380950 }, { "epoch": 0.7109883624641066, "grad_norm": 1.0283795595169067, "learning_rate": 0.00019523239449415668, "loss": 4.6904, "step": 381000 }, { "epoch": 0.7109883624641066, "eval_loss": 4.738658905029297, "eval_runtime": 230.5123, "eval_samples_per_second": 11.314, "eval_steps_per_second": 11.314, "eval_tts_loss": 7.550678487040068, "step": 381000 }, { "epoch": 0.711081668023485, "grad_norm": 1.0346516370773315, "learning_rate": 0.0001952311127383996, "loss": 4.3708, "step": 381050 }, { "epoch": 0.7111749735828635, "grad_norm": 1.0621981620788574, "learning_rate": 0.00019522983081457634, "loss": 4.642, "step": 381100 }, { "epoch": 0.711268279142242, "grad_norm": 1.1409293413162231, "learning_rate": 0.00019522854872268923, "loss": 4.4803, "step": 381150 }, { "epoch": 0.7113615847016205, "grad_norm": 0.9242346286773682, "learning_rate": 0.00019522726646274046, "loss": 4.4169, "step": 381200 }, { "epoch": 0.711454890260999, "grad_norm": 1.2605868577957153, "learning_rate": 0.00019522598403473232, "loss": 4.7021, "step": 381250 }, { "epoch": 0.7115481958203774, "grad_norm": 1.2490514516830444, "learning_rate": 0.00019522470143866708, "loss": 4.6978, "step": 381300 }, { "epoch": 0.711641501379756, "grad_norm": 1.114847183227539, "learning_rate": 0.000195223418674547, "loss": 4.5004, "step": 381350 }, { "epoch": 0.7117348069391345, "grad_norm": 1.267319679260254, "learning_rate": 0.00019522213574237434, "loss": 4.6773, "step": 381400 }, { "epoch": 0.711828112498513, "grad_norm": 1.009610652923584, "learning_rate": 0.00019522085264215135, "loss": 4.4301, "step": 381450 }, { "epoch": 0.7119214180578914, "grad_norm": 0.7319119572639465, "learning_rate": 0.00019521956937388032, "loss": 4.5613, "step": 381500 }, { "epoch": 0.7120147236172699, "grad_norm": 1.0810520648956299, "learning_rate": 0.00019521828593756352, "loss": 4.5926, "step": 381550 }, { "epoch": 0.7121080291766484, "grad_norm": 1.1929056644439697, "learning_rate": 0.00019521700233320318, "loss": 4.623, "step": 381600 }, { "epoch": 0.7122013347360269, "grad_norm": 1.0149575471878052, "learning_rate": 0.0001952157185608016, "loss": 4.4202, "step": 381650 }, { "epoch": 0.7122946402954053, "grad_norm": 0.918735146522522, "learning_rate": 0.000195214434620361, "loss": 4.6246, "step": 381700 }, { "epoch": 0.7123879458547839, "grad_norm": 1.206817388534546, "learning_rate": 0.0001952131505118837, "loss": 4.5127, "step": 381750 }, { "epoch": 0.7124812514141624, "grad_norm": 1.0245983600616455, "learning_rate": 0.00019521186623537196, "loss": 4.6317, "step": 381800 }, { "epoch": 0.7125745569735409, "grad_norm": 1.0655715465545654, "learning_rate": 0.000195210581790828, "loss": 4.3029, "step": 381850 }, { "epoch": 0.7126678625329194, "grad_norm": 1.1398463249206543, "learning_rate": 0.00019520929717825413, "loss": 4.5782, "step": 381900 }, { "epoch": 0.7127611680922978, "grad_norm": 1.129570484161377, "learning_rate": 0.00019520801239765258, "loss": 4.6278, "step": 381950 }, { "epoch": 0.7128544736516763, "grad_norm": 1.1474553346633911, "learning_rate": 0.00019520672744902567, "loss": 4.4637, "step": 382000 }, { "epoch": 0.7129477792110548, "grad_norm": 1.088990330696106, "learning_rate": 0.00019520544233237561, "loss": 4.4453, "step": 382050 }, { "epoch": 0.7130410847704334, "grad_norm": 0.9756652116775513, "learning_rate": 0.00019520415704770474, "loss": 4.4904, "step": 382100 }, { "epoch": 0.7131343903298119, "grad_norm": 1.2050561904907227, "learning_rate": 0.00019520287159501524, "loss": 4.6689, "step": 382150 }, { "epoch": 0.7132276958891903, "grad_norm": 1.0057941675186157, "learning_rate": 0.00019520158597430943, "loss": 4.5429, "step": 382200 }, { "epoch": 0.7133210014485688, "grad_norm": 0.9556266665458679, "learning_rate": 0.0001952003001855896, "loss": 4.4695, "step": 382250 }, { "epoch": 0.7134143070079473, "grad_norm": 0.8169947862625122, "learning_rate": 0.00019519901422885796, "loss": 4.6116, "step": 382300 }, { "epoch": 0.7135076125673258, "grad_norm": 1.0497891902923584, "learning_rate": 0.00019519772810411683, "loss": 4.5701, "step": 382350 }, { "epoch": 0.7136009181267042, "grad_norm": 0.8632709383964539, "learning_rate": 0.00019519644181136843, "loss": 4.47, "step": 382400 }, { "epoch": 0.7136942236860828, "grad_norm": 1.1212148666381836, "learning_rate": 0.0001951951553506151, "loss": 4.4381, "step": 382450 }, { "epoch": 0.7137875292454613, "grad_norm": 1.2033298015594482, "learning_rate": 0.00019519386872185907, "loss": 4.9272, "step": 382500 }, { "epoch": 0.7138808348048398, "grad_norm": 1.0442355871200562, "learning_rate": 0.0001951925819251026, "loss": 4.5303, "step": 382550 }, { "epoch": 0.7139741403642182, "grad_norm": 1.5087108612060547, "learning_rate": 0.00019519129496034797, "loss": 4.59, "step": 382600 }, { "epoch": 0.7140674459235967, "grad_norm": 1.2250038385391235, "learning_rate": 0.00019519000782759744, "loss": 4.5591, "step": 382650 }, { "epoch": 0.7141607514829752, "grad_norm": 1.4251493215560913, "learning_rate": 0.0001951887205268533, "loss": 4.4625, "step": 382700 }, { "epoch": 0.7142540570423537, "grad_norm": 1.1542210578918457, "learning_rate": 0.00019518743305811784, "loss": 4.4542, "step": 382750 }, { "epoch": 0.7143473626017323, "grad_norm": 1.1355654001235962, "learning_rate": 0.0001951861454213933, "loss": 4.3306, "step": 382800 }, { "epoch": 0.7144406681611107, "grad_norm": 0.8172572255134583, "learning_rate": 0.00019518485761668198, "loss": 4.6956, "step": 382850 }, { "epoch": 0.7145339737204892, "grad_norm": 1.0452672243118286, "learning_rate": 0.0001951835696439861, "loss": 4.577, "step": 382900 }, { "epoch": 0.7146272792798677, "grad_norm": 1.2720973491668701, "learning_rate": 0.00019518228150330798, "loss": 4.5383, "step": 382950 }, { "epoch": 0.7147205848392462, "grad_norm": 0.8681235313415527, "learning_rate": 0.00019518099319464985, "loss": 4.5858, "step": 383000 }, { "epoch": 0.7148138903986246, "grad_norm": 1.3290435075759888, "learning_rate": 0.00019517970471801408, "loss": 4.8388, "step": 383050 }, { "epoch": 0.7149071959580031, "grad_norm": 0.9616467356681824, "learning_rate": 0.00019517841607340284, "loss": 4.4234, "step": 383100 }, { "epoch": 0.7150005015173817, "grad_norm": 1.1521810293197632, "learning_rate": 0.00019517712726081844, "loss": 4.5619, "step": 383150 }, { "epoch": 0.7150938070767602, "grad_norm": 0.8024904131889343, "learning_rate": 0.00019517583828026314, "loss": 4.7557, "step": 383200 }, { "epoch": 0.7151871126361387, "grad_norm": 1.0209003686904907, "learning_rate": 0.00019517454913173927, "loss": 4.6532, "step": 383250 }, { "epoch": 0.7152804181955171, "grad_norm": 1.0985634326934814, "learning_rate": 0.00019517325981524905, "loss": 4.4483, "step": 383300 }, { "epoch": 0.7153737237548956, "grad_norm": 1.0770183801651, "learning_rate": 0.00019517197033079478, "loss": 4.7606, "step": 383350 }, { "epoch": 0.7154670293142741, "grad_norm": 0.8601162433624268, "learning_rate": 0.00019517068067837872, "loss": 4.439, "step": 383400 }, { "epoch": 0.7155603348736526, "grad_norm": 1.282110571861267, "learning_rate": 0.00019516939085800316, "loss": 4.4998, "step": 383450 }, { "epoch": 0.7156536404330311, "grad_norm": 1.1574286222457886, "learning_rate": 0.00019516810086967038, "loss": 4.5013, "step": 383500 }, { "epoch": 0.7157469459924096, "grad_norm": 1.2425411939620972, "learning_rate": 0.00019516681071338264, "loss": 4.6996, "step": 383550 }, { "epoch": 0.7158402515517881, "grad_norm": 1.076066255569458, "learning_rate": 0.00019516552038914222, "loss": 4.6142, "step": 383600 }, { "epoch": 0.7159335571111666, "grad_norm": 1.0551059246063232, "learning_rate": 0.0001951642298969514, "loss": 4.7165, "step": 383650 }, { "epoch": 0.716026862670545, "grad_norm": 1.227857232093811, "learning_rate": 0.00019516293923681246, "loss": 4.5505, "step": 383700 }, { "epoch": 0.7161201682299235, "grad_norm": 1.0538171529769897, "learning_rate": 0.00019516164840872768, "loss": 4.5371, "step": 383750 }, { "epoch": 0.716213473789302, "grad_norm": 0.9511804580688477, "learning_rate": 0.00019516035741269934, "loss": 4.5966, "step": 383800 }, { "epoch": 0.7163067793486806, "grad_norm": 0.7090573310852051, "learning_rate": 0.0001951590662487297, "loss": 4.6471, "step": 383850 }, { "epoch": 0.7164000849080591, "grad_norm": 1.0298354625701904, "learning_rate": 0.00019515777491682106, "loss": 4.5427, "step": 383900 }, { "epoch": 0.7164933904674375, "grad_norm": 1.163970947265625, "learning_rate": 0.00019515648341697568, "loss": 4.4544, "step": 383950 }, { "epoch": 0.716586696026816, "grad_norm": 1.0745140314102173, "learning_rate": 0.00019515519174919586, "loss": 4.3617, "step": 384000 }, { "epoch": 0.716586696026816, "eval_loss": 4.7351579666137695, "eval_runtime": 229.8886, "eval_samples_per_second": 11.345, "eval_steps_per_second": 11.345, "eval_tts_loss": 7.557877517531003, "step": 384000 }, { "epoch": 0.7166800015861945, "grad_norm": 0.9457404613494873, "learning_rate": 0.00019515389991348387, "loss": 4.3421, "step": 384050 }, { "epoch": 0.716773307145573, "grad_norm": 1.0447351932525635, "learning_rate": 0.000195152607909842, "loss": 4.5032, "step": 384100 }, { "epoch": 0.7168666127049514, "grad_norm": 1.032181978225708, "learning_rate": 0.0001951513157382725, "loss": 4.7111, "step": 384150 }, { "epoch": 0.71695991826433, "grad_norm": 0.8992804884910583, "learning_rate": 0.00019515002339877765, "loss": 4.5748, "step": 384200 }, { "epoch": 0.7170532238237085, "grad_norm": 1.2443760633468628, "learning_rate": 0.0001951487308913598, "loss": 4.5732, "step": 384250 }, { "epoch": 0.717146529383087, "grad_norm": 1.1091217994689941, "learning_rate": 0.00019514743821602118, "loss": 4.7897, "step": 384300 }, { "epoch": 0.7172398349424655, "grad_norm": 1.1647604703903198, "learning_rate": 0.00019514614537276403, "loss": 4.3145, "step": 384350 }, { "epoch": 0.7173331405018439, "grad_norm": 0.9686366319656372, "learning_rate": 0.0001951448523615907, "loss": 4.5574, "step": 384400 }, { "epoch": 0.7174264460612224, "grad_norm": 1.0548909902572632, "learning_rate": 0.00019514355918250344, "loss": 4.7254, "step": 384450 }, { "epoch": 0.7175197516206009, "grad_norm": 1.2261728048324585, "learning_rate": 0.00019514226583550458, "loss": 4.524, "step": 384500 }, { "epoch": 0.7176130571799795, "grad_norm": 1.1724244356155396, "learning_rate": 0.0001951409723205963, "loss": 4.6381, "step": 384550 }, { "epoch": 0.717706362739358, "grad_norm": 1.0948948860168457, "learning_rate": 0.00019513967863778097, "loss": 4.6415, "step": 384600 }, { "epoch": 0.7177996682987364, "grad_norm": 1.1220853328704834, "learning_rate": 0.00019513838478706086, "loss": 4.4522, "step": 384650 }, { "epoch": 0.7178929738581149, "grad_norm": 0.9094553589820862, "learning_rate": 0.00019513709076843822, "loss": 4.7201, "step": 384700 }, { "epoch": 0.7179862794174934, "grad_norm": 1.1259245872497559, "learning_rate": 0.00019513579658191537, "loss": 4.4064, "step": 384750 }, { "epoch": 0.7180795849768719, "grad_norm": 1.1270395517349243, "learning_rate": 0.00019513450222749456, "loss": 4.5983, "step": 384800 }, { "epoch": 0.7181728905362503, "grad_norm": 1.206202507019043, "learning_rate": 0.00019513320770517812, "loss": 4.5298, "step": 384850 }, { "epoch": 0.7182661960956289, "grad_norm": 1.294735312461853, "learning_rate": 0.0001951319130149683, "loss": 4.6011, "step": 384900 }, { "epoch": 0.7183595016550074, "grad_norm": 1.2166738510131836, "learning_rate": 0.00019513061815686737, "loss": 4.7282, "step": 384950 }, { "epoch": 0.7184528072143859, "grad_norm": 1.1213105916976929, "learning_rate": 0.00019512932313087765, "loss": 4.4658, "step": 385000 }, { "epoch": 0.7185461127737643, "grad_norm": 1.0961776971817017, "learning_rate": 0.00019512802793700143, "loss": 4.4198, "step": 385050 }, { "epoch": 0.7186394183331428, "grad_norm": 1.1053520441055298, "learning_rate": 0.00019512673257524098, "loss": 4.6368, "step": 385100 }, { "epoch": 0.7187327238925213, "grad_norm": 1.132785677909851, "learning_rate": 0.00019512543704559856, "loss": 4.6449, "step": 385150 }, { "epoch": 0.7188260294518998, "grad_norm": 0.7992883920669556, "learning_rate": 0.0001951241413480765, "loss": 4.5777, "step": 385200 }, { "epoch": 0.7189193350112784, "grad_norm": 1.2569609880447388, "learning_rate": 0.00019512284548267705, "loss": 4.7781, "step": 385250 }, { "epoch": 0.7190126405706568, "grad_norm": 0.7372807264328003, "learning_rate": 0.00019512154944940254, "loss": 4.6389, "step": 385300 }, { "epoch": 0.7191059461300353, "grad_norm": 0.9772287011146545, "learning_rate": 0.00019512025324825517, "loss": 4.5089, "step": 385350 }, { "epoch": 0.7191992516894138, "grad_norm": 1.157492995262146, "learning_rate": 0.00019511895687923737, "loss": 4.6116, "step": 385400 }, { "epoch": 0.7192925572487923, "grad_norm": 1.1132258176803589, "learning_rate": 0.0001951176603423513, "loss": 4.305, "step": 385450 }, { "epoch": 0.7193858628081707, "grad_norm": 1.1786819696426392, "learning_rate": 0.00019511636363759931, "loss": 4.6685, "step": 385500 }, { "epoch": 0.7194791683675492, "grad_norm": 1.2146203517913818, "learning_rate": 0.00019511506676498367, "loss": 4.3969, "step": 385550 }, { "epoch": 0.7195724739269278, "grad_norm": 1.0019917488098145, "learning_rate": 0.00019511376972450667, "loss": 4.7664, "step": 385600 }, { "epoch": 0.7196657794863063, "grad_norm": 1.107676386833191, "learning_rate": 0.00019511247251617064, "loss": 4.5236, "step": 385650 }, { "epoch": 0.7197590850456848, "grad_norm": 1.2492763996124268, "learning_rate": 0.0001951111751399778, "loss": 4.787, "step": 385700 }, { "epoch": 0.7198523906050632, "grad_norm": 0.9558393955230713, "learning_rate": 0.00019510987759593047, "loss": 4.4466, "step": 385750 }, { "epoch": 0.7199456961644417, "grad_norm": 1.5168671607971191, "learning_rate": 0.00019510857988403095, "loss": 4.6665, "step": 385800 }, { "epoch": 0.7200390017238202, "grad_norm": 1.0870764255523682, "learning_rate": 0.0001951072820042815, "loss": 4.4197, "step": 385850 }, { "epoch": 0.7201323072831987, "grad_norm": 1.1981571912765503, "learning_rate": 0.00019510598395668444, "loss": 4.5403, "step": 385900 }, { "epoch": 0.7202256128425772, "grad_norm": 1.0989803075790405, "learning_rate": 0.00019510468574124206, "loss": 4.6166, "step": 385950 }, { "epoch": 0.7203189184019557, "grad_norm": 1.1044586896896362, "learning_rate": 0.00019510338735795668, "loss": 4.6017, "step": 386000 }, { "epoch": 0.7204122239613342, "grad_norm": 1.1001883745193481, "learning_rate": 0.0001951020888068305, "loss": 4.6469, "step": 386050 }, { "epoch": 0.7205055295207127, "grad_norm": 1.145124077796936, "learning_rate": 0.0001951007900878659, "loss": 4.4491, "step": 386100 }, { "epoch": 0.7205988350800911, "grad_norm": 1.18637216091156, "learning_rate": 0.00019509949120106513, "loss": 4.4879, "step": 386150 }, { "epoch": 0.7206921406394696, "grad_norm": 0.5531867146492004, "learning_rate": 0.0001950981921464305, "loss": 4.6821, "step": 386200 }, { "epoch": 0.7207854461988481, "grad_norm": 1.187780737876892, "learning_rate": 0.00019509689292396428, "loss": 4.6521, "step": 386250 }, { "epoch": 0.7208787517582267, "grad_norm": 0.8865212798118591, "learning_rate": 0.00019509559353366878, "loss": 4.5438, "step": 386300 }, { "epoch": 0.7209720573176052, "grad_norm": 0.9871026277542114, "learning_rate": 0.00019509429397554628, "loss": 4.5224, "step": 386350 }, { "epoch": 0.7210653628769836, "grad_norm": 1.2513176202774048, "learning_rate": 0.00019509299424959908, "loss": 4.4853, "step": 386400 }, { "epoch": 0.7211586684363621, "grad_norm": 1.108982801437378, "learning_rate": 0.0001950916943558295, "loss": 4.7211, "step": 386450 }, { "epoch": 0.7212519739957406, "grad_norm": 1.206200361251831, "learning_rate": 0.00019509039429423979, "loss": 4.6938, "step": 386500 }, { "epoch": 0.7213452795551191, "grad_norm": 0.9535238742828369, "learning_rate": 0.00019508909406483227, "loss": 4.5804, "step": 386550 }, { "epoch": 0.7214385851144975, "grad_norm": 1.1816784143447876, "learning_rate": 0.00019508779366760926, "loss": 4.5578, "step": 386600 }, { "epoch": 0.7215318906738761, "grad_norm": 1.0024590492248535, "learning_rate": 0.00019508649310257298, "loss": 4.613, "step": 386650 }, { "epoch": 0.7216251962332546, "grad_norm": 1.145411729812622, "learning_rate": 0.0001950851923697258, "loss": 4.7277, "step": 386700 }, { "epoch": 0.7217185017926331, "grad_norm": 0.9904378056526184, "learning_rate": 0.00019508389146906998, "loss": 4.4863, "step": 386750 }, { "epoch": 0.7218118073520116, "grad_norm": 0.7934876680374146, "learning_rate": 0.0001950825904006078, "loss": 4.6157, "step": 386800 }, { "epoch": 0.72190511291139, "grad_norm": 0.9897130727767944, "learning_rate": 0.0001950812891643416, "loss": 4.5527, "step": 386850 }, { "epoch": 0.7219984184707685, "grad_norm": 1.23386549949646, "learning_rate": 0.0001950799877602737, "loss": 4.579, "step": 386900 }, { "epoch": 0.722091724030147, "grad_norm": 1.0667505264282227, "learning_rate": 0.00019507868618840629, "loss": 4.708, "step": 386950 }, { "epoch": 0.7221850295895256, "grad_norm": 1.2649896144866943, "learning_rate": 0.00019507738444874174, "loss": 4.4828, "step": 387000 }, { "epoch": 0.7221850295895256, "eval_loss": 4.725461483001709, "eval_runtime": 231.3684, "eval_samples_per_second": 11.272, "eval_steps_per_second": 11.272, "eval_tts_loss": 7.58398832848147, "step": 387000 }, { "epoch": 0.722278335148904, "grad_norm": 0.9351997971534729, "learning_rate": 0.00019507608254128232, "loss": 4.5677, "step": 387050 }, { "epoch": 0.7223716407082825, "grad_norm": 1.1455060243606567, "learning_rate": 0.00019507478046603037, "loss": 4.6401, "step": 387100 }, { "epoch": 0.722464946267661, "grad_norm": 0.8970882296562195, "learning_rate": 0.00019507347822298815, "loss": 4.4225, "step": 387150 }, { "epoch": 0.7225582518270395, "grad_norm": 1.546656847000122, "learning_rate": 0.00019507217581215798, "loss": 4.5408, "step": 387200 }, { "epoch": 0.722651557386418, "grad_norm": 1.1023125648498535, "learning_rate": 0.00019507087323354213, "loss": 4.6823, "step": 387250 }, { "epoch": 0.7227448629457964, "grad_norm": 0.9081626534461975, "learning_rate": 0.00019506957048714293, "loss": 4.3673, "step": 387300 }, { "epoch": 0.7228381685051749, "grad_norm": 1.059166669845581, "learning_rate": 0.00019506826757296266, "loss": 4.6348, "step": 387350 }, { "epoch": 0.7229314740645535, "grad_norm": 1.273818016052246, "learning_rate": 0.00019506696449100364, "loss": 4.6823, "step": 387400 }, { "epoch": 0.723024779623932, "grad_norm": 1.3179587125778198, "learning_rate": 0.00019506566124126814, "loss": 4.7256, "step": 387450 }, { "epoch": 0.7231180851833104, "grad_norm": 1.1236895322799683, "learning_rate": 0.0001950643578237585, "loss": 4.7807, "step": 387500 }, { "epoch": 0.7232113907426889, "grad_norm": 1.0101490020751953, "learning_rate": 0.00019506305423847698, "loss": 4.6532, "step": 387550 }, { "epoch": 0.7233046963020674, "grad_norm": 1.0887134075164795, "learning_rate": 0.00019506175048542587, "loss": 4.5312, "step": 387600 }, { "epoch": 0.7233980018614459, "grad_norm": 0.9442769885063171, "learning_rate": 0.00019506044656460752, "loss": 4.4949, "step": 387650 }, { "epoch": 0.7234913074208243, "grad_norm": 1.0077662467956543, "learning_rate": 0.00019505914247602423, "loss": 4.5552, "step": 387700 }, { "epoch": 0.7235846129802029, "grad_norm": 1.100584626197815, "learning_rate": 0.00019505783821967824, "loss": 4.5031, "step": 387750 }, { "epoch": 0.7236779185395814, "grad_norm": 0.8957000374794006, "learning_rate": 0.0001950565337955719, "loss": 4.5858, "step": 387800 }, { "epoch": 0.7237712240989599, "grad_norm": 1.2696412801742554, "learning_rate": 0.00019505522920370752, "loss": 4.4523, "step": 387850 }, { "epoch": 0.7238645296583384, "grad_norm": 1.0976306200027466, "learning_rate": 0.00019505392444408738, "loss": 4.574, "step": 387900 }, { "epoch": 0.7239578352177168, "grad_norm": 1.000716209411621, "learning_rate": 0.0001950526195167138, "loss": 4.4646, "step": 387950 }, { "epoch": 0.7240511407770953, "grad_norm": 0.972341775894165, "learning_rate": 0.00019505131442158907, "loss": 4.5893, "step": 388000 }, { "epoch": 0.7241444463364738, "grad_norm": 1.0794826745986938, "learning_rate": 0.00019505000915871548, "loss": 4.7191, "step": 388050 }, { "epoch": 0.7242377518958524, "grad_norm": 0.9431530833244324, "learning_rate": 0.00019504870372809535, "loss": 4.718, "step": 388100 }, { "epoch": 0.7243310574552309, "grad_norm": 1.200202465057373, "learning_rate": 0.00019504739812973098, "loss": 4.7026, "step": 388150 }, { "epoch": 0.7244243630146093, "grad_norm": 1.0569508075714111, "learning_rate": 0.00019504609236362469, "loss": 4.4948, "step": 388200 }, { "epoch": 0.7245176685739878, "grad_norm": 0.9381723999977112, "learning_rate": 0.00019504478642977876, "loss": 4.5383, "step": 388250 }, { "epoch": 0.7246109741333663, "grad_norm": 1.137317419052124, "learning_rate": 0.00019504348032819548, "loss": 4.6089, "step": 388300 }, { "epoch": 0.7247042796927448, "grad_norm": 1.215718388557434, "learning_rate": 0.0001950421740588772, "loss": 4.7085, "step": 388350 }, { "epoch": 0.7247975852521232, "grad_norm": 1.2167165279388428, "learning_rate": 0.00019504086762182622, "loss": 4.6836, "step": 388400 }, { "epoch": 0.7248908908115018, "grad_norm": 1.219132423400879, "learning_rate": 0.00019503956101704485, "loss": 4.5349, "step": 388450 }, { "epoch": 0.7249841963708803, "grad_norm": 0.8871475458145142, "learning_rate": 0.00019503825424453534, "loss": 4.5926, "step": 388500 }, { "epoch": 0.7250775019302588, "grad_norm": 1.1598362922668457, "learning_rate": 0.00019503694730430007, "loss": 4.6355, "step": 388550 }, { "epoch": 0.7251708074896372, "grad_norm": 0.7973438501358032, "learning_rate": 0.00019503564019634126, "loss": 4.4744, "step": 388600 }, { "epoch": 0.7252641130490157, "grad_norm": 1.1789216995239258, "learning_rate": 0.0001950343329206613, "loss": 4.6298, "step": 388650 }, { "epoch": 0.7253574186083942, "grad_norm": 1.0197705030441284, "learning_rate": 0.00019503302547726244, "loss": 4.4355, "step": 388700 }, { "epoch": 0.7254507241677727, "grad_norm": 1.1121525764465332, "learning_rate": 0.00019503171786614706, "loss": 4.2835, "step": 388750 }, { "epoch": 0.7255440297271513, "grad_norm": 0.8549951910972595, "learning_rate": 0.00019503041008731738, "loss": 4.3967, "step": 388800 }, { "epoch": 0.7256373352865297, "grad_norm": 0.8149085640907288, "learning_rate": 0.00019502910214077576, "loss": 4.4414, "step": 388850 }, { "epoch": 0.7257306408459082, "grad_norm": 1.0984678268432617, "learning_rate": 0.00019502779402652448, "loss": 4.3974, "step": 388900 }, { "epoch": 0.7258239464052867, "grad_norm": 1.0284241437911987, "learning_rate": 0.00019502648574456588, "loss": 4.5263, "step": 388950 }, { "epoch": 0.7259172519646652, "grad_norm": 1.1385748386383057, "learning_rate": 0.00019502517729490227, "loss": 4.6607, "step": 389000 }, { "epoch": 0.7260105575240436, "grad_norm": 1.0035887956619263, "learning_rate": 0.00019502386867753592, "loss": 4.7158, "step": 389050 }, { "epoch": 0.7261038630834221, "grad_norm": 0.9715141654014587, "learning_rate": 0.00019502255989246916, "loss": 4.4282, "step": 389100 }, { "epoch": 0.7261971686428007, "grad_norm": 1.1229904890060425, "learning_rate": 0.0001950212509397043, "loss": 4.3938, "step": 389150 }, { "epoch": 0.7262904742021792, "grad_norm": 0.7794562578201294, "learning_rate": 0.00019501994181924366, "loss": 4.4609, "step": 389200 }, { "epoch": 0.7263837797615577, "grad_norm": 1.1683100461959839, "learning_rate": 0.00019501863253108953, "loss": 4.5351, "step": 389250 }, { "epoch": 0.7264770853209361, "grad_norm": 1.011674404144287, "learning_rate": 0.00019501732307524426, "loss": 4.5928, "step": 389300 }, { "epoch": 0.7265703908803146, "grad_norm": 0.9872075319290161, "learning_rate": 0.0001950160134517101, "loss": 4.7019, "step": 389350 }, { "epoch": 0.7266636964396931, "grad_norm": 1.1409260034561157, "learning_rate": 0.00019501470366048941, "loss": 4.4976, "step": 389400 }, { "epoch": 0.7267570019990716, "grad_norm": 0.9765341877937317, "learning_rate": 0.0001950133937015845, "loss": 4.5095, "step": 389450 }, { "epoch": 0.7268503075584501, "grad_norm": 1.293187141418457, "learning_rate": 0.00019501208357499764, "loss": 4.7893, "step": 389500 }, { "epoch": 0.7269436131178286, "grad_norm": 0.8450595736503601, "learning_rate": 0.0001950107732807312, "loss": 4.5728, "step": 389550 }, { "epoch": 0.7270369186772071, "grad_norm": 0.9158079028129578, "learning_rate": 0.00019500946281878742, "loss": 4.4127, "step": 389600 }, { "epoch": 0.7271302242365856, "grad_norm": 1.1306376457214355, "learning_rate": 0.0001950081521891687, "loss": 4.5249, "step": 389650 }, { "epoch": 0.727223529795964, "grad_norm": 1.2033452987670898, "learning_rate": 0.0001950068413918773, "loss": 4.6137, "step": 389700 }, { "epoch": 0.7273168353553425, "grad_norm": 1.193956732749939, "learning_rate": 0.00019500553042691552, "loss": 4.715, "step": 389750 }, { "epoch": 0.727410140914721, "grad_norm": 1.084000825881958, "learning_rate": 0.00019500421929428573, "loss": 4.5281, "step": 389800 }, { "epoch": 0.7275034464740996, "grad_norm": 1.1593761444091797, "learning_rate": 0.00019500290799399016, "loss": 4.6083, "step": 389850 }, { "epoch": 0.7275967520334781, "grad_norm": 1.0730018615722656, "learning_rate": 0.00019500159652603122, "loss": 4.5715, "step": 389900 }, { "epoch": 0.7276900575928565, "grad_norm": 0.9677862524986267, "learning_rate": 0.00019500028489041114, "loss": 4.5463, "step": 389950 }, { "epoch": 0.727783363152235, "grad_norm": 1.1932060718536377, "learning_rate": 0.00019499897308713232, "loss": 4.5934, "step": 390000 }, { "epoch": 0.727783363152235, "eval_loss": 4.735870361328125, "eval_runtime": 230.6162, "eval_samples_per_second": 11.309, "eval_steps_per_second": 11.309, "eval_tts_loss": 7.53584766990728, "step": 390000 }, { "epoch": 0.7278766687116135, "grad_norm": 0.6686156988143921, "learning_rate": 0.000194997661116197, "loss": 4.5504, "step": 390050 }, { "epoch": 0.727969974270992, "grad_norm": 1.496148705482483, "learning_rate": 0.00019499634897760751, "loss": 4.4838, "step": 390100 }, { "epoch": 0.7280632798303704, "grad_norm": 0.8240087032318115, "learning_rate": 0.0001949950366713662, "loss": 4.5827, "step": 390150 }, { "epoch": 0.728156585389749, "grad_norm": 1.0409563779830933, "learning_rate": 0.00019499372419747534, "loss": 4.6779, "step": 390200 }, { "epoch": 0.7282498909491275, "grad_norm": 0.863301157951355, "learning_rate": 0.0001949924115559373, "loss": 4.7497, "step": 390250 }, { "epoch": 0.728343196508506, "grad_norm": 1.2422351837158203, "learning_rate": 0.00019499109874675436, "loss": 4.697, "step": 390300 }, { "epoch": 0.7284365020678845, "grad_norm": 1.0232807397842407, "learning_rate": 0.00019498978576992885, "loss": 4.6335, "step": 390350 }, { "epoch": 0.7285298076272629, "grad_norm": 0.8831896185874939, "learning_rate": 0.00019498847262546307, "loss": 4.4275, "step": 390400 }, { "epoch": 0.7286231131866414, "grad_norm": 1.2371958494186401, "learning_rate": 0.0001949871593133594, "loss": 4.4267, "step": 390450 }, { "epoch": 0.7287164187460199, "grad_norm": 0.9164360165596008, "learning_rate": 0.00019498584583362003, "loss": 4.498, "step": 390500 }, { "epoch": 0.7288097243053985, "grad_norm": 1.0729354619979858, "learning_rate": 0.0001949845321862474, "loss": 4.8734, "step": 390550 }, { "epoch": 0.728903029864777, "grad_norm": 0.5111291408538818, "learning_rate": 0.00019498321837124375, "loss": 4.4942, "step": 390600 }, { "epoch": 0.7289963354241554, "grad_norm": 0.9883329272270203, "learning_rate": 0.00019498190438861147, "loss": 4.4397, "step": 390650 }, { "epoch": 0.7290896409835339, "grad_norm": 1.101575255393982, "learning_rate": 0.0001949805902383528, "loss": 4.6804, "step": 390700 }, { "epoch": 0.7291829465429124, "grad_norm": 0.9370002746582031, "learning_rate": 0.00019497927592047014, "loss": 4.8269, "step": 390750 }, { "epoch": 0.7292762521022909, "grad_norm": 1.1477497816085815, "learning_rate": 0.00019497796143496575, "loss": 4.6717, "step": 390800 }, { "epoch": 0.7293695576616693, "grad_norm": 1.1353437900543213, "learning_rate": 0.000194976646781842, "loss": 4.7019, "step": 390850 }, { "epoch": 0.7294628632210479, "grad_norm": 0.694190263748169, "learning_rate": 0.00019497533196110114, "loss": 4.5756, "step": 390900 }, { "epoch": 0.7295561687804264, "grad_norm": 1.288985013961792, "learning_rate": 0.00019497401697274555, "loss": 4.6886, "step": 390950 }, { "epoch": 0.7296494743398049, "grad_norm": 0.5960861444473267, "learning_rate": 0.0001949727018167775, "loss": 4.4684, "step": 391000 }, { "epoch": 0.7297427798991833, "grad_norm": 1.026583194732666, "learning_rate": 0.0001949713864931994, "loss": 4.5915, "step": 391050 }, { "epoch": 0.7298360854585618, "grad_norm": 0.8665286302566528, "learning_rate": 0.00019497007100201347, "loss": 4.5365, "step": 391100 }, { "epoch": 0.7299293910179403, "grad_norm": 1.13361656665802, "learning_rate": 0.0001949687553432221, "loss": 4.5007, "step": 391150 }, { "epoch": 0.7300226965773188, "grad_norm": 1.010580062866211, "learning_rate": 0.0001949674395168276, "loss": 4.7294, "step": 391200 }, { "epoch": 0.7301160021366974, "grad_norm": 0.966787576675415, "learning_rate": 0.00019496612352283224, "loss": 4.5658, "step": 391250 }, { "epoch": 0.7302093076960758, "grad_norm": 1.1436415910720825, "learning_rate": 0.00019496480736123838, "loss": 4.7356, "step": 391300 }, { "epoch": 0.7303026132554543, "grad_norm": 1.1993733644485474, "learning_rate": 0.00019496349103204837, "loss": 4.7252, "step": 391350 }, { "epoch": 0.7303959188148328, "grad_norm": 0.901866614818573, "learning_rate": 0.00019496217453526452, "loss": 4.5208, "step": 391400 }, { "epoch": 0.7304892243742113, "grad_norm": 1.084151029586792, "learning_rate": 0.00019496085787088912, "loss": 4.5361, "step": 391450 }, { "epoch": 0.7305825299335897, "grad_norm": 0.8758310675621033, "learning_rate": 0.0001949595410389245, "loss": 4.807, "step": 391500 }, { "epoch": 0.7306758354929682, "grad_norm": 1.0187368392944336, "learning_rate": 0.000194958224039373, "loss": 4.4729, "step": 391550 }, { "epoch": 0.7307691410523468, "grad_norm": 1.1745333671569824, "learning_rate": 0.00019495690687223695, "loss": 4.3888, "step": 391600 }, { "epoch": 0.7308624466117253, "grad_norm": 0.8803514838218689, "learning_rate": 0.0001949555895375187, "loss": 4.5403, "step": 391650 }, { "epoch": 0.7309557521711038, "grad_norm": 0.7681301236152649, "learning_rate": 0.00019495427203522053, "loss": 4.5776, "step": 391700 }, { "epoch": 0.7310490577304822, "grad_norm": 0.8530518412590027, "learning_rate": 0.00019495295436534473, "loss": 4.4472, "step": 391750 }, { "epoch": 0.7311423632898607, "grad_norm": 1.137397050857544, "learning_rate": 0.00019495163652789372, "loss": 4.6771, "step": 391800 }, { "epoch": 0.7312356688492392, "grad_norm": 0.6580849289894104, "learning_rate": 0.00019495031852286978, "loss": 4.477, "step": 391850 }, { "epoch": 0.7313289744086177, "grad_norm": 1.0122594833374023, "learning_rate": 0.0001949490003502752, "loss": 4.6751, "step": 391900 }, { "epoch": 0.7314222799679962, "grad_norm": 1.0346349477767944, "learning_rate": 0.00019494768201011237, "loss": 4.465, "step": 391950 }, { "epoch": 0.7315155855273747, "grad_norm": 0.912458598613739, "learning_rate": 0.0001949463635023836, "loss": 4.5299, "step": 392000 }, { "epoch": 0.7316088910867532, "grad_norm": 1.1435613632202148, "learning_rate": 0.00019494504482709117, "loss": 4.6511, "step": 392050 }, { "epoch": 0.7317021966461317, "grad_norm": 1.081440806388855, "learning_rate": 0.00019494372598423745, "loss": 4.5773, "step": 392100 }, { "epoch": 0.7317955022055102, "grad_norm": 0.9449694156646729, "learning_rate": 0.0001949424069738248, "loss": 4.4573, "step": 392150 }, { "epoch": 0.7318888077648886, "grad_norm": 1.1139907836914062, "learning_rate": 0.00019494108779585545, "loss": 4.4003, "step": 392200 }, { "epoch": 0.7319821133242671, "grad_norm": 1.0709891319274902, "learning_rate": 0.0001949397684503318, "loss": 4.7699, "step": 392250 }, { "epoch": 0.7320754188836457, "grad_norm": 1.0212353467941284, "learning_rate": 0.00019493844893725618, "loss": 4.6232, "step": 392300 }, { "epoch": 0.7321687244430242, "grad_norm": 1.0591708421707153, "learning_rate": 0.00019493712925663088, "loss": 4.6856, "step": 392350 }, { "epoch": 0.7322620300024026, "grad_norm": 1.2571759223937988, "learning_rate": 0.00019493580940845828, "loss": 4.6591, "step": 392400 }, { "epoch": 0.7323553355617811, "grad_norm": 1.0696747303009033, "learning_rate": 0.00019493448939274068, "loss": 4.5348, "step": 392450 }, { "epoch": 0.7324486411211596, "grad_norm": 1.1913007497787476, "learning_rate": 0.00019493316920948038, "loss": 4.5864, "step": 392500 }, { "epoch": 0.7325419466805381, "grad_norm": 1.1159700155258179, "learning_rate": 0.00019493184885867977, "loss": 4.5561, "step": 392550 }, { "epoch": 0.7326352522399165, "grad_norm": 1.200270652770996, "learning_rate": 0.00019493052834034113, "loss": 4.4971, "step": 392600 }, { "epoch": 0.732728557799295, "grad_norm": 0.9367539286613464, "learning_rate": 0.00019492920765446683, "loss": 4.497, "step": 392650 }, { "epoch": 0.7328218633586736, "grad_norm": 1.0923216342926025, "learning_rate": 0.00019492788680105917, "loss": 4.69, "step": 392700 }, { "epoch": 0.7329151689180521, "grad_norm": 1.0680104494094849, "learning_rate": 0.0001949265657801205, "loss": 4.5658, "step": 392750 }, { "epoch": 0.7330084744774306, "grad_norm": 1.0249441862106323, "learning_rate": 0.00019492524459165315, "loss": 4.6498, "step": 392800 }, { "epoch": 0.733101780036809, "grad_norm": 1.0899511575698853, "learning_rate": 0.0001949239232356594, "loss": 4.6632, "step": 392850 }, { "epoch": 0.7331950855961875, "grad_norm": 0.9146956205368042, "learning_rate": 0.00019492260171214166, "loss": 4.7565, "step": 392900 }, { "epoch": 0.733288391155566, "grad_norm": 1.1608656644821167, "learning_rate": 0.00019492128002110225, "loss": 4.5894, "step": 392950 }, { "epoch": 0.7333816967149445, "grad_norm": 0.9399400353431702, "learning_rate": 0.00019491995816254347, "loss": 4.4536, "step": 393000 }, { "epoch": 0.7333816967149445, "eval_loss": 4.722550868988037, "eval_runtime": 230.5993, "eval_samples_per_second": 11.31, "eval_steps_per_second": 11.31, "eval_tts_loss": 7.581275009621574, "step": 393000 }, { "epoch": 0.733475002274323, "grad_norm": 0.8586224317550659, "learning_rate": 0.00019491863613646765, "loss": 4.4582, "step": 393050 }, { "epoch": 0.7335683078337015, "grad_norm": 1.012885332107544, "learning_rate": 0.00019491731394287717, "loss": 4.6452, "step": 393100 }, { "epoch": 0.73366161339308, "grad_norm": 1.2605918645858765, "learning_rate": 0.0001949159915817743, "loss": 4.3936, "step": 393150 }, { "epoch": 0.7337549189524585, "grad_norm": 0.7984983325004578, "learning_rate": 0.00019491466905316142, "loss": 4.6295, "step": 393200 }, { "epoch": 0.733848224511837, "grad_norm": 0.8521361947059631, "learning_rate": 0.00019491334635704085, "loss": 4.6162, "step": 393250 }, { "epoch": 0.7339415300712154, "grad_norm": 1.0116236209869385, "learning_rate": 0.00019491202349341494, "loss": 4.517, "step": 393300 }, { "epoch": 0.7340348356305939, "grad_norm": 1.4281280040740967, "learning_rate": 0.00019491070046228595, "loss": 4.5894, "step": 393350 }, { "epoch": 0.7341281411899725, "grad_norm": 1.2629574537277222, "learning_rate": 0.00019490937726365632, "loss": 4.4543, "step": 393400 }, { "epoch": 0.734221446749351, "grad_norm": 1.2775226831436157, "learning_rate": 0.00019490805389752835, "loss": 4.5062, "step": 393450 }, { "epoch": 0.7343147523087294, "grad_norm": 1.2514355182647705, "learning_rate": 0.00019490673036390435, "loss": 4.7704, "step": 393500 }, { "epoch": 0.7344080578681079, "grad_norm": 1.2405158281326294, "learning_rate": 0.00019490540666278666, "loss": 4.4944, "step": 393550 }, { "epoch": 0.7345013634274864, "grad_norm": 0.9776833057403564, "learning_rate": 0.00019490408279417766, "loss": 4.7371, "step": 393600 }, { "epoch": 0.7345946689868649, "grad_norm": 1.1885216236114502, "learning_rate": 0.0001949027587580796, "loss": 4.511, "step": 393650 }, { "epoch": 0.7346879745462433, "grad_norm": 1.2549670934677124, "learning_rate": 0.0001949014345544949, "loss": 4.5568, "step": 393700 }, { "epoch": 0.7347812801056219, "grad_norm": 1.0409431457519531, "learning_rate": 0.00019490011018342585, "loss": 4.5221, "step": 393750 }, { "epoch": 0.7348745856650004, "grad_norm": 0.9980065822601318, "learning_rate": 0.00019489878564487483, "loss": 4.6895, "step": 393800 }, { "epoch": 0.7349678912243789, "grad_norm": 1.0543571710586548, "learning_rate": 0.00019489746093884414, "loss": 4.7759, "step": 393850 }, { "epoch": 0.7350611967837574, "grad_norm": 1.1471644639968872, "learning_rate": 0.00019489613606533612, "loss": 4.5118, "step": 393900 }, { "epoch": 0.7351545023431358, "grad_norm": 0.9260897040367126, "learning_rate": 0.00019489481102435315, "loss": 4.512, "step": 393950 }, { "epoch": 0.7352478079025143, "grad_norm": 0.8788979649543762, "learning_rate": 0.0001948934858158975, "loss": 4.5552, "step": 394000 }, { "epoch": 0.7353411134618928, "grad_norm": 1.2206834554672241, "learning_rate": 0.00019489216043997156, "loss": 4.4762, "step": 394050 }, { "epoch": 0.7354344190212714, "grad_norm": 0.9536528587341309, "learning_rate": 0.00019489083489657765, "loss": 4.6025, "step": 394100 }, { "epoch": 0.7355277245806499, "grad_norm": 1.0591988563537598, "learning_rate": 0.0001948895091857181, "loss": 4.6237, "step": 394150 }, { "epoch": 0.7356210301400283, "grad_norm": 0.9356663227081299, "learning_rate": 0.00019488818330739527, "loss": 4.3627, "step": 394200 }, { "epoch": 0.7357143356994068, "grad_norm": 1.0228075981140137, "learning_rate": 0.00019488685726161151, "loss": 4.5646, "step": 394250 }, { "epoch": 0.7358076412587853, "grad_norm": 0.7471048831939697, "learning_rate": 0.00019488553104836913, "loss": 4.6334, "step": 394300 }, { "epoch": 0.7359009468181638, "grad_norm": 0.9984940886497498, "learning_rate": 0.00019488420466767047, "loss": 4.4695, "step": 394350 }, { "epoch": 0.7359942523775422, "grad_norm": 1.1658512353897095, "learning_rate": 0.00019488287811951792, "loss": 4.4986, "step": 394400 }, { "epoch": 0.7360875579369208, "grad_norm": 0.700202226638794, "learning_rate": 0.00019488155140391376, "loss": 4.4986, "step": 394450 }, { "epoch": 0.7361808634962993, "grad_norm": 1.2018382549285889, "learning_rate": 0.00019488022452086035, "loss": 4.473, "step": 394500 }, { "epoch": 0.7362741690556778, "grad_norm": 1.2141975164413452, "learning_rate": 0.00019487889747036005, "loss": 4.5214, "step": 394550 }, { "epoch": 0.7363674746150563, "grad_norm": 0.9552044868469238, "learning_rate": 0.00019487757025241518, "loss": 4.6345, "step": 394600 }, { "epoch": 0.7364607801744347, "grad_norm": 0.8278501629829407, "learning_rate": 0.0001948762428670281, "loss": 4.6117, "step": 394650 }, { "epoch": 0.7365540857338132, "grad_norm": 1.293843150138855, "learning_rate": 0.00019487491531420113, "loss": 4.5266, "step": 394700 }, { "epoch": 0.7366473912931917, "grad_norm": 1.0281263589859009, "learning_rate": 0.00019487358759393664, "loss": 4.4981, "step": 394750 }, { "epoch": 0.7367406968525703, "grad_norm": 1.3931729793548584, "learning_rate": 0.000194872259706237, "loss": 4.7601, "step": 394800 }, { "epoch": 0.7368340024119487, "grad_norm": 1.1016684770584106, "learning_rate": 0.00019487093165110447, "loss": 4.5277, "step": 394850 }, { "epoch": 0.7369273079713272, "grad_norm": 0.9391690492630005, "learning_rate": 0.00019486960342854145, "loss": 4.613, "step": 394900 }, { "epoch": 0.7370206135307057, "grad_norm": 1.0593945980072021, "learning_rate": 0.00019486827503855026, "loss": 4.53, "step": 394950 }, { "epoch": 0.7371139190900842, "grad_norm": 0.8746975660324097, "learning_rate": 0.0001948669464811333, "loss": 4.5007, "step": 395000 }, { "epoch": 0.7372072246494626, "grad_norm": 0.904080867767334, "learning_rate": 0.00019486561775629283, "loss": 4.5248, "step": 395050 }, { "epoch": 0.7373005302088411, "grad_norm": 0.9073638916015625, "learning_rate": 0.00019486428886403124, "loss": 4.2864, "step": 395100 }, { "epoch": 0.7373938357682197, "grad_norm": 1.2093853950500488, "learning_rate": 0.0001948629598043509, "loss": 4.6007, "step": 395150 }, { "epoch": 0.7374871413275982, "grad_norm": 0.8730159997940063, "learning_rate": 0.00019486163057725407, "loss": 4.5062, "step": 395200 }, { "epoch": 0.7375804468869767, "grad_norm": 0.858184278011322, "learning_rate": 0.0001948603011827432, "loss": 4.5427, "step": 395250 }, { "epoch": 0.7376737524463551, "grad_norm": 1.0963269472122192, "learning_rate": 0.0001948589716208206, "loss": 4.5644, "step": 395300 }, { "epoch": 0.7377670580057336, "grad_norm": 0.8483244776725769, "learning_rate": 0.00019485764189148855, "loss": 4.6353, "step": 395350 }, { "epoch": 0.7378603635651121, "grad_norm": 0.8771704435348511, "learning_rate": 0.0001948563119947495, "loss": 4.4817, "step": 395400 }, { "epoch": 0.7379536691244906, "grad_norm": 1.1800220012664795, "learning_rate": 0.00019485498193060576, "loss": 4.4817, "step": 395450 }, { "epoch": 0.7380469746838692, "grad_norm": 1.2803168296813965, "learning_rate": 0.00019485365169905964, "loss": 4.6603, "step": 395500 }, { "epoch": 0.7381402802432476, "grad_norm": 1.1339360475540161, "learning_rate": 0.00019485232130011352, "loss": 4.5782, "step": 395550 }, { "epoch": 0.7382335858026261, "grad_norm": 0.944157063961029, "learning_rate": 0.00019485099073376976, "loss": 4.4737, "step": 395600 }, { "epoch": 0.7383268913620046, "grad_norm": 0.9531108140945435, "learning_rate": 0.00019484966000003065, "loss": 4.571, "step": 395650 }, { "epoch": 0.738420196921383, "grad_norm": 1.2628448009490967, "learning_rate": 0.00019484832909889862, "loss": 4.6907, "step": 395700 }, { "epoch": 0.7385135024807615, "grad_norm": 0.8827887773513794, "learning_rate": 0.00019484699803037596, "loss": 4.5015, "step": 395750 }, { "epoch": 0.73860680804014, "grad_norm": 1.2830276489257812, "learning_rate": 0.00019484566679446504, "loss": 4.5319, "step": 395800 }, { "epoch": 0.7387001135995186, "grad_norm": 1.3190330266952515, "learning_rate": 0.0001948443353911682, "loss": 4.7166, "step": 395850 }, { "epoch": 0.7387934191588971, "grad_norm": 1.0245732069015503, "learning_rate": 0.0001948430038204878, "loss": 4.6743, "step": 395900 }, { "epoch": 0.7388867247182755, "grad_norm": 1.0815318822860718, "learning_rate": 0.0001948416720824262, "loss": 4.4301, "step": 395950 }, { "epoch": 0.738980030277654, "grad_norm": 1.1338343620300293, "learning_rate": 0.00019484034017698573, "loss": 4.5976, "step": 396000 }, { "epoch": 0.738980030277654, "eval_loss": 4.7241644859313965, "eval_runtime": 231.6116, "eval_samples_per_second": 11.26, "eval_steps_per_second": 11.26, "eval_tts_loss": 7.610429070300017, "step": 396000 }, { "epoch": 0.7390733358370325, "grad_norm": 1.0788607597351074, "learning_rate": 0.00019483900810416874, "loss": 4.5512, "step": 396050 }, { "epoch": 0.739166641396411, "grad_norm": 1.434680700302124, "learning_rate": 0.00019483767586397758, "loss": 4.4577, "step": 396100 }, { "epoch": 0.7392599469557894, "grad_norm": 1.0275232791900635, "learning_rate": 0.0001948363434564146, "loss": 4.6689, "step": 396150 }, { "epoch": 0.739353252515168, "grad_norm": 1.1067668199539185, "learning_rate": 0.0001948350108814822, "loss": 4.6546, "step": 396200 }, { "epoch": 0.7394465580745465, "grad_norm": 1.112076759338379, "learning_rate": 0.00019483367813918266, "loss": 4.646, "step": 396250 }, { "epoch": 0.739539863633925, "grad_norm": 1.1287716627120972, "learning_rate": 0.00019483234522951837, "loss": 4.5277, "step": 396300 }, { "epoch": 0.7396331691933035, "grad_norm": 0.8934891223907471, "learning_rate": 0.0001948310121524917, "loss": 4.748, "step": 396350 }, { "epoch": 0.7397264747526819, "grad_norm": 1.2289316654205322, "learning_rate": 0.000194829678908105, "loss": 4.5759, "step": 396400 }, { "epoch": 0.7398197803120604, "grad_norm": 0.8652569651603699, "learning_rate": 0.00019482834549636055, "loss": 4.6176, "step": 396450 }, { "epoch": 0.7399130858714389, "grad_norm": 1.4067641496658325, "learning_rate": 0.00019482701191726078, "loss": 4.5168, "step": 396500 }, { "epoch": 0.7400063914308175, "grad_norm": 1.1128288507461548, "learning_rate": 0.00019482567817080798, "loss": 4.381, "step": 396550 }, { "epoch": 0.740099696990196, "grad_norm": 0.8590818643569946, "learning_rate": 0.0001948243442570046, "loss": 4.4131, "step": 396600 }, { "epoch": 0.7401930025495744, "grad_norm": 1.0324068069458008, "learning_rate": 0.00019482301017585292, "loss": 4.6837, "step": 396650 }, { "epoch": 0.7402863081089529, "grad_norm": 0.7649784684181213, "learning_rate": 0.0001948216759273553, "loss": 4.5271, "step": 396700 }, { "epoch": 0.7403796136683314, "grad_norm": 1.2677695751190186, "learning_rate": 0.00019482034151151412, "loss": 4.5032, "step": 396750 }, { "epoch": 0.7404729192277099, "grad_norm": 1.2309435606002808, "learning_rate": 0.0001948190069283317, "loss": 4.597, "step": 396800 }, { "epoch": 0.7405662247870883, "grad_norm": 1.3792515993118286, "learning_rate": 0.00019481767217781042, "loss": 4.5447, "step": 396850 }, { "epoch": 0.7406595303464669, "grad_norm": 1.1079990863800049, "learning_rate": 0.00019481633725995266, "loss": 4.5244, "step": 396900 }, { "epoch": 0.7407528359058454, "grad_norm": 1.1744472980499268, "learning_rate": 0.0001948150021747607, "loss": 4.7786, "step": 396950 }, { "epoch": 0.7408461414652239, "grad_norm": 1.1463979482650757, "learning_rate": 0.00019481366692223698, "loss": 4.5897, "step": 397000 }, { "epoch": 0.7409394470246023, "grad_norm": 1.1247175931930542, "learning_rate": 0.00019481233150238383, "loss": 4.5714, "step": 397050 }, { "epoch": 0.7410327525839808, "grad_norm": 0.952838122844696, "learning_rate": 0.00019481099591520356, "loss": 4.4864, "step": 397100 }, { "epoch": 0.7411260581433593, "grad_norm": 1.1566669940948486, "learning_rate": 0.00019480966016069862, "loss": 4.3836, "step": 397150 }, { "epoch": 0.7412193637027378, "grad_norm": 1.0998328924179077, "learning_rate": 0.00019480832423887128, "loss": 4.4351, "step": 397200 }, { "epoch": 0.7413126692621164, "grad_norm": 0.9825552701950073, "learning_rate": 0.00019480698814972392, "loss": 4.7547, "step": 397250 }, { "epoch": 0.7414059748214948, "grad_norm": 0.9544752836227417, "learning_rate": 0.00019480565189325893, "loss": 4.3914, "step": 397300 }, { "epoch": 0.7414992803808733, "grad_norm": 1.7373327016830444, "learning_rate": 0.0001948043154694786, "loss": 4.5653, "step": 397350 }, { "epoch": 0.7415925859402518, "grad_norm": 0.8258759379386902, "learning_rate": 0.00019480297887838538, "loss": 4.7044, "step": 397400 }, { "epoch": 0.7416858914996303, "grad_norm": 0.9235637187957764, "learning_rate": 0.00019480164211998157, "loss": 4.8239, "step": 397450 }, { "epoch": 0.7417791970590087, "grad_norm": 1.1709929704666138, "learning_rate": 0.00019480030519426956, "loss": 4.797, "step": 397500 }, { "epoch": 0.7418725026183872, "grad_norm": 0.8909640908241272, "learning_rate": 0.00019479896810125166, "loss": 4.5543, "step": 397550 }, { "epoch": 0.7419658081777658, "grad_norm": 0.995737612247467, "learning_rate": 0.00019479763084093027, "loss": 4.5107, "step": 397600 }, { "epoch": 0.7420591137371443, "grad_norm": 0.8248655796051025, "learning_rate": 0.00019479629341330777, "loss": 4.3587, "step": 397650 }, { "epoch": 0.7421524192965228, "grad_norm": 0.9279022812843323, "learning_rate": 0.00019479495581838644, "loss": 4.5868, "step": 397700 }, { "epoch": 0.7422457248559012, "grad_norm": 1.4893839359283447, "learning_rate": 0.00019479361805616873, "loss": 4.4135, "step": 397750 }, { "epoch": 0.7423390304152797, "grad_norm": 0.8427762389183044, "learning_rate": 0.00019479228012665698, "loss": 4.452, "step": 397800 }, { "epoch": 0.7424323359746582, "grad_norm": 1.1434813737869263, "learning_rate": 0.00019479094202985348, "loss": 4.5496, "step": 397850 }, { "epoch": 0.7425256415340367, "grad_norm": 0.9122126698493958, "learning_rate": 0.0001947896037657607, "loss": 4.5626, "step": 397900 }, { "epoch": 0.7426189470934151, "grad_norm": 1.0076179504394531, "learning_rate": 0.00019478826533438092, "loss": 4.4451, "step": 397950 }, { "epoch": 0.7427122526527937, "grad_norm": 1.1911746263504028, "learning_rate": 0.00019478692673571656, "loss": 4.6376, "step": 398000 }, { "epoch": 0.7428055582121722, "grad_norm": 0.8936728835105896, "learning_rate": 0.0001947855879697699, "loss": 4.5737, "step": 398050 }, { "epoch": 0.7428988637715507, "grad_norm": 1.344441294670105, "learning_rate": 0.0001947842490365434, "loss": 4.6513, "step": 398100 }, { "epoch": 0.7429921693309292, "grad_norm": 1.1638767719268799, "learning_rate": 0.00019478290993603937, "loss": 4.5577, "step": 398150 }, { "epoch": 0.7430854748903076, "grad_norm": 1.0375152826309204, "learning_rate": 0.00019478157066826017, "loss": 4.432, "step": 398200 }, { "epoch": 0.7431787804496861, "grad_norm": 0.875194251537323, "learning_rate": 0.00019478023123320817, "loss": 4.5255, "step": 398250 }, { "epoch": 0.7432720860090646, "grad_norm": 1.1665624380111694, "learning_rate": 0.00019477889163088574, "loss": 4.595, "step": 398300 }, { "epoch": 0.7433653915684432, "grad_norm": 1.2813527584075928, "learning_rate": 0.00019477755186129528, "loss": 4.5058, "step": 398350 }, { "epoch": 0.7434586971278216, "grad_norm": 1.1091448068618774, "learning_rate": 0.00019477621192443906, "loss": 4.2865, "step": 398400 }, { "epoch": 0.7435520026872001, "grad_norm": 1.17946457862854, "learning_rate": 0.00019477487182031954, "loss": 4.3896, "step": 398450 }, { "epoch": 0.7436453082465786, "grad_norm": 1.1359678506851196, "learning_rate": 0.00019477353154893906, "loss": 4.3753, "step": 398500 }, { "epoch": 0.7437386138059571, "grad_norm": 0.7580427527427673, "learning_rate": 0.00019477219111029993, "loss": 4.6792, "step": 398550 }, { "epoch": 0.7438319193653355, "grad_norm": 1.2027475833892822, "learning_rate": 0.00019477085050440457, "loss": 4.4967, "step": 398600 }, { "epoch": 0.743925224924714, "grad_norm": 1.2368563413619995, "learning_rate": 0.00019476950973125535, "loss": 4.6587, "step": 398650 }, { "epoch": 0.7440185304840926, "grad_norm": 1.3931561708450317, "learning_rate": 0.0001947681687908546, "loss": 4.6635, "step": 398700 }, { "epoch": 0.7441118360434711, "grad_norm": 1.1592406034469604, "learning_rate": 0.00019476682768320474, "loss": 4.4236, "step": 398750 }, { "epoch": 0.7442051416028496, "grad_norm": 1.1834975481033325, "learning_rate": 0.00019476548640830806, "loss": 4.6033, "step": 398800 }, { "epoch": 0.744298447162228, "grad_norm": 1.156468391418457, "learning_rate": 0.00019476414496616702, "loss": 4.7581, "step": 398850 }, { "epoch": 0.7443917527216065, "grad_norm": 1.147061824798584, "learning_rate": 0.00019476280335678392, "loss": 4.5531, "step": 398900 }, { "epoch": 0.744485058280985, "grad_norm": 1.1417890787124634, "learning_rate": 0.00019476146158016114, "loss": 4.767, "step": 398950 }, { "epoch": 0.7445783638403635, "grad_norm": 0.9738073348999023, "learning_rate": 0.00019476011963630104, "loss": 4.6031, "step": 399000 }, { "epoch": 0.7445783638403635, "eval_loss": 4.733906269073486, "eval_runtime": 231.4593, "eval_samples_per_second": 11.268, "eval_steps_per_second": 11.268, "eval_tts_loss": 7.55335535614557, "step": 399000 }, { "epoch": 0.744671669399742, "grad_norm": 0.9165759682655334, "learning_rate": 0.00019475877752520604, "loss": 4.6827, "step": 399050 }, { "epoch": 0.7447649749591205, "grad_norm": 1.19134521484375, "learning_rate": 0.00019475743524687843, "loss": 4.5243, "step": 399100 }, { "epoch": 0.744858280518499, "grad_norm": 1.1840969324111938, "learning_rate": 0.00019475609280132067, "loss": 4.5984, "step": 399150 }, { "epoch": 0.7449515860778775, "grad_norm": 1.0910753011703491, "learning_rate": 0.00019475475018853502, "loss": 4.4959, "step": 399200 }, { "epoch": 0.745044891637256, "grad_norm": 0.8953019380569458, "learning_rate": 0.00019475340740852395, "loss": 4.6832, "step": 399250 }, { "epoch": 0.7451381971966344, "grad_norm": 0.8007261157035828, "learning_rate": 0.00019475206446128976, "loss": 4.6535, "step": 399300 }, { "epoch": 0.7452315027560129, "grad_norm": 0.9712862968444824, "learning_rate": 0.00019475072134683488, "loss": 4.6529, "step": 399350 }, { "epoch": 0.7453248083153915, "grad_norm": 1.3298040628433228, "learning_rate": 0.0001947493780651616, "loss": 4.5217, "step": 399400 }, { "epoch": 0.74541811387477, "grad_norm": 1.211425542831421, "learning_rate": 0.00019474803461627237, "loss": 4.452, "step": 399450 }, { "epoch": 0.7455114194341484, "grad_norm": 1.0443273782730103, "learning_rate": 0.00019474669100016953, "loss": 4.5222, "step": 399500 }, { "epoch": 0.7456047249935269, "grad_norm": 1.1909388303756714, "learning_rate": 0.00019474534721685547, "loss": 4.84, "step": 399550 }, { "epoch": 0.7456980305529054, "grad_norm": 1.1435961723327637, "learning_rate": 0.00019474400326633252, "loss": 4.589, "step": 399600 }, { "epoch": 0.7457913361122839, "grad_norm": 0.9750648140907288, "learning_rate": 0.00019474265914860308, "loss": 4.4519, "step": 399650 }, { "epoch": 0.7458846416716624, "grad_norm": 1.2055881023406982, "learning_rate": 0.00019474131486366952, "loss": 4.5154, "step": 399700 }, { "epoch": 0.7459779472310409, "grad_norm": 0.9073157906532288, "learning_rate": 0.0001947399704115342, "loss": 4.5807, "step": 399750 }, { "epoch": 0.7460712527904194, "grad_norm": 1.0163545608520508, "learning_rate": 0.00019473862579219952, "loss": 4.6481, "step": 399800 }, { "epoch": 0.7461645583497979, "grad_norm": 1.2088340520858765, "learning_rate": 0.00019473728100566782, "loss": 4.5471, "step": 399850 }, { "epoch": 0.7462578639091764, "grad_norm": 0.8323642015457153, "learning_rate": 0.0001947359360519415, "loss": 4.5136, "step": 399900 }, { "epoch": 0.7463511694685548, "grad_norm": 0.7911824584007263, "learning_rate": 0.0001947345909310229, "loss": 4.6413, "step": 399950 }, { "epoch": 0.7464444750279333, "grad_norm": 1.021842122077942, "learning_rate": 0.00019473324564291445, "loss": 4.4617, "step": 400000 }, { "epoch": 0.7465377805873118, "grad_norm": 1.359620213508606, "learning_rate": 0.00019473190018761847, "loss": 4.7336, "step": 400050 }, { "epoch": 0.7466310861466904, "grad_norm": 1.1204358339309692, "learning_rate": 0.00019473055456513736, "loss": 4.5005, "step": 400100 }, { "epoch": 0.7467243917060689, "grad_norm": 1.1570340394973755, "learning_rate": 0.00019472920877547348, "loss": 4.6138, "step": 400150 }, { "epoch": 0.7468176972654473, "grad_norm": 1.142628788948059, "learning_rate": 0.00019472786281862921, "loss": 4.3871, "step": 400200 }, { "epoch": 0.7469110028248258, "grad_norm": 1.0545388460159302, "learning_rate": 0.00019472651669460693, "loss": 4.594, "step": 400250 }, { "epoch": 0.7470043083842043, "grad_norm": 1.013261079788208, "learning_rate": 0.00019472517040340902, "loss": 4.7622, "step": 400300 }, { "epoch": 0.7470976139435828, "grad_norm": 1.195189118385315, "learning_rate": 0.00019472382394503787, "loss": 4.6786, "step": 400350 }, { "epoch": 0.7471909195029612, "grad_norm": 1.234176516532898, "learning_rate": 0.00019472247731949581, "loss": 4.5959, "step": 400400 }, { "epoch": 0.7472842250623398, "grad_norm": 0.8004019260406494, "learning_rate": 0.00019472113052678528, "loss": 4.6037, "step": 400450 }, { "epoch": 0.7473775306217183, "grad_norm": 0.8816819190979004, "learning_rate": 0.0001947197835669086, "loss": 4.5606, "step": 400500 }, { "epoch": 0.7474708361810968, "grad_norm": 0.9571092128753662, "learning_rate": 0.0001947184364398682, "loss": 4.4627, "step": 400550 }, { "epoch": 0.7475641417404753, "grad_norm": 0.8714603781700134, "learning_rate": 0.00019471708914566637, "loss": 4.621, "step": 400600 }, { "epoch": 0.7476574472998537, "grad_norm": 1.0243371725082397, "learning_rate": 0.0001947157416843056, "loss": 4.4409, "step": 400650 }, { "epoch": 0.7477507528592322, "grad_norm": 0.9247996807098389, "learning_rate": 0.00019471439405578816, "loss": 4.5011, "step": 400700 }, { "epoch": 0.7478440584186107, "grad_norm": 1.2378648519515991, "learning_rate": 0.0001947130462601165, "loss": 4.6031, "step": 400750 }, { "epoch": 0.7479373639779893, "grad_norm": 1.0939531326293945, "learning_rate": 0.00019471169829729295, "loss": 4.8021, "step": 400800 }, { "epoch": 0.7480306695373677, "grad_norm": 1.0560319423675537, "learning_rate": 0.00019471035016731998, "loss": 4.4943, "step": 400850 }, { "epoch": 0.7481239750967462, "grad_norm": 0.8283106684684753, "learning_rate": 0.00019470900187019987, "loss": 4.4933, "step": 400900 }, { "epoch": 0.7482172806561247, "grad_norm": 0.7455849051475525, "learning_rate": 0.00019470765340593503, "loss": 4.613, "step": 400950 }, { "epoch": 0.7483105862155032, "grad_norm": 1.328385829925537, "learning_rate": 0.00019470630477452786, "loss": 4.4819, "step": 401000 }, { "epoch": 0.7484038917748816, "grad_norm": 1.1238863468170166, "learning_rate": 0.00019470495597598072, "loss": 4.5988, "step": 401050 }, { "epoch": 0.7484971973342601, "grad_norm": 1.2375730276107788, "learning_rate": 0.00019470360701029601, "loss": 4.6046, "step": 401100 }, { "epoch": 0.7485905028936387, "grad_norm": 0.9606476426124573, "learning_rate": 0.00019470225787747607, "loss": 4.6423, "step": 401150 }, { "epoch": 0.7486838084530172, "grad_norm": 1.2989537715911865, "learning_rate": 0.00019470090857752332, "loss": 4.5824, "step": 401200 }, { "epoch": 0.7487771140123957, "grad_norm": 1.0816088914871216, "learning_rate": 0.00019469955911044016, "loss": 4.5863, "step": 401250 }, { "epoch": 0.7488704195717741, "grad_norm": 0.8996260762214661, "learning_rate": 0.0001946982094762289, "loss": 4.399, "step": 401300 }, { "epoch": 0.7489637251311526, "grad_norm": 0.9724602103233337, "learning_rate": 0.00019469685967489196, "loss": 4.6596, "step": 401350 }, { "epoch": 0.7490570306905311, "grad_norm": 1.0567646026611328, "learning_rate": 0.00019469550970643176, "loss": 4.5988, "step": 401400 }, { "epoch": 0.7491503362499096, "grad_norm": 1.2678072452545166, "learning_rate": 0.00019469415957085065, "loss": 4.4416, "step": 401450 }, { "epoch": 0.7492436418092882, "grad_norm": 0.6493592262268066, "learning_rate": 0.000194692809268151, "loss": 4.5179, "step": 401500 }, { "epoch": 0.7493369473686666, "grad_norm": 1.1661994457244873, "learning_rate": 0.0001946914587983352, "loss": 4.4027, "step": 401550 }, { "epoch": 0.7494302529280451, "grad_norm": 1.3520020246505737, "learning_rate": 0.0001946901081614056, "loss": 4.4424, "step": 401600 }, { "epoch": 0.7495235584874236, "grad_norm": 1.0197575092315674, "learning_rate": 0.00019468875735736467, "loss": 4.4455, "step": 401650 }, { "epoch": 0.7496168640468021, "grad_norm": 1.0030943155288696, "learning_rate": 0.0001946874063862147, "loss": 4.584, "step": 401700 }, { "epoch": 0.7497101696061805, "grad_norm": 1.0498521327972412, "learning_rate": 0.00019468605524795816, "loss": 4.6093, "step": 401750 }, { "epoch": 0.749803475165559, "grad_norm": 1.0407713651657104, "learning_rate": 0.00019468470394259736, "loss": 4.6358, "step": 401800 }, { "epoch": 0.7498967807249376, "grad_norm": 1.0960047245025635, "learning_rate": 0.00019468335247013475, "loss": 4.3322, "step": 401850 }, { "epoch": 0.7499900862843161, "grad_norm": 1.2660539150238037, "learning_rate": 0.00019468200083057268, "loss": 4.5874, "step": 401900 }, { "epoch": 0.7500833918436945, "grad_norm": 1.0143240690231323, "learning_rate": 0.00019468064902391352, "loss": 4.5931, "step": 401950 }, { "epoch": 0.750176697403073, "grad_norm": 1.3088220357894897, "learning_rate": 0.00019467929705015968, "loss": 4.4499, "step": 402000 }, { "epoch": 0.750176697403073, "eval_loss": 4.724297523498535, "eval_runtime": 230.2641, "eval_samples_per_second": 11.326, "eval_steps_per_second": 11.326, "eval_tts_loss": 7.560537891713845, "step": 402000 }, { "epoch": 0.7502700029624515, "grad_norm": 1.0270438194274902, "learning_rate": 0.00019467794490931356, "loss": 4.4617, "step": 402050 }, { "epoch": 0.75036330852183, "grad_norm": 1.3051048517227173, "learning_rate": 0.00019467659260137747, "loss": 4.4863, "step": 402100 }, { "epoch": 0.7504566140812085, "grad_norm": 0.9769117832183838, "learning_rate": 0.0001946752401263539, "loss": 4.656, "step": 402150 }, { "epoch": 0.750549919640587, "grad_norm": 1.2200247049331665, "learning_rate": 0.0001946738874842452, "loss": 4.4802, "step": 402200 }, { "epoch": 0.7506432251999655, "grad_norm": 0.6177666187286377, "learning_rate": 0.0001946725346750537, "loss": 4.658, "step": 402250 }, { "epoch": 0.750736530759344, "grad_norm": 0.9986745119094849, "learning_rate": 0.00019467118169878185, "loss": 4.4032, "step": 402300 }, { "epoch": 0.7508298363187225, "grad_norm": 1.0253387689590454, "learning_rate": 0.00019466982855543204, "loss": 4.5996, "step": 402350 }, { "epoch": 0.7509231418781009, "grad_norm": 0.9981570839881897, "learning_rate": 0.00019466847524500665, "loss": 4.5612, "step": 402400 }, { "epoch": 0.7510164474374794, "grad_norm": 1.055335283279419, "learning_rate": 0.00019466712176750806, "loss": 4.5569, "step": 402450 }, { "epoch": 0.7511097529968579, "grad_norm": 1.0270355939865112, "learning_rate": 0.0001946657681229386, "loss": 4.6409, "step": 402500 }, { "epoch": 0.7512030585562365, "grad_norm": 0.9899023175239563, "learning_rate": 0.00019466441431130078, "loss": 4.4688, "step": 402550 }, { "epoch": 0.751296364115615, "grad_norm": 1.2617480754852295, "learning_rate": 0.00019466306033259689, "loss": 4.5089, "step": 402600 }, { "epoch": 0.7513896696749934, "grad_norm": 0.9508602619171143, "learning_rate": 0.00019466170618682936, "loss": 4.8497, "step": 402650 }, { "epoch": 0.7514829752343719, "grad_norm": 1.1826248168945312, "learning_rate": 0.00019466035187400057, "loss": 4.517, "step": 402700 }, { "epoch": 0.7515762807937504, "grad_norm": 0.763058066368103, "learning_rate": 0.00019465899739411294, "loss": 4.4908, "step": 402750 }, { "epoch": 0.7516695863531289, "grad_norm": 1.045179843902588, "learning_rate": 0.00019465764274716882, "loss": 4.4248, "step": 402800 }, { "epoch": 0.7517628919125073, "grad_norm": 0.8762446641921997, "learning_rate": 0.00019465628793317063, "loss": 4.538, "step": 402850 }, { "epoch": 0.7518561974718859, "grad_norm": 0.9932625889778137, "learning_rate": 0.00019465493295212075, "loss": 4.6375, "step": 402900 }, { "epoch": 0.7519495030312644, "grad_norm": 1.0016560554504395, "learning_rate": 0.00019465357780402154, "loss": 4.3052, "step": 402950 }, { "epoch": 0.7520428085906429, "grad_norm": 1.1759898662567139, "learning_rate": 0.00019465222248887543, "loss": 4.4555, "step": 403000 }, { "epoch": 0.7521361141500214, "grad_norm": 0.9538491368293762, "learning_rate": 0.0001946508670066848, "loss": 4.4514, "step": 403050 }, { "epoch": 0.7522294197093998, "grad_norm": 1.1780041456222534, "learning_rate": 0.00019464951135745207, "loss": 4.6608, "step": 403100 }, { "epoch": 0.7523227252687783, "grad_norm": 1.1293666362762451, "learning_rate": 0.00019464815554117956, "loss": 4.425, "step": 403150 }, { "epoch": 0.7524160308281568, "grad_norm": 1.1419936418533325, "learning_rate": 0.00019464679955786976, "loss": 4.7019, "step": 403200 }, { "epoch": 0.7525093363875354, "grad_norm": 1.1940276622772217, "learning_rate": 0.00019464544340752496, "loss": 4.3547, "step": 403250 }, { "epoch": 0.7526026419469138, "grad_norm": 0.8929638862609863, "learning_rate": 0.00019464408709014762, "loss": 4.6242, "step": 403300 }, { "epoch": 0.7526959475062923, "grad_norm": 1.017694354057312, "learning_rate": 0.00019464273060574016, "loss": 4.6063, "step": 403350 }, { "epoch": 0.7527892530656708, "grad_norm": 1.4753926992416382, "learning_rate": 0.0001946413739543049, "loss": 4.4256, "step": 403400 }, { "epoch": 0.7528825586250493, "grad_norm": 0.9293118119239807, "learning_rate": 0.00019464001713584428, "loss": 4.5837, "step": 403450 }, { "epoch": 0.7529758641844277, "grad_norm": 0.9962372779846191, "learning_rate": 0.00019463866015036066, "loss": 4.6717, "step": 403500 }, { "epoch": 0.7530691697438062, "grad_norm": 1.1656463146209717, "learning_rate": 0.0001946373029978565, "loss": 4.6189, "step": 403550 }, { "epoch": 0.7531624753031847, "grad_norm": 1.0788663625717163, "learning_rate": 0.0001946359456783341, "loss": 4.3787, "step": 403600 }, { "epoch": 0.7532557808625633, "grad_norm": 1.0301463603973389, "learning_rate": 0.00019463458819179595, "loss": 4.4878, "step": 403650 }, { "epoch": 0.7533490864219418, "grad_norm": 0.9593393802642822, "learning_rate": 0.00019463323053824438, "loss": 4.4218, "step": 403700 }, { "epoch": 0.7534423919813202, "grad_norm": 0.7830232977867126, "learning_rate": 0.00019463187271768183, "loss": 4.4766, "step": 403750 }, { "epoch": 0.7535356975406987, "grad_norm": 0.786916971206665, "learning_rate": 0.00019463051473011064, "loss": 4.4523, "step": 403800 }, { "epoch": 0.7536290031000772, "grad_norm": 1.4050339460372925, "learning_rate": 0.00019462915657553326, "loss": 4.5962, "step": 403850 }, { "epoch": 0.7537223086594557, "grad_norm": 0.8983107805252075, "learning_rate": 0.00019462779825395207, "loss": 4.4027, "step": 403900 }, { "epoch": 0.7538156142188341, "grad_norm": 1.0345656871795654, "learning_rate": 0.00019462643976536948, "loss": 4.4476, "step": 403950 }, { "epoch": 0.7539089197782127, "grad_norm": 1.4160012006759644, "learning_rate": 0.00019462508110978786, "loss": 4.4757, "step": 404000 }, { "epoch": 0.7540022253375912, "grad_norm": 1.0370961427688599, "learning_rate": 0.0001946237222872096, "loss": 4.6786, "step": 404050 }, { "epoch": 0.7540955308969697, "grad_norm": 0.8926733732223511, "learning_rate": 0.00019462236329763718, "loss": 4.4315, "step": 404100 }, { "epoch": 0.7541888364563482, "grad_norm": 1.3053297996520996, "learning_rate": 0.0001946210041410729, "loss": 4.6435, "step": 404150 }, { "epoch": 0.7542821420157266, "grad_norm": 1.331666350364685, "learning_rate": 0.00019461964481751916, "loss": 4.4299, "step": 404200 }, { "epoch": 0.7543754475751051, "grad_norm": 1.0818222761154175, "learning_rate": 0.00019461828532697846, "loss": 4.6642, "step": 404250 }, { "epoch": 0.7544687531344836, "grad_norm": 1.265306830406189, "learning_rate": 0.00019461692566945308, "loss": 4.5585, "step": 404300 }, { "epoch": 0.7545620586938622, "grad_norm": 1.1073338985443115, "learning_rate": 0.0001946155658449455, "loss": 4.6652, "step": 404350 }, { "epoch": 0.7546553642532406, "grad_norm": 1.1014400720596313, "learning_rate": 0.00019461420585345813, "loss": 4.4495, "step": 404400 }, { "epoch": 0.7547486698126191, "grad_norm": 1.0301647186279297, "learning_rate": 0.0001946128456949933, "loss": 4.6083, "step": 404450 }, { "epoch": 0.7548419753719976, "grad_norm": 1.091760277748108, "learning_rate": 0.00019461148536955344, "loss": 4.4661, "step": 404500 }, { "epoch": 0.7549352809313761, "grad_norm": 1.064077615737915, "learning_rate": 0.00019461012487714097, "loss": 4.4569, "step": 404550 }, { "epoch": 0.7550285864907545, "grad_norm": 1.3203434944152832, "learning_rate": 0.00019460876421775825, "loss": 4.6581, "step": 404600 }, { "epoch": 0.755121892050133, "grad_norm": 0.9822664856910706, "learning_rate": 0.00019460740339140775, "loss": 4.5907, "step": 404650 }, { "epoch": 0.7552151976095116, "grad_norm": 0.9781966209411621, "learning_rate": 0.00019460604239809182, "loss": 4.5604, "step": 404700 }, { "epoch": 0.7553085031688901, "grad_norm": 1.252949595451355, "learning_rate": 0.00019460468123781285, "loss": 4.4489, "step": 404750 }, { "epoch": 0.7554018087282686, "grad_norm": 0.8946507573127747, "learning_rate": 0.00019460331991057327, "loss": 4.6669, "step": 404800 }, { "epoch": 0.755495114287647, "grad_norm": 1.1925349235534668, "learning_rate": 0.00019460195841637548, "loss": 4.6114, "step": 404850 }, { "epoch": 0.7555884198470255, "grad_norm": 0.9882096648216248, "learning_rate": 0.00019460059675522188, "loss": 4.6776, "step": 404900 }, { "epoch": 0.755681725406404, "grad_norm": 1.3011587858200073, "learning_rate": 0.00019459923492711487, "loss": 4.5147, "step": 404950 }, { "epoch": 0.7557750309657825, "grad_norm": 1.0943289995193481, "learning_rate": 0.00019459787293205685, "loss": 4.6341, "step": 405000 }, { "epoch": 0.7557750309657825, "eval_loss": 4.72869348526001, "eval_runtime": 232.6461, "eval_samples_per_second": 11.21, "eval_steps_per_second": 11.21, "eval_tts_loss": 7.544485853397876, "step": 405000 }, { "epoch": 0.755868336525161, "grad_norm": 1.1097865104675293, "learning_rate": 0.00019459651077005023, "loss": 4.4529, "step": 405050 }, { "epoch": 0.7559616420845395, "grad_norm": 1.0926594734191895, "learning_rate": 0.00019459514844109743, "loss": 4.4716, "step": 405100 }, { "epoch": 0.756054947643918, "grad_norm": 1.0199013948440552, "learning_rate": 0.0001945937859452008, "loss": 4.6114, "step": 405150 }, { "epoch": 0.7561482532032965, "grad_norm": 1.1694191694259644, "learning_rate": 0.00019459242328236282, "loss": 4.3158, "step": 405200 }, { "epoch": 0.756241558762675, "grad_norm": 1.148744821548462, "learning_rate": 0.00019459106045258587, "loss": 4.5701, "step": 405250 }, { "epoch": 0.7563348643220534, "grad_norm": 0.6980789303779602, "learning_rate": 0.0001945896974558723, "loss": 4.5219, "step": 405300 }, { "epoch": 0.7564281698814319, "grad_norm": 1.2195786237716675, "learning_rate": 0.00019458833429222459, "loss": 4.3908, "step": 405350 }, { "epoch": 0.7565214754408105, "grad_norm": 1.0867304801940918, "learning_rate": 0.00019458697096164508, "loss": 4.4848, "step": 405400 }, { "epoch": 0.756614781000189, "grad_norm": 1.2966946363449097, "learning_rate": 0.00019458560746413622, "loss": 4.6711, "step": 405450 }, { "epoch": 0.7567080865595674, "grad_norm": 1.1345088481903076, "learning_rate": 0.00019458424379970043, "loss": 4.941, "step": 405500 }, { "epoch": 0.7568013921189459, "grad_norm": 1.2740004062652588, "learning_rate": 0.00019458287996834009, "loss": 4.666, "step": 405550 }, { "epoch": 0.7568946976783244, "grad_norm": 0.915218710899353, "learning_rate": 0.00019458151597005757, "loss": 4.3748, "step": 405600 }, { "epoch": 0.7569880032377029, "grad_norm": 1.0948460102081299, "learning_rate": 0.00019458015180485536, "loss": 4.595, "step": 405650 }, { "epoch": 0.7570813087970814, "grad_norm": 1.0221774578094482, "learning_rate": 0.0001945787874727358, "loss": 4.3731, "step": 405700 }, { "epoch": 0.7571746143564599, "grad_norm": 1.1602940559387207, "learning_rate": 0.00019457742297370132, "loss": 4.5798, "step": 405750 }, { "epoch": 0.7572679199158384, "grad_norm": 1.1894081830978394, "learning_rate": 0.00019457605830775437, "loss": 4.5092, "step": 405800 }, { "epoch": 0.7573612254752169, "grad_norm": 1.1323769092559814, "learning_rate": 0.00019457469347489728, "loss": 4.3593, "step": 405850 }, { "epoch": 0.7574545310345954, "grad_norm": 1.1201635599136353, "learning_rate": 0.00019457332847513248, "loss": 4.6032, "step": 405900 }, { "epoch": 0.7575478365939738, "grad_norm": 0.9292293190956116, "learning_rate": 0.00019457196330846242, "loss": 4.4532, "step": 405950 }, { "epoch": 0.7576411421533523, "grad_norm": 1.1106488704681396, "learning_rate": 0.00019457059797488947, "loss": 4.5502, "step": 406000 }, { "epoch": 0.7577344477127308, "grad_norm": 1.143852710723877, "learning_rate": 0.0001945692324744161, "loss": 4.5175, "step": 406050 }, { "epoch": 0.7578277532721094, "grad_norm": 0.8539165258407593, "learning_rate": 0.0001945678668070446, "loss": 4.5666, "step": 406100 }, { "epoch": 0.7579210588314879, "grad_norm": 1.088355541229248, "learning_rate": 0.0001945665009727775, "loss": 4.4709, "step": 406150 }, { "epoch": 0.7580143643908663, "grad_norm": 1.0392314195632935, "learning_rate": 0.0001945651349716172, "loss": 4.5914, "step": 406200 }, { "epoch": 0.7581076699502448, "grad_norm": 0.8739420175552368, "learning_rate": 0.000194563768803566, "loss": 4.5308, "step": 406250 }, { "epoch": 0.7582009755096233, "grad_norm": 1.187957525253296, "learning_rate": 0.00019456240246862644, "loss": 4.6386, "step": 406300 }, { "epoch": 0.7582942810690018, "grad_norm": 0.9893680810928345, "learning_rate": 0.00019456103596680088, "loss": 4.3799, "step": 406350 }, { "epoch": 0.7583875866283802, "grad_norm": 1.0255956649780273, "learning_rate": 0.0001945596692980917, "loss": 4.7057, "step": 406400 }, { "epoch": 0.7584808921877588, "grad_norm": 1.1427487134933472, "learning_rate": 0.00019455830246250136, "loss": 4.5165, "step": 406450 }, { "epoch": 0.7585741977471373, "grad_norm": 1.0257514715194702, "learning_rate": 0.00019455693546003227, "loss": 4.6241, "step": 406500 }, { "epoch": 0.7586675033065158, "grad_norm": 1.1918649673461914, "learning_rate": 0.0001945555682906868, "loss": 4.6769, "step": 406550 }, { "epoch": 0.7587608088658943, "grad_norm": 0.8059489727020264, "learning_rate": 0.0001945542009544674, "loss": 4.5509, "step": 406600 }, { "epoch": 0.7588541144252727, "grad_norm": 1.0606913566589355, "learning_rate": 0.00019455283345137648, "loss": 4.4686, "step": 406650 }, { "epoch": 0.7589474199846512, "grad_norm": 1.049088954925537, "learning_rate": 0.00019455146578141642, "loss": 4.5742, "step": 406700 }, { "epoch": 0.7590407255440297, "grad_norm": 1.4002255201339722, "learning_rate": 0.00019455009794458968, "loss": 4.6296, "step": 406750 }, { "epoch": 0.7591340311034083, "grad_norm": 1.4870983362197876, "learning_rate": 0.00019454872994089867, "loss": 4.567, "step": 406800 }, { "epoch": 0.7592273366627867, "grad_norm": 0.9239751696586609, "learning_rate": 0.00019454736177034574, "loss": 4.4658, "step": 406850 }, { "epoch": 0.7593206422221652, "grad_norm": 1.0361567735671997, "learning_rate": 0.00019454599343293337, "loss": 4.4316, "step": 406900 }, { "epoch": 0.7594139477815437, "grad_norm": 0.99744713306427, "learning_rate": 0.000194544624928664, "loss": 4.6061, "step": 406950 }, { "epoch": 0.7595072533409222, "grad_norm": 0.8480738997459412, "learning_rate": 0.00019454325625753997, "loss": 4.2468, "step": 407000 }, { "epoch": 0.7596005589003006, "grad_norm": 1.0402275323867798, "learning_rate": 0.0001945418874195637, "loss": 4.6353, "step": 407050 }, { "epoch": 0.7596938644596791, "grad_norm": 1.025551676750183, "learning_rate": 0.00019454051841473767, "loss": 4.6685, "step": 407100 }, { "epoch": 0.7597871700190577, "grad_norm": 1.0114656686782837, "learning_rate": 0.00019453914924306425, "loss": 4.7492, "step": 407150 }, { "epoch": 0.7598804755784362, "grad_norm": 1.142307162284851, "learning_rate": 0.00019453777990454586, "loss": 4.7868, "step": 407200 }, { "epoch": 0.7599737811378147, "grad_norm": 1.160451889038086, "learning_rate": 0.00019453641039918494, "loss": 4.7449, "step": 407250 }, { "epoch": 0.7600670866971931, "grad_norm": 1.1598868370056152, "learning_rate": 0.00019453504072698387, "loss": 4.6213, "step": 407300 }, { "epoch": 0.7601603922565716, "grad_norm": 0.8196390867233276, "learning_rate": 0.0001945336708879451, "loss": 4.4612, "step": 407350 }, { "epoch": 0.7602536978159501, "grad_norm": 1.2065792083740234, "learning_rate": 0.00019453230088207105, "loss": 4.5206, "step": 407400 }, { "epoch": 0.7603470033753286, "grad_norm": 1.200914740562439, "learning_rate": 0.00019453093070936407, "loss": 4.574, "step": 407450 }, { "epoch": 0.7604403089347072, "grad_norm": 1.0380594730377197, "learning_rate": 0.0001945295603698267, "loss": 4.5967, "step": 407500 }, { "epoch": 0.7605336144940856, "grad_norm": 0.948436439037323, "learning_rate": 0.00019452818986346123, "loss": 4.4912, "step": 407550 }, { "epoch": 0.7606269200534641, "grad_norm": 1.1930091381072998, "learning_rate": 0.00019452681919027018, "loss": 4.2002, "step": 407600 }, { "epoch": 0.7607202256128426, "grad_norm": 0.9641384482383728, "learning_rate": 0.0001945254483502559, "loss": 4.5361, "step": 407650 }, { "epoch": 0.7608135311722211, "grad_norm": 0.9800957441329956, "learning_rate": 0.00019452407734342085, "loss": 4.7418, "step": 407700 }, { "epoch": 0.7609068367315995, "grad_norm": 0.9638832807540894, "learning_rate": 0.00019452270616976742, "loss": 4.6648, "step": 407750 }, { "epoch": 0.761000142290978, "grad_norm": 0.9217967987060547, "learning_rate": 0.00019452133482929807, "loss": 4.4051, "step": 407800 }, { "epoch": 0.7610934478503566, "grad_norm": 0.972507894039154, "learning_rate": 0.00019451996332201516, "loss": 4.5565, "step": 407850 }, { "epoch": 0.7611867534097351, "grad_norm": 1.100412130355835, "learning_rate": 0.0001945185916479212, "loss": 4.6013, "step": 407900 }, { "epoch": 0.7612800589691135, "grad_norm": 1.2424074411392212, "learning_rate": 0.00019451721980701851, "loss": 4.5051, "step": 407950 }, { "epoch": 0.761373364528492, "grad_norm": 0.6012529134750366, "learning_rate": 0.00019451584779930955, "loss": 4.5495, "step": 408000 }, { "epoch": 0.761373364528492, "eval_loss": 4.730047225952148, "eval_runtime": 231.8619, "eval_samples_per_second": 11.248, "eval_steps_per_second": 11.248, "eval_tts_loss": 7.581340771529858, "step": 408000 }, { "epoch": 0.7614666700878705, "grad_norm": 1.240382432937622, "learning_rate": 0.00019451447562479678, "loss": 4.44, "step": 408050 }, { "epoch": 0.761559975647249, "grad_norm": 0.8447127938270569, "learning_rate": 0.0001945131032834826, "loss": 4.5198, "step": 408100 }, { "epoch": 0.7616532812066275, "grad_norm": 1.0848270654678345, "learning_rate": 0.00019451173077536938, "loss": 4.6223, "step": 408150 }, { "epoch": 0.761746586766006, "grad_norm": 1.1647281646728516, "learning_rate": 0.0001945103581004596, "loss": 4.4722, "step": 408200 }, { "epoch": 0.7618398923253845, "grad_norm": 1.1008877754211426, "learning_rate": 0.0001945089852587557, "loss": 4.5822, "step": 408250 }, { "epoch": 0.761933197884763, "grad_norm": 1.0027493238449097, "learning_rate": 0.00019450761225026005, "loss": 4.5399, "step": 408300 }, { "epoch": 0.7620265034441415, "grad_norm": 1.3000314235687256, "learning_rate": 0.0001945062390749751, "loss": 4.5339, "step": 408350 }, { "epoch": 0.7621198090035199, "grad_norm": 1.1075400114059448, "learning_rate": 0.00019450486573290325, "loss": 4.2773, "step": 408400 }, { "epoch": 0.7622131145628984, "grad_norm": 1.1882368326187134, "learning_rate": 0.00019450349222404695, "loss": 4.5993, "step": 408450 }, { "epoch": 0.7623064201222769, "grad_norm": 0.9934725761413574, "learning_rate": 0.0001945021185484086, "loss": 4.6568, "step": 408500 }, { "epoch": 0.7623997256816555, "grad_norm": 0.9820505976676941, "learning_rate": 0.00019450074470599068, "loss": 4.4453, "step": 408550 }, { "epoch": 0.762493031241034, "grad_norm": 1.356992244720459, "learning_rate": 0.00019449937069679555, "loss": 4.5526, "step": 408600 }, { "epoch": 0.7625863368004124, "grad_norm": 0.980800986289978, "learning_rate": 0.00019449799652082562, "loss": 4.497, "step": 408650 }, { "epoch": 0.7626796423597909, "grad_norm": 1.1779906749725342, "learning_rate": 0.0001944966221780834, "loss": 4.5632, "step": 408700 }, { "epoch": 0.7627729479191694, "grad_norm": 0.7027290463447571, "learning_rate": 0.00019449524766857127, "loss": 4.5865, "step": 408750 }, { "epoch": 0.7628662534785479, "grad_norm": 0.9588069915771484, "learning_rate": 0.00019449387299229164, "loss": 4.6278, "step": 408800 }, { "epoch": 0.7629595590379263, "grad_norm": 1.077543020248413, "learning_rate": 0.00019449249814924694, "loss": 4.5214, "step": 408850 }, { "epoch": 0.7630528645973048, "grad_norm": 0.7556289434432983, "learning_rate": 0.00019449112313943962, "loss": 4.3555, "step": 408900 }, { "epoch": 0.7631461701566834, "grad_norm": 1.1389154195785522, "learning_rate": 0.0001944897479628721, "loss": 4.6045, "step": 408950 }, { "epoch": 0.7632394757160619, "grad_norm": 1.0199394226074219, "learning_rate": 0.00019448837261954678, "loss": 4.3561, "step": 409000 }, { "epoch": 0.7633327812754404, "grad_norm": 1.3064322471618652, "learning_rate": 0.00019448699710946613, "loss": 4.4714, "step": 409050 }, { "epoch": 0.7634260868348188, "grad_norm": 1.4018571376800537, "learning_rate": 0.00019448562143263256, "loss": 4.4632, "step": 409100 }, { "epoch": 0.7635193923941973, "grad_norm": 1.2147985696792603, "learning_rate": 0.00019448424558904848, "loss": 4.8225, "step": 409150 }, { "epoch": 0.7636126979535758, "grad_norm": 1.0787280797958374, "learning_rate": 0.00019448286957871632, "loss": 4.4554, "step": 409200 }, { "epoch": 0.7637060035129543, "grad_norm": 1.1411551237106323, "learning_rate": 0.0001944814934016385, "loss": 4.4555, "step": 409250 }, { "epoch": 0.7637993090723328, "grad_norm": 0.8726068735122681, "learning_rate": 0.00019448011705781754, "loss": 4.648, "step": 409300 }, { "epoch": 0.7638926146317113, "grad_norm": 0.9773291945457458, "learning_rate": 0.00019447874054725578, "loss": 4.6669, "step": 409350 }, { "epoch": 0.7639859201910898, "grad_norm": 0.9928725361824036, "learning_rate": 0.00019447736386995563, "loss": 4.7582, "step": 409400 }, { "epoch": 0.7640792257504683, "grad_norm": 1.1772263050079346, "learning_rate": 0.00019447598702591957, "loss": 4.7123, "step": 409450 }, { "epoch": 0.7641725313098467, "grad_norm": 1.0281718969345093, "learning_rate": 0.00019447461001515003, "loss": 4.5314, "step": 409500 }, { "epoch": 0.7642658368692252, "grad_norm": 1.2277331352233887, "learning_rate": 0.0001944732328376494, "loss": 4.6218, "step": 409550 }, { "epoch": 0.7643591424286037, "grad_norm": 0.8721914887428284, "learning_rate": 0.00019447185549342018, "loss": 4.5136, "step": 409600 }, { "epoch": 0.7644524479879823, "grad_norm": 0.7213169932365417, "learning_rate": 0.00019447047798246472, "loss": 4.5493, "step": 409650 }, { "epoch": 0.7645457535473608, "grad_norm": 1.10368812084198, "learning_rate": 0.00019446910030478548, "loss": 4.6382, "step": 409700 }, { "epoch": 0.7646390591067392, "grad_norm": 1.156548261642456, "learning_rate": 0.00019446772246038493, "loss": 4.8191, "step": 409750 }, { "epoch": 0.7647323646661177, "grad_norm": 1.1590478420257568, "learning_rate": 0.00019446634444926546, "loss": 4.7934, "step": 409800 }, { "epoch": 0.7648256702254962, "grad_norm": 0.8049784302711487, "learning_rate": 0.00019446496627142953, "loss": 4.5612, "step": 409850 }, { "epoch": 0.7649189757848747, "grad_norm": 1.069915771484375, "learning_rate": 0.00019446358792687954, "loss": 4.5778, "step": 409900 }, { "epoch": 0.7650122813442531, "grad_norm": 1.0776913166046143, "learning_rate": 0.0001944622094156179, "loss": 4.7124, "step": 409950 }, { "epoch": 0.7651055869036317, "grad_norm": 1.3020472526550293, "learning_rate": 0.00019446083073764716, "loss": 4.4051, "step": 410000 }, { "epoch": 0.7651988924630102, "grad_norm": 1.0878071784973145, "learning_rate": 0.00019445945189296965, "loss": 4.6188, "step": 410050 }, { "epoch": 0.7652921980223887, "grad_norm": 1.1448755264282227, "learning_rate": 0.00019445807288158778, "loss": 4.6531, "step": 410100 }, { "epoch": 0.7653855035817672, "grad_norm": 0.8183618783950806, "learning_rate": 0.0001944566937035041, "loss": 4.751, "step": 410150 }, { "epoch": 0.7654788091411456, "grad_norm": 0.969052791595459, "learning_rate": 0.00019445531435872092, "loss": 4.5996, "step": 410200 }, { "epoch": 0.7655721147005241, "grad_norm": 0.8166422247886658, "learning_rate": 0.00019445393484724077, "loss": 4.5969, "step": 410250 }, { "epoch": 0.7656654202599026, "grad_norm": 0.8773173093795776, "learning_rate": 0.000194452555169066, "loss": 4.5966, "step": 410300 }, { "epoch": 0.7657587258192812, "grad_norm": 1.2333333492279053, "learning_rate": 0.00019445117532419914, "loss": 4.5009, "step": 410350 }, { "epoch": 0.7658520313786596, "grad_norm": 1.1725988388061523, "learning_rate": 0.00019444979531264256, "loss": 4.6075, "step": 410400 }, { "epoch": 0.7659453369380381, "grad_norm": 1.0543071031570435, "learning_rate": 0.00019444841513439868, "loss": 4.6056, "step": 410450 }, { "epoch": 0.7660386424974166, "grad_norm": 1.2511577606201172, "learning_rate": 0.00019444703478947, "loss": 4.5412, "step": 410500 }, { "epoch": 0.7661319480567951, "grad_norm": 3.9236738681793213, "learning_rate": 0.0001944456542778589, "loss": 4.5994, "step": 410550 }, { "epoch": 0.7662252536161736, "grad_norm": 1.1855190992355347, "learning_rate": 0.00019444427359956788, "loss": 4.4763, "step": 410600 }, { "epoch": 0.766318559175552, "grad_norm": 0.8772841095924377, "learning_rate": 0.00019444289275459927, "loss": 4.7665, "step": 410650 }, { "epoch": 0.7664118647349306, "grad_norm": 0.7112861275672913, "learning_rate": 0.0001944415117429556, "loss": 4.5654, "step": 410700 }, { "epoch": 0.7665051702943091, "grad_norm": 1.1377826929092407, "learning_rate": 0.0001944401305646393, "loss": 4.5007, "step": 410750 }, { "epoch": 0.7665984758536876, "grad_norm": 0.8896861672401428, "learning_rate": 0.00019443874921965277, "loss": 4.5797, "step": 410800 }, { "epoch": 0.766691781413066, "grad_norm": 1.1979306936264038, "learning_rate": 0.00019443736770799849, "loss": 4.4931, "step": 410850 }, { "epoch": 0.7667850869724445, "grad_norm": 1.0301620960235596, "learning_rate": 0.00019443598602967882, "loss": 4.4661, "step": 410900 }, { "epoch": 0.766878392531823, "grad_norm": 1.393868088722229, "learning_rate": 0.00019443460418469628, "loss": 4.5384, "step": 410950 }, { "epoch": 0.7669716980912015, "grad_norm": 1.014648199081421, "learning_rate": 0.0001944332221730533, "loss": 4.4657, "step": 411000 }, { "epoch": 0.7669716980912015, "eval_loss": 4.729247093200684, "eval_runtime": 234.8612, "eval_samples_per_second": 11.104, "eval_steps_per_second": 11.104, "eval_tts_loss": 7.555574542404284, "step": 411000 }, { "epoch": 0.7670650036505801, "grad_norm": 1.115856647491455, "learning_rate": 0.00019443183999475229, "loss": 4.71, "step": 411050 }, { "epoch": 0.7671583092099585, "grad_norm": 0.8463857769966125, "learning_rate": 0.00019443045764979568, "loss": 4.4989, "step": 411100 }, { "epoch": 0.767251614769337, "grad_norm": 0.9952225089073181, "learning_rate": 0.0001944290751381859, "loss": 4.7664, "step": 411150 }, { "epoch": 0.7673449203287155, "grad_norm": 0.9273759126663208, "learning_rate": 0.00019442769245992547, "loss": 4.602, "step": 411200 }, { "epoch": 0.767438225888094, "grad_norm": 1.1283340454101562, "learning_rate": 0.00019442630961501677, "loss": 4.5756, "step": 411250 }, { "epoch": 0.7675315314474724, "grad_norm": 1.040757417678833, "learning_rate": 0.00019442492660346223, "loss": 4.5112, "step": 411300 }, { "epoch": 0.7676248370068509, "grad_norm": 1.1037769317626953, "learning_rate": 0.00019442354342526433, "loss": 4.6146, "step": 411350 }, { "epoch": 0.7677181425662295, "grad_norm": 1.0502253770828247, "learning_rate": 0.00019442216008042545, "loss": 4.6156, "step": 411400 }, { "epoch": 0.767811448125608, "grad_norm": 0.9768388271331787, "learning_rate": 0.00019442077656894812, "loss": 4.4807, "step": 411450 }, { "epoch": 0.7679047536849865, "grad_norm": 1.213956356048584, "learning_rate": 0.0001944193928908347, "loss": 4.4356, "step": 411500 }, { "epoch": 0.7679980592443649, "grad_norm": 1.2873953580856323, "learning_rate": 0.00019441800904608769, "loss": 4.5072, "step": 411550 }, { "epoch": 0.7680913648037434, "grad_norm": 0.7170881628990173, "learning_rate": 0.0001944166250347095, "loss": 4.4458, "step": 411600 }, { "epoch": 0.7681846703631219, "grad_norm": 0.972938060760498, "learning_rate": 0.00019441524085670258, "loss": 4.5461, "step": 411650 }, { "epoch": 0.7682779759225004, "grad_norm": 1.1991459131240845, "learning_rate": 0.00019441385651206936, "loss": 4.4705, "step": 411700 }, { "epoch": 0.7683712814818789, "grad_norm": 1.1323275566101074, "learning_rate": 0.0001944124720008123, "loss": 4.6942, "step": 411750 }, { "epoch": 0.7684645870412574, "grad_norm": 1.1658388376235962, "learning_rate": 0.00019441108732293384, "loss": 4.7014, "step": 411800 }, { "epoch": 0.7685578926006359, "grad_norm": 0.9996901750564575, "learning_rate": 0.00019440970247843644, "loss": 4.5796, "step": 411850 }, { "epoch": 0.7686511981600144, "grad_norm": 0.9950489401817322, "learning_rate": 0.0001944083174673225, "loss": 4.6371, "step": 411900 }, { "epoch": 0.7687445037193928, "grad_norm": 0.9656080007553101, "learning_rate": 0.00019440693228959453, "loss": 4.6424, "step": 411950 }, { "epoch": 0.7688378092787713, "grad_norm": 1.0694754123687744, "learning_rate": 0.0001944055469452549, "loss": 4.4645, "step": 412000 }, { "epoch": 0.7689311148381498, "grad_norm": 1.3352670669555664, "learning_rate": 0.0001944041614343061, "loss": 4.5704, "step": 412050 }, { "epoch": 0.7690244203975284, "grad_norm": 1.2701009511947632, "learning_rate": 0.00019440277575675057, "loss": 4.7115, "step": 412100 }, { "epoch": 0.7691177259569069, "grad_norm": 1.1434177160263062, "learning_rate": 0.00019440138991259076, "loss": 4.6429, "step": 412150 }, { "epoch": 0.7692110315162853, "grad_norm": 0.9898224472999573, "learning_rate": 0.0001944000039018291, "loss": 4.4657, "step": 412200 }, { "epoch": 0.7693043370756638, "grad_norm": 1.0095897912979126, "learning_rate": 0.00019439861772446802, "loss": 4.3326, "step": 412250 }, { "epoch": 0.7693976426350423, "grad_norm": 1.093815565109253, "learning_rate": 0.00019439723138051004, "loss": 4.5972, "step": 412300 }, { "epoch": 0.7694909481944208, "grad_norm": 1.092368483543396, "learning_rate": 0.0001943958448699575, "loss": 4.5492, "step": 412350 }, { "epoch": 0.7695842537537992, "grad_norm": 1.0995482206344604, "learning_rate": 0.00019439445819281292, "loss": 4.5641, "step": 412400 }, { "epoch": 0.7696775593131778, "grad_norm": 1.3554096221923828, "learning_rate": 0.00019439307134907872, "loss": 4.415, "step": 412450 }, { "epoch": 0.7697708648725563, "grad_norm": 1.2155762910842896, "learning_rate": 0.0001943916843387574, "loss": 4.4498, "step": 412500 }, { "epoch": 0.7698641704319348, "grad_norm": 1.114269733428955, "learning_rate": 0.00019439029716185135, "loss": 4.5365, "step": 412550 }, { "epoch": 0.7699574759913133, "grad_norm": 1.1325422525405884, "learning_rate": 0.00019438890981836302, "loss": 4.5969, "step": 412600 }, { "epoch": 0.7700507815506917, "grad_norm": 1.136763095855713, "learning_rate": 0.0001943875223082949, "loss": 4.6564, "step": 412650 }, { "epoch": 0.7701440871100702, "grad_norm": 1.1705725193023682, "learning_rate": 0.0001943861346316494, "loss": 4.3569, "step": 412700 }, { "epoch": 0.7702373926694487, "grad_norm": 1.3166948556900024, "learning_rate": 0.00019438474678842895, "loss": 4.2721, "step": 412750 }, { "epoch": 0.7703306982288273, "grad_norm": 1.262305736541748, "learning_rate": 0.00019438335877863604, "loss": 4.6105, "step": 412800 }, { "epoch": 0.7704240037882057, "grad_norm": 1.311882734298706, "learning_rate": 0.00019438197060227312, "loss": 4.5957, "step": 412850 }, { "epoch": 0.7705173093475842, "grad_norm": 1.2204020023345947, "learning_rate": 0.00019438058225934265, "loss": 4.5673, "step": 412900 }, { "epoch": 0.7706106149069627, "grad_norm": 1.0517381429672241, "learning_rate": 0.000194379193749847, "loss": 4.6357, "step": 412950 }, { "epoch": 0.7707039204663412, "grad_norm": 1.0536929368972778, "learning_rate": 0.00019437780507378872, "loss": 4.4254, "step": 413000 }, { "epoch": 0.7707972260257197, "grad_norm": 1.018850326538086, "learning_rate": 0.00019437641623117023, "loss": 4.5578, "step": 413050 }, { "epoch": 0.7708905315850981, "grad_norm": 1.2499680519104004, "learning_rate": 0.00019437502722199397, "loss": 4.6343, "step": 413100 }, { "epoch": 0.7709838371444767, "grad_norm": 1.065850853919983, "learning_rate": 0.00019437363804626235, "loss": 4.5045, "step": 413150 }, { "epoch": 0.7710771427038552, "grad_norm": 1.150632619857788, "learning_rate": 0.00019437224870397793, "loss": 4.613, "step": 413200 }, { "epoch": 0.7711704482632337, "grad_norm": 0.9182499647140503, "learning_rate": 0.00019437085919514306, "loss": 4.5223, "step": 413250 }, { "epoch": 0.7712637538226121, "grad_norm": 1.01868736743927, "learning_rate": 0.00019436946951976024, "loss": 4.6569, "step": 413300 }, { "epoch": 0.7713570593819906, "grad_norm": 1.1553199291229248, "learning_rate": 0.0001943680796778319, "loss": 4.7115, "step": 413350 }, { "epoch": 0.7714503649413691, "grad_norm": 1.235852599143982, "learning_rate": 0.00019436668966936048, "loss": 4.5456, "step": 413400 }, { "epoch": 0.7715436705007476, "grad_norm": 1.1578147411346436, "learning_rate": 0.00019436529949434848, "loss": 4.5546, "step": 413450 }, { "epoch": 0.7716369760601262, "grad_norm": 1.095115303993225, "learning_rate": 0.00019436390915279837, "loss": 4.6062, "step": 413500 }, { "epoch": 0.7717302816195046, "grad_norm": 0.9559109807014465, "learning_rate": 0.0001943625186447125, "loss": 4.5222, "step": 413550 }, { "epoch": 0.7718235871788831, "grad_norm": 0.964905321598053, "learning_rate": 0.00019436112797009342, "loss": 4.7261, "step": 413600 }, { "epoch": 0.7719168927382616, "grad_norm": 1.1534135341644287, "learning_rate": 0.00019435973712894356, "loss": 4.575, "step": 413650 }, { "epoch": 0.7720101982976401, "grad_norm": 0.8084390163421631, "learning_rate": 0.00019435834612126535, "loss": 4.6461, "step": 413700 }, { "epoch": 0.7721035038570185, "grad_norm": 0.9302859306335449, "learning_rate": 0.00019435695494706128, "loss": 4.5341, "step": 413750 }, { "epoch": 0.772196809416397, "grad_norm": 0.7920522689819336, "learning_rate": 0.00019435556360633375, "loss": 4.3171, "step": 413800 }, { "epoch": 0.7722901149757756, "grad_norm": 1.0716264247894287, "learning_rate": 0.0001943541720990853, "loss": 4.7698, "step": 413850 }, { "epoch": 0.7723834205351541, "grad_norm": 0.8790009021759033, "learning_rate": 0.0001943527804253183, "loss": 4.5874, "step": 413900 }, { "epoch": 0.7724767260945326, "grad_norm": 1.0554299354553223, "learning_rate": 0.00019435138858503527, "loss": 4.4623, "step": 413950 }, { "epoch": 0.772570031653911, "grad_norm": 0.9873991012573242, "learning_rate": 0.00019434999657823863, "loss": 4.4922, "step": 414000 }, { "epoch": 0.772570031653911, "eval_loss": 4.728475093841553, "eval_runtime": 234.8475, "eval_samples_per_second": 11.105, "eval_steps_per_second": 11.105, "eval_tts_loss": 7.568875565229304, "step": 414000 }, { "epoch": 0.7726633372132895, "grad_norm": 1.185647964477539, "learning_rate": 0.00019434860440493084, "loss": 4.5529, "step": 414050 }, { "epoch": 0.772756642772668, "grad_norm": 1.039781093597412, "learning_rate": 0.00019434721206511437, "loss": 4.3436, "step": 414100 }, { "epoch": 0.7728499483320465, "grad_norm": 1.0758379697799683, "learning_rate": 0.00019434581955879166, "loss": 4.652, "step": 414150 }, { "epoch": 0.7729432538914249, "grad_norm": 1.0461156368255615, "learning_rate": 0.0001943444268859652, "loss": 4.6459, "step": 414200 }, { "epoch": 0.7730365594508035, "grad_norm": 1.0226956605911255, "learning_rate": 0.00019434303404663745, "loss": 4.3713, "step": 414250 }, { "epoch": 0.773129865010182, "grad_norm": 0.9251855611801147, "learning_rate": 0.0001943416410408108, "loss": 4.6817, "step": 414300 }, { "epoch": 0.7732231705695605, "grad_norm": 0.9334062337875366, "learning_rate": 0.00019434024786848777, "loss": 4.4471, "step": 414350 }, { "epoch": 0.773316476128939, "grad_norm": 1.0316689014434814, "learning_rate": 0.0001943388545296708, "loss": 4.5474, "step": 414400 }, { "epoch": 0.7734097816883174, "grad_norm": 1.1120091676712036, "learning_rate": 0.00019433746102436237, "loss": 4.699, "step": 414450 }, { "epoch": 0.7735030872476959, "grad_norm": 0.9110586047172546, "learning_rate": 0.00019433606735256488, "loss": 4.5102, "step": 414500 }, { "epoch": 0.7735963928070744, "grad_norm": 1.002780795097351, "learning_rate": 0.00019433467351428086, "loss": 4.6266, "step": 414550 }, { "epoch": 0.773689698366453, "grad_norm": 1.3112727403640747, "learning_rate": 0.00019433327950951274, "loss": 4.6851, "step": 414600 }, { "epoch": 0.7737830039258314, "grad_norm": 1.1229636669158936, "learning_rate": 0.00019433188533826297, "loss": 4.6857, "step": 414650 }, { "epoch": 0.7738763094852099, "grad_norm": 0.9074994325637817, "learning_rate": 0.00019433049100053403, "loss": 4.656, "step": 414700 }, { "epoch": 0.7739696150445884, "grad_norm": 0.9077115654945374, "learning_rate": 0.00019432909649632835, "loss": 4.3477, "step": 414750 }, { "epoch": 0.7740629206039669, "grad_norm": 0.8067662119865417, "learning_rate": 0.00019432770182564843, "loss": 4.5333, "step": 414800 }, { "epoch": 0.7741562261633453, "grad_norm": 1.110912799835205, "learning_rate": 0.00019432630698849673, "loss": 4.6428, "step": 414850 }, { "epoch": 0.7742495317227238, "grad_norm": 1.0513705015182495, "learning_rate": 0.00019432491198487567, "loss": 4.5769, "step": 414900 }, { "epoch": 0.7743428372821024, "grad_norm": 1.2311010360717773, "learning_rate": 0.00019432351681478773, "loss": 4.7393, "step": 414950 }, { "epoch": 0.7744361428414809, "grad_norm": 1.2063549757003784, "learning_rate": 0.0001943221214782354, "loss": 4.3179, "step": 415000 }, { "epoch": 0.7745294484008594, "grad_norm": 1.0355026721954346, "learning_rate": 0.0001943207259752211, "loss": 4.4265, "step": 415050 }, { "epoch": 0.7746227539602378, "grad_norm": 0.9668757915496826, "learning_rate": 0.00019431933030574733, "loss": 4.5367, "step": 415100 }, { "epoch": 0.7747160595196163, "grad_norm": 1.165673017501831, "learning_rate": 0.00019431793446981655, "loss": 4.5728, "step": 415150 }, { "epoch": 0.7748093650789948, "grad_norm": 0.811471700668335, "learning_rate": 0.00019431653846743122, "loss": 4.5724, "step": 415200 }, { "epoch": 0.7749026706383733, "grad_norm": 1.1002155542373657, "learning_rate": 0.00019431514229859375, "loss": 4.4974, "step": 415250 }, { "epoch": 0.7749959761977518, "grad_norm": 0.9775769710540771, "learning_rate": 0.00019431374596330666, "loss": 4.135, "step": 415300 }, { "epoch": 0.7750892817571303, "grad_norm": 0.9746343493461609, "learning_rate": 0.00019431234946157245, "loss": 4.503, "step": 415350 }, { "epoch": 0.7751825873165088, "grad_norm": 1.0700674057006836, "learning_rate": 0.00019431095279339347, "loss": 4.6995, "step": 415400 }, { "epoch": 0.7752758928758873, "grad_norm": 1.1561682224273682, "learning_rate": 0.0001943095559587723, "loss": 4.4893, "step": 415450 }, { "epoch": 0.7753691984352657, "grad_norm": 0.9776704907417297, "learning_rate": 0.00019430815895771136, "loss": 4.368, "step": 415500 }, { "epoch": 0.7754625039946442, "grad_norm": 1.591834306716919, "learning_rate": 0.0001943067617902131, "loss": 4.7457, "step": 415550 }, { "epoch": 0.7755558095540227, "grad_norm": 1.0882806777954102, "learning_rate": 0.00019430536445628, "loss": 4.618, "step": 415600 }, { "epoch": 0.7756491151134013, "grad_norm": 0.9813977479934692, "learning_rate": 0.00019430396695591454, "loss": 4.6303, "step": 415650 }, { "epoch": 0.7757424206727798, "grad_norm": 1.1258888244628906, "learning_rate": 0.0001943025692891191, "loss": 4.4641, "step": 415700 }, { "epoch": 0.7758357262321582, "grad_norm": 1.2065447568893433, "learning_rate": 0.0001943011714558963, "loss": 4.573, "step": 415750 }, { "epoch": 0.7759290317915367, "grad_norm": 0.7935813665390015, "learning_rate": 0.0001942997734562485, "loss": 4.7439, "step": 415800 }, { "epoch": 0.7760223373509152, "grad_norm": 0.7973695993423462, "learning_rate": 0.00019429837529017817, "loss": 4.5464, "step": 415850 }, { "epoch": 0.7761156429102937, "grad_norm": 1.237538456916809, "learning_rate": 0.00019429697695768784, "loss": 4.5519, "step": 415900 }, { "epoch": 0.7762089484696721, "grad_norm": 1.0046755075454712, "learning_rate": 0.00019429557845877994, "loss": 4.584, "step": 415950 }, { "epoch": 0.7763022540290507, "grad_norm": 1.007865071296692, "learning_rate": 0.0001942941797934569, "loss": 4.5836, "step": 416000 }, { "epoch": 0.7763955595884292, "grad_norm": 1.0284510850906372, "learning_rate": 0.00019429278096172125, "loss": 4.5441, "step": 416050 }, { "epoch": 0.7764888651478077, "grad_norm": 1.0060741901397705, "learning_rate": 0.0001942913819635754, "loss": 4.4357, "step": 416100 }, { "epoch": 0.7765821707071862, "grad_norm": 1.2664929628372192, "learning_rate": 0.00019428998279902188, "loss": 4.3837, "step": 416150 }, { "epoch": 0.7766754762665646, "grad_norm": 1.3339954614639282, "learning_rate": 0.00019428858346806312, "loss": 4.5083, "step": 416200 }, { "epoch": 0.7767687818259431, "grad_norm": 1.8308089971542358, "learning_rate": 0.00019428718397070162, "loss": 4.6836, "step": 416250 }, { "epoch": 0.7768620873853216, "grad_norm": 1.1787583827972412, "learning_rate": 0.0001942857843069398, "loss": 4.4676, "step": 416300 }, { "epoch": 0.7769553929447002, "grad_norm": 1.283090591430664, "learning_rate": 0.0001942843844767802, "loss": 4.5517, "step": 416350 }, { "epoch": 0.7770486985040786, "grad_norm": 0.8927239179611206, "learning_rate": 0.00019428298448022523, "loss": 4.7133, "step": 416400 }, { "epoch": 0.7771420040634571, "grad_norm": 1.2088508605957031, "learning_rate": 0.00019428158431727738, "loss": 4.5483, "step": 416450 }, { "epoch": 0.7772353096228356, "grad_norm": 0.8650713562965393, "learning_rate": 0.00019428018398793913, "loss": 4.6464, "step": 416500 }, { "epoch": 0.7773286151822141, "grad_norm": 0.8104282021522522, "learning_rate": 0.00019427878349221293, "loss": 4.2484, "step": 416550 }, { "epoch": 0.7774219207415926, "grad_norm": 1.0021339654922485, "learning_rate": 0.00019427738283010128, "loss": 4.5643, "step": 416600 }, { "epoch": 0.777515226300971, "grad_norm": 1.0103039741516113, "learning_rate": 0.00019427598200160662, "loss": 4.4514, "step": 416650 }, { "epoch": 0.7776085318603496, "grad_norm": 1.0550456047058105, "learning_rate": 0.00019427458100673147, "loss": 4.4182, "step": 416700 }, { "epoch": 0.7777018374197281, "grad_norm": 0.9796375036239624, "learning_rate": 0.00019427317984547823, "loss": 4.7182, "step": 416750 }, { "epoch": 0.7777951429791066, "grad_norm": 0.9060117602348328, "learning_rate": 0.00019427177851784943, "loss": 4.7177, "step": 416800 }, { "epoch": 0.777888448538485, "grad_norm": 1.0749459266662598, "learning_rate": 0.00019427037702384755, "loss": 4.6499, "step": 416850 }, { "epoch": 0.7779817540978635, "grad_norm": 1.1595515012741089, "learning_rate": 0.00019426897536347504, "loss": 4.6817, "step": 416900 }, { "epoch": 0.778075059657242, "grad_norm": 1.2265338897705078, "learning_rate": 0.00019426757353673436, "loss": 4.6216, "step": 416950 }, { "epoch": 0.7781683652166205, "grad_norm": 0.9206924438476562, "learning_rate": 0.000194266171543628, "loss": 4.4135, "step": 417000 }, { "epoch": 0.7781683652166205, "eval_loss": 4.7275614738464355, "eval_runtime": 233.8432, "eval_samples_per_second": 11.153, "eval_steps_per_second": 11.153, "eval_tts_loss": 7.560000432109563, "step": 417000 }, { "epoch": 0.7782616707759991, "grad_norm": 1.1508795022964478, "learning_rate": 0.00019426476938415844, "loss": 4.4384, "step": 417050 }, { "epoch": 0.7783549763353775, "grad_norm": 1.052635669708252, "learning_rate": 0.00019426336705832817, "loss": 4.5719, "step": 417100 }, { "epoch": 0.778448281894756, "grad_norm": 1.231458306312561, "learning_rate": 0.0001942619645661396, "loss": 4.7393, "step": 417150 }, { "epoch": 0.7785415874541345, "grad_norm": 0.6703341603279114, "learning_rate": 0.00019426056190759527, "loss": 4.3872, "step": 417200 }, { "epoch": 0.778634893013513, "grad_norm": 1.0658252239227295, "learning_rate": 0.00019425915908269766, "loss": 4.4999, "step": 417250 }, { "epoch": 0.7787281985728914, "grad_norm": 1.0017874240875244, "learning_rate": 0.0001942577560914492, "loss": 4.766, "step": 417300 }, { "epoch": 0.7788215041322699, "grad_norm": 1.4141888618469238, "learning_rate": 0.00019425635293385236, "loss": 4.4379, "step": 417350 }, { "epoch": 0.7789148096916485, "grad_norm": 1.4482828378677368, "learning_rate": 0.00019425494960990965, "loss": 4.5795, "step": 417400 }, { "epoch": 0.779008115251027, "grad_norm": 1.072346568107605, "learning_rate": 0.00019425354611962353, "loss": 4.4124, "step": 417450 }, { "epoch": 0.7791014208104055, "grad_norm": 1.0275367498397827, "learning_rate": 0.00019425214246299651, "loss": 4.6065, "step": 417500 }, { "epoch": 0.7791947263697839, "grad_norm": 1.1461256742477417, "learning_rate": 0.00019425073864003106, "loss": 4.5195, "step": 417550 }, { "epoch": 0.7792880319291624, "grad_norm": 1.0955123901367188, "learning_rate": 0.00019424933465072962, "loss": 4.6588, "step": 417600 }, { "epoch": 0.7793813374885409, "grad_norm": 1.1902095079421997, "learning_rate": 0.0001942479304950947, "loss": 4.5148, "step": 417650 }, { "epoch": 0.7794746430479194, "grad_norm": 1.45046865940094, "learning_rate": 0.00019424652617312872, "loss": 4.676, "step": 417700 }, { "epoch": 0.779567948607298, "grad_norm": 1.0053950548171997, "learning_rate": 0.00019424512168483425, "loss": 4.5951, "step": 417750 }, { "epoch": 0.7796612541666764, "grad_norm": 1.1858372688293457, "learning_rate": 0.00019424371703021372, "loss": 4.4666, "step": 417800 }, { "epoch": 0.7797545597260549, "grad_norm": 1.2483307123184204, "learning_rate": 0.00019424231220926957, "loss": 4.4132, "step": 417850 }, { "epoch": 0.7798478652854334, "grad_norm": 1.159818410873413, "learning_rate": 0.0001942409072220044, "loss": 4.296, "step": 417900 }, { "epoch": 0.7799411708448118, "grad_norm": 0.9254298806190491, "learning_rate": 0.00019423950206842055, "loss": 4.5937, "step": 417950 }, { "epoch": 0.7800344764041903, "grad_norm": 0.9275083541870117, "learning_rate": 0.00019423809674852058, "loss": 4.5276, "step": 418000 }, { "epoch": 0.7801277819635688, "grad_norm": 0.9060862064361572, "learning_rate": 0.00019423669126230695, "loss": 4.6078, "step": 418050 }, { "epoch": 0.7802210875229474, "grad_norm": 0.9902940988540649, "learning_rate": 0.00019423528560978212, "loss": 4.6819, "step": 418100 }, { "epoch": 0.7803143930823259, "grad_norm": 0.9466397166252136, "learning_rate": 0.0001942338797909486, "loss": 4.578, "step": 418150 }, { "epoch": 0.7804076986417043, "grad_norm": 1.061043381690979, "learning_rate": 0.0001942324738058089, "loss": 4.6206, "step": 418200 }, { "epoch": 0.7805010042010828, "grad_norm": 0.9925414323806763, "learning_rate": 0.00019423106765436543, "loss": 4.4128, "step": 418250 }, { "epoch": 0.7805943097604613, "grad_norm": 1.138055443763733, "learning_rate": 0.0001942296613366207, "loss": 4.5836, "step": 418300 }, { "epoch": 0.7806876153198398, "grad_norm": 0.9431651830673218, "learning_rate": 0.0001942282548525772, "loss": 4.488, "step": 418350 }, { "epoch": 0.7807809208792182, "grad_norm": 1.1069378852844238, "learning_rate": 0.00019422684820223743, "loss": 4.3909, "step": 418400 }, { "epoch": 0.7808742264385968, "grad_norm": 1.473007321357727, "learning_rate": 0.00019422544138560387, "loss": 4.7102, "step": 418450 }, { "epoch": 0.7809675319979753, "grad_norm": 1.063603162765503, "learning_rate": 0.00019422403440267892, "loss": 4.4908, "step": 418500 }, { "epoch": 0.7810608375573538, "grad_norm": 0.9246348142623901, "learning_rate": 0.0001942226272534652, "loss": 4.5294, "step": 418550 }, { "epoch": 0.7811541431167323, "grad_norm": 1.083772897720337, "learning_rate": 0.00019422121993796507, "loss": 4.6143, "step": 418600 }, { "epoch": 0.7812474486761107, "grad_norm": 1.3056730031967163, "learning_rate": 0.0001942198124561811, "loss": 4.6124, "step": 418650 }, { "epoch": 0.7813407542354892, "grad_norm": 0.9770277142524719, "learning_rate": 0.00019421840480811573, "loss": 4.3734, "step": 418700 }, { "epoch": 0.7814340597948677, "grad_norm": 1.2718496322631836, "learning_rate": 0.00019421699699377146, "loss": 4.6823, "step": 418750 }, { "epoch": 0.7815273653542463, "grad_norm": 0.8844640254974365, "learning_rate": 0.00019421558901315072, "loss": 4.5507, "step": 418800 }, { "epoch": 0.7816206709136247, "grad_norm": 0.9227458834648132, "learning_rate": 0.0001942141808662561, "loss": 4.3661, "step": 418850 }, { "epoch": 0.7817139764730032, "grad_norm": 1.2249149084091187, "learning_rate": 0.00019421277255309, "loss": 4.4293, "step": 418900 }, { "epoch": 0.7818072820323817, "grad_norm": 1.1062428951263428, "learning_rate": 0.00019421136407365497, "loss": 4.6696, "step": 418950 }, { "epoch": 0.7819005875917602, "grad_norm": 0.9474209547042847, "learning_rate": 0.00019420995542795342, "loss": 4.5205, "step": 419000 }, { "epoch": 0.7819938931511387, "grad_norm": 1.0487412214279175, "learning_rate": 0.00019420854661598787, "loss": 4.5216, "step": 419050 }, { "epoch": 0.7820871987105171, "grad_norm": 0.9993675947189331, "learning_rate": 0.00019420713763776084, "loss": 4.6082, "step": 419100 }, { "epoch": 0.7821805042698957, "grad_norm": 0.8808112740516663, "learning_rate": 0.0001942057284932748, "loss": 4.505, "step": 419150 }, { "epoch": 0.7822738098292742, "grad_norm": 1.1803745031356812, "learning_rate": 0.0001942043191825322, "loss": 4.5235, "step": 419200 }, { "epoch": 0.7823671153886527, "grad_norm": 0.852111279964447, "learning_rate": 0.00019420290970553552, "loss": 4.6029, "step": 419250 }, { "epoch": 0.7824604209480311, "grad_norm": 1.1742595434188843, "learning_rate": 0.00019420150006228733, "loss": 4.6342, "step": 419300 }, { "epoch": 0.7825537265074096, "grad_norm": 0.799828290939331, "learning_rate": 0.00019420009025279005, "loss": 4.6569, "step": 419350 }, { "epoch": 0.7826470320667881, "grad_norm": 1.0880378484725952, "learning_rate": 0.00019419868027704617, "loss": 4.4393, "step": 419400 }, { "epoch": 0.7827403376261666, "grad_norm": 1.1590566635131836, "learning_rate": 0.00019419727013505824, "loss": 4.6774, "step": 419450 }, { "epoch": 0.782833643185545, "grad_norm": 0.9538164138793945, "learning_rate": 0.00019419585982682867, "loss": 4.5779, "step": 419500 }, { "epoch": 0.7829269487449236, "grad_norm": 1.166024923324585, "learning_rate": 0.00019419444935235997, "loss": 4.6071, "step": 419550 }, { "epoch": 0.7830202543043021, "grad_norm": 1.2705340385437012, "learning_rate": 0.00019419303871165465, "loss": 4.5342, "step": 419600 }, { "epoch": 0.7831135598636806, "grad_norm": 1.2403308153152466, "learning_rate": 0.0001941916279047152, "loss": 4.4299, "step": 419650 }, { "epoch": 0.7832068654230591, "grad_norm": 0.6722421050071716, "learning_rate": 0.0001941902169315441, "loss": 4.7047, "step": 419700 }, { "epoch": 0.7833001709824375, "grad_norm": 1.151554822921753, "learning_rate": 0.00019418880579214384, "loss": 4.6488, "step": 419750 }, { "epoch": 0.783393476541816, "grad_norm": 0.9347752928733826, "learning_rate": 0.00019418739448651693, "loss": 4.3575, "step": 419800 }, { "epoch": 0.7834867821011945, "grad_norm": 1.110811471939087, "learning_rate": 0.00019418598301466583, "loss": 4.5186, "step": 419850 }, { "epoch": 0.7835800876605731, "grad_norm": 1.1932270526885986, "learning_rate": 0.00019418457137659304, "loss": 4.6109, "step": 419900 }, { "epoch": 0.7836733932199516, "grad_norm": 1.063226342201233, "learning_rate": 0.00019418315957230102, "loss": 4.4918, "step": 419950 }, { "epoch": 0.78376669877933, "grad_norm": 1.0015634298324585, "learning_rate": 0.00019418174760179232, "loss": 4.5527, "step": 420000 }, { "epoch": 0.78376669877933, "eval_loss": 4.721020221710205, "eval_runtime": 233.7819, "eval_samples_per_second": 11.156, "eval_steps_per_second": 11.156, "eval_tts_loss": 7.572293719996704, "step": 420000 }, { "epoch": 0.7838600043387085, "grad_norm": 1.0639441013336182, "learning_rate": 0.00019418033546506942, "loss": 4.7157, "step": 420050 }, { "epoch": 0.783953309898087, "grad_norm": 1.0074838399887085, "learning_rate": 0.0001941789231621348, "loss": 4.7882, "step": 420100 }, { "epoch": 0.7840466154574655, "grad_norm": 0.9637343883514404, "learning_rate": 0.00019417751069299093, "loss": 4.5151, "step": 420150 }, { "epoch": 0.7841399210168439, "grad_norm": 0.9664762020111084, "learning_rate": 0.00019417609805764033, "loss": 4.7028, "step": 420200 }, { "epoch": 0.7842332265762225, "grad_norm": 1.0026178359985352, "learning_rate": 0.00019417468525608552, "loss": 4.5529, "step": 420250 }, { "epoch": 0.784326532135601, "grad_norm": 0.8280647993087769, "learning_rate": 0.00019417327228832891, "loss": 4.6032, "step": 420300 }, { "epoch": 0.7844198376949795, "grad_norm": 1.0198395252227783, "learning_rate": 0.0001941718591543731, "loss": 4.5357, "step": 420350 }, { "epoch": 0.784513143254358, "grad_norm": 1.0759913921356201, "learning_rate": 0.00019417044585422048, "loss": 4.6143, "step": 420400 }, { "epoch": 0.7846064488137364, "grad_norm": 1.1430230140686035, "learning_rate": 0.0001941690323878736, "loss": 4.6088, "step": 420450 }, { "epoch": 0.7846997543731149, "grad_norm": 1.0650012493133545, "learning_rate": 0.00019416761875533495, "loss": 4.6748, "step": 420500 }, { "epoch": 0.7847930599324934, "grad_norm": 0.8512048721313477, "learning_rate": 0.00019416620495660704, "loss": 4.5298, "step": 420550 }, { "epoch": 0.784886365491872, "grad_norm": 1.2103396654129028, "learning_rate": 0.00019416479099169235, "loss": 4.6461, "step": 420600 }, { "epoch": 0.7849796710512504, "grad_norm": 0.8439053893089294, "learning_rate": 0.00019416337686059337, "loss": 4.4296, "step": 420650 }, { "epoch": 0.7850729766106289, "grad_norm": 1.0167698860168457, "learning_rate": 0.00019416196256331257, "loss": 4.6308, "step": 420700 }, { "epoch": 0.7851662821700074, "grad_norm": 0.9046244025230408, "learning_rate": 0.0001941605480998525, "loss": 4.4991, "step": 420750 }, { "epoch": 0.7852595877293859, "grad_norm": 0.9066252708435059, "learning_rate": 0.0001941591334702156, "loss": 4.6585, "step": 420800 }, { "epoch": 0.7853528932887643, "grad_norm": 1.1947590112686157, "learning_rate": 0.00019415771867440444, "loss": 4.6536, "step": 420850 }, { "epoch": 0.7854461988481428, "grad_norm": 0.8964353799819946, "learning_rate": 0.00019415630371242144, "loss": 4.5808, "step": 420900 }, { "epoch": 0.7855395044075214, "grad_norm": 1.4293516874313354, "learning_rate": 0.00019415488858426917, "loss": 4.4272, "step": 420950 }, { "epoch": 0.7856328099668999, "grad_norm": 1.0083626508712769, "learning_rate": 0.00019415347328995006, "loss": 4.4853, "step": 421000 }, { "epoch": 0.7857261155262784, "grad_norm": 1.1188188791275024, "learning_rate": 0.00019415205782946665, "loss": 4.3596, "step": 421050 }, { "epoch": 0.7858194210856568, "grad_norm": 1.024791955947876, "learning_rate": 0.0001941506422028214, "loss": 4.5746, "step": 421100 }, { "epoch": 0.7859127266450353, "grad_norm": 1.1521062850952148, "learning_rate": 0.00019414922641001686, "loss": 4.5167, "step": 421150 }, { "epoch": 0.7860060322044138, "grad_norm": 1.025851845741272, "learning_rate": 0.00019414781045105548, "loss": 4.4505, "step": 421200 }, { "epoch": 0.7860993377637923, "grad_norm": 0.8077706694602966, "learning_rate": 0.0001941463943259398, "loss": 4.4218, "step": 421250 }, { "epoch": 0.7861926433231708, "grad_norm": 1.1391069889068604, "learning_rate": 0.00019414497803467228, "loss": 4.7806, "step": 421300 }, { "epoch": 0.7862859488825493, "grad_norm": 0.9531813859939575, "learning_rate": 0.00019414356157725546, "loss": 4.4662, "step": 421350 }, { "epoch": 0.7863792544419278, "grad_norm": 0.9496504664421082, "learning_rate": 0.0001941421449536918, "loss": 4.4881, "step": 421400 }, { "epoch": 0.7864725600013063, "grad_norm": 1.1776880025863647, "learning_rate": 0.00019414072816398384, "loss": 4.5501, "step": 421450 }, { "epoch": 0.7865658655606848, "grad_norm": 1.0474661588668823, "learning_rate": 0.00019413931120813404, "loss": 4.6452, "step": 421500 }, { "epoch": 0.7866591711200632, "grad_norm": 1.3503599166870117, "learning_rate": 0.00019413789408614492, "loss": 4.7286, "step": 421550 }, { "epoch": 0.7867524766794417, "grad_norm": 1.51902174949646, "learning_rate": 0.00019413647679801898, "loss": 4.4683, "step": 421600 }, { "epoch": 0.7868457822388203, "grad_norm": 1.1486494541168213, "learning_rate": 0.00019413505934375876, "loss": 4.4698, "step": 421650 }, { "epoch": 0.7869390877981988, "grad_norm": 1.0835721492767334, "learning_rate": 0.00019413364172336667, "loss": 4.4485, "step": 421700 }, { "epoch": 0.7870323933575772, "grad_norm": 1.2232539653778076, "learning_rate": 0.0001941322239368453, "loss": 4.7299, "step": 421750 }, { "epoch": 0.7871256989169557, "grad_norm": 1.496379017829895, "learning_rate": 0.0001941308059841971, "loss": 4.5478, "step": 421800 }, { "epoch": 0.7872190044763342, "grad_norm": 0.9647490382194519, "learning_rate": 0.0001941293878654246, "loss": 4.6043, "step": 421850 }, { "epoch": 0.7873123100357127, "grad_norm": 1.1284325122833252, "learning_rate": 0.0001941279695805303, "loss": 4.5567, "step": 421900 }, { "epoch": 0.7874056155950911, "grad_norm": 1.2496758699417114, "learning_rate": 0.00019412655112951668, "loss": 4.4419, "step": 421950 }, { "epoch": 0.7874989211544697, "grad_norm": 1.2127939462661743, "learning_rate": 0.00019412513251238625, "loss": 4.5831, "step": 422000 }, { "epoch": 0.7875922267138482, "grad_norm": 1.3008806705474854, "learning_rate": 0.00019412371372914155, "loss": 4.582, "step": 422050 }, { "epoch": 0.7876855322732267, "grad_norm": 1.2855007648468018, "learning_rate": 0.00019412229477978505, "loss": 4.5988, "step": 422100 }, { "epoch": 0.7877788378326052, "grad_norm": 1.103704571723938, "learning_rate": 0.00019412087566431926, "loss": 4.7151, "step": 422150 }, { "epoch": 0.7878721433919836, "grad_norm": 0.8745959997177124, "learning_rate": 0.0001941194563827467, "loss": 4.6391, "step": 422200 }, { "epoch": 0.7879654489513621, "grad_norm": 2.252020835876465, "learning_rate": 0.00019411803693506982, "loss": 4.7215, "step": 422250 }, { "epoch": 0.7880587545107406, "grad_norm": 1.0912048816680908, "learning_rate": 0.00019411661732129118, "loss": 4.7557, "step": 422300 }, { "epoch": 0.7881520600701192, "grad_norm": 1.004123568534851, "learning_rate": 0.00019411519754141326, "loss": 4.4193, "step": 422350 }, { "epoch": 0.7882453656294977, "grad_norm": 1.1392632722854614, "learning_rate": 0.0001941137775954386, "loss": 4.379, "step": 422400 }, { "epoch": 0.7883386711888761, "grad_norm": 1.1152008771896362, "learning_rate": 0.00019411235748336968, "loss": 4.6683, "step": 422450 }, { "epoch": 0.7884319767482546, "grad_norm": 1.1675337553024292, "learning_rate": 0.000194110937205209, "loss": 4.6349, "step": 422500 }, { "epoch": 0.7885252823076331, "grad_norm": 1.2631796598434448, "learning_rate": 0.0001941095167609591, "loss": 4.4409, "step": 422550 }, { "epoch": 0.7886185878670116, "grad_norm": 1.056110143661499, "learning_rate": 0.00019410809615062245, "loss": 4.5492, "step": 422600 }, { "epoch": 0.78871189342639, "grad_norm": 1.232704520225525, "learning_rate": 0.00019410667537420157, "loss": 4.7121, "step": 422650 }, { "epoch": 0.7888051989857686, "grad_norm": 0.8401553630828857, "learning_rate": 0.00019410525443169896, "loss": 4.3471, "step": 422700 }, { "epoch": 0.7888985045451471, "grad_norm": 1.2476882934570312, "learning_rate": 0.00019410383332311713, "loss": 4.6391, "step": 422750 }, { "epoch": 0.7889918101045256, "grad_norm": 0.9736795425415039, "learning_rate": 0.0001941024120484586, "loss": 4.6042, "step": 422800 }, { "epoch": 0.789085115663904, "grad_norm": 1.1006672382354736, "learning_rate": 0.00019410099060772584, "loss": 4.6395, "step": 422850 }, { "epoch": 0.7891784212232825, "grad_norm": 1.1191246509552002, "learning_rate": 0.00019409956900092143, "loss": 4.5684, "step": 422900 }, { "epoch": 0.789271726782661, "grad_norm": 1.0905956029891968, "learning_rate": 0.0001940981472280478, "loss": 4.5317, "step": 422950 }, { "epoch": 0.7893650323420395, "grad_norm": 0.9905209541320801, "learning_rate": 0.00019409672528910755, "loss": 4.6819, "step": 423000 }, { "epoch": 0.7893650323420395, "eval_loss": 4.721234321594238, "eval_runtime": 234.0422, "eval_samples_per_second": 11.143, "eval_steps_per_second": 11.143, "eval_tts_loss": 7.554684614714098, "step": 423000 }, { "epoch": 0.7894583379014181, "grad_norm": 1.1053640842437744, "learning_rate": 0.00019409530318410313, "loss": 4.7404, "step": 423050 }, { "epoch": 0.7895516434607965, "grad_norm": 1.2071901559829712, "learning_rate": 0.000194093880913037, "loss": 4.4602, "step": 423100 }, { "epoch": 0.789644949020175, "grad_norm": 0.8920719027519226, "learning_rate": 0.00019409245847591176, "loss": 4.4808, "step": 423150 }, { "epoch": 0.7897382545795535, "grad_norm": 1.0834848880767822, "learning_rate": 0.00019409103587272989, "loss": 4.3641, "step": 423200 }, { "epoch": 0.789831560138932, "grad_norm": 1.171592116355896, "learning_rate": 0.00019408961310349388, "loss": 4.7109, "step": 423250 }, { "epoch": 0.7899248656983104, "grad_norm": 1.1035345792770386, "learning_rate": 0.0001940881901682063, "loss": 4.5078, "step": 423300 }, { "epoch": 0.7900181712576889, "grad_norm": 1.0811997652053833, "learning_rate": 0.0001940867670668696, "loss": 4.5882, "step": 423350 }, { "epoch": 0.7901114768170675, "grad_norm": 1.1453145742416382, "learning_rate": 0.00019408534379948629, "loss": 4.6713, "step": 423400 }, { "epoch": 0.790204782376446, "grad_norm": 0.9412307739257812, "learning_rate": 0.00019408392036605892, "loss": 4.5075, "step": 423450 }, { "epoch": 0.7902980879358245, "grad_norm": 1.229821801185608, "learning_rate": 0.00019408249676658998, "loss": 4.507, "step": 423500 }, { "epoch": 0.7903913934952029, "grad_norm": 1.3128808736801147, "learning_rate": 0.000194081073001082, "loss": 4.5877, "step": 423550 }, { "epoch": 0.7904846990545814, "grad_norm": 1.1726655960083008, "learning_rate": 0.00019407964906953745, "loss": 4.5936, "step": 423600 }, { "epoch": 0.7905780046139599, "grad_norm": 1.0003212690353394, "learning_rate": 0.0001940782249719589, "loss": 4.4855, "step": 423650 }, { "epoch": 0.7906713101733384, "grad_norm": 1.2849147319793701, "learning_rate": 0.00019407680070834883, "loss": 4.7337, "step": 423700 }, { "epoch": 0.790764615732717, "grad_norm": 1.2079015970230103, "learning_rate": 0.00019407537627870976, "loss": 4.4842, "step": 423750 }, { "epoch": 0.7908579212920954, "grad_norm": 1.1246628761291504, "learning_rate": 0.0001940739516830442, "loss": 4.4204, "step": 423800 }, { "epoch": 0.7909512268514739, "grad_norm": 0.8938810229301453, "learning_rate": 0.00019407252692135466, "loss": 4.5199, "step": 423850 }, { "epoch": 0.7910445324108524, "grad_norm": 1.0785510540008545, "learning_rate": 0.00019407110199364368, "loss": 4.4074, "step": 423900 }, { "epoch": 0.7911378379702309, "grad_norm": 0.9102846384048462, "learning_rate": 0.00019406967689991378, "loss": 4.42, "step": 423950 }, { "epoch": 0.7912311435296093, "grad_norm": 0.9787639379501343, "learning_rate": 0.00019406825164016743, "loss": 4.5467, "step": 424000 }, { "epoch": 0.7913244490889878, "grad_norm": 1.0713123083114624, "learning_rate": 0.00019406682621440715, "loss": 4.7614, "step": 424050 }, { "epoch": 0.7914177546483664, "grad_norm": 0.9125092625617981, "learning_rate": 0.00019406540062263548, "loss": 4.6056, "step": 424100 }, { "epoch": 0.7915110602077449, "grad_norm": 1.1230067014694214, "learning_rate": 0.00019406397486485493, "loss": 4.499, "step": 424150 }, { "epoch": 0.7916043657671233, "grad_norm": 0.9560022354125977, "learning_rate": 0.00019406254894106803, "loss": 4.3798, "step": 424200 }, { "epoch": 0.7916976713265018, "grad_norm": 0.8355482816696167, "learning_rate": 0.00019406112285127726, "loss": 4.3809, "step": 424250 }, { "epoch": 0.7917909768858803, "grad_norm": 1.1256998777389526, "learning_rate": 0.00019405969659548519, "loss": 4.7232, "step": 424300 }, { "epoch": 0.7918842824452588, "grad_norm": 1.2029259204864502, "learning_rate": 0.00019405827017369428, "loss": 4.7615, "step": 424350 }, { "epoch": 0.7919775880046372, "grad_norm": 1.0264108180999756, "learning_rate": 0.00019405684358590706, "loss": 4.4148, "step": 424400 }, { "epoch": 0.7920708935640158, "grad_norm": 0.945449709892273, "learning_rate": 0.00019405541683212608, "loss": 4.569, "step": 424450 }, { "epoch": 0.7921641991233943, "grad_norm": 1.1476784944534302, "learning_rate": 0.00019405398991235385, "loss": 4.6362, "step": 424500 }, { "epoch": 0.7922575046827728, "grad_norm": 1.0508506298065186, "learning_rate": 0.00019405256282659287, "loss": 4.7232, "step": 424550 }, { "epoch": 0.7923508102421513, "grad_norm": 0.9733095765113831, "learning_rate": 0.00019405113557484564, "loss": 4.5333, "step": 424600 }, { "epoch": 0.7924441158015297, "grad_norm": 1.1830148696899414, "learning_rate": 0.00019404970815711475, "loss": 4.5946, "step": 424650 }, { "epoch": 0.7925374213609082, "grad_norm": 0.968312680721283, "learning_rate": 0.00019404828057340264, "loss": 4.3865, "step": 424700 }, { "epoch": 0.7926307269202867, "grad_norm": 1.0268644094467163, "learning_rate": 0.00019404685282371187, "loss": 4.6542, "step": 424750 }, { "epoch": 0.7927240324796653, "grad_norm": 0.8817425966262817, "learning_rate": 0.00019404542490804495, "loss": 4.4903, "step": 424800 }, { "epoch": 0.7928173380390438, "grad_norm": 1.0740628242492676, "learning_rate": 0.0001940439968264044, "loss": 4.7789, "step": 424850 }, { "epoch": 0.7929106435984222, "grad_norm": 1.2251063585281372, "learning_rate": 0.00019404256857879278, "loss": 4.6793, "step": 424900 }, { "epoch": 0.7930039491578007, "grad_norm": 1.330265998840332, "learning_rate": 0.00019404114016521255, "loss": 4.6689, "step": 424950 }, { "epoch": 0.7930972547171792, "grad_norm": 1.1080074310302734, "learning_rate": 0.00019403971158566625, "loss": 4.5088, "step": 425000 }, { "epoch": 0.7931905602765577, "grad_norm": 1.2231508493423462, "learning_rate": 0.00019403828284015638, "loss": 4.38, "step": 425050 }, { "epoch": 0.7932838658359361, "grad_norm": 1.0425812005996704, "learning_rate": 0.00019403685392868552, "loss": 4.5114, "step": 425100 }, { "epoch": 0.7933771713953146, "grad_norm": 0.8140989542007446, "learning_rate": 0.00019403542485125615, "loss": 4.6362, "step": 425150 }, { "epoch": 0.7934704769546932, "grad_norm": 1.0934182405471802, "learning_rate": 0.0001940339956078708, "loss": 4.3847, "step": 425200 }, { "epoch": 0.7935637825140717, "grad_norm": 1.2771217823028564, "learning_rate": 0.000194032566198532, "loss": 4.5295, "step": 425250 }, { "epoch": 0.7936570880734501, "grad_norm": 1.1933695077896118, "learning_rate": 0.0001940311366232423, "loss": 4.687, "step": 425300 }, { "epoch": 0.7937503936328286, "grad_norm": 1.0973583459854126, "learning_rate": 0.00019402970688200413, "loss": 4.6445, "step": 425350 }, { "epoch": 0.7938436991922071, "grad_norm": 1.0848637819290161, "learning_rate": 0.0001940282769748201, "loss": 4.4759, "step": 425400 }, { "epoch": 0.7939370047515856, "grad_norm": 0.9990158081054688, "learning_rate": 0.00019402684690169268, "loss": 4.5484, "step": 425450 }, { "epoch": 0.794030310310964, "grad_norm": 0.8308629989624023, "learning_rate": 0.00019402541666262445, "loss": 4.3825, "step": 425500 }, { "epoch": 0.7941236158703426, "grad_norm": 1.1615198850631714, "learning_rate": 0.0001940239862576179, "loss": 4.6229, "step": 425550 }, { "epoch": 0.7942169214297211, "grad_norm": 1.0980902910232544, "learning_rate": 0.00019402255568667557, "loss": 4.5118, "step": 425600 }, { "epoch": 0.7943102269890996, "grad_norm": 1.2873594760894775, "learning_rate": 0.00019402112494979994, "loss": 4.404, "step": 425650 }, { "epoch": 0.7944035325484781, "grad_norm": 0.7118760347366333, "learning_rate": 0.00019401969404699357, "loss": 4.3621, "step": 425700 }, { "epoch": 0.7944968381078565, "grad_norm": 0.9714242219924927, "learning_rate": 0.00019401826297825903, "loss": 4.5299, "step": 425750 }, { "epoch": 0.794590143667235, "grad_norm": 1.098937749862671, "learning_rate": 0.00019401683174359875, "loss": 4.6598, "step": 425800 }, { "epoch": 0.7946834492266135, "grad_norm": 0.9259284734725952, "learning_rate": 0.00019401540034301532, "loss": 4.2878, "step": 425850 }, { "epoch": 0.7947767547859921, "grad_norm": 0.9571802020072937, "learning_rate": 0.00019401396877651123, "loss": 4.5047, "step": 425900 }, { "epoch": 0.7948700603453706, "grad_norm": 1.0199216604232788, "learning_rate": 0.00019401253704408905, "loss": 4.6847, "step": 425950 }, { "epoch": 0.794963365904749, "grad_norm": 1.0073426961898804, "learning_rate": 0.0001940111051457513, "loss": 4.6102, "step": 426000 }, { "epoch": 0.794963365904749, "eval_loss": 4.718998432159424, "eval_runtime": 231.5463, "eval_samples_per_second": 11.263, "eval_steps_per_second": 11.263, "eval_tts_loss": 7.552527759125616, "step": 426000 }, { "epoch": 0.7950566714641275, "grad_norm": 0.927761971950531, "learning_rate": 0.00019400967308150045, "loss": 4.6581, "step": 426050 }, { "epoch": 0.795149977023506, "grad_norm": 1.1682859659194946, "learning_rate": 0.0001940082408513391, "loss": 4.5859, "step": 426100 }, { "epoch": 0.7952432825828845, "grad_norm": 0.9072175621986389, "learning_rate": 0.00019400680845526976, "loss": 4.4219, "step": 426150 }, { "epoch": 0.7953365881422629, "grad_norm": 0.6768325567245483, "learning_rate": 0.00019400537589329492, "loss": 4.5796, "step": 426200 }, { "epoch": 0.7954298937016415, "grad_norm": 1.156038522720337, "learning_rate": 0.00019400394316541712, "loss": 4.7026, "step": 426250 }, { "epoch": 0.79552319926102, "grad_norm": 1.071454405784607, "learning_rate": 0.00019400251027163894, "loss": 4.6118, "step": 426300 }, { "epoch": 0.7956165048203985, "grad_norm": 1.1358624696731567, "learning_rate": 0.00019400107721196283, "loss": 4.511, "step": 426350 }, { "epoch": 0.795709810379777, "grad_norm": 1.1195452213287354, "learning_rate": 0.00019399964398639138, "loss": 4.4004, "step": 426400 }, { "epoch": 0.7958031159391554, "grad_norm": 1.0327094793319702, "learning_rate": 0.0001939982105949271, "loss": 4.6984, "step": 426450 }, { "epoch": 0.7958964214985339, "grad_norm": 1.180101990699768, "learning_rate": 0.00019399677703757251, "loss": 4.701, "step": 426500 }, { "epoch": 0.7959897270579124, "grad_norm": 0.8951889276504517, "learning_rate": 0.00019399534331433015, "loss": 4.5, "step": 426550 }, { "epoch": 0.796083032617291, "grad_norm": 1.1613125801086426, "learning_rate": 0.00019399390942520254, "loss": 4.6163, "step": 426600 }, { "epoch": 0.7961763381766694, "grad_norm": 3.089205503463745, "learning_rate": 0.00019399247537019222, "loss": 4.5153, "step": 426650 }, { "epoch": 0.7962696437360479, "grad_norm": 1.1623356342315674, "learning_rate": 0.00019399104114930173, "loss": 4.5372, "step": 426700 }, { "epoch": 0.7963629492954264, "grad_norm": 0.8393343687057495, "learning_rate": 0.00019398960676253357, "loss": 4.3724, "step": 426750 }, { "epoch": 0.7964562548548049, "grad_norm": 1.1424648761749268, "learning_rate": 0.0001939881722098903, "loss": 4.6238, "step": 426800 }, { "epoch": 0.7965495604141833, "grad_norm": 1.0593382120132446, "learning_rate": 0.00019398673749137447, "loss": 4.5704, "step": 426850 }, { "epoch": 0.7966428659735618, "grad_norm": 0.9022440910339355, "learning_rate": 0.00019398530260698857, "loss": 4.3505, "step": 426900 }, { "epoch": 0.7967361715329404, "grad_norm": 1.0963038206100464, "learning_rate": 0.00019398386755673514, "loss": 4.5951, "step": 426950 }, { "epoch": 0.7968294770923189, "grad_norm": 1.1834969520568848, "learning_rate": 0.00019398243234061674, "loss": 4.453, "step": 427000 }, { "epoch": 0.7969227826516974, "grad_norm": 0.8198856711387634, "learning_rate": 0.00019398099695863588, "loss": 4.5635, "step": 427050 }, { "epoch": 0.7970160882110758, "grad_norm": 0.9053031802177429, "learning_rate": 0.00019397956141079504, "loss": 4.6795, "step": 427100 }, { "epoch": 0.7971093937704543, "grad_norm": 1.189697265625, "learning_rate": 0.00019397812569709691, "loss": 4.7513, "step": 427150 }, { "epoch": 0.7972026993298328, "grad_norm": 0.9244701862335205, "learning_rate": 0.00019397668981754385, "loss": 4.4181, "step": 427200 }, { "epoch": 0.7972960048892113, "grad_norm": 0.9748154878616333, "learning_rate": 0.0001939752537721385, "loss": 4.6368, "step": 427250 }, { "epoch": 0.7973893104485898, "grad_norm": 1.1173096895217896, "learning_rate": 0.00019397381756088334, "loss": 4.6749, "step": 427300 }, { "epoch": 0.7974826160079683, "grad_norm": 0.9567530751228333, "learning_rate": 0.00019397238118378093, "loss": 4.4394, "step": 427350 }, { "epoch": 0.7975759215673468, "grad_norm": 0.9598793387413025, "learning_rate": 0.0001939709446408338, "loss": 4.5758, "step": 427400 }, { "epoch": 0.7976692271267253, "grad_norm": 1.2359800338745117, "learning_rate": 0.00019396950793204451, "loss": 4.4107, "step": 427450 }, { "epoch": 0.7977625326861038, "grad_norm": 1.8980746269226074, "learning_rate": 0.00019396807105741558, "loss": 4.5808, "step": 427500 }, { "epoch": 0.7978558382454822, "grad_norm": 0.9553906917572021, "learning_rate": 0.0001939666340169495, "loss": 4.4524, "step": 427550 }, { "epoch": 0.7979491438048607, "grad_norm": 1.077438235282898, "learning_rate": 0.00019396519681064885, "loss": 4.516, "step": 427600 }, { "epoch": 0.7980424493642393, "grad_norm": 0.6697947382926941, "learning_rate": 0.00019396375943851618, "loss": 4.5095, "step": 427650 }, { "epoch": 0.7981357549236178, "grad_norm": 1.3011325597763062, "learning_rate": 0.000193962321900554, "loss": 4.5906, "step": 427700 }, { "epoch": 0.7982290604829962, "grad_norm": 1.0033787488937378, "learning_rate": 0.00019396088419676487, "loss": 4.5734, "step": 427750 }, { "epoch": 0.7983223660423747, "grad_norm": 1.2484372854232788, "learning_rate": 0.0001939594463271513, "loss": 4.724, "step": 427800 }, { "epoch": 0.7984156716017532, "grad_norm": 1.0030156373977661, "learning_rate": 0.00019395800829171585, "loss": 4.5814, "step": 427850 }, { "epoch": 0.7985089771611317, "grad_norm": 0.8476771116256714, "learning_rate": 0.00019395657009046104, "loss": 4.7281, "step": 427900 }, { "epoch": 0.7986022827205101, "grad_norm": 1.0120536088943481, "learning_rate": 0.0001939551317233894, "loss": 4.5716, "step": 427950 }, { "epoch": 0.7986955882798887, "grad_norm": 1.2245855331420898, "learning_rate": 0.00019395369319050352, "loss": 4.6486, "step": 428000 }, { "epoch": 0.7987888938392672, "grad_norm": 1.0530380010604858, "learning_rate": 0.0001939522544918059, "loss": 4.8216, "step": 428050 }, { "epoch": 0.7988821993986457, "grad_norm": 1.4846291542053223, "learning_rate": 0.00019395081562729904, "loss": 4.6106, "step": 428100 }, { "epoch": 0.7989755049580242, "grad_norm": 1.0235326290130615, "learning_rate": 0.00019394937659698554, "loss": 4.5297, "step": 428150 }, { "epoch": 0.7990688105174026, "grad_norm": 1.066236138343811, "learning_rate": 0.00019394793740086793, "loss": 4.663, "step": 428200 }, { "epoch": 0.7991621160767811, "grad_norm": 0.7847051620483398, "learning_rate": 0.00019394649803894875, "loss": 4.5661, "step": 428250 }, { "epoch": 0.7992554216361596, "grad_norm": 0.9233770370483398, "learning_rate": 0.00019394505851123053, "loss": 4.5384, "step": 428300 }, { "epoch": 0.7993487271955382, "grad_norm": 1.0995171070098877, "learning_rate": 0.0001939436188177158, "loss": 4.6251, "step": 428350 }, { "epoch": 0.7994420327549167, "grad_norm": 1.2371164560317993, "learning_rate": 0.0001939421789584071, "loss": 4.5591, "step": 428400 }, { "epoch": 0.7995353383142951, "grad_norm": 1.12126886844635, "learning_rate": 0.000193940738933307, "loss": 4.5319, "step": 428450 }, { "epoch": 0.7996286438736736, "grad_norm": 1.0614880323410034, "learning_rate": 0.000193939298742418, "loss": 4.6363, "step": 428500 }, { "epoch": 0.7997219494330521, "grad_norm": 0.83480304479599, "learning_rate": 0.0001939378583857427, "loss": 4.6423, "step": 428550 }, { "epoch": 0.7998152549924306, "grad_norm": 1.0235836505889893, "learning_rate": 0.0001939364178632836, "loss": 4.5759, "step": 428600 }, { "epoch": 0.799908560551809, "grad_norm": 1.0326169729232788, "learning_rate": 0.00019393497717504326, "loss": 4.4497, "step": 428650 }, { "epoch": 0.8000018661111876, "grad_norm": 0.876973032951355, "learning_rate": 0.00019393353632102421, "loss": 4.3632, "step": 428700 }, { "epoch": 0.8000951716705661, "grad_norm": 1.080389380455017, "learning_rate": 0.000193932095301229, "loss": 4.6233, "step": 428750 }, { "epoch": 0.8001884772299446, "grad_norm": 0.9304258227348328, "learning_rate": 0.00019393065411566016, "loss": 4.5492, "step": 428800 }, { "epoch": 0.800281782789323, "grad_norm": 1.1061395406723022, "learning_rate": 0.00019392921276432023, "loss": 4.5125, "step": 428850 }, { "epoch": 0.8003750883487015, "grad_norm": 1.3603273630142212, "learning_rate": 0.00019392777124721178, "loss": 4.3903, "step": 428900 }, { "epoch": 0.80046839390808, "grad_norm": 1.0902111530303955, "learning_rate": 0.00019392632956433733, "loss": 4.5412, "step": 428950 }, { "epoch": 0.8005616994674585, "grad_norm": 0.8139966726303101, "learning_rate": 0.00019392488771569945, "loss": 4.4993, "step": 429000 }, { "epoch": 0.8005616994674585, "eval_loss": 4.7199482917785645, "eval_runtime": 232.6367, "eval_samples_per_second": 11.211, "eval_steps_per_second": 11.211, "eval_tts_loss": 7.580851806817645, "step": 429000 }, { "epoch": 0.8006550050268371, "grad_norm": 1.0927164554595947, "learning_rate": 0.0001939234457013007, "loss": 4.5409, "step": 429050 }, { "epoch": 0.8007483105862155, "grad_norm": 0.8889446258544922, "learning_rate": 0.00019392200352114355, "loss": 4.5445, "step": 429100 }, { "epoch": 0.800841616145594, "grad_norm": 1.0192301273345947, "learning_rate": 0.0001939205611752306, "loss": 4.633, "step": 429150 }, { "epoch": 0.8009349217049725, "grad_norm": 0.770257294178009, "learning_rate": 0.00019391911866356435, "loss": 4.4848, "step": 429200 }, { "epoch": 0.801028227264351, "grad_norm": 1.1421622037887573, "learning_rate": 0.00019391767598614745, "loss": 4.544, "step": 429250 }, { "epoch": 0.8011215328237294, "grad_norm": 0.9731292128562927, "learning_rate": 0.00019391623314298233, "loss": 4.406, "step": 429300 }, { "epoch": 0.8012148383831079, "grad_norm": 0.8407917022705078, "learning_rate": 0.00019391479013407162, "loss": 4.7693, "step": 429350 }, { "epoch": 0.8013081439424865, "grad_norm": 0.8591404557228088, "learning_rate": 0.00019391334695941778, "loss": 4.5128, "step": 429400 }, { "epoch": 0.801401449501865, "grad_norm": 0.9232378602027893, "learning_rate": 0.00019391190361902342, "loss": 4.5328, "step": 429450 }, { "epoch": 0.8014947550612435, "grad_norm": 1.314203143119812, "learning_rate": 0.0001939104601128911, "loss": 4.5138, "step": 429500 }, { "epoch": 0.8015880606206219, "grad_norm": 0.9116572141647339, "learning_rate": 0.00019390901644102335, "loss": 4.7407, "step": 429550 }, { "epoch": 0.8016813661800004, "grad_norm": 1.0177239179611206, "learning_rate": 0.00019390757260342268, "loss": 4.5381, "step": 429600 }, { "epoch": 0.8017746717393789, "grad_norm": 0.8840582966804504, "learning_rate": 0.00019390612860009169, "loss": 4.4708, "step": 429650 }, { "epoch": 0.8018679772987574, "grad_norm": 1.3547614812850952, "learning_rate": 0.00019390468443103288, "loss": 4.4395, "step": 429700 }, { "epoch": 0.801961282858136, "grad_norm": 1.4510341882705688, "learning_rate": 0.00019390324009624886, "loss": 4.5966, "step": 429750 }, { "epoch": 0.8020545884175144, "grad_norm": 0.9955864548683167, "learning_rate": 0.0001939017955957421, "loss": 4.4213, "step": 429800 }, { "epoch": 0.8021478939768929, "grad_norm": 1.103618860244751, "learning_rate": 0.00019390035092951523, "loss": 4.6235, "step": 429850 }, { "epoch": 0.8022411995362714, "grad_norm": 1.115838885307312, "learning_rate": 0.00019389890609757075, "loss": 4.643, "step": 429900 }, { "epoch": 0.8023345050956499, "grad_norm": 0.9047982692718506, "learning_rate": 0.00019389746109991121, "loss": 4.4769, "step": 429950 }, { "epoch": 0.8024278106550283, "grad_norm": 1.1972705125808716, "learning_rate": 0.00019389601593653919, "loss": 4.3868, "step": 430000 }, { "epoch": 0.8025211162144068, "grad_norm": 0.8004233837127686, "learning_rate": 0.00019389457060745724, "loss": 4.7359, "step": 430050 }, { "epoch": 0.8026144217737854, "grad_norm": 1.0934377908706665, "learning_rate": 0.00019389312511266786, "loss": 4.5359, "step": 430100 }, { "epoch": 0.8027077273331639, "grad_norm": 1.0269830226898193, "learning_rate": 0.00019389167945217366, "loss": 4.4079, "step": 430150 }, { "epoch": 0.8028010328925423, "grad_norm": 1.061034917831421, "learning_rate": 0.00019389023362597713, "loss": 4.5938, "step": 430200 }, { "epoch": 0.8028943384519208, "grad_norm": 1.0427353382110596, "learning_rate": 0.00019388878763408092, "loss": 4.5804, "step": 430250 }, { "epoch": 0.8029876440112993, "grad_norm": 1.1512060165405273, "learning_rate": 0.00019388734147648745, "loss": 4.432, "step": 430300 }, { "epoch": 0.8030809495706778, "grad_norm": 0.8705995678901672, "learning_rate": 0.0001938858951531994, "loss": 4.5563, "step": 430350 }, { "epoch": 0.8031742551300562, "grad_norm": 0.9586858153343201, "learning_rate": 0.00019388444866421924, "loss": 4.4358, "step": 430400 }, { "epoch": 0.8032675606894347, "grad_norm": 1.1046531200408936, "learning_rate": 0.00019388300200954954, "loss": 4.5264, "step": 430450 }, { "epoch": 0.8033608662488133, "grad_norm": 1.192919373512268, "learning_rate": 0.00019388155518919287, "loss": 4.6478, "step": 430500 }, { "epoch": 0.8034541718081918, "grad_norm": 1.0320477485656738, "learning_rate": 0.0001938801082031518, "loss": 4.6415, "step": 430550 }, { "epoch": 0.8035474773675703, "grad_norm": 0.8191885948181152, "learning_rate": 0.0001938786610514288, "loss": 4.418, "step": 430600 }, { "epoch": 0.8036407829269487, "grad_norm": 1.1384339332580566, "learning_rate": 0.00019387721373402653, "loss": 4.6896, "step": 430650 }, { "epoch": 0.8037340884863272, "grad_norm": 0.8888775706291199, "learning_rate": 0.00019387576625094747, "loss": 4.5304, "step": 430700 }, { "epoch": 0.8038273940457057, "grad_norm": 1.0733976364135742, "learning_rate": 0.0001938743186021942, "loss": 4.5005, "step": 430750 }, { "epoch": 0.8039206996050842, "grad_norm": 1.3514388799667358, "learning_rate": 0.00019387287078776927, "loss": 4.7052, "step": 430800 }, { "epoch": 0.8040140051644628, "grad_norm": 1.0057579278945923, "learning_rate": 0.00019387142280767527, "loss": 4.5062, "step": 430850 }, { "epoch": 0.8041073107238412, "grad_norm": 1.267685055732727, "learning_rate": 0.00019386997466191473, "loss": 4.4404, "step": 430900 }, { "epoch": 0.8042006162832197, "grad_norm": 1.0608516931533813, "learning_rate": 0.00019386852635049017, "loss": 4.5119, "step": 430950 }, { "epoch": 0.8042939218425982, "grad_norm": 1.2537400722503662, "learning_rate": 0.0001938670778734042, "loss": 4.7115, "step": 431000 }, { "epoch": 0.8043872274019767, "grad_norm": 1.269854187965393, "learning_rate": 0.00019386562923065935, "loss": 4.7092, "step": 431050 }, { "epoch": 0.8044805329613551, "grad_norm": 1.3239473104476929, "learning_rate": 0.00019386418042225816, "loss": 4.5927, "step": 431100 }, { "epoch": 0.8045738385207336, "grad_norm": 1.200350046157837, "learning_rate": 0.00019386273144820324, "loss": 4.8531, "step": 431150 }, { "epoch": 0.8046671440801122, "grad_norm": 0.8199599981307983, "learning_rate": 0.0001938612823084971, "loss": 4.5809, "step": 431200 }, { "epoch": 0.8047604496394907, "grad_norm": 1.2164291143417358, "learning_rate": 0.0001938598330031423, "loss": 4.5588, "step": 431250 }, { "epoch": 0.8048537551988691, "grad_norm": 0.743262767791748, "learning_rate": 0.00019385838353214142, "loss": 4.711, "step": 431300 }, { "epoch": 0.8049470607582476, "grad_norm": 1.2544654607772827, "learning_rate": 0.000193856933895497, "loss": 4.4243, "step": 431350 }, { "epoch": 0.8050403663176261, "grad_norm": 1.6423524618148804, "learning_rate": 0.0001938554840932116, "loss": 4.401, "step": 431400 }, { "epoch": 0.8051336718770046, "grad_norm": 0.9570430517196655, "learning_rate": 0.00019385403412528783, "loss": 4.5196, "step": 431450 }, { "epoch": 0.805226977436383, "grad_norm": 0.5382121205329895, "learning_rate": 0.00019385258399172813, "loss": 4.4504, "step": 431500 }, { "epoch": 0.8053202829957616, "grad_norm": 0.977709174156189, "learning_rate": 0.0001938511336925352, "loss": 4.4906, "step": 431550 }, { "epoch": 0.8054135885551401, "grad_norm": 1.1496464014053345, "learning_rate": 0.0001938496832277115, "loss": 4.644, "step": 431600 }, { "epoch": 0.8055068941145186, "grad_norm": 0.735632598400116, "learning_rate": 0.00019384823259725964, "loss": 4.4586, "step": 431650 }, { "epoch": 0.8056001996738971, "grad_norm": 1.154396653175354, "learning_rate": 0.00019384678180118213, "loss": 4.4867, "step": 431700 }, { "epoch": 0.8056935052332755, "grad_norm": 0.8699199557304382, "learning_rate": 0.0001938453308394816, "loss": 4.5057, "step": 431750 }, { "epoch": 0.805786810792654, "grad_norm": 1.3764482736587524, "learning_rate": 0.0001938438797121606, "loss": 4.4259, "step": 431800 }, { "epoch": 0.8058801163520325, "grad_norm": 1.190111756324768, "learning_rate": 0.0001938424284192216, "loss": 4.6185, "step": 431850 }, { "epoch": 0.8059734219114111, "grad_norm": 0.7815505862236023, "learning_rate": 0.00019384097696066723, "loss": 4.3942, "step": 431900 }, { "epoch": 0.8060667274707896, "grad_norm": 1.2037734985351562, "learning_rate": 0.00019383952533650006, "loss": 4.6106, "step": 431950 }, { "epoch": 0.806160033030168, "grad_norm": 0.9880151748657227, "learning_rate": 0.00019383807354672266, "loss": 4.6179, "step": 432000 }, { "epoch": 0.806160033030168, "eval_loss": 4.717226505279541, "eval_runtime": 232.5206, "eval_samples_per_second": 11.216, "eval_steps_per_second": 11.216, "eval_tts_loss": 7.522934058765406, "step": 432000 }, { "epoch": 0.8062533385895465, "grad_norm": 1.1217437982559204, "learning_rate": 0.00019383662159133756, "loss": 4.5934, "step": 432050 }, { "epoch": 0.806346644148925, "grad_norm": 1.386207103729248, "learning_rate": 0.00019383516947034735, "loss": 4.7539, "step": 432100 }, { "epoch": 0.8064399497083035, "grad_norm": 1.0124891996383667, "learning_rate": 0.00019383371718375455, "loss": 4.3381, "step": 432150 }, { "epoch": 0.8065332552676819, "grad_norm": 0.9883818626403809, "learning_rate": 0.00019383226473156174, "loss": 4.3125, "step": 432200 }, { "epoch": 0.8066265608270605, "grad_norm": 0.9008289575576782, "learning_rate": 0.00019383081211377153, "loss": 4.3435, "step": 432250 }, { "epoch": 0.806719866386439, "grad_norm": 1.0534547567367554, "learning_rate": 0.00019382935933038643, "loss": 4.6482, "step": 432300 }, { "epoch": 0.8068131719458175, "grad_norm": 0.7578873634338379, "learning_rate": 0.000193827906381409, "loss": 4.4737, "step": 432350 }, { "epoch": 0.806906477505196, "grad_norm": 1.3317744731903076, "learning_rate": 0.00019382645326684186, "loss": 4.7364, "step": 432400 }, { "epoch": 0.8069997830645744, "grad_norm": 1.2212891578674316, "learning_rate": 0.0001938249999866875, "loss": 4.4826, "step": 432450 }, { "epoch": 0.8070930886239529, "grad_norm": 1.2318320274353027, "learning_rate": 0.00019382354654094857, "loss": 4.7266, "step": 432500 }, { "epoch": 0.8071863941833314, "grad_norm": 1.1068717241287231, "learning_rate": 0.00019382209292962755, "loss": 4.4529, "step": 432550 }, { "epoch": 0.80727969974271, "grad_norm": 0.8774948716163635, "learning_rate": 0.00019382063915272707, "loss": 4.5036, "step": 432600 }, { "epoch": 0.8073730053020884, "grad_norm": 1.1480398178100586, "learning_rate": 0.00019381918521024967, "loss": 4.3559, "step": 432650 }, { "epoch": 0.8074663108614669, "grad_norm": 1.0650641918182373, "learning_rate": 0.0001938177311021979, "loss": 4.6795, "step": 432700 }, { "epoch": 0.8075596164208454, "grad_norm": 1.1158818006515503, "learning_rate": 0.00019381627682857436, "loss": 4.5262, "step": 432750 }, { "epoch": 0.8076529219802239, "grad_norm": 0.9662179350852966, "learning_rate": 0.00019381482238938158, "loss": 4.5011, "step": 432800 }, { "epoch": 0.8077462275396023, "grad_norm": 1.0558754205703735, "learning_rate": 0.00019381336778462216, "loss": 4.4317, "step": 432850 }, { "epoch": 0.8078395330989808, "grad_norm": 1.1265623569488525, "learning_rate": 0.00019381191301429864, "loss": 4.5669, "step": 432900 }, { "epoch": 0.8079328386583594, "grad_norm": 1.1973166465759277, "learning_rate": 0.0001938104580784136, "loss": 4.4845, "step": 432950 }, { "epoch": 0.8080261442177379, "grad_norm": 1.4364298582077026, "learning_rate": 0.0001938090029769696, "loss": 4.4275, "step": 433000 }, { "epoch": 0.8081194497771164, "grad_norm": 1.2793713808059692, "learning_rate": 0.00019380754770996925, "loss": 4.6811, "step": 433050 }, { "epoch": 0.8082127553364948, "grad_norm": 0.9621243476867676, "learning_rate": 0.00019380609227741507, "loss": 4.4556, "step": 433100 }, { "epoch": 0.8083060608958733, "grad_norm": 1.181762933731079, "learning_rate": 0.00019380463667930965, "loss": 4.4645, "step": 433150 }, { "epoch": 0.8083993664552518, "grad_norm": 1.160107135772705, "learning_rate": 0.00019380318091565552, "loss": 4.4719, "step": 433200 }, { "epoch": 0.8084926720146303, "grad_norm": 1.064186692237854, "learning_rate": 0.0001938017249864553, "loss": 4.5754, "step": 433250 }, { "epoch": 0.8085859775740089, "grad_norm": 0.985802173614502, "learning_rate": 0.00019380026889171154, "loss": 4.7013, "step": 433300 }, { "epoch": 0.8086792831333873, "grad_norm": 1.314144253730774, "learning_rate": 0.00019379881263142682, "loss": 4.4962, "step": 433350 }, { "epoch": 0.8087725886927658, "grad_norm": 1.0957282781600952, "learning_rate": 0.0001937973562056037, "loss": 4.5735, "step": 433400 }, { "epoch": 0.8088658942521443, "grad_norm": 1.0866283178329468, "learning_rate": 0.00019379589961424475, "loss": 4.5805, "step": 433450 }, { "epoch": 0.8089591998115228, "grad_norm": 1.0170081853866577, "learning_rate": 0.0001937944428573525, "loss": 4.6924, "step": 433500 }, { "epoch": 0.8090525053709012, "grad_norm": 1.194145917892456, "learning_rate": 0.0001937929859349296, "loss": 4.4469, "step": 433550 }, { "epoch": 0.8091458109302797, "grad_norm": 0.969913125038147, "learning_rate": 0.00019379152884697857, "loss": 4.4026, "step": 433600 }, { "epoch": 0.8092391164896583, "grad_norm": 1.0304129123687744, "learning_rate": 0.00019379007159350201, "loss": 4.4583, "step": 433650 }, { "epoch": 0.8093324220490368, "grad_norm": 0.905687153339386, "learning_rate": 0.00019378861417450248, "loss": 4.5727, "step": 433700 }, { "epoch": 0.8094257276084152, "grad_norm": 1.0900845527648926, "learning_rate": 0.00019378715658998252, "loss": 4.5235, "step": 433750 }, { "epoch": 0.8095190331677937, "grad_norm": 1.0319151878356934, "learning_rate": 0.00019378569883994476, "loss": 4.563, "step": 433800 }, { "epoch": 0.8096123387271722, "grad_norm": 0.8259133100509644, "learning_rate": 0.00019378424092439172, "loss": 4.5616, "step": 433850 }, { "epoch": 0.8097056442865507, "grad_norm": 0.8913447260856628, "learning_rate": 0.00019378278284332603, "loss": 4.6049, "step": 433900 }, { "epoch": 0.8097989498459292, "grad_norm": 0.9409626722335815, "learning_rate": 0.00019378132459675022, "loss": 4.5054, "step": 433950 }, { "epoch": 0.8098922554053077, "grad_norm": 1.0639225244522095, "learning_rate": 0.00019377986618466687, "loss": 4.5129, "step": 434000 }, { "epoch": 0.8099855609646862, "grad_norm": 1.1352790594100952, "learning_rate": 0.00019377840760707852, "loss": 4.5635, "step": 434050 }, { "epoch": 0.8100788665240647, "grad_norm": 0.8985616564750671, "learning_rate": 0.00019377694886398784, "loss": 4.4284, "step": 434100 }, { "epoch": 0.8101721720834432, "grad_norm": 1.2385417222976685, "learning_rate": 0.0001937754899553973, "loss": 4.5906, "step": 434150 }, { "epoch": 0.8102654776428216, "grad_norm": 1.389296293258667, "learning_rate": 0.00019377403088130955, "loss": 4.5582, "step": 434200 }, { "epoch": 0.8103587832022001, "grad_norm": 0.6670157313346863, "learning_rate": 0.0001937725716417271, "loss": 4.426, "step": 434250 }, { "epoch": 0.8104520887615786, "grad_norm": 0.9546065330505371, "learning_rate": 0.0001937711122366526, "loss": 4.632, "step": 434300 }, { "epoch": 0.8105453943209572, "grad_norm": 1.1238577365875244, "learning_rate": 0.00019376965266608854, "loss": 4.6131, "step": 434350 }, { "epoch": 0.8106386998803357, "grad_norm": 1.2199645042419434, "learning_rate": 0.0001937681929300376, "loss": 4.6498, "step": 434400 }, { "epoch": 0.8107320054397141, "grad_norm": 1.0700517892837524, "learning_rate": 0.00019376673302850227, "loss": 4.7002, "step": 434450 }, { "epoch": 0.8108253109990926, "grad_norm": 1.0681084394454956, "learning_rate": 0.00019376527296148513, "loss": 4.4733, "step": 434500 }, { "epoch": 0.8109186165584711, "grad_norm": 0.9579760432243347, "learning_rate": 0.00019376381272898883, "loss": 4.376, "step": 434550 }, { "epoch": 0.8110119221178496, "grad_norm": 1.084481954574585, "learning_rate": 0.00019376235233101586, "loss": 4.5135, "step": 434600 }, { "epoch": 0.811105227677228, "grad_norm": 1.28203547000885, "learning_rate": 0.00019376089176756888, "loss": 4.5879, "step": 434650 }, { "epoch": 0.8111985332366066, "grad_norm": 1.1159179210662842, "learning_rate": 0.0001937594310386504, "loss": 4.5683, "step": 434700 }, { "epoch": 0.8112918387959851, "grad_norm": 1.1986109018325806, "learning_rate": 0.00019375797014426301, "loss": 4.5916, "step": 434750 }, { "epoch": 0.8113851443553636, "grad_norm": 1.1072922945022583, "learning_rate": 0.0001937565090844093, "loss": 4.4843, "step": 434800 }, { "epoch": 0.811478449914742, "grad_norm": 0.9535547494888306, "learning_rate": 0.00019375504785909184, "loss": 4.3743, "step": 434850 }, { "epoch": 0.8115717554741205, "grad_norm": 1.0558990240097046, "learning_rate": 0.00019375358646831324, "loss": 4.5874, "step": 434900 }, { "epoch": 0.811665061033499, "grad_norm": 1.1166068315505981, "learning_rate": 0.00019375212491207607, "loss": 4.4974, "step": 434950 }, { "epoch": 0.8117583665928775, "grad_norm": 0.9038137197494507, "learning_rate": 0.00019375066319038287, "loss": 4.6755, "step": 435000 }, { "epoch": 0.8117583665928775, "eval_loss": 4.713137149810791, "eval_runtime": 233.558, "eval_samples_per_second": 11.166, "eval_steps_per_second": 11.166, "eval_tts_loss": 7.611853272671257, "step": 435000 }, { "epoch": 0.8118516721522561, "grad_norm": 1.1564393043518066, "learning_rate": 0.00019374920130323624, "loss": 4.5115, "step": 435050 }, { "epoch": 0.8119449777116345, "grad_norm": 1.019436001777649, "learning_rate": 0.00019374773925063879, "loss": 4.5726, "step": 435100 }, { "epoch": 0.812038283271013, "grad_norm": 1.456741213798523, "learning_rate": 0.00019374627703259306, "loss": 4.5041, "step": 435150 }, { "epoch": 0.8121315888303915, "grad_norm": 1.1122186183929443, "learning_rate": 0.00019374481464910163, "loss": 4.4887, "step": 435200 }, { "epoch": 0.81222489438977, "grad_norm": 0.973937451839447, "learning_rate": 0.00019374335210016713, "loss": 4.6768, "step": 435250 }, { "epoch": 0.8123181999491484, "grad_norm": 1.2644262313842773, "learning_rate": 0.00019374188938579208, "loss": 4.625, "step": 435300 }, { "epoch": 0.8124115055085269, "grad_norm": 0.8474568724632263, "learning_rate": 0.0001937404265059791, "loss": 4.4576, "step": 435350 }, { "epoch": 0.8125048110679055, "grad_norm": 1.687286615371704, "learning_rate": 0.00019373896346073076, "loss": 4.6964, "step": 435400 }, { "epoch": 0.812598116627284, "grad_norm": 1.0351998805999756, "learning_rate": 0.00019373750025004967, "loss": 4.5602, "step": 435450 }, { "epoch": 0.8126914221866625, "grad_norm": 1.0308910608291626, "learning_rate": 0.00019373603687393835, "loss": 4.6834, "step": 435500 }, { "epoch": 0.8127847277460409, "grad_norm": 1.1440367698669434, "learning_rate": 0.00019373457333239944, "loss": 4.4099, "step": 435550 }, { "epoch": 0.8128780333054194, "grad_norm": 0.7390778064727783, "learning_rate": 0.0001937331096254355, "loss": 4.6605, "step": 435600 }, { "epoch": 0.8129713388647979, "grad_norm": 0.8651524782180786, "learning_rate": 0.0001937316457530491, "loss": 4.692, "step": 435650 }, { "epoch": 0.8130646444241764, "grad_norm": 0.9918853640556335, "learning_rate": 0.00019373018171524285, "loss": 4.5824, "step": 435700 }, { "epoch": 0.8131579499835548, "grad_norm": 1.238487720489502, "learning_rate": 0.0001937287175120193, "loss": 4.5105, "step": 435750 }, { "epoch": 0.8132512555429334, "grad_norm": 1.0910192728042603, "learning_rate": 0.0001937272531433811, "loss": 4.4074, "step": 435800 }, { "epoch": 0.8133445611023119, "grad_norm": 0.9015228748321533, "learning_rate": 0.00019372578860933076, "loss": 4.7231, "step": 435850 }, { "epoch": 0.8134378666616904, "grad_norm": 0.8462371826171875, "learning_rate": 0.0001937243239098709, "loss": 4.4896, "step": 435900 }, { "epoch": 0.8135311722210689, "grad_norm": 0.8238480687141418, "learning_rate": 0.00019372285904500406, "loss": 4.5205, "step": 435950 }, { "epoch": 0.8136244777804473, "grad_norm": 0.7250654101371765, "learning_rate": 0.00019372139401473292, "loss": 4.5316, "step": 436000 }, { "epoch": 0.8137177833398258, "grad_norm": 1.0538784265518188, "learning_rate": 0.00019371992881906001, "loss": 4.5203, "step": 436050 }, { "epoch": 0.8138110888992043, "grad_norm": 0.9435088038444519, "learning_rate": 0.00019371846345798787, "loss": 4.4478, "step": 436100 }, { "epoch": 0.8139043944585829, "grad_norm": 1.0073535442352295, "learning_rate": 0.00019371699793151916, "loss": 4.685, "step": 436150 }, { "epoch": 0.8139977000179613, "grad_norm": 1.3032444715499878, "learning_rate": 0.00019371553223965644, "loss": 4.6852, "step": 436200 }, { "epoch": 0.8140910055773398, "grad_norm": 1.096046805381775, "learning_rate": 0.0001937140663824023, "loss": 4.4044, "step": 436250 }, { "epoch": 0.8141843111367183, "grad_norm": 0.7758470773696899, "learning_rate": 0.00019371260035975932, "loss": 4.3741, "step": 436300 }, { "epoch": 0.8142776166960968, "grad_norm": 0.9498625993728638, "learning_rate": 0.0001937111341717301, "loss": 4.7392, "step": 436350 }, { "epoch": 0.8143709222554752, "grad_norm": 1.3436883687973022, "learning_rate": 0.0001937096678183172, "loss": 4.4038, "step": 436400 }, { "epoch": 0.8144642278148537, "grad_norm": 0.9300359487533569, "learning_rate": 0.00019370820129952323, "loss": 4.5385, "step": 436450 }, { "epoch": 0.8145575333742323, "grad_norm": 1.2873274087905884, "learning_rate": 0.00019370673461535076, "loss": 4.6742, "step": 436500 }, { "epoch": 0.8146508389336108, "grad_norm": 1.1295846700668335, "learning_rate": 0.0001937052677658024, "loss": 4.5639, "step": 436550 }, { "epoch": 0.8147441444929893, "grad_norm": 1.1773862838745117, "learning_rate": 0.00019370380075088074, "loss": 4.6209, "step": 436600 }, { "epoch": 0.8148374500523677, "grad_norm": 1.0682029724121094, "learning_rate": 0.00019370233357058836, "loss": 4.5695, "step": 436650 }, { "epoch": 0.8149307556117462, "grad_norm": 0.7991012334823608, "learning_rate": 0.00019370086622492782, "loss": 4.475, "step": 436700 }, { "epoch": 0.8150240611711247, "grad_norm": 1.3142451047897339, "learning_rate": 0.0001936993987139018, "loss": 4.5237, "step": 436750 }, { "epoch": 0.8151173667305032, "grad_norm": 1.2649680376052856, "learning_rate": 0.00019369793103751276, "loss": 4.6097, "step": 436800 }, { "epoch": 0.8152106722898818, "grad_norm": 0.9599008560180664, "learning_rate": 0.0001936964631957634, "loss": 4.583, "step": 436850 }, { "epoch": 0.8153039778492602, "grad_norm": 1.1062766313552856, "learning_rate": 0.00019369499518865623, "loss": 4.479, "step": 436900 }, { "epoch": 0.8153972834086387, "grad_norm": 0.7987418174743652, "learning_rate": 0.0001936935270161939, "loss": 4.4334, "step": 436950 }, { "epoch": 0.8154905889680172, "grad_norm": 0.9704287648200989, "learning_rate": 0.000193692058678379, "loss": 4.4151, "step": 437000 }, { "epoch": 0.8155838945273957, "grad_norm": 0.8681563138961792, "learning_rate": 0.0001936905901752141, "loss": 4.6867, "step": 437050 }, { "epoch": 0.8156772000867741, "grad_norm": 0.8741142153739929, "learning_rate": 0.00019368912150670176, "loss": 4.4993, "step": 437100 }, { "epoch": 0.8157705056461526, "grad_norm": 0.6997228264808655, "learning_rate": 0.00019368765267284465, "loss": 4.5395, "step": 437150 }, { "epoch": 0.8158638112055312, "grad_norm": 1.2820758819580078, "learning_rate": 0.0001936861836736453, "loss": 4.2663, "step": 437200 }, { "epoch": 0.8159571167649097, "grad_norm": 1.075132131576538, "learning_rate": 0.0001936847145091063, "loss": 4.6971, "step": 437250 }, { "epoch": 0.8160504223242881, "grad_norm": 1.4322841167449951, "learning_rate": 0.0001936832451792303, "loss": 4.5894, "step": 437300 }, { "epoch": 0.8161437278836666, "grad_norm": 1.2873334884643555, "learning_rate": 0.00019368177568401983, "loss": 4.549, "step": 437350 }, { "epoch": 0.8162370334430451, "grad_norm": 1.1981624364852905, "learning_rate": 0.00019368030602347748, "loss": 4.7408, "step": 437400 }, { "epoch": 0.8163303390024236, "grad_norm": 1.266121506690979, "learning_rate": 0.00019367883619760595, "loss": 4.4947, "step": 437450 }, { "epoch": 0.816423644561802, "grad_norm": 1.034895896911621, "learning_rate": 0.0001936773662064077, "loss": 4.5056, "step": 437500 }, { "epoch": 0.8165169501211806, "grad_norm": 0.8305200338363647, "learning_rate": 0.00019367589604988542, "loss": 4.3609, "step": 437550 }, { "epoch": 0.8166102556805591, "grad_norm": 1.0634491443634033, "learning_rate": 0.00019367442572804162, "loss": 4.5135, "step": 437600 }, { "epoch": 0.8167035612399376, "grad_norm": 1.3011289834976196, "learning_rate": 0.00019367295524087895, "loss": 4.4633, "step": 437650 }, { "epoch": 0.8167968667993161, "grad_norm": 1.2609500885009766, "learning_rate": 0.0001936714845884, "loss": 4.6228, "step": 437700 }, { "epoch": 0.8168901723586945, "grad_norm": 0.9911806583404541, "learning_rate": 0.00019367001377060738, "loss": 4.5142, "step": 437750 }, { "epoch": 0.816983477918073, "grad_norm": 1.2613277435302734, "learning_rate": 0.0001936685427875037, "loss": 4.6284, "step": 437800 }, { "epoch": 0.8170767834774515, "grad_norm": 1.2594963312149048, "learning_rate": 0.00019366707163909146, "loss": 4.4815, "step": 437850 }, { "epoch": 0.8171700890368301, "grad_norm": 0.746685802936554, "learning_rate": 0.00019366560032537335, "loss": 4.4356, "step": 437900 }, { "epoch": 0.8172633945962086, "grad_norm": 1.3757548332214355, "learning_rate": 0.00019366412884635193, "loss": 4.59, "step": 437950 }, { "epoch": 0.817356700155587, "grad_norm": 1.0559455156326294, "learning_rate": 0.0001936626572020298, "loss": 4.545, "step": 438000 }, { "epoch": 0.817356700155587, "eval_loss": 4.7206220626831055, "eval_runtime": 230.9923, "eval_samples_per_second": 11.29, "eval_steps_per_second": 11.29, "eval_tts_loss": 7.6207517155023305, "step": 438000 }, { "epoch": 0.8174500057149655, "grad_norm": 0.9019753336906433, "learning_rate": 0.00019366118539240957, "loss": 4.6121, "step": 438050 }, { "epoch": 0.817543311274344, "grad_norm": 0.8623765110969543, "learning_rate": 0.00019365971341749383, "loss": 4.4661, "step": 438100 }, { "epoch": 0.8176366168337225, "grad_norm": 1.2834802865982056, "learning_rate": 0.00019365824127728516, "loss": 4.5664, "step": 438150 }, { "epoch": 0.8177299223931009, "grad_norm": 1.1096104383468628, "learning_rate": 0.00019365676897178618, "loss": 4.4834, "step": 438200 }, { "epoch": 0.8178232279524795, "grad_norm": 1.050219178199768, "learning_rate": 0.00019365529650099947, "loss": 4.3719, "step": 438250 }, { "epoch": 0.817916533511858, "grad_norm": 1.2993484735488892, "learning_rate": 0.00019365382386492765, "loss": 4.4403, "step": 438300 }, { "epoch": 0.8180098390712365, "grad_norm": 1.1241735219955444, "learning_rate": 0.0001936523510635733, "loss": 4.475, "step": 438350 }, { "epoch": 0.818103144630615, "grad_norm": 1.0164114236831665, "learning_rate": 0.00019365087809693906, "loss": 4.5896, "step": 438400 }, { "epoch": 0.8181964501899934, "grad_norm": 1.322221279144287, "learning_rate": 0.00019364940496502747, "loss": 4.6524, "step": 438450 }, { "epoch": 0.8182897557493719, "grad_norm": 1.0273584127426147, "learning_rate": 0.00019364793166784118, "loss": 4.5349, "step": 438500 }, { "epoch": 0.8183830613087504, "grad_norm": 1.0621941089630127, "learning_rate": 0.00019364645820538277, "loss": 4.4498, "step": 438550 }, { "epoch": 0.818476366868129, "grad_norm": 1.1163345575332642, "learning_rate": 0.00019364498457765482, "loss": 4.4727, "step": 438600 }, { "epoch": 0.8185696724275074, "grad_norm": 1.2609418630599976, "learning_rate": 0.00019364351078465997, "loss": 4.6392, "step": 438650 }, { "epoch": 0.8186629779868859, "grad_norm": 1.1600068807601929, "learning_rate": 0.0001936420368264008, "loss": 4.347, "step": 438700 }, { "epoch": 0.8187562835462644, "grad_norm": 0.9375897645950317, "learning_rate": 0.0001936405627028799, "loss": 4.6244, "step": 438750 }, { "epoch": 0.8188495891056429, "grad_norm": 1.4730111360549927, "learning_rate": 0.00019363908841409986, "loss": 4.6538, "step": 438800 }, { "epoch": 0.8189428946650213, "grad_norm": 0.9291381239891052, "learning_rate": 0.00019363761396006332, "loss": 4.5404, "step": 438850 }, { "epoch": 0.8190362002243998, "grad_norm": 1.0139800310134888, "learning_rate": 0.0001936361393407729, "loss": 4.6218, "step": 438900 }, { "epoch": 0.8191295057837784, "grad_norm": 1.3425135612487793, "learning_rate": 0.00019363466455623113, "loss": 4.5433, "step": 438950 }, { "epoch": 0.8192228113431569, "grad_norm": 0.9264724850654602, "learning_rate": 0.00019363318960644068, "loss": 4.6103, "step": 439000 }, { "epoch": 0.8193161169025354, "grad_norm": 1.0425794124603271, "learning_rate": 0.0001936317144914041, "loss": 4.595, "step": 439050 }, { "epoch": 0.8194094224619138, "grad_norm": 1.0097206830978394, "learning_rate": 0.00019363023921112405, "loss": 4.4658, "step": 439100 }, { "epoch": 0.8195027280212923, "grad_norm": 1.0204837322235107, "learning_rate": 0.00019362876376560306, "loss": 4.4472, "step": 439150 }, { "epoch": 0.8195960335806708, "grad_norm": 1.0780389308929443, "learning_rate": 0.0001936272881548438, "loss": 4.6381, "step": 439200 }, { "epoch": 0.8196893391400493, "grad_norm": 1.4371711015701294, "learning_rate": 0.00019362581237884885, "loss": 4.5083, "step": 439250 }, { "epoch": 0.8197826446994279, "grad_norm": 1.2701109647750854, "learning_rate": 0.0001936243364376208, "loss": 4.6079, "step": 439300 }, { "epoch": 0.8198759502588063, "grad_norm": 1.156437635421753, "learning_rate": 0.00019362286033116233, "loss": 4.6392, "step": 439350 }, { "epoch": 0.8199692558181848, "grad_norm": 1.3134959936141968, "learning_rate": 0.00019362138405947592, "loss": 4.5783, "step": 439400 }, { "epoch": 0.8200625613775633, "grad_norm": 0.8655474781990051, "learning_rate": 0.00019361990762256425, "loss": 4.5355, "step": 439450 }, { "epoch": 0.8201558669369418, "grad_norm": 1.014864206314087, "learning_rate": 0.00019361843102042996, "loss": 4.3339, "step": 439500 }, { "epoch": 0.8202491724963202, "grad_norm": 0.9473825097084045, "learning_rate": 0.00019361695425307556, "loss": 4.6462, "step": 439550 }, { "epoch": 0.8203424780556987, "grad_norm": 1.0384535789489746, "learning_rate": 0.00019361547732050374, "loss": 4.6507, "step": 439600 }, { "epoch": 0.8204357836150773, "grad_norm": 0.7609256505966187, "learning_rate": 0.00019361400022271705, "loss": 4.3934, "step": 439650 }, { "epoch": 0.8205290891744558, "grad_norm": 0.8633242845535278, "learning_rate": 0.00019361252295971813, "loss": 4.441, "step": 439700 }, { "epoch": 0.8206223947338342, "grad_norm": 0.8905895352363586, "learning_rate": 0.0001936110455315096, "loss": 4.562, "step": 439750 }, { "epoch": 0.8207157002932127, "grad_norm": 0.65688157081604, "learning_rate": 0.000193609567938094, "loss": 4.6056, "step": 439800 }, { "epoch": 0.8208090058525912, "grad_norm": 0.895016610622406, "learning_rate": 0.00019360809017947403, "loss": 4.4881, "step": 439850 }, { "epoch": 0.8209023114119697, "grad_norm": 1.1564828157424927, "learning_rate": 0.00019360661225565224, "loss": 4.8174, "step": 439900 }, { "epoch": 0.8209956169713482, "grad_norm": 1.0620310306549072, "learning_rate": 0.00019360513416663124, "loss": 4.648, "step": 439950 }, { "epoch": 0.8210889225307267, "grad_norm": 1.1277257204055786, "learning_rate": 0.00019360365591241364, "loss": 4.4791, "step": 440000 }, { "epoch": 0.8211822280901052, "grad_norm": 1.2287598848342896, "learning_rate": 0.0001936021774930021, "loss": 4.5892, "step": 440050 }, { "epoch": 0.8212755336494837, "grad_norm": 0.9048253297805786, "learning_rate": 0.00019360069890839913, "loss": 4.4444, "step": 440100 }, { "epoch": 0.8213688392088622, "grad_norm": 0.9036170244216919, "learning_rate": 0.00019359922015860744, "loss": 4.7466, "step": 440150 }, { "epoch": 0.8214621447682406, "grad_norm": 1.0846253633499146, "learning_rate": 0.00019359774124362957, "loss": 4.5556, "step": 440200 }, { "epoch": 0.8215554503276191, "grad_norm": 1.0259082317352295, "learning_rate": 0.00019359626216346816, "loss": 4.447, "step": 440250 }, { "epoch": 0.8216487558869976, "grad_norm": 1.0791857242584229, "learning_rate": 0.00019359478291812581, "loss": 4.604, "step": 440300 }, { "epoch": 0.8217420614463762, "grad_norm": 1.085229754447937, "learning_rate": 0.00019359330350760516, "loss": 4.6088, "step": 440350 }, { "epoch": 0.8218353670057547, "grad_norm": 0.9532181620597839, "learning_rate": 0.00019359182393190877, "loss": 4.5741, "step": 440400 }, { "epoch": 0.8219286725651331, "grad_norm": 1.000110149383545, "learning_rate": 0.0001935903441910393, "loss": 4.5447, "step": 440450 }, { "epoch": 0.8220219781245116, "grad_norm": 1.0213303565979004, "learning_rate": 0.00019358886428499935, "loss": 4.7264, "step": 440500 }, { "epoch": 0.8221152836838901, "grad_norm": 1.0686661005020142, "learning_rate": 0.0001935873842137915, "loss": 4.5832, "step": 440550 }, { "epoch": 0.8222085892432686, "grad_norm": 0.8314534425735474, "learning_rate": 0.0001935859039774184, "loss": 4.5014, "step": 440600 }, { "epoch": 0.822301894802647, "grad_norm": 1.1675671339035034, "learning_rate": 0.00019358442357588263, "loss": 4.6003, "step": 440650 }, { "epoch": 0.8223952003620256, "grad_norm": 1.1414916515350342, "learning_rate": 0.00019358294300918683, "loss": 4.4115, "step": 440700 }, { "epoch": 0.8224885059214041, "grad_norm": 1.1694287061691284, "learning_rate": 0.0001935814622773336, "loss": 4.4839, "step": 440750 }, { "epoch": 0.8225818114807826, "grad_norm": 0.8491055369377136, "learning_rate": 0.00019357998138032554, "loss": 4.3969, "step": 440800 }, { "epoch": 0.822675117040161, "grad_norm": 0.9321352243423462, "learning_rate": 0.00019357850031816528, "loss": 4.6168, "step": 440850 }, { "epoch": 0.8227684225995395, "grad_norm": 1.036413550376892, "learning_rate": 0.00019357701909085545, "loss": 4.4682, "step": 440900 }, { "epoch": 0.822861728158918, "grad_norm": 1.0816128253936768, "learning_rate": 0.00019357553769839864, "loss": 4.7283, "step": 440950 }, { "epoch": 0.8229550337182965, "grad_norm": 1.1739145517349243, "learning_rate": 0.00019357405614079748, "loss": 4.47, "step": 441000 }, { "epoch": 0.8229550337182965, "eval_loss": 4.712830066680908, "eval_runtime": 232.585, "eval_samples_per_second": 11.213, "eval_steps_per_second": 11.213, "eval_tts_loss": 7.592319561862758, "step": 441000 }, { "epoch": 0.8230483392776751, "grad_norm": 1.0819491147994995, "learning_rate": 0.00019357257441805455, "loss": 4.574, "step": 441050 }, { "epoch": 0.8231416448370535, "grad_norm": 1.0389968156814575, "learning_rate": 0.00019357109253017254, "loss": 4.5989, "step": 441100 }, { "epoch": 0.823234950396432, "grad_norm": 1.2745685577392578, "learning_rate": 0.00019356961047715395, "loss": 4.4777, "step": 441150 }, { "epoch": 0.8233282559558105, "grad_norm": 1.6084719896316528, "learning_rate": 0.00019356812825900154, "loss": 4.3616, "step": 441200 }, { "epoch": 0.823421561515189, "grad_norm": 1.1243865489959717, "learning_rate": 0.0001935666458757178, "loss": 4.6195, "step": 441250 }, { "epoch": 0.8235148670745674, "grad_norm": 1.3631200790405273, "learning_rate": 0.00019356516332730536, "loss": 4.4103, "step": 441300 }, { "epoch": 0.8236081726339459, "grad_norm": 1.0318950414657593, "learning_rate": 0.00019356368061376693, "loss": 4.4224, "step": 441350 }, { "epoch": 0.8237014781933244, "grad_norm": 1.1379915475845337, "learning_rate": 0.00019356219773510502, "loss": 4.8016, "step": 441400 }, { "epoch": 0.823794783752703, "grad_norm": 0.9971288442611694, "learning_rate": 0.00019356071469132233, "loss": 4.4852, "step": 441450 }, { "epoch": 0.8238880893120815, "grad_norm": 0.6444038152694702, "learning_rate": 0.00019355923148242141, "loss": 4.5769, "step": 441500 }, { "epoch": 0.8239813948714599, "grad_norm": 0.9256520867347717, "learning_rate": 0.00019355774810840494, "loss": 4.5454, "step": 441550 }, { "epoch": 0.8240747004308384, "grad_norm": 1.0571428537368774, "learning_rate": 0.00019355626456927548, "loss": 4.5266, "step": 441600 }, { "epoch": 0.8241680059902169, "grad_norm": 1.3807666301727295, "learning_rate": 0.00019355478086503572, "loss": 4.7166, "step": 441650 }, { "epoch": 0.8242613115495954, "grad_norm": 0.9384042024612427, "learning_rate": 0.00019355329699568817, "loss": 4.5248, "step": 441700 }, { "epoch": 0.8243546171089738, "grad_norm": 0.9789107441902161, "learning_rate": 0.00019355181296123558, "loss": 4.4164, "step": 441750 }, { "epoch": 0.8244479226683524, "grad_norm": 0.6336019039154053, "learning_rate": 0.00019355032876168044, "loss": 4.5301, "step": 441800 }, { "epoch": 0.8245412282277309, "grad_norm": 0.892485499382019, "learning_rate": 0.00019354884439702548, "loss": 4.7402, "step": 441850 }, { "epoch": 0.8246345337871094, "grad_norm": 1.0237869024276733, "learning_rate": 0.00019354735986727324, "loss": 4.5871, "step": 441900 }, { "epoch": 0.8247278393464879, "grad_norm": 1.1821402311325073, "learning_rate": 0.00019354587517242635, "loss": 4.619, "step": 441950 }, { "epoch": 0.8248211449058663, "grad_norm": 1.0676908493041992, "learning_rate": 0.00019354439031248747, "loss": 4.331, "step": 442000 }, { "epoch": 0.8249144504652448, "grad_norm": 0.8039354681968689, "learning_rate": 0.00019354290528745922, "loss": 4.8554, "step": 442050 }, { "epoch": 0.8250077560246233, "grad_norm": 1.2127957344055176, "learning_rate": 0.00019354142009734418, "loss": 4.601, "step": 442100 }, { "epoch": 0.8251010615840019, "grad_norm": 1.4200031757354736, "learning_rate": 0.00019353993474214502, "loss": 4.71, "step": 442150 }, { "epoch": 0.8251943671433803, "grad_norm": 1.0917919874191284, "learning_rate": 0.0001935384492218643, "loss": 4.524, "step": 442200 }, { "epoch": 0.8252876727027588, "grad_norm": 1.008728265762329, "learning_rate": 0.0001935369635365047, "loss": 4.4806, "step": 442250 }, { "epoch": 0.8253809782621373, "grad_norm": 1.1972861289978027, "learning_rate": 0.0001935354776860688, "loss": 4.4702, "step": 442300 }, { "epoch": 0.8254742838215158, "grad_norm": 0.9816621541976929, "learning_rate": 0.00019353399167055925, "loss": 4.6638, "step": 442350 }, { "epoch": 0.8255675893808943, "grad_norm": 1.1558983325958252, "learning_rate": 0.00019353250548997865, "loss": 4.6179, "step": 442400 }, { "epoch": 0.8256608949402727, "grad_norm": 0.9303699731826782, "learning_rate": 0.00019353101914432966, "loss": 4.618, "step": 442450 }, { "epoch": 0.8257542004996513, "grad_norm": 1.1025595664978027, "learning_rate": 0.00019352953263361485, "loss": 4.7275, "step": 442500 }, { "epoch": 0.8258475060590298, "grad_norm": 1.0343772172927856, "learning_rate": 0.0001935280459578369, "loss": 4.5344, "step": 442550 }, { "epoch": 0.8259408116184083, "grad_norm": 0.9679977893829346, "learning_rate": 0.0001935265591169984, "loss": 4.5776, "step": 442600 }, { "epoch": 0.8260341171777867, "grad_norm": 1.3095729351043701, "learning_rate": 0.00019352507211110195, "loss": 4.5481, "step": 442650 }, { "epoch": 0.8261274227371652, "grad_norm": 0.7980509400367737, "learning_rate": 0.00019352358494015024, "loss": 4.3728, "step": 442700 }, { "epoch": 0.8262207282965437, "grad_norm": 1.1941235065460205, "learning_rate": 0.00019352209760414583, "loss": 4.4981, "step": 442750 }, { "epoch": 0.8263140338559222, "grad_norm": 1.2921700477600098, "learning_rate": 0.0001935206101030914, "loss": 4.5507, "step": 442800 }, { "epoch": 0.8264073394153008, "grad_norm": 1.0769251585006714, "learning_rate": 0.00019351912243698952, "loss": 4.5446, "step": 442850 }, { "epoch": 0.8265006449746792, "grad_norm": 0.9345519542694092, "learning_rate": 0.00019351763460584288, "loss": 4.583, "step": 442900 }, { "epoch": 0.8265939505340577, "grad_norm": 1.0591378211975098, "learning_rate": 0.00019351614660965405, "loss": 4.5456, "step": 442950 }, { "epoch": 0.8266872560934362, "grad_norm": 1.0100284814834595, "learning_rate": 0.00019351465844842566, "loss": 4.4957, "step": 443000 }, { "epoch": 0.8267805616528147, "grad_norm": 1.1169507503509521, "learning_rate": 0.00019351317012216037, "loss": 4.4462, "step": 443050 }, { "epoch": 0.8268738672121931, "grad_norm": 1.3374896049499512, "learning_rate": 0.00019351168163086077, "loss": 4.6489, "step": 443100 }, { "epoch": 0.8269671727715716, "grad_norm": 0.8487604856491089, "learning_rate": 0.00019351019297452954, "loss": 4.5124, "step": 443150 }, { "epoch": 0.8270604783309502, "grad_norm": 0.9288005232810974, "learning_rate": 0.00019350870415316926, "loss": 4.8345, "step": 443200 }, { "epoch": 0.8271537838903287, "grad_norm": 1.0807114839553833, "learning_rate": 0.00019350721516678257, "loss": 4.6849, "step": 443250 }, { "epoch": 0.8272470894497072, "grad_norm": 1.2470729351043701, "learning_rate": 0.0001935057260153721, "loss": 4.6842, "step": 443300 }, { "epoch": 0.8273403950090856, "grad_norm": 1.235500454902649, "learning_rate": 0.00019350423669894047, "loss": 4.5028, "step": 443350 }, { "epoch": 0.8274337005684641, "grad_norm": 0.8499687314033508, "learning_rate": 0.00019350274721749032, "loss": 4.4574, "step": 443400 }, { "epoch": 0.8275270061278426, "grad_norm": 1.2059576511383057, "learning_rate": 0.00019350125757102427, "loss": 4.5983, "step": 443450 }, { "epoch": 0.827620311687221, "grad_norm": 0.9806116819381714, "learning_rate": 0.00019349976775954493, "loss": 4.594, "step": 443500 }, { "epoch": 0.8277136172465996, "grad_norm": 0.8075485229492188, "learning_rate": 0.00019349827778305498, "loss": 4.536, "step": 443550 }, { "epoch": 0.8278069228059781, "grad_norm": 0.8367143869400024, "learning_rate": 0.00019349678764155704, "loss": 4.5917, "step": 443600 }, { "epoch": 0.8279002283653566, "grad_norm": 0.9745396375656128, "learning_rate": 0.00019349529733505368, "loss": 4.3854, "step": 443650 }, { "epoch": 0.8279935339247351, "grad_norm": 1.4076462984085083, "learning_rate": 0.00019349380686354759, "loss": 4.539, "step": 443700 }, { "epoch": 0.8280868394841135, "grad_norm": 1.1576460599899292, "learning_rate": 0.00019349231622704135, "loss": 4.5147, "step": 443750 }, { "epoch": 0.828180145043492, "grad_norm": 0.9356547594070435, "learning_rate": 0.00019349082542553767, "loss": 4.8188, "step": 443800 }, { "epoch": 0.8282734506028705, "grad_norm": 0.9233013987541199, "learning_rate": 0.0001934893344590391, "loss": 4.6146, "step": 443850 }, { "epoch": 0.8283667561622491, "grad_norm": 1.172934651374817, "learning_rate": 0.00019348784332754834, "loss": 4.5364, "step": 443900 }, { "epoch": 0.8284600617216276, "grad_norm": 1.0102208852767944, "learning_rate": 0.00019348635203106796, "loss": 4.5929, "step": 443950 }, { "epoch": 0.828553367281006, "grad_norm": 1.102264165878296, "learning_rate": 0.0001934848605696006, "loss": 4.6786, "step": 444000 }, { "epoch": 0.828553367281006, "eval_loss": 4.721865177154541, "eval_runtime": 232.6378, "eval_samples_per_second": 11.211, "eval_steps_per_second": 11.211, "eval_tts_loss": 7.58550966499899, "step": 444000 }, { "epoch": 0.8286466728403845, "grad_norm": 0.9398770332336426, "learning_rate": 0.00019348336894314894, "loss": 4.5993, "step": 444050 }, { "epoch": 0.828739978399763, "grad_norm": 0.9913604259490967, "learning_rate": 0.0001934818771517156, "loss": 4.244, "step": 444100 }, { "epoch": 0.8288332839591415, "grad_norm": 1.2712340354919434, "learning_rate": 0.00019348038519530317, "loss": 4.6613, "step": 444150 }, { "epoch": 0.8289265895185199, "grad_norm": 0.976055920124054, "learning_rate": 0.0001934788930739143, "loss": 4.5837, "step": 444200 }, { "epoch": 0.8290198950778985, "grad_norm": 1.0469322204589844, "learning_rate": 0.00019347740078755164, "loss": 4.2978, "step": 444250 }, { "epoch": 0.829113200637277, "grad_norm": 0.8743473887443542, "learning_rate": 0.00019347590833621782, "loss": 4.439, "step": 444300 }, { "epoch": 0.8292065061966555, "grad_norm": 1.1809208393096924, "learning_rate": 0.00019347441571991548, "loss": 4.388, "step": 444350 }, { "epoch": 0.829299811756034, "grad_norm": 1.614380121231079, "learning_rate": 0.0001934729229386472, "loss": 4.3623, "step": 444400 }, { "epoch": 0.8293931173154124, "grad_norm": 0.9954262971878052, "learning_rate": 0.00019347142999241572, "loss": 4.6601, "step": 444450 }, { "epoch": 0.8294864228747909, "grad_norm": 0.9364942312240601, "learning_rate": 0.00019346993688122358, "loss": 4.693, "step": 444500 }, { "epoch": 0.8295797284341694, "grad_norm": 1.0278654098510742, "learning_rate": 0.00019346844360507344, "loss": 4.6674, "step": 444550 }, { "epoch": 0.829673033993548, "grad_norm": 1.1544060707092285, "learning_rate": 0.00019346695016396795, "loss": 4.4789, "step": 444600 }, { "epoch": 0.8297663395529264, "grad_norm": 1.6491734981536865, "learning_rate": 0.00019346545655790976, "loss": 4.7639, "step": 444650 }, { "epoch": 0.8298596451123049, "grad_norm": 0.6557138562202454, "learning_rate": 0.00019346396278690145, "loss": 4.3905, "step": 444700 }, { "epoch": 0.8299529506716834, "grad_norm": 1.094346523284912, "learning_rate": 0.0001934624688509457, "loss": 4.4474, "step": 444750 }, { "epoch": 0.8300462562310619, "grad_norm": 1.0265722274780273, "learning_rate": 0.00019346097475004515, "loss": 4.5761, "step": 444800 }, { "epoch": 0.8301395617904404, "grad_norm": 1.1581714153289795, "learning_rate": 0.00019345948048420244, "loss": 4.4582, "step": 444850 }, { "epoch": 0.8302328673498188, "grad_norm": 1.1250214576721191, "learning_rate": 0.00019345798605342016, "loss": 4.5792, "step": 444900 }, { "epoch": 0.8303261729091974, "grad_norm": 1.219117283821106, "learning_rate": 0.00019345649145770098, "loss": 4.642, "step": 444950 }, { "epoch": 0.8304194784685759, "grad_norm": 0.9941939115524292, "learning_rate": 0.00019345499669704754, "loss": 4.4172, "step": 445000 }, { "epoch": 0.8305127840279544, "grad_norm": 1.004757046699524, "learning_rate": 0.0001934535017714625, "loss": 4.4208, "step": 445050 }, { "epoch": 0.8306060895873328, "grad_norm": 1.1099998950958252, "learning_rate": 0.00019345200668094846, "loss": 4.6247, "step": 445100 }, { "epoch": 0.8306993951467113, "grad_norm": 1.1278831958770752, "learning_rate": 0.00019345051142550808, "loss": 4.6416, "step": 445150 }, { "epoch": 0.8307927007060898, "grad_norm": 1.3656575679779053, "learning_rate": 0.00019344901600514396, "loss": 4.7044, "step": 445200 }, { "epoch": 0.8308860062654683, "grad_norm": 1.220974326133728, "learning_rate": 0.00019344752041985878, "loss": 4.3872, "step": 445250 }, { "epoch": 0.8309793118248469, "grad_norm": 1.3309504985809326, "learning_rate": 0.00019344602466965518, "loss": 4.5627, "step": 445300 }, { "epoch": 0.8310726173842253, "grad_norm": 1.1383757591247559, "learning_rate": 0.0001934445287545358, "loss": 4.5751, "step": 445350 }, { "epoch": 0.8311659229436038, "grad_norm": 1.1477620601654053, "learning_rate": 0.00019344303267450322, "loss": 4.5198, "step": 445400 }, { "epoch": 0.8312592285029823, "grad_norm": 0.9861552119255066, "learning_rate": 0.00019344153642956018, "loss": 4.247, "step": 445450 }, { "epoch": 0.8313525340623608, "grad_norm": 0.9693005084991455, "learning_rate": 0.00019344004001970924, "loss": 4.6788, "step": 445500 }, { "epoch": 0.8314458396217392, "grad_norm": 1.1747108697891235, "learning_rate": 0.0001934385434449531, "loss": 4.5564, "step": 445550 }, { "epoch": 0.8315391451811177, "grad_norm": 0.9544690251350403, "learning_rate": 0.00019343704670529433, "loss": 4.4388, "step": 445600 }, { "epoch": 0.8316324507404963, "grad_norm": 1.0575664043426514, "learning_rate": 0.00019343554980073567, "loss": 4.4988, "step": 445650 }, { "epoch": 0.8317257562998748, "grad_norm": 1.320414662361145, "learning_rate": 0.00019343405273127966, "loss": 4.6764, "step": 445700 }, { "epoch": 0.8318190618592533, "grad_norm": 1.0772422552108765, "learning_rate": 0.000193432555496929, "loss": 4.5405, "step": 445750 }, { "epoch": 0.8319123674186317, "grad_norm": 0.8643651604652405, "learning_rate": 0.00019343105809768631, "loss": 4.3712, "step": 445800 }, { "epoch": 0.8320056729780102, "grad_norm": 1.0002784729003906, "learning_rate": 0.00019342956053355429, "loss": 4.6529, "step": 445850 }, { "epoch": 0.8320989785373887, "grad_norm": 1.0368218421936035, "learning_rate": 0.00019342806280453546, "loss": 4.5132, "step": 445900 }, { "epoch": 0.8321922840967672, "grad_norm": 1.2993515729904175, "learning_rate": 0.0001934265649106326, "loss": 4.7953, "step": 445950 }, { "epoch": 0.8322855896561457, "grad_norm": 0.9636172652244568, "learning_rate": 0.00019342506685184825, "loss": 4.433, "step": 446000 }, { "epoch": 0.8323788952155242, "grad_norm": 1.0712823867797852, "learning_rate": 0.00019342356862818511, "loss": 4.4565, "step": 446050 }, { "epoch": 0.8324722007749027, "grad_norm": 1.0290355682373047, "learning_rate": 0.00019342207023964584, "loss": 4.5186, "step": 446100 }, { "epoch": 0.8325655063342812, "grad_norm": 0.8416425585746765, "learning_rate": 0.000193420571686233, "loss": 4.6492, "step": 446150 }, { "epoch": 0.8326588118936596, "grad_norm": 1.026302695274353, "learning_rate": 0.00019341907296794932, "loss": 4.4113, "step": 446200 }, { "epoch": 0.8327521174530381, "grad_norm": 1.206878423690796, "learning_rate": 0.00019341757408479742, "loss": 4.4899, "step": 446250 }, { "epoch": 0.8328454230124166, "grad_norm": 0.9927154779434204, "learning_rate": 0.00019341607503677992, "loss": 4.6825, "step": 446300 }, { "epoch": 0.8329387285717952, "grad_norm": 1.098044753074646, "learning_rate": 0.0001934145758238995, "loss": 4.584, "step": 446350 }, { "epoch": 0.8330320341311737, "grad_norm": 1.096203088760376, "learning_rate": 0.00019341307644615878, "loss": 4.5713, "step": 446400 }, { "epoch": 0.8331253396905521, "grad_norm": 1.0846002101898193, "learning_rate": 0.00019341157690356043, "loss": 4.5434, "step": 446450 }, { "epoch": 0.8332186452499306, "grad_norm": 1.1647289991378784, "learning_rate": 0.00019341007719610707, "loss": 4.4887, "step": 446500 }, { "epoch": 0.8333119508093091, "grad_norm": 1.1023021936416626, "learning_rate": 0.00019340857732380134, "loss": 4.4969, "step": 446550 }, { "epoch": 0.8334052563686876, "grad_norm": 0.9889746308326721, "learning_rate": 0.00019340707728664596, "loss": 4.468, "step": 446600 }, { "epoch": 0.833498561928066, "grad_norm": 1.0913546085357666, "learning_rate": 0.00019340557708464348, "loss": 4.5224, "step": 446650 }, { "epoch": 0.8335918674874445, "grad_norm": 1.1410549879074097, "learning_rate": 0.0001934040767177966, "loss": 4.5776, "step": 446700 }, { "epoch": 0.8336851730468231, "grad_norm": 0.9060096740722656, "learning_rate": 0.00019340257618610798, "loss": 4.5885, "step": 446750 }, { "epoch": 0.8337784786062016, "grad_norm": 1.1176276206970215, "learning_rate": 0.00019340107548958022, "loss": 4.4966, "step": 446800 }, { "epoch": 0.83387178416558, "grad_norm": 1.2471281290054321, "learning_rate": 0.00019339957462821602, "loss": 4.4346, "step": 446850 }, { "epoch": 0.8339650897249585, "grad_norm": 1.1243484020233154, "learning_rate": 0.00019339807360201802, "loss": 4.7867, "step": 446900 }, { "epoch": 0.834058395284337, "grad_norm": 1.058711051940918, "learning_rate": 0.00019339657241098878, "loss": 4.8493, "step": 446950 }, { "epoch": 0.8341517008437155, "grad_norm": 1.22663152217865, "learning_rate": 0.0001933950710551311, "loss": 4.4856, "step": 447000 }, { "epoch": 0.8341517008437155, "eval_loss": 4.712674140930176, "eval_runtime": 232.2896, "eval_samples_per_second": 11.227, "eval_steps_per_second": 11.227, "eval_tts_loss": 7.630600184322181, "step": 447000 }, { "epoch": 0.834245006403094, "grad_norm": 1.2073787450790405, "learning_rate": 0.00019339356953444753, "loss": 4.6652, "step": 447050 }, { "epoch": 0.8343383119624725, "grad_norm": 1.0807472467422485, "learning_rate": 0.0001933920678489407, "loss": 4.6302, "step": 447100 }, { "epoch": 0.834431617521851, "grad_norm": 0.8505199551582336, "learning_rate": 0.00019339056599861336, "loss": 4.6837, "step": 447150 }, { "epoch": 0.8345249230812295, "grad_norm": 1.016205906867981, "learning_rate": 0.00019338906398346807, "loss": 4.6417, "step": 447200 }, { "epoch": 0.834618228640608, "grad_norm": 1.0800621509552002, "learning_rate": 0.00019338756180350755, "loss": 4.5582, "step": 447250 }, { "epoch": 0.8347115341999864, "grad_norm": 0.9196884036064148, "learning_rate": 0.0001933860594587344, "loss": 4.4846, "step": 447300 }, { "epoch": 0.8348048397593649, "grad_norm": 1.1872416734695435, "learning_rate": 0.00019338455694915127, "loss": 4.5343, "step": 447350 }, { "epoch": 0.8348981453187434, "grad_norm": 1.0017876625061035, "learning_rate": 0.00019338305427476083, "loss": 4.3577, "step": 447400 }, { "epoch": 0.834991450878122, "grad_norm": 1.4030224084854126, "learning_rate": 0.00019338155143556573, "loss": 4.6123, "step": 447450 }, { "epoch": 0.8350847564375005, "grad_norm": 1.8109701871871948, "learning_rate": 0.00019338004843156865, "loss": 4.5135, "step": 447500 }, { "epoch": 0.8351780619968789, "grad_norm": 1.2797409296035767, "learning_rate": 0.00019337854526277218, "loss": 4.4158, "step": 447550 }, { "epoch": 0.8352713675562574, "grad_norm": 0.9986948370933533, "learning_rate": 0.00019337704192917904, "loss": 4.6389, "step": 447600 }, { "epoch": 0.8353646731156359, "grad_norm": 1.0423827171325684, "learning_rate": 0.00019337553843079184, "loss": 4.4192, "step": 447650 }, { "epoch": 0.8354579786750144, "grad_norm": 0.9322047233581543, "learning_rate": 0.00019337403476761325, "loss": 4.5607, "step": 447700 }, { "epoch": 0.8355512842343928, "grad_norm": 1.076505422592163, "learning_rate": 0.0001933725309396459, "loss": 4.8305, "step": 447750 }, { "epoch": 0.8356445897937714, "grad_norm": 1.1497913599014282, "learning_rate": 0.00019337102694689246, "loss": 4.4388, "step": 447800 }, { "epoch": 0.8357378953531499, "grad_norm": 1.237391710281372, "learning_rate": 0.00019336952278935563, "loss": 4.6497, "step": 447850 }, { "epoch": 0.8358312009125284, "grad_norm": 1.1664336919784546, "learning_rate": 0.00019336801846703798, "loss": 4.4486, "step": 447900 }, { "epoch": 0.8359245064719069, "grad_norm": 1.0214166641235352, "learning_rate": 0.0001933665139799422, "loss": 4.6264, "step": 447950 }, { "epoch": 0.8360178120312853, "grad_norm": 1.5139281749725342, "learning_rate": 0.00019336500932807098, "loss": 4.7044, "step": 448000 }, { "epoch": 0.8361111175906638, "grad_norm": 1.286423683166504, "learning_rate": 0.00019336350451142694, "loss": 4.4449, "step": 448050 }, { "epoch": 0.8362044231500423, "grad_norm": 0.8456434011459351, "learning_rate": 0.00019336199953001273, "loss": 4.3987, "step": 448100 }, { "epoch": 0.8362977287094209, "grad_norm": 0.9992003440856934, "learning_rate": 0.00019336049438383103, "loss": 4.6912, "step": 448150 }, { "epoch": 0.8363910342687993, "grad_norm": 1.1905733346939087, "learning_rate": 0.00019335898907288447, "loss": 4.6875, "step": 448200 }, { "epoch": 0.8364843398281778, "grad_norm": 0.9549011588096619, "learning_rate": 0.00019335748359717574, "loss": 4.6345, "step": 448250 }, { "epoch": 0.8365776453875563, "grad_norm": 1.0515804290771484, "learning_rate": 0.00019335597795670746, "loss": 4.4891, "step": 448300 }, { "epoch": 0.8366709509469348, "grad_norm": 1.0489537715911865, "learning_rate": 0.0001933544721514823, "loss": 4.7241, "step": 448350 }, { "epoch": 0.8367642565063133, "grad_norm": 1.0642502307891846, "learning_rate": 0.00019335296618150296, "loss": 4.6523, "step": 448400 }, { "epoch": 0.8368575620656917, "grad_norm": 1.2421493530273438, "learning_rate": 0.00019335146004677203, "loss": 4.5251, "step": 448450 }, { "epoch": 0.8369508676250703, "grad_norm": 1.2859892845153809, "learning_rate": 0.0001933499537472922, "loss": 4.5942, "step": 448500 }, { "epoch": 0.8370441731844488, "grad_norm": 1.2976332902908325, "learning_rate": 0.00019334844728306614, "loss": 4.4986, "step": 448550 }, { "epoch": 0.8371374787438273, "grad_norm": 0.7612985372543335, "learning_rate": 0.00019334694065409647, "loss": 4.601, "step": 448600 }, { "epoch": 0.8372307843032057, "grad_norm": 1.1025867462158203, "learning_rate": 0.00019334543386038592, "loss": 4.4959, "step": 448650 }, { "epoch": 0.8373240898625842, "grad_norm": 1.1251939535140991, "learning_rate": 0.00019334392690193708, "loss": 4.6645, "step": 448700 }, { "epoch": 0.8374173954219627, "grad_norm": 1.149271845817566, "learning_rate": 0.0001933424197787526, "loss": 4.5686, "step": 448750 }, { "epoch": 0.8375107009813412, "grad_norm": 1.1676266193389893, "learning_rate": 0.00019334091249083524, "loss": 4.5399, "step": 448800 }, { "epoch": 0.8376040065407198, "grad_norm": 0.7529256343841553, "learning_rate": 0.00019333940503818756, "loss": 4.4607, "step": 448850 }, { "epoch": 0.8376973121000982, "grad_norm": 1.1759973764419556, "learning_rate": 0.00019333789742081222, "loss": 4.3374, "step": 448900 }, { "epoch": 0.8377906176594767, "grad_norm": 1.1348012685775757, "learning_rate": 0.00019333638963871198, "loss": 4.5555, "step": 448950 }, { "epoch": 0.8378839232188552, "grad_norm": 1.3209272623062134, "learning_rate": 0.00019333488169188937, "loss": 4.5982, "step": 449000 }, { "epoch": 0.8379772287782337, "grad_norm": 1.3228648900985718, "learning_rate": 0.00019333337358034715, "loss": 4.4296, "step": 449050 }, { "epoch": 0.8380705343376121, "grad_norm": 1.0187352895736694, "learning_rate": 0.00019333186530408797, "loss": 4.7661, "step": 449100 }, { "epoch": 0.8381638398969906, "grad_norm": 1.0453397035598755, "learning_rate": 0.00019333035686311444, "loss": 4.5145, "step": 449150 }, { "epoch": 0.8382571454563692, "grad_norm": 1.3582311868667603, "learning_rate": 0.00019332884825742924, "loss": 4.6269, "step": 449200 }, { "epoch": 0.8383504510157477, "grad_norm": 1.093062400817871, "learning_rate": 0.00019332733948703506, "loss": 4.619, "step": 449250 }, { "epoch": 0.8384437565751262, "grad_norm": 0.9017040133476257, "learning_rate": 0.00019332583055193454, "loss": 4.6322, "step": 449300 }, { "epoch": 0.8385370621345046, "grad_norm": 1.103871464729309, "learning_rate": 0.00019332432145213038, "loss": 4.5635, "step": 449350 }, { "epoch": 0.8386303676938831, "grad_norm": 1.134921908378601, "learning_rate": 0.0001933228121876252, "loss": 4.4465, "step": 449400 }, { "epoch": 0.8387236732532616, "grad_norm": 1.078826665878296, "learning_rate": 0.00019332130275842167, "loss": 4.466, "step": 449450 }, { "epoch": 0.8388169788126401, "grad_norm": 1.0170928239822388, "learning_rate": 0.00019331979316452243, "loss": 4.4511, "step": 449500 }, { "epoch": 0.8389102843720186, "grad_norm": 1.1686103343963623, "learning_rate": 0.0001933182834059302, "loss": 4.4879, "step": 449550 }, { "epoch": 0.8390035899313971, "grad_norm": 1.040514349937439, "learning_rate": 0.00019331677348264766, "loss": 4.5254, "step": 449600 }, { "epoch": 0.8390968954907756, "grad_norm": 0.905274510383606, "learning_rate": 0.0001933152633946774, "loss": 4.6875, "step": 449650 }, { "epoch": 0.8391902010501541, "grad_norm": 0.9785981774330139, "learning_rate": 0.0001933137531420221, "loss": 4.6597, "step": 449700 }, { "epoch": 0.8392835066095325, "grad_norm": 1.2745859622955322, "learning_rate": 0.00019331224272468446, "loss": 4.4091, "step": 449750 }, { "epoch": 0.839376812168911, "grad_norm": 1.7750924825668335, "learning_rate": 0.00019331073214266716, "loss": 4.564, "step": 449800 }, { "epoch": 0.8394701177282895, "grad_norm": 1.0283859968185425, "learning_rate": 0.0001933092213959728, "loss": 4.5776, "step": 449850 }, { "epoch": 0.8395634232876681, "grad_norm": 1.2207344770431519, "learning_rate": 0.0001933077104846041, "loss": 4.6865, "step": 449900 }, { "epoch": 0.8396567288470466, "grad_norm": 1.1184113025665283, "learning_rate": 0.00019330619940856369, "loss": 4.6266, "step": 449950 }, { "epoch": 0.839750034406425, "grad_norm": 1.1503829956054688, "learning_rate": 0.00019330468816785426, "loss": 4.5217, "step": 450000 }, { "epoch": 0.839750034406425, "eval_loss": 4.710883617401123, "eval_runtime": 232.9307, "eval_samples_per_second": 11.196, "eval_steps_per_second": 11.196, "eval_tts_loss": 7.568327237897627, "step": 450000 }, { "epoch": 0.8398433399658035, "grad_norm": 1.1530593633651733, "learning_rate": 0.0001933031767624785, "loss": 4.6247, "step": 450050 }, { "epoch": 0.839936645525182, "grad_norm": 0.7560440301895142, "learning_rate": 0.00019330166519243905, "loss": 4.5636, "step": 450100 }, { "epoch": 0.8400299510845605, "grad_norm": 1.126522183418274, "learning_rate": 0.00019330015345773855, "loss": 4.6178, "step": 450150 }, { "epoch": 0.8401232566439389, "grad_norm": 1.2579351663589478, "learning_rate": 0.0001932986415583797, "loss": 4.8406, "step": 450200 }, { "epoch": 0.8402165622033175, "grad_norm": 1.1700242757797241, "learning_rate": 0.00019329712949436517, "loss": 4.5789, "step": 450250 }, { "epoch": 0.840309867762696, "grad_norm": 1.1610465049743652, "learning_rate": 0.00019329561726569763, "loss": 4.7092, "step": 450300 }, { "epoch": 0.8404031733220745, "grad_norm": 0.921472430229187, "learning_rate": 0.0001932941048723797, "loss": 4.4653, "step": 450350 }, { "epoch": 0.840496478881453, "grad_norm": 1.2933341264724731, "learning_rate": 0.00019329259231441414, "loss": 4.5035, "step": 450400 }, { "epoch": 0.8405897844408314, "grad_norm": 0.9860426187515259, "learning_rate": 0.00019329107959180358, "loss": 4.5394, "step": 450450 }, { "epoch": 0.8406830900002099, "grad_norm": 1.5691264867782593, "learning_rate": 0.00019328956670455064, "loss": 4.4438, "step": 450500 }, { "epoch": 0.8407763955595884, "grad_norm": 1.0524669885635376, "learning_rate": 0.00019328805365265806, "loss": 4.5818, "step": 450550 }, { "epoch": 0.840869701118967, "grad_norm": 1.1248055696487427, "learning_rate": 0.00019328654043612844, "loss": 4.6727, "step": 450600 }, { "epoch": 0.8409630066783454, "grad_norm": 0.8522668480873108, "learning_rate": 0.00019328502705496451, "loss": 4.4992, "step": 450650 }, { "epoch": 0.8410563122377239, "grad_norm": 0.8684114217758179, "learning_rate": 0.00019328351350916893, "loss": 4.5486, "step": 450700 }, { "epoch": 0.8411496177971024, "grad_norm": 1.2275203466415405, "learning_rate": 0.00019328199979874437, "loss": 4.5316, "step": 450750 }, { "epoch": 0.8412429233564809, "grad_norm": 1.0412169694900513, "learning_rate": 0.00019328048592369345, "loss": 4.2662, "step": 450800 }, { "epoch": 0.8413362289158594, "grad_norm": 0.8822068572044373, "learning_rate": 0.00019327897188401894, "loss": 4.6678, "step": 450850 }, { "epoch": 0.8414295344752378, "grad_norm": 1.0500564575195312, "learning_rate": 0.00019327745767972344, "loss": 4.6266, "step": 450900 }, { "epoch": 0.8415228400346164, "grad_norm": 1.1681212186813354, "learning_rate": 0.00019327594331080961, "loss": 4.5943, "step": 450950 }, { "epoch": 0.8416161455939949, "grad_norm": 1.0938808917999268, "learning_rate": 0.00019327442877728018, "loss": 4.5703, "step": 451000 }, { "epoch": 0.8417094511533734, "grad_norm": 0.9540387988090515, "learning_rate": 0.0001932729140791378, "loss": 4.5122, "step": 451050 }, { "epoch": 0.8418027567127518, "grad_norm": 0.9375895857810974, "learning_rate": 0.00019327139921638514, "loss": 4.5586, "step": 451100 }, { "epoch": 0.8418960622721303, "grad_norm": 1.1446915864944458, "learning_rate": 0.00019326988418902485, "loss": 4.5502, "step": 451150 }, { "epoch": 0.8419893678315088, "grad_norm": 1.0916074514389038, "learning_rate": 0.00019326836899705965, "loss": 4.654, "step": 451200 }, { "epoch": 0.8420826733908873, "grad_norm": 1.0063576698303223, "learning_rate": 0.00019326685364049219, "loss": 4.4077, "step": 451250 }, { "epoch": 0.8421759789502659, "grad_norm": 1.0716850757598877, "learning_rate": 0.0001932653381193251, "loss": 4.6463, "step": 451300 }, { "epoch": 0.8422692845096443, "grad_norm": 1.2872897386550903, "learning_rate": 0.00019326382243356114, "loss": 4.4904, "step": 451350 }, { "epoch": 0.8423625900690228, "grad_norm": 0.8096011877059937, "learning_rate": 0.00019326230658320292, "loss": 4.6754, "step": 451400 }, { "epoch": 0.8424558956284013, "grad_norm": 1.3478927612304688, "learning_rate": 0.00019326079056825317, "loss": 4.5584, "step": 451450 }, { "epoch": 0.8425492011877798, "grad_norm": 1.3702021837234497, "learning_rate": 0.0001932592743887145, "loss": 4.5173, "step": 451500 }, { "epoch": 0.8426425067471582, "grad_norm": 0.9578404426574707, "learning_rate": 0.00019325775804458965, "loss": 4.7119, "step": 451550 }, { "epoch": 0.8427358123065367, "grad_norm": 1.2032324075698853, "learning_rate": 0.00019325624153588127, "loss": 4.5057, "step": 451600 }, { "epoch": 0.8428291178659153, "grad_norm": 1.288559079170227, "learning_rate": 0.000193254724862592, "loss": 4.5697, "step": 451650 }, { "epoch": 0.8429224234252938, "grad_norm": 0.963632345199585, "learning_rate": 0.00019325320802472456, "loss": 4.4117, "step": 451700 }, { "epoch": 0.8430157289846723, "grad_norm": 1.1363961696624756, "learning_rate": 0.00019325169102228166, "loss": 4.5567, "step": 451750 }, { "epoch": 0.8431090345440507, "grad_norm": 1.131595492362976, "learning_rate": 0.00019325017385526586, "loss": 4.3638, "step": 451800 }, { "epoch": 0.8432023401034292, "grad_norm": 0.9933236241340637, "learning_rate": 0.00019324865652367996, "loss": 4.4048, "step": 451850 }, { "epoch": 0.8432956456628077, "grad_norm": 0.9534032344818115, "learning_rate": 0.00019324713902752657, "loss": 4.5009, "step": 451900 }, { "epoch": 0.8433889512221862, "grad_norm": 1.0373858213424683, "learning_rate": 0.00019324562136680839, "loss": 4.7383, "step": 451950 }, { "epoch": 0.8434822567815646, "grad_norm": 0.8199633955955505, "learning_rate": 0.00019324410354152808, "loss": 4.5314, "step": 452000 }, { "epoch": 0.8435755623409432, "grad_norm": 1.0837599039077759, "learning_rate": 0.00019324258555168837, "loss": 4.5894, "step": 452050 }, { "epoch": 0.8436688679003217, "grad_norm": 1.2364952564239502, "learning_rate": 0.00019324106739729188, "loss": 4.6582, "step": 452100 }, { "epoch": 0.8437621734597002, "grad_norm": 0.7789540886878967, "learning_rate": 0.00019323954907834133, "loss": 4.5806, "step": 452150 }, { "epoch": 0.8438554790190786, "grad_norm": 1.2450273036956787, "learning_rate": 0.00019323803059483936, "loss": 4.6015, "step": 452200 }, { "epoch": 0.8439487845784571, "grad_norm": 1.2874577045440674, "learning_rate": 0.00019323651194678868, "loss": 4.5458, "step": 452250 }, { "epoch": 0.8440420901378356, "grad_norm": 1.3656573295593262, "learning_rate": 0.00019323499313419196, "loss": 4.6391, "step": 452300 }, { "epoch": 0.8441353956972141, "grad_norm": 1.064369797706604, "learning_rate": 0.00019323347415705188, "loss": 4.4416, "step": 452350 }, { "epoch": 0.8442287012565927, "grad_norm": 1.1734198331832886, "learning_rate": 0.00019323195501537115, "loss": 4.362, "step": 452400 }, { "epoch": 0.8443220068159711, "grad_norm": 1.1021287441253662, "learning_rate": 0.00019323043570915237, "loss": 4.678, "step": 452450 }, { "epoch": 0.8444153123753496, "grad_norm": 1.1119120121002197, "learning_rate": 0.00019322891623839834, "loss": 4.7595, "step": 452500 }, { "epoch": 0.8445086179347281, "grad_norm": 1.0412399768829346, "learning_rate": 0.00019322739660311165, "loss": 4.6088, "step": 452550 }, { "epoch": 0.8446019234941066, "grad_norm": 0.7403721213340759, "learning_rate": 0.00019322587680329497, "loss": 4.5932, "step": 452600 }, { "epoch": 0.844695229053485, "grad_norm": 1.14961576461792, "learning_rate": 0.00019322435683895108, "loss": 4.5157, "step": 452650 }, { "epoch": 0.8447885346128635, "grad_norm": 1.1009234189987183, "learning_rate": 0.00019322283671008257, "loss": 4.5013, "step": 452700 }, { "epoch": 0.8448818401722421, "grad_norm": 0.8400731086730957, "learning_rate": 0.00019322131641669218, "loss": 4.5795, "step": 452750 }, { "epoch": 0.8449751457316206, "grad_norm": 1.0518465042114258, "learning_rate": 0.00019321979595878254, "loss": 4.6606, "step": 452800 }, { "epoch": 0.8450684512909991, "grad_norm": 1.2242203950881958, "learning_rate": 0.00019321827533635638, "loss": 4.4912, "step": 452850 }, { "epoch": 0.8451617568503775, "grad_norm": 1.163169026374817, "learning_rate": 0.00019321675454941634, "loss": 4.5729, "step": 452900 }, { "epoch": 0.845255062409756, "grad_norm": 1.1667548418045044, "learning_rate": 0.00019321523359796516, "loss": 4.5619, "step": 452950 }, { "epoch": 0.8453483679691345, "grad_norm": 1.4125514030456543, "learning_rate": 0.00019321371248200552, "loss": 4.5351, "step": 453000 }, { "epoch": 0.8453483679691345, "eval_loss": 4.710817813873291, "eval_runtime": 233.9958, "eval_samples_per_second": 11.145, "eval_steps_per_second": 11.145, "eval_tts_loss": 7.61211348108753, "step": 453000 }, { "epoch": 0.845441673528513, "grad_norm": 1.372028112411499, "learning_rate": 0.00019321219120154, "loss": 4.5285, "step": 453050 }, { "epoch": 0.8455349790878915, "grad_norm": 1.0454567670822144, "learning_rate": 0.00019321066975657144, "loss": 4.4837, "step": 453100 }, { "epoch": 0.84562828464727, "grad_norm": 0.9676375985145569, "learning_rate": 0.00019320914814710243, "loss": 4.4702, "step": 453150 }, { "epoch": 0.8457215902066485, "grad_norm": 1.0043405294418335, "learning_rate": 0.00019320762637313566, "loss": 4.6094, "step": 453200 }, { "epoch": 0.845814895766027, "grad_norm": 0.9126599431037903, "learning_rate": 0.00019320610443467382, "loss": 4.3521, "step": 453250 }, { "epoch": 0.8459082013254055, "grad_norm": 1.2024579048156738, "learning_rate": 0.00019320458233171962, "loss": 4.3249, "step": 453300 }, { "epoch": 0.8460015068847839, "grad_norm": 1.1454980373382568, "learning_rate": 0.00019320306006427574, "loss": 4.4368, "step": 453350 }, { "epoch": 0.8460948124441624, "grad_norm": 0.9680221080780029, "learning_rate": 0.00019320153763234483, "loss": 4.6728, "step": 453400 }, { "epoch": 0.846188118003541, "grad_norm": 0.8100181221961975, "learning_rate": 0.00019320001503592965, "loss": 4.4589, "step": 453450 }, { "epoch": 0.8462814235629195, "grad_norm": 1.1991851329803467, "learning_rate": 0.0001931984922750328, "loss": 4.637, "step": 453500 }, { "epoch": 0.8463747291222979, "grad_norm": 1.2254598140716553, "learning_rate": 0.00019319696934965702, "loss": 4.6401, "step": 453550 }, { "epoch": 0.8464680346816764, "grad_norm": 0.8437310457229614, "learning_rate": 0.000193195446259805, "loss": 4.661, "step": 453600 }, { "epoch": 0.8465613402410549, "grad_norm": 1.3140015602111816, "learning_rate": 0.00019319392300547943, "loss": 4.6409, "step": 453650 }, { "epoch": 0.8466546458004334, "grad_norm": 1.2739115953445435, "learning_rate": 0.00019319239958668293, "loss": 4.6791, "step": 453700 }, { "epoch": 0.8467479513598118, "grad_norm": 1.2224656343460083, "learning_rate": 0.0001931908760034183, "loss": 4.5679, "step": 453750 }, { "epoch": 0.8468412569191904, "grad_norm": 1.1305352449417114, "learning_rate": 0.00019318935225568816, "loss": 4.4716, "step": 453800 }, { "epoch": 0.8469345624785689, "grad_norm": 1.1846295595169067, "learning_rate": 0.0001931878283434952, "loss": 4.5026, "step": 453850 }, { "epoch": 0.8470278680379474, "grad_norm": 0.8586726188659668, "learning_rate": 0.0001931863042668421, "loss": 4.4602, "step": 453900 }, { "epoch": 0.8471211735973259, "grad_norm": 1.3571677207946777, "learning_rate": 0.00019318478002573159, "loss": 4.439, "step": 453950 }, { "epoch": 0.8472144791567043, "grad_norm": 1.2450292110443115, "learning_rate": 0.00019318325562016635, "loss": 4.5915, "step": 454000 }, { "epoch": 0.8473077847160828, "grad_norm": 1.3634917736053467, "learning_rate": 0.00019318173105014902, "loss": 4.5351, "step": 454050 }, { "epoch": 0.8474010902754613, "grad_norm": 0.88820481300354, "learning_rate": 0.00019318020631568236, "loss": 4.4644, "step": 454100 }, { "epoch": 0.8474943958348399, "grad_norm": 1.2433198690414429, "learning_rate": 0.00019317868141676905, "loss": 4.5574, "step": 454150 }, { "epoch": 0.8475877013942184, "grad_norm": 1.1844819784164429, "learning_rate": 0.00019317715635341171, "loss": 4.4774, "step": 454200 }, { "epoch": 0.8476810069535968, "grad_norm": 0.9785028696060181, "learning_rate": 0.0001931756311256131, "loss": 4.64, "step": 454250 }, { "epoch": 0.8477743125129753, "grad_norm": 0.9226474165916443, "learning_rate": 0.00019317410573337594, "loss": 4.5598, "step": 454300 }, { "epoch": 0.8478676180723538, "grad_norm": 0.9986966848373413, "learning_rate": 0.00019317258017670285, "loss": 4.5249, "step": 454350 }, { "epoch": 0.8479609236317323, "grad_norm": 1.232153296470642, "learning_rate": 0.00019317105445559656, "loss": 4.4745, "step": 454400 }, { "epoch": 0.8480542291911107, "grad_norm": 1.3236726522445679, "learning_rate": 0.00019316952857005974, "loss": 4.5652, "step": 454450 }, { "epoch": 0.8481475347504893, "grad_norm": 1.0787049531936646, "learning_rate": 0.00019316800252009507, "loss": 4.5882, "step": 454500 }, { "epoch": 0.8482408403098678, "grad_norm": 1.0206072330474854, "learning_rate": 0.0001931664763057053, "loss": 4.5725, "step": 454550 }, { "epoch": 0.8483341458692463, "grad_norm": 0.9839275479316711, "learning_rate": 0.00019316494992689312, "loss": 4.3779, "step": 454600 }, { "epoch": 0.8484274514286247, "grad_norm": 1.3855035305023193, "learning_rate": 0.00019316342338366116, "loss": 4.6458, "step": 454650 }, { "epoch": 0.8485207569880032, "grad_norm": 0.9846870303153992, "learning_rate": 0.00019316189667601216, "loss": 4.3942, "step": 454700 }, { "epoch": 0.8486140625473817, "grad_norm": 0.9406638741493225, "learning_rate": 0.0001931603698039488, "loss": 4.4591, "step": 454750 }, { "epoch": 0.8487073681067602, "grad_norm": 0.9582647681236267, "learning_rate": 0.00019315884276747378, "loss": 4.4388, "step": 454800 }, { "epoch": 0.8488006736661388, "grad_norm": 1.097686767578125, "learning_rate": 0.0001931573155665898, "loss": 4.464, "step": 454850 }, { "epoch": 0.8488939792255172, "grad_norm": 1.1097980737686157, "learning_rate": 0.00019315578820129953, "loss": 4.3796, "step": 454900 }, { "epoch": 0.8489872847848957, "grad_norm": 0.9290542602539062, "learning_rate": 0.00019315426067160568, "loss": 4.6605, "step": 454950 }, { "epoch": 0.8490805903442742, "grad_norm": 1.1880428791046143, "learning_rate": 0.000193152732977511, "loss": 4.6119, "step": 455000 }, { "epoch": 0.8491738959036527, "grad_norm": 1.0108721256256104, "learning_rate": 0.00019315120511901812, "loss": 4.6721, "step": 455050 }, { "epoch": 0.8492672014630311, "grad_norm": 1.0492184162139893, "learning_rate": 0.00019314967709612972, "loss": 4.5184, "step": 455100 }, { "epoch": 0.8493605070224096, "grad_norm": 0.9968841075897217, "learning_rate": 0.00019314814890884855, "loss": 4.6032, "step": 455150 }, { "epoch": 0.8494538125817882, "grad_norm": 1.4183242321014404, "learning_rate": 0.00019314662055717728, "loss": 4.4636, "step": 455200 }, { "epoch": 0.8495471181411667, "grad_norm": 1.1357654333114624, "learning_rate": 0.00019314509204111863, "loss": 4.4497, "step": 455250 }, { "epoch": 0.8496404237005452, "grad_norm": 1.1006675958633423, "learning_rate": 0.0001931435633606753, "loss": 4.3582, "step": 455300 }, { "epoch": 0.8497337292599236, "grad_norm": 1.1002415418624878, "learning_rate": 0.00019314203451584994, "loss": 4.5873, "step": 455350 }, { "epoch": 0.8498270348193021, "grad_norm": 1.0014864206314087, "learning_rate": 0.00019314050550664526, "loss": 4.2742, "step": 455400 }, { "epoch": 0.8499203403786806, "grad_norm": 0.887138843536377, "learning_rate": 0.000193138976333064, "loss": 4.7558, "step": 455450 }, { "epoch": 0.8500136459380591, "grad_norm": 0.9670901894569397, "learning_rate": 0.00019313744699510883, "loss": 4.4171, "step": 455500 }, { "epoch": 0.8501069514974376, "grad_norm": 0.9820638298988342, "learning_rate": 0.00019313591749278247, "loss": 4.355, "step": 455550 }, { "epoch": 0.8502002570568161, "grad_norm": 1.3597747087478638, "learning_rate": 0.00019313438782608757, "loss": 4.6693, "step": 455600 }, { "epoch": 0.8502935626161946, "grad_norm": 1.0340396165847778, "learning_rate": 0.0001931328579950269, "loss": 4.695, "step": 455650 }, { "epoch": 0.8503868681755731, "grad_norm": 1.1341750621795654, "learning_rate": 0.00019313132799960312, "loss": 4.4944, "step": 455700 }, { "epoch": 0.8504801737349516, "grad_norm": 1.0759532451629639, "learning_rate": 0.0001931297978398189, "loss": 4.6385, "step": 455750 }, { "epoch": 0.85057347929433, "grad_norm": 1.6166722774505615, "learning_rate": 0.00019312826751567702, "loss": 4.6034, "step": 455800 }, { "epoch": 0.8506667848537085, "grad_norm": 1.1615766286849976, "learning_rate": 0.0001931267370271801, "loss": 4.6077, "step": 455850 }, { "epoch": 0.8507600904130871, "grad_norm": 1.085879921913147, "learning_rate": 0.00019312520637433086, "loss": 4.5408, "step": 455900 }, { "epoch": 0.8508533959724656, "grad_norm": 1.3078773021697998, "learning_rate": 0.00019312367555713203, "loss": 4.4441, "step": 455950 }, { "epoch": 0.850946701531844, "grad_norm": 0.8836840391159058, "learning_rate": 0.00019312214457558634, "loss": 4.5579, "step": 456000 }, { "epoch": 0.850946701531844, "eval_loss": 4.705729007720947, "eval_runtime": 234.949, "eval_samples_per_second": 11.1, "eval_steps_per_second": 11.1, "eval_tts_loss": 7.585697766739098, "step": 456000 }, { "epoch": 0.8510400070912225, "grad_norm": 0.922694206237793, "learning_rate": 0.0001931206134296964, "loss": 4.4405, "step": 456050 }, { "epoch": 0.851133312650601, "grad_norm": 1.6349189281463623, "learning_rate": 0.000193119082119465, "loss": 4.6537, "step": 456100 }, { "epoch": 0.8512266182099795, "grad_norm": 0.9771808981895447, "learning_rate": 0.00019311755064489478, "loss": 4.5145, "step": 456150 }, { "epoch": 0.851319923769358, "grad_norm": 0.9934489727020264, "learning_rate": 0.00019311601900598845, "loss": 4.5699, "step": 456200 }, { "epoch": 0.8514132293287365, "grad_norm": 0.9607605338096619, "learning_rate": 0.00019311448720274877, "loss": 4.4473, "step": 456250 }, { "epoch": 0.851506534888115, "grad_norm": 1.040191650390625, "learning_rate": 0.0001931129552351784, "loss": 4.3806, "step": 456300 }, { "epoch": 0.8515998404474935, "grad_norm": 1.0230000019073486, "learning_rate": 0.00019311142310328002, "loss": 4.5976, "step": 456350 }, { "epoch": 0.851693146006872, "grad_norm": 1.1846024990081787, "learning_rate": 0.00019310989080705636, "loss": 4.4676, "step": 456400 }, { "epoch": 0.8517864515662504, "grad_norm": 1.0081008672714233, "learning_rate": 0.00019310835834651018, "loss": 4.5692, "step": 456450 }, { "epoch": 0.8518797571256289, "grad_norm": 1.0581300258636475, "learning_rate": 0.0001931068257216441, "loss": 4.4911, "step": 456500 }, { "epoch": 0.8519730626850074, "grad_norm": 1.159439206123352, "learning_rate": 0.00019310529293246081, "loss": 4.8498, "step": 456550 }, { "epoch": 0.852066368244386, "grad_norm": 1.0190571546554565, "learning_rate": 0.00019310375997896314, "loss": 4.5151, "step": 456600 }, { "epoch": 0.8521596738037645, "grad_norm": 1.3881120681762695, "learning_rate": 0.00019310222686115365, "loss": 4.4332, "step": 456650 }, { "epoch": 0.8522529793631429, "grad_norm": 1.0735052824020386, "learning_rate": 0.00019310069357903514, "loss": 4.6731, "step": 456700 }, { "epoch": 0.8523462849225214, "grad_norm": 1.28987455368042, "learning_rate": 0.00019309916013261029, "loss": 4.5543, "step": 456750 }, { "epoch": 0.8524395904818999, "grad_norm": 1.0508424043655396, "learning_rate": 0.0001930976265218818, "loss": 4.5146, "step": 456800 }, { "epoch": 0.8525328960412784, "grad_norm": 1.033884882926941, "learning_rate": 0.00019309609274685238, "loss": 4.4683, "step": 456850 }, { "epoch": 0.8526262016006568, "grad_norm": 0.9419358372688293, "learning_rate": 0.0001930945588075247, "loss": 4.6409, "step": 456900 }, { "epoch": 0.8527195071600354, "grad_norm": 1.1568870544433594, "learning_rate": 0.00019309302470390155, "loss": 4.6434, "step": 456950 }, { "epoch": 0.8528128127194139, "grad_norm": 0.8612609505653381, "learning_rate": 0.0001930914904359856, "loss": 4.5828, "step": 457000 }, { "epoch": 0.8529061182787924, "grad_norm": 1.1326133012771606, "learning_rate": 0.0001930899560037795, "loss": 4.4556, "step": 457050 }, { "epoch": 0.8529994238381708, "grad_norm": 1.1270262002944946, "learning_rate": 0.00019308842140728608, "loss": 4.6409, "step": 457100 }, { "epoch": 0.8530927293975493, "grad_norm": 1.13180673122406, "learning_rate": 0.00019308688664650792, "loss": 4.5052, "step": 457150 }, { "epoch": 0.8531860349569278, "grad_norm": 0.6030737161636353, "learning_rate": 0.0001930853517214478, "loss": 4.3616, "step": 457200 }, { "epoch": 0.8532793405163063, "grad_norm": 1.1248189210891724, "learning_rate": 0.0001930838166321084, "loss": 4.4918, "step": 457250 }, { "epoch": 0.8533726460756849, "grad_norm": 0.838978111743927, "learning_rate": 0.00019308228137849244, "loss": 4.5434, "step": 457300 }, { "epoch": 0.8534659516350633, "grad_norm": 0.9123812913894653, "learning_rate": 0.00019308074596060263, "loss": 4.503, "step": 457350 }, { "epoch": 0.8535592571944418, "grad_norm": 1.452649712562561, "learning_rate": 0.0001930792103784417, "loss": 4.5399, "step": 457400 }, { "epoch": 0.8536525627538203, "grad_norm": 1.1544631719589233, "learning_rate": 0.00019307767463201233, "loss": 4.636, "step": 457450 }, { "epoch": 0.8537458683131988, "grad_norm": 1.042115569114685, "learning_rate": 0.00019307613872131726, "loss": 4.4964, "step": 457500 }, { "epoch": 0.8538391738725772, "grad_norm": 1.2403693199157715, "learning_rate": 0.0001930746026463592, "loss": 4.5582, "step": 457550 }, { "epoch": 0.8539324794319557, "grad_norm": 1.259184718132019, "learning_rate": 0.0001930730664071408, "loss": 4.4463, "step": 457600 }, { "epoch": 0.8540257849913342, "grad_norm": 0.9611954092979431, "learning_rate": 0.00019307153000366483, "loss": 4.4916, "step": 457650 }, { "epoch": 0.8541190905507128, "grad_norm": 1.051138997077942, "learning_rate": 0.000193069993435934, "loss": 4.5431, "step": 457700 }, { "epoch": 0.8542123961100913, "grad_norm": 0.987114667892456, "learning_rate": 0.00019306845670395099, "loss": 4.5658, "step": 457750 }, { "epoch": 0.8543057016694697, "grad_norm": 0.884158730506897, "learning_rate": 0.00019306691980771853, "loss": 4.4915, "step": 457800 }, { "epoch": 0.8543990072288482, "grad_norm": 1.4324150085449219, "learning_rate": 0.00019306538274723933, "loss": 4.6267, "step": 457850 }, { "epoch": 0.8544923127882267, "grad_norm": 1.1447198390960693, "learning_rate": 0.0001930638455225161, "loss": 4.5824, "step": 457900 }, { "epoch": 0.8545856183476052, "grad_norm": 1.421771764755249, "learning_rate": 0.00019306230813355155, "loss": 4.291, "step": 457950 }, { "epoch": 0.8546789239069836, "grad_norm": 1.0389111042022705, "learning_rate": 0.00019306077058034844, "loss": 4.588, "step": 458000 }, { "epoch": 0.8547722294663622, "grad_norm": 1.0093166828155518, "learning_rate": 0.0001930592328629094, "loss": 4.4956, "step": 458050 }, { "epoch": 0.8548655350257407, "grad_norm": 0.8882410526275635, "learning_rate": 0.00019305769498123722, "loss": 4.5256, "step": 458100 }, { "epoch": 0.8549588405851192, "grad_norm": 1.124986171722412, "learning_rate": 0.0001930561569353346, "loss": 4.6541, "step": 458150 }, { "epoch": 0.8550521461444976, "grad_norm": 0.8576604127883911, "learning_rate": 0.0001930546187252042, "loss": 4.329, "step": 458200 }, { "epoch": 0.8551454517038761, "grad_norm": 1.1762045621871948, "learning_rate": 0.00019305308035084877, "loss": 4.4463, "step": 458250 }, { "epoch": 0.8552387572632546, "grad_norm": 1.1113213300704956, "learning_rate": 0.00019305154181227107, "loss": 4.379, "step": 458300 }, { "epoch": 0.8553320628226331, "grad_norm": 1.1153367757797241, "learning_rate": 0.0001930500031094737, "loss": 4.5141, "step": 458350 }, { "epoch": 0.8554253683820117, "grad_norm": 1.1676336526870728, "learning_rate": 0.00019304846424245947, "loss": 4.5917, "step": 458400 }, { "epoch": 0.8555186739413901, "grad_norm": 0.8630878925323486, "learning_rate": 0.0001930469252112311, "loss": 4.4346, "step": 458450 }, { "epoch": 0.8556119795007686, "grad_norm": 1.2297073602676392, "learning_rate": 0.0001930453860157913, "loss": 4.7697, "step": 458500 }, { "epoch": 0.8557052850601471, "grad_norm": 1.3272281885147095, "learning_rate": 0.0001930438466561427, "loss": 4.352, "step": 458550 }, { "epoch": 0.8557985906195256, "grad_norm": 1.030927062034607, "learning_rate": 0.00019304230713228812, "loss": 4.4467, "step": 458600 }, { "epoch": 0.855891896178904, "grad_norm": 1.0438718795776367, "learning_rate": 0.00019304076744423024, "loss": 4.5988, "step": 458650 }, { "epoch": 0.8559852017382825, "grad_norm": 1.0471688508987427, "learning_rate": 0.00019303922759197176, "loss": 4.6565, "step": 458700 }, { "epoch": 0.8560785072976611, "grad_norm": 1.117255687713623, "learning_rate": 0.00019303768757551545, "loss": 4.6486, "step": 458750 }, { "epoch": 0.8561718128570396, "grad_norm": 1.063703179359436, "learning_rate": 0.00019303614739486395, "loss": 4.4944, "step": 458800 }, { "epoch": 0.8562651184164181, "grad_norm": 1.191904902458191, "learning_rate": 0.00019303460705002006, "loss": 4.5925, "step": 458850 }, { "epoch": 0.8563584239757965, "grad_norm": 1.1104278564453125, "learning_rate": 0.00019303306654098643, "loss": 4.5023, "step": 458900 }, { "epoch": 0.856451729535175, "grad_norm": 0.9294896721839905, "learning_rate": 0.00019303152586776584, "loss": 4.486, "step": 458950 }, { "epoch": 0.8565450350945535, "grad_norm": 1.0294172763824463, "learning_rate": 0.0001930299850303609, "loss": 4.6008, "step": 459000 }, { "epoch": 0.8565450350945535, "eval_loss": 4.719043731689453, "eval_runtime": 234.5789, "eval_samples_per_second": 11.118, "eval_steps_per_second": 11.118, "eval_tts_loss": 7.604671585362153, "step": 459000 }, { "epoch": 0.856638340653932, "grad_norm": 1.1844731569290161, "learning_rate": 0.00019302844402877448, "loss": 4.4275, "step": 459050 }, { "epoch": 0.8567316462133105, "grad_norm": 0.5739268064498901, "learning_rate": 0.00019302690286300921, "loss": 4.5983, "step": 459100 }, { "epoch": 0.856824951772689, "grad_norm": 0.8256612420082092, "learning_rate": 0.0001930253615330678, "loss": 4.5875, "step": 459150 }, { "epoch": 0.8569182573320675, "grad_norm": 1.1579002141952515, "learning_rate": 0.00019302382003895301, "loss": 4.3835, "step": 459200 }, { "epoch": 0.857011562891446, "grad_norm": 1.6398296356201172, "learning_rate": 0.00019302227838066755, "loss": 4.5398, "step": 459250 }, { "epoch": 0.8571048684508245, "grad_norm": 1.1075565814971924, "learning_rate": 0.00019302073655821412, "loss": 4.5899, "step": 459300 }, { "epoch": 0.8571981740102029, "grad_norm": 1.1114987134933472, "learning_rate": 0.0001930191945715955, "loss": 4.6631, "step": 459350 }, { "epoch": 0.8572914795695814, "grad_norm": 1.2715694904327393, "learning_rate": 0.00019301765242081434, "loss": 4.4969, "step": 459400 }, { "epoch": 0.85738478512896, "grad_norm": 1.1257323026657104, "learning_rate": 0.0001930161101058734, "loss": 4.3626, "step": 459450 }, { "epoch": 0.8574780906883385, "grad_norm": 1.3674393892288208, "learning_rate": 0.00019301456762677537, "loss": 4.7802, "step": 459500 }, { "epoch": 0.8575713962477169, "grad_norm": 1.3454407453536987, "learning_rate": 0.000193013024983523, "loss": 4.6992, "step": 459550 }, { "epoch": 0.8576647018070954, "grad_norm": 0.8767492771148682, "learning_rate": 0.00019301148217611903, "loss": 4.54, "step": 459600 }, { "epoch": 0.8577580073664739, "grad_norm": 0.8753656148910522, "learning_rate": 0.00019300993920456612, "loss": 4.5982, "step": 459650 }, { "epoch": 0.8578513129258524, "grad_norm": 1.1201082468032837, "learning_rate": 0.00019300839606886707, "loss": 4.6188, "step": 459700 }, { "epoch": 0.8579446184852308, "grad_norm": 0.9129759073257446, "learning_rate": 0.00019300685276902455, "loss": 4.334, "step": 459750 }, { "epoch": 0.8580379240446094, "grad_norm": 0.828029215335846, "learning_rate": 0.0001930053093050413, "loss": 4.5788, "step": 459800 }, { "epoch": 0.8581312296039879, "grad_norm": 1.2626936435699463, "learning_rate": 0.00019300376567692006, "loss": 4.1639, "step": 459850 }, { "epoch": 0.8582245351633664, "grad_norm": 1.095350980758667, "learning_rate": 0.00019300222188466354, "loss": 4.4777, "step": 459900 }, { "epoch": 0.8583178407227449, "grad_norm": 0.9461998343467712, "learning_rate": 0.00019300067792827442, "loss": 4.7225, "step": 459950 }, { "epoch": 0.8584111462821233, "grad_norm": 1.1858104467391968, "learning_rate": 0.0001929991338077555, "loss": 4.8325, "step": 460000 }, { "epoch": 0.8585044518415018, "grad_norm": 0.8696582317352295, "learning_rate": 0.00019299758952310946, "loss": 4.623, "step": 460050 }, { "epoch": 0.8585977574008803, "grad_norm": 1.4265830516815186, "learning_rate": 0.00019299604507433906, "loss": 4.5486, "step": 460100 }, { "epoch": 0.8586910629602589, "grad_norm": 1.3984472751617432, "learning_rate": 0.00019299450046144699, "loss": 4.4441, "step": 460150 }, { "epoch": 0.8587843685196374, "grad_norm": 1.2104437351226807, "learning_rate": 0.00019299295568443595, "loss": 4.6945, "step": 460200 }, { "epoch": 0.8588776740790158, "grad_norm": 0.9072539806365967, "learning_rate": 0.00019299141074330875, "loss": 4.6262, "step": 460250 }, { "epoch": 0.8589709796383943, "grad_norm": 0.9867424964904785, "learning_rate": 0.00019298986563806808, "loss": 4.4876, "step": 460300 }, { "epoch": 0.8590642851977728, "grad_norm": 1.104433536529541, "learning_rate": 0.00019298832036871667, "loss": 4.5798, "step": 460350 }, { "epoch": 0.8591575907571513, "grad_norm": 1.264764666557312, "learning_rate": 0.00019298677493525718, "loss": 4.7017, "step": 460400 }, { "epoch": 0.8592508963165297, "grad_norm": 1.002766728401184, "learning_rate": 0.00019298522933769246, "loss": 4.5531, "step": 460450 }, { "epoch": 0.8593442018759083, "grad_norm": 1.2102422714233398, "learning_rate": 0.0001929836835760251, "loss": 4.6989, "step": 460500 }, { "epoch": 0.8594375074352868, "grad_norm": 1.1256135702133179, "learning_rate": 0.00019298213765025794, "loss": 4.5006, "step": 460550 }, { "epoch": 0.8595308129946653, "grad_norm": 1.0683096647262573, "learning_rate": 0.00019298059156039367, "loss": 4.5041, "step": 460600 }, { "epoch": 0.8596241185540437, "grad_norm": 1.422804832458496, "learning_rate": 0.00019297904530643503, "loss": 4.5532, "step": 460650 }, { "epoch": 0.8597174241134222, "grad_norm": 1.1306761503219604, "learning_rate": 0.0001929774988883847, "loss": 4.6819, "step": 460700 }, { "epoch": 0.8598107296728007, "grad_norm": 0.88563472032547, "learning_rate": 0.00019297595230624546, "loss": 4.6199, "step": 460750 }, { "epoch": 0.8599040352321792, "grad_norm": 0.8759470582008362, "learning_rate": 0.00019297440556002006, "loss": 4.5261, "step": 460800 }, { "epoch": 0.8599973407915578, "grad_norm": 1.230621337890625, "learning_rate": 0.00019297285864971114, "loss": 4.5109, "step": 460850 }, { "epoch": 0.8600906463509362, "grad_norm": 1.1666408777236938, "learning_rate": 0.00019297131157532148, "loss": 4.3958, "step": 460900 }, { "epoch": 0.8601839519103147, "grad_norm": 1.119201898574829, "learning_rate": 0.00019296976433685385, "loss": 4.4851, "step": 460950 }, { "epoch": 0.8602772574696932, "grad_norm": 0.9621995687484741, "learning_rate": 0.00019296821693431094, "loss": 4.438, "step": 461000 }, { "epoch": 0.8603705630290717, "grad_norm": 1.2571659088134766, "learning_rate": 0.00019296666936769548, "loss": 4.7484, "step": 461050 }, { "epoch": 0.8604638685884501, "grad_norm": 1.1245945692062378, "learning_rate": 0.00019296512163701016, "loss": 4.6354, "step": 461100 }, { "epoch": 0.8605571741478286, "grad_norm": 1.136570692062378, "learning_rate": 0.00019296357374225785, "loss": 4.4911, "step": 461150 }, { "epoch": 0.8606504797072072, "grad_norm": 1.0873862504959106, "learning_rate": 0.0001929620256834411, "loss": 4.527, "step": 461200 }, { "epoch": 0.8607437852665857, "grad_norm": 0.8894184827804565, "learning_rate": 0.00019296047746056278, "loss": 4.5169, "step": 461250 }, { "epoch": 0.8608370908259642, "grad_norm": 1.1038322448730469, "learning_rate": 0.00019295892907362556, "loss": 4.6536, "step": 461300 }, { "epoch": 0.8609303963853426, "grad_norm": 0.9065510034561157, "learning_rate": 0.0001929573805226322, "loss": 4.5633, "step": 461350 }, { "epoch": 0.8610237019447211, "grad_norm": 1.192690372467041, "learning_rate": 0.00019295583180758543, "loss": 4.4378, "step": 461400 }, { "epoch": 0.8611170075040996, "grad_norm": 1.1816054582595825, "learning_rate": 0.00019295428292848794, "loss": 4.5303, "step": 461450 }, { "epoch": 0.8612103130634781, "grad_norm": 0.892443060874939, "learning_rate": 0.00019295273388534253, "loss": 4.4727, "step": 461500 }, { "epoch": 0.8613036186228566, "grad_norm": 1.0818899869918823, "learning_rate": 0.00019295118467815186, "loss": 4.4628, "step": 461550 }, { "epoch": 0.8613969241822351, "grad_norm": 1.1730269193649292, "learning_rate": 0.0001929496353069187, "loss": 4.5145, "step": 461600 }, { "epoch": 0.8614902297416136, "grad_norm": 0.8327314853668213, "learning_rate": 0.00019294808577164586, "loss": 4.6123, "step": 461650 }, { "epoch": 0.8615835353009921, "grad_norm": 0.870633602142334, "learning_rate": 0.00019294653607233595, "loss": 4.523, "step": 461700 }, { "epoch": 0.8616768408603706, "grad_norm": 1.139284372329712, "learning_rate": 0.00019294498620899176, "loss": 4.8048, "step": 461750 }, { "epoch": 0.861770146419749, "grad_norm": 1.2574660778045654, "learning_rate": 0.00019294343618161603, "loss": 4.5516, "step": 461800 }, { "epoch": 0.8618634519791275, "grad_norm": 1.016002893447876, "learning_rate": 0.0001929418859902115, "loss": 4.5009, "step": 461850 }, { "epoch": 0.8619567575385061, "grad_norm": 1.5868070125579834, "learning_rate": 0.00019294033563478088, "loss": 4.9067, "step": 461900 }, { "epoch": 0.8620500630978846, "grad_norm": 1.2562463283538818, "learning_rate": 0.0001929387851153269, "loss": 4.4525, "step": 461950 }, { "epoch": 0.862143368657263, "grad_norm": 1.1887893676757812, "learning_rate": 0.00019293723443185236, "loss": 4.6491, "step": 462000 }, { "epoch": 0.862143368657263, "eval_loss": 4.7134013175964355, "eval_runtime": 234.3748, "eval_samples_per_second": 11.127, "eval_steps_per_second": 11.127, "eval_tts_loss": 7.602741665749645, "step": 462000 }, { "epoch": 0.8622366742166415, "grad_norm": 1.0132383108139038, "learning_rate": 0.00019293568358435994, "loss": 4.721, "step": 462050 }, { "epoch": 0.86232997977602, "grad_norm": 1.1651688814163208, "learning_rate": 0.00019293413257285237, "loss": 4.5922, "step": 462100 }, { "epoch": 0.8624232853353985, "grad_norm": 0.6764702796936035, "learning_rate": 0.00019293258139733245, "loss": 4.5047, "step": 462150 }, { "epoch": 0.862516590894777, "grad_norm": 0.9960355162620544, "learning_rate": 0.00019293103005780284, "loss": 4.622, "step": 462200 }, { "epoch": 0.8626098964541555, "grad_norm": 1.2576119899749756, "learning_rate": 0.00019292947855426633, "loss": 4.6511, "step": 462250 }, { "epoch": 0.862703202013534, "grad_norm": 1.1315255165100098, "learning_rate": 0.00019292792688672562, "loss": 4.5326, "step": 462300 }, { "epoch": 0.8627965075729125, "grad_norm": 1.0933852195739746, "learning_rate": 0.00019292637505518353, "loss": 4.3809, "step": 462350 }, { "epoch": 0.862889813132291, "grad_norm": 1.229644775390625, "learning_rate": 0.00019292482305964268, "loss": 4.4325, "step": 462400 }, { "epoch": 0.8629831186916694, "grad_norm": 1.0292531251907349, "learning_rate": 0.00019292327090010588, "loss": 4.5731, "step": 462450 }, { "epoch": 0.8630764242510479, "grad_norm": 1.2701035737991333, "learning_rate": 0.00019292171857657585, "loss": 4.6353, "step": 462500 }, { "epoch": 0.8631697298104264, "grad_norm": 0.9864139556884766, "learning_rate": 0.00019292016608905534, "loss": 4.5678, "step": 462550 }, { "epoch": 0.863263035369805, "grad_norm": 1.189972162246704, "learning_rate": 0.00019291861343754708, "loss": 4.5319, "step": 462600 }, { "epoch": 0.8633563409291835, "grad_norm": 1.2956823110580444, "learning_rate": 0.00019291706062205387, "loss": 4.5879, "step": 462650 }, { "epoch": 0.8634496464885619, "grad_norm": 1.01268470287323, "learning_rate": 0.00019291550764257833, "loss": 4.6386, "step": 462700 }, { "epoch": 0.8635429520479404, "grad_norm": 0.85806804895401, "learning_rate": 0.00019291395449912334, "loss": 4.7307, "step": 462750 }, { "epoch": 0.8636362576073189, "grad_norm": 1.0039422512054443, "learning_rate": 0.0001929124011916915, "loss": 4.4267, "step": 462800 }, { "epoch": 0.8637295631666974, "grad_norm": 0.9849223494529724, "learning_rate": 0.00019291084772028565, "loss": 4.4856, "step": 462850 }, { "epoch": 0.8638228687260758, "grad_norm": 1.2357029914855957, "learning_rate": 0.00019290929408490848, "loss": 4.462, "step": 462900 }, { "epoch": 0.8639161742854543, "grad_norm": 1.0906906127929688, "learning_rate": 0.0001929077402855628, "loss": 4.3928, "step": 462950 }, { "epoch": 0.8640094798448329, "grad_norm": 1.1652013063430786, "learning_rate": 0.00019290618632225126, "loss": 4.4009, "step": 463000 }, { "epoch": 0.8641027854042114, "grad_norm": 1.3629333972930908, "learning_rate": 0.00019290463219497666, "loss": 4.4741, "step": 463050 }, { "epoch": 0.8641960909635898, "grad_norm": 0.8302815556526184, "learning_rate": 0.00019290307790374175, "loss": 4.5197, "step": 463100 }, { "epoch": 0.8642893965229683, "grad_norm": 0.9268425107002258, "learning_rate": 0.00019290152344854925, "loss": 4.3876, "step": 463150 }, { "epoch": 0.8643827020823468, "grad_norm": 1.082864761352539, "learning_rate": 0.0001928999688294019, "loss": 4.6077, "step": 463200 }, { "epoch": 0.8644760076417253, "grad_norm": 1.1909546852111816, "learning_rate": 0.00019289841404630242, "loss": 4.5246, "step": 463250 }, { "epoch": 0.8645693132011038, "grad_norm": 0.8759803175926208, "learning_rate": 0.00019289685909925364, "loss": 4.6791, "step": 463300 }, { "epoch": 0.8646626187604823, "grad_norm": 1.3796002864837646, "learning_rate": 0.0001928953039882582, "loss": 4.6118, "step": 463350 }, { "epoch": 0.8647559243198608, "grad_norm": 1.1816750764846802, "learning_rate": 0.00019289374871331893, "loss": 4.4584, "step": 463400 }, { "epoch": 0.8648492298792393, "grad_norm": 1.023587703704834, "learning_rate": 0.00019289219327443854, "loss": 4.5076, "step": 463450 }, { "epoch": 0.8649425354386178, "grad_norm": 0.8140079975128174, "learning_rate": 0.00019289063767161975, "loss": 4.3738, "step": 463500 }, { "epoch": 0.8650358409979962, "grad_norm": 0.8466084599494934, "learning_rate": 0.00019288908190486534, "loss": 4.4616, "step": 463550 }, { "epoch": 0.8651291465573747, "grad_norm": 0.9516765475273132, "learning_rate": 0.00019288752597417804, "loss": 4.4458, "step": 463600 }, { "epoch": 0.8652224521167532, "grad_norm": 0.8681322336196899, "learning_rate": 0.00019288596987956062, "loss": 4.4382, "step": 463650 }, { "epoch": 0.8653157576761318, "grad_norm": 1.0693433284759521, "learning_rate": 0.0001928844136210158, "loss": 4.6493, "step": 463700 }, { "epoch": 0.8654090632355103, "grad_norm": 0.9446530938148499, "learning_rate": 0.00019288285719854632, "loss": 4.6797, "step": 463750 }, { "epoch": 0.8655023687948887, "grad_norm": 1.0062452554702759, "learning_rate": 0.00019288130061215496, "loss": 4.5555, "step": 463800 }, { "epoch": 0.8655956743542672, "grad_norm": 1.0414299964904785, "learning_rate": 0.00019287974386184443, "loss": 4.5787, "step": 463850 }, { "epoch": 0.8656889799136457, "grad_norm": 1.0510808229446411, "learning_rate": 0.0001928781869476175, "loss": 4.5636, "step": 463900 }, { "epoch": 0.8657822854730242, "grad_norm": 1.231963872909546, "learning_rate": 0.0001928766298694769, "loss": 4.7286, "step": 463950 }, { "epoch": 0.8658755910324026, "grad_norm": 1.066776156425476, "learning_rate": 0.0001928750726274254, "loss": 4.538, "step": 464000 }, { "epoch": 0.8659688965917812, "grad_norm": 1.0529029369354248, "learning_rate": 0.00019287351522146576, "loss": 4.3238, "step": 464050 }, { "epoch": 0.8660622021511597, "grad_norm": 1.0820529460906982, "learning_rate": 0.00019287195765160066, "loss": 4.5946, "step": 464100 }, { "epoch": 0.8661555077105382, "grad_norm": 1.049566626548767, "learning_rate": 0.00019287039991783297, "loss": 4.4558, "step": 464150 }, { "epoch": 0.8662488132699167, "grad_norm": 1.1727852821350098, "learning_rate": 0.00019286884202016527, "loss": 4.397, "step": 464200 }, { "epoch": 0.8663421188292951, "grad_norm": 0.9440301060676575, "learning_rate": 0.00019286728395860047, "loss": 4.5329, "step": 464250 }, { "epoch": 0.8664354243886736, "grad_norm": 0.8551458716392517, "learning_rate": 0.00019286572573314124, "loss": 4.5481, "step": 464300 }, { "epoch": 0.8665287299480521, "grad_norm": 1.333892822265625, "learning_rate": 0.00019286416734379036, "loss": 4.5486, "step": 464350 }, { "epoch": 0.8666220355074307, "grad_norm": 0.996533215045929, "learning_rate": 0.00019286260879055052, "loss": 4.5251, "step": 464400 }, { "epoch": 0.8667153410668091, "grad_norm": 1.046530842781067, "learning_rate": 0.00019286105007342456, "loss": 4.6038, "step": 464450 }, { "epoch": 0.8668086466261876, "grad_norm": 1.0920482873916626, "learning_rate": 0.00019285949119241517, "loss": 4.8125, "step": 464500 }, { "epoch": 0.8669019521855661, "grad_norm": 1.100996971130371, "learning_rate": 0.0001928579321475251, "loss": 4.4519, "step": 464550 }, { "epoch": 0.8669952577449446, "grad_norm": 1.4868696928024292, "learning_rate": 0.00019285637293875714, "loss": 4.5632, "step": 464600 }, { "epoch": 0.867088563304323, "grad_norm": 1.1399754285812378, "learning_rate": 0.000192854813566114, "loss": 4.4723, "step": 464650 }, { "epoch": 0.8671818688637015, "grad_norm": 1.156640887260437, "learning_rate": 0.00019285325402959848, "loss": 4.4813, "step": 464700 }, { "epoch": 0.8672751744230801, "grad_norm": 0.8303513526916504, "learning_rate": 0.0001928516943292133, "loss": 4.417, "step": 464750 }, { "epoch": 0.8673684799824586, "grad_norm": 0.8029341101646423, "learning_rate": 0.0001928501344649612, "loss": 4.4222, "step": 464800 }, { "epoch": 0.8674617855418371, "grad_norm": 0.8799920082092285, "learning_rate": 0.00019284857443684495, "loss": 4.7166, "step": 464850 }, { "epoch": 0.8675550911012155, "grad_norm": 1.060721516609192, "learning_rate": 0.00019284701424486734, "loss": 4.6998, "step": 464900 }, { "epoch": 0.867648396660594, "grad_norm": 1.1703776121139526, "learning_rate": 0.00019284545388903107, "loss": 4.612, "step": 464950 }, { "epoch": 0.8677417022199725, "grad_norm": 0.948229193687439, "learning_rate": 0.00019284389336933887, "loss": 4.6585, "step": 465000 }, { "epoch": 0.8677417022199725, "eval_loss": 4.716647624969482, "eval_runtime": 233.7784, "eval_samples_per_second": 11.156, "eval_steps_per_second": 11.156, "eval_tts_loss": 7.5927892732234, "step": 465000 }, { "epoch": 0.867835007779351, "grad_norm": 1.1838222742080688, "learning_rate": 0.00019284233268579358, "loss": 4.3513, "step": 465050 }, { "epoch": 0.8679283133387296, "grad_norm": 1.0094722509384155, "learning_rate": 0.00019284077183839789, "loss": 4.48, "step": 465100 }, { "epoch": 0.868021618898108, "grad_norm": 1.1781718730926514, "learning_rate": 0.00019283921082715458, "loss": 4.5282, "step": 465150 }, { "epoch": 0.8681149244574865, "grad_norm": 0.8361363410949707, "learning_rate": 0.0001928376496520664, "loss": 4.4695, "step": 465200 }, { "epoch": 0.868208230016865, "grad_norm": 0.8535106182098389, "learning_rate": 0.00019283608831313607, "loss": 4.4963, "step": 465250 }, { "epoch": 0.8683015355762435, "grad_norm": 1.2140179872512817, "learning_rate": 0.00019283452681036643, "loss": 4.6092, "step": 465300 }, { "epoch": 0.8683948411356219, "grad_norm": 1.2857625484466553, "learning_rate": 0.00019283296514376015, "loss": 4.5237, "step": 465350 }, { "epoch": 0.8684881466950004, "grad_norm": 1.263896107673645, "learning_rate": 0.00019283140331332005, "loss": 4.6654, "step": 465400 }, { "epoch": 0.868581452254379, "grad_norm": 0.9182162880897522, "learning_rate": 0.00019282984131904882, "loss": 4.3799, "step": 465450 }, { "epoch": 0.8686747578137575, "grad_norm": 1.1222580671310425, "learning_rate": 0.00019282827916094927, "loss": 4.6969, "step": 465500 }, { "epoch": 0.868768063373136, "grad_norm": 1.0498583316802979, "learning_rate": 0.00019282671683902415, "loss": 4.6806, "step": 465550 }, { "epoch": 0.8688613689325144, "grad_norm": 1.0600595474243164, "learning_rate": 0.00019282515435327618, "loss": 4.5403, "step": 465600 }, { "epoch": 0.8689546744918929, "grad_norm": 1.0661931037902832, "learning_rate": 0.00019282359170370817, "loss": 4.6093, "step": 465650 }, { "epoch": 0.8690479800512714, "grad_norm": 0.9582245945930481, "learning_rate": 0.00019282202889032288, "loss": 4.6067, "step": 465700 }, { "epoch": 0.8691412856106498, "grad_norm": 1.061657190322876, "learning_rate": 0.000192820465913123, "loss": 4.5828, "step": 465750 }, { "epoch": 0.8692345911700284, "grad_norm": 0.9799156785011292, "learning_rate": 0.00019281890277211136, "loss": 4.5106, "step": 465800 }, { "epoch": 0.8693278967294069, "grad_norm": 1.1217502355575562, "learning_rate": 0.00019281733946729067, "loss": 4.5386, "step": 465850 }, { "epoch": 0.8694212022887854, "grad_norm": 1.0424093008041382, "learning_rate": 0.00019281577599866373, "loss": 4.721, "step": 465900 }, { "epoch": 0.8695145078481639, "grad_norm": 1.0570974349975586, "learning_rate": 0.00019281421236623324, "loss": 4.5282, "step": 465950 }, { "epoch": 0.8696078134075423, "grad_norm": 1.1005921363830566, "learning_rate": 0.000192812648570002, "loss": 4.4005, "step": 466000 }, { "epoch": 0.8697011189669208, "grad_norm": 1.060054898262024, "learning_rate": 0.0001928110846099728, "loss": 4.3224, "step": 466050 }, { "epoch": 0.8697944245262993, "grad_norm": 1.0495976209640503, "learning_rate": 0.00019280952048614833, "loss": 4.545, "step": 466100 }, { "epoch": 0.8698877300856779, "grad_norm": 0.8841475248336792, "learning_rate": 0.00019280795619853143, "loss": 4.5245, "step": 466150 }, { "epoch": 0.8699810356450564, "grad_norm": 1.0591237545013428, "learning_rate": 0.0001928063917471248, "loss": 4.6912, "step": 466200 }, { "epoch": 0.8700743412044348, "grad_norm": 1.3591599464416504, "learning_rate": 0.0001928048271319312, "loss": 4.6817, "step": 466250 }, { "epoch": 0.8701676467638133, "grad_norm": 1.3419016599655151, "learning_rate": 0.0001928032623529534, "loss": 4.6041, "step": 466300 }, { "epoch": 0.8702609523231918, "grad_norm": 0.8852062821388245, "learning_rate": 0.00019280169741019417, "loss": 4.4491, "step": 466350 }, { "epoch": 0.8703542578825703, "grad_norm": 1.100988507270813, "learning_rate": 0.0001928001323036563, "loss": 4.6465, "step": 466400 }, { "epoch": 0.8704475634419487, "grad_norm": 1.0544284582138062, "learning_rate": 0.00019279856703334253, "loss": 4.5577, "step": 466450 }, { "epoch": 0.8705408690013273, "grad_norm": 1.1751656532287598, "learning_rate": 0.0001927970015992556, "loss": 4.4888, "step": 466500 }, { "epoch": 0.8706341745607058, "grad_norm": 0.8412641286849976, "learning_rate": 0.00019279543600139831, "loss": 4.4661, "step": 466550 }, { "epoch": 0.8707274801200843, "grad_norm": 1.0064772367477417, "learning_rate": 0.0001927938702397734, "loss": 4.4996, "step": 466600 }, { "epoch": 0.8708207856794627, "grad_norm": 0.9518139958381653, "learning_rate": 0.0001927923043143836, "loss": 4.4364, "step": 466650 }, { "epoch": 0.8709140912388412, "grad_norm": 1.2723345756530762, "learning_rate": 0.00019279073822523176, "loss": 4.5561, "step": 466700 }, { "epoch": 0.8710073967982197, "grad_norm": 0.9318612217903137, "learning_rate": 0.00019278917197232057, "loss": 4.4146, "step": 466750 }, { "epoch": 0.8711007023575982, "grad_norm": 1.12967848777771, "learning_rate": 0.0001927876055556528, "loss": 4.6484, "step": 466800 }, { "epoch": 0.8711940079169768, "grad_norm": 1.0170807838439941, "learning_rate": 0.00019278603897523127, "loss": 4.5308, "step": 466850 }, { "epoch": 0.8712873134763552, "grad_norm": 1.1193941831588745, "learning_rate": 0.0001927844722310587, "loss": 4.4862, "step": 466900 }, { "epoch": 0.8713806190357337, "grad_norm": 1.1991682052612305, "learning_rate": 0.00019278290532313784, "loss": 4.6256, "step": 466950 }, { "epoch": 0.8714739245951122, "grad_norm": 1.2194184064865112, "learning_rate": 0.00019278133825147148, "loss": 4.6567, "step": 467000 }, { "epoch": 0.8715672301544907, "grad_norm": 0.9851527810096741, "learning_rate": 0.0001927797710160624, "loss": 4.5432, "step": 467050 }, { "epoch": 0.8716605357138691, "grad_norm": 1.0576553344726562, "learning_rate": 0.00019277820361691332, "loss": 4.5246, "step": 467100 }, { "epoch": 0.8717538412732476, "grad_norm": 0.8049138784408569, "learning_rate": 0.00019277663605402707, "loss": 4.3887, "step": 467150 }, { "epoch": 0.8718471468326262, "grad_norm": 1.1640795469284058, "learning_rate": 0.00019277506832740638, "loss": 4.5693, "step": 467200 }, { "epoch": 0.8719404523920047, "grad_norm": 1.2224133014678955, "learning_rate": 0.000192773500437054, "loss": 4.5694, "step": 467250 }, { "epoch": 0.8720337579513832, "grad_norm": 1.3226362466812134, "learning_rate": 0.00019277193238297272, "loss": 4.527, "step": 467300 }, { "epoch": 0.8721270635107616, "grad_norm": 0.8425313234329224, "learning_rate": 0.0001927703641651653, "loss": 4.5368, "step": 467350 }, { "epoch": 0.8722203690701401, "grad_norm": 0.8112648725509644, "learning_rate": 0.00019276879578363453, "loss": 4.6862, "step": 467400 }, { "epoch": 0.8723136746295186, "grad_norm": 1.2792651653289795, "learning_rate": 0.00019276722723838313, "loss": 4.5265, "step": 467450 }, { "epoch": 0.8724069801888971, "grad_norm": 1.0517545938491821, "learning_rate": 0.00019276565852941393, "loss": 4.4714, "step": 467500 }, { "epoch": 0.8725002857482757, "grad_norm": 1.0767686367034912, "learning_rate": 0.00019276408965672963, "loss": 4.1743, "step": 467550 }, { "epoch": 0.8725935913076541, "grad_norm": 0.7544139623641968, "learning_rate": 0.00019276252062033305, "loss": 4.5288, "step": 467600 }, { "epoch": 0.8726868968670326, "grad_norm": 1.112797498703003, "learning_rate": 0.00019276095142022696, "loss": 4.6371, "step": 467650 }, { "epoch": 0.8727802024264111, "grad_norm": 1.1193571090698242, "learning_rate": 0.0001927593820564141, "loss": 4.4925, "step": 467700 }, { "epoch": 0.8728735079857896, "grad_norm": 1.1630029678344727, "learning_rate": 0.00019275781252889724, "loss": 4.5043, "step": 467750 }, { "epoch": 0.872966813545168, "grad_norm": 0.8309341669082642, "learning_rate": 0.00019275624283767917, "loss": 4.6645, "step": 467800 }, { "epoch": 0.8730601191045465, "grad_norm": 1.0450292825698853, "learning_rate": 0.00019275467298276265, "loss": 4.6422, "step": 467850 }, { "epoch": 0.8731534246639251, "grad_norm": 1.1610996723175049, "learning_rate": 0.00019275310296415046, "loss": 4.4824, "step": 467900 }, { "epoch": 0.8732467302233036, "grad_norm": 1.1184924840927124, "learning_rate": 0.00019275153278184535, "loss": 4.6829, "step": 467950 }, { "epoch": 0.873340035782682, "grad_norm": 1.1254422664642334, "learning_rate": 0.00019274996243585013, "loss": 4.4835, "step": 468000 }, { "epoch": 0.873340035782682, "eval_loss": 4.718130111694336, "eval_runtime": 231.0488, "eval_samples_per_second": 11.288, "eval_steps_per_second": 11.288, "eval_tts_loss": 7.57261305500359, "step": 468000 }, { "epoch": 0.8734333413420605, "grad_norm": 1.3211454153060913, "learning_rate": 0.00019274839192616754, "loss": 4.5638, "step": 468050 }, { "epoch": 0.873526646901439, "grad_norm": 1.1221386194229126, "learning_rate": 0.00019274682125280034, "loss": 4.5564, "step": 468100 }, { "epoch": 0.8736199524608175, "grad_norm": 1.482282042503357, "learning_rate": 0.00019274525041575134, "loss": 4.3688, "step": 468150 }, { "epoch": 0.873713258020196, "grad_norm": 1.0802263021469116, "learning_rate": 0.00019274367941502326, "loss": 4.6402, "step": 468200 }, { "epoch": 0.8738065635795744, "grad_norm": 1.2762898206710815, "learning_rate": 0.00019274210825061892, "loss": 4.4803, "step": 468250 }, { "epoch": 0.873899869138953, "grad_norm": 0.9525300860404968, "learning_rate": 0.0001927405369225411, "loss": 4.6589, "step": 468300 }, { "epoch": 0.8739931746983315, "grad_norm": 1.0424537658691406, "learning_rate": 0.00019273896543079252, "loss": 4.6008, "step": 468350 }, { "epoch": 0.87408648025771, "grad_norm": 1.0094565153121948, "learning_rate": 0.000192737393775376, "loss": 4.5487, "step": 468400 }, { "epoch": 0.8741797858170884, "grad_norm": 0.9425789713859558, "learning_rate": 0.0001927358219562943, "loss": 4.4299, "step": 468450 }, { "epoch": 0.8742730913764669, "grad_norm": 1.1396279335021973, "learning_rate": 0.00019273424997355018, "loss": 4.4761, "step": 468500 }, { "epoch": 0.8743663969358454, "grad_norm": 0.8912315368652344, "learning_rate": 0.00019273267782714645, "loss": 4.5606, "step": 468550 }, { "epoch": 0.8744597024952239, "grad_norm": 1.1854628324508667, "learning_rate": 0.00019273110551708581, "loss": 4.8195, "step": 468600 }, { "epoch": 0.8745530080546025, "grad_norm": 1.257662057876587, "learning_rate": 0.00019272953304337115, "loss": 4.539, "step": 468650 }, { "epoch": 0.8746463136139809, "grad_norm": 0.9374984502792358, "learning_rate": 0.00019272796040600514, "loss": 4.4405, "step": 468700 }, { "epoch": 0.8747396191733594, "grad_norm": 1.203815221786499, "learning_rate": 0.0001927263876049906, "loss": 4.5042, "step": 468750 }, { "epoch": 0.8748329247327379, "grad_norm": 0.884100079536438, "learning_rate": 0.0001927248146403303, "loss": 4.5122, "step": 468800 }, { "epoch": 0.8749262302921164, "grad_norm": 1.130448818206787, "learning_rate": 0.00019272324151202704, "loss": 4.6089, "step": 468850 }, { "epoch": 0.8750195358514948, "grad_norm": 1.006072998046875, "learning_rate": 0.00019272166822008354, "loss": 4.5081, "step": 468900 }, { "epoch": 0.8751128414108733, "grad_norm": 0.9598018527030945, "learning_rate": 0.00019272009476450265, "loss": 4.5977, "step": 468950 }, { "epoch": 0.8752061469702519, "grad_norm": 1.005086064338684, "learning_rate": 0.0001927185211452871, "loss": 4.5544, "step": 469000 }, { "epoch": 0.8752994525296304, "grad_norm": 1.1870617866516113, "learning_rate": 0.00019271694736243964, "loss": 4.4127, "step": 469050 }, { "epoch": 0.8753927580890088, "grad_norm": 1.27022385597229, "learning_rate": 0.0001927153734159631, "loss": 4.5925, "step": 469100 }, { "epoch": 0.8754860636483873, "grad_norm": 1.1191506385803223, "learning_rate": 0.00019271379930586025, "loss": 4.4178, "step": 469150 }, { "epoch": 0.8755793692077658, "grad_norm": 0.8403235077857971, "learning_rate": 0.00019271222503213384, "loss": 4.6836, "step": 469200 }, { "epoch": 0.8756726747671443, "grad_norm": 0.7562884092330933, "learning_rate": 0.0001927106505947867, "loss": 4.5828, "step": 469250 }, { "epoch": 0.8757659803265228, "grad_norm": 1.0119215250015259, "learning_rate": 0.00019270907599382153, "loss": 4.3888, "step": 469300 }, { "epoch": 0.8758592858859013, "grad_norm": 1.1594666242599487, "learning_rate": 0.00019270750122924116, "loss": 4.5207, "step": 469350 }, { "epoch": 0.8759525914452798, "grad_norm": 1.2950938940048218, "learning_rate": 0.0001927059263010484, "loss": 4.2749, "step": 469400 }, { "epoch": 0.8760458970046583, "grad_norm": 1.2516156435012817, "learning_rate": 0.00019270435120924596, "loss": 4.4011, "step": 469450 }, { "epoch": 0.8761392025640368, "grad_norm": 1.2195113897323608, "learning_rate": 0.00019270277595383667, "loss": 4.469, "step": 469500 }, { "epoch": 0.8762325081234152, "grad_norm": 1.1938484907150269, "learning_rate": 0.00019270120053482327, "loss": 4.5412, "step": 469550 }, { "epoch": 0.8763258136827937, "grad_norm": 1.0063114166259766, "learning_rate": 0.00019269962495220856, "loss": 4.4837, "step": 469600 }, { "epoch": 0.8764191192421722, "grad_norm": 1.1846363544464111, "learning_rate": 0.00019269804920599536, "loss": 4.5601, "step": 469650 }, { "epoch": 0.8765124248015508, "grad_norm": 1.1089617013931274, "learning_rate": 0.00019269647329618638, "loss": 4.6128, "step": 469700 }, { "epoch": 0.8766057303609293, "grad_norm": 0.6164752840995789, "learning_rate": 0.00019269489722278444, "loss": 4.3259, "step": 469750 }, { "epoch": 0.8766990359203077, "grad_norm": 1.0538560152053833, "learning_rate": 0.00019269332098579233, "loss": 4.3699, "step": 469800 }, { "epoch": 0.8767923414796862, "grad_norm": 1.1497721672058105, "learning_rate": 0.0001926917445852128, "loss": 4.6973, "step": 469850 }, { "epoch": 0.8768856470390647, "grad_norm": 1.0393162965774536, "learning_rate": 0.00019269016802104868, "loss": 4.503, "step": 469900 }, { "epoch": 0.8769789525984432, "grad_norm": 1.3194526433944702, "learning_rate": 0.0001926885912933027, "loss": 4.5962, "step": 469950 }, { "epoch": 0.8770722581578216, "grad_norm": 1.1103098392486572, "learning_rate": 0.00019268701440197766, "loss": 4.5427, "step": 470000 }, { "epoch": 0.8771655637172002, "grad_norm": 1.1723588705062866, "learning_rate": 0.0001926854373470764, "loss": 4.6448, "step": 470050 }, { "epoch": 0.8772588692765787, "grad_norm": 0.9415161609649658, "learning_rate": 0.0001926838601286016, "loss": 4.5021, "step": 470100 }, { "epoch": 0.8773521748359572, "grad_norm": 0.9576285481452942, "learning_rate": 0.0001926822827465561, "loss": 4.4223, "step": 470150 }, { "epoch": 0.8774454803953357, "grad_norm": 1.1041916608810425, "learning_rate": 0.00019268070520094268, "loss": 4.508, "step": 470200 }, { "epoch": 0.8775387859547141, "grad_norm": 0.9985270500183105, "learning_rate": 0.00019267912749176412, "loss": 4.6515, "step": 470250 }, { "epoch": 0.8776320915140926, "grad_norm": 1.2307016849517822, "learning_rate": 0.0001926775496190232, "loss": 4.5489, "step": 470300 }, { "epoch": 0.8777253970734711, "grad_norm": 0.8963343501091003, "learning_rate": 0.00019267597158272276, "loss": 4.382, "step": 470350 }, { "epoch": 0.8778187026328497, "grad_norm": 1.1342476606369019, "learning_rate": 0.0001926743933828655, "loss": 4.7212, "step": 470400 }, { "epoch": 0.8779120081922281, "grad_norm": 1.0419046878814697, "learning_rate": 0.00019267281501945426, "loss": 4.4446, "step": 470450 }, { "epoch": 0.8780053137516066, "grad_norm": 1.182022213935852, "learning_rate": 0.00019267123649249177, "loss": 4.3155, "step": 470500 }, { "epoch": 0.8780986193109851, "grad_norm": 1.6235601902008057, "learning_rate": 0.00019266965780198088, "loss": 4.8045, "step": 470550 }, { "epoch": 0.8781919248703636, "grad_norm": 1.2540438175201416, "learning_rate": 0.00019266807894792435, "loss": 4.4914, "step": 470600 }, { "epoch": 0.878285230429742, "grad_norm": 1.2668167352676392, "learning_rate": 0.00019266649993032498, "loss": 4.842, "step": 470650 }, { "epoch": 0.8783785359891205, "grad_norm": 0.9757852554321289, "learning_rate": 0.00019266492074918552, "loss": 4.568, "step": 470700 }, { "epoch": 0.8784718415484991, "grad_norm": 0.9917505979537964, "learning_rate": 0.00019266334140450876, "loss": 4.6792, "step": 470750 }, { "epoch": 0.8785651471078776, "grad_norm": 0.8268977999687195, "learning_rate": 0.00019266176189629755, "loss": 4.6346, "step": 470800 }, { "epoch": 0.8786584526672561, "grad_norm": 1.233041524887085, "learning_rate": 0.0001926601822245546, "loss": 4.4082, "step": 470850 }, { "epoch": 0.8787517582266345, "grad_norm": 1.2684880495071411, "learning_rate": 0.00019265860238928278, "loss": 4.5912, "step": 470900 }, { "epoch": 0.878845063786013, "grad_norm": 1.0635311603546143, "learning_rate": 0.0001926570223904848, "loss": 4.5625, "step": 470950 }, { "epoch": 0.8789383693453915, "grad_norm": 1.100784182548523, "learning_rate": 0.00019265544222816345, "loss": 4.5234, "step": 471000 }, { "epoch": 0.8789383693453915, "eval_loss": 4.717212677001953, "eval_runtime": 233.0282, "eval_samples_per_second": 11.192, "eval_steps_per_second": 11.192, "eval_tts_loss": 7.581809554389032, "step": 471000 }, { "epoch": 0.87903167490477, "grad_norm": 0.8170973658561707, "learning_rate": 0.00019265386190232158, "loss": 4.5747, "step": 471050 }, { "epoch": 0.8791249804641486, "grad_norm": 1.1318987607955933, "learning_rate": 0.00019265228141296197, "loss": 4.4058, "step": 471100 }, { "epoch": 0.879218286023527, "grad_norm": 1.101556658744812, "learning_rate": 0.00019265070076008732, "loss": 4.5053, "step": 471150 }, { "epoch": 0.8793115915829055, "grad_norm": 0.9891483783721924, "learning_rate": 0.00019264911994370057, "loss": 4.5975, "step": 471200 }, { "epoch": 0.879404897142284, "grad_norm": 0.858481228351593, "learning_rate": 0.00019264753896380436, "loss": 4.4779, "step": 471250 }, { "epoch": 0.8794982027016625, "grad_norm": 0.9057176113128662, "learning_rate": 0.00019264595782040157, "loss": 4.6907, "step": 471300 }, { "epoch": 0.8795915082610409, "grad_norm": 1.0125700235366821, "learning_rate": 0.00019264437651349495, "loss": 4.3944, "step": 471350 }, { "epoch": 0.8796848138204194, "grad_norm": 1.1874668598175049, "learning_rate": 0.00019264279504308733, "loss": 4.5879, "step": 471400 }, { "epoch": 0.879778119379798, "grad_norm": 1.0236549377441406, "learning_rate": 0.00019264121340918145, "loss": 4.4136, "step": 471450 }, { "epoch": 0.8798714249391765, "grad_norm": 1.091837763786316, "learning_rate": 0.00019263963161178017, "loss": 4.4745, "step": 471500 }, { "epoch": 0.879964730498555, "grad_norm": 1.1850827932357788, "learning_rate": 0.0001926380496508862, "loss": 4.721, "step": 471550 }, { "epoch": 0.8800580360579334, "grad_norm": 1.138193130493164, "learning_rate": 0.0001926364675265024, "loss": 4.4365, "step": 471600 }, { "epoch": 0.8801513416173119, "grad_norm": 1.1351327896118164, "learning_rate": 0.00019263488523863154, "loss": 4.497, "step": 471650 }, { "epoch": 0.8802446471766904, "grad_norm": 1.2687690258026123, "learning_rate": 0.00019263330278727637, "loss": 4.7238, "step": 471700 }, { "epoch": 0.8803379527360689, "grad_norm": 1.1612459421157837, "learning_rate": 0.00019263172017243974, "loss": 4.5756, "step": 471750 }, { "epoch": 0.8804312582954474, "grad_norm": 1.3393118381500244, "learning_rate": 0.00019263013739412443, "loss": 4.5649, "step": 471800 }, { "epoch": 0.8805245638548259, "grad_norm": 0.9592518210411072, "learning_rate": 0.00019262855445233322, "loss": 4.506, "step": 471850 }, { "epoch": 0.8806178694142044, "grad_norm": 1.2129544019699097, "learning_rate": 0.0001926269713470689, "loss": 4.7577, "step": 471900 }, { "epoch": 0.8807111749735829, "grad_norm": 0.8999137878417969, "learning_rate": 0.00019262538807833429, "loss": 4.5339, "step": 471950 }, { "epoch": 0.8808044805329613, "grad_norm": 1.5264843702316284, "learning_rate": 0.00019262380464613217, "loss": 4.6379, "step": 472000 }, { "epoch": 0.8808977860923398, "grad_norm": 1.079666018486023, "learning_rate": 0.0001926222210504653, "loss": 4.5105, "step": 472050 }, { "epoch": 0.8809910916517183, "grad_norm": 1.0563101768493652, "learning_rate": 0.00019262063729133655, "loss": 4.4208, "step": 472100 }, { "epoch": 0.8810843972110969, "grad_norm": 0.9801346063613892, "learning_rate": 0.00019261905336874865, "loss": 4.5998, "step": 472150 }, { "epoch": 0.8811777027704754, "grad_norm": 1.2503702640533447, "learning_rate": 0.00019261746928270443, "loss": 4.5336, "step": 472200 }, { "epoch": 0.8812710083298538, "grad_norm": 1.2237844467163086, "learning_rate": 0.00019261588503320667, "loss": 4.5463, "step": 472250 }, { "epoch": 0.8813643138892323, "grad_norm": 1.0349375009536743, "learning_rate": 0.00019261430062025815, "loss": 4.48, "step": 472300 }, { "epoch": 0.8814576194486108, "grad_norm": 1.0323736667633057, "learning_rate": 0.00019261271604386167, "loss": 4.6317, "step": 472350 }, { "epoch": 0.8815509250079893, "grad_norm": 1.0419288873672485, "learning_rate": 0.00019261113130402012, "loss": 4.3148, "step": 472400 }, { "epoch": 0.8816442305673677, "grad_norm": 1.0934524536132812, "learning_rate": 0.00019260954640073615, "loss": 4.3474, "step": 472450 }, { "epoch": 0.8817375361267463, "grad_norm": 0.9106244444847107, "learning_rate": 0.00019260796133401263, "loss": 4.6065, "step": 472500 }, { "epoch": 0.8818308416861248, "grad_norm": 0.8649581074714661, "learning_rate": 0.00019260637610385239, "loss": 4.6287, "step": 472550 }, { "epoch": 0.8819241472455033, "grad_norm": 1.1462935209274292, "learning_rate": 0.00019260479071025818, "loss": 4.7667, "step": 472600 }, { "epoch": 0.8820174528048818, "grad_norm": 1.1925376653671265, "learning_rate": 0.0001926032051532328, "loss": 4.5393, "step": 472650 }, { "epoch": 0.8821107583642602, "grad_norm": 1.1471788883209229, "learning_rate": 0.00019260161943277906, "loss": 4.5883, "step": 472700 }, { "epoch": 0.8822040639236387, "grad_norm": 1.2094416618347168, "learning_rate": 0.00019260003354889973, "loss": 4.5594, "step": 472750 }, { "epoch": 0.8822973694830172, "grad_norm": 1.0781878232955933, "learning_rate": 0.00019259844750159767, "loss": 4.4519, "step": 472800 }, { "epoch": 0.8823906750423958, "grad_norm": 0.9969003796577454, "learning_rate": 0.0001925968612908756, "loss": 4.2835, "step": 472850 }, { "epoch": 0.8824839806017742, "grad_norm": 1.024268627166748, "learning_rate": 0.0001925952749167364, "loss": 4.4837, "step": 472900 }, { "epoch": 0.8825772861611527, "grad_norm": 1.0069246292114258, "learning_rate": 0.00019259368837918283, "loss": 4.4673, "step": 472950 }, { "epoch": 0.8826705917205312, "grad_norm": 1.0265321731567383, "learning_rate": 0.0001925921016782177, "loss": 4.6297, "step": 473000 }, { "epoch": 0.8827638972799097, "grad_norm": 1.1687591075897217, "learning_rate": 0.0001925905148138438, "loss": 4.3603, "step": 473050 }, { "epoch": 0.8828572028392881, "grad_norm": 0.6202795505523682, "learning_rate": 0.00019258892778606387, "loss": 4.427, "step": 473100 }, { "epoch": 0.8829505083986666, "grad_norm": 1.0767467021942139, "learning_rate": 0.00019258734059488086, "loss": 4.468, "step": 473150 }, { "epoch": 0.8830438139580452, "grad_norm": 1.0983110666275024, "learning_rate": 0.00019258575324029745, "loss": 4.5833, "step": 473200 }, { "epoch": 0.8831371195174237, "grad_norm": 0.9307078123092651, "learning_rate": 0.00019258416572231647, "loss": 4.4465, "step": 473250 }, { "epoch": 0.8832304250768022, "grad_norm": 1.1207062005996704, "learning_rate": 0.00019258257804094072, "loss": 4.5858, "step": 473300 }, { "epoch": 0.8833237306361806, "grad_norm": 0.8835721015930176, "learning_rate": 0.00019258099019617305, "loss": 4.4705, "step": 473350 }, { "epoch": 0.8834170361955591, "grad_norm": 0.870012640953064, "learning_rate": 0.0001925794021880162, "loss": 4.5602, "step": 473400 }, { "epoch": 0.8835103417549376, "grad_norm": 1.3526321649551392, "learning_rate": 0.000192577814016473, "loss": 4.644, "step": 473450 }, { "epoch": 0.8836036473143161, "grad_norm": 0.9841532111167908, "learning_rate": 0.0001925762256815462, "loss": 4.5527, "step": 473500 }, { "epoch": 0.8836969528736947, "grad_norm": 0.9762299656867981, "learning_rate": 0.00019257463718323872, "loss": 4.5266, "step": 473550 }, { "epoch": 0.8837902584330731, "grad_norm": 1.3573507070541382, "learning_rate": 0.00019257304852155326, "loss": 4.5429, "step": 473600 }, { "epoch": 0.8838835639924516, "grad_norm": 1.0638320446014404, "learning_rate": 0.00019257145969649267, "loss": 4.6049, "step": 473650 }, { "epoch": 0.8839768695518301, "grad_norm": 1.035857081413269, "learning_rate": 0.00019256987070805974, "loss": 4.7841, "step": 473700 }, { "epoch": 0.8840701751112086, "grad_norm": 1.7121716737747192, "learning_rate": 0.00019256828155625726, "loss": 4.6054, "step": 473750 }, { "epoch": 0.884163480670587, "grad_norm": 0.9061555862426758, "learning_rate": 0.00019256669224108807, "loss": 4.4984, "step": 473800 }, { "epoch": 0.8842567862299655, "grad_norm": 1.1283390522003174, "learning_rate": 0.00019256510276255493, "loss": 4.5969, "step": 473850 }, { "epoch": 0.884350091789344, "grad_norm": 1.2580294609069824, "learning_rate": 0.00019256351312066069, "loss": 4.6179, "step": 473900 }, { "epoch": 0.8844433973487226, "grad_norm": 0.7480835318565369, "learning_rate": 0.00019256192331540815, "loss": 4.4874, "step": 473950 }, { "epoch": 0.884536702908101, "grad_norm": 1.211553692817688, "learning_rate": 0.0001925603333468001, "loss": 4.5896, "step": 474000 }, { "epoch": 0.884536702908101, "eval_loss": 4.7165398597717285, "eval_runtime": 228.5334, "eval_samples_per_second": 11.412, "eval_steps_per_second": 11.412, "eval_tts_loss": 7.573559369761897, "step": 474000 }, { "epoch": 0.8846300084674795, "grad_norm": 0.8259678483009338, "learning_rate": 0.00019255874321483934, "loss": 4.6858, "step": 474050 }, { "epoch": 0.884723314026858, "grad_norm": 1.3462460041046143, "learning_rate": 0.00019255715291952868, "loss": 4.6262, "step": 474100 }, { "epoch": 0.8848166195862365, "grad_norm": 0.9073434472084045, "learning_rate": 0.00019255556246087092, "loss": 4.5647, "step": 474150 }, { "epoch": 0.884909925145615, "grad_norm": 0.9772023558616638, "learning_rate": 0.00019255397183886889, "loss": 4.5206, "step": 474200 }, { "epoch": 0.8850032307049934, "grad_norm": 1.4259358644485474, "learning_rate": 0.00019255238105352537, "loss": 4.6257, "step": 474250 }, { "epoch": 0.885096536264372, "grad_norm": 1.1639713048934937, "learning_rate": 0.0001925507901048432, "loss": 4.5753, "step": 474300 }, { "epoch": 0.8851898418237505, "grad_norm": 0.9388675093650818, "learning_rate": 0.0001925491989928252, "loss": 4.4998, "step": 474350 }, { "epoch": 0.885283147383129, "grad_norm": 1.1765773296356201, "learning_rate": 0.0001925476077174741, "loss": 4.5805, "step": 474400 }, { "epoch": 0.8853764529425074, "grad_norm": 1.0637716054916382, "learning_rate": 0.00019254601627879278, "loss": 4.5969, "step": 474450 }, { "epoch": 0.8854697585018859, "grad_norm": 1.190686821937561, "learning_rate": 0.000192544424676784, "loss": 4.5881, "step": 474500 }, { "epoch": 0.8855630640612644, "grad_norm": 0.9555903077125549, "learning_rate": 0.0001925428329114506, "loss": 4.5869, "step": 474550 }, { "epoch": 0.8856563696206429, "grad_norm": 0.8813297152519226, "learning_rate": 0.00019254124098279542, "loss": 4.5683, "step": 474600 }, { "epoch": 0.8857496751800215, "grad_norm": 1.0782814025878906, "learning_rate": 0.0001925396488908212, "loss": 4.5528, "step": 474650 }, { "epoch": 0.8858429807393999, "grad_norm": 1.0104131698608398, "learning_rate": 0.00019253805663553078, "loss": 4.5352, "step": 474700 }, { "epoch": 0.8859362862987784, "grad_norm": 1.0496783256530762, "learning_rate": 0.00019253646421692698, "loss": 4.6384, "step": 474750 }, { "epoch": 0.8860295918581569, "grad_norm": 0.9454028010368347, "learning_rate": 0.00019253487163501263, "loss": 4.5314, "step": 474800 }, { "epoch": 0.8861228974175354, "grad_norm": 1.2104294300079346, "learning_rate": 0.0001925332788897905, "loss": 4.6148, "step": 474850 }, { "epoch": 0.8862162029769138, "grad_norm": 1.0959908962249756, "learning_rate": 0.0001925316859812634, "loss": 4.6213, "step": 474900 }, { "epoch": 0.8863095085362923, "grad_norm": 1.1584961414337158, "learning_rate": 0.00019253009290943416, "loss": 4.3093, "step": 474950 }, { "epoch": 0.8864028140956709, "grad_norm": 1.216267704963684, "learning_rate": 0.00019252849967430556, "loss": 4.6484, "step": 475000 }, { "epoch": 0.8864961196550494, "grad_norm": 1.154923439025879, "learning_rate": 0.0001925269062758805, "loss": 4.4519, "step": 475050 }, { "epoch": 0.8865894252144279, "grad_norm": 0.9570915102958679, "learning_rate": 0.0001925253127141617, "loss": 4.364, "step": 475100 }, { "epoch": 0.8866827307738063, "grad_norm": 1.0998491048812866, "learning_rate": 0.00019252371898915202, "loss": 4.4283, "step": 475150 }, { "epoch": 0.8867760363331848, "grad_norm": 0.9555975198745728, "learning_rate": 0.00019252212510085422, "loss": 4.6768, "step": 475200 }, { "epoch": 0.8868693418925633, "grad_norm": 1.030609369277954, "learning_rate": 0.00019252053104927116, "loss": 4.5128, "step": 475250 }, { "epoch": 0.8869626474519418, "grad_norm": 1.2955753803253174, "learning_rate": 0.00019251893683440563, "loss": 4.5385, "step": 475300 }, { "epoch": 0.8870559530113203, "grad_norm": 1.2106226682662964, "learning_rate": 0.00019251734245626047, "loss": 4.5328, "step": 475350 }, { "epoch": 0.8871492585706988, "grad_norm": 0.6228367686271667, "learning_rate": 0.0001925157479148385, "loss": 4.3444, "step": 475400 }, { "epoch": 0.8872425641300773, "grad_norm": 1.0899059772491455, "learning_rate": 0.0001925141532101425, "loss": 4.677, "step": 475450 }, { "epoch": 0.8873358696894558, "grad_norm": 1.1146882772445679, "learning_rate": 0.00019251255834217527, "loss": 4.4954, "step": 475500 }, { "epoch": 0.8874291752488342, "grad_norm": 0.9069426655769348, "learning_rate": 0.0001925109633109397, "loss": 4.4805, "step": 475550 }, { "epoch": 0.8875224808082127, "grad_norm": 0.8840944766998291, "learning_rate": 0.00019250936811643853, "loss": 4.6011, "step": 475600 }, { "epoch": 0.8876157863675912, "grad_norm": 1.4863438606262207, "learning_rate": 0.0001925077727586746, "loss": 4.4559, "step": 475650 }, { "epoch": 0.8877090919269698, "grad_norm": 0.9979962706565857, "learning_rate": 0.00019250617723765077, "loss": 4.6728, "step": 475700 }, { "epoch": 0.8878023974863483, "grad_norm": 1.2860273122787476, "learning_rate": 0.00019250458155336975, "loss": 4.315, "step": 475750 }, { "epoch": 0.8878957030457267, "grad_norm": 1.4837069511413574, "learning_rate": 0.00019250298570583447, "loss": 4.6044, "step": 475800 }, { "epoch": 0.8879890086051052, "grad_norm": 1.2858986854553223, "learning_rate": 0.00019250138969504764, "loss": 4.4225, "step": 475850 }, { "epoch": 0.8880823141644837, "grad_norm": 0.9045429229736328, "learning_rate": 0.0001924997935210122, "loss": 4.5414, "step": 475900 }, { "epoch": 0.8881756197238622, "grad_norm": 1.3608227968215942, "learning_rate": 0.0001924981971837309, "loss": 4.4982, "step": 475950 }, { "epoch": 0.8882689252832406, "grad_norm": 1.0310899019241333, "learning_rate": 0.0001924966006832065, "loss": 4.5696, "step": 476000 }, { "epoch": 0.8883622308426192, "grad_norm": 1.006451964378357, "learning_rate": 0.0001924950040194419, "loss": 4.5891, "step": 476050 }, { "epoch": 0.8884555364019977, "grad_norm": 1.0919299125671387, "learning_rate": 0.00019249340719243988, "loss": 4.613, "step": 476100 }, { "epoch": 0.8885488419613762, "grad_norm": 1.140575885772705, "learning_rate": 0.0001924918102022033, "loss": 4.4866, "step": 476150 }, { "epoch": 0.8886421475207547, "grad_norm": 1.1656068563461304, "learning_rate": 0.00019249021304873493, "loss": 4.7198, "step": 476200 }, { "epoch": 0.8887354530801331, "grad_norm": 0.9861620664596558, "learning_rate": 0.0001924886157320376, "loss": 4.3626, "step": 476250 }, { "epoch": 0.8888287586395116, "grad_norm": 0.9631748199462891, "learning_rate": 0.00019248701825211418, "loss": 4.6335, "step": 476300 }, { "epoch": 0.8889220641988901, "grad_norm": 1.142419457435608, "learning_rate": 0.0001924854206089674, "loss": 4.6723, "step": 476350 }, { "epoch": 0.8890153697582687, "grad_norm": 0.992556631565094, "learning_rate": 0.00019248382280260016, "loss": 4.6588, "step": 476400 }, { "epoch": 0.8891086753176471, "grad_norm": 1.15220308303833, "learning_rate": 0.00019248222483301525, "loss": 4.2161, "step": 476450 }, { "epoch": 0.8892019808770256, "grad_norm": 1.1019169092178345, "learning_rate": 0.00019248062670021546, "loss": 4.7114, "step": 476500 }, { "epoch": 0.8892952864364041, "grad_norm": 1.3275052309036255, "learning_rate": 0.00019247902840420364, "loss": 4.6296, "step": 476550 }, { "epoch": 0.8893885919957826, "grad_norm": 1.34925377368927, "learning_rate": 0.00019247742994498262, "loss": 4.534, "step": 476600 }, { "epoch": 0.889481897555161, "grad_norm": 0.9820336103439331, "learning_rate": 0.00019247583132255518, "loss": 4.6693, "step": 476650 }, { "epoch": 0.8895752031145395, "grad_norm": 1.1470335721969604, "learning_rate": 0.0001924742325369242, "loss": 4.7839, "step": 476700 }, { "epoch": 0.8896685086739181, "grad_norm": 0.916946291923523, "learning_rate": 0.00019247263358809246, "loss": 4.3625, "step": 476750 }, { "epoch": 0.8897618142332966, "grad_norm": 1.0986276865005493, "learning_rate": 0.0001924710344760628, "loss": 4.624, "step": 476800 }, { "epoch": 0.8898551197926751, "grad_norm": 1.1948660612106323, "learning_rate": 0.000192469435200838, "loss": 4.6542, "step": 476850 }, { "epoch": 0.8899484253520535, "grad_norm": 1.3349121809005737, "learning_rate": 0.00019246783576242095, "loss": 4.524, "step": 476900 }, { "epoch": 0.890041730911432, "grad_norm": 1.0939390659332275, "learning_rate": 0.00019246623616081443, "loss": 4.6037, "step": 476950 }, { "epoch": 0.8901350364708105, "grad_norm": 1.1679463386535645, "learning_rate": 0.00019246463639602127, "loss": 4.6764, "step": 477000 }, { "epoch": 0.8901350364708105, "eval_loss": 4.721210956573486, "eval_runtime": 229.3445, "eval_samples_per_second": 11.372, "eval_steps_per_second": 11.372, "eval_tts_loss": 7.548798499464635, "step": 477000 }, { "epoch": 0.890228342030189, "grad_norm": 0.8751832842826843, "learning_rate": 0.0001924630364680443, "loss": 4.5684, "step": 477050 }, { "epoch": 0.8903216475895676, "grad_norm": 0.9397994875907898, "learning_rate": 0.00019246143637688637, "loss": 4.4774, "step": 477100 }, { "epoch": 0.890414953148946, "grad_norm": 1.231675386428833, "learning_rate": 0.00019245983612255023, "loss": 4.2469, "step": 477150 }, { "epoch": 0.8905082587083245, "grad_norm": 1.0264906883239746, "learning_rate": 0.00019245823570503875, "loss": 4.7321, "step": 477200 }, { "epoch": 0.890601564267703, "grad_norm": 1.2082854509353638, "learning_rate": 0.00019245663512435478, "loss": 4.5844, "step": 477250 }, { "epoch": 0.8906948698270815, "grad_norm": 0.859123945236206, "learning_rate": 0.0001924550343805011, "loss": 4.4322, "step": 477300 }, { "epoch": 0.8907881753864599, "grad_norm": 1.140683889389038, "learning_rate": 0.00019245343347348056, "loss": 4.3265, "step": 477350 }, { "epoch": 0.8908814809458384, "grad_norm": 0.8733227252960205, "learning_rate": 0.00019245183240329595, "loss": 4.2769, "step": 477400 }, { "epoch": 0.890974786505217, "grad_norm": 0.8868075609207153, "learning_rate": 0.00019245023116995017, "loss": 4.4867, "step": 477450 }, { "epoch": 0.8910680920645955, "grad_norm": 1.0979993343353271, "learning_rate": 0.00019244862977344597, "loss": 4.4667, "step": 477500 }, { "epoch": 0.891161397623974, "grad_norm": 0.9554575085639954, "learning_rate": 0.00019244702821378623, "loss": 4.5473, "step": 477550 }, { "epoch": 0.8912547031833524, "grad_norm": 1.1561933755874634, "learning_rate": 0.0001924454264909737, "loss": 4.4871, "step": 477600 }, { "epoch": 0.8913480087427309, "grad_norm": 1.0048874616622925, "learning_rate": 0.00019244382460501132, "loss": 4.4055, "step": 477650 }, { "epoch": 0.8914413143021094, "grad_norm": 1.0914218425750732, "learning_rate": 0.00019244222255590182, "loss": 4.7203, "step": 477700 }, { "epoch": 0.8915346198614879, "grad_norm": 1.171195149421692, "learning_rate": 0.00019244062034364808, "loss": 4.5342, "step": 477750 }, { "epoch": 0.8916279254208664, "grad_norm": 0.9125056266784668, "learning_rate": 0.00019243901796825287, "loss": 4.7096, "step": 477800 }, { "epoch": 0.8917212309802449, "grad_norm": 1.0592591762542725, "learning_rate": 0.0001924374154297191, "loss": 4.8141, "step": 477850 }, { "epoch": 0.8918145365396234, "grad_norm": 1.0421098470687866, "learning_rate": 0.00019243581272804957, "loss": 4.4832, "step": 477900 }, { "epoch": 0.8919078420990019, "grad_norm": 1.03078293800354, "learning_rate": 0.00019243420986324704, "loss": 4.6159, "step": 477950 }, { "epoch": 0.8920011476583803, "grad_norm": 1.0375412702560425, "learning_rate": 0.00019243260683531442, "loss": 4.5823, "step": 478000 }, { "epoch": 0.8920944532177588, "grad_norm": 1.0121946334838867, "learning_rate": 0.00019243100364425454, "loss": 4.663, "step": 478050 }, { "epoch": 0.8921877587771373, "grad_norm": 1.274235725402832, "learning_rate": 0.00019242940029007017, "loss": 4.3813, "step": 478100 }, { "epoch": 0.8922810643365159, "grad_norm": 1.0843164920806885, "learning_rate": 0.0001924277967727642, "loss": 4.6905, "step": 478150 }, { "epoch": 0.8923743698958944, "grad_norm": 1.0751274824142456, "learning_rate": 0.0001924261930923394, "loss": 4.5964, "step": 478200 }, { "epoch": 0.8924676754552728, "grad_norm": 0.9134946465492249, "learning_rate": 0.00019242458924879867, "loss": 4.5149, "step": 478250 }, { "epoch": 0.8925609810146513, "grad_norm": 1.4338033199310303, "learning_rate": 0.00019242298524214477, "loss": 4.5166, "step": 478300 }, { "epoch": 0.8926542865740298, "grad_norm": 1.1398868560791016, "learning_rate": 0.00019242138107238057, "loss": 4.6697, "step": 478350 }, { "epoch": 0.8927475921334083, "grad_norm": 1.146594762802124, "learning_rate": 0.00019241977673950888, "loss": 4.402, "step": 478400 }, { "epoch": 0.8928408976927867, "grad_norm": 1.1539602279663086, "learning_rate": 0.0001924181722435326, "loss": 4.5183, "step": 478450 }, { "epoch": 0.8929342032521653, "grad_norm": 0.8117108345031738, "learning_rate": 0.00019241656758445442, "loss": 4.4973, "step": 478500 }, { "epoch": 0.8930275088115438, "grad_norm": 1.1321312189102173, "learning_rate": 0.00019241496276227734, "loss": 4.742, "step": 478550 }, { "epoch": 0.8931208143709223, "grad_norm": 0.8790409564971924, "learning_rate": 0.00019241335777700407, "loss": 4.2764, "step": 478600 }, { "epoch": 0.8932141199303008, "grad_norm": 0.9671831727027893, "learning_rate": 0.0001924117526286375, "loss": 4.4868, "step": 478650 }, { "epoch": 0.8933074254896792, "grad_norm": 1.0273115634918213, "learning_rate": 0.00019241014731718044, "loss": 4.721, "step": 478700 }, { "epoch": 0.8934007310490577, "grad_norm": 1.0558019876480103, "learning_rate": 0.00019240854184263576, "loss": 4.5739, "step": 478750 }, { "epoch": 0.8934940366084362, "grad_norm": 0.8809823989868164, "learning_rate": 0.00019240693620500623, "loss": 4.3934, "step": 478800 }, { "epoch": 0.8935873421678148, "grad_norm": 0.9690717458724976, "learning_rate": 0.0001924053304042947, "loss": 4.3148, "step": 478850 }, { "epoch": 0.8936806477271932, "grad_norm": 1.4946317672729492, "learning_rate": 0.00019240372444050404, "loss": 4.532, "step": 478900 }, { "epoch": 0.8937739532865717, "grad_norm": 1.072158932685852, "learning_rate": 0.0001924021183136371, "loss": 4.5442, "step": 478950 }, { "epoch": 0.8938672588459502, "grad_norm": 1.1257818937301636, "learning_rate": 0.00019240051202369665, "loss": 4.5442, "step": 479000 }, { "epoch": 0.8939605644053287, "grad_norm": 1.0502753257751465, "learning_rate": 0.00019239890557068553, "loss": 4.6461, "step": 479050 }, { "epoch": 0.8940538699647071, "grad_norm": 1.0350525379180908, "learning_rate": 0.00019239729895460662, "loss": 4.427, "step": 479100 }, { "epoch": 0.8941471755240856, "grad_norm": 1.0828386545181274, "learning_rate": 0.00019239569217546274, "loss": 4.2337, "step": 479150 }, { "epoch": 0.8942404810834641, "grad_norm": 1.095841884613037, "learning_rate": 0.00019239408523325673, "loss": 4.5154, "step": 479200 }, { "epoch": 0.8943337866428427, "grad_norm": 1.6823346614837646, "learning_rate": 0.0001923924781279914, "loss": 4.4984, "step": 479250 }, { "epoch": 0.8944270922022212, "grad_norm": 1.506416916847229, "learning_rate": 0.0001923908708596696, "loss": 4.5688, "step": 479300 }, { "epoch": 0.8945203977615996, "grad_norm": 0.8799375891685486, "learning_rate": 0.00019238926342829418, "loss": 4.5935, "step": 479350 }, { "epoch": 0.8946137033209781, "grad_norm": 0.7702896595001221, "learning_rate": 0.00019238765583386798, "loss": 4.6526, "step": 479400 }, { "epoch": 0.8947070088803566, "grad_norm": 0.8651245832443237, "learning_rate": 0.0001923860480763938, "loss": 4.399, "step": 479450 }, { "epoch": 0.8948003144397351, "grad_norm": 1.3018206357955933, "learning_rate": 0.0001923844401558745, "loss": 4.4704, "step": 479500 }, { "epoch": 0.8948936199991135, "grad_norm": 1.1552797555923462, "learning_rate": 0.00019238283207231294, "loss": 4.5859, "step": 479550 }, { "epoch": 0.8949869255584921, "grad_norm": 1.1593358516693115, "learning_rate": 0.00019238122382571194, "loss": 4.5231, "step": 479600 }, { "epoch": 0.8950802311178706, "grad_norm": 0.9813491106033325, "learning_rate": 0.0001923796154160743, "loss": 4.4514, "step": 479650 }, { "epoch": 0.8951735366772491, "grad_norm": 1.2231067419052124, "learning_rate": 0.0001923780068434029, "loss": 4.5498, "step": 479700 }, { "epoch": 0.8952668422366276, "grad_norm": 1.049607515335083, "learning_rate": 0.0001923763981077006, "loss": 4.6033, "step": 479750 }, { "epoch": 0.895360147796006, "grad_norm": 0.9342153668403625, "learning_rate": 0.00019237478920897021, "loss": 4.6114, "step": 479800 }, { "epoch": 0.8954534533553845, "grad_norm": 1.157514214515686, "learning_rate": 0.00019237318014721455, "loss": 4.7431, "step": 479850 }, { "epoch": 0.895546758914763, "grad_norm": 0.933319091796875, "learning_rate": 0.0001923715709224365, "loss": 4.4688, "step": 479900 }, { "epoch": 0.8956400644741416, "grad_norm": 0.7947107553482056, "learning_rate": 0.00019236996153463887, "loss": 4.323, "step": 479950 }, { "epoch": 0.89573337003352, "grad_norm": 0.724888265132904, "learning_rate": 0.0001923683519838245, "loss": 4.4641, "step": 480000 }, { "epoch": 0.89573337003352, "eval_loss": 4.723694324493408, "eval_runtime": 231.6982, "eval_samples_per_second": 11.256, "eval_steps_per_second": 11.256, "eval_tts_loss": 7.5714161622300855, "step": 480000 }, { "epoch": 0.8958266755928985, "grad_norm": 0.945941150188446, "learning_rate": 0.00019236674226999625, "loss": 4.4814, "step": 480050 }, { "epoch": 0.895919981152277, "grad_norm": 1.1227316856384277, "learning_rate": 0.00019236513239315696, "loss": 4.4636, "step": 480100 }, { "epoch": 0.8960132867116555, "grad_norm": 1.5031123161315918, "learning_rate": 0.00019236352235330945, "loss": 4.412, "step": 480150 }, { "epoch": 0.896106592271034, "grad_norm": 1.1095486879348755, "learning_rate": 0.0001923619121504566, "loss": 4.4956, "step": 480200 }, { "epoch": 0.8961998978304124, "grad_norm": 1.297509789466858, "learning_rate": 0.00019236030178460117, "loss": 4.5368, "step": 480250 }, { "epoch": 0.896293203389791, "grad_norm": 1.0762945413589478, "learning_rate": 0.00019235869125574612, "loss": 4.6397, "step": 480300 }, { "epoch": 0.8963865089491695, "grad_norm": 1.0022764205932617, "learning_rate": 0.00019235708056389422, "loss": 4.596, "step": 480350 }, { "epoch": 0.896479814508548, "grad_norm": 1.223881483078003, "learning_rate": 0.0001923554697090483, "loss": 4.4484, "step": 480400 }, { "epoch": 0.8965731200679264, "grad_norm": 1.0800113677978516, "learning_rate": 0.00019235385869121124, "loss": 4.6962, "step": 480450 }, { "epoch": 0.8966664256273049, "grad_norm": 1.145907998085022, "learning_rate": 0.00019235224751038588, "loss": 4.5875, "step": 480500 }, { "epoch": 0.8967597311866834, "grad_norm": 1.1217701435089111, "learning_rate": 0.00019235063616657505, "loss": 4.6838, "step": 480550 }, { "epoch": 0.8968530367460619, "grad_norm": 0.9556708335876465, "learning_rate": 0.00019234902465978157, "loss": 4.5025, "step": 480600 }, { "epoch": 0.8969463423054405, "grad_norm": 0.7880034446716309, "learning_rate": 0.00019234741299000834, "loss": 4.5818, "step": 480650 }, { "epoch": 0.8970396478648189, "grad_norm": 1.0314735174179077, "learning_rate": 0.00019234580115725814, "loss": 4.5423, "step": 480700 }, { "epoch": 0.8971329534241974, "grad_norm": 0.7877832055091858, "learning_rate": 0.0001923441891615339, "loss": 4.5966, "step": 480750 }, { "epoch": 0.8972262589835759, "grad_norm": 1.1641756296157837, "learning_rate": 0.00019234257700283838, "loss": 4.6077, "step": 480800 }, { "epoch": 0.8973195645429544, "grad_norm": 1.1258875131607056, "learning_rate": 0.00019234096468117445, "loss": 4.5919, "step": 480850 }, { "epoch": 0.8974128701023328, "grad_norm": 1.0736488103866577, "learning_rate": 0.000192339352196545, "loss": 4.7133, "step": 480900 }, { "epoch": 0.8975061756617113, "grad_norm": 1.4151861667633057, "learning_rate": 0.0001923377395489528, "loss": 4.7003, "step": 480950 }, { "epoch": 0.8975994812210899, "grad_norm": 1.0960739850997925, "learning_rate": 0.0001923361267384008, "loss": 4.6995, "step": 481000 }, { "epoch": 0.8976927867804684, "grad_norm": 0.7709876298904419, "learning_rate": 0.00019233451376489172, "loss": 4.6444, "step": 481050 }, { "epoch": 0.8977860923398469, "grad_norm": 1.220925211906433, "learning_rate": 0.0001923329006284285, "loss": 4.4667, "step": 481100 }, { "epoch": 0.8978793978992253, "grad_norm": 1.09330415725708, "learning_rate": 0.00019233128732901395, "loss": 4.6498, "step": 481150 }, { "epoch": 0.8979727034586038, "grad_norm": 1.0120528936386108, "learning_rate": 0.00019232967386665094, "loss": 4.6587, "step": 481200 }, { "epoch": 0.8980660090179823, "grad_norm": 1.14932119846344, "learning_rate": 0.00019232806024134226, "loss": 4.6079, "step": 481250 }, { "epoch": 0.8981593145773608, "grad_norm": 0.8780590295791626, "learning_rate": 0.00019232644645309085, "loss": 4.4652, "step": 481300 }, { "epoch": 0.8982526201367393, "grad_norm": 1.0543479919433594, "learning_rate": 0.00019232483250189949, "loss": 4.5014, "step": 481350 }, { "epoch": 0.8983459256961178, "grad_norm": 0.7677117586135864, "learning_rate": 0.00019232321838777102, "loss": 4.3553, "step": 481400 }, { "epoch": 0.8984392312554963, "grad_norm": 1.0454624891281128, "learning_rate": 0.00019232160411070833, "loss": 4.3821, "step": 481450 }, { "epoch": 0.8985325368148748, "grad_norm": 1.2625417709350586, "learning_rate": 0.00019231998967071426, "loss": 4.5055, "step": 481500 }, { "epoch": 0.8986258423742532, "grad_norm": 0.7533479332923889, "learning_rate": 0.00019231837506779166, "loss": 4.708, "step": 481550 }, { "epoch": 0.8987191479336317, "grad_norm": 1.1038877964019775, "learning_rate": 0.00019231676030194333, "loss": 4.4323, "step": 481600 }, { "epoch": 0.8988124534930102, "grad_norm": 1.19779634475708, "learning_rate": 0.00019231514537317224, "loss": 4.5693, "step": 481650 }, { "epoch": 0.8989057590523888, "grad_norm": 1.1061819791793823, "learning_rate": 0.00019231353028148108, "loss": 4.5401, "step": 481700 }, { "epoch": 0.8989990646117673, "grad_norm": 1.0342915058135986, "learning_rate": 0.00019231191502687284, "loss": 4.3222, "step": 481750 }, { "epoch": 0.8990923701711457, "grad_norm": 1.1219396591186523, "learning_rate": 0.00019231029960935028, "loss": 4.6153, "step": 481800 }, { "epoch": 0.8991856757305242, "grad_norm": 1.1243091821670532, "learning_rate": 0.00019230868402891628, "loss": 4.384, "step": 481850 }, { "epoch": 0.8992789812899027, "grad_norm": 0.8768875598907471, "learning_rate": 0.00019230706828557372, "loss": 4.7314, "step": 481900 }, { "epoch": 0.8993722868492812, "grad_norm": 0.9880782961845398, "learning_rate": 0.00019230545237932542, "loss": 4.3457, "step": 481950 }, { "epoch": 0.8994655924086596, "grad_norm": 1.1840072870254517, "learning_rate": 0.00019230383631017422, "loss": 4.5494, "step": 482000 }, { "epoch": 0.8995588979680382, "grad_norm": 1.0306917428970337, "learning_rate": 0.000192302220078123, "loss": 4.4978, "step": 482050 }, { "epoch": 0.8996522035274167, "grad_norm": 1.2749229669570923, "learning_rate": 0.00019230060368317461, "loss": 4.5083, "step": 482100 }, { "epoch": 0.8997455090867952, "grad_norm": 0.9335286021232605, "learning_rate": 0.0001922989871253319, "loss": 4.4035, "step": 482150 }, { "epoch": 0.8998388146461737, "grad_norm": 1.0909782648086548, "learning_rate": 0.0001922973704045977, "loss": 4.5027, "step": 482200 }, { "epoch": 0.8999321202055521, "grad_norm": 1.3790338039398193, "learning_rate": 0.00019229575352097488, "loss": 4.6202, "step": 482250 }, { "epoch": 0.9000254257649306, "grad_norm": 1.0716173648834229, "learning_rate": 0.00019229413647446633, "loss": 4.5974, "step": 482300 }, { "epoch": 0.9001187313243091, "grad_norm": 0.884589672088623, "learning_rate": 0.00019229251926507486, "loss": 4.5097, "step": 482350 }, { "epoch": 0.9002120368836877, "grad_norm": 0.9533361196517944, "learning_rate": 0.0001922909018928033, "loss": 4.4418, "step": 482400 }, { "epoch": 0.9003053424430661, "grad_norm": 1.0003838539123535, "learning_rate": 0.00019228928435765456, "loss": 4.6466, "step": 482450 }, { "epoch": 0.9003986480024446, "grad_norm": 1.1342599391937256, "learning_rate": 0.00019228766665963147, "loss": 4.4474, "step": 482500 }, { "epoch": 0.9004919535618231, "grad_norm": 1.0784145593643188, "learning_rate": 0.0001922860487987369, "loss": 4.4095, "step": 482550 }, { "epoch": 0.9005852591212016, "grad_norm": 0.9923028349876404, "learning_rate": 0.00019228443077497364, "loss": 4.6017, "step": 482600 }, { "epoch": 0.90067856468058, "grad_norm": 1.0136337280273438, "learning_rate": 0.00019228281258834465, "loss": 4.3578, "step": 482650 }, { "epoch": 0.9007718702399585, "grad_norm": 1.2821561098098755, "learning_rate": 0.00019228119423885272, "loss": 4.7055, "step": 482700 }, { "epoch": 0.9008651757993371, "grad_norm": 0.7629845142364502, "learning_rate": 0.00019227957572650071, "loss": 4.3653, "step": 482750 }, { "epoch": 0.9009584813587156, "grad_norm": 1.0491293668746948, "learning_rate": 0.0001922779570512915, "loss": 4.495, "step": 482800 }, { "epoch": 0.9010517869180941, "grad_norm": 1.2240322828292847, "learning_rate": 0.0001922763382132279, "loss": 4.6833, "step": 482850 }, { "epoch": 0.9011450924774725, "grad_norm": 1.1035583019256592, "learning_rate": 0.00019227471921231284, "loss": 4.5753, "step": 482900 }, { "epoch": 0.901238398036851, "grad_norm": 1.0582516193389893, "learning_rate": 0.0001922731000485491, "loss": 4.5063, "step": 482950 }, { "epoch": 0.9013317035962295, "grad_norm": 1.1088228225708008, "learning_rate": 0.00019227148072193962, "loss": 4.6792, "step": 483000 }, { "epoch": 0.9013317035962295, "eval_loss": 4.713059902191162, "eval_runtime": 232.9395, "eval_samples_per_second": 11.196, "eval_steps_per_second": 11.196, "eval_tts_loss": 7.533985597230037, "step": 483000 }, { "epoch": 0.901425009155608, "grad_norm": 1.0973457098007202, "learning_rate": 0.0001922698612324872, "loss": 4.6279, "step": 483050 }, { "epoch": 0.9015183147149866, "grad_norm": 0.6816809773445129, "learning_rate": 0.00019226824158019467, "loss": 4.5961, "step": 483100 }, { "epoch": 0.901611620274365, "grad_norm": 1.4455472230911255, "learning_rate": 0.00019226662176506496, "loss": 4.3883, "step": 483150 }, { "epoch": 0.9017049258337435, "grad_norm": 1.094477653503418, "learning_rate": 0.0001922650017871009, "loss": 4.7538, "step": 483200 }, { "epoch": 0.901798231393122, "grad_norm": 0.9077624082565308, "learning_rate": 0.00019226338164630535, "loss": 4.4936, "step": 483250 }, { "epoch": 0.9018915369525005, "grad_norm": 1.3698067665100098, "learning_rate": 0.00019226176134268114, "loss": 4.6575, "step": 483300 }, { "epoch": 0.9019848425118789, "grad_norm": 1.1963516473770142, "learning_rate": 0.00019226014087623118, "loss": 4.5284, "step": 483350 }, { "epoch": 0.9020781480712574, "grad_norm": 1.1139675378799438, "learning_rate": 0.00019225852024695828, "loss": 4.3756, "step": 483400 }, { "epoch": 0.902171453630636, "grad_norm": 1.2275303602218628, "learning_rate": 0.00019225689945486535, "loss": 4.5184, "step": 483450 }, { "epoch": 0.9022647591900145, "grad_norm": 0.9934277534484863, "learning_rate": 0.00019225527849995523, "loss": 4.6678, "step": 483500 }, { "epoch": 0.902358064749393, "grad_norm": 1.0899839401245117, "learning_rate": 0.00019225365738223077, "loss": 4.4406, "step": 483550 }, { "epoch": 0.9024513703087714, "grad_norm": 1.3048666715621948, "learning_rate": 0.0001922520361016948, "loss": 4.6652, "step": 483600 }, { "epoch": 0.9025446758681499, "grad_norm": 0.9421764016151428, "learning_rate": 0.00019225041465835028, "loss": 4.4843, "step": 483650 }, { "epoch": 0.9026379814275284, "grad_norm": 0.9295839071273804, "learning_rate": 0.00019224879305219995, "loss": 4.4873, "step": 483700 }, { "epoch": 0.9027312869869069, "grad_norm": 1.1828733682632446, "learning_rate": 0.00019224717128324677, "loss": 4.4558, "step": 483750 }, { "epoch": 0.9028245925462854, "grad_norm": 0.7866224646568298, "learning_rate": 0.00019224554935149353, "loss": 4.5954, "step": 483800 }, { "epoch": 0.9029178981056639, "grad_norm": 1.3195923566818237, "learning_rate": 0.00019224392725694317, "loss": 4.5692, "step": 483850 }, { "epoch": 0.9030112036650424, "grad_norm": 1.182767391204834, "learning_rate": 0.0001922423049995985, "loss": 4.3976, "step": 483900 }, { "epoch": 0.9031045092244209, "grad_norm": 0.8278961777687073, "learning_rate": 0.0001922406825794624, "loss": 4.5223, "step": 483950 }, { "epoch": 0.9031978147837993, "grad_norm": 1.4472931623458862, "learning_rate": 0.00019223905999653769, "loss": 4.7193, "step": 484000 }, { "epoch": 0.9032911203431778, "grad_norm": 1.030922770500183, "learning_rate": 0.00019223743725082728, "loss": 4.5075, "step": 484050 }, { "epoch": 0.9033844259025563, "grad_norm": 1.1027861833572388, "learning_rate": 0.00019223581434233404, "loss": 4.6115, "step": 484100 }, { "epoch": 0.9034777314619349, "grad_norm": 1.1541696786880493, "learning_rate": 0.0001922341912710608, "loss": 4.7336, "step": 484150 }, { "epoch": 0.9035710370213134, "grad_norm": 1.1733508110046387, "learning_rate": 0.00019223256803701047, "loss": 4.6465, "step": 484200 }, { "epoch": 0.9036643425806918, "grad_norm": 1.0827579498291016, "learning_rate": 0.00019223094464018584, "loss": 4.5089, "step": 484250 }, { "epoch": 0.9037576481400703, "grad_norm": 1.2341217994689941, "learning_rate": 0.00019222932108058983, "loss": 4.5899, "step": 484300 }, { "epoch": 0.9038509536994488, "grad_norm": 0.9782057404518127, "learning_rate": 0.00019222769735822534, "loss": 4.6837, "step": 484350 }, { "epoch": 0.9039442592588273, "grad_norm": 1.100797414779663, "learning_rate": 0.0001922260734730952, "loss": 4.6549, "step": 484400 }, { "epoch": 0.9040375648182057, "grad_norm": 0.7023369669914246, "learning_rate": 0.00019222444942520222, "loss": 4.6172, "step": 484450 }, { "epoch": 0.9041308703775842, "grad_norm": 1.3590797185897827, "learning_rate": 0.0001922228252145493, "loss": 4.4214, "step": 484500 }, { "epoch": 0.9042241759369628, "grad_norm": 1.2530616521835327, "learning_rate": 0.00019222120084113936, "loss": 4.5571, "step": 484550 }, { "epoch": 0.9043174814963413, "grad_norm": 0.9176886677742004, "learning_rate": 0.00019221957630497522, "loss": 4.261, "step": 484600 }, { "epoch": 0.9044107870557198, "grad_norm": 1.4142929315567017, "learning_rate": 0.00019221795160605972, "loss": 4.7505, "step": 484650 }, { "epoch": 0.9045040926150982, "grad_norm": 0.9088461995124817, "learning_rate": 0.0001922163267443958, "loss": 4.678, "step": 484700 }, { "epoch": 0.9045973981744767, "grad_norm": 1.0148144960403442, "learning_rate": 0.00019221470171998628, "loss": 4.5586, "step": 484750 }, { "epoch": 0.9046907037338552, "grad_norm": 1.0496110916137695, "learning_rate": 0.00019221307653283405, "loss": 4.6296, "step": 484800 }, { "epoch": 0.9047840092932337, "grad_norm": 1.1175556182861328, "learning_rate": 0.00019221145118294194, "loss": 4.5289, "step": 484850 }, { "epoch": 0.9048773148526122, "grad_norm": 1.1451400518417358, "learning_rate": 0.00019220982567031287, "loss": 4.6588, "step": 484900 }, { "epoch": 0.9049706204119907, "grad_norm": 0.876974880695343, "learning_rate": 0.00019220819999494966, "loss": 4.3353, "step": 484950 }, { "epoch": 0.9050639259713692, "grad_norm": 1.1902146339416504, "learning_rate": 0.00019220657415685525, "loss": 4.6245, "step": 485000 }, { "epoch": 0.9051572315307477, "grad_norm": 1.2148951292037964, "learning_rate": 0.0001922049481560324, "loss": 4.7467, "step": 485050 }, { "epoch": 0.9052505370901262, "grad_norm": 0.8775200843811035, "learning_rate": 0.00019220332199248406, "loss": 4.7016, "step": 485100 }, { "epoch": 0.9053438426495046, "grad_norm": 1.0701316595077515, "learning_rate": 0.00019220169566621306, "loss": 4.3375, "step": 485150 }, { "epoch": 0.9054371482088831, "grad_norm": 0.9478065967559814, "learning_rate": 0.00019220006917722234, "loss": 4.3327, "step": 485200 }, { "epoch": 0.9055304537682617, "grad_norm": 0.8785211443901062, "learning_rate": 0.0001921984425255147, "loss": 4.674, "step": 485250 }, { "epoch": 0.9056237593276402, "grad_norm": 1.0310087203979492, "learning_rate": 0.00019219681571109302, "loss": 4.4732, "step": 485300 }, { "epoch": 0.9057170648870186, "grad_norm": 1.1273384094238281, "learning_rate": 0.0001921951887339602, "loss": 4.5746, "step": 485350 }, { "epoch": 0.9058103704463971, "grad_norm": 1.277830719947815, "learning_rate": 0.00019219356159411906, "loss": 4.4078, "step": 485400 }, { "epoch": 0.9059036760057756, "grad_norm": 1.1626276969909668, "learning_rate": 0.00019219193429157252, "loss": 4.5481, "step": 485450 }, { "epoch": 0.9059969815651541, "grad_norm": 1.0158990621566772, "learning_rate": 0.00019219030682632345, "loss": 4.454, "step": 485500 }, { "epoch": 0.9060902871245325, "grad_norm": 1.0277559757232666, "learning_rate": 0.00019218867919837472, "loss": 4.7083, "step": 485550 }, { "epoch": 0.9061835926839111, "grad_norm": 1.1720540523529053, "learning_rate": 0.00019218705140772917, "loss": 4.3923, "step": 485600 }, { "epoch": 0.9062768982432896, "grad_norm": 1.3606492280960083, "learning_rate": 0.0001921854234543897, "loss": 4.6602, "step": 485650 }, { "epoch": 0.9063702038026681, "grad_norm": 1.3768712282180786, "learning_rate": 0.00019218379533835918, "loss": 4.2959, "step": 485700 }, { "epoch": 0.9064635093620466, "grad_norm": 0.9553027749061584, "learning_rate": 0.00019218216705964048, "loss": 4.5227, "step": 485750 }, { "epoch": 0.906556814921425, "grad_norm": 1.0254628658294678, "learning_rate": 0.00019218053861823648, "loss": 4.433, "step": 485800 }, { "epoch": 0.9066501204808035, "grad_norm": 1.1287095546722412, "learning_rate": 0.00019217891001415004, "loss": 4.5025, "step": 485850 }, { "epoch": 0.906743426040182, "grad_norm": 1.2820013761520386, "learning_rate": 0.00019217728124738405, "loss": 4.6763, "step": 485900 }, { "epoch": 0.9068367315995606, "grad_norm": 1.1369165182113647, "learning_rate": 0.00019217565231794139, "loss": 4.4951, "step": 485950 }, { "epoch": 0.906930037158939, "grad_norm": 1.1727133989334106, "learning_rate": 0.00019217402322582492, "loss": 4.5462, "step": 486000 }, { "epoch": 0.906930037158939, "eval_loss": 4.708469867706299, "eval_runtime": 233.4449, "eval_samples_per_second": 11.172, "eval_steps_per_second": 11.172, "eval_tts_loss": 7.5771399114875155, "step": 486000 }, { "epoch": 0.9070233427183175, "grad_norm": 0.9646222591400146, "learning_rate": 0.0001921723939710375, "loss": 4.5943, "step": 486050 }, { "epoch": 0.907116648277696, "grad_norm": 1.20554780960083, "learning_rate": 0.00019217076455358204, "loss": 4.6885, "step": 486100 }, { "epoch": 0.9072099538370745, "grad_norm": 0.9490997195243835, "learning_rate": 0.00019216913497346137, "loss": 4.6716, "step": 486150 }, { "epoch": 0.907303259396453, "grad_norm": 0.9042333364486694, "learning_rate": 0.00019216750523067843, "loss": 4.6691, "step": 486200 }, { "epoch": 0.9073965649558314, "grad_norm": 0.9545812606811523, "learning_rate": 0.00019216587532523604, "loss": 4.4417, "step": 486250 }, { "epoch": 0.90748987051521, "grad_norm": 0.8847161531448364, "learning_rate": 0.0001921642452571371, "loss": 4.5915, "step": 486300 }, { "epoch": 0.9075831760745885, "grad_norm": 0.8875687122344971, "learning_rate": 0.00019216261502638448, "loss": 4.5054, "step": 486350 }, { "epoch": 0.907676481633967, "grad_norm": 1.1414605379104614, "learning_rate": 0.00019216098463298108, "loss": 4.4134, "step": 486400 }, { "epoch": 0.9077697871933454, "grad_norm": 1.0377355813980103, "learning_rate": 0.00019215935407692974, "loss": 4.4815, "step": 486450 }, { "epoch": 0.9078630927527239, "grad_norm": 1.0772016048431396, "learning_rate": 0.00019215772335823338, "loss": 4.5782, "step": 486500 }, { "epoch": 0.9079563983121024, "grad_norm": 1.034775972366333, "learning_rate": 0.00019215609247689482, "loss": 4.7385, "step": 486550 }, { "epoch": 0.9080497038714809, "grad_norm": 1.3947745561599731, "learning_rate": 0.00019215446143291699, "loss": 4.7174, "step": 486600 }, { "epoch": 0.9081430094308595, "grad_norm": 1.2977344989776611, "learning_rate": 0.00019215283022630276, "loss": 4.463, "step": 486650 }, { "epoch": 0.9082363149902379, "grad_norm": 1.2110722064971924, "learning_rate": 0.00019215119885705498, "loss": 4.5484, "step": 486700 }, { "epoch": 0.9083296205496164, "grad_norm": 0.9305819869041443, "learning_rate": 0.00019214956732517655, "loss": 4.3747, "step": 486750 }, { "epoch": 0.9084229261089949, "grad_norm": 1.0282576084136963, "learning_rate": 0.00019214793563067034, "loss": 4.6036, "step": 486800 }, { "epoch": 0.9085162316683734, "grad_norm": 1.0480084419250488, "learning_rate": 0.00019214630377353925, "loss": 4.6764, "step": 486850 }, { "epoch": 0.9086095372277518, "grad_norm": 1.1996289491653442, "learning_rate": 0.00019214467175378616, "loss": 4.5765, "step": 486900 }, { "epoch": 0.9087028427871303, "grad_norm": 1.3886234760284424, "learning_rate": 0.00019214303957141394, "loss": 4.5513, "step": 486950 }, { "epoch": 0.9087961483465089, "grad_norm": 1.0197093486785889, "learning_rate": 0.00019214140722642543, "loss": 4.6382, "step": 487000 }, { "epoch": 0.9088894539058874, "grad_norm": 1.2810677289962769, "learning_rate": 0.00019213977471882357, "loss": 4.4208, "step": 487050 }, { "epoch": 0.9089827594652659, "grad_norm": 1.0516036748886108, "learning_rate": 0.00019213814204861125, "loss": 4.5122, "step": 487100 }, { "epoch": 0.9090760650246443, "grad_norm": 0.9273803234100342, "learning_rate": 0.00019213650921579128, "loss": 4.6194, "step": 487150 }, { "epoch": 0.9091693705840228, "grad_norm": 0.9306402802467346, "learning_rate": 0.0001921348762203666, "loss": 4.5089, "step": 487200 }, { "epoch": 0.9092626761434013, "grad_norm": 0.8739670515060425, "learning_rate": 0.00019213324306234007, "loss": 4.482, "step": 487250 }, { "epoch": 0.9093559817027798, "grad_norm": 1.0457218885421753, "learning_rate": 0.0001921316097417146, "loss": 4.5354, "step": 487300 }, { "epoch": 0.9094492872621583, "grad_norm": 1.2138333320617676, "learning_rate": 0.000192129976258493, "loss": 4.6182, "step": 487350 }, { "epoch": 0.9095425928215368, "grad_norm": 1.4671598672866821, "learning_rate": 0.00019212834261267823, "loss": 4.7004, "step": 487400 }, { "epoch": 0.9096358983809153, "grad_norm": 1.1629071235656738, "learning_rate": 0.00019212670880427318, "loss": 4.5454, "step": 487450 }, { "epoch": 0.9097292039402938, "grad_norm": 0.575943648815155, "learning_rate": 0.00019212507483328065, "loss": 4.4984, "step": 487500 }, { "epoch": 0.9098225094996722, "grad_norm": 0.9166334867477417, "learning_rate": 0.00019212344069970358, "loss": 4.5206, "step": 487550 }, { "epoch": 0.9099158150590507, "grad_norm": 1.0846489667892456, "learning_rate": 0.00019212180640354486, "loss": 4.6879, "step": 487600 }, { "epoch": 0.9100091206184292, "grad_norm": 1.2294113636016846, "learning_rate": 0.00019212017194480735, "loss": 4.6703, "step": 487650 }, { "epoch": 0.9101024261778078, "grad_norm": 0.7781298160552979, "learning_rate": 0.000192118537323494, "loss": 4.7883, "step": 487700 }, { "epoch": 0.9101957317371863, "grad_norm": 1.2383092641830444, "learning_rate": 0.00019211690253960757, "loss": 4.4692, "step": 487750 }, { "epoch": 0.9102890372965647, "grad_norm": 1.1804494857788086, "learning_rate": 0.00019211526759315103, "loss": 4.5754, "step": 487800 }, { "epoch": 0.9103823428559432, "grad_norm": 0.8236009478569031, "learning_rate": 0.00019211363248412727, "loss": 4.327, "step": 487850 }, { "epoch": 0.9104756484153217, "grad_norm": 1.024800181388855, "learning_rate": 0.00019211199721253915, "loss": 4.64, "step": 487900 }, { "epoch": 0.9105689539747002, "grad_norm": 1.1174505949020386, "learning_rate": 0.00019211036177838956, "loss": 4.8015, "step": 487950 }, { "epoch": 0.9106622595340786, "grad_norm": 1.6258537769317627, "learning_rate": 0.00019210872618168138, "loss": 4.6144, "step": 488000 }, { "epoch": 0.9107555650934572, "grad_norm": 1.170012354850769, "learning_rate": 0.00019210709042241753, "loss": 4.3449, "step": 488050 }, { "epoch": 0.9108488706528357, "grad_norm": 1.1477054357528687, "learning_rate": 0.00019210545450060084, "loss": 4.6412, "step": 488100 }, { "epoch": 0.9109421762122142, "grad_norm": 0.8709313869476318, "learning_rate": 0.00019210381841623425, "loss": 4.7908, "step": 488150 }, { "epoch": 0.9110354817715927, "grad_norm": 1.116733431816101, "learning_rate": 0.0001921021821693206, "loss": 4.5387, "step": 488200 }, { "epoch": 0.9111287873309711, "grad_norm": 1.0928171873092651, "learning_rate": 0.00019210054575986285, "loss": 4.7364, "step": 488250 }, { "epoch": 0.9112220928903496, "grad_norm": 1.1159753799438477, "learning_rate": 0.0001920989091878638, "loss": 4.3054, "step": 488300 }, { "epoch": 0.9113153984497281, "grad_norm": 1.1690645217895508, "learning_rate": 0.0001920972724533264, "loss": 4.5056, "step": 488350 }, { "epoch": 0.9114087040091067, "grad_norm": 1.0336902141571045, "learning_rate": 0.00019209563555625355, "loss": 4.4958, "step": 488400 }, { "epoch": 0.9115020095684851, "grad_norm": 0.9411969184875488, "learning_rate": 0.00019209399849664805, "loss": 4.7455, "step": 488450 }, { "epoch": 0.9115953151278636, "grad_norm": 0.7945715188980103, "learning_rate": 0.0001920923612745129, "loss": 4.6542, "step": 488500 }, { "epoch": 0.9116886206872421, "grad_norm": 0.8895405530929565, "learning_rate": 0.0001920907238898509, "loss": 4.5231, "step": 488550 }, { "epoch": 0.9117819262466206, "grad_norm": 1.0384875535964966, "learning_rate": 0.00019208908634266498, "loss": 4.6397, "step": 488600 }, { "epoch": 0.911875231805999, "grad_norm": 1.0071812868118286, "learning_rate": 0.00019208744863295803, "loss": 4.6988, "step": 488650 }, { "epoch": 0.9119685373653775, "grad_norm": 1.1119264364242554, "learning_rate": 0.00019208581076073292, "loss": 4.6237, "step": 488700 }, { "epoch": 0.9120618429247561, "grad_norm": 0.9369913339614868, "learning_rate": 0.0001920841727259926, "loss": 4.4234, "step": 488750 }, { "epoch": 0.9121551484841346, "grad_norm": 1.347872018814087, "learning_rate": 0.0001920825345287399, "loss": 4.5331, "step": 488800 }, { "epoch": 0.9122484540435131, "grad_norm": 0.8222604990005493, "learning_rate": 0.00019208089616897772, "loss": 4.5802, "step": 488850 }, { "epoch": 0.9123417596028915, "grad_norm": 0.984792172908783, "learning_rate": 0.00019207925764670898, "loss": 4.4384, "step": 488900 }, { "epoch": 0.91243506516227, "grad_norm": 1.0539358854293823, "learning_rate": 0.00019207761896193653, "loss": 4.6409, "step": 488950 }, { "epoch": 0.9125283707216485, "grad_norm": 0.9865897297859192, "learning_rate": 0.00019207598011466328, "loss": 4.372, "step": 489000 }, { "epoch": 0.9125283707216485, "eval_loss": 4.713347911834717, "eval_runtime": 230.0663, "eval_samples_per_second": 11.336, "eval_steps_per_second": 11.336, "eval_tts_loss": 7.604087135577403, "step": 489000 }, { "epoch": 0.912621676281027, "grad_norm": 1.0789505243301392, "learning_rate": 0.00019207434110489214, "loss": 4.6304, "step": 489050 }, { "epoch": 0.9127149818404056, "grad_norm": 1.298210620880127, "learning_rate": 0.00019207270193262598, "loss": 4.4513, "step": 489100 }, { "epoch": 0.912808287399784, "grad_norm": 1.0639249086380005, "learning_rate": 0.0001920710625978677, "loss": 4.337, "step": 489150 }, { "epoch": 0.9129015929591625, "grad_norm": 1.4610494375228882, "learning_rate": 0.0001920694231006202, "loss": 4.5831, "step": 489200 }, { "epoch": 0.912994898518541, "grad_norm": 0.9978083372116089, "learning_rate": 0.00019206778344088639, "loss": 4.6509, "step": 489250 }, { "epoch": 0.9130882040779195, "grad_norm": 1.506140112876892, "learning_rate": 0.00019206614361866914, "loss": 4.4705, "step": 489300 }, { "epoch": 0.9131815096372979, "grad_norm": 1.5171911716461182, "learning_rate": 0.00019206450363397132, "loss": 4.5114, "step": 489350 }, { "epoch": 0.9132748151966764, "grad_norm": 1.2756547927856445, "learning_rate": 0.00019206286348679585, "loss": 4.5548, "step": 489400 }, { "epoch": 0.913368120756055, "grad_norm": 1.1202374696731567, "learning_rate": 0.00019206122317714562, "loss": 4.3207, "step": 489450 }, { "epoch": 0.9134614263154335, "grad_norm": 0.7241478562355042, "learning_rate": 0.00019205958270502357, "loss": 4.3381, "step": 489500 }, { "epoch": 0.913554731874812, "grad_norm": 1.2047520875930786, "learning_rate": 0.0001920579420704325, "loss": 4.6308, "step": 489550 }, { "epoch": 0.9136480374341904, "grad_norm": 0.908880352973938, "learning_rate": 0.00019205630127337538, "loss": 4.5044, "step": 489600 }, { "epoch": 0.9137413429935689, "grad_norm": 1.1000475883483887, "learning_rate": 0.0001920546603138551, "loss": 4.4585, "step": 489650 }, { "epoch": 0.9138346485529474, "grad_norm": 1.1672555208206177, "learning_rate": 0.00019205301919187452, "loss": 4.4368, "step": 489700 }, { "epoch": 0.9139279541123259, "grad_norm": 1.267440676689148, "learning_rate": 0.0001920513779074366, "loss": 4.6477, "step": 489750 }, { "epoch": 0.9140212596717044, "grad_norm": 1.3579221963882446, "learning_rate": 0.00019204973646054415, "loss": 4.5113, "step": 489800 }, { "epoch": 0.9141145652310829, "grad_norm": 1.0702447891235352, "learning_rate": 0.00019204809485120014, "loss": 4.5416, "step": 489850 }, { "epoch": 0.9142078707904614, "grad_norm": 1.4126346111297607, "learning_rate": 0.00019204645307940746, "loss": 4.5249, "step": 489900 }, { "epoch": 0.9143011763498399, "grad_norm": 1.3368160724639893, "learning_rate": 0.00019204481114516893, "loss": 4.6445, "step": 489950 }, { "epoch": 0.9143944819092183, "grad_norm": 1.0472965240478516, "learning_rate": 0.00019204316904848753, "loss": 4.4813, "step": 490000 }, { "epoch": 0.9144877874685968, "grad_norm": 0.9504802227020264, "learning_rate": 0.00019204152678936614, "loss": 4.5216, "step": 490050 }, { "epoch": 0.9145810930279753, "grad_norm": 0.8336538076400757, "learning_rate": 0.00019203988436780763, "loss": 4.669, "step": 490100 }, { "epoch": 0.9146743985873538, "grad_norm": 1.171618938446045, "learning_rate": 0.00019203824178381495, "loss": 4.6981, "step": 490150 }, { "epoch": 0.9147677041467324, "grad_norm": 1.1843730211257935, "learning_rate": 0.00019203659903739096, "loss": 4.5826, "step": 490200 }, { "epoch": 0.9148610097061108, "grad_norm": 0.9480854272842407, "learning_rate": 0.0001920349561285385, "loss": 4.4697, "step": 490250 }, { "epoch": 0.9149543152654893, "grad_norm": 1.13724946975708, "learning_rate": 0.00019203331305726065, "loss": 4.4815, "step": 490300 }, { "epoch": 0.9150476208248678, "grad_norm": 1.2858607769012451, "learning_rate": 0.00019203166982356012, "loss": 4.468, "step": 490350 }, { "epoch": 0.9151409263842463, "grad_norm": 0.9288567900657654, "learning_rate": 0.0001920300264274399, "loss": 4.4118, "step": 490400 }, { "epoch": 0.9152342319436247, "grad_norm": 1.0951497554779053, "learning_rate": 0.0001920283828689029, "loss": 4.3865, "step": 490450 }, { "epoch": 0.9153275375030032, "grad_norm": 1.4390220642089844, "learning_rate": 0.00019202673914795197, "loss": 4.6388, "step": 490500 }, { "epoch": 0.9154208430623818, "grad_norm": 1.319702386856079, "learning_rate": 0.00019202509526459005, "loss": 4.6216, "step": 490550 }, { "epoch": 0.9155141486217603, "grad_norm": 1.0547899007797241, "learning_rate": 0.00019202345121882003, "loss": 4.683, "step": 490600 }, { "epoch": 0.9156074541811388, "grad_norm": 1.025553822517395, "learning_rate": 0.00019202180701064482, "loss": 4.4436, "step": 490650 }, { "epoch": 0.9157007597405172, "grad_norm": 1.3679935932159424, "learning_rate": 0.0001920201626400673, "loss": 4.4812, "step": 490700 }, { "epoch": 0.9157940652998957, "grad_norm": 1.3404130935668945, "learning_rate": 0.00019201851810709037, "loss": 4.5201, "step": 490750 }, { "epoch": 0.9158873708592742, "grad_norm": 1.2220560312271118, "learning_rate": 0.00019201687341171696, "loss": 4.796, "step": 490800 }, { "epoch": 0.9159806764186527, "grad_norm": 0.9640553593635559, "learning_rate": 0.00019201522855394996, "loss": 4.5743, "step": 490850 }, { "epoch": 0.9160739819780312, "grad_norm": 0.7948808670043945, "learning_rate": 0.00019201358353379227, "loss": 4.6452, "step": 490900 }, { "epoch": 0.9161672875374097, "grad_norm": 0.9232187271118164, "learning_rate": 0.0001920119383512468, "loss": 4.4965, "step": 490950 }, { "epoch": 0.9162605930967882, "grad_norm": 1.1507916450500488, "learning_rate": 0.00019201029300631645, "loss": 4.3744, "step": 491000 }, { "epoch": 0.9163538986561667, "grad_norm": 1.1561031341552734, "learning_rate": 0.00019200864749900412, "loss": 4.5918, "step": 491050 }, { "epoch": 0.9164472042155452, "grad_norm": 0.92462557554245, "learning_rate": 0.0001920070018293127, "loss": 4.4749, "step": 491100 }, { "epoch": 0.9165405097749236, "grad_norm": 1.0105730295181274, "learning_rate": 0.00019200535599724514, "loss": 4.4683, "step": 491150 }, { "epoch": 0.9166338153343021, "grad_norm": 0.9027535915374756, "learning_rate": 0.0001920037100028043, "loss": 4.6471, "step": 491200 }, { "epoch": 0.9167271208936807, "grad_norm": 1.2318564653396606, "learning_rate": 0.00019200206384599312, "loss": 4.5611, "step": 491250 }, { "epoch": 0.9168204264530592, "grad_norm": 1.2528573274612427, "learning_rate": 0.00019200041752681445, "loss": 4.6653, "step": 491300 }, { "epoch": 0.9169137320124376, "grad_norm": 1.1168265342712402, "learning_rate": 0.00019199877104527125, "loss": 4.5361, "step": 491350 }, { "epoch": 0.9170070375718161, "grad_norm": 1.177459716796875, "learning_rate": 0.00019199712440136642, "loss": 4.6327, "step": 491400 }, { "epoch": 0.9171003431311946, "grad_norm": 1.0752363204956055, "learning_rate": 0.00019199547759510282, "loss": 4.579, "step": 491450 }, { "epoch": 0.9171936486905731, "grad_norm": 0.9750145673751831, "learning_rate": 0.00019199383062648342, "loss": 4.502, "step": 491500 }, { "epoch": 0.9172869542499515, "grad_norm": 1.1898130178451538, "learning_rate": 0.0001919921834955111, "loss": 4.6407, "step": 491550 }, { "epoch": 0.9173802598093301, "grad_norm": 0.9509438276290894, "learning_rate": 0.00019199053620218873, "loss": 4.4782, "step": 491600 }, { "epoch": 0.9174735653687086, "grad_norm": 0.9310864806175232, "learning_rate": 0.00019198888874651928, "loss": 4.602, "step": 491650 }, { "epoch": 0.9175668709280871, "grad_norm": 1.163161277770996, "learning_rate": 0.0001919872411285056, "loss": 4.5539, "step": 491700 }, { "epoch": 0.9176601764874656, "grad_norm": 1.0423190593719482, "learning_rate": 0.00019198559334815065, "loss": 4.6556, "step": 491750 }, { "epoch": 0.917753482046844, "grad_norm": 1.4016902446746826, "learning_rate": 0.00019198394540545727, "loss": 4.5981, "step": 491800 }, { "epoch": 0.9178467876062225, "grad_norm": 1.2689636945724487, "learning_rate": 0.00019198229730042846, "loss": 4.4365, "step": 491850 }, { "epoch": 0.917940093165601, "grad_norm": 1.1335569620132446, "learning_rate": 0.00019198064903306708, "loss": 4.551, "step": 491900 }, { "epoch": 0.9180333987249796, "grad_norm": 1.1723552942276, "learning_rate": 0.00019197900060337602, "loss": 4.8645, "step": 491950 }, { "epoch": 0.918126704284358, "grad_norm": 2.616619825363159, "learning_rate": 0.0001919773520113582, "loss": 4.4552, "step": 492000 }, { "epoch": 0.918126704284358, "eval_loss": 4.706608772277832, "eval_runtime": 233.3538, "eval_samples_per_second": 11.176, "eval_steps_per_second": 11.176, "eval_tts_loss": 7.570185992345385, "step": 492000 }, { "epoch": 0.9182200098437365, "grad_norm": 1.1619545221328735, "learning_rate": 0.00019197570325701655, "loss": 4.3797, "step": 492050 }, { "epoch": 0.918313315403115, "grad_norm": 1.2034484148025513, "learning_rate": 0.00019197405434035397, "loss": 4.6197, "step": 492100 }, { "epoch": 0.9184066209624935, "grad_norm": 1.3257561922073364, "learning_rate": 0.00019197240526137334, "loss": 4.5576, "step": 492150 }, { "epoch": 0.918499926521872, "grad_norm": 1.2478523254394531, "learning_rate": 0.00019197075602007764, "loss": 4.5804, "step": 492200 }, { "epoch": 0.9185932320812504, "grad_norm": 1.2818803787231445, "learning_rate": 0.00019196910661646973, "loss": 4.5755, "step": 492250 }, { "epoch": 0.918686537640629, "grad_norm": 1.2143594026565552, "learning_rate": 0.00019196745705055252, "loss": 4.4089, "step": 492300 }, { "epoch": 0.9187798432000075, "grad_norm": 1.419231653213501, "learning_rate": 0.0001919658073223289, "loss": 4.5005, "step": 492350 }, { "epoch": 0.918873148759386, "grad_norm": 1.2244318723678589, "learning_rate": 0.00019196415743180184, "loss": 4.6664, "step": 492400 }, { "epoch": 0.9189664543187644, "grad_norm": 0.7943194508552551, "learning_rate": 0.00019196250737897423, "loss": 4.5967, "step": 492450 }, { "epoch": 0.9190597598781429, "grad_norm": 1.0210673809051514, "learning_rate": 0.00019196085716384899, "loss": 4.6502, "step": 492500 }, { "epoch": 0.9191530654375214, "grad_norm": 0.9071007966995239, "learning_rate": 0.000191959206786429, "loss": 4.5059, "step": 492550 }, { "epoch": 0.9192463709968999, "grad_norm": 0.9264189004898071, "learning_rate": 0.00019195755624671715, "loss": 4.607, "step": 492600 }, { "epoch": 0.9193396765562785, "grad_norm": 1.0069105625152588, "learning_rate": 0.00019195590554471647, "loss": 4.626, "step": 492650 }, { "epoch": 0.9194329821156569, "grad_norm": 1.1050623655319214, "learning_rate": 0.00019195425468042973, "loss": 4.5318, "step": 492700 }, { "epoch": 0.9195262876750354, "grad_norm": 1.3213204145431519, "learning_rate": 0.00019195260365385995, "loss": 4.6927, "step": 492750 }, { "epoch": 0.9196195932344139, "grad_norm": 1.3675942420959473, "learning_rate": 0.00019195095246501, "loss": 4.7745, "step": 492800 }, { "epoch": 0.9197128987937924, "grad_norm": 1.082534909248352, "learning_rate": 0.00019194930111388279, "loss": 4.7244, "step": 492850 }, { "epoch": 0.9198062043531708, "grad_norm": 0.8137260675430298, "learning_rate": 0.00019194764960048125, "loss": 4.639, "step": 492900 }, { "epoch": 0.9198995099125493, "grad_norm": 0.9784749150276184, "learning_rate": 0.0001919459979248083, "loss": 4.4088, "step": 492950 }, { "epoch": 0.9199928154719279, "grad_norm": 1.2249281406402588, "learning_rate": 0.0001919443460868668, "loss": 4.5662, "step": 493000 }, { "epoch": 0.9200861210313064, "grad_norm": 1.009429931640625, "learning_rate": 0.00019194269408665972, "loss": 4.6046, "step": 493050 }, { "epoch": 0.9201794265906849, "grad_norm": 1.3822860717773438, "learning_rate": 0.00019194104192419, "loss": 4.4931, "step": 493100 }, { "epoch": 0.9202727321500633, "grad_norm": 1.1014047861099243, "learning_rate": 0.00019193938959946046, "loss": 4.1832, "step": 493150 }, { "epoch": 0.9203660377094418, "grad_norm": 1.1175732612609863, "learning_rate": 0.00019193773711247413, "loss": 4.4333, "step": 493200 }, { "epoch": 0.9204593432688203, "grad_norm": 1.3878298997879028, "learning_rate": 0.00019193608446323385, "loss": 4.4315, "step": 493250 }, { "epoch": 0.9205526488281988, "grad_norm": 1.1050724983215332, "learning_rate": 0.00019193443165174253, "loss": 4.5137, "step": 493300 }, { "epoch": 0.9206459543875773, "grad_norm": 0.9894223213195801, "learning_rate": 0.00019193277867800317, "loss": 4.6598, "step": 493350 }, { "epoch": 0.9207392599469558, "grad_norm": 1.0854759216308594, "learning_rate": 0.0001919311255420186, "loss": 4.4817, "step": 493400 }, { "epoch": 0.9208325655063343, "grad_norm": 0.7749813199043274, "learning_rate": 0.00019192947224379179, "loss": 4.6342, "step": 493450 }, { "epoch": 0.9209258710657128, "grad_norm": 1.0218850374221802, "learning_rate": 0.00019192781878332558, "loss": 4.528, "step": 493500 }, { "epoch": 0.9210191766250913, "grad_norm": 0.9555392861366272, "learning_rate": 0.000191926165160623, "loss": 4.6554, "step": 493550 }, { "epoch": 0.9211124821844697, "grad_norm": 0.7390783429145813, "learning_rate": 0.0001919245113756869, "loss": 4.1388, "step": 493600 }, { "epoch": 0.9212057877438482, "grad_norm": 1.1266878843307495, "learning_rate": 0.00019192285742852022, "loss": 4.6678, "step": 493650 }, { "epoch": 0.9212990933032268, "grad_norm": 0.7407153844833374, "learning_rate": 0.00019192120331912587, "loss": 4.363, "step": 493700 }, { "epoch": 0.9213923988626053, "grad_norm": 1.3411778211593628, "learning_rate": 0.00019191954904750676, "loss": 4.5881, "step": 493750 }, { "epoch": 0.9214857044219837, "grad_norm": 0.9631601572036743, "learning_rate": 0.0001919178946136658, "loss": 4.5874, "step": 493800 }, { "epoch": 0.9215790099813622, "grad_norm": 1.037002444267273, "learning_rate": 0.00019191624001760595, "loss": 4.4229, "step": 493850 }, { "epoch": 0.9216723155407407, "grad_norm": 1.460612177848816, "learning_rate": 0.0001919145852593301, "loss": 4.488, "step": 493900 }, { "epoch": 0.9217656211001192, "grad_norm": 1.2926445007324219, "learning_rate": 0.00019191293033884118, "loss": 4.697, "step": 493950 }, { "epoch": 0.9218589266594976, "grad_norm": 0.9376246333122253, "learning_rate": 0.00019191127525614214, "loss": 4.3617, "step": 494000 }, { "epoch": 0.9219522322188762, "grad_norm": 1.3558824062347412, "learning_rate": 0.00019190962001123586, "loss": 4.3911, "step": 494050 }, { "epoch": 0.9220455377782547, "grad_norm": 1.4340498447418213, "learning_rate": 0.00019190796460412526, "loss": 4.5853, "step": 494100 }, { "epoch": 0.9221388433376332, "grad_norm": 1.2456148862838745, "learning_rate": 0.00019190630903481327, "loss": 4.5026, "step": 494150 }, { "epoch": 0.9222321488970117, "grad_norm": 1.39982271194458, "learning_rate": 0.00019190465330330282, "loss": 4.5963, "step": 494200 }, { "epoch": 0.9223254544563901, "grad_norm": 1.4766526222229004, "learning_rate": 0.00019190299740959682, "loss": 4.666, "step": 494250 }, { "epoch": 0.9224187600157686, "grad_norm": 1.0576071739196777, "learning_rate": 0.00019190134135369823, "loss": 4.3156, "step": 494300 }, { "epoch": 0.9225120655751471, "grad_norm": 1.4292043447494507, "learning_rate": 0.0001918996851356099, "loss": 4.5257, "step": 494350 }, { "epoch": 0.9226053711345257, "grad_norm": 1.069832682609558, "learning_rate": 0.00019189802875533484, "loss": 4.5969, "step": 494400 }, { "epoch": 0.9226986766939042, "grad_norm": 1.1232844591140747, "learning_rate": 0.00019189637221287588, "loss": 4.7108, "step": 494450 }, { "epoch": 0.9227919822532826, "grad_norm": 1.187986135482788, "learning_rate": 0.00019189471550823606, "loss": 4.3577, "step": 494500 }, { "epoch": 0.9228852878126611, "grad_norm": 0.9981232285499573, "learning_rate": 0.00019189305864141817, "loss": 4.4981, "step": 494550 }, { "epoch": 0.9229785933720396, "grad_norm": 1.035632848739624, "learning_rate": 0.00019189140161242525, "loss": 4.4728, "step": 494600 }, { "epoch": 0.9230718989314181, "grad_norm": 1.0567675828933716, "learning_rate": 0.00019188974442126014, "loss": 4.4881, "step": 494650 }, { "epoch": 0.9231652044907965, "grad_norm": 1.0472192764282227, "learning_rate": 0.00019188808706792582, "loss": 4.3789, "step": 494700 }, { "epoch": 0.9232585100501751, "grad_norm": 1.0989373922348022, "learning_rate": 0.0001918864295524252, "loss": 4.5442, "step": 494750 }, { "epoch": 0.9233518156095536, "grad_norm": 1.022678017616272, "learning_rate": 0.00019188477187476118, "loss": 4.6974, "step": 494800 }, { "epoch": 0.9234451211689321, "grad_norm": 1.061224341392517, "learning_rate": 0.00019188311403493672, "loss": 4.4406, "step": 494850 }, { "epoch": 0.9235384267283105, "grad_norm": 0.9928637742996216, "learning_rate": 0.0001918814560329547, "loss": 4.454, "step": 494900 }, { "epoch": 0.923631732287689, "grad_norm": 1.087742805480957, "learning_rate": 0.0001918797978688181, "loss": 4.5653, "step": 494950 }, { "epoch": 0.9237250378470675, "grad_norm": 1.2408620119094849, "learning_rate": 0.00019187813954252986, "loss": 4.6998, "step": 495000 }, { "epoch": 0.9237250378470675, "eval_loss": 4.711771011352539, "eval_runtime": 233.7544, "eval_samples_per_second": 11.157, "eval_steps_per_second": 11.157, "eval_tts_loss": 7.639049441676495, "step": 495000 }, { "epoch": 0.923818343406446, "grad_norm": 1.035805344581604, "learning_rate": 0.0001918764810540928, "loss": 4.4604, "step": 495050 }, { "epoch": 0.9239116489658246, "grad_norm": 0.8934062719345093, "learning_rate": 0.00019187482240350998, "loss": 4.59, "step": 495100 }, { "epoch": 0.924004954525203, "grad_norm": 0.9868542551994324, "learning_rate": 0.0001918731635907842, "loss": 4.7011, "step": 495150 }, { "epoch": 0.9240982600845815, "grad_norm": 0.9394553303718567, "learning_rate": 0.00019187150461591853, "loss": 4.6778, "step": 495200 }, { "epoch": 0.92419156564396, "grad_norm": 1.1220202445983887, "learning_rate": 0.00019186984547891575, "loss": 4.7294, "step": 495250 }, { "epoch": 0.9242848712033385, "grad_norm": 1.3254166841506958, "learning_rate": 0.0001918681861797789, "loss": 4.6153, "step": 495300 }, { "epoch": 0.9243781767627169, "grad_norm": 1.0898741483688354, "learning_rate": 0.00019186652671851086, "loss": 4.6697, "step": 495350 }, { "epoch": 0.9244714823220954, "grad_norm": 1.141361951828003, "learning_rate": 0.00019186486709511455, "loss": 4.5562, "step": 495400 }, { "epoch": 0.9245647878814739, "grad_norm": 0.9436092972755432, "learning_rate": 0.0001918632073095929, "loss": 4.4245, "step": 495450 }, { "epoch": 0.9246580934408525, "grad_norm": 1.0880069732666016, "learning_rate": 0.0001918615473619489, "loss": 4.5883, "step": 495500 }, { "epoch": 0.924751399000231, "grad_norm": 0.9273828864097595, "learning_rate": 0.0001918598872521854, "loss": 4.3435, "step": 495550 }, { "epoch": 0.9248447045596094, "grad_norm": 1.133474349975586, "learning_rate": 0.00019185822698030537, "loss": 4.478, "step": 495600 }, { "epoch": 0.9249380101189879, "grad_norm": 1.0183912515640259, "learning_rate": 0.00019185656654631172, "loss": 4.5214, "step": 495650 }, { "epoch": 0.9250313156783664, "grad_norm": 1.0907320976257324, "learning_rate": 0.00019185490595020741, "loss": 4.3954, "step": 495700 }, { "epoch": 0.9251246212377449, "grad_norm": 1.2996900081634521, "learning_rate": 0.00019185324519199534, "loss": 4.7421, "step": 495750 }, { "epoch": 0.9252179267971233, "grad_norm": 1.375575304031372, "learning_rate": 0.00019185158427167847, "loss": 4.534, "step": 495800 }, { "epoch": 0.9253112323565019, "grad_norm": 1.1907944679260254, "learning_rate": 0.0001918499231892597, "loss": 4.7619, "step": 495850 }, { "epoch": 0.9254045379158804, "grad_norm": 0.9488288164138794, "learning_rate": 0.000191848261944742, "loss": 4.2826, "step": 495900 }, { "epoch": 0.9254978434752589, "grad_norm": 0.9436691999435425, "learning_rate": 0.00019184660053812825, "loss": 4.369, "step": 495950 }, { "epoch": 0.9255911490346374, "grad_norm": 1.1006758213043213, "learning_rate": 0.0001918449389694214, "loss": 4.3983, "step": 496000 }, { "epoch": 0.9256844545940158, "grad_norm": 1.0657426118850708, "learning_rate": 0.00019184327723862445, "loss": 4.42, "step": 496050 }, { "epoch": 0.9257777601533943, "grad_norm": 1.1576955318450928, "learning_rate": 0.00019184161534574023, "loss": 4.5842, "step": 496100 }, { "epoch": 0.9258710657127728, "grad_norm": 1.4153337478637695, "learning_rate": 0.00019183995329077175, "loss": 4.5285, "step": 496150 }, { "epoch": 0.9259643712721514, "grad_norm": 1.1001230478286743, "learning_rate": 0.00019183829107372188, "loss": 4.504, "step": 496200 }, { "epoch": 0.9260576768315298, "grad_norm": 0.9774655699729919, "learning_rate": 0.0001918366286945936, "loss": 4.6236, "step": 496250 }, { "epoch": 0.9261509823909083, "grad_norm": 0.8979067206382751, "learning_rate": 0.00019183496615338984, "loss": 4.6179, "step": 496300 }, { "epoch": 0.9262442879502868, "grad_norm": 0.8782025575637817, "learning_rate": 0.0001918333034501135, "loss": 4.5561, "step": 496350 }, { "epoch": 0.9263375935096653, "grad_norm": 1.248464584350586, "learning_rate": 0.00019183164058476755, "loss": 4.5225, "step": 496400 }, { "epoch": 0.9264308990690437, "grad_norm": 1.1266189813613892, "learning_rate": 0.0001918299775573549, "loss": 4.7256, "step": 496450 }, { "epoch": 0.9265242046284222, "grad_norm": 0.9100159406661987, "learning_rate": 0.00019182831436787852, "loss": 4.4962, "step": 496500 }, { "epoch": 0.9266175101878008, "grad_norm": 1.190032958984375, "learning_rate": 0.00019182665101634134, "loss": 4.5482, "step": 496550 }, { "epoch": 0.9267108157471793, "grad_norm": 1.243491530418396, "learning_rate": 0.00019182498750274624, "loss": 4.5592, "step": 496600 }, { "epoch": 0.9268041213065578, "grad_norm": 1.1773312091827393, "learning_rate": 0.0001918233238270962, "loss": 4.6388, "step": 496650 }, { "epoch": 0.9268974268659362, "grad_norm": 0.9503074288368225, "learning_rate": 0.00019182165998939417, "loss": 4.5778, "step": 496700 }, { "epoch": 0.9269907324253147, "grad_norm": 1.125603437423706, "learning_rate": 0.00019181999598964303, "loss": 4.5352, "step": 496750 }, { "epoch": 0.9270840379846932, "grad_norm": 1.1518375873565674, "learning_rate": 0.00019181833182784579, "loss": 4.5762, "step": 496800 }, { "epoch": 0.9271773435440717, "grad_norm": 1.1666361093521118, "learning_rate": 0.00019181666750400537, "loss": 4.6755, "step": 496850 }, { "epoch": 0.9272706491034503, "grad_norm": 0.942234218120575, "learning_rate": 0.00019181500301812462, "loss": 4.5134, "step": 496900 }, { "epoch": 0.9273639546628287, "grad_norm": 1.1006569862365723, "learning_rate": 0.0001918133383702066, "loss": 4.6687, "step": 496950 }, { "epoch": 0.9274572602222072, "grad_norm": 1.3982492685317993, "learning_rate": 0.00019181167356025416, "loss": 4.5265, "step": 497000 }, { "epoch": 0.9275505657815857, "grad_norm": 0.8414306044578552, "learning_rate": 0.00019181000858827028, "loss": 4.3879, "step": 497050 }, { "epoch": 0.9276438713409642, "grad_norm": 1.173119306564331, "learning_rate": 0.0001918083434542579, "loss": 4.4646, "step": 497100 }, { "epoch": 0.9277371769003426, "grad_norm": 1.2540926933288574, "learning_rate": 0.00019180667815821994, "loss": 4.4725, "step": 497150 }, { "epoch": 0.9278304824597211, "grad_norm": 1.2728825807571411, "learning_rate": 0.00019180501270015934, "loss": 4.7252, "step": 497200 }, { "epoch": 0.9279237880190997, "grad_norm": 1.2974334955215454, "learning_rate": 0.00019180334708007905, "loss": 4.6825, "step": 497250 }, { "epoch": 0.9280170935784782, "grad_norm": 1.004394769668579, "learning_rate": 0.00019180168129798202, "loss": 4.4509, "step": 497300 }, { "epoch": 0.9281103991378566, "grad_norm": 1.0203449726104736, "learning_rate": 0.00019180001535387114, "loss": 4.3362, "step": 497350 }, { "epoch": 0.9282037046972351, "grad_norm": 1.0894291400909424, "learning_rate": 0.00019179834924774942, "loss": 4.6014, "step": 497400 }, { "epoch": 0.9282970102566136, "grad_norm": 1.0922901630401611, "learning_rate": 0.00019179668297961974, "loss": 4.461, "step": 497450 }, { "epoch": 0.9283903158159921, "grad_norm": 1.242812156677246, "learning_rate": 0.00019179501654948512, "loss": 4.5075, "step": 497500 }, { "epoch": 0.9284836213753705, "grad_norm": 0.9364404082298279, "learning_rate": 0.00019179334995734842, "loss": 4.432, "step": 497550 }, { "epoch": 0.9285769269347491, "grad_norm": 1.0533199310302734, "learning_rate": 0.00019179168320321258, "loss": 4.7015, "step": 497600 }, { "epoch": 0.9286702324941276, "grad_norm": 1.1974238157272339, "learning_rate": 0.00019179001628708058, "loss": 4.6011, "step": 497650 }, { "epoch": 0.9287635380535061, "grad_norm": 0.8727584481239319, "learning_rate": 0.00019178834920895537, "loss": 4.643, "step": 497700 }, { "epoch": 0.9288568436128846, "grad_norm": 1.238200306892395, "learning_rate": 0.00019178668196883988, "loss": 4.5221, "step": 497750 }, { "epoch": 0.928950149172263, "grad_norm": 1.2333307266235352, "learning_rate": 0.00019178501456673704, "loss": 4.5676, "step": 497800 }, { "epoch": 0.9290434547316415, "grad_norm": 0.8074859380722046, "learning_rate": 0.00019178334700264976, "loss": 4.5119, "step": 497850 }, { "epoch": 0.92913676029102, "grad_norm": 1.1961002349853516, "learning_rate": 0.00019178167927658105, "loss": 4.3987, "step": 497900 }, { "epoch": 0.9292300658503986, "grad_norm": 1.1720647811889648, "learning_rate": 0.00019178001138853384, "loss": 4.5519, "step": 497950 }, { "epoch": 0.929323371409777, "grad_norm": 1.145991563796997, "learning_rate": 0.00019177834333851106, "loss": 4.4437, "step": 498000 }, { "epoch": 0.929323371409777, "eval_loss": 4.707544803619385, "eval_runtime": 230.5995, "eval_samples_per_second": 11.31, "eval_steps_per_second": 11.31, "eval_tts_loss": 7.598612629631363, "step": 498000 }, { "epoch": 0.9294166769691555, "grad_norm": 1.1381690502166748, "learning_rate": 0.00019177667512651562, "loss": 4.7779, "step": 498050 }, { "epoch": 0.929509982528534, "grad_norm": 1.2675760984420776, "learning_rate": 0.00019177500675255054, "loss": 4.6912, "step": 498100 }, { "epoch": 0.9296032880879125, "grad_norm": 1.0090388059616089, "learning_rate": 0.0001917733382166187, "loss": 4.4757, "step": 498150 }, { "epoch": 0.929696593647291, "grad_norm": 1.1202852725982666, "learning_rate": 0.00019177166951872305, "loss": 4.61, "step": 498200 }, { "epoch": 0.9297898992066694, "grad_norm": 1.106505036354065, "learning_rate": 0.00019177000065886656, "loss": 4.481, "step": 498250 }, { "epoch": 0.929883204766048, "grad_norm": 0.8593153953552246, "learning_rate": 0.0001917683316370522, "loss": 4.5034, "step": 498300 }, { "epoch": 0.9299765103254265, "grad_norm": 1.0426411628723145, "learning_rate": 0.00019176666245328286, "loss": 4.6896, "step": 498350 }, { "epoch": 0.930069815884805, "grad_norm": 1.1071492433547974, "learning_rate": 0.0001917649931075615, "loss": 4.5192, "step": 498400 }, { "epoch": 0.9301631214441834, "grad_norm": 0.8001417517662048, "learning_rate": 0.00019176332359989106, "loss": 4.6989, "step": 498450 }, { "epoch": 0.9302564270035619, "grad_norm": 1.0328518152236938, "learning_rate": 0.00019176165393027455, "loss": 4.5664, "step": 498500 }, { "epoch": 0.9303497325629404, "grad_norm": 0.6290932297706604, "learning_rate": 0.0001917599840987148, "loss": 4.7094, "step": 498550 }, { "epoch": 0.9304430381223189, "grad_norm": 1.155130386352539, "learning_rate": 0.00019175831410521486, "loss": 4.5629, "step": 498600 }, { "epoch": 0.9305363436816975, "grad_norm": 1.0391170978546143, "learning_rate": 0.00019175664394977763, "loss": 4.6798, "step": 498650 }, { "epoch": 0.9306296492410759, "grad_norm": 1.1518107652664185, "learning_rate": 0.0001917549736324061, "loss": 4.5742, "step": 498700 }, { "epoch": 0.9307229548004544, "grad_norm": 1.2177016735076904, "learning_rate": 0.00019175330315310317, "loss": 4.5263, "step": 498750 }, { "epoch": 0.9308162603598329, "grad_norm": 1.0102006196975708, "learning_rate": 0.00019175163251187178, "loss": 4.5388, "step": 498800 }, { "epoch": 0.9309095659192114, "grad_norm": 1.1485563516616821, "learning_rate": 0.00019174996170871493, "loss": 4.3893, "step": 498850 }, { "epoch": 0.9310028714785898, "grad_norm": 1.1185194253921509, "learning_rate": 0.00019174829074363551, "loss": 4.4194, "step": 498900 }, { "epoch": 0.9310961770379683, "grad_norm": 0.6963541507720947, "learning_rate": 0.00019174661961663655, "loss": 4.5452, "step": 498950 }, { "epoch": 0.9311894825973469, "grad_norm": 1.1910724639892578, "learning_rate": 0.00019174494832772092, "loss": 4.5614, "step": 499000 }, { "epoch": 0.9312827881567254, "grad_norm": 1.1797943115234375, "learning_rate": 0.0001917432768768916, "loss": 4.6847, "step": 499050 }, { "epoch": 0.9313760937161039, "grad_norm": 1.0755647420883179, "learning_rate": 0.00019174160526415154, "loss": 4.559, "step": 499100 }, { "epoch": 0.9314693992754823, "grad_norm": 0.9277364611625671, "learning_rate": 0.00019173993348950367, "loss": 4.7794, "step": 499150 }, { "epoch": 0.9315627048348608, "grad_norm": 1.1691107749938965, "learning_rate": 0.000191738261552951, "loss": 4.5802, "step": 499200 }, { "epoch": 0.9316560103942393, "grad_norm": 1.1780486106872559, "learning_rate": 0.0001917365894544964, "loss": 4.6327, "step": 499250 }, { "epoch": 0.9317493159536178, "grad_norm": 1.3142284154891968, "learning_rate": 0.0001917349171941429, "loss": 4.453, "step": 499300 }, { "epoch": 0.9318426215129963, "grad_norm": 0.9346514940261841, "learning_rate": 0.0001917332447718934, "loss": 4.3799, "step": 499350 }, { "epoch": 0.9319359270723748, "grad_norm": 1.1572456359863281, "learning_rate": 0.00019173157218775085, "loss": 4.5406, "step": 499400 }, { "epoch": 0.9320292326317533, "grad_norm": 1.312286615371704, "learning_rate": 0.00019172989944171823, "loss": 4.5077, "step": 499450 }, { "epoch": 0.9321225381911318, "grad_norm": 1.1337004899978638, "learning_rate": 0.00019172822653379848, "loss": 4.5138, "step": 499500 }, { "epoch": 0.9322158437505103, "grad_norm": 1.1699273586273193, "learning_rate": 0.00019172655346399452, "loss": 4.4352, "step": 499550 }, { "epoch": 0.9323091493098887, "grad_norm": 1.3052034378051758, "learning_rate": 0.00019172488023230937, "loss": 4.4956, "step": 499600 }, { "epoch": 0.9324024548692672, "grad_norm": 1.0680040121078491, "learning_rate": 0.00019172320683874593, "loss": 4.4326, "step": 499650 }, { "epoch": 0.9324957604286458, "grad_norm": 1.084257960319519, "learning_rate": 0.00019172153328330717, "loss": 4.678, "step": 499700 }, { "epoch": 0.9325890659880243, "grad_norm": 0.9289731383323669, "learning_rate": 0.00019171985956599602, "loss": 4.566, "step": 499750 }, { "epoch": 0.9326823715474027, "grad_norm": 0.9422695636749268, "learning_rate": 0.0001917181856868155, "loss": 4.6121, "step": 499800 }, { "epoch": 0.9327756771067812, "grad_norm": 1.1771169900894165, "learning_rate": 0.00019171651164576848, "loss": 4.6019, "step": 499850 }, { "epoch": 0.9328689826661597, "grad_norm": 1.3393751382827759, "learning_rate": 0.00019171483744285798, "loss": 4.5295, "step": 499900 }, { "epoch": 0.9329622882255382, "grad_norm": 0.9513429403305054, "learning_rate": 0.00019171316307808694, "loss": 4.4012, "step": 499950 }, { "epoch": 0.9330555937849166, "grad_norm": 0.9342679381370544, "learning_rate": 0.00019171148855145828, "loss": 4.6776, "step": 500000 }, { "epoch": 0.9331488993442952, "grad_norm": 1.0279916524887085, "learning_rate": 0.00019170981386297498, "loss": 4.4113, "step": 500050 }, { "epoch": 0.9332422049036737, "grad_norm": 1.0957508087158203, "learning_rate": 0.00019170813901264, "loss": 4.338, "step": 500100 }, { "epoch": 0.9333355104630522, "grad_norm": 1.4782789945602417, "learning_rate": 0.0001917064640004563, "loss": 4.5581, "step": 500150 }, { "epoch": 0.9334288160224307, "grad_norm": 1.0908095836639404, "learning_rate": 0.00019170478882642682, "loss": 4.5593, "step": 500200 }, { "epoch": 0.9335221215818091, "grad_norm": 1.0782676935195923, "learning_rate": 0.0001917031134905545, "loss": 4.753, "step": 500250 }, { "epoch": 0.9336154271411876, "grad_norm": 1.0441406965255737, "learning_rate": 0.00019170143799284237, "loss": 4.3509, "step": 500300 }, { "epoch": 0.9337087327005661, "grad_norm": 0.9400619268417358, "learning_rate": 0.00019169976233329332, "loss": 4.6646, "step": 500350 }, { "epoch": 0.9338020382599447, "grad_norm": 0.8745513558387756, "learning_rate": 0.0001916980865119103, "loss": 4.6288, "step": 500400 }, { "epoch": 0.9338953438193232, "grad_norm": 1.2456843852996826, "learning_rate": 0.0001916964105286963, "loss": 4.4571, "step": 500450 }, { "epoch": 0.9339886493787016, "grad_norm": 0.8797690868377686, "learning_rate": 0.0001916947343836543, "loss": 4.5039, "step": 500500 }, { "epoch": 0.9340819549380801, "grad_norm": 1.189086675643921, "learning_rate": 0.0001916930580767872, "loss": 4.4947, "step": 500550 }, { "epoch": 0.9341752604974586, "grad_norm": 1.2526785135269165, "learning_rate": 0.00019169138160809798, "loss": 4.6672, "step": 500600 }, { "epoch": 0.9342685660568371, "grad_norm": 0.8574811220169067, "learning_rate": 0.00019168970497758964, "loss": 4.8201, "step": 500650 }, { "epoch": 0.9343618716162155, "grad_norm": 0.9610766768455505, "learning_rate": 0.00019168802818526507, "loss": 4.621, "step": 500700 }, { "epoch": 0.934455177175594, "grad_norm": 1.250868797302246, "learning_rate": 0.0001916863512311273, "loss": 4.5293, "step": 500750 }, { "epoch": 0.9345484827349726, "grad_norm": 1.1095539331436157, "learning_rate": 0.0001916846741151792, "loss": 4.4734, "step": 500800 }, { "epoch": 0.9346417882943511, "grad_norm": 1.1623324155807495, "learning_rate": 0.00019168299683742382, "loss": 4.4767, "step": 500850 }, { "epoch": 0.9347350938537295, "grad_norm": 0.9767505526542664, "learning_rate": 0.00019168131939786406, "loss": 4.4991, "step": 500900 }, { "epoch": 0.934828399413108, "grad_norm": 1.1820361614227295, "learning_rate": 0.00019167964179650295, "loss": 4.5178, "step": 500950 }, { "epoch": 0.9349217049724865, "grad_norm": 1.129621982574463, "learning_rate": 0.00019167796403334335, "loss": 4.4402, "step": 501000 }, { "epoch": 0.9349217049724865, "eval_loss": 4.727338790893555, "eval_runtime": 231.5457, "eval_samples_per_second": 11.263, "eval_steps_per_second": 11.263, "eval_tts_loss": 7.550874634724441, "step": 501000 }, { "epoch": 0.935015010531865, "grad_norm": 1.4686557054519653, "learning_rate": 0.0001916762861083883, "loss": 4.6708, "step": 501050 }, { "epoch": 0.9351083160912435, "grad_norm": 1.1647988557815552, "learning_rate": 0.00019167460802164073, "loss": 4.4997, "step": 501100 }, { "epoch": 0.935201621650622, "grad_norm": 0.9778158068656921, "learning_rate": 0.0001916729297731036, "loss": 4.6296, "step": 501150 }, { "epoch": 0.9352949272100005, "grad_norm": 1.249403953552246, "learning_rate": 0.0001916712513627799, "loss": 4.4585, "step": 501200 }, { "epoch": 0.935388232769379, "grad_norm": 1.0672589540481567, "learning_rate": 0.00019166957279067256, "loss": 4.5543, "step": 501250 }, { "epoch": 0.9354815383287575, "grad_norm": 1.0421271324157715, "learning_rate": 0.00019166789405678455, "loss": 4.6291, "step": 501300 }, { "epoch": 0.9355748438881359, "grad_norm": 1.1030302047729492, "learning_rate": 0.00019166621516111885, "loss": 4.5344, "step": 501350 }, { "epoch": 0.9356681494475144, "grad_norm": 1.1490404605865479, "learning_rate": 0.0001916645361036784, "loss": 4.3964, "step": 501400 }, { "epoch": 0.9357614550068929, "grad_norm": 1.1173632144927979, "learning_rate": 0.00019166285688446616, "loss": 4.6468, "step": 501450 }, { "epoch": 0.9358547605662715, "grad_norm": 1.1404368877410889, "learning_rate": 0.00019166117750348512, "loss": 4.77, "step": 501500 }, { "epoch": 0.93594806612565, "grad_norm": 1.083148717880249, "learning_rate": 0.00019165949796073824, "loss": 4.4888, "step": 501550 }, { "epoch": 0.9360413716850284, "grad_norm": 0.9196500182151794, "learning_rate": 0.00019165781825622845, "loss": 4.3259, "step": 501600 }, { "epoch": 0.9361346772444069, "grad_norm": 1.3665529489517212, "learning_rate": 0.00019165613838995876, "loss": 4.5907, "step": 501650 }, { "epoch": 0.9362279828037854, "grad_norm": 1.239429235458374, "learning_rate": 0.0001916544583619321, "loss": 4.6542, "step": 501700 }, { "epoch": 0.9363212883631639, "grad_norm": 0.9900445938110352, "learning_rate": 0.00019165277817215145, "loss": 4.6058, "step": 501750 }, { "epoch": 0.9364145939225423, "grad_norm": 1.1552355289459229, "learning_rate": 0.00019165109782061978, "loss": 4.5155, "step": 501800 }, { "epoch": 0.9365078994819209, "grad_norm": 0.8719364404678345, "learning_rate": 0.00019164941730734003, "loss": 4.3866, "step": 501850 }, { "epoch": 0.9366012050412994, "grad_norm": 1.1679394245147705, "learning_rate": 0.0001916477366323152, "loss": 4.4166, "step": 501900 }, { "epoch": 0.9366945106006779, "grad_norm": 0.9125158190727234, "learning_rate": 0.00019164605579554826, "loss": 4.5252, "step": 501950 }, { "epoch": 0.9367878161600564, "grad_norm": 1.115824818611145, "learning_rate": 0.00019164437479704213, "loss": 4.3309, "step": 502000 }, { "epoch": 0.9368811217194348, "grad_norm": 0.9249014258384705, "learning_rate": 0.0001916426936367998, "loss": 4.4925, "step": 502050 }, { "epoch": 0.9369744272788133, "grad_norm": 1.4526195526123047, "learning_rate": 0.00019164101231482428, "loss": 4.4564, "step": 502100 }, { "epoch": 0.9370677328381918, "grad_norm": 0.9519857168197632, "learning_rate": 0.00019163933083111845, "loss": 4.4864, "step": 502150 }, { "epoch": 0.9371610383975704, "grad_norm": 1.1214677095413208, "learning_rate": 0.00019163764918568536, "loss": 4.5473, "step": 502200 }, { "epoch": 0.9372543439569488, "grad_norm": 1.0648483037948608, "learning_rate": 0.00019163596737852793, "loss": 4.5636, "step": 502250 }, { "epoch": 0.9373476495163273, "grad_norm": 0.9157907962799072, "learning_rate": 0.0001916342854096491, "loss": 4.4576, "step": 502300 }, { "epoch": 0.9374409550757058, "grad_norm": 1.015963077545166, "learning_rate": 0.00019163260327905194, "loss": 4.4946, "step": 502350 }, { "epoch": 0.9375342606350843, "grad_norm": 1.1713849306106567, "learning_rate": 0.00019163092098673933, "loss": 4.6891, "step": 502400 }, { "epoch": 0.9376275661944627, "grad_norm": 1.0429054498672485, "learning_rate": 0.0001916292385327143, "loss": 4.4217, "step": 502450 }, { "epoch": 0.9377208717538412, "grad_norm": 0.9693223834037781, "learning_rate": 0.00019162755591697973, "loss": 4.6839, "step": 502500 }, { "epoch": 0.9378141773132198, "grad_norm": 1.195693016052246, "learning_rate": 0.0001916258731395387, "loss": 4.8632, "step": 502550 }, { "epoch": 0.9379074828725983, "grad_norm": 1.2385532855987549, "learning_rate": 0.0001916241902003941, "loss": 4.6459, "step": 502600 }, { "epoch": 0.9380007884319768, "grad_norm": 1.3130062818527222, "learning_rate": 0.00019162250709954895, "loss": 4.3449, "step": 502650 }, { "epoch": 0.9380940939913552, "grad_norm": 0.7653967142105103, "learning_rate": 0.00019162082383700617, "loss": 4.49, "step": 502700 }, { "epoch": 0.9381873995507337, "grad_norm": 0.954108476638794, "learning_rate": 0.00019161914041276873, "loss": 4.5921, "step": 502750 }, { "epoch": 0.9382807051101122, "grad_norm": 1.064396858215332, "learning_rate": 0.00019161745682683967, "loss": 4.4968, "step": 502800 }, { "epoch": 0.9383740106694907, "grad_norm": 1.0272167921066284, "learning_rate": 0.00019161577307922194, "loss": 4.5518, "step": 502850 }, { "epoch": 0.9384673162288693, "grad_norm": 1.2548187971115112, "learning_rate": 0.00019161408916991846, "loss": 4.6563, "step": 502900 }, { "epoch": 0.9385606217882477, "grad_norm": 1.3265324831008911, "learning_rate": 0.0001916124050989322, "loss": 4.7048, "step": 502950 }, { "epoch": 0.9386539273476262, "grad_norm": 1.2669545412063599, "learning_rate": 0.00019161072086626622, "loss": 4.6942, "step": 503000 }, { "epoch": 0.9387472329070047, "grad_norm": 0.9460360407829285, "learning_rate": 0.0001916090364719234, "loss": 4.363, "step": 503050 }, { "epoch": 0.9388405384663832, "grad_norm": 1.3682780265808105, "learning_rate": 0.00019160735191590679, "loss": 4.4761, "step": 503100 }, { "epoch": 0.9389338440257616, "grad_norm": 0.9808215498924255, "learning_rate": 0.00019160566719821927, "loss": 4.4847, "step": 503150 }, { "epoch": 0.9390271495851401, "grad_norm": 0.9927389621734619, "learning_rate": 0.0001916039823188639, "loss": 4.4859, "step": 503200 }, { "epoch": 0.9391204551445187, "grad_norm": 1.0215957164764404, "learning_rate": 0.00019160229727784364, "loss": 4.6765, "step": 503250 }, { "epoch": 0.9392137607038972, "grad_norm": 1.2800140380859375, "learning_rate": 0.0001916006120751614, "loss": 4.5646, "step": 503300 }, { "epoch": 0.9393070662632756, "grad_norm": 1.234557867050171, "learning_rate": 0.00019159892671082021, "loss": 4.5173, "step": 503350 }, { "epoch": 0.9394003718226541, "grad_norm": 1.2853304147720337, "learning_rate": 0.00019159724118482304, "loss": 4.5073, "step": 503400 }, { "epoch": 0.9394936773820326, "grad_norm": 0.987296462059021, "learning_rate": 0.00019159555549717286, "loss": 4.7587, "step": 503450 }, { "epoch": 0.9395869829414111, "grad_norm": 1.3254731893539429, "learning_rate": 0.00019159386964787264, "loss": 4.5207, "step": 503500 }, { "epoch": 0.9396802885007896, "grad_norm": 1.1575664281845093, "learning_rate": 0.00019159218363692536, "loss": 4.4809, "step": 503550 }, { "epoch": 0.9397735940601681, "grad_norm": 1.0227736234664917, "learning_rate": 0.00019159049746433397, "loss": 4.5333, "step": 503600 }, { "epoch": 0.9398668996195466, "grad_norm": 1.0817604064941406, "learning_rate": 0.0001915888111301015, "loss": 4.5315, "step": 503650 }, { "epoch": 0.9399602051789251, "grad_norm": 1.1532713174819946, "learning_rate": 0.00019158712463423084, "loss": 4.393, "step": 503700 }, { "epoch": 0.9400535107383036, "grad_norm": 1.0893367528915405, "learning_rate": 0.00019158543797672507, "loss": 4.7046, "step": 503750 }, { "epoch": 0.940146816297682, "grad_norm": 1.1007519960403442, "learning_rate": 0.0001915837511575871, "loss": 4.481, "step": 503800 }, { "epoch": 0.9402401218570605, "grad_norm": 0.9720431566238403, "learning_rate": 0.00019158206417681993, "loss": 4.5839, "step": 503850 }, { "epoch": 0.940333427416439, "grad_norm": 1.0621325969696045, "learning_rate": 0.00019158037703442653, "loss": 4.4535, "step": 503900 }, { "epoch": 0.9404267329758176, "grad_norm": 0.9895194172859192, "learning_rate": 0.00019157868973040986, "loss": 4.416, "step": 503950 }, { "epoch": 0.9405200385351961, "grad_norm": 0.8388993740081787, "learning_rate": 0.00019157700226477294, "loss": 4.6315, "step": 504000 }, { "epoch": 0.9405200385351961, "eval_loss": 4.70612096786499, "eval_runtime": 229.0705, "eval_samples_per_second": 11.385, "eval_steps_per_second": 11.385, "eval_tts_loss": 7.55402650199602, "step": 504000 }, { "epoch": 0.9406133440945745, "grad_norm": 1.0822992324829102, "learning_rate": 0.0001915753146375187, "loss": 4.5764, "step": 504050 }, { "epoch": 0.940706649653953, "grad_norm": 0.8459479808807373, "learning_rate": 0.0001915736268486502, "loss": 4.4599, "step": 504100 }, { "epoch": 0.9407999552133315, "grad_norm": 0.9300702810287476, "learning_rate": 0.0001915719388981703, "loss": 4.7675, "step": 504150 }, { "epoch": 0.94089326077271, "grad_norm": 1.2536556720733643, "learning_rate": 0.00019157025078608207, "loss": 4.5551, "step": 504200 }, { "epoch": 0.9409865663320884, "grad_norm": 0.7732415199279785, "learning_rate": 0.00019156856251238843, "loss": 4.4968, "step": 504250 }, { "epoch": 0.941079871891467, "grad_norm": 0.9805217981338501, "learning_rate": 0.00019156687407709242, "loss": 4.7133, "step": 504300 }, { "epoch": 0.9411731774508455, "grad_norm": 1.0305819511413574, "learning_rate": 0.00019156518548019697, "loss": 4.6051, "step": 504350 }, { "epoch": 0.941266483010224, "grad_norm": 1.1708893775939941, "learning_rate": 0.0001915634967217051, "loss": 4.3428, "step": 504400 }, { "epoch": 0.9413597885696025, "grad_norm": 1.1750810146331787, "learning_rate": 0.00019156180780161974, "loss": 4.5026, "step": 504450 }, { "epoch": 0.9414530941289809, "grad_norm": 1.595369577407837, "learning_rate": 0.00019156011871994394, "loss": 4.4609, "step": 504500 }, { "epoch": 0.9415463996883594, "grad_norm": 0.9636746644973755, "learning_rate": 0.0001915584294766806, "loss": 4.6637, "step": 504550 }, { "epoch": 0.9416397052477379, "grad_norm": 1.0972245931625366, "learning_rate": 0.00019155674007183276, "loss": 4.5445, "step": 504600 }, { "epoch": 0.9417330108071165, "grad_norm": 1.2407582998275757, "learning_rate": 0.0001915550505054034, "loss": 4.4367, "step": 504650 }, { "epoch": 0.9418263163664949, "grad_norm": 1.0888290405273438, "learning_rate": 0.00019155336077739547, "loss": 4.2476, "step": 504700 }, { "epoch": 0.9419196219258734, "grad_norm": 0.8973804116249084, "learning_rate": 0.00019155167088781198, "loss": 4.5612, "step": 504750 }, { "epoch": 0.9420129274852519, "grad_norm": 1.0939947366714478, "learning_rate": 0.00019154998083665588, "loss": 4.7406, "step": 504800 }, { "epoch": 0.9421062330446304, "grad_norm": 1.0775772333145142, "learning_rate": 0.00019154829062393016, "loss": 4.537, "step": 504850 }, { "epoch": 0.9421995386040088, "grad_norm": 1.0950112342834473, "learning_rate": 0.00019154660024963782, "loss": 4.4696, "step": 504900 }, { "epoch": 0.9422928441633873, "grad_norm": 1.3029534816741943, "learning_rate": 0.00019154490971378188, "loss": 4.5937, "step": 504950 }, { "epoch": 0.9423861497227659, "grad_norm": 0.7080609202384949, "learning_rate": 0.00019154321901636526, "loss": 4.567, "step": 505000 }, { "epoch": 0.9424794552821444, "grad_norm": 1.184503197669983, "learning_rate": 0.00019154152815739094, "loss": 4.5611, "step": 505050 }, { "epoch": 0.9425727608415229, "grad_norm": 1.1705070734024048, "learning_rate": 0.00019153983713686196, "loss": 4.3698, "step": 505100 }, { "epoch": 0.9426660664009013, "grad_norm": 0.9056841731071472, "learning_rate": 0.00019153814595478127, "loss": 4.3095, "step": 505150 }, { "epoch": 0.9427593719602798, "grad_norm": 1.3515528440475464, "learning_rate": 0.00019153645461115185, "loss": 4.4063, "step": 505200 }, { "epoch": 0.9428526775196583, "grad_norm": 1.2400225400924683, "learning_rate": 0.0001915347631059767, "loss": 4.4565, "step": 505250 }, { "epoch": 0.9429459830790368, "grad_norm": 1.181801676750183, "learning_rate": 0.00019153307143925878, "loss": 4.5905, "step": 505300 }, { "epoch": 0.9430392886384154, "grad_norm": 0.9441163539886475, "learning_rate": 0.00019153137961100113, "loss": 4.43, "step": 505350 }, { "epoch": 0.9431325941977938, "grad_norm": 0.9969195127487183, "learning_rate": 0.00019152968762120665, "loss": 4.5199, "step": 505400 }, { "epoch": 0.9432258997571723, "grad_norm": 0.986419141292572, "learning_rate": 0.00019152799546987844, "loss": 4.6016, "step": 505450 }, { "epoch": 0.9433192053165508, "grad_norm": 1.0366374254226685, "learning_rate": 0.0001915263031570194, "loss": 4.3475, "step": 505500 }, { "epoch": 0.9434125108759293, "grad_norm": 1.1309645175933838, "learning_rate": 0.00019152461068263252, "loss": 4.7278, "step": 505550 }, { "epoch": 0.9435058164353077, "grad_norm": 0.9096659421920776, "learning_rate": 0.00019152291804672082, "loss": 4.4641, "step": 505600 }, { "epoch": 0.9435991219946862, "grad_norm": 1.5472477674484253, "learning_rate": 0.0001915212252492873, "loss": 4.5692, "step": 505650 }, { "epoch": 0.9436924275540648, "grad_norm": 1.0727791786193848, "learning_rate": 0.00019151953229033488, "loss": 4.3292, "step": 505700 }, { "epoch": 0.9437857331134433, "grad_norm": 0.9225097894668579, "learning_rate": 0.0001915178391698666, "loss": 4.5358, "step": 505750 }, { "epoch": 0.9438790386728217, "grad_norm": 1.1473532915115356, "learning_rate": 0.00019151614588788543, "loss": 4.527, "step": 505800 }, { "epoch": 0.9439723442322002, "grad_norm": 0.9864091277122498, "learning_rate": 0.00019151445244439436, "loss": 4.4697, "step": 505850 }, { "epoch": 0.9440656497915787, "grad_norm": 1.0338512659072876, "learning_rate": 0.00019151275883939642, "loss": 4.6161, "step": 505900 }, { "epoch": 0.9441589553509572, "grad_norm": 1.1502354145050049, "learning_rate": 0.00019151106507289453, "loss": 4.5233, "step": 505950 }, { "epoch": 0.9442522609103357, "grad_norm": 1.1455802917480469, "learning_rate": 0.00019150937114489173, "loss": 4.4803, "step": 506000 }, { "epoch": 0.9443455664697142, "grad_norm": 0.6245649456977844, "learning_rate": 0.00019150767705539097, "loss": 4.6095, "step": 506050 }, { "epoch": 0.9444388720290927, "grad_norm": 1.0817911624908447, "learning_rate": 0.00019150598280439528, "loss": 4.5613, "step": 506100 }, { "epoch": 0.9445321775884712, "grad_norm": 1.5298781394958496, "learning_rate": 0.00019150428839190763, "loss": 4.6667, "step": 506150 }, { "epoch": 0.9446254831478497, "grad_norm": 0.7002063393592834, "learning_rate": 0.000191502593817931, "loss": 4.5497, "step": 506200 }, { "epoch": 0.9447187887072281, "grad_norm": 1.5078949928283691, "learning_rate": 0.0001915008990824684, "loss": 4.599, "step": 506250 }, { "epoch": 0.9448120942666066, "grad_norm": 1.227243185043335, "learning_rate": 0.0001914992041855228, "loss": 4.5282, "step": 506300 }, { "epoch": 0.9449053998259851, "grad_norm": 1.1678166389465332, "learning_rate": 0.00019149750912709722, "loss": 4.2659, "step": 506350 }, { "epoch": 0.9449987053853636, "grad_norm": 1.213831901550293, "learning_rate": 0.00019149581390719465, "loss": 4.5668, "step": 506400 }, { "epoch": 0.9450920109447422, "grad_norm": 1.337912678718567, "learning_rate": 0.00019149411852581807, "loss": 4.4615, "step": 506450 }, { "epoch": 0.9451853165041206, "grad_norm": 1.1013575792312622, "learning_rate": 0.00019149242298297043, "loss": 4.3458, "step": 506500 }, { "epoch": 0.9452786220634991, "grad_norm": 0.9433706998825073, "learning_rate": 0.0001914907272786548, "loss": 4.5446, "step": 506550 }, { "epoch": 0.9453719276228776, "grad_norm": 1.216244101524353, "learning_rate": 0.0001914890314128741, "loss": 4.677, "step": 506600 }, { "epoch": 0.9454652331822561, "grad_norm": 1.3864295482635498, "learning_rate": 0.00019148733538563138, "loss": 4.4114, "step": 506650 }, { "epoch": 0.9455585387416345, "grad_norm": 1.1006184816360474, "learning_rate": 0.0001914856391969296, "loss": 4.6923, "step": 506700 }, { "epoch": 0.945651844301013, "grad_norm": 0.9698508381843567, "learning_rate": 0.0001914839428467718, "loss": 4.616, "step": 506750 }, { "epoch": 0.9457451498603916, "grad_norm": 1.0982531309127808, "learning_rate": 0.0001914822463351609, "loss": 4.5469, "step": 506800 }, { "epoch": 0.9458384554197701, "grad_norm": 1.2480262517929077, "learning_rate": 0.00019148054966209994, "loss": 4.6309, "step": 506850 }, { "epoch": 0.9459317609791486, "grad_norm": 0.8199253678321838, "learning_rate": 0.0001914788528275919, "loss": 4.596, "step": 506900 }, { "epoch": 0.946025066538527, "grad_norm": 0.9453520774841309, "learning_rate": 0.00019147715583163978, "loss": 4.6424, "step": 506950 }, { "epoch": 0.9461183720979055, "grad_norm": 0.9872258901596069, "learning_rate": 0.0001914754586742466, "loss": 4.5295, "step": 507000 }, { "epoch": 0.9461183720979055, "eval_loss": 4.7034807205200195, "eval_runtime": 229.4578, "eval_samples_per_second": 11.366, "eval_steps_per_second": 11.366, "eval_tts_loss": 7.652594781353601, "step": 507000 }, { "epoch": 0.946211677657284, "grad_norm": 1.1275808811187744, "learning_rate": 0.0001914737613554153, "loss": 4.6108, "step": 507050 }, { "epoch": 0.9463049832166625, "grad_norm": 1.3099130392074585, "learning_rate": 0.00019147206387514894, "loss": 4.6085, "step": 507100 }, { "epoch": 0.946398288776041, "grad_norm": 1.145595908164978, "learning_rate": 0.00019147036623345045, "loss": 4.4127, "step": 507150 }, { "epoch": 0.9464915943354195, "grad_norm": 1.1868587732315063, "learning_rate": 0.00019146866843032287, "loss": 4.584, "step": 507200 }, { "epoch": 0.946584899894798, "grad_norm": 1.085736870765686, "learning_rate": 0.0001914669704657692, "loss": 4.6588, "step": 507250 }, { "epoch": 0.9466782054541765, "grad_norm": 1.0987493991851807, "learning_rate": 0.0001914652723397924, "loss": 4.4645, "step": 507300 }, { "epoch": 0.946771511013555, "grad_norm": 1.005104660987854, "learning_rate": 0.0001914635740523955, "loss": 4.3681, "step": 507350 }, { "epoch": 0.9468648165729334, "grad_norm": 1.098920226097107, "learning_rate": 0.00019146187560358148, "loss": 4.4385, "step": 507400 }, { "epoch": 0.9469581221323119, "grad_norm": 1.0033411979675293, "learning_rate": 0.00019146017699335334, "loss": 4.6221, "step": 507450 }, { "epoch": 0.9470514276916905, "grad_norm": 0.9110998511314392, "learning_rate": 0.0001914584782217141, "loss": 4.3511, "step": 507500 }, { "epoch": 0.947144733251069, "grad_norm": 0.86753910779953, "learning_rate": 0.0001914567792886667, "loss": 4.6094, "step": 507550 }, { "epoch": 0.9472380388104474, "grad_norm": 1.3776609897613525, "learning_rate": 0.00019145508019421421, "loss": 4.5604, "step": 507600 }, { "epoch": 0.9473313443698259, "grad_norm": 1.0817846059799194, "learning_rate": 0.0001914533809383596, "loss": 4.7804, "step": 507650 }, { "epoch": 0.9474246499292044, "grad_norm": 1.044704794883728, "learning_rate": 0.0001914516815211058, "loss": 4.6772, "step": 507700 }, { "epoch": 0.9475179554885829, "grad_norm": 1.2120769023895264, "learning_rate": 0.00019144998194245595, "loss": 4.6616, "step": 507750 }, { "epoch": 0.9476112610479613, "grad_norm": 1.1874451637268066, "learning_rate": 0.00019144828220241295, "loss": 4.6984, "step": 507800 }, { "epoch": 0.9477045666073399, "grad_norm": 1.2390722036361694, "learning_rate": 0.00019144658230097982, "loss": 4.5209, "step": 507850 }, { "epoch": 0.9477978721667184, "grad_norm": 1.2326321601867676, "learning_rate": 0.00019144488223815954, "loss": 4.5146, "step": 507900 }, { "epoch": 0.9478911777260969, "grad_norm": 1.2687432765960693, "learning_rate": 0.0001914431820139552, "loss": 4.6694, "step": 507950 }, { "epoch": 0.9479844832854754, "grad_norm": 1.1790771484375, "learning_rate": 0.00019144148162836966, "loss": 4.3071, "step": 508000 }, { "epoch": 0.9480777888448538, "grad_norm": 0.9642509818077087, "learning_rate": 0.00019143978108140602, "loss": 4.403, "step": 508050 }, { "epoch": 0.9481710944042323, "grad_norm": 1.1100856065750122, "learning_rate": 0.00019143808037306727, "loss": 4.4633, "step": 508100 }, { "epoch": 0.9482643999636108, "grad_norm": 1.048416018486023, "learning_rate": 0.00019143637950335638, "loss": 4.5011, "step": 508150 }, { "epoch": 0.9483577055229894, "grad_norm": 1.2268264293670654, "learning_rate": 0.0001914346784722764, "loss": 4.4618, "step": 508200 }, { "epoch": 0.9484510110823678, "grad_norm": 1.0577186346054077, "learning_rate": 0.00019143297727983028, "loss": 4.4674, "step": 508250 }, { "epoch": 0.9485443166417463, "grad_norm": 1.46774160861969, "learning_rate": 0.000191431275926021, "loss": 4.7837, "step": 508300 }, { "epoch": 0.9486376222011248, "grad_norm": 1.2539163827896118, "learning_rate": 0.0001914295744108517, "loss": 4.6426, "step": 508350 }, { "epoch": 0.9487309277605033, "grad_norm": 1.2751028537750244, "learning_rate": 0.0001914278727343252, "loss": 4.6279, "step": 508400 }, { "epoch": 0.9488242333198817, "grad_norm": 1.018558144569397, "learning_rate": 0.0001914261708964446, "loss": 4.6285, "step": 508450 }, { "epoch": 0.9489175388792602, "grad_norm": 1.1424463987350464, "learning_rate": 0.00019142446889721296, "loss": 4.6577, "step": 508500 }, { "epoch": 0.9490108444386388, "grad_norm": 1.2826651334762573, "learning_rate": 0.00019142276673663315, "loss": 4.6548, "step": 508550 }, { "epoch": 0.9491041499980173, "grad_norm": 1.3990439176559448, "learning_rate": 0.0001914210644147083, "loss": 4.8168, "step": 508600 }, { "epoch": 0.9491974555573958, "grad_norm": 1.088784098625183, "learning_rate": 0.00019141936193144132, "loss": 4.5858, "step": 508650 }, { "epoch": 0.9492907611167742, "grad_norm": 1.2218523025512695, "learning_rate": 0.00019141765928683526, "loss": 4.6277, "step": 508700 }, { "epoch": 0.9493840666761527, "grad_norm": 0.9638603925704956, "learning_rate": 0.00019141595648089311, "loss": 4.4823, "step": 508750 }, { "epoch": 0.9494773722355312, "grad_norm": 0.7992116212844849, "learning_rate": 0.00019141425351361792, "loss": 4.4877, "step": 508800 }, { "epoch": 0.9495706777949097, "grad_norm": 0.948126494884491, "learning_rate": 0.0001914125503850126, "loss": 4.6357, "step": 508850 }, { "epoch": 0.9496639833542883, "grad_norm": 1.2118310928344727, "learning_rate": 0.00019141084709508026, "loss": 4.4274, "step": 508900 }, { "epoch": 0.9497572889136667, "grad_norm": 1.1753696203231812, "learning_rate": 0.00019140914364382383, "loss": 4.3724, "step": 508950 }, { "epoch": 0.9498505944730452, "grad_norm": 1.201859951019287, "learning_rate": 0.00019140744003124635, "loss": 4.5276, "step": 509000 }, { "epoch": 0.9499439000324237, "grad_norm": 1.1688952445983887, "learning_rate": 0.00019140573625735084, "loss": 4.5749, "step": 509050 }, { "epoch": 0.9500372055918022, "grad_norm": 1.2634717226028442, "learning_rate": 0.00019140403232214027, "loss": 4.7231, "step": 509100 }, { "epoch": 0.9501305111511806, "grad_norm": 1.114612102508545, "learning_rate": 0.0001914023282256177, "loss": 4.598, "step": 509150 }, { "epoch": 0.9502238167105591, "grad_norm": 0.9612153172492981, "learning_rate": 0.00019140062396778604, "loss": 4.5885, "step": 509200 }, { "epoch": 0.9503171222699377, "grad_norm": 1.1103726625442505, "learning_rate": 0.0001913989195486484, "loss": 4.3738, "step": 509250 }, { "epoch": 0.9504104278293162, "grad_norm": 1.1408940553665161, "learning_rate": 0.00019139721496820776, "loss": 4.3502, "step": 509300 }, { "epoch": 0.9505037333886946, "grad_norm": 1.1231387853622437, "learning_rate": 0.0001913955102264671, "loss": 4.6939, "step": 509350 }, { "epoch": 0.9505970389480731, "grad_norm": 1.0511226654052734, "learning_rate": 0.00019139380532342943, "loss": 4.3805, "step": 509400 }, { "epoch": 0.9506903445074516, "grad_norm": 1.2479896545410156, "learning_rate": 0.0001913921002590978, "loss": 4.5346, "step": 509450 }, { "epoch": 0.9507836500668301, "grad_norm": 1.035717487335205, "learning_rate": 0.0001913903950334752, "loss": 4.527, "step": 509500 }, { "epoch": 0.9508769556262086, "grad_norm": 1.0333224534988403, "learning_rate": 0.00019138868964656462, "loss": 4.5744, "step": 509550 }, { "epoch": 0.9509702611855871, "grad_norm": 0.8429966568946838, "learning_rate": 0.00019138698409836908, "loss": 4.378, "step": 509600 }, { "epoch": 0.9510635667449656, "grad_norm": 0.958810031414032, "learning_rate": 0.0001913852783888916, "loss": 4.4409, "step": 509650 }, { "epoch": 0.9511568723043441, "grad_norm": 1.4274485111236572, "learning_rate": 0.0001913835725181352, "loss": 4.6393, "step": 509700 }, { "epoch": 0.9512501778637226, "grad_norm": 1.2070671319961548, "learning_rate": 0.00019138186648610286, "loss": 4.5333, "step": 509750 }, { "epoch": 0.951343483423101, "grad_norm": 1.112189769744873, "learning_rate": 0.0001913801602927976, "loss": 4.6378, "step": 509800 }, { "epoch": 0.9514367889824795, "grad_norm": 1.0322123765945435, "learning_rate": 0.00019137845393822243, "loss": 4.6306, "step": 509850 }, { "epoch": 0.951530094541858, "grad_norm": 0.9617332816123962, "learning_rate": 0.00019137674742238035, "loss": 4.4933, "step": 509900 }, { "epoch": 0.9516234001012366, "grad_norm": 0.9105497002601624, "learning_rate": 0.00019137504074527443, "loss": 4.5947, "step": 509950 }, { "epoch": 0.9517167056606151, "grad_norm": 1.0770008563995361, "learning_rate": 0.0001913733339069076, "loss": 4.6508, "step": 510000 }, { "epoch": 0.9517167056606151, "eval_loss": 4.700717926025391, "eval_runtime": 228.3952, "eval_samples_per_second": 11.419, "eval_steps_per_second": 11.419, "eval_tts_loss": 7.58605571152311, "step": 510000 }, { "epoch": 0.9518100112199935, "grad_norm": 1.2147283554077148, "learning_rate": 0.00019137162690728294, "loss": 4.7256, "step": 510050 }, { "epoch": 0.951903316779372, "grad_norm": 1.20662522315979, "learning_rate": 0.00019136991974640343, "loss": 4.6374, "step": 510100 }, { "epoch": 0.9519966223387505, "grad_norm": 1.095915675163269, "learning_rate": 0.00019136821242427213, "loss": 4.4668, "step": 510150 }, { "epoch": 0.952089927898129, "grad_norm": 0.9449310302734375, "learning_rate": 0.00019136650494089197, "loss": 4.5404, "step": 510200 }, { "epoch": 0.9521832334575074, "grad_norm": 1.139679193496704, "learning_rate": 0.000191364797296266, "loss": 4.6257, "step": 510250 }, { "epoch": 0.952276539016886, "grad_norm": 1.191811442375183, "learning_rate": 0.00019136308949039722, "loss": 4.549, "step": 510300 }, { "epoch": 0.9523698445762645, "grad_norm": 0.7286585569381714, "learning_rate": 0.0001913613815232887, "loss": 4.5917, "step": 510350 }, { "epoch": 0.952463150135643, "grad_norm": 1.2073794603347778, "learning_rate": 0.00019135967339494343, "loss": 4.5037, "step": 510400 }, { "epoch": 0.9525564556950215, "grad_norm": 0.8102605938911438, "learning_rate": 0.00019135796510536438, "loss": 4.4903, "step": 510450 }, { "epoch": 0.9526497612543999, "grad_norm": 1.4343971014022827, "learning_rate": 0.0001913562566545546, "loss": 4.5497, "step": 510500 }, { "epoch": 0.9527430668137784, "grad_norm": 1.0295119285583496, "learning_rate": 0.0001913545480425171, "loss": 4.5604, "step": 510550 }, { "epoch": 0.9528363723731569, "grad_norm": 1.0169522762298584, "learning_rate": 0.00019135283926925492, "loss": 4.5347, "step": 510600 }, { "epoch": 0.9529296779325355, "grad_norm": 0.8969013094902039, "learning_rate": 0.00019135113033477103, "loss": 4.6475, "step": 510650 }, { "epoch": 0.9530229834919139, "grad_norm": 1.1601159572601318, "learning_rate": 0.00019134942123906848, "loss": 4.6288, "step": 510700 }, { "epoch": 0.9531162890512924, "grad_norm": 1.1090325117111206, "learning_rate": 0.00019134771198215026, "loss": 4.6869, "step": 510750 }, { "epoch": 0.9532095946106709, "grad_norm": 1.1559101343154907, "learning_rate": 0.00019134600256401943, "loss": 4.5961, "step": 510800 }, { "epoch": 0.9533029001700494, "grad_norm": 1.1502505540847778, "learning_rate": 0.00019134429298467895, "loss": 4.6023, "step": 510850 }, { "epoch": 0.9533962057294278, "grad_norm": 1.173335313796997, "learning_rate": 0.0001913425832441319, "loss": 4.4694, "step": 510900 }, { "epoch": 0.9534895112888063, "grad_norm": 1.4289402961730957, "learning_rate": 0.0001913408733423812, "loss": 4.5789, "step": 510950 }, { "epoch": 0.9535828168481849, "grad_norm": 1.192551851272583, "learning_rate": 0.00019133916327943, "loss": 4.6357, "step": 511000 }, { "epoch": 0.9536761224075634, "grad_norm": 1.3199892044067383, "learning_rate": 0.00019133745305528123, "loss": 4.627, "step": 511050 }, { "epoch": 0.9537694279669419, "grad_norm": 1.1246048212051392, "learning_rate": 0.00019133574266993791, "loss": 4.3647, "step": 511100 }, { "epoch": 0.9538627335263203, "grad_norm": 0.8767019510269165, "learning_rate": 0.00019133403212340304, "loss": 4.633, "step": 511150 }, { "epoch": 0.9539560390856988, "grad_norm": 1.2108631134033203, "learning_rate": 0.00019133232141567975, "loss": 4.6334, "step": 511200 }, { "epoch": 0.9540493446450773, "grad_norm": 1.1174519062042236, "learning_rate": 0.00019133061054677094, "loss": 4.4731, "step": 511250 }, { "epoch": 0.9541426502044558, "grad_norm": 0.9424172043800354, "learning_rate": 0.00019132889951667967, "loss": 4.4385, "step": 511300 }, { "epoch": 0.9542359557638344, "grad_norm": 1.3405001163482666, "learning_rate": 0.00019132718832540896, "loss": 4.4583, "step": 511350 }, { "epoch": 0.9543292613232128, "grad_norm": 1.055501937866211, "learning_rate": 0.00019132547697296186, "loss": 4.713, "step": 511400 }, { "epoch": 0.9544225668825913, "grad_norm": 1.0255764722824097, "learning_rate": 0.00019132376545934134, "loss": 4.7617, "step": 511450 }, { "epoch": 0.9545158724419698, "grad_norm": 0.8952128291130066, "learning_rate": 0.00019132205378455047, "loss": 4.4883, "step": 511500 }, { "epoch": 0.9546091780013483, "grad_norm": 1.2182574272155762, "learning_rate": 0.0001913203419485922, "loss": 4.4766, "step": 511550 }, { "epoch": 0.9547024835607267, "grad_norm": 0.8846496939659119, "learning_rate": 0.0001913186299514696, "loss": 4.3567, "step": 511600 }, { "epoch": 0.9547957891201052, "grad_norm": 1.2655670642852783, "learning_rate": 0.0001913169177931857, "loss": 4.4254, "step": 511650 }, { "epoch": 0.9548890946794837, "grad_norm": 0.9712004661560059, "learning_rate": 0.00019131520547374352, "loss": 4.453, "step": 511700 }, { "epoch": 0.9549824002388623, "grad_norm": 1.2166794538497925, "learning_rate": 0.00019131349299314606, "loss": 4.3828, "step": 511750 }, { "epoch": 0.9550757057982407, "grad_norm": 1.251700520515442, "learning_rate": 0.00019131178035139637, "loss": 4.5611, "step": 511800 }, { "epoch": 0.9551690113576192, "grad_norm": 0.8383100032806396, "learning_rate": 0.00019131006754849743, "loss": 4.6724, "step": 511850 }, { "epoch": 0.9552623169169977, "grad_norm": 1.186093807220459, "learning_rate": 0.0001913083545844523, "loss": 4.4113, "step": 511900 }, { "epoch": 0.9553556224763762, "grad_norm": 0.9625120162963867, "learning_rate": 0.000191306641459264, "loss": 4.5281, "step": 511950 }, { "epoch": 0.9554489280357547, "grad_norm": 1.259109377861023, "learning_rate": 0.0001913049281729355, "loss": 4.5746, "step": 512000 }, { "epoch": 0.9555422335951331, "grad_norm": 0.9541958570480347, "learning_rate": 0.00019130321472546992, "loss": 4.5897, "step": 512050 }, { "epoch": 0.9556355391545117, "grad_norm": 1.281565546989441, "learning_rate": 0.00019130150111687017, "loss": 4.5235, "step": 512100 }, { "epoch": 0.9557288447138902, "grad_norm": 1.0066953897476196, "learning_rate": 0.0001912997873471394, "loss": 4.601, "step": 512150 }, { "epoch": 0.9558221502732687, "grad_norm": 1.0612179040908813, "learning_rate": 0.00019129807341628055, "loss": 4.6265, "step": 512200 }, { "epoch": 0.9559154558326471, "grad_norm": 1.2616524696350098, "learning_rate": 0.00019129635932429664, "loss": 4.5701, "step": 512250 }, { "epoch": 0.9560087613920256, "grad_norm": 1.064116358757019, "learning_rate": 0.00019129464507119072, "loss": 4.4622, "step": 512300 }, { "epoch": 0.9561020669514041, "grad_norm": 1.2274696826934814, "learning_rate": 0.00019129293065696585, "loss": 4.6811, "step": 512350 }, { "epoch": 0.9561953725107826, "grad_norm": 1.2299728393554688, "learning_rate": 0.00019129121608162501, "loss": 4.7158, "step": 512400 }, { "epoch": 0.9562886780701612, "grad_norm": 0.8861685991287231, "learning_rate": 0.00019128950134517122, "loss": 4.7097, "step": 512450 }, { "epoch": 0.9563819836295396, "grad_norm": 1.276901364326477, "learning_rate": 0.00019128778644760753, "loss": 4.5695, "step": 512500 }, { "epoch": 0.9564752891889181, "grad_norm": 1.0338551998138428, "learning_rate": 0.00019128607138893698, "loss": 4.5535, "step": 512550 }, { "epoch": 0.9565685947482966, "grad_norm": 1.0839725732803345, "learning_rate": 0.00019128435616916253, "loss": 4.631, "step": 512600 }, { "epoch": 0.9566619003076751, "grad_norm": 1.2017780542373657, "learning_rate": 0.00019128264078828732, "loss": 4.5005, "step": 512650 }, { "epoch": 0.9567552058670535, "grad_norm": 1.1648411750793457, "learning_rate": 0.00019128092524631427, "loss": 4.5858, "step": 512700 }, { "epoch": 0.956848511426432, "grad_norm": 1.1855037212371826, "learning_rate": 0.00019127920954324643, "loss": 4.5213, "step": 512750 }, { "epoch": 0.9569418169858106, "grad_norm": 0.7989550232887268, "learning_rate": 0.00019127749367908687, "loss": 4.417, "step": 512800 }, { "epoch": 0.9570351225451891, "grad_norm": 1.4755687713623047, "learning_rate": 0.00019127577765383862, "loss": 4.3803, "step": 512850 }, { "epoch": 0.9571284281045676, "grad_norm": 1.0128858089447021, "learning_rate": 0.00019127406146750463, "loss": 4.5694, "step": 512900 }, { "epoch": 0.957221733663946, "grad_norm": 1.1745485067367554, "learning_rate": 0.000191272345120088, "loss": 4.7241, "step": 512950 }, { "epoch": 0.9573150392233245, "grad_norm": 1.0665655136108398, "learning_rate": 0.00019127062861159177, "loss": 4.5249, "step": 513000 }, { "epoch": 0.9573150392233245, "eval_loss": 4.706900596618652, "eval_runtime": 228.0016, "eval_samples_per_second": 11.439, "eval_steps_per_second": 11.439, "eval_tts_loss": 7.588165010949825, "step": 513000 }, { "epoch": 0.957408344782703, "grad_norm": 0.9406837821006775, "learning_rate": 0.00019126891194201891, "loss": 4.587, "step": 513050 }, { "epoch": 0.9575016503420815, "grad_norm": 1.0566860437393188, "learning_rate": 0.0001912671951113725, "loss": 4.5948, "step": 513100 }, { "epoch": 0.95759495590146, "grad_norm": 1.175111174583435, "learning_rate": 0.00019126547811965553, "loss": 4.4508, "step": 513150 }, { "epoch": 0.9576882614608385, "grad_norm": 1.3218798637390137, "learning_rate": 0.00019126376096687103, "loss": 4.6365, "step": 513200 }, { "epoch": 0.957781567020217, "grad_norm": 0.8846462965011597, "learning_rate": 0.0001912620436530221, "loss": 4.4954, "step": 513250 }, { "epoch": 0.9578748725795955, "grad_norm": 1.115067481994629, "learning_rate": 0.00019126032617811168, "loss": 4.4764, "step": 513300 }, { "epoch": 0.957968178138974, "grad_norm": 0.9618848562240601, "learning_rate": 0.00019125860854214287, "loss": 4.5889, "step": 513350 }, { "epoch": 0.9580614836983524, "grad_norm": 1.0808846950531006, "learning_rate": 0.00019125689074511864, "loss": 4.3513, "step": 513400 }, { "epoch": 0.9581547892577309, "grad_norm": 1.0568145513534546, "learning_rate": 0.0001912551727870421, "loss": 4.4852, "step": 513450 }, { "epoch": 0.9582480948171095, "grad_norm": 1.2343864440917969, "learning_rate": 0.00019125345466791622, "loss": 4.5879, "step": 513500 }, { "epoch": 0.958341400376488, "grad_norm": 1.3668192625045776, "learning_rate": 0.00019125173638774405, "loss": 4.6623, "step": 513550 }, { "epoch": 0.9584347059358664, "grad_norm": 1.00023353099823, "learning_rate": 0.0001912500179465286, "loss": 4.6536, "step": 513600 }, { "epoch": 0.9585280114952449, "grad_norm": 1.1797987222671509, "learning_rate": 0.00019124829934427296, "loss": 4.5728, "step": 513650 }, { "epoch": 0.9586213170546234, "grad_norm": 0.835677981376648, "learning_rate": 0.00019124658058098008, "loss": 4.4838, "step": 513700 }, { "epoch": 0.9587146226140019, "grad_norm": 1.1132127046585083, "learning_rate": 0.00019124486165665307, "loss": 4.5901, "step": 513750 }, { "epoch": 0.9588079281733803, "grad_norm": 0.9234887361526489, "learning_rate": 0.00019124314257129495, "loss": 4.4613, "step": 513800 }, { "epoch": 0.9589012337327589, "grad_norm": 0.8118153214454651, "learning_rate": 0.00019124142332490872, "loss": 4.4436, "step": 513850 }, { "epoch": 0.9589945392921374, "grad_norm": 1.2153241634368896, "learning_rate": 0.00019123970391749743, "loss": 4.4398, "step": 513900 }, { "epoch": 0.9590878448515159, "grad_norm": 1.2988272905349731, "learning_rate": 0.0001912379843490641, "loss": 4.5709, "step": 513950 }, { "epoch": 0.9591811504108944, "grad_norm": 1.180262565612793, "learning_rate": 0.00019123626461961182, "loss": 4.5929, "step": 514000 }, { "epoch": 0.9592744559702728, "grad_norm": 1.1091892719268799, "learning_rate": 0.00019123454472914358, "loss": 4.7876, "step": 514050 }, { "epoch": 0.9593677615296513, "grad_norm": 1.2552402019500732, "learning_rate": 0.0001912328246776624, "loss": 4.5823, "step": 514100 }, { "epoch": 0.9594610670890298, "grad_norm": 1.3410017490386963, "learning_rate": 0.00019123110446517134, "loss": 4.6766, "step": 514150 }, { "epoch": 0.9595543726484084, "grad_norm": 1.0101282596588135, "learning_rate": 0.00019122938409167343, "loss": 4.4324, "step": 514200 }, { "epoch": 0.9596476782077868, "grad_norm": 1.3024471998214722, "learning_rate": 0.00019122766355717175, "loss": 4.5987, "step": 514250 }, { "epoch": 0.9597409837671653, "grad_norm": 1.286859154701233, "learning_rate": 0.00019122594286166925, "loss": 4.6245, "step": 514300 }, { "epoch": 0.9598342893265438, "grad_norm": 0.948573112487793, "learning_rate": 0.00019122422200516903, "loss": 4.5755, "step": 514350 }, { "epoch": 0.9599275948859223, "grad_norm": 1.2615385055541992, "learning_rate": 0.00019122250098767412, "loss": 4.5497, "step": 514400 }, { "epoch": 0.9600209004453008, "grad_norm": 1.0237566232681274, "learning_rate": 0.00019122077980918752, "loss": 4.5499, "step": 514450 }, { "epoch": 0.9601142060046792, "grad_norm": 1.3502542972564697, "learning_rate": 0.00019121905846971232, "loss": 4.6049, "step": 514500 }, { "epoch": 0.9602075115640578, "grad_norm": 1.2443954944610596, "learning_rate": 0.0001912173369692515, "loss": 4.5759, "step": 514550 }, { "epoch": 0.9603008171234363, "grad_norm": 0.953113853931427, "learning_rate": 0.00019121561530780813, "loss": 4.4111, "step": 514600 }, { "epoch": 0.9603941226828148, "grad_norm": 1.4755457639694214, "learning_rate": 0.0001912138934853853, "loss": 4.7263, "step": 514650 }, { "epoch": 0.9604874282421932, "grad_norm": 1.3381918668746948, "learning_rate": 0.00019121217150198596, "loss": 4.7537, "step": 514700 }, { "epoch": 0.9605807338015717, "grad_norm": 1.1979091167449951, "learning_rate": 0.0001912104493576132, "loss": 4.5042, "step": 514750 }, { "epoch": 0.9606740393609502, "grad_norm": 1.1829605102539062, "learning_rate": 0.00019120872705227004, "loss": 4.5382, "step": 514800 }, { "epoch": 0.9607673449203287, "grad_norm": 1.1124389171600342, "learning_rate": 0.00019120700458595953, "loss": 4.456, "step": 514850 }, { "epoch": 0.9608606504797073, "grad_norm": 1.1636784076690674, "learning_rate": 0.0001912052819586847, "loss": 4.5455, "step": 514900 }, { "epoch": 0.9609539560390857, "grad_norm": 1.2731540203094482, "learning_rate": 0.00019120355917044858, "loss": 4.6987, "step": 514950 }, { "epoch": 0.9610472615984642, "grad_norm": 1.4770314693450928, "learning_rate": 0.00019120183622125424, "loss": 4.7013, "step": 515000 }, { "epoch": 0.9611405671578427, "grad_norm": 0.9899406433105469, "learning_rate": 0.0001912001131111047, "loss": 4.5914, "step": 515050 }, { "epoch": 0.9612338727172212, "grad_norm": 0.9768618941307068, "learning_rate": 0.00019119838984000303, "loss": 4.5281, "step": 515100 }, { "epoch": 0.9613271782765996, "grad_norm": 0.9205986261367798, "learning_rate": 0.0001911966664079522, "loss": 4.5482, "step": 515150 }, { "epoch": 0.9614204838359781, "grad_norm": 1.0015758275985718, "learning_rate": 0.00019119494281495537, "loss": 4.3742, "step": 515200 }, { "epoch": 0.9615137893953567, "grad_norm": 1.2049912214279175, "learning_rate": 0.00019119321906101546, "loss": 4.5792, "step": 515250 }, { "epoch": 0.9616070949547352, "grad_norm": 1.0098450183868408, "learning_rate": 0.00019119149514613558, "loss": 4.477, "step": 515300 }, { "epoch": 0.9617004005141137, "grad_norm": 1.1096776723861694, "learning_rate": 0.00019118977107031876, "loss": 4.2704, "step": 515350 }, { "epoch": 0.9617937060734921, "grad_norm": 1.2406129837036133, "learning_rate": 0.00019118804683356807, "loss": 4.4237, "step": 515400 }, { "epoch": 0.9618870116328706, "grad_norm": 1.6609387397766113, "learning_rate": 0.00019118632243588648, "loss": 4.6336, "step": 515450 }, { "epoch": 0.9619803171922491, "grad_norm": 1.6383963823318481, "learning_rate": 0.00019118459787727707, "loss": 4.4773, "step": 515500 }, { "epoch": 0.9620736227516276, "grad_norm": 1.0180213451385498, "learning_rate": 0.0001911828731577429, "loss": 4.5634, "step": 515550 }, { "epoch": 0.9621669283110061, "grad_norm": 1.4522619247436523, "learning_rate": 0.00019118114827728703, "loss": 4.5308, "step": 515600 }, { "epoch": 0.9622602338703846, "grad_norm": 1.0256778001785278, "learning_rate": 0.00019117942323591244, "loss": 4.5014, "step": 515650 }, { "epoch": 0.9623535394297631, "grad_norm": 1.0212059020996094, "learning_rate": 0.00019117769803362223, "loss": 4.6578, "step": 515700 }, { "epoch": 0.9624468449891416, "grad_norm": 1.0357530117034912, "learning_rate": 0.0001911759726704194, "loss": 4.6828, "step": 515750 }, { "epoch": 0.96254015054852, "grad_norm": 1.6250239610671997, "learning_rate": 0.00019117424714630706, "loss": 4.5253, "step": 515800 }, { "epoch": 0.9626334561078985, "grad_norm": 1.1354687213897705, "learning_rate": 0.0001911725214612882, "loss": 4.4391, "step": 515850 }, { "epoch": 0.962726761667277, "grad_norm": 0.9825906157493591, "learning_rate": 0.00019117079561536588, "loss": 4.4771, "step": 515900 }, { "epoch": 0.9628200672266556, "grad_norm": 0.9522579908370972, "learning_rate": 0.00019116906960854314, "loss": 4.6786, "step": 515950 }, { "epoch": 0.9629133727860341, "grad_norm": 1.2736667394638062, "learning_rate": 0.00019116734344082304, "loss": 4.4803, "step": 516000 }, { "epoch": 0.9629133727860341, "eval_loss": 4.6960129737854, "eval_runtime": 226.8369, "eval_samples_per_second": 11.497, "eval_steps_per_second": 11.497, "eval_tts_loss": 7.630245616222429, "step": 516000 }, { "epoch": 0.9630066783454125, "grad_norm": 1.0559250116348267, "learning_rate": 0.00019116561711220864, "loss": 4.5395, "step": 516050 }, { "epoch": 0.963099983904791, "grad_norm": 1.3147692680358887, "learning_rate": 0.00019116389062270295, "loss": 4.6774, "step": 516100 }, { "epoch": 0.9631932894641695, "grad_norm": 1.0479824542999268, "learning_rate": 0.00019116216397230905, "loss": 4.6161, "step": 516150 }, { "epoch": 0.963286595023548, "grad_norm": 1.3919200897216797, "learning_rate": 0.00019116043716102993, "loss": 4.332, "step": 516200 }, { "epoch": 0.9633799005829264, "grad_norm": 1.5771493911743164, "learning_rate": 0.00019115871018886873, "loss": 4.6193, "step": 516250 }, { "epoch": 0.963473206142305, "grad_norm": 1.1621042490005493, "learning_rate": 0.00019115698305582842, "loss": 4.7933, "step": 516300 }, { "epoch": 0.9635665117016835, "grad_norm": 1.318016767501831, "learning_rate": 0.00019115525576191206, "loss": 4.5998, "step": 516350 }, { "epoch": 0.963659817261062, "grad_norm": 1.270268201828003, "learning_rate": 0.00019115352830712272, "loss": 4.4825, "step": 516400 }, { "epoch": 0.9637531228204405, "grad_norm": 1.2554879188537598, "learning_rate": 0.00019115180069146346, "loss": 4.5602, "step": 516450 }, { "epoch": 0.9638464283798189, "grad_norm": 1.1198776960372925, "learning_rate": 0.00019115007291493728, "loss": 4.6057, "step": 516500 }, { "epoch": 0.9639397339391974, "grad_norm": 1.1703027486801147, "learning_rate": 0.0001911483449775473, "loss": 4.6311, "step": 516550 }, { "epoch": 0.9640330394985759, "grad_norm": 0.8470339775085449, "learning_rate": 0.00019114661687929648, "loss": 4.5826, "step": 516600 }, { "epoch": 0.9641263450579545, "grad_norm": 1.2922124862670898, "learning_rate": 0.00019114488862018796, "loss": 4.6169, "step": 516650 }, { "epoch": 0.964219650617333, "grad_norm": 0.991391122341156, "learning_rate": 0.00019114316020022476, "loss": 4.621, "step": 516700 }, { "epoch": 0.9643129561767114, "grad_norm": 0.8454660773277283, "learning_rate": 0.00019114143161940987, "loss": 4.4598, "step": 516750 }, { "epoch": 0.9644062617360899, "grad_norm": 1.271307110786438, "learning_rate": 0.00019113970287774643, "loss": 4.4723, "step": 516800 }, { "epoch": 0.9644995672954684, "grad_norm": 1.0154337882995605, "learning_rate": 0.00019113797397523743, "loss": 4.6468, "step": 516850 }, { "epoch": 0.9645928728548469, "grad_norm": 0.8643420338630676, "learning_rate": 0.00019113624491188595, "loss": 4.4208, "step": 516900 }, { "epoch": 0.9646861784142253, "grad_norm": 1.2459940910339355, "learning_rate": 0.00019113451568769505, "loss": 4.4535, "step": 516950 }, { "epoch": 0.9647794839736038, "grad_norm": 1.0978434085845947, "learning_rate": 0.00019113278630266772, "loss": 4.4128, "step": 517000 }, { "epoch": 0.9648727895329824, "grad_norm": 0.6621046662330627, "learning_rate": 0.0001911310567568071, "loss": 4.4595, "step": 517050 }, { "epoch": 0.9649660950923609, "grad_norm": 1.2009626626968384, "learning_rate": 0.00019112932705011616, "loss": 4.3229, "step": 517100 }, { "epoch": 0.9650594006517393, "grad_norm": 0.8467291593551636, "learning_rate": 0.00019112759718259805, "loss": 4.679, "step": 517150 }, { "epoch": 0.9651527062111178, "grad_norm": 1.3191983699798584, "learning_rate": 0.00019112586715425573, "loss": 4.6211, "step": 517200 }, { "epoch": 0.9652460117704963, "grad_norm": 1.162018060684204, "learning_rate": 0.00019112413696509227, "loss": 4.5509, "step": 517250 }, { "epoch": 0.9653393173298748, "grad_norm": 0.7355647683143616, "learning_rate": 0.00019112240661511078, "loss": 4.4843, "step": 517300 }, { "epoch": 0.9654326228892532, "grad_norm": 0.7065161466598511, "learning_rate": 0.00019112067610431424, "loss": 4.4872, "step": 517350 }, { "epoch": 0.9655259284486318, "grad_norm": 1.284676194190979, "learning_rate": 0.00019111894543270575, "loss": 4.4103, "step": 517400 }, { "epoch": 0.9656192340080103, "grad_norm": 1.0690058469772339, "learning_rate": 0.0001911172146002884, "loss": 4.6988, "step": 517450 }, { "epoch": 0.9657125395673888, "grad_norm": 1.0841615200042725, "learning_rate": 0.00019111548360706518, "loss": 4.4285, "step": 517500 }, { "epoch": 0.9658058451267673, "grad_norm": 1.3093359470367432, "learning_rate": 0.00019111375245303914, "loss": 4.6408, "step": 517550 }, { "epoch": 0.9658991506861457, "grad_norm": 1.1756441593170166, "learning_rate": 0.00019111202113821337, "loss": 4.5503, "step": 517600 }, { "epoch": 0.9659924562455242, "grad_norm": 1.0677144527435303, "learning_rate": 0.00019111028966259087, "loss": 4.5982, "step": 517650 }, { "epoch": 0.9660857618049027, "grad_norm": 1.2211114168167114, "learning_rate": 0.0001911085580261748, "loss": 4.4679, "step": 517700 }, { "epoch": 0.9661790673642813, "grad_norm": 1.4065744876861572, "learning_rate": 0.00019110682622896813, "loss": 4.4371, "step": 517750 }, { "epoch": 0.9662723729236598, "grad_norm": 0.9702979326248169, "learning_rate": 0.00019110509427097396, "loss": 4.5696, "step": 517800 }, { "epoch": 0.9663656784830382, "grad_norm": 1.0108451843261719, "learning_rate": 0.0001911033621521953, "loss": 4.5064, "step": 517850 }, { "epoch": 0.9664589840424167, "grad_norm": 1.2534197568893433, "learning_rate": 0.00019110162987263528, "loss": 4.8135, "step": 517900 }, { "epoch": 0.9665522896017952, "grad_norm": 1.1588776111602783, "learning_rate": 0.00019109989743229687, "loss": 4.7165, "step": 517950 }, { "epoch": 0.9666455951611737, "grad_norm": 0.7583629488945007, "learning_rate": 0.00019109816483118319, "loss": 4.6447, "step": 518000 }, { "epoch": 0.9667389007205521, "grad_norm": 0.9244435429573059, "learning_rate": 0.00019109643206929725, "loss": 4.6123, "step": 518050 }, { "epoch": 0.9668322062799307, "grad_norm": 1.2333121299743652, "learning_rate": 0.0001910946991466422, "loss": 4.5914, "step": 518100 }, { "epoch": 0.9669255118393092, "grad_norm": 1.1857706308364868, "learning_rate": 0.00019109296606322098, "loss": 4.6203, "step": 518150 }, { "epoch": 0.9670188173986877, "grad_norm": 1.2005411386489868, "learning_rate": 0.0001910912328190367, "loss": 4.5829, "step": 518200 }, { "epoch": 0.9671121229580661, "grad_norm": 1.2851636409759521, "learning_rate": 0.00019108949941409246, "loss": 4.624, "step": 518250 }, { "epoch": 0.9672054285174446, "grad_norm": 1.3155996799468994, "learning_rate": 0.00019108776584839126, "loss": 4.49, "step": 518300 }, { "epoch": 0.9672987340768231, "grad_norm": 1.2079349756240845, "learning_rate": 0.0001910860321219362, "loss": 4.408, "step": 518350 }, { "epoch": 0.9673920396362016, "grad_norm": 1.1507331132888794, "learning_rate": 0.0001910842982347303, "loss": 4.5914, "step": 518400 }, { "epoch": 0.9674853451955802, "grad_norm": 1.406695008277893, "learning_rate": 0.0001910825641867766, "loss": 4.6636, "step": 518450 }, { "epoch": 0.9675786507549586, "grad_norm": 0.7086337208747864, "learning_rate": 0.00019108082997807828, "loss": 4.3636, "step": 518500 }, { "epoch": 0.9676719563143371, "grad_norm": 1.0646345615386963, "learning_rate": 0.00019107909560863825, "loss": 4.417, "step": 518550 }, { "epoch": 0.9677652618737156, "grad_norm": 1.0562654733657837, "learning_rate": 0.0001910773610784597, "loss": 4.5462, "step": 518600 }, { "epoch": 0.9678585674330941, "grad_norm": 0.9988654851913452, "learning_rate": 0.00019107562638754562, "loss": 4.3818, "step": 518650 }, { "epoch": 0.9679518729924725, "grad_norm": 0.8831391930580139, "learning_rate": 0.00019107389153589908, "loss": 4.5092, "step": 518700 }, { "epoch": 0.968045178551851, "grad_norm": 1.177599549293518, "learning_rate": 0.00019107215652352312, "loss": 4.4949, "step": 518750 }, { "epoch": 0.9681384841112296, "grad_norm": 1.1170774698257446, "learning_rate": 0.00019107042135042086, "loss": 4.3362, "step": 518800 }, { "epoch": 0.9682317896706081, "grad_norm": 1.195797324180603, "learning_rate": 0.00019106868601659533, "loss": 4.5056, "step": 518850 }, { "epoch": 0.9683250952299866, "grad_norm": 1.0617939233779907, "learning_rate": 0.00019106695052204957, "loss": 4.461, "step": 518900 }, { "epoch": 0.968418400789365, "grad_norm": 1.143950343132019, "learning_rate": 0.0001910652148667867, "loss": 4.5851, "step": 518950 }, { "epoch": 0.9685117063487435, "grad_norm": 1.59396493434906, "learning_rate": 0.00019106347905080973, "loss": 4.5557, "step": 519000 }, { "epoch": 0.9685117063487435, "eval_loss": 4.7001214027404785, "eval_runtime": 228.5739, "eval_samples_per_second": 11.41, "eval_steps_per_second": 11.41, "eval_tts_loss": 7.622987998685188, "step": 519000 }, { "epoch": 0.968605011908122, "grad_norm": 1.0155067443847656, "learning_rate": 0.00019106174307412173, "loss": 4.6652, "step": 519050 }, { "epoch": 0.9686983174675005, "grad_norm": 1.2388640642166138, "learning_rate": 0.0001910600069367258, "loss": 4.616, "step": 519100 }, { "epoch": 0.968791623026879, "grad_norm": 1.195953130722046, "learning_rate": 0.00019105827063862495, "loss": 4.5419, "step": 519150 }, { "epoch": 0.9688849285862575, "grad_norm": 0.9257772564888, "learning_rate": 0.0001910565341798223, "loss": 4.5923, "step": 519200 }, { "epoch": 0.968978234145636, "grad_norm": 1.1395663022994995, "learning_rate": 0.00019105479756032086, "loss": 4.4718, "step": 519250 }, { "epoch": 0.9690715397050145, "grad_norm": 1.1108932495117188, "learning_rate": 0.00019105306078012377, "loss": 4.5627, "step": 519300 }, { "epoch": 0.969164845264393, "grad_norm": 1.2264920473098755, "learning_rate": 0.00019105132383923402, "loss": 4.7129, "step": 519350 }, { "epoch": 0.9692581508237714, "grad_norm": 1.001344084739685, "learning_rate": 0.0001910495867376547, "loss": 4.5912, "step": 519400 }, { "epoch": 0.9693514563831499, "grad_norm": 1.2849265336990356, "learning_rate": 0.0001910478494753889, "loss": 4.6034, "step": 519450 }, { "epoch": 0.9694447619425285, "grad_norm": 0.9491596817970276, "learning_rate": 0.00019104611205243965, "loss": 4.754, "step": 519500 }, { "epoch": 0.969538067501907, "grad_norm": 1.0722806453704834, "learning_rate": 0.00019104437446881004, "loss": 4.6485, "step": 519550 }, { "epoch": 0.9696313730612854, "grad_norm": 1.2749931812286377, "learning_rate": 0.00019104263672450314, "loss": 4.3542, "step": 519600 }, { "epoch": 0.9697246786206639, "grad_norm": 1.2010455131530762, "learning_rate": 0.000191040898819522, "loss": 4.7002, "step": 519650 }, { "epoch": 0.9698179841800424, "grad_norm": 1.2712091207504272, "learning_rate": 0.00019103916075386966, "loss": 4.5621, "step": 519700 }, { "epoch": 0.9699112897394209, "grad_norm": 1.0799859762191772, "learning_rate": 0.00019103742252754927, "loss": 4.6556, "step": 519750 }, { "epoch": 0.9700045952987993, "grad_norm": 1.2823902368545532, "learning_rate": 0.0001910356841405638, "loss": 4.598, "step": 519800 }, { "epoch": 0.9700979008581779, "grad_norm": 1.025672197341919, "learning_rate": 0.0001910339455929164, "loss": 4.4827, "step": 519850 }, { "epoch": 0.9701912064175564, "grad_norm": 1.4269847869873047, "learning_rate": 0.00019103220688461008, "loss": 4.5906, "step": 519900 }, { "epoch": 0.9702845119769349, "grad_norm": 1.6765121221542358, "learning_rate": 0.00019103046801564796, "loss": 4.57, "step": 519950 }, { "epoch": 0.9703778175363134, "grad_norm": 1.5583916902542114, "learning_rate": 0.00019102872898603307, "loss": 4.5254, "step": 520000 }, { "epoch": 0.9704711230956918, "grad_norm": 1.3943850994110107, "learning_rate": 0.00019102698979576847, "loss": 4.7866, "step": 520050 }, { "epoch": 0.9705644286550703, "grad_norm": 1.132477045059204, "learning_rate": 0.00019102525044485726, "loss": 4.4869, "step": 520100 }, { "epoch": 0.9706577342144488, "grad_norm": 0.8886978030204773, "learning_rate": 0.00019102351093330252, "loss": 4.5587, "step": 520150 }, { "epoch": 0.9707510397738274, "grad_norm": 0.7341157793998718, "learning_rate": 0.00019102177126110727, "loss": 4.6586, "step": 520200 }, { "epoch": 0.9708443453332058, "grad_norm": 1.2957510948181152, "learning_rate": 0.00019102003142827464, "loss": 4.6656, "step": 520250 }, { "epoch": 0.9709376508925843, "grad_norm": 1.2569067478179932, "learning_rate": 0.00019101829143480765, "loss": 4.4138, "step": 520300 }, { "epoch": 0.9710309564519628, "grad_norm": 1.1801834106445312, "learning_rate": 0.00019101655128070938, "loss": 4.558, "step": 520350 }, { "epoch": 0.9711242620113413, "grad_norm": 1.3396977186203003, "learning_rate": 0.00019101481096598294, "loss": 4.6046, "step": 520400 }, { "epoch": 0.9712175675707198, "grad_norm": 1.2308201789855957, "learning_rate": 0.00019101307049063134, "loss": 4.3835, "step": 520450 }, { "epoch": 0.9713108731300982, "grad_norm": 1.3665229082107544, "learning_rate": 0.00019101132985465774, "loss": 4.4739, "step": 520500 }, { "epoch": 0.9714041786894768, "grad_norm": 0.778360903263092, "learning_rate": 0.00019100958905806512, "loss": 4.6401, "step": 520550 }, { "epoch": 0.9714974842488553, "grad_norm": 1.2904661893844604, "learning_rate": 0.0001910078481008566, "loss": 4.5424, "step": 520600 }, { "epoch": 0.9715907898082338, "grad_norm": 1.339087963104248, "learning_rate": 0.00019100610698303522, "loss": 4.5129, "step": 520650 }, { "epoch": 0.9716840953676122, "grad_norm": 1.1398874521255493, "learning_rate": 0.00019100436570460407, "loss": 4.6786, "step": 520700 }, { "epoch": 0.9717774009269907, "grad_norm": 1.1007639169692993, "learning_rate": 0.00019100262426556625, "loss": 4.6878, "step": 520750 }, { "epoch": 0.9718707064863692, "grad_norm": 0.8794023394584656, "learning_rate": 0.00019100088266592481, "loss": 4.5158, "step": 520800 }, { "epoch": 0.9719640120457477, "grad_norm": 0.9615235924720764, "learning_rate": 0.00019099914090568284, "loss": 4.347, "step": 520850 }, { "epoch": 0.9720573176051263, "grad_norm": 1.2793523073196411, "learning_rate": 0.00019099739898484333, "loss": 4.2946, "step": 520900 }, { "epoch": 0.9721506231645047, "grad_norm": 1.1405634880065918, "learning_rate": 0.00019099565690340947, "loss": 4.623, "step": 520950 }, { "epoch": 0.9722439287238832, "grad_norm": 1.1924285888671875, "learning_rate": 0.0001909939146613843, "loss": 4.284, "step": 521000 }, { "epoch": 0.9723372342832617, "grad_norm": 1.3644477128982544, "learning_rate": 0.00019099217225877085, "loss": 4.4425, "step": 521050 }, { "epoch": 0.9724305398426402, "grad_norm": 0.867011547088623, "learning_rate": 0.00019099042969557227, "loss": 4.6923, "step": 521100 }, { "epoch": 0.9725238454020186, "grad_norm": 1.1317694187164307, "learning_rate": 0.00019098868697179158, "loss": 4.5613, "step": 521150 }, { "epoch": 0.9726171509613971, "grad_norm": 1.280672311782837, "learning_rate": 0.00019098694408743185, "loss": 4.6894, "step": 521200 }, { "epoch": 0.9727104565207757, "grad_norm": 1.0753579139709473, "learning_rate": 0.00019098520104249617, "loss": 4.6305, "step": 521250 }, { "epoch": 0.9728037620801542, "grad_norm": 1.0739895105361938, "learning_rate": 0.0001909834578369876, "loss": 4.4128, "step": 521300 }, { "epoch": 0.9728970676395327, "grad_norm": 1.0510057210922241, "learning_rate": 0.00019098171447090928, "loss": 4.5856, "step": 521350 }, { "epoch": 0.9729903731989111, "grad_norm": 1.1727142333984375, "learning_rate": 0.00019097997094426423, "loss": 4.5719, "step": 521400 }, { "epoch": 0.9730836787582896, "grad_norm": 1.5315583944320679, "learning_rate": 0.00019097822725705555, "loss": 4.4798, "step": 521450 }, { "epoch": 0.9731769843176681, "grad_norm": 1.0818549394607544, "learning_rate": 0.0001909764834092863, "loss": 4.5298, "step": 521500 }, { "epoch": 0.9732702898770466, "grad_norm": 1.3374048471450806, "learning_rate": 0.00019097473940095952, "loss": 4.5685, "step": 521550 }, { "epoch": 0.9733635954364251, "grad_norm": 1.1182202100753784, "learning_rate": 0.00019097299523207836, "loss": 4.4299, "step": 521600 }, { "epoch": 0.9734569009958036, "grad_norm": 1.2719510793685913, "learning_rate": 0.0001909712509026459, "loss": 4.4454, "step": 521650 }, { "epoch": 0.9735502065551821, "grad_norm": 1.1283625364303589, "learning_rate": 0.00019096950641266517, "loss": 4.5501, "step": 521700 }, { "epoch": 0.9736435121145606, "grad_norm": 1.059170126914978, "learning_rate": 0.00019096776176213926, "loss": 4.5142, "step": 521750 }, { "epoch": 0.973736817673939, "grad_norm": 1.1073553562164307, "learning_rate": 0.00019096601695107126, "loss": 4.4595, "step": 521800 }, { "epoch": 0.9738301232333175, "grad_norm": 0.8375008702278137, "learning_rate": 0.00019096427197946424, "loss": 4.5453, "step": 521850 }, { "epoch": 0.973923428792696, "grad_norm": 0.9027486443519592, "learning_rate": 0.0001909625268473213, "loss": 4.6543, "step": 521900 }, { "epoch": 0.9740167343520746, "grad_norm": 0.9850625395774841, "learning_rate": 0.0001909607815546455, "loss": 4.5029, "step": 521950 }, { "epoch": 0.9741100399114531, "grad_norm": 0.9993777275085449, "learning_rate": 0.00019095903610143994, "loss": 4.6469, "step": 522000 }, { "epoch": 0.9741100399114531, "eval_loss": 4.7097978591918945, "eval_runtime": 230.53, "eval_samples_per_second": 11.313, "eval_steps_per_second": 11.313, "eval_tts_loss": 7.578122001564119, "step": 522000 }, { "epoch": 0.9742033454708315, "grad_norm": 1.300405502319336, "learning_rate": 0.00019095729048770765, "loss": 4.583, "step": 522050 }, { "epoch": 0.97429665103021, "grad_norm": 0.9611183404922485, "learning_rate": 0.00019095554471345176, "loss": 4.4744, "step": 522100 }, { "epoch": 0.9743899565895885, "grad_norm": 1.297658920288086, "learning_rate": 0.00019095379877867536, "loss": 4.5024, "step": 522150 }, { "epoch": 0.974483262148967, "grad_norm": 1.0517747402191162, "learning_rate": 0.0001909520526833815, "loss": 4.4372, "step": 522200 }, { "epoch": 0.9745765677083454, "grad_norm": 1.0412423610687256, "learning_rate": 0.0001909503064275733, "loss": 4.6539, "step": 522250 }, { "epoch": 0.974669873267724, "grad_norm": 1.1834392547607422, "learning_rate": 0.00019094856001125373, "loss": 4.8113, "step": 522300 }, { "epoch": 0.9747631788271025, "grad_norm": 1.0815627574920654, "learning_rate": 0.00019094681343442602, "loss": 4.533, "step": 522350 }, { "epoch": 0.974856484386481, "grad_norm": 1.2627404928207397, "learning_rate": 0.00019094506669709314, "loss": 4.5659, "step": 522400 }, { "epoch": 0.9749497899458595, "grad_norm": 1.175206184387207, "learning_rate": 0.0001909433197992583, "loss": 4.514, "step": 522450 }, { "epoch": 0.9750430955052379, "grad_norm": 1.1156954765319824, "learning_rate": 0.0001909415727409244, "loss": 4.7133, "step": 522500 }, { "epoch": 0.9751364010646164, "grad_norm": 1.33116614818573, "learning_rate": 0.00019093982552209472, "loss": 4.5982, "step": 522550 }, { "epoch": 0.9752297066239949, "grad_norm": 1.1394720077514648, "learning_rate": 0.00019093807814277218, "loss": 4.5595, "step": 522600 }, { "epoch": 0.9753230121833734, "grad_norm": 1.0634437799453735, "learning_rate": 0.00019093633060295994, "loss": 4.5115, "step": 522650 }, { "epoch": 0.975416317742752, "grad_norm": 1.186028003692627, "learning_rate": 0.00019093458290266111, "loss": 4.5716, "step": 522700 }, { "epoch": 0.9755096233021304, "grad_norm": 1.4477113485336304, "learning_rate": 0.00019093283504187874, "loss": 4.6151, "step": 522750 }, { "epoch": 0.9756029288615089, "grad_norm": 1.169278860092163, "learning_rate": 0.0001909310870206159, "loss": 4.6071, "step": 522800 }, { "epoch": 0.9756962344208874, "grad_norm": 1.1335707902908325, "learning_rate": 0.0001909293388388757, "loss": 4.5348, "step": 522850 }, { "epoch": 0.9757895399802659, "grad_norm": 1.218106746673584, "learning_rate": 0.0001909275904966612, "loss": 4.4443, "step": 522900 }, { "epoch": 0.9758828455396443, "grad_norm": 1.0890352725982666, "learning_rate": 0.0001909258419939755, "loss": 4.5643, "step": 522950 }, { "epoch": 0.9759761510990228, "grad_norm": 1.2253206968307495, "learning_rate": 0.0001909240933308217, "loss": 4.5474, "step": 523000 }, { "epoch": 0.9760694566584014, "grad_norm": 0.8894392251968384, "learning_rate": 0.0001909223445072029, "loss": 4.3845, "step": 523050 }, { "epoch": 0.9761627622177799, "grad_norm": 1.2921968698501587, "learning_rate": 0.00019092059552312213, "loss": 4.6033, "step": 523100 }, { "epoch": 0.9762560677771583, "grad_norm": 1.240365743637085, "learning_rate": 0.00019091884637858253, "loss": 4.3429, "step": 523150 }, { "epoch": 0.9763493733365368, "grad_norm": 1.24266517162323, "learning_rate": 0.00019091709707358713, "loss": 4.4915, "step": 523200 }, { "epoch": 0.9764426788959153, "grad_norm": 1.0493535995483398, "learning_rate": 0.00019091534760813907, "loss": 4.516, "step": 523250 }, { "epoch": 0.9765359844552938, "grad_norm": 0.922390341758728, "learning_rate": 0.0001909135979822414, "loss": 4.5715, "step": 523300 }, { "epoch": 0.9766292900146722, "grad_norm": 1.4100518226623535, "learning_rate": 0.00019091184819589727, "loss": 4.5929, "step": 523350 }, { "epoch": 0.9767225955740508, "grad_norm": 0.7337448000907898, "learning_rate": 0.0001909100982491097, "loss": 4.6677, "step": 523400 }, { "epoch": 0.9768159011334293, "grad_norm": 1.5000475645065308, "learning_rate": 0.00019090834814188177, "loss": 4.2484, "step": 523450 }, { "epoch": 0.9769092066928078, "grad_norm": 0.9067786335945129, "learning_rate": 0.00019090659787421665, "loss": 4.645, "step": 523500 }, { "epoch": 0.9770025122521863, "grad_norm": 1.1780048608779907, "learning_rate": 0.00019090484744611734, "loss": 4.5151, "step": 523550 }, { "epoch": 0.9770958178115647, "grad_norm": 0.8858790993690491, "learning_rate": 0.000190903096857587, "loss": 4.5761, "step": 523600 }, { "epoch": 0.9771891233709432, "grad_norm": 1.1873397827148438, "learning_rate": 0.00019090134610862868, "loss": 4.552, "step": 523650 }, { "epoch": 0.9772824289303217, "grad_norm": 0.9411844611167908, "learning_rate": 0.00019089959519924548, "loss": 4.545, "step": 523700 }, { "epoch": 0.9773757344897003, "grad_norm": 0.8996508717536926, "learning_rate": 0.00019089784412944049, "loss": 4.6331, "step": 523750 }, { "epoch": 0.9774690400490788, "grad_norm": 1.1739059686660767, "learning_rate": 0.0001908960928992168, "loss": 4.4864, "step": 523800 }, { "epoch": 0.9775623456084572, "grad_norm": 1.0767582654953003, "learning_rate": 0.0001908943415085775, "loss": 4.6048, "step": 523850 }, { "epoch": 0.9776556511678357, "grad_norm": 1.2113652229309082, "learning_rate": 0.00019089258995752566, "loss": 4.5867, "step": 523900 }, { "epoch": 0.9777489567272142, "grad_norm": 0.9723443984985352, "learning_rate": 0.00019089083824606442, "loss": 4.3316, "step": 523950 }, { "epoch": 0.9778422622865927, "grad_norm": 1.060306429862976, "learning_rate": 0.00019088908637419682, "loss": 4.4673, "step": 524000 }, { "epoch": 0.9779355678459711, "grad_norm": 1.0762046575546265, "learning_rate": 0.00019088733434192602, "loss": 4.6635, "step": 524050 }, { "epoch": 0.9780288734053497, "grad_norm": 1.2389363050460815, "learning_rate": 0.000190885582149255, "loss": 4.3978, "step": 524100 }, { "epoch": 0.9781221789647282, "grad_norm": 1.1130664348602295, "learning_rate": 0.00019088382979618696, "loss": 4.5513, "step": 524150 }, { "epoch": 0.9782154845241067, "grad_norm": 1.1711992025375366, "learning_rate": 0.00019088207728272494, "loss": 4.6135, "step": 524200 }, { "epoch": 0.9783087900834851, "grad_norm": 1.3873505592346191, "learning_rate": 0.00019088032460887205, "loss": 4.5133, "step": 524250 }, { "epoch": 0.9784020956428636, "grad_norm": 1.0164330005645752, "learning_rate": 0.00019087857177463137, "loss": 4.3928, "step": 524300 }, { "epoch": 0.9784954012022421, "grad_norm": 0.922845184803009, "learning_rate": 0.00019087681878000601, "loss": 4.4376, "step": 524350 }, { "epoch": 0.9785887067616206, "grad_norm": 0.9557695984840393, "learning_rate": 0.00019087506562499905, "loss": 4.4592, "step": 524400 }, { "epoch": 0.9786820123209992, "grad_norm": 1.2097173929214478, "learning_rate": 0.0001908733123096136, "loss": 4.7187, "step": 524450 }, { "epoch": 0.9787753178803776, "grad_norm": 1.1994881629943848, "learning_rate": 0.00019087155883385274, "loss": 4.5767, "step": 524500 }, { "epoch": 0.9788686234397561, "grad_norm": 1.1564199924468994, "learning_rate": 0.00019086980519771957, "loss": 4.4908, "step": 524550 }, { "epoch": 0.9789619289991346, "grad_norm": 1.0967894792556763, "learning_rate": 0.00019086805140121715, "loss": 4.4639, "step": 524600 }, { "epoch": 0.9790552345585131, "grad_norm": 1.059061050415039, "learning_rate": 0.00019086629744434865, "loss": 4.5158, "step": 524650 }, { "epoch": 0.9791485401178915, "grad_norm": 0.7575185894966125, "learning_rate": 0.0001908645433271171, "loss": 4.5341, "step": 524700 }, { "epoch": 0.97924184567727, "grad_norm": 1.2447662353515625, "learning_rate": 0.00019086278904952565, "loss": 4.546, "step": 524750 }, { "epoch": 0.9793351512366486, "grad_norm": 1.3063735961914062, "learning_rate": 0.00019086103461157734, "loss": 4.4303, "step": 524800 }, { "epoch": 0.9794284567960271, "grad_norm": 0.9031000137329102, "learning_rate": 0.0001908592800132753, "loss": 4.5639, "step": 524850 }, { "epoch": 0.9795217623554056, "grad_norm": 1.2312612533569336, "learning_rate": 0.0001908575252546226, "loss": 4.3601, "step": 524900 }, { "epoch": 0.979615067914784, "grad_norm": 0.9371641874313354, "learning_rate": 0.00019085577033562236, "loss": 4.4598, "step": 524950 }, { "epoch": 0.9797083734741625, "grad_norm": 1.0243632793426514, "learning_rate": 0.00019085401525627767, "loss": 4.7131, "step": 525000 }, { "epoch": 0.9797083734741625, "eval_loss": 4.694164276123047, "eval_runtime": 230.3774, "eval_samples_per_second": 11.321, "eval_steps_per_second": 11.321, "eval_tts_loss": 7.6419614720246845, "step": 525000 }, { "epoch": 0.979801679033541, "grad_norm": 0.9438709616661072, "learning_rate": 0.00019085226001659165, "loss": 4.7328, "step": 525050 }, { "epoch": 0.9798949845929195, "grad_norm": 1.038021445274353, "learning_rate": 0.00019085050461656734, "loss": 4.4494, "step": 525100 }, { "epoch": 0.979988290152298, "grad_norm": 0.7887255549430847, "learning_rate": 0.00019084874905620793, "loss": 4.6494, "step": 525150 }, { "epoch": 0.9800815957116765, "grad_norm": 0.9669321179389954, "learning_rate": 0.00019084699333551643, "loss": 4.6906, "step": 525200 }, { "epoch": 0.980174901271055, "grad_norm": 1.0607885122299194, "learning_rate": 0.00019084523745449596, "loss": 4.4508, "step": 525250 }, { "epoch": 0.9802682068304335, "grad_norm": 0.9444970488548279, "learning_rate": 0.00019084348141314966, "loss": 4.6233, "step": 525300 }, { "epoch": 0.980361512389812, "grad_norm": 1.1765512228012085, "learning_rate": 0.0001908417252114806, "loss": 4.716, "step": 525350 }, { "epoch": 0.9804548179491904, "grad_norm": 1.0271841287612915, "learning_rate": 0.00019083996884949187, "loss": 4.5474, "step": 525400 }, { "epoch": 0.9805481235085689, "grad_norm": 1.081523060798645, "learning_rate": 0.00019083821232718655, "loss": 4.5634, "step": 525450 }, { "epoch": 0.9806414290679475, "grad_norm": 1.1915758848190308, "learning_rate": 0.0001908364556445678, "loss": 4.3078, "step": 525500 }, { "epoch": 0.980734734627326, "grad_norm": 1.2185860872268677, "learning_rate": 0.0001908346988016387, "loss": 4.4354, "step": 525550 }, { "epoch": 0.9808280401867044, "grad_norm": 1.02804434299469, "learning_rate": 0.0001908329417984023, "loss": 4.4834, "step": 525600 }, { "epoch": 0.9809213457460829, "grad_norm": 1.025506854057312, "learning_rate": 0.00019083118463486176, "loss": 4.4818, "step": 525650 }, { "epoch": 0.9810146513054614, "grad_norm": 1.221799373626709, "learning_rate": 0.00019082942731102017, "loss": 4.4144, "step": 525700 }, { "epoch": 0.9811079568648399, "grad_norm": 0.9400854110717773, "learning_rate": 0.00019082766982688061, "loss": 4.4667, "step": 525750 }, { "epoch": 0.9812012624242183, "grad_norm": 1.1538151502609253, "learning_rate": 0.00019082591218244619, "loss": 4.4497, "step": 525800 }, { "epoch": 0.9812945679835969, "grad_norm": 1.0266728401184082, "learning_rate": 0.00019082415437772003, "loss": 4.4458, "step": 525850 }, { "epoch": 0.9813878735429754, "grad_norm": 1.5398420095443726, "learning_rate": 0.0001908223964127052, "loss": 4.5238, "step": 525900 }, { "epoch": 0.9814811791023539, "grad_norm": 1.1759055852890015, "learning_rate": 0.00019082063828740483, "loss": 4.6566, "step": 525950 }, { "epoch": 0.9815744846617324, "grad_norm": 1.3355110883712769, "learning_rate": 0.00019081888000182205, "loss": 4.5036, "step": 526000 }, { "epoch": 0.9816677902211108, "grad_norm": 0.9322126507759094, "learning_rate": 0.0001908171215559599, "loss": 4.4897, "step": 526050 }, { "epoch": 0.9817610957804893, "grad_norm": 1.1785377264022827, "learning_rate": 0.0001908153629498215, "loss": 4.36, "step": 526100 }, { "epoch": 0.9818544013398678, "grad_norm": 0.8922885656356812, "learning_rate": 0.00019081360418340997, "loss": 4.6905, "step": 526150 }, { "epoch": 0.9819477068992464, "grad_norm": 1.1009005308151245, "learning_rate": 0.0001908118452567284, "loss": 4.5811, "step": 526200 }, { "epoch": 0.9820410124586249, "grad_norm": 0.8094876408576965, "learning_rate": 0.00019081008616977991, "loss": 4.4383, "step": 526250 }, { "epoch": 0.9821343180180033, "grad_norm": 1.6933599710464478, "learning_rate": 0.0001908083269225676, "loss": 4.4826, "step": 526300 }, { "epoch": 0.9822276235773818, "grad_norm": 1.1940139532089233, "learning_rate": 0.00019080656751509456, "loss": 4.5107, "step": 526350 }, { "epoch": 0.9823209291367603, "grad_norm": 1.3585516214370728, "learning_rate": 0.00019080480794736392, "loss": 4.456, "step": 526400 }, { "epoch": 0.9824142346961388, "grad_norm": 1.0569653511047363, "learning_rate": 0.00019080304821937876, "loss": 4.5069, "step": 526450 }, { "epoch": 0.9825075402555172, "grad_norm": 0.8311865925788879, "learning_rate": 0.0001908012883311422, "loss": 4.5033, "step": 526500 }, { "epoch": 0.9826008458148958, "grad_norm": 1.108024001121521, "learning_rate": 0.00019079952828265735, "loss": 4.5833, "step": 526550 }, { "epoch": 0.9826941513742743, "grad_norm": 1.1564241647720337, "learning_rate": 0.00019079776807392728, "loss": 4.5738, "step": 526600 }, { "epoch": 0.9827874569336528, "grad_norm": 1.261348843574524, "learning_rate": 0.00019079600770495518, "loss": 4.6863, "step": 526650 }, { "epoch": 0.9828807624930312, "grad_norm": 1.1307748556137085, "learning_rate": 0.00019079424717574405, "loss": 4.5751, "step": 526700 }, { "epoch": 0.9829740680524097, "grad_norm": 1.1750500202178955, "learning_rate": 0.0001907924864862971, "loss": 4.6139, "step": 526750 }, { "epoch": 0.9830673736117882, "grad_norm": 1.1854808330535889, "learning_rate": 0.00019079072563661734, "loss": 4.706, "step": 526800 }, { "epoch": 0.9831606791711667, "grad_norm": 0.9059605598449707, "learning_rate": 0.00019078896462670794, "loss": 4.4532, "step": 526850 }, { "epoch": 0.9832539847305453, "grad_norm": 1.2157238721847534, "learning_rate": 0.000190787203456572, "loss": 4.5562, "step": 526900 }, { "epoch": 0.9833472902899237, "grad_norm": 1.356483817100525, "learning_rate": 0.0001907854421262126, "loss": 4.5016, "step": 526950 }, { "epoch": 0.9834405958493022, "grad_norm": 1.2194933891296387, "learning_rate": 0.0001907836806356329, "loss": 4.6496, "step": 527000 }, { "epoch": 0.9835339014086807, "grad_norm": 1.0702190399169922, "learning_rate": 0.00019078191898483598, "loss": 4.6524, "step": 527050 }, { "epoch": 0.9836272069680592, "grad_norm": 1.1717504262924194, "learning_rate": 0.0001907801571738249, "loss": 4.5169, "step": 527100 }, { "epoch": 0.9837205125274376, "grad_norm": 1.2796210050582886, "learning_rate": 0.00019077839520260286, "loss": 4.5833, "step": 527150 }, { "epoch": 0.9838138180868161, "grad_norm": 1.2936681509017944, "learning_rate": 0.00019077663307117294, "loss": 4.4274, "step": 527200 }, { "epoch": 0.9839071236461947, "grad_norm": 1.2062686681747437, "learning_rate": 0.00019077487077953818, "loss": 4.5781, "step": 527250 }, { "epoch": 0.9840004292055732, "grad_norm": 1.1827361583709717, "learning_rate": 0.00019077310832770177, "loss": 4.5158, "step": 527300 }, { "epoch": 0.9840937347649517, "grad_norm": 1.073013186454773, "learning_rate": 0.0001907713457156668, "loss": 4.4073, "step": 527350 }, { "epoch": 0.9841870403243301, "grad_norm": 1.0383636951446533, "learning_rate": 0.00019076958294343636, "loss": 4.444, "step": 527400 }, { "epoch": 0.9842803458837086, "grad_norm": 1.004257082939148, "learning_rate": 0.0001907678200110136, "loss": 4.7476, "step": 527450 }, { "epoch": 0.9843736514430871, "grad_norm": 1.1027897596359253, "learning_rate": 0.0001907660569184016, "loss": 4.3951, "step": 527500 }, { "epoch": 0.9844669570024656, "grad_norm": 1.162660837173462, "learning_rate": 0.00019076429366560348, "loss": 4.6682, "step": 527550 }, { "epoch": 0.9845602625618441, "grad_norm": 1.189071536064148, "learning_rate": 0.00019076253025262236, "loss": 4.4873, "step": 527600 }, { "epoch": 0.9846535681212226, "grad_norm": 1.4174994230270386, "learning_rate": 0.00019076076667946134, "loss": 4.3936, "step": 527650 }, { "epoch": 0.9847468736806011, "grad_norm": 0.9970645308494568, "learning_rate": 0.00019075900294612352, "loss": 4.5345, "step": 527700 }, { "epoch": 0.9848401792399796, "grad_norm": 0.7957653999328613, "learning_rate": 0.00019075723905261205, "loss": 4.4642, "step": 527750 }, { "epoch": 0.984933484799358, "grad_norm": 1.2149927616119385, "learning_rate": 0.00019075547499893, "loss": 4.3071, "step": 527800 }, { "epoch": 0.9850267903587365, "grad_norm": 1.3680000305175781, "learning_rate": 0.00019075371078508053, "loss": 4.4464, "step": 527850 }, { "epoch": 0.985120095918115, "grad_norm": 1.4044183492660522, "learning_rate": 0.0001907519464110667, "loss": 4.4879, "step": 527900 }, { "epoch": 0.9852134014774935, "grad_norm": 1.0351190567016602, "learning_rate": 0.0001907501818768917, "loss": 4.8115, "step": 527950 }, { "epoch": 0.9853067070368721, "grad_norm": 1.2702745199203491, "learning_rate": 0.00019074841718255857, "loss": 4.4789, "step": 528000 }, { "epoch": 0.9853067070368721, "eval_loss": 4.701737880706787, "eval_runtime": 231.8581, "eval_samples_per_second": 11.248, "eval_steps_per_second": 11.248, "eval_tts_loss": 7.641061514590302, "step": 528000 }, { "epoch": 0.9854000125962505, "grad_norm": 1.175429105758667, "learning_rate": 0.00019074665232807043, "loss": 4.5269, "step": 528050 }, { "epoch": 0.985493318155629, "grad_norm": 1.2303099632263184, "learning_rate": 0.00019074488731343046, "loss": 4.3899, "step": 528100 }, { "epoch": 0.9855866237150075, "grad_norm": 1.229248285293579, "learning_rate": 0.0001907431221386417, "loss": 4.3248, "step": 528150 }, { "epoch": 0.985679929274386, "grad_norm": 0.8154194355010986, "learning_rate": 0.00019074135680370732, "loss": 4.5129, "step": 528200 }, { "epoch": 0.9857732348337644, "grad_norm": 1.0199432373046875, "learning_rate": 0.00019073959130863036, "loss": 4.7699, "step": 528250 }, { "epoch": 0.9858665403931429, "grad_norm": 1.0072429180145264, "learning_rate": 0.00019073782565341403, "loss": 4.5529, "step": 528300 }, { "epoch": 0.9859598459525215, "grad_norm": 1.3674672842025757, "learning_rate": 0.0001907360598380614, "loss": 4.5244, "step": 528350 }, { "epoch": 0.9860531515119, "grad_norm": 0.9725618362426758, "learning_rate": 0.00019073429386257557, "loss": 4.6635, "step": 528400 }, { "epoch": 0.9861464570712785, "grad_norm": 0.9817070364952087, "learning_rate": 0.0001907325277269597, "loss": 4.3839, "step": 528450 }, { "epoch": 0.9862397626306569, "grad_norm": 1.056870937347412, "learning_rate": 0.0001907307614312169, "loss": 4.5474, "step": 528500 }, { "epoch": 0.9863330681900354, "grad_norm": 1.2692755460739136, "learning_rate": 0.00019072899497535022, "loss": 4.6293, "step": 528550 }, { "epoch": 0.9864263737494139, "grad_norm": 1.1193509101867676, "learning_rate": 0.00019072722835936285, "loss": 4.428, "step": 528600 }, { "epoch": 0.9865196793087924, "grad_norm": 0.9396880865097046, "learning_rate": 0.0001907254615832579, "loss": 4.4101, "step": 528650 }, { "epoch": 0.986612984868171, "grad_norm": 1.2640496492385864, "learning_rate": 0.00019072369464703847, "loss": 4.5506, "step": 528700 }, { "epoch": 0.9867062904275494, "grad_norm": 1.1889392137527466, "learning_rate": 0.0001907219275507077, "loss": 4.4693, "step": 528750 }, { "epoch": 0.9867995959869279, "grad_norm": 1.404665470123291, "learning_rate": 0.00019072016029426865, "loss": 4.5982, "step": 528800 }, { "epoch": 0.9868929015463064, "grad_norm": 1.02499520778656, "learning_rate": 0.00019071839287772448, "loss": 4.5424, "step": 528850 }, { "epoch": 0.9869862071056849, "grad_norm": 1.14762544631958, "learning_rate": 0.0001907166253010783, "loss": 4.5078, "step": 528900 }, { "epoch": 0.9870795126650633, "grad_norm": 1.1336801052093506, "learning_rate": 0.0001907148575643333, "loss": 4.4453, "step": 528950 }, { "epoch": 0.9871728182244418, "grad_norm": 1.0139964818954468, "learning_rate": 0.0001907130896674925, "loss": 4.6717, "step": 529000 }, { "epoch": 0.9872661237838204, "grad_norm": 1.1291979551315308, "learning_rate": 0.00019071132161055908, "loss": 4.6179, "step": 529050 }, { "epoch": 0.9873594293431989, "grad_norm": 1.1060199737548828, "learning_rate": 0.0001907095533935361, "loss": 4.5175, "step": 529100 }, { "epoch": 0.9874527349025773, "grad_norm": 1.0623533725738525, "learning_rate": 0.00019070778501642676, "loss": 4.7316, "step": 529150 }, { "epoch": 0.9875460404619558, "grad_norm": 1.0466336011886597, "learning_rate": 0.00019070601647923411, "loss": 4.6024, "step": 529200 }, { "epoch": 0.9876393460213343, "grad_norm": 1.2481396198272705, "learning_rate": 0.0001907042477819613, "loss": 4.5658, "step": 529250 }, { "epoch": 0.9877326515807128, "grad_norm": 1.0156023502349854, "learning_rate": 0.00019070247892461145, "loss": 4.3863, "step": 529300 }, { "epoch": 0.9878259571400912, "grad_norm": 1.1544495820999146, "learning_rate": 0.00019070070990718772, "loss": 4.4979, "step": 529350 }, { "epoch": 0.9879192626994698, "grad_norm": 1.5071150064468384, "learning_rate": 0.00019069894072969317, "loss": 4.6878, "step": 529400 }, { "epoch": 0.9880125682588483, "grad_norm": 1.299371600151062, "learning_rate": 0.00019069717139213094, "loss": 4.5968, "step": 529450 }, { "epoch": 0.9881058738182268, "grad_norm": 1.0916576385498047, "learning_rate": 0.0001906954018945042, "loss": 4.5247, "step": 529500 }, { "epoch": 0.9881991793776053, "grad_norm": 0.9396972060203552, "learning_rate": 0.00019069363223681599, "loss": 4.5981, "step": 529550 }, { "epoch": 0.9882924849369837, "grad_norm": 1.1531301736831665, "learning_rate": 0.0001906918624190695, "loss": 4.6371, "step": 529600 }, { "epoch": 0.9883857904963622, "grad_norm": 1.1413004398345947, "learning_rate": 0.00019069009244126784, "loss": 4.608, "step": 529650 }, { "epoch": 0.9884790960557407, "grad_norm": 1.3086955547332764, "learning_rate": 0.0001906883223034141, "loss": 4.6226, "step": 529700 }, { "epoch": 0.9885724016151193, "grad_norm": 1.0513914823532104, "learning_rate": 0.00019068655200551142, "loss": 4.4234, "step": 529750 }, { "epoch": 0.9886657071744978, "grad_norm": 1.0332573652267456, "learning_rate": 0.00019068478154756296, "loss": 4.6262, "step": 529800 }, { "epoch": 0.9887590127338762, "grad_norm": 1.0589451789855957, "learning_rate": 0.0001906830109295718, "loss": 4.6415, "step": 529850 }, { "epoch": 0.9888523182932547, "grad_norm": 1.223704218864441, "learning_rate": 0.0001906812401515411, "loss": 4.4208, "step": 529900 }, { "epoch": 0.9889456238526332, "grad_norm": 1.0677813291549683, "learning_rate": 0.00019067946921347395, "loss": 4.6721, "step": 529950 }, { "epoch": 0.9890389294120117, "grad_norm": 0.8295161724090576, "learning_rate": 0.0001906776981153735, "loss": 4.6259, "step": 530000 }, { "epoch": 0.9891322349713901, "grad_norm": 0.9274109601974487, "learning_rate": 0.00019067592685724285, "loss": 4.7403, "step": 530050 }, { "epoch": 0.9892255405307687, "grad_norm": 1.0937379598617554, "learning_rate": 0.00019067415543908518, "loss": 4.5758, "step": 530100 }, { "epoch": 0.9893188460901472, "grad_norm": 0.9244849681854248, "learning_rate": 0.00019067238386090356, "loss": 4.3224, "step": 530150 }, { "epoch": 0.9894121516495257, "grad_norm": 0.9020528793334961, "learning_rate": 0.00019067061212270115, "loss": 4.4759, "step": 530200 }, { "epoch": 0.9895054572089041, "grad_norm": 1.0903435945510864, "learning_rate": 0.00019066884022448104, "loss": 4.4313, "step": 530250 }, { "epoch": 0.9895987627682826, "grad_norm": 1.2521004676818848, "learning_rate": 0.00019066706816624638, "loss": 4.6391, "step": 530300 }, { "epoch": 0.9896920683276611, "grad_norm": 1.0018556118011475, "learning_rate": 0.0001906652959480003, "loss": 4.4181, "step": 530350 }, { "epoch": 0.9897853738870396, "grad_norm": 1.1250208616256714, "learning_rate": 0.00019066352356974595, "loss": 4.6277, "step": 530400 }, { "epoch": 0.9898786794464182, "grad_norm": 1.228994369506836, "learning_rate": 0.00019066175103148641, "loss": 4.5979, "step": 530450 }, { "epoch": 0.9899719850057966, "grad_norm": 1.1446768045425415, "learning_rate": 0.00019065997833322485, "loss": 4.4695, "step": 530500 }, { "epoch": 0.9900652905651751, "grad_norm": 1.1885581016540527, "learning_rate": 0.00019065820547496437, "loss": 4.4892, "step": 530550 }, { "epoch": 0.9901585961245536, "grad_norm": 0.977424144744873, "learning_rate": 0.0001906564324567081, "loss": 4.622, "step": 530600 }, { "epoch": 0.9902519016839321, "grad_norm": 1.01764976978302, "learning_rate": 0.00019065465927845917, "loss": 4.52, "step": 530650 }, { "epoch": 0.9903452072433105, "grad_norm": 1.1252986192703247, "learning_rate": 0.00019065288594022074, "loss": 4.3796, "step": 530700 }, { "epoch": 0.990438512802689, "grad_norm": 1.1534461975097656, "learning_rate": 0.00019065111244199594, "loss": 4.3811, "step": 530750 }, { "epoch": 0.9905318183620676, "grad_norm": 0.98378986120224, "learning_rate": 0.00019064933878378784, "loss": 4.6197, "step": 530800 }, { "epoch": 0.9906251239214461, "grad_norm": 1.5495555400848389, "learning_rate": 0.00019064756496559962, "loss": 4.8127, "step": 530850 }, { "epoch": 0.9907184294808246, "grad_norm": 1.0124245882034302, "learning_rate": 0.0001906457909874344, "loss": 4.5536, "step": 530900 }, { "epoch": 0.990811735040203, "grad_norm": 0.8453911542892456, "learning_rate": 0.0001906440168492953, "loss": 4.5716, "step": 530950 }, { "epoch": 0.9909050405995815, "grad_norm": 1.2097727060317993, "learning_rate": 0.00019064224255118544, "loss": 4.5778, "step": 531000 }, { "epoch": 0.9909050405995815, "eval_loss": 4.699122905731201, "eval_runtime": 231.7485, "eval_samples_per_second": 11.254, "eval_steps_per_second": 11.254, "eval_tts_loss": 7.6238937171853145, "step": 531000 }, { "epoch": 0.99099834615896, "grad_norm": 1.0384348630905151, "learning_rate": 0.00019064046809310803, "loss": 4.5944, "step": 531050 }, { "epoch": 0.9910916517183385, "grad_norm": 1.4319239854812622, "learning_rate": 0.00019063869347506608, "loss": 4.6203, "step": 531100 }, { "epoch": 0.991184957277717, "grad_norm": 1.4958970546722412, "learning_rate": 0.0001906369186970628, "loss": 4.5215, "step": 531150 }, { "epoch": 0.9912782628370955, "grad_norm": 1.0489661693572998, "learning_rate": 0.00019063514375910134, "loss": 4.6591, "step": 531200 }, { "epoch": 0.991371568396474, "grad_norm": 0.996589720249176, "learning_rate": 0.00019063336866118478, "loss": 4.6233, "step": 531250 }, { "epoch": 0.9914648739558525, "grad_norm": 1.0587917566299438, "learning_rate": 0.00019063159340331626, "loss": 4.4618, "step": 531300 }, { "epoch": 0.991558179515231, "grad_norm": 0.8412649035453796, "learning_rate": 0.00019062981798549894, "loss": 4.4867, "step": 531350 }, { "epoch": 0.9916514850746094, "grad_norm": 0.9184856414794922, "learning_rate": 0.00019062804240773594, "loss": 4.5567, "step": 531400 }, { "epoch": 0.9917447906339879, "grad_norm": 1.1070401668548584, "learning_rate": 0.00019062626667003038, "loss": 4.5658, "step": 531450 }, { "epoch": 0.9918380961933665, "grad_norm": 1.1344000101089478, "learning_rate": 0.0001906244907723854, "loss": 4.5737, "step": 531500 }, { "epoch": 0.991931401752745, "grad_norm": 0.8413671255111694, "learning_rate": 0.00019062271471480415, "loss": 4.7232, "step": 531550 }, { "epoch": 0.9920247073121234, "grad_norm": 1.0313146114349365, "learning_rate": 0.00019062093849728974, "loss": 4.6848, "step": 531600 }, { "epoch": 0.9921180128715019, "grad_norm": 1.0481282472610474, "learning_rate": 0.00019061916211984535, "loss": 4.6459, "step": 531650 }, { "epoch": 0.9922113184308804, "grad_norm": 1.2196465730667114, "learning_rate": 0.00019061738558247407, "loss": 4.6273, "step": 531700 }, { "epoch": 0.9923046239902589, "grad_norm": 1.1619541645050049, "learning_rate": 0.00019061560888517903, "loss": 4.633, "step": 531750 }, { "epoch": 0.9923979295496373, "grad_norm": 1.1147663593292236, "learning_rate": 0.00019061383202796339, "loss": 4.635, "step": 531800 }, { "epoch": 0.9924912351090159, "grad_norm": 1.0607409477233887, "learning_rate": 0.0001906120550108303, "loss": 4.5172, "step": 531850 }, { "epoch": 0.9925845406683944, "grad_norm": 1.0881463289260864, "learning_rate": 0.00019061027783378286, "loss": 4.6757, "step": 531900 }, { "epoch": 0.9926778462277729, "grad_norm": 1.1087841987609863, "learning_rate": 0.00019060850049682423, "loss": 4.5579, "step": 531950 }, { "epoch": 0.9927711517871514, "grad_norm": 1.0826126337051392, "learning_rate": 0.00019060672299995755, "loss": 4.6, "step": 532000 }, { "epoch": 0.9928644573465298, "grad_norm": 1.1662094593048096, "learning_rate": 0.00019060494534318593, "loss": 4.4341, "step": 532050 }, { "epoch": 0.9929577629059083, "grad_norm": 1.3940582275390625, "learning_rate": 0.00019060316752651254, "loss": 4.5606, "step": 532100 }, { "epoch": 0.9930510684652868, "grad_norm": 1.2326139211654663, "learning_rate": 0.00019060138954994047, "loss": 4.5611, "step": 532150 }, { "epoch": 0.9931443740246654, "grad_norm": 1.0579559803009033, "learning_rate": 0.00019059961141347292, "loss": 4.5872, "step": 532200 }, { "epoch": 0.9932376795840439, "grad_norm": 1.110518217086792, "learning_rate": 0.00019059783311711298, "loss": 4.4326, "step": 532250 }, { "epoch": 0.9933309851434223, "grad_norm": 1.2886223793029785, "learning_rate": 0.0001905960546608638, "loss": 4.505, "step": 532300 }, { "epoch": 0.9934242907028008, "grad_norm": 1.0703747272491455, "learning_rate": 0.00019059427604472854, "loss": 4.6028, "step": 532350 }, { "epoch": 0.9935175962621793, "grad_norm": 1.0797709226608276, "learning_rate": 0.00019059249726871035, "loss": 4.5237, "step": 532400 }, { "epoch": 0.9936109018215578, "grad_norm": 0.8941507935523987, "learning_rate": 0.0001905907183328123, "loss": 4.3898, "step": 532450 }, { "epoch": 0.9937042073809362, "grad_norm": 1.381885290145874, "learning_rate": 0.0001905889392370376, "loss": 4.2965, "step": 532500 }, { "epoch": 0.9937975129403148, "grad_norm": 1.0937845706939697, "learning_rate": 0.0001905871599813893, "loss": 4.5507, "step": 532550 }, { "epoch": 0.9938908184996933, "grad_norm": 1.1042110919952393, "learning_rate": 0.00019058538056587068, "loss": 4.4801, "step": 532600 }, { "epoch": 0.9939841240590718, "grad_norm": 1.0986593961715698, "learning_rate": 0.00019058360099048476, "loss": 4.5203, "step": 532650 }, { "epoch": 0.9940774296184502, "grad_norm": 1.116207242012024, "learning_rate": 0.00019058182125523474, "loss": 4.2824, "step": 532700 }, { "epoch": 0.9941707351778287, "grad_norm": 1.0907038450241089, "learning_rate": 0.00019058004136012375, "loss": 4.5044, "step": 532750 }, { "epoch": 0.9942640407372072, "grad_norm": 0.9307441711425781, "learning_rate": 0.0001905782613051549, "loss": 4.5677, "step": 532800 }, { "epoch": 0.9943573462965857, "grad_norm": 1.0902098417282104, "learning_rate": 0.00019057648109033136, "loss": 4.6141, "step": 532850 }, { "epoch": 0.9944506518559643, "grad_norm": 1.2520498037338257, "learning_rate": 0.00019057470071565628, "loss": 4.6639, "step": 532900 }, { "epoch": 0.9945439574153427, "grad_norm": 1.171056866645813, "learning_rate": 0.0001905729201811328, "loss": 4.7123, "step": 532950 }, { "epoch": 0.9946372629747212, "grad_norm": 1.2704112529754639, "learning_rate": 0.00019057113948676405, "loss": 4.5451, "step": 533000 }, { "epoch": 0.9947305685340997, "grad_norm": 0.9767695069313049, "learning_rate": 0.00019056935863255315, "loss": 4.4256, "step": 533050 }, { "epoch": 0.9948238740934782, "grad_norm": 1.0574843883514404, "learning_rate": 0.0001905675776185033, "loss": 4.5258, "step": 533100 }, { "epoch": 0.9949171796528566, "grad_norm": 1.0145986080169678, "learning_rate": 0.00019056579644461758, "loss": 4.5905, "step": 533150 }, { "epoch": 0.9950104852122351, "grad_norm": 1.249972939491272, "learning_rate": 0.00019056401511089919, "loss": 4.6597, "step": 533200 }, { "epoch": 0.9951037907716136, "grad_norm": 1.0792731046676636, "learning_rate": 0.00019056223361735124, "loss": 4.5722, "step": 533250 }, { "epoch": 0.9951970963309922, "grad_norm": 1.297702670097351, "learning_rate": 0.00019056045196397687, "loss": 4.2897, "step": 533300 }, { "epoch": 0.9952904018903707, "grad_norm": 1.4102619886398315, "learning_rate": 0.00019055867015077923, "loss": 4.4805, "step": 533350 }, { "epoch": 0.9953837074497491, "grad_norm": 1.1016347408294678, "learning_rate": 0.00019055688817776152, "loss": 4.4792, "step": 533400 }, { "epoch": 0.9954770130091276, "grad_norm": 1.1190767288208008, "learning_rate": 0.0001905551060449268, "loss": 4.6041, "step": 533450 }, { "epoch": 0.9955703185685061, "grad_norm": 1.33827805519104, "learning_rate": 0.00019055332375227826, "loss": 4.3679, "step": 533500 }, { "epoch": 0.9956636241278846, "grad_norm": 0.9772111773490906, "learning_rate": 0.00019055154129981903, "loss": 4.549, "step": 533550 }, { "epoch": 0.995756929687263, "grad_norm": 0.7483047842979431, "learning_rate": 0.00019054975868755224, "loss": 4.4554, "step": 533600 }, { "epoch": 0.9958502352466416, "grad_norm": 1.0748953819274902, "learning_rate": 0.00019054797591548113, "loss": 4.6084, "step": 533650 }, { "epoch": 0.9959435408060201, "grad_norm": 1.2168477773666382, "learning_rate": 0.0001905461929836087, "loss": 4.5336, "step": 533700 }, { "epoch": 0.9960368463653986, "grad_norm": 0.8975369930267334, "learning_rate": 0.00019054440989193822, "loss": 4.6317, "step": 533750 }, { "epoch": 0.996130151924777, "grad_norm": 1.256788730621338, "learning_rate": 0.00019054262664047275, "loss": 4.5925, "step": 533800 }, { "epoch": 0.9962234574841555, "grad_norm": 1.3946073055267334, "learning_rate": 0.0001905408432292155, "loss": 4.4788, "step": 533850 }, { "epoch": 0.996316763043534, "grad_norm": 1.403526782989502, "learning_rate": 0.00019053905965816959, "loss": 4.602, "step": 533900 }, { "epoch": 0.9964100686029125, "grad_norm": 1.2732264995574951, "learning_rate": 0.0001905372759273382, "loss": 4.7035, "step": 533950 }, { "epoch": 0.9965033741622911, "grad_norm": 0.9456918835639954, "learning_rate": 0.00019053549203672438, "loss": 4.5725, "step": 534000 }, { "epoch": 0.9965033741622911, "eval_loss": 4.707389831542969, "eval_runtime": 230.3561, "eval_samples_per_second": 11.322, "eval_steps_per_second": 11.322, "eval_tts_loss": 7.574313958996815, "step": 534000 }, { "epoch": 0.9965966797216695, "grad_norm": 1.0162501335144043, "learning_rate": 0.0001905337079863314, "loss": 4.4362, "step": 534050 }, { "epoch": 0.996689985281048, "grad_norm": 1.0250506401062012, "learning_rate": 0.00019053192377616233, "loss": 4.3679, "step": 534100 }, { "epoch": 0.9967832908404265, "grad_norm": 1.0628612041473389, "learning_rate": 0.00019053013940622035, "loss": 4.479, "step": 534150 }, { "epoch": 0.996876596399805, "grad_norm": 1.0884239673614502, "learning_rate": 0.0001905283548765086, "loss": 4.5048, "step": 534200 }, { "epoch": 0.9969699019591834, "grad_norm": 1.1981223821640015, "learning_rate": 0.00019052657018703027, "loss": 4.3252, "step": 534250 }, { "epoch": 0.9970632075185619, "grad_norm": 1.1008185148239136, "learning_rate": 0.00019052478533778845, "loss": 4.1987, "step": 534300 }, { "epoch": 0.9971565130779405, "grad_norm": 1.2123324871063232, "learning_rate": 0.0001905230003287863, "loss": 4.3969, "step": 534350 }, { "epoch": 0.997249818637319, "grad_norm": 1.4960330724716187, "learning_rate": 0.00019052121516002698, "loss": 4.6826, "step": 534400 }, { "epoch": 0.9973431241966975, "grad_norm": 0.847585141658783, "learning_rate": 0.00019051942983151364, "loss": 4.7524, "step": 534450 }, { "epoch": 0.9974364297560759, "grad_norm": 0.9599868655204773, "learning_rate": 0.00019051764434324943, "loss": 4.5751, "step": 534500 }, { "epoch": 0.9975297353154544, "grad_norm": 0.7774083018302917, "learning_rate": 0.00019051585869523754, "loss": 4.5372, "step": 534550 }, { "epoch": 0.9976230408748329, "grad_norm": 0.8145257830619812, "learning_rate": 0.00019051407288748107, "loss": 4.4877, "step": 534600 }, { "epoch": 0.9977163464342114, "grad_norm": 1.031836748123169, "learning_rate": 0.00019051228691998318, "loss": 4.5667, "step": 534650 }, { "epoch": 0.99780965199359, "grad_norm": 0.8227076530456543, "learning_rate": 0.00019051050079274702, "loss": 4.4369, "step": 534700 }, { "epoch": 0.9979029575529684, "grad_norm": 0.9858543276786804, "learning_rate": 0.00019050871450577577, "loss": 4.4834, "step": 534750 }, { "epoch": 0.9979962631123469, "grad_norm": 1.207655429840088, "learning_rate": 0.00019050692805907255, "loss": 4.5887, "step": 534800 }, { "epoch": 0.9980895686717254, "grad_norm": 1.222407341003418, "learning_rate": 0.00019050514145264056, "loss": 4.7608, "step": 534850 }, { "epoch": 0.9981828742311039, "grad_norm": 1.037797451019287, "learning_rate": 0.00019050335468648286, "loss": 4.319, "step": 534900 }, { "epoch": 0.9982761797904823, "grad_norm": 1.1223154067993164, "learning_rate": 0.0001905015677606027, "loss": 4.4363, "step": 534950 }, { "epoch": 0.9983694853498608, "grad_norm": 1.1128987073898315, "learning_rate": 0.0001904997806750032, "loss": 4.4908, "step": 535000 }, { "epoch": 0.9984627909092394, "grad_norm": 0.9986346960067749, "learning_rate": 0.0001904979934296875, "loss": 4.3684, "step": 535050 }, { "epoch": 0.9985560964686179, "grad_norm": 1.190338134765625, "learning_rate": 0.0001904962060246588, "loss": 4.5316, "step": 535100 }, { "epoch": 0.9986494020279963, "grad_norm": 0.8209508657455444, "learning_rate": 0.00019049441845992018, "loss": 4.6394, "step": 535150 }, { "epoch": 0.9987427075873748, "grad_norm": 1.3837367296218872, "learning_rate": 0.00019049263073547484, "loss": 4.232, "step": 535200 }, { "epoch": 0.9988360131467533, "grad_norm": 0.9641016125679016, "learning_rate": 0.00019049084285132595, "loss": 4.4795, "step": 535250 }, { "epoch": 0.9989293187061318, "grad_norm": 1.1519343852996826, "learning_rate": 0.00019048905480747663, "loss": 4.5802, "step": 535300 }, { "epoch": 0.9990226242655103, "grad_norm": 0.9062324166297913, "learning_rate": 0.00019048726660393005, "loss": 4.6197, "step": 535350 }, { "epoch": 0.9991159298248888, "grad_norm": 1.1203495264053345, "learning_rate": 0.00019048547824068938, "loss": 4.4523, "step": 535400 }, { "epoch": 0.9992092353842673, "grad_norm": 1.2169877290725708, "learning_rate": 0.00019048368971775773, "loss": 4.7709, "step": 535450 }, { "epoch": 0.9993025409436458, "grad_norm": 1.1131130456924438, "learning_rate": 0.00019048190103513831, "loss": 4.6224, "step": 535500 }, { "epoch": 0.9993958465030243, "grad_norm": 1.1744643449783325, "learning_rate": 0.00019048011219283427, "loss": 4.4821, "step": 535550 }, { "epoch": 0.9994891520624027, "grad_norm": 1.290284514427185, "learning_rate": 0.00019047832319084873, "loss": 4.708, "step": 535600 }, { "epoch": 0.9995824576217812, "grad_norm": 0.7323465347290039, "learning_rate": 0.00019047653402918488, "loss": 4.694, "step": 535650 }, { "epoch": 0.9996757631811597, "grad_norm": 1.8891117572784424, "learning_rate": 0.00019047474470784587, "loss": 4.4259, "step": 535700 }, { "epoch": 0.9997690687405383, "grad_norm": 0.9453818202018738, "learning_rate": 0.00019047295522683482, "loss": 4.5088, "step": 535750 }, { "epoch": 0.9998623742999168, "grad_norm": 0.7688091397285461, "learning_rate": 0.00019047116558615498, "loss": 4.4709, "step": 535800 }, { "epoch": 0.9999556798592952, "grad_norm": 0.906826913356781, "learning_rate": 0.00019046937578580942, "loss": 4.5063, "step": 535850 }, { "epoch": 1.0000485188908768, "grad_norm": 1.1050300598144531, "learning_rate": 0.00019046758582580133, "loss": 4.8328, "step": 535900 }, { "epoch": 1.0001418244502553, "grad_norm": 1.0780692100524902, "learning_rate": 0.00019046579570613386, "loss": 4.6246, "step": 535950 }, { "epoch": 1.0002351300096337, "grad_norm": 0.8668692111968994, "learning_rate": 0.0001904640054268102, "loss": 4.5249, "step": 536000 }, { "epoch": 1.0003284355690123, "grad_norm": 0.7647411823272705, "learning_rate": 0.00019046221498783346, "loss": 4.6676, "step": 536050 }, { "epoch": 1.0004217411283909, "grad_norm": 0.787149965763092, "learning_rate": 0.00019046042438920687, "loss": 4.4565, "step": 536100 }, { "epoch": 1.0005150466877692, "grad_norm": 1.0924606323242188, "learning_rate": 0.00019045863363093351, "loss": 4.3474, "step": 536150 }, { "epoch": 1.0006083522471478, "grad_norm": 1.0612322092056274, "learning_rate": 0.0001904568427130166, "loss": 4.6581, "step": 536200 }, { "epoch": 1.0007016578065262, "grad_norm": 1.1464825868606567, "learning_rate": 0.00019045505163545927, "loss": 4.7031, "step": 536250 }, { "epoch": 1.0007949633659048, "grad_norm": 1.002851128578186, "learning_rate": 0.0001904532603982647, "loss": 4.4208, "step": 536300 }, { "epoch": 1.0008882689252832, "grad_norm": 1.1369976997375488, "learning_rate": 0.00019045146900143603, "loss": 4.4431, "step": 536350 }, { "epoch": 1.0009815744846617, "grad_norm": 0.9887504577636719, "learning_rate": 0.00019044967744497645, "loss": 4.526, "step": 536400 }, { "epoch": 1.0010748800440403, "grad_norm": 1.428369402885437, "learning_rate": 0.00019044788572888908, "loss": 4.4447, "step": 536450 }, { "epoch": 1.0011681856034187, "grad_norm": 0.7506327629089355, "learning_rate": 0.00019044609385317711, "loss": 4.3349, "step": 536500 }, { "epoch": 1.0012614911627973, "grad_norm": 1.258374571800232, "learning_rate": 0.00019044430181784373, "loss": 4.464, "step": 536550 }, { "epoch": 1.0013547967221756, "grad_norm": 0.9526757001876831, "learning_rate": 0.00019044250962289204, "loss": 4.5613, "step": 536600 }, { "epoch": 1.0014481022815542, "grad_norm": 0.8064827919006348, "learning_rate": 0.00019044071726832522, "loss": 4.2519, "step": 536650 }, { "epoch": 1.0015414078409326, "grad_norm": 1.100570797920227, "learning_rate": 0.0001904389247541465, "loss": 4.3619, "step": 536700 }, { "epoch": 1.0016347134003112, "grad_norm": 1.0979149341583252, "learning_rate": 0.00019043713208035894, "loss": 4.8011, "step": 536750 }, { "epoch": 1.0017280189596898, "grad_norm": 1.1031086444854736, "learning_rate": 0.00019043533924696578, "loss": 4.5644, "step": 536800 }, { "epoch": 1.0018213245190681, "grad_norm": 1.5263746976852417, "learning_rate": 0.00019043354625397014, "loss": 4.5467, "step": 536850 }, { "epoch": 1.0019146300784467, "grad_norm": 0.9368390440940857, "learning_rate": 0.00019043175310137523, "loss": 4.4881, "step": 536900 }, { "epoch": 1.002007935637825, "grad_norm": 1.1151654720306396, "learning_rate": 0.00019042995978918415, "loss": 4.4038, "step": 536950 }, { "epoch": 1.0021012411972037, "grad_norm": 1.0742931365966797, "learning_rate": 0.00019042816631740015, "loss": 4.5626, "step": 537000 }, { "epoch": 1.0021012411972037, "eval_loss": 4.699240207672119, "eval_runtime": 231.6566, "eval_samples_per_second": 11.258, "eval_steps_per_second": 11.258, "eval_tts_loss": 7.5812140480983725, "step": 537000 }, { "epoch": 1.002194546756582, "grad_norm": 1.2127902507781982, "learning_rate": 0.00019042637268602632, "loss": 4.2878, "step": 537050 }, { "epoch": 1.0022878523159606, "grad_norm": 1.0882904529571533, "learning_rate": 0.00019042457889506585, "loss": 4.4729, "step": 537100 }, { "epoch": 1.0023811578753392, "grad_norm": 1.0783560276031494, "learning_rate": 0.00019042278494452194, "loss": 4.415, "step": 537150 }, { "epoch": 1.0024744634347176, "grad_norm": 1.005277156829834, "learning_rate": 0.0001904209908343977, "loss": 4.4799, "step": 537200 }, { "epoch": 1.0025677689940962, "grad_norm": 1.3093390464782715, "learning_rate": 0.00019041919656469632, "loss": 4.3449, "step": 537250 }, { "epoch": 1.0026610745534745, "grad_norm": 1.0098280906677246, "learning_rate": 0.00019041740213542095, "loss": 4.6261, "step": 537300 }, { "epoch": 1.0027543801128531, "grad_norm": 1.3620668649673462, "learning_rate": 0.0001904156075465748, "loss": 4.5349, "step": 537350 }, { "epoch": 1.0028476856722315, "grad_norm": 1.1838512420654297, "learning_rate": 0.000190413812798161, "loss": 4.3803, "step": 537400 }, { "epoch": 1.00294099123161, "grad_norm": 0.8748452067375183, "learning_rate": 0.00019041201789018274, "loss": 4.3642, "step": 537450 }, { "epoch": 1.0030342967909887, "grad_norm": 1.1544240713119507, "learning_rate": 0.0001904102228226432, "loss": 4.497, "step": 537500 }, { "epoch": 1.003127602350367, "grad_norm": 1.1001613140106201, "learning_rate": 0.0001904084275955455, "loss": 4.4069, "step": 537550 }, { "epoch": 1.0032209079097456, "grad_norm": 0.9362643957138062, "learning_rate": 0.00019040663220889283, "loss": 4.417, "step": 537600 }, { "epoch": 1.003314213469124, "grad_norm": 1.3126659393310547, "learning_rate": 0.00019040483666268835, "loss": 4.563, "step": 537650 }, { "epoch": 1.0034075190285026, "grad_norm": 0.9009743332862854, "learning_rate": 0.00019040304095693526, "loss": 4.5927, "step": 537700 }, { "epoch": 1.003500824587881, "grad_norm": 1.1675734519958496, "learning_rate": 0.0001904012450916367, "loss": 4.5082, "step": 537750 }, { "epoch": 1.0035941301472595, "grad_norm": 0.9650899767875671, "learning_rate": 0.00019039944906679585, "loss": 4.5691, "step": 537800 }, { "epoch": 1.003687435706638, "grad_norm": 1.1398006677627563, "learning_rate": 0.00019039765288241588, "loss": 4.5381, "step": 537850 }, { "epoch": 1.0037807412660165, "grad_norm": 1.1838109493255615, "learning_rate": 0.00019039585653849995, "loss": 4.5974, "step": 537900 }, { "epoch": 1.003874046825395, "grad_norm": 0.671570897102356, "learning_rate": 0.00019039406003505126, "loss": 4.3424, "step": 537950 }, { "epoch": 1.0039673523847734, "grad_norm": 1.2386478185653687, "learning_rate": 0.00019039226337207293, "loss": 4.4034, "step": 538000 }, { "epoch": 1.004060657944152, "grad_norm": 1.0624420642852783, "learning_rate": 0.0001903904665495682, "loss": 4.5649, "step": 538050 }, { "epoch": 1.0041539635035304, "grad_norm": 1.6138900518417358, "learning_rate": 0.0001903886695675402, "loss": 4.479, "step": 538100 }, { "epoch": 1.004247269062909, "grad_norm": 1.0794661045074463, "learning_rate": 0.00019038687242599206, "loss": 4.6885, "step": 538150 }, { "epoch": 1.0043405746222875, "grad_norm": 1.1537190675735474, "learning_rate": 0.00019038507512492703, "loss": 4.4598, "step": 538200 }, { "epoch": 1.004433880181666, "grad_norm": 1.0201255083084106, "learning_rate": 0.00019038327766434822, "loss": 4.3721, "step": 538250 }, { "epoch": 1.0045271857410445, "grad_norm": 0.9818448424339294, "learning_rate": 0.00019038148004425884, "loss": 4.4722, "step": 538300 }, { "epoch": 1.0046204913004229, "grad_norm": 1.2369880676269531, "learning_rate": 0.00019037968226466204, "loss": 4.3214, "step": 538350 }, { "epoch": 1.0047137968598014, "grad_norm": 1.2787377834320068, "learning_rate": 0.00019037788432556103, "loss": 4.5827, "step": 538400 }, { "epoch": 1.0048071024191798, "grad_norm": 1.1561073064804077, "learning_rate": 0.00019037608622695894, "loss": 4.6638, "step": 538450 }, { "epoch": 1.0049004079785584, "grad_norm": 1.4122798442840576, "learning_rate": 0.00019037428796885895, "loss": 4.6426, "step": 538500 }, { "epoch": 1.004993713537937, "grad_norm": 1.176689863204956, "learning_rate": 0.00019037248955126429, "loss": 4.5449, "step": 538550 }, { "epoch": 1.0050870190973153, "grad_norm": 1.0620105266571045, "learning_rate": 0.00019037069097417803, "loss": 4.4693, "step": 538600 }, { "epoch": 1.005180324656694, "grad_norm": 1.0586485862731934, "learning_rate": 0.00019036889223760344, "loss": 4.4946, "step": 538650 }, { "epoch": 1.0052736302160723, "grad_norm": 0.9944150447845459, "learning_rate": 0.00019036709334154363, "loss": 4.7105, "step": 538700 }, { "epoch": 1.0053669357754509, "grad_norm": 1.0901190042495728, "learning_rate": 0.00019036529428600183, "loss": 4.5091, "step": 538750 }, { "epoch": 1.0054602413348293, "grad_norm": 1.039443850517273, "learning_rate": 0.00019036349507098115, "loss": 4.4052, "step": 538800 }, { "epoch": 1.0055535468942078, "grad_norm": 1.127867341041565, "learning_rate": 0.00019036169569648482, "loss": 4.4724, "step": 538850 }, { "epoch": 1.0056468524535864, "grad_norm": 1.2713344097137451, "learning_rate": 0.000190359896162516, "loss": 4.5228, "step": 538900 }, { "epoch": 1.0057401580129648, "grad_norm": 1.0708993673324585, "learning_rate": 0.00019035809646907785, "loss": 4.3425, "step": 538950 }, { "epoch": 1.0058334635723434, "grad_norm": 1.1331545114517212, "learning_rate": 0.00019035629661617355, "loss": 4.7897, "step": 539000 }, { "epoch": 1.0059267691317217, "grad_norm": 1.0157674551010132, "learning_rate": 0.0001903544966038063, "loss": 4.3779, "step": 539050 }, { "epoch": 1.0060200746911003, "grad_norm": 1.569435954093933, "learning_rate": 0.00019035269643197927, "loss": 4.4831, "step": 539100 }, { "epoch": 1.0061133802504787, "grad_norm": 1.0539754629135132, "learning_rate": 0.00019035089610069562, "loss": 4.5353, "step": 539150 }, { "epoch": 1.0062066858098573, "grad_norm": 1.0248615741729736, "learning_rate": 0.00019034909560995854, "loss": 4.4455, "step": 539200 }, { "epoch": 1.0062999913692359, "grad_norm": 1.3914874792099, "learning_rate": 0.0001903472949597712, "loss": 4.465, "step": 539250 }, { "epoch": 1.0063932969286142, "grad_norm": 1.2889328002929688, "learning_rate": 0.0001903454941501368, "loss": 4.3567, "step": 539300 }, { "epoch": 1.0064866024879928, "grad_norm": 1.4189263582229614, "learning_rate": 0.0001903436931810585, "loss": 4.589, "step": 539350 }, { "epoch": 1.0065799080473712, "grad_norm": 0.944640576839447, "learning_rate": 0.00019034189205253946, "loss": 4.6261, "step": 539400 }, { "epoch": 1.0066732136067498, "grad_norm": 0.943041205406189, "learning_rate": 0.00019034009076458289, "loss": 4.4165, "step": 539450 }, { "epoch": 1.0067665191661281, "grad_norm": 0.9356258511543274, "learning_rate": 0.00019033828931719194, "loss": 4.4372, "step": 539500 }, { "epoch": 1.0068598247255067, "grad_norm": 0.8639683127403259, "learning_rate": 0.0001903364877103698, "loss": 4.5028, "step": 539550 }, { "epoch": 1.006953130284885, "grad_norm": 1.059850811958313, "learning_rate": 0.00019033468594411966, "loss": 4.542, "step": 539600 }, { "epoch": 1.0070464358442637, "grad_norm": 0.9037283062934875, "learning_rate": 0.0001903328840184447, "loss": 4.5928, "step": 539650 }, { "epoch": 1.0071397414036423, "grad_norm": 1.2375746965408325, "learning_rate": 0.00019033108193334806, "loss": 4.5933, "step": 539700 }, { "epoch": 1.0072330469630206, "grad_norm": 1.3181354999542236, "learning_rate": 0.000190329279688833, "loss": 4.3941, "step": 539750 }, { "epoch": 1.0073263525223992, "grad_norm": 1.1965779066085815, "learning_rate": 0.00019032747728490266, "loss": 4.7212, "step": 539800 }, { "epoch": 1.0074196580817776, "grad_norm": 0.9987531304359436, "learning_rate": 0.00019032567472156018, "loss": 4.5759, "step": 539850 }, { "epoch": 1.0075129636411562, "grad_norm": 0.9417375922203064, "learning_rate": 0.0001903238719988088, "loss": 4.5505, "step": 539900 }, { "epoch": 1.0076062692005345, "grad_norm": 0.9793846607208252, "learning_rate": 0.00019032206911665168, "loss": 4.6156, "step": 539950 }, { "epoch": 1.0076995747599131, "grad_norm": 1.3332839012145996, "learning_rate": 0.00019032026607509198, "loss": 4.5355, "step": 540000 }, { "epoch": 1.0076995747599131, "eval_loss": 4.701556205749512, "eval_runtime": 230.3121, "eval_samples_per_second": 11.324, "eval_steps_per_second": 11.324, "eval_tts_loss": 7.582207388784508, "step": 540000 }, { "epoch": 1.0077928803192917, "grad_norm": 1.0059635639190674, "learning_rate": 0.00019031846287413293, "loss": 4.7126, "step": 540050 }, { "epoch": 1.00788618587867, "grad_norm": 1.0994254350662231, "learning_rate": 0.00019031665951377765, "loss": 4.6395, "step": 540100 }, { "epoch": 1.0079794914380487, "grad_norm": 1.2932167053222656, "learning_rate": 0.0001903148559940294, "loss": 4.4634, "step": 540150 }, { "epoch": 1.008072796997427, "grad_norm": 1.0917683839797974, "learning_rate": 0.0001903130523148913, "loss": 4.4227, "step": 540200 }, { "epoch": 1.0081661025568056, "grad_norm": 0.9450581669807434, "learning_rate": 0.00019031124847636655, "loss": 4.3612, "step": 540250 }, { "epoch": 1.008259408116184, "grad_norm": 1.3047562837600708, "learning_rate": 0.00019030944447845832, "loss": 4.7172, "step": 540300 }, { "epoch": 1.0083527136755626, "grad_norm": 0.9530224204063416, "learning_rate": 0.00019030764032116986, "loss": 4.6142, "step": 540350 }, { "epoch": 1.0084460192349411, "grad_norm": 1.4468913078308105, "learning_rate": 0.0001903058360045043, "loss": 4.364, "step": 540400 }, { "epoch": 1.0085393247943195, "grad_norm": 1.2973716259002686, "learning_rate": 0.00019030403152846476, "loss": 4.5699, "step": 540450 }, { "epoch": 1.008632630353698, "grad_norm": 1.0247949361801147, "learning_rate": 0.00019030222689305455, "loss": 4.5638, "step": 540500 }, { "epoch": 1.0087259359130765, "grad_norm": 0.9376242756843567, "learning_rate": 0.00019030042209827679, "loss": 4.4467, "step": 540550 }, { "epoch": 1.008819241472455, "grad_norm": 1.213931918144226, "learning_rate": 0.0001902986171441347, "loss": 4.5531, "step": 540600 }, { "epoch": 1.0089125470318334, "grad_norm": 1.3004121780395508, "learning_rate": 0.00019029681203063138, "loss": 4.7505, "step": 540650 }, { "epoch": 1.009005852591212, "grad_norm": 1.2604354619979858, "learning_rate": 0.00019029500675777012, "loss": 4.6139, "step": 540700 }, { "epoch": 1.0090991581505906, "grad_norm": 1.1701959371566772, "learning_rate": 0.00019029320132555406, "loss": 4.4938, "step": 540750 }, { "epoch": 1.009192463709969, "grad_norm": 1.0890642404556274, "learning_rate": 0.0001902913957339864, "loss": 4.4423, "step": 540800 }, { "epoch": 1.0092857692693475, "grad_norm": 1.1168172359466553, "learning_rate": 0.00019028958998307025, "loss": 4.517, "step": 540850 }, { "epoch": 1.009379074828726, "grad_norm": 1.2921905517578125, "learning_rate": 0.0001902877840728089, "loss": 4.6458, "step": 540900 }, { "epoch": 1.0094723803881045, "grad_norm": 0.931175708770752, "learning_rate": 0.0001902859780032055, "loss": 4.5139, "step": 540950 }, { "epoch": 1.0095656859474829, "grad_norm": 1.4524635076522827, "learning_rate": 0.00019028417177426325, "loss": 4.6079, "step": 541000 }, { "epoch": 1.0096589915068614, "grad_norm": 1.0124688148498535, "learning_rate": 0.0001902823653859853, "loss": 4.5461, "step": 541050 }, { "epoch": 1.00975229706624, "grad_norm": 0.9658783674240112, "learning_rate": 0.00019028055883837486, "loss": 4.6389, "step": 541100 }, { "epoch": 1.0098456026256184, "grad_norm": 0.8920276165008545, "learning_rate": 0.00019027875213143515, "loss": 4.4377, "step": 541150 }, { "epoch": 1.009938908184997, "grad_norm": 0.9267619848251343, "learning_rate": 0.00019027694526516927, "loss": 4.7028, "step": 541200 }, { "epoch": 1.0100322137443754, "grad_norm": 1.1771366596221924, "learning_rate": 0.0001902751382395805, "loss": 4.4498, "step": 541250 }, { "epoch": 1.010125519303754, "grad_norm": 0.9027936458587646, "learning_rate": 0.000190273331054672, "loss": 4.5482, "step": 541300 }, { "epoch": 1.0102188248631323, "grad_norm": 1.108690619468689, "learning_rate": 0.00019027152371044697, "loss": 4.4529, "step": 541350 }, { "epoch": 1.010312130422511, "grad_norm": 1.2799545526504517, "learning_rate": 0.00019026971620690855, "loss": 4.474, "step": 541400 }, { "epoch": 1.0104054359818895, "grad_norm": 1.0903300046920776, "learning_rate": 0.00019026790854405999, "loss": 4.6141, "step": 541450 }, { "epoch": 1.0104987415412678, "grad_norm": 1.0114182233810425, "learning_rate": 0.00019026610072190442, "loss": 4.4377, "step": 541500 }, { "epoch": 1.0105920471006464, "grad_norm": 1.0331299304962158, "learning_rate": 0.00019026429274044508, "loss": 4.3912, "step": 541550 }, { "epoch": 1.0106853526600248, "grad_norm": 1.3294506072998047, "learning_rate": 0.00019026248459968516, "loss": 4.6991, "step": 541600 }, { "epoch": 1.0107786582194034, "grad_norm": 1.011147141456604, "learning_rate": 0.00019026067629962783, "loss": 4.417, "step": 541650 }, { "epoch": 1.0108719637787817, "grad_norm": 1.153851866722107, "learning_rate": 0.00019025886784027632, "loss": 4.4814, "step": 541700 }, { "epoch": 1.0109652693381603, "grad_norm": 1.225648045539856, "learning_rate": 0.00019025705922163372, "loss": 4.6247, "step": 541750 }, { "epoch": 1.011058574897539, "grad_norm": 1.3746927976608276, "learning_rate": 0.00019025525044370336, "loss": 4.644, "step": 541800 }, { "epoch": 1.0111518804569173, "grad_norm": 0.8752259612083435, "learning_rate": 0.0001902534415064883, "loss": 4.4952, "step": 541850 }, { "epoch": 1.0112451860162959, "grad_norm": 1.1165878772735596, "learning_rate": 0.00019025163240999184, "loss": 4.3048, "step": 541900 }, { "epoch": 1.0113384915756742, "grad_norm": 1.19025456905365, "learning_rate": 0.00019024982315421713, "loss": 4.6627, "step": 541950 }, { "epoch": 1.0114317971350528, "grad_norm": 0.9901291131973267, "learning_rate": 0.00019024801373916734, "loss": 4.6646, "step": 542000 }, { "epoch": 1.0115251026944312, "grad_norm": 1.3689393997192383, "learning_rate": 0.0001902462041648457, "loss": 4.6475, "step": 542050 }, { "epoch": 1.0116184082538098, "grad_norm": 1.0982152223587036, "learning_rate": 0.0001902443944312554, "loss": 4.5955, "step": 542100 }, { "epoch": 1.0117117138131884, "grad_norm": 1.2972334623336792, "learning_rate": 0.00019024258453839962, "loss": 4.5909, "step": 542150 }, { "epoch": 1.0118050193725667, "grad_norm": 1.2537150382995605, "learning_rate": 0.00019024077448628153, "loss": 4.4323, "step": 542200 }, { "epoch": 1.0118983249319453, "grad_norm": 1.2513378858566284, "learning_rate": 0.00019023896427490438, "loss": 4.5902, "step": 542250 }, { "epoch": 1.0119916304913237, "grad_norm": 0.9631972908973694, "learning_rate": 0.00019023715390427133, "loss": 4.603, "step": 542300 }, { "epoch": 1.0120849360507023, "grad_norm": 1.20443594455719, "learning_rate": 0.00019023534337438556, "loss": 4.2743, "step": 542350 }, { "epoch": 1.0121782416100806, "grad_norm": 1.109481692314148, "learning_rate": 0.00019023353268525029, "loss": 4.6316, "step": 542400 }, { "epoch": 1.0122715471694592, "grad_norm": 1.1215111017227173, "learning_rate": 0.00019023172183686872, "loss": 4.5253, "step": 542450 }, { "epoch": 1.0123648527288378, "grad_norm": 1.0365400314331055, "learning_rate": 0.00019022991082924404, "loss": 4.4663, "step": 542500 }, { "epoch": 1.0124581582882162, "grad_norm": 1.0256403684616089, "learning_rate": 0.00019022809966237945, "loss": 4.4957, "step": 542550 }, { "epoch": 1.0125514638475948, "grad_norm": 1.0503144264221191, "learning_rate": 0.00019022628833627813, "loss": 4.6072, "step": 542600 }, { "epoch": 1.0126447694069731, "grad_norm": 1.0918593406677246, "learning_rate": 0.0001902244768509433, "loss": 4.3269, "step": 542650 }, { "epoch": 1.0127380749663517, "grad_norm": 1.0993248224258423, "learning_rate": 0.00019022266520637814, "loss": 4.5809, "step": 542700 }, { "epoch": 1.01283138052573, "grad_norm": 1.193779706954956, "learning_rate": 0.00019022085340258583, "loss": 4.5613, "step": 542750 }, { "epoch": 1.0129246860851087, "grad_norm": 1.0740245580673218, "learning_rate": 0.00019021904143956962, "loss": 4.5994, "step": 542800 }, { "epoch": 1.0130179916444872, "grad_norm": 1.144930124282837, "learning_rate": 0.00019021722931733263, "loss": 4.6738, "step": 542850 }, { "epoch": 1.0131112972038656, "grad_norm": 1.0984407663345337, "learning_rate": 0.00019021541703587814, "loss": 4.3794, "step": 542900 }, { "epoch": 1.0132046027632442, "grad_norm": 1.0142338275909424, "learning_rate": 0.00019021360459520935, "loss": 4.4659, "step": 542950 }, { "epoch": 1.0132979083226226, "grad_norm": 1.2005324363708496, "learning_rate": 0.00019021179199532937, "loss": 4.5428, "step": 543000 }, { "epoch": 1.0132979083226226, "eval_loss": 4.697453022003174, "eval_runtime": 230.5056, "eval_samples_per_second": 11.314, "eval_steps_per_second": 11.314, "eval_tts_loss": 7.589998391816861, "step": 543000 }, { "epoch": 1.0133912138820012, "grad_norm": 1.1389976739883423, "learning_rate": 0.00019020997923624143, "loss": 4.5679, "step": 543050 }, { "epoch": 1.0134845194413795, "grad_norm": 1.1948825120925903, "learning_rate": 0.0001902081663179488, "loss": 4.7387, "step": 543100 }, { "epoch": 1.013577825000758, "grad_norm": 1.2106578350067139, "learning_rate": 0.0001902063532404546, "loss": 4.4878, "step": 543150 }, { "epoch": 1.0136711305601367, "grad_norm": 0.9923614263534546, "learning_rate": 0.00019020454000376205, "loss": 4.4992, "step": 543200 }, { "epoch": 1.013764436119515, "grad_norm": 1.3534685373306274, "learning_rate": 0.0001902027266078744, "loss": 4.5303, "step": 543250 }, { "epoch": 1.0138577416788936, "grad_norm": 0.9535532593727112, "learning_rate": 0.00019020091305279476, "loss": 4.2836, "step": 543300 }, { "epoch": 1.013951047238272, "grad_norm": 1.635932207107544, "learning_rate": 0.00019019909933852645, "loss": 4.5393, "step": 543350 }, { "epoch": 1.0140443527976506, "grad_norm": 0.9862998723983765, "learning_rate": 0.00019019728546507253, "loss": 4.5978, "step": 543400 }, { "epoch": 1.014137658357029, "grad_norm": 1.0948961973190308, "learning_rate": 0.0001901954714324363, "loss": 4.7948, "step": 543450 }, { "epoch": 1.0142309639164075, "grad_norm": 1.0864793062210083, "learning_rate": 0.00019019365724062094, "loss": 4.6331, "step": 543500 }, { "epoch": 1.0143242694757861, "grad_norm": 0.96950763463974, "learning_rate": 0.0001901918428896296, "loss": 4.1941, "step": 543550 }, { "epoch": 1.0144175750351645, "grad_norm": 0.6845828294754028, "learning_rate": 0.00019019002837946557, "loss": 4.5144, "step": 543600 }, { "epoch": 1.014510880594543, "grad_norm": 0.7557372450828552, "learning_rate": 0.00019018821371013202, "loss": 4.5403, "step": 543650 }, { "epoch": 1.0146041861539215, "grad_norm": 0.9517985582351685, "learning_rate": 0.00019018639888163214, "loss": 4.488, "step": 543700 }, { "epoch": 1.0146974917133, "grad_norm": 1.3639390468597412, "learning_rate": 0.00019018458389396913, "loss": 4.4132, "step": 543750 }, { "epoch": 1.0147907972726784, "grad_norm": 1.215621829032898, "learning_rate": 0.00019018276874714616, "loss": 4.5032, "step": 543800 }, { "epoch": 1.014884102832057, "grad_norm": 0.9738672971725464, "learning_rate": 0.0001901809534411665, "loss": 4.3032, "step": 543850 }, { "epoch": 1.0149774083914356, "grad_norm": 1.0050828456878662, "learning_rate": 0.00019017913797603333, "loss": 4.4844, "step": 543900 }, { "epoch": 1.015070713950814, "grad_norm": 1.283862590789795, "learning_rate": 0.00019017732235174984, "loss": 4.5341, "step": 543950 }, { "epoch": 1.0151640195101925, "grad_norm": 0.8861822485923767, "learning_rate": 0.00019017550656831925, "loss": 4.5741, "step": 544000 }, { "epoch": 1.015257325069571, "grad_norm": 1.1378333568572998, "learning_rate": 0.00019017369062574474, "loss": 4.7118, "step": 544050 }, { "epoch": 1.0153506306289495, "grad_norm": 0.9245370030403137, "learning_rate": 0.00019017187452402955, "loss": 4.5186, "step": 544100 }, { "epoch": 1.0154439361883278, "grad_norm": 1.2554208040237427, "learning_rate": 0.00019017005826317687, "loss": 4.6057, "step": 544150 }, { "epoch": 1.0155372417477064, "grad_norm": 1.0860042572021484, "learning_rate": 0.0001901682418431899, "loss": 4.4693, "step": 544200 }, { "epoch": 1.015630547307085, "grad_norm": 0.8820282220840454, "learning_rate": 0.00019016642526407184, "loss": 4.3332, "step": 544250 }, { "epoch": 1.0157238528664634, "grad_norm": 1.1365478038787842, "learning_rate": 0.0001901646085258259, "loss": 4.477, "step": 544300 }, { "epoch": 1.015817158425842, "grad_norm": 1.242470383644104, "learning_rate": 0.00019016279162845532, "loss": 4.6677, "step": 544350 }, { "epoch": 1.0159104639852203, "grad_norm": 0.9678322076797485, "learning_rate": 0.00019016097457196326, "loss": 4.7006, "step": 544400 }, { "epoch": 1.016003769544599, "grad_norm": 1.325256109237671, "learning_rate": 0.00019015915735635296, "loss": 4.3841, "step": 544450 }, { "epoch": 1.0160970751039773, "grad_norm": 1.1439032554626465, "learning_rate": 0.00019015733998162756, "loss": 4.4862, "step": 544500 }, { "epoch": 1.0161903806633559, "grad_norm": 1.0289452075958252, "learning_rate": 0.00019015552244779036, "loss": 4.5438, "step": 544550 }, { "epoch": 1.0162836862227345, "grad_norm": 0.9760018587112427, "learning_rate": 0.0001901537047548445, "loss": 4.5027, "step": 544600 }, { "epoch": 1.0163769917821128, "grad_norm": 0.9549455642700195, "learning_rate": 0.00019015188690279326, "loss": 4.3358, "step": 544650 }, { "epoch": 1.0164702973414914, "grad_norm": 0.9723955988883972, "learning_rate": 0.00019015006889163978, "loss": 4.4062, "step": 544700 }, { "epoch": 1.0165636029008698, "grad_norm": 1.0864965915679932, "learning_rate": 0.0001901482507213873, "loss": 4.6014, "step": 544750 }, { "epoch": 1.0166569084602484, "grad_norm": 1.0674396753311157, "learning_rate": 0.000190146432392039, "loss": 4.6498, "step": 544800 }, { "epoch": 1.0167502140196267, "grad_norm": 1.3030675649642944, "learning_rate": 0.0001901446139035981, "loss": 4.6099, "step": 544850 }, { "epoch": 1.0168435195790053, "grad_norm": 1.2110140323638916, "learning_rate": 0.00019014279525606784, "loss": 4.7885, "step": 544900 }, { "epoch": 1.016936825138384, "grad_norm": 1.2192643880844116, "learning_rate": 0.0001901409764494514, "loss": 4.6666, "step": 544950 }, { "epoch": 1.0170301306977623, "grad_norm": 1.3506648540496826, "learning_rate": 0.00019013915748375198, "loss": 4.64, "step": 545000 }, { "epoch": 1.0171234362571409, "grad_norm": 1.3187509775161743, "learning_rate": 0.00019013733835897284, "loss": 4.5459, "step": 545050 }, { "epoch": 1.0172167418165192, "grad_norm": 1.0731501579284668, "learning_rate": 0.00019013551907511714, "loss": 4.3975, "step": 545100 }, { "epoch": 1.0173100473758978, "grad_norm": 1.1390810012817383, "learning_rate": 0.0001901336996321881, "loss": 4.6323, "step": 545150 }, { "epoch": 1.0174033529352762, "grad_norm": 1.2760993242263794, "learning_rate": 0.00019013188003018895, "loss": 4.5361, "step": 545200 }, { "epoch": 1.0174966584946548, "grad_norm": 1.1871904134750366, "learning_rate": 0.0001901300602691229, "loss": 4.3984, "step": 545250 }, { "epoch": 1.0175899640540333, "grad_norm": 1.1329501867294312, "learning_rate": 0.00019012824034899314, "loss": 4.4322, "step": 545300 }, { "epoch": 1.0176832696134117, "grad_norm": 1.0897506475448608, "learning_rate": 0.0001901264202698029, "loss": 4.5046, "step": 545350 }, { "epoch": 1.0177765751727903, "grad_norm": 1.0522172451019287, "learning_rate": 0.00019012460003155538, "loss": 4.3528, "step": 545400 }, { "epoch": 1.0178698807321687, "grad_norm": 1.3402506113052368, "learning_rate": 0.0001901227796342538, "loss": 4.5892, "step": 545450 }, { "epoch": 1.0179631862915473, "grad_norm": 1.1988232135772705, "learning_rate": 0.00019012095907790138, "loss": 4.6001, "step": 545500 }, { "epoch": 1.0180564918509256, "grad_norm": 1.0484787225723267, "learning_rate": 0.00019011913836250132, "loss": 4.4381, "step": 545550 }, { "epoch": 1.0181497974103042, "grad_norm": 1.1192346811294556, "learning_rate": 0.00019011731748805682, "loss": 4.7114, "step": 545600 }, { "epoch": 1.0182431029696828, "grad_norm": 0.8509928584098816, "learning_rate": 0.00019011549645457114, "loss": 4.6135, "step": 545650 }, { "epoch": 1.0183364085290612, "grad_norm": 0.9683635830879211, "learning_rate": 0.00019011367526204745, "loss": 4.6374, "step": 545700 }, { "epoch": 1.0184297140884397, "grad_norm": 1.0340948104858398, "learning_rate": 0.000190111853910489, "loss": 4.4883, "step": 545750 }, { "epoch": 1.018523019647818, "grad_norm": 1.220017910003662, "learning_rate": 0.00019011003239989896, "loss": 4.4929, "step": 545800 }, { "epoch": 1.0186163252071967, "grad_norm": 1.0294854640960693, "learning_rate": 0.0001901082107302806, "loss": 4.5348, "step": 545850 }, { "epoch": 1.018709630766575, "grad_norm": 1.0364654064178467, "learning_rate": 0.00019010638890163705, "loss": 4.6181, "step": 545900 }, { "epoch": 1.0188029363259536, "grad_norm": 1.1076903343200684, "learning_rate": 0.0001901045669139716, "loss": 4.5875, "step": 545950 }, { "epoch": 1.0188962418853322, "grad_norm": 1.228792667388916, "learning_rate": 0.00019010274476728746, "loss": 4.4998, "step": 546000 }, { "epoch": 1.0188962418853322, "eval_loss": 4.690388202667236, "eval_runtime": 230.4943, "eval_samples_per_second": 11.315, "eval_steps_per_second": 11.315, "eval_tts_loss": 7.6077654536162145, "step": 546000 }, { "epoch": 1.0189895474447106, "grad_norm": 1.1098546981811523, "learning_rate": 0.00019010092246158781, "loss": 4.4673, "step": 546050 }, { "epoch": 1.0190828530040892, "grad_norm": 1.1192878484725952, "learning_rate": 0.00019009909999687593, "loss": 4.6404, "step": 546100 }, { "epoch": 1.0191761585634675, "grad_norm": 1.19114089012146, "learning_rate": 0.00019009727737315495, "loss": 4.5135, "step": 546150 }, { "epoch": 1.0192694641228461, "grad_norm": 1.0214899778366089, "learning_rate": 0.00019009545459042811, "loss": 4.5668, "step": 546200 }, { "epoch": 1.0193627696822245, "grad_norm": 1.4133801460266113, "learning_rate": 0.0001900936316486987, "loss": 4.5919, "step": 546250 }, { "epoch": 1.019456075241603, "grad_norm": 1.2368242740631104, "learning_rate": 0.00019009180854796985, "loss": 4.4759, "step": 546300 }, { "epoch": 1.0195493808009817, "grad_norm": 1.151662826538086, "learning_rate": 0.0001900899852882448, "loss": 4.4599, "step": 546350 }, { "epoch": 1.01964268636036, "grad_norm": 1.1606634855270386, "learning_rate": 0.0001900881618695268, "loss": 4.5594, "step": 546400 }, { "epoch": 1.0197359919197386, "grad_norm": 1.0892674922943115, "learning_rate": 0.00019008633829181905, "loss": 4.3974, "step": 546450 }, { "epoch": 1.019829297479117, "grad_norm": 1.345313549041748, "learning_rate": 0.00019008451455512475, "loss": 4.7025, "step": 546500 }, { "epoch": 1.0199226030384956, "grad_norm": 1.3450990915298462, "learning_rate": 0.00019008269065944711, "loss": 4.5025, "step": 546550 }, { "epoch": 1.020015908597874, "grad_norm": 0.8099167943000793, "learning_rate": 0.0001900808666047894, "loss": 4.7273, "step": 546600 }, { "epoch": 1.0201092141572525, "grad_norm": 1.254998803138733, "learning_rate": 0.0001900790423911548, "loss": 4.4077, "step": 546650 }, { "epoch": 1.0202025197166311, "grad_norm": 1.1391496658325195, "learning_rate": 0.00019007721801854658, "loss": 4.5197, "step": 546700 }, { "epoch": 1.0202958252760095, "grad_norm": 0.8898797035217285, "learning_rate": 0.00019007539348696786, "loss": 4.5667, "step": 546750 }, { "epoch": 1.020389130835388, "grad_norm": 1.1788212060928345, "learning_rate": 0.00019007356879642196, "loss": 4.6814, "step": 546800 }, { "epoch": 1.0204824363947664, "grad_norm": 0.8147475123405457, "learning_rate": 0.00019007174394691204, "loss": 4.5981, "step": 546850 }, { "epoch": 1.020575741954145, "grad_norm": 1.359390139579773, "learning_rate": 0.00019006991893844136, "loss": 4.3825, "step": 546900 }, { "epoch": 1.0206690475135234, "grad_norm": 1.0567032098770142, "learning_rate": 0.0001900680937710131, "loss": 4.469, "step": 546950 }, { "epoch": 1.020762353072902, "grad_norm": 0.9742462038993835, "learning_rate": 0.0001900662684446305, "loss": 4.6335, "step": 547000 }, { "epoch": 1.0208556586322806, "grad_norm": 1.1637524366378784, "learning_rate": 0.0001900644429592968, "loss": 4.6856, "step": 547050 }, { "epoch": 1.020948964191659, "grad_norm": 1.4293187856674194, "learning_rate": 0.0001900626173150152, "loss": 4.5786, "step": 547100 }, { "epoch": 1.0210422697510375, "grad_norm": 1.0260276794433594, "learning_rate": 0.00019006079151178892, "loss": 4.4986, "step": 547150 }, { "epoch": 1.0211355753104159, "grad_norm": 1.3049365282058716, "learning_rate": 0.0001900589655496212, "loss": 4.4575, "step": 547200 }, { "epoch": 1.0212288808697945, "grad_norm": 1.1264426708221436, "learning_rate": 0.00019005713942851524, "loss": 4.4474, "step": 547250 }, { "epoch": 1.0213221864291728, "grad_norm": 1.0150907039642334, "learning_rate": 0.00019005531314847428, "loss": 4.4076, "step": 547300 }, { "epoch": 1.0214154919885514, "grad_norm": 1.0107377767562866, "learning_rate": 0.00019005348670950154, "loss": 4.6729, "step": 547350 }, { "epoch": 1.02150879754793, "grad_norm": 0.9894163012504578, "learning_rate": 0.00019005166011160026, "loss": 4.3435, "step": 547400 }, { "epoch": 1.0216021031073084, "grad_norm": 1.1041195392608643, "learning_rate": 0.0001900498333547736, "loss": 4.5799, "step": 547450 }, { "epoch": 1.021695408666687, "grad_norm": 1.2388858795166016, "learning_rate": 0.00019004800643902488, "loss": 4.6362, "step": 547500 }, { "epoch": 1.0217887142260653, "grad_norm": 1.2287755012512207, "learning_rate": 0.00019004617936435727, "loss": 4.5006, "step": 547550 }, { "epoch": 1.021882019785444, "grad_norm": 1.1951005458831787, "learning_rate": 0.00019004435213077397, "loss": 4.5295, "step": 547600 }, { "epoch": 1.0219753253448223, "grad_norm": 1.1389034986495972, "learning_rate": 0.00019004252473827821, "loss": 4.4791, "step": 547650 }, { "epoch": 1.0220686309042009, "grad_norm": 1.2237263917922974, "learning_rate": 0.0001900406971868733, "loss": 4.4285, "step": 547700 }, { "epoch": 1.0221619364635794, "grad_norm": 0.8839457035064697, "learning_rate": 0.00019003886947656235, "loss": 4.4377, "step": 547750 }, { "epoch": 1.0222552420229578, "grad_norm": 0.9023349285125732, "learning_rate": 0.00019003704160734867, "loss": 4.5573, "step": 547800 }, { "epoch": 1.0223485475823364, "grad_norm": 1.3150593042373657, "learning_rate": 0.00019003521357923543, "loss": 4.557, "step": 547850 }, { "epoch": 1.0224418531417148, "grad_norm": 1.4899663925170898, "learning_rate": 0.0001900333853922259, "loss": 4.7248, "step": 547900 }, { "epoch": 1.0225351587010933, "grad_norm": 0.972459614276886, "learning_rate": 0.00019003155704632328, "loss": 4.5011, "step": 547950 }, { "epoch": 1.0226284642604717, "grad_norm": 1.3415731191635132, "learning_rate": 0.0001900297285415308, "loss": 4.4368, "step": 548000 }, { "epoch": 1.0227217698198503, "grad_norm": 1.1737490892410278, "learning_rate": 0.0001900278998778517, "loss": 4.5888, "step": 548050 }, { "epoch": 1.0228150753792289, "grad_norm": 1.2002341747283936, "learning_rate": 0.0001900260710552892, "loss": 4.3857, "step": 548100 }, { "epoch": 1.0229083809386073, "grad_norm": 1.052138090133667, "learning_rate": 0.0001900242420738465, "loss": 4.491, "step": 548150 }, { "epoch": 1.0230016864979858, "grad_norm": 1.0072627067565918, "learning_rate": 0.00019002241293352687, "loss": 4.6409, "step": 548200 }, { "epoch": 1.0230949920573642, "grad_norm": 1.2276058197021484, "learning_rate": 0.00019002058363433352, "loss": 4.5174, "step": 548250 }, { "epoch": 1.0231882976167428, "grad_norm": 0.861262857913971, "learning_rate": 0.0001900187541762697, "loss": 4.5847, "step": 548300 }, { "epoch": 1.0232816031761212, "grad_norm": 0.8748422265052795, "learning_rate": 0.00019001692455933856, "loss": 4.5348, "step": 548350 }, { "epoch": 1.0233749087354997, "grad_norm": 1.2001471519470215, "learning_rate": 0.00019001509478354346, "loss": 4.6703, "step": 548400 }, { "epoch": 1.0234682142948783, "grad_norm": 1.0208408832550049, "learning_rate": 0.0001900132648488875, "loss": 4.4133, "step": 548450 }, { "epoch": 1.0235615198542567, "grad_norm": 1.0386520624160767, "learning_rate": 0.000190011434755374, "loss": 4.3354, "step": 548500 }, { "epoch": 1.0236548254136353, "grad_norm": 1.4474914073944092, "learning_rate": 0.00019000960450300613, "loss": 4.7387, "step": 548550 }, { "epoch": 1.0237481309730136, "grad_norm": 1.2579118013381958, "learning_rate": 0.00019000777409178716, "loss": 4.3727, "step": 548600 }, { "epoch": 1.0238414365323922, "grad_norm": 0.9952757358551025, "learning_rate": 0.0001900059435217203, "loss": 4.7372, "step": 548650 }, { "epoch": 1.0239347420917706, "grad_norm": 1.212010145187378, "learning_rate": 0.0001900041127928088, "loss": 4.5565, "step": 548700 }, { "epoch": 1.0240280476511492, "grad_norm": 1.0180038213729858, "learning_rate": 0.00019000228190505587, "loss": 4.4266, "step": 548750 }, { "epoch": 1.0241213532105278, "grad_norm": 1.1852561235427856, "learning_rate": 0.00019000045085846475, "loss": 4.4221, "step": 548800 }, { "epoch": 1.0242146587699061, "grad_norm": 1.3834125995635986, "learning_rate": 0.00018999861965303866, "loss": 4.53, "step": 548850 }, { "epoch": 1.0243079643292847, "grad_norm": 0.8798831701278687, "learning_rate": 0.00018999678828878085, "loss": 4.4843, "step": 548900 }, { "epoch": 1.024401269888663, "grad_norm": 1.1726356744766235, "learning_rate": 0.00018999495676569455, "loss": 4.5477, "step": 548950 }, { "epoch": 1.0244945754480417, "grad_norm": 1.201399326324463, "learning_rate": 0.00018999312508378297, "loss": 4.4439, "step": 549000 }, { "epoch": 1.0244945754480417, "eval_loss": 4.692353248596191, "eval_runtime": 234.9081, "eval_samples_per_second": 11.102, "eval_steps_per_second": 11.102, "eval_tts_loss": 7.582008384266708, "step": 549000 }, { "epoch": 1.02458788100742, "grad_norm": 0.9765532612800598, "learning_rate": 0.0001899912932430494, "loss": 4.6308, "step": 549050 }, { "epoch": 1.0246811865667986, "grad_norm": 1.1887390613555908, "learning_rate": 0.00018998946124349698, "loss": 4.6674, "step": 549100 }, { "epoch": 1.0247744921261772, "grad_norm": 0.9697726368904114, "learning_rate": 0.000189987629085129, "loss": 4.374, "step": 549150 }, { "epoch": 1.0248677976855556, "grad_norm": 1.1879429817199707, "learning_rate": 0.00018998579676794873, "loss": 4.4175, "step": 549200 }, { "epoch": 1.0249611032449342, "grad_norm": 0.9584805965423584, "learning_rate": 0.00018998396429195934, "loss": 4.4396, "step": 549250 }, { "epoch": 1.0250544088043125, "grad_norm": 0.89393150806427, "learning_rate": 0.00018998213165716406, "loss": 4.4157, "step": 549300 }, { "epoch": 1.0251477143636911, "grad_norm": 1.1879955530166626, "learning_rate": 0.00018998029886356616, "loss": 4.4587, "step": 549350 }, { "epoch": 1.0252410199230695, "grad_norm": 1.3241792917251587, "learning_rate": 0.0001899784659111689, "loss": 4.5936, "step": 549400 }, { "epoch": 1.025334325482448, "grad_norm": 1.1142139434814453, "learning_rate": 0.00018997663279997546, "loss": 4.6565, "step": 549450 }, { "epoch": 1.0254276310418264, "grad_norm": 1.0738365650177002, "learning_rate": 0.0001899747995299891, "loss": 4.451, "step": 549500 }, { "epoch": 1.025520936601205, "grad_norm": 0.9906955361366272, "learning_rate": 0.00018997296610121303, "loss": 4.571, "step": 549550 }, { "epoch": 1.0256142421605836, "grad_norm": 1.2519268989562988, "learning_rate": 0.00018997113251365054, "loss": 4.6531, "step": 549600 }, { "epoch": 1.025707547719962, "grad_norm": 1.2590681314468384, "learning_rate": 0.00018996929876730477, "loss": 4.5396, "step": 549650 }, { "epoch": 1.0258008532793406, "grad_norm": 1.121073603630066, "learning_rate": 0.00018996746486217906, "loss": 4.5207, "step": 549700 }, { "epoch": 1.025894158838719, "grad_norm": 1.0509285926818848, "learning_rate": 0.00018996563079827662, "loss": 4.4286, "step": 549750 }, { "epoch": 1.0259874643980975, "grad_norm": 0.942859411239624, "learning_rate": 0.00018996379657560067, "loss": 4.5319, "step": 549800 }, { "epoch": 1.0260807699574759, "grad_norm": 1.353922724723816, "learning_rate": 0.00018996196219415442, "loss": 4.5655, "step": 549850 }, { "epoch": 1.0261740755168545, "grad_norm": 1.0995395183563232, "learning_rate": 0.00018996012765394116, "loss": 4.6327, "step": 549900 }, { "epoch": 1.026267381076233, "grad_norm": 0.9851589798927307, "learning_rate": 0.0001899582929549641, "loss": 4.6047, "step": 549950 }, { "epoch": 1.0263606866356114, "grad_norm": 1.0693026781082153, "learning_rate": 0.0001899564580972265, "loss": 4.5257, "step": 550000 }, { "epoch": 1.02645399219499, "grad_norm": 1.1348246335983276, "learning_rate": 0.00018995462308073153, "loss": 4.4607, "step": 550050 }, { "epoch": 1.0265472977543684, "grad_norm": 1.1245973110198975, "learning_rate": 0.0001899527879054825, "loss": 4.2938, "step": 550100 }, { "epoch": 1.026640603313747, "grad_norm": 0.7979853749275208, "learning_rate": 0.00018995095257148264, "loss": 4.4723, "step": 550150 }, { "epoch": 1.0267339088731253, "grad_norm": 1.2790230512619019, "learning_rate": 0.00018994911707873518, "loss": 4.4899, "step": 550200 }, { "epoch": 1.026827214432504, "grad_norm": 1.3592939376831055, "learning_rate": 0.00018994728142724334, "loss": 4.4838, "step": 550250 }, { "epoch": 1.0269205199918825, "grad_norm": 1.0678589344024658, "learning_rate": 0.00018994544561701034, "loss": 4.5184, "step": 550300 }, { "epoch": 1.0270138255512609, "grad_norm": 1.1794860363006592, "learning_rate": 0.00018994360964803952, "loss": 4.4968, "step": 550350 }, { "epoch": 1.0271071311106394, "grad_norm": 1.2433096170425415, "learning_rate": 0.00018994177352033402, "loss": 4.4661, "step": 550400 }, { "epoch": 1.0272004366700178, "grad_norm": 1.5378727912902832, "learning_rate": 0.00018993993723389713, "loss": 4.5227, "step": 550450 }, { "epoch": 1.0272937422293964, "grad_norm": 1.1838160753250122, "learning_rate": 0.00018993810078873205, "loss": 4.7627, "step": 550500 }, { "epoch": 1.0273870477887748, "grad_norm": 0.9702062606811523, "learning_rate": 0.00018993626418484209, "loss": 4.4266, "step": 550550 }, { "epoch": 1.0274803533481534, "grad_norm": 1.1267657279968262, "learning_rate": 0.00018993442742223042, "loss": 4.4647, "step": 550600 }, { "epoch": 1.027573658907532, "grad_norm": 1.2197514772415161, "learning_rate": 0.00018993259050090032, "loss": 4.6111, "step": 550650 }, { "epoch": 1.0276669644669103, "grad_norm": 1.0420445203781128, "learning_rate": 0.00018993075342085502, "loss": 4.5919, "step": 550700 }, { "epoch": 1.027760270026289, "grad_norm": 1.1075886487960815, "learning_rate": 0.00018992891618209776, "loss": 4.3646, "step": 550750 }, { "epoch": 1.0278535755856673, "grad_norm": 1.2440576553344727, "learning_rate": 0.00018992707878463179, "loss": 4.5511, "step": 550800 }, { "epoch": 1.0279468811450458, "grad_norm": 1.0751181840896606, "learning_rate": 0.00018992524122846032, "loss": 4.4715, "step": 550850 }, { "epoch": 1.0280401867044242, "grad_norm": 1.2242587804794312, "learning_rate": 0.00018992340351358666, "loss": 4.588, "step": 550900 }, { "epoch": 1.0281334922638028, "grad_norm": 0.8009092807769775, "learning_rate": 0.000189921565640014, "loss": 4.3882, "step": 550950 }, { "epoch": 1.0282267978231814, "grad_norm": 1.0183814764022827, "learning_rate": 0.0001899197276077456, "loss": 4.715, "step": 551000 }, { "epoch": 1.0283201033825597, "grad_norm": 1.318489670753479, "learning_rate": 0.00018991788941678472, "loss": 4.4744, "step": 551050 }, { "epoch": 1.0284134089419383, "grad_norm": 0.7976197600364685, "learning_rate": 0.00018991605106713456, "loss": 4.5125, "step": 551100 }, { "epoch": 1.0285067145013167, "grad_norm": 1.1699376106262207, "learning_rate": 0.0001899142125587984, "loss": 4.4062, "step": 551150 }, { "epoch": 1.0286000200606953, "grad_norm": 1.148091197013855, "learning_rate": 0.0001899123738917795, "loss": 4.6467, "step": 551200 }, { "epoch": 1.0286933256200737, "grad_norm": 1.1050102710723877, "learning_rate": 0.00018991053506608104, "loss": 4.4554, "step": 551250 }, { "epoch": 1.0287866311794522, "grad_norm": 1.2489887475967407, "learning_rate": 0.00018990869608170635, "loss": 4.5437, "step": 551300 }, { "epoch": 1.0288799367388308, "grad_norm": 1.226762056350708, "learning_rate": 0.0001899068569386586, "loss": 4.8283, "step": 551350 }, { "epoch": 1.0289732422982092, "grad_norm": 0.9568690061569214, "learning_rate": 0.00018990501763694107, "loss": 4.3061, "step": 551400 }, { "epoch": 1.0290665478575878, "grad_norm": 0.7701925039291382, "learning_rate": 0.000189903178176557, "loss": 4.5815, "step": 551450 }, { "epoch": 1.0291598534169661, "grad_norm": 1.1537120342254639, "learning_rate": 0.00018990133855750963, "loss": 4.5177, "step": 551500 }, { "epoch": 1.0292531589763447, "grad_norm": 1.6827070713043213, "learning_rate": 0.00018989949877980226, "loss": 4.4596, "step": 551550 }, { "epoch": 1.029346464535723, "grad_norm": 1.1412551403045654, "learning_rate": 0.00018989765884343807, "loss": 4.5545, "step": 551600 }, { "epoch": 1.0294397700951017, "grad_norm": 1.2005754709243774, "learning_rate": 0.00018989581874842034, "loss": 4.4322, "step": 551650 }, { "epoch": 1.0295330756544803, "grad_norm": 0.8345110416412354, "learning_rate": 0.00018989397849475227, "loss": 4.3902, "step": 551700 }, { "epoch": 1.0296263812138586, "grad_norm": 0.9611878991127014, "learning_rate": 0.00018989213808243718, "loss": 4.5405, "step": 551750 }, { "epoch": 1.0297196867732372, "grad_norm": 1.098541498184204, "learning_rate": 0.00018989029751147828, "loss": 4.1599, "step": 551800 }, { "epoch": 1.0298129923326156, "grad_norm": 1.1365591287612915, "learning_rate": 0.0001898884567818788, "loss": 4.6236, "step": 551850 }, { "epoch": 1.0299062978919942, "grad_norm": 0.9525176882743835, "learning_rate": 0.000189886615893642, "loss": 4.5895, "step": 551900 }, { "epoch": 1.0299996034513725, "grad_norm": 1.2500810623168945, "learning_rate": 0.0001898847748467712, "loss": 4.6422, "step": 551950 }, { "epoch": 1.0300929090107511, "grad_norm": 0.9244524836540222, "learning_rate": 0.00018988293364126952, "loss": 4.4584, "step": 552000 }, { "epoch": 1.0300929090107511, "eval_loss": 4.697979927062988, "eval_runtime": 232.8011, "eval_samples_per_second": 11.203, "eval_steps_per_second": 11.203, "eval_tts_loss": 7.645190617760952, "step": 552000 }, { "epoch": 1.0301862145701297, "grad_norm": 1.1201730966567993, "learning_rate": 0.0001898810922771403, "loss": 4.5343, "step": 552050 }, { "epoch": 1.030279520129508, "grad_norm": 1.2420499324798584, "learning_rate": 0.0001898792507543868, "loss": 4.5632, "step": 552100 }, { "epoch": 1.0303728256888867, "grad_norm": 1.192976474761963, "learning_rate": 0.00018987740907301222, "loss": 4.5806, "step": 552150 }, { "epoch": 1.030466131248265, "grad_norm": 0.9666003584861755, "learning_rate": 0.00018987556723301983, "loss": 4.4386, "step": 552200 }, { "epoch": 1.0305594368076436, "grad_norm": 0.8053319454193115, "learning_rate": 0.00018987372523441285, "loss": 4.585, "step": 552250 }, { "epoch": 1.030652742367022, "grad_norm": 0.997139573097229, "learning_rate": 0.00018987188307719458, "loss": 4.5064, "step": 552300 }, { "epoch": 1.0307460479264006, "grad_norm": 0.9804717302322388, "learning_rate": 0.00018987004076136824, "loss": 4.5395, "step": 552350 }, { "epoch": 1.0308393534857792, "grad_norm": 0.9765677452087402, "learning_rate": 0.00018986819828693713, "loss": 4.5284, "step": 552400 }, { "epoch": 1.0309326590451575, "grad_norm": 0.749104917049408, "learning_rate": 0.0001898663556539044, "loss": 4.4265, "step": 552450 }, { "epoch": 1.031025964604536, "grad_norm": 1.351374626159668, "learning_rate": 0.0001898645128622734, "loss": 4.4012, "step": 552500 }, { "epoch": 1.0311192701639145, "grad_norm": 0.7262341976165771, "learning_rate": 0.00018986266991204734, "loss": 4.4933, "step": 552550 }, { "epoch": 1.031212575723293, "grad_norm": 1.0371111631393433, "learning_rate": 0.0001898608268032295, "loss": 4.4978, "step": 552600 }, { "epoch": 1.0313058812826714, "grad_norm": 1.2455886602401733, "learning_rate": 0.00018985898353582307, "loss": 4.5111, "step": 552650 }, { "epoch": 1.03139918684205, "grad_norm": 1.0494903326034546, "learning_rate": 0.00018985714010983138, "loss": 4.5405, "step": 552700 }, { "epoch": 1.0314924924014286, "grad_norm": 1.4863853454589844, "learning_rate": 0.0001898552965252576, "loss": 4.6239, "step": 552750 }, { "epoch": 1.031585797960807, "grad_norm": 1.0769399404525757, "learning_rate": 0.0001898534527821051, "loss": 4.6296, "step": 552800 }, { "epoch": 1.0316791035201855, "grad_norm": 1.1758257150650024, "learning_rate": 0.00018985160888037705, "loss": 4.6943, "step": 552850 }, { "epoch": 1.031772409079564, "grad_norm": 0.8951691389083862, "learning_rate": 0.00018984976482007667, "loss": 4.5039, "step": 552900 }, { "epoch": 1.0318657146389425, "grad_norm": 0.9229235053062439, "learning_rate": 0.00018984792060120734, "loss": 4.4887, "step": 552950 }, { "epoch": 1.0319590201983209, "grad_norm": 0.8344191312789917, "learning_rate": 0.0001898460762237722, "loss": 4.7193, "step": 553000 }, { "epoch": 1.0320523257576995, "grad_norm": 1.2222954034805298, "learning_rate": 0.0001898442316877745, "loss": 4.5866, "step": 553050 }, { "epoch": 1.032145631317078, "grad_norm": 1.2271044254302979, "learning_rate": 0.00018984238699321763, "loss": 4.6491, "step": 553100 }, { "epoch": 1.0322389368764564, "grad_norm": 0.8897698521614075, "learning_rate": 0.00018984054214010469, "loss": 4.6259, "step": 553150 }, { "epoch": 1.032332242435835, "grad_norm": 1.2667361497879028, "learning_rate": 0.00018983869712843902, "loss": 4.5809, "step": 553200 }, { "epoch": 1.0324255479952134, "grad_norm": 1.1170142889022827, "learning_rate": 0.00018983685195822386, "loss": 4.5109, "step": 553250 }, { "epoch": 1.032518853554592, "grad_norm": 0.8995992541313171, "learning_rate": 0.00018983500662946248, "loss": 4.4187, "step": 553300 }, { "epoch": 1.0326121591139703, "grad_norm": 0.9850976467132568, "learning_rate": 0.00018983316114215809, "loss": 4.5642, "step": 553350 }, { "epoch": 1.032705464673349, "grad_norm": 1.0501117706298828, "learning_rate": 0.00018983131549631398, "loss": 4.7431, "step": 553400 }, { "epoch": 1.0327987702327275, "grad_norm": 0.5963752865791321, "learning_rate": 0.00018982946969193343, "loss": 4.5484, "step": 553450 }, { "epoch": 1.0328920757921058, "grad_norm": 1.4002224206924438, "learning_rate": 0.00018982762372901966, "loss": 4.3498, "step": 553500 }, { "epoch": 1.0329853813514844, "grad_norm": 1.3687794208526611, "learning_rate": 0.00018982577760757592, "loss": 4.597, "step": 553550 }, { "epoch": 1.0330786869108628, "grad_norm": 1.0715551376342773, "learning_rate": 0.0001898239313276055, "loss": 4.7076, "step": 553600 }, { "epoch": 1.0331719924702414, "grad_norm": 1.2118879556655884, "learning_rate": 0.00018982208488911164, "loss": 4.5442, "step": 553650 }, { "epoch": 1.0332652980296197, "grad_norm": 1.1751309633255005, "learning_rate": 0.00018982023829209764, "loss": 4.5322, "step": 553700 }, { "epoch": 1.0333586035889983, "grad_norm": 0.9542876482009888, "learning_rate": 0.0001898183915365667, "loss": 4.5454, "step": 553750 }, { "epoch": 1.033451909148377, "grad_norm": 1.0695343017578125, "learning_rate": 0.00018981654462252206, "loss": 4.5214, "step": 553800 }, { "epoch": 1.0335452147077553, "grad_norm": 1.3467059135437012, "learning_rate": 0.00018981469754996707, "loss": 4.7626, "step": 553850 }, { "epoch": 1.0336385202671339, "grad_norm": 1.1576812267303467, "learning_rate": 0.00018981285031890495, "loss": 4.5459, "step": 553900 }, { "epoch": 1.0337318258265122, "grad_norm": 1.2458592653274536, "learning_rate": 0.00018981100292933894, "loss": 4.5422, "step": 553950 }, { "epoch": 1.0338251313858908, "grad_norm": 1.2345342636108398, "learning_rate": 0.00018980915538127231, "loss": 4.6187, "step": 554000 }, { "epoch": 1.0339184369452692, "grad_norm": 1.18948495388031, "learning_rate": 0.0001898073076747083, "loss": 4.5004, "step": 554050 }, { "epoch": 1.0340117425046478, "grad_norm": 1.0282189846038818, "learning_rate": 0.00018980545980965024, "loss": 4.5688, "step": 554100 }, { "epoch": 1.0341050480640264, "grad_norm": 1.166161060333252, "learning_rate": 0.0001898036117861013, "loss": 4.6386, "step": 554150 }, { "epoch": 1.0341983536234047, "grad_norm": 0.9692121744155884, "learning_rate": 0.00018980176360406482, "loss": 4.432, "step": 554200 }, { "epoch": 1.0342916591827833, "grad_norm": 1.222730040550232, "learning_rate": 0.00018979991526354404, "loss": 4.5393, "step": 554250 }, { "epoch": 1.0343849647421617, "grad_norm": 0.9999300837516785, "learning_rate": 0.00018979806676454217, "loss": 4.5835, "step": 554300 }, { "epoch": 1.0344782703015403, "grad_norm": 1.437756061553955, "learning_rate": 0.00018979621810706255, "loss": 4.6452, "step": 554350 }, { "epoch": 1.0345715758609186, "grad_norm": 1.035550594329834, "learning_rate": 0.00018979436929110837, "loss": 4.4114, "step": 554400 }, { "epoch": 1.0346648814202972, "grad_norm": 1.122685432434082, "learning_rate": 0.00018979252031668296, "loss": 4.6043, "step": 554450 }, { "epoch": 1.0347581869796758, "grad_norm": 1.1198245286941528, "learning_rate": 0.00018979067118378952, "loss": 4.4459, "step": 554500 }, { "epoch": 1.0348514925390542, "grad_norm": 1.0280388593673706, "learning_rate": 0.00018978882189243134, "loss": 4.4387, "step": 554550 }, { "epoch": 1.0349447980984328, "grad_norm": 1.0488851070404053, "learning_rate": 0.00018978697244261173, "loss": 4.4168, "step": 554600 }, { "epoch": 1.0350381036578111, "grad_norm": 1.5133382081985474, "learning_rate": 0.00018978512283433386, "loss": 4.6665, "step": 554650 }, { "epoch": 1.0351314092171897, "grad_norm": 0.999513566493988, "learning_rate": 0.00018978327306760112, "loss": 4.7277, "step": 554700 }, { "epoch": 1.035224714776568, "grad_norm": 1.3199903964996338, "learning_rate": 0.00018978142314241665, "loss": 4.5005, "step": 554750 }, { "epoch": 1.0353180203359467, "grad_norm": 1.1765607595443726, "learning_rate": 0.00018977957305878375, "loss": 4.4154, "step": 554800 }, { "epoch": 1.0354113258953253, "grad_norm": 1.4772642850875854, "learning_rate": 0.00018977772281670574, "loss": 4.38, "step": 554850 }, { "epoch": 1.0355046314547036, "grad_norm": 1.2143290042877197, "learning_rate": 0.0001897758724161858, "loss": 4.6778, "step": 554900 }, { "epoch": 1.0355979370140822, "grad_norm": 1.1253890991210938, "learning_rate": 0.0001897740218572273, "loss": 4.5541, "step": 554950 }, { "epoch": 1.0356912425734606, "grad_norm": 0.9873462319374084, "learning_rate": 0.00018977217113983342, "loss": 4.5327, "step": 555000 }, { "epoch": 1.0356912425734606, "eval_loss": 4.694144248962402, "eval_runtime": 232.7899, "eval_samples_per_second": 11.203, "eval_steps_per_second": 11.203, "eval_tts_loss": 7.576006341847003, "step": 555000 }, { "epoch": 1.0357845481328392, "grad_norm": 0.9340860247612, "learning_rate": 0.00018977032026400745, "loss": 4.4378, "step": 555050 }, { "epoch": 1.0358778536922175, "grad_norm": 1.1068315505981445, "learning_rate": 0.00018976846922975264, "loss": 4.6226, "step": 555100 }, { "epoch": 1.035971159251596, "grad_norm": 0.9125676155090332, "learning_rate": 0.0001897666180370723, "loss": 4.2313, "step": 555150 }, { "epoch": 1.0360644648109747, "grad_norm": 0.7258133292198181, "learning_rate": 0.0001897647666859697, "loss": 4.3426, "step": 555200 }, { "epoch": 1.036157770370353, "grad_norm": 1.3283452987670898, "learning_rate": 0.00018976291517644803, "loss": 4.6055, "step": 555250 }, { "epoch": 1.0362510759297316, "grad_norm": 1.1347540616989136, "learning_rate": 0.00018976106350851067, "loss": 4.5741, "step": 555300 }, { "epoch": 1.03634438148911, "grad_norm": 1.2645660638809204, "learning_rate": 0.00018975921168216074, "loss": 4.4495, "step": 555350 }, { "epoch": 1.0364376870484886, "grad_norm": 1.0976849794387817, "learning_rate": 0.00018975735969740167, "loss": 4.5656, "step": 555400 }, { "epoch": 1.036530992607867, "grad_norm": 1.1340277194976807, "learning_rate": 0.00018975550755423662, "loss": 4.502, "step": 555450 }, { "epoch": 1.0366242981672456, "grad_norm": 1.1174198389053345, "learning_rate": 0.0001897536552526689, "loss": 4.5445, "step": 555500 }, { "epoch": 1.0367176037266241, "grad_norm": 0.8465671539306641, "learning_rate": 0.0001897518027927018, "loss": 4.5581, "step": 555550 }, { "epoch": 1.0368109092860025, "grad_norm": 0.7383942604064941, "learning_rate": 0.00018974995017433853, "loss": 4.6791, "step": 555600 }, { "epoch": 1.036904214845381, "grad_norm": 0.9870350956916809, "learning_rate": 0.0001897480973975824, "loss": 4.5087, "step": 555650 }, { "epoch": 1.0369975204047595, "grad_norm": 1.0926276445388794, "learning_rate": 0.00018974624446243664, "loss": 4.4633, "step": 555700 }, { "epoch": 1.037090825964138, "grad_norm": 1.1819891929626465, "learning_rate": 0.0001897443913689046, "loss": 4.4491, "step": 555750 }, { "epoch": 1.0371841315235164, "grad_norm": 1.1450855731964111, "learning_rate": 0.00018974253811698946, "loss": 4.7746, "step": 555800 }, { "epoch": 1.037277437082895, "grad_norm": 0.9828746318817139, "learning_rate": 0.00018974068470669454, "loss": 4.4467, "step": 555850 }, { "epoch": 1.0373707426422736, "grad_norm": 1.193833827972412, "learning_rate": 0.00018973883113802308, "loss": 4.7393, "step": 555900 }, { "epoch": 1.037464048201652, "grad_norm": 1.0269185304641724, "learning_rate": 0.00018973697741097843, "loss": 4.2026, "step": 555950 }, { "epoch": 1.0375573537610305, "grad_norm": 1.2609224319458008, "learning_rate": 0.00018973512352556379, "loss": 4.4624, "step": 556000 }, { "epoch": 1.037650659320409, "grad_norm": 1.2936489582061768, "learning_rate": 0.0001897332694817824, "loss": 4.5701, "step": 556050 }, { "epoch": 1.0377439648797875, "grad_norm": 0.8027763962745667, "learning_rate": 0.0001897314152796376, "loss": 4.4779, "step": 556100 }, { "epoch": 1.0378372704391658, "grad_norm": 1.2781606912612915, "learning_rate": 0.00018972956091913265, "loss": 4.434, "step": 556150 }, { "epoch": 1.0379305759985444, "grad_norm": 1.1502928733825684, "learning_rate": 0.00018972770640027085, "loss": 4.5025, "step": 556200 }, { "epoch": 1.038023881557923, "grad_norm": 1.1656410694122314, "learning_rate": 0.00018972585172305538, "loss": 4.5134, "step": 556250 }, { "epoch": 1.0381171871173014, "grad_norm": 1.2848443984985352, "learning_rate": 0.0001897239968874896, "loss": 4.4367, "step": 556300 }, { "epoch": 1.03821049267668, "grad_norm": 1.2250791788101196, "learning_rate": 0.00018972214189357672, "loss": 4.6023, "step": 556350 }, { "epoch": 1.0383037982360583, "grad_norm": 1.2145739793777466, "learning_rate": 0.00018972028674132008, "loss": 4.8139, "step": 556400 }, { "epoch": 1.038397103795437, "grad_norm": 0.9680354595184326, "learning_rate": 0.0001897184314307229, "loss": 4.4938, "step": 556450 }, { "epoch": 1.0384904093548153, "grad_norm": 1.0261461734771729, "learning_rate": 0.0001897165759617885, "loss": 4.6593, "step": 556500 }, { "epoch": 1.0385837149141939, "grad_norm": 0.9852768778800964, "learning_rate": 0.0001897147203345201, "loss": 4.6186, "step": 556550 }, { "epoch": 1.0386770204735725, "grad_norm": 0.9749924540519714, "learning_rate": 0.000189712864548921, "loss": 4.5509, "step": 556600 }, { "epoch": 1.0387703260329508, "grad_norm": 1.1581624746322632, "learning_rate": 0.00018971100860499452, "loss": 4.6174, "step": 556650 }, { "epoch": 1.0388636315923294, "grad_norm": 1.1250700950622559, "learning_rate": 0.00018970915250274387, "loss": 4.5048, "step": 556700 }, { "epoch": 1.0389569371517078, "grad_norm": 1.299763798713684, "learning_rate": 0.00018970729624217235, "loss": 4.5505, "step": 556750 }, { "epoch": 1.0390502427110864, "grad_norm": 1.0283108949661255, "learning_rate": 0.00018970543982328324, "loss": 4.4908, "step": 556800 }, { "epoch": 1.0391435482704647, "grad_norm": 1.2390642166137695, "learning_rate": 0.0001897035832460798, "loss": 4.3632, "step": 556850 }, { "epoch": 1.0392368538298433, "grad_norm": 1.1650816202163696, "learning_rate": 0.00018970172651056535, "loss": 4.6495, "step": 556900 }, { "epoch": 1.039330159389222, "grad_norm": 1.0848872661590576, "learning_rate": 0.0001896998696167431, "loss": 4.5114, "step": 556950 }, { "epoch": 1.0394234649486003, "grad_norm": 1.2461999654769897, "learning_rate": 0.0001896980125646164, "loss": 4.5043, "step": 557000 }, { "epoch": 1.0395167705079789, "grad_norm": 1.3505102396011353, "learning_rate": 0.00018969615535418847, "loss": 4.7409, "step": 557050 }, { "epoch": 1.0396100760673572, "grad_norm": 1.2618919610977173, "learning_rate": 0.0001896942979854626, "loss": 4.4529, "step": 557100 }, { "epoch": 1.0397033816267358, "grad_norm": 1.1932255029678345, "learning_rate": 0.0001896924404584421, "loss": 4.6241, "step": 557150 }, { "epoch": 1.0397966871861142, "grad_norm": 1.3994927406311035, "learning_rate": 0.00018969058277313022, "loss": 4.5202, "step": 557200 }, { "epoch": 1.0398899927454928, "grad_norm": 1.1713100671768188, "learning_rate": 0.0001896887249295302, "loss": 4.4738, "step": 557250 }, { "epoch": 1.0399832983048714, "grad_norm": 1.0338475704193115, "learning_rate": 0.0001896868669276454, "loss": 4.7082, "step": 557300 }, { "epoch": 1.0400766038642497, "grad_norm": 1.1523065567016602, "learning_rate": 0.00018968500876747905, "loss": 4.5592, "step": 557350 }, { "epoch": 1.0401699094236283, "grad_norm": 0.8810263872146606, "learning_rate": 0.00018968315044903445, "loss": 4.5853, "step": 557400 }, { "epoch": 1.0402632149830067, "grad_norm": 1.048643708229065, "learning_rate": 0.00018968129197231484, "loss": 4.5802, "step": 557450 }, { "epoch": 1.0403565205423853, "grad_norm": 1.163019061088562, "learning_rate": 0.0001896794333373236, "loss": 4.5207, "step": 557500 }, { "epoch": 1.0404498261017636, "grad_norm": 1.1989173889160156, "learning_rate": 0.0001896775745440639, "loss": 4.775, "step": 557550 }, { "epoch": 1.0405431316611422, "grad_norm": 1.218392252922058, "learning_rate": 0.00018967571559253904, "loss": 4.5981, "step": 557600 }, { "epoch": 1.0406364372205208, "grad_norm": 3.4598512649536133, "learning_rate": 0.00018967385648275233, "loss": 4.6136, "step": 557650 }, { "epoch": 1.0407297427798992, "grad_norm": 1.2795765399932861, "learning_rate": 0.00018967199721470706, "loss": 4.4636, "step": 557700 }, { "epoch": 1.0408230483392777, "grad_norm": 0.908409595489502, "learning_rate": 0.0001896701377884065, "loss": 4.5152, "step": 557750 }, { "epoch": 1.040916353898656, "grad_norm": 1.2649221420288086, "learning_rate": 0.00018966827820385388, "loss": 4.4376, "step": 557800 }, { "epoch": 1.0410096594580347, "grad_norm": 0.7400497794151306, "learning_rate": 0.00018966641846105255, "loss": 4.5623, "step": 557850 }, { "epoch": 1.041102965017413, "grad_norm": 1.1603256464004517, "learning_rate": 0.0001896645585600058, "loss": 4.5737, "step": 557900 }, { "epoch": 1.0411962705767916, "grad_norm": 0.8541489839553833, "learning_rate": 0.00018966269850071682, "loss": 4.6021, "step": 557950 }, { "epoch": 1.0412895761361702, "grad_norm": 1.0671254396438599, "learning_rate": 0.00018966083828318898, "loss": 4.5099, "step": 558000 }, { "epoch": 1.0412895761361702, "eval_loss": 4.686485290527344, "eval_runtime": 231.5073, "eval_samples_per_second": 11.265, "eval_steps_per_second": 11.265, "eval_tts_loss": 7.619232694609834, "step": 558000 }, { "epoch": 1.0413828816955486, "grad_norm": 1.209587812423706, "learning_rate": 0.00018965897790742555, "loss": 4.5227, "step": 558050 }, { "epoch": 1.0414761872549272, "grad_norm": 1.2582651376724243, "learning_rate": 0.0001896571173734298, "loss": 4.7177, "step": 558100 }, { "epoch": 1.0415694928143056, "grad_norm": 1.1203304529190063, "learning_rate": 0.00018965525668120502, "loss": 4.3358, "step": 558150 }, { "epoch": 1.0416627983736841, "grad_norm": 1.3835645914077759, "learning_rate": 0.00018965339583075445, "loss": 4.4995, "step": 558200 }, { "epoch": 1.0417561039330625, "grad_norm": 1.066826581954956, "learning_rate": 0.00018965153482208144, "loss": 4.5885, "step": 558250 }, { "epoch": 1.041849409492441, "grad_norm": 0.9376452565193176, "learning_rate": 0.00018964967365518924, "loss": 4.5685, "step": 558300 }, { "epoch": 1.0419427150518197, "grad_norm": 1.0500354766845703, "learning_rate": 0.00018964781233008117, "loss": 4.589, "step": 558350 }, { "epoch": 1.042036020611198, "grad_norm": 1.278267741203308, "learning_rate": 0.00018964595084676044, "loss": 4.5226, "step": 558400 }, { "epoch": 1.0421293261705766, "grad_norm": 1.0431292057037354, "learning_rate": 0.00018964408920523038, "loss": 4.5272, "step": 558450 }, { "epoch": 1.042222631729955, "grad_norm": 1.1851284503936768, "learning_rate": 0.00018964222740549432, "loss": 4.6326, "step": 558500 }, { "epoch": 1.0423159372893336, "grad_norm": 1.442199468612671, "learning_rate": 0.00018964036544755544, "loss": 4.4882, "step": 558550 }, { "epoch": 1.042409242848712, "grad_norm": 1.006096363067627, "learning_rate": 0.00018963850333141716, "loss": 4.2993, "step": 558600 }, { "epoch": 1.0425025484080905, "grad_norm": 1.1521356105804443, "learning_rate": 0.00018963664105708263, "loss": 4.5155, "step": 558650 }, { "epoch": 1.0425958539674691, "grad_norm": 1.2330262660980225, "learning_rate": 0.00018963477862455523, "loss": 4.4454, "step": 558700 }, { "epoch": 1.0426891595268475, "grad_norm": 0.9843170046806335, "learning_rate": 0.0001896329160338382, "loss": 4.3208, "step": 558750 }, { "epoch": 1.042782465086226, "grad_norm": 1.3005472421646118, "learning_rate": 0.0001896310532849349, "loss": 4.5191, "step": 558800 }, { "epoch": 1.0428757706456044, "grad_norm": 1.2429771423339844, "learning_rate": 0.0001896291903778485, "loss": 4.5921, "step": 558850 }, { "epoch": 1.042969076204983, "grad_norm": 1.1753979921340942, "learning_rate": 0.00018962732731258236, "loss": 4.6101, "step": 558900 }, { "epoch": 1.0430623817643614, "grad_norm": 1.2317447662353516, "learning_rate": 0.00018962546408913977, "loss": 4.3712, "step": 558950 }, { "epoch": 1.04315568732374, "grad_norm": 1.6157867908477783, "learning_rate": 0.00018962360070752402, "loss": 4.4975, "step": 559000 }, { "epoch": 1.0432489928831186, "grad_norm": 1.175130844116211, "learning_rate": 0.00018962173716773837, "loss": 4.7446, "step": 559050 }, { "epoch": 1.043342298442497, "grad_norm": 1.1215260028839111, "learning_rate": 0.0001896198734697861, "loss": 4.3115, "step": 559100 }, { "epoch": 1.0434356040018755, "grad_norm": 0.9790464639663696, "learning_rate": 0.00018961800961367055, "loss": 4.4471, "step": 559150 }, { "epoch": 1.0435289095612539, "grad_norm": 1.004855990409851, "learning_rate": 0.000189616145599395, "loss": 4.5773, "step": 559200 }, { "epoch": 1.0436222151206325, "grad_norm": 0.9849260449409485, "learning_rate": 0.00018961428142696266, "loss": 4.359, "step": 559250 }, { "epoch": 1.0437155206800108, "grad_norm": 1.3343024253845215, "learning_rate": 0.00018961241709637694, "loss": 4.4028, "step": 559300 }, { "epoch": 1.0438088262393894, "grad_norm": 0.9690690636634827, "learning_rate": 0.00018961055260764104, "loss": 4.5623, "step": 559350 }, { "epoch": 1.043902131798768, "grad_norm": 1.2736104726791382, "learning_rate": 0.0001896086879607583, "loss": 4.6301, "step": 559400 }, { "epoch": 1.0439954373581464, "grad_norm": 1.5723506212234497, "learning_rate": 0.00018960682315573199, "loss": 4.6686, "step": 559450 }, { "epoch": 1.044088742917525, "grad_norm": 1.0768077373504639, "learning_rate": 0.0001896049581925654, "loss": 4.4611, "step": 559500 }, { "epoch": 1.0441820484769033, "grad_norm": 0.940334677696228, "learning_rate": 0.00018960309307126182, "loss": 4.368, "step": 559550 }, { "epoch": 1.044275354036282, "grad_norm": 1.264877438545227, "learning_rate": 0.00018960122779182456, "loss": 4.8875, "step": 559600 }, { "epoch": 1.0443686595956603, "grad_norm": 1.06669282913208, "learning_rate": 0.00018959936235425688, "loss": 4.5684, "step": 559650 }, { "epoch": 1.0444619651550389, "grad_norm": 1.0286917686462402, "learning_rate": 0.0001895974967585621, "loss": 4.5213, "step": 559700 }, { "epoch": 1.0445552707144174, "grad_norm": 1.3345227241516113, "learning_rate": 0.00018959563100474352, "loss": 4.5032, "step": 559750 }, { "epoch": 1.0446485762737958, "grad_norm": 1.071852684020996, "learning_rate": 0.0001895937650928044, "loss": 4.4457, "step": 559800 }, { "epoch": 1.0447418818331744, "grad_norm": 1.2425581216812134, "learning_rate": 0.00018959189902274806, "loss": 4.3231, "step": 559850 }, { "epoch": 1.0448351873925528, "grad_norm": 1.019054889678955, "learning_rate": 0.00018959003279457777, "loss": 4.6841, "step": 559900 }, { "epoch": 1.0449284929519314, "grad_norm": 1.1893304586410522, "learning_rate": 0.00018958816640829684, "loss": 4.7017, "step": 559950 }, { "epoch": 1.0450217985113097, "grad_norm": 1.3405216932296753, "learning_rate": 0.00018958629986390855, "loss": 4.4659, "step": 560000 }, { "epoch": 1.0451151040706883, "grad_norm": 1.1335903406143188, "learning_rate": 0.0001895844331614162, "loss": 4.3849, "step": 560050 }, { "epoch": 1.045208409630067, "grad_norm": 1.388495683670044, "learning_rate": 0.0001895825663008231, "loss": 4.5064, "step": 560100 }, { "epoch": 1.0453017151894453, "grad_norm": 1.1342506408691406, "learning_rate": 0.00018958069928213256, "loss": 4.5683, "step": 560150 }, { "epoch": 1.0453950207488238, "grad_norm": 1.268939733505249, "learning_rate": 0.00018957883210534782, "loss": 4.2853, "step": 560200 }, { "epoch": 1.0454883263082022, "grad_norm": 0.9538944363594055, "learning_rate": 0.0001895769647704722, "loss": 4.4406, "step": 560250 }, { "epoch": 1.0455816318675808, "grad_norm": 1.121256709098816, "learning_rate": 0.000189575097277509, "loss": 4.4468, "step": 560300 }, { "epoch": 1.0456749374269592, "grad_norm": 1.0234959125518799, "learning_rate": 0.00018957322962646152, "loss": 4.5058, "step": 560350 }, { "epoch": 1.0457682429863377, "grad_norm": 0.9980145692825317, "learning_rate": 0.00018957136181733304, "loss": 4.6537, "step": 560400 }, { "epoch": 1.0458615485457163, "grad_norm": 1.3222782611846924, "learning_rate": 0.0001895694938501269, "loss": 4.7943, "step": 560450 }, { "epoch": 1.0459548541050947, "grad_norm": 1.0320600271224976, "learning_rate": 0.0001895676257248463, "loss": 4.6738, "step": 560500 }, { "epoch": 1.0460481596644733, "grad_norm": 0.9366627931594849, "learning_rate": 0.00018956575744149466, "loss": 4.6358, "step": 560550 }, { "epoch": 1.0461414652238517, "grad_norm": 1.1141005754470825, "learning_rate": 0.00018956388900007518, "loss": 4.5083, "step": 560600 }, { "epoch": 1.0462347707832302, "grad_norm": 1.1361972093582153, "learning_rate": 0.00018956202040059124, "loss": 4.5352, "step": 560650 }, { "epoch": 1.0463280763426086, "grad_norm": 1.0605624914169312, "learning_rate": 0.00018956015164304604, "loss": 4.2743, "step": 560700 }, { "epoch": 1.0464213819019872, "grad_norm": 1.068372130393982, "learning_rate": 0.00018955828272744296, "loss": 4.4751, "step": 560750 }, { "epoch": 1.0465146874613658, "grad_norm": 1.1838358640670776, "learning_rate": 0.00018955641365378528, "loss": 4.4285, "step": 560800 }, { "epoch": 1.0466079930207441, "grad_norm": 1.1569998264312744, "learning_rate": 0.00018955454442207628, "loss": 4.3729, "step": 560850 }, { "epoch": 1.0467012985801227, "grad_norm": 1.3380413055419922, "learning_rate": 0.00018955267503231924, "loss": 4.4339, "step": 560900 }, { "epoch": 1.046794604139501, "grad_norm": 1.1564974784851074, "learning_rate": 0.00018955080548451752, "loss": 4.6892, "step": 560950 }, { "epoch": 1.0468879096988797, "grad_norm": 1.1108636856079102, "learning_rate": 0.00018954893577867435, "loss": 4.54, "step": 561000 }, { "epoch": 1.0468879096988797, "eval_loss": 4.693432331085205, "eval_runtime": 231.6978, "eval_samples_per_second": 11.256, "eval_steps_per_second": 11.256, "eval_tts_loss": 7.616177150733952, "step": 561000 }, { "epoch": 1.046981215258258, "grad_norm": 1.0902169942855835, "learning_rate": 0.0001895470659147931, "loss": 4.4799, "step": 561050 }, { "epoch": 1.0470745208176366, "grad_norm": 0.9893404841423035, "learning_rate": 0.00018954519589287704, "loss": 4.5642, "step": 561100 }, { "epoch": 1.0471678263770152, "grad_norm": 1.3182376623153687, "learning_rate": 0.00018954332571292944, "loss": 4.5593, "step": 561150 }, { "epoch": 1.0472611319363936, "grad_norm": 1.230796456336975, "learning_rate": 0.0001895414553749536, "loss": 4.5276, "step": 561200 }, { "epoch": 1.0473544374957722, "grad_norm": 1.0458203554153442, "learning_rate": 0.0001895395848789529, "loss": 4.4789, "step": 561250 }, { "epoch": 1.0474477430551505, "grad_norm": 1.4411078691482544, "learning_rate": 0.00018953771422493057, "loss": 4.4952, "step": 561300 }, { "epoch": 1.0475410486145291, "grad_norm": 1.0232350826263428, "learning_rate": 0.00018953584341288994, "loss": 4.3662, "step": 561350 }, { "epoch": 1.0476343541739075, "grad_norm": 1.2663980722427368, "learning_rate": 0.00018953397244283427, "loss": 4.6556, "step": 561400 }, { "epoch": 1.047727659733286, "grad_norm": 1.2295639514923096, "learning_rate": 0.00018953210131476692, "loss": 4.6358, "step": 561450 }, { "epoch": 1.0478209652926647, "grad_norm": 0.9961820840835571, "learning_rate": 0.00018953023002869115, "loss": 4.5961, "step": 561500 }, { "epoch": 1.047914270852043, "grad_norm": 1.3407602310180664, "learning_rate": 0.00018952835858461028, "loss": 4.4658, "step": 561550 }, { "epoch": 1.0480075764114216, "grad_norm": 1.0445343255996704, "learning_rate": 0.0001895264869825276, "loss": 4.4863, "step": 561600 }, { "epoch": 1.0481008819708, "grad_norm": 1.2506422996520996, "learning_rate": 0.00018952461522244644, "loss": 4.6661, "step": 561650 }, { "epoch": 1.0481941875301786, "grad_norm": 1.6274755001068115, "learning_rate": 0.0001895227433043701, "loss": 4.5066, "step": 561700 }, { "epoch": 1.048287493089557, "grad_norm": 0.9504886269569397, "learning_rate": 0.0001895208712283018, "loss": 4.4973, "step": 561750 }, { "epoch": 1.0483807986489355, "grad_norm": 0.9459189176559448, "learning_rate": 0.000189518998994245, "loss": 4.4917, "step": 561800 }, { "epoch": 1.048474104208314, "grad_norm": 1.1334755420684814, "learning_rate": 0.0001895171266022029, "loss": 4.4682, "step": 561850 }, { "epoch": 1.0485674097676925, "grad_norm": 1.276906967163086, "learning_rate": 0.00018951525405217878, "loss": 4.5799, "step": 561900 }, { "epoch": 1.048660715327071, "grad_norm": 1.2976340055465698, "learning_rate": 0.00018951338134417603, "loss": 4.378, "step": 561950 }, { "epoch": 1.0487540208864494, "grad_norm": 1.3985769748687744, "learning_rate": 0.0001895115084781979, "loss": 4.6261, "step": 562000 }, { "epoch": 1.048847326445828, "grad_norm": 1.263553261756897, "learning_rate": 0.00018950963545424768, "loss": 4.543, "step": 562050 }, { "epoch": 1.0489406320052064, "grad_norm": 1.1268646717071533, "learning_rate": 0.00018950776227232873, "loss": 4.5693, "step": 562100 }, { "epoch": 1.049033937564585, "grad_norm": 0.989986777305603, "learning_rate": 0.00018950588893244433, "loss": 4.389, "step": 562150 }, { "epoch": 1.0491272431239633, "grad_norm": 1.0866034030914307, "learning_rate": 0.00018950401543459778, "loss": 4.3803, "step": 562200 }, { "epoch": 1.049220548683342, "grad_norm": 0.9227002263069153, "learning_rate": 0.00018950214177879238, "loss": 4.5133, "step": 562250 }, { "epoch": 1.0493138542427205, "grad_norm": 0.9953296184539795, "learning_rate": 0.00018950026796503146, "loss": 4.723, "step": 562300 }, { "epoch": 1.0494071598020989, "grad_norm": 0.8850206136703491, "learning_rate": 0.00018949839399331832, "loss": 4.6411, "step": 562350 }, { "epoch": 1.0495004653614775, "grad_norm": 1.1648188829421997, "learning_rate": 0.00018949651986365626, "loss": 4.5264, "step": 562400 }, { "epoch": 1.0495937709208558, "grad_norm": 0.9056611657142639, "learning_rate": 0.00018949464557604858, "loss": 4.4404, "step": 562450 }, { "epoch": 1.0496870764802344, "grad_norm": 0.9841514825820923, "learning_rate": 0.0001894927711304986, "loss": 4.3411, "step": 562500 }, { "epoch": 1.0497803820396128, "grad_norm": 1.2094756364822388, "learning_rate": 0.00018949089652700964, "loss": 4.6422, "step": 562550 }, { "epoch": 1.0498736875989914, "grad_norm": 1.14655601978302, "learning_rate": 0.00018948902176558498, "loss": 4.604, "step": 562600 }, { "epoch": 1.04996699315837, "grad_norm": 0.940509557723999, "learning_rate": 0.00018948714684622795, "loss": 4.5093, "step": 562650 }, { "epoch": 1.0500602987177483, "grad_norm": 1.0163979530334473, "learning_rate": 0.00018948527176894183, "loss": 4.4825, "step": 562700 }, { "epoch": 1.050153604277127, "grad_norm": 0.9425746202468872, "learning_rate": 0.00018948339653372998, "loss": 4.3171, "step": 562750 }, { "epoch": 1.0502469098365053, "grad_norm": 0.961168646812439, "learning_rate": 0.00018948152114059568, "loss": 4.6963, "step": 562800 }, { "epoch": 1.0503402153958838, "grad_norm": 0.8502451777458191, "learning_rate": 0.00018947964558954224, "loss": 4.3919, "step": 562850 }, { "epoch": 1.0504335209552622, "grad_norm": 1.1700360774993896, "learning_rate": 0.00018947776988057296, "loss": 4.6709, "step": 562900 }, { "epoch": 1.0505268265146408, "grad_norm": 1.3634721040725708, "learning_rate": 0.00018947589401369116, "loss": 4.5583, "step": 562950 }, { "epoch": 1.0506201320740194, "grad_norm": 1.1737244129180908, "learning_rate": 0.00018947401798890013, "loss": 4.3817, "step": 563000 }, { "epoch": 1.0507134376333978, "grad_norm": 1.0025185346603394, "learning_rate": 0.00018947214180620322, "loss": 4.6577, "step": 563050 }, { "epoch": 1.0508067431927763, "grad_norm": 1.3499510288238525, "learning_rate": 0.00018947026546560372, "loss": 4.5578, "step": 563100 }, { "epoch": 1.0509000487521547, "grad_norm": 2.3306143283843994, "learning_rate": 0.00018946838896710498, "loss": 4.4672, "step": 563150 }, { "epoch": 1.0509933543115333, "grad_norm": 1.2126907110214233, "learning_rate": 0.00018946651231071022, "loss": 4.6436, "step": 563200 }, { "epoch": 1.0510866598709117, "grad_norm": 1.1317464113235474, "learning_rate": 0.00018946463549642285, "loss": 4.6496, "step": 563250 }, { "epoch": 1.0511799654302902, "grad_norm": 1.2876487970352173, "learning_rate": 0.0001894627585242461, "loss": 4.4664, "step": 563300 }, { "epoch": 1.0512732709896688, "grad_norm": 0.8973978161811829, "learning_rate": 0.00018946088139418335, "loss": 4.2282, "step": 563350 }, { "epoch": 1.0513665765490472, "grad_norm": 1.2594698667526245, "learning_rate": 0.0001894590041062379, "loss": 4.7603, "step": 563400 }, { "epoch": 1.0514598821084258, "grad_norm": 0.948663592338562, "learning_rate": 0.00018945712666041303, "loss": 4.4227, "step": 563450 }, { "epoch": 1.0515531876678041, "grad_norm": 1.5037624835968018, "learning_rate": 0.00018945524905671207, "loss": 4.4737, "step": 563500 }, { "epoch": 1.0516464932271827, "grad_norm": 1.2594573497772217, "learning_rate": 0.00018945337129513834, "loss": 4.7468, "step": 563550 }, { "epoch": 1.051739798786561, "grad_norm": 1.0872416496276855, "learning_rate": 0.00018945149337569516, "loss": 4.6402, "step": 563600 }, { "epoch": 1.0518331043459397, "grad_norm": 1.3019628524780273, "learning_rate": 0.0001894496152983858, "loss": 4.7495, "step": 563650 }, { "epoch": 1.0519264099053183, "grad_norm": 0.8815997242927551, "learning_rate": 0.00018944773706321364, "loss": 4.4985, "step": 563700 }, { "epoch": 1.0520197154646966, "grad_norm": 1.0509639978408813, "learning_rate": 0.00018944585867018194, "loss": 4.5311, "step": 563750 }, { "epoch": 1.0521130210240752, "grad_norm": 1.4877582788467407, "learning_rate": 0.00018944398011929404, "loss": 4.5572, "step": 563800 }, { "epoch": 1.0522063265834536, "grad_norm": 1.2185157537460327, "learning_rate": 0.00018944210141055327, "loss": 4.6767, "step": 563850 }, { "epoch": 1.0522996321428322, "grad_norm": 0.8431780934333801, "learning_rate": 0.00018944022254396291, "loss": 4.5187, "step": 563900 }, { "epoch": 1.0523929377022105, "grad_norm": 0.9985096454620361, "learning_rate": 0.00018943834351952633, "loss": 4.5292, "step": 563950 }, { "epoch": 1.0524862432615891, "grad_norm": 0.9899874329566956, "learning_rate": 0.0001894364643372468, "loss": 4.5206, "step": 564000 }, { "epoch": 1.0524862432615891, "eval_loss": 4.686426639556885, "eval_runtime": 231.3036, "eval_samples_per_second": 11.275, "eval_steps_per_second": 11.275, "eval_tts_loss": 7.612957475466962, "step": 564000 }, { "epoch": 1.0525795488209677, "grad_norm": 1.1439553499221802, "learning_rate": 0.0001894345849971276, "loss": 4.542, "step": 564050 }, { "epoch": 1.052672854380346, "grad_norm": 1.119964838027954, "learning_rate": 0.00018943270549917214, "loss": 4.3492, "step": 564100 }, { "epoch": 1.0527661599397247, "grad_norm": 1.3427046537399292, "learning_rate": 0.00018943082584338368, "loss": 4.5136, "step": 564150 }, { "epoch": 1.052859465499103, "grad_norm": 1.234615445137024, "learning_rate": 0.00018942894602976554, "loss": 4.6634, "step": 564200 }, { "epoch": 1.0529527710584816, "grad_norm": 1.080623745918274, "learning_rate": 0.00018942706605832105, "loss": 4.5613, "step": 564250 }, { "epoch": 1.05304607661786, "grad_norm": 1.2452552318572998, "learning_rate": 0.00018942518592905355, "loss": 4.6778, "step": 564300 }, { "epoch": 1.0531393821772386, "grad_norm": 1.4397739171981812, "learning_rate": 0.00018942330564196628, "loss": 4.5713, "step": 564350 }, { "epoch": 1.0532326877366172, "grad_norm": 1.0161422491073608, "learning_rate": 0.00018942142519706265, "loss": 4.4963, "step": 564400 }, { "epoch": 1.0533259932959955, "grad_norm": 1.0763418674468994, "learning_rate": 0.00018941954459434593, "loss": 4.5815, "step": 564450 }, { "epoch": 1.053419298855374, "grad_norm": 1.0191566944122314, "learning_rate": 0.00018941766383381945, "loss": 4.6922, "step": 564500 }, { "epoch": 1.0535126044147525, "grad_norm": 1.1929208040237427, "learning_rate": 0.00018941578291548653, "loss": 4.5326, "step": 564550 }, { "epoch": 1.053605909974131, "grad_norm": 0.7376588582992554, "learning_rate": 0.00018941390183935047, "loss": 4.535, "step": 564600 }, { "epoch": 1.0536992155335094, "grad_norm": 1.2888697385787964, "learning_rate": 0.00018941202060541461, "loss": 4.5828, "step": 564650 }, { "epoch": 1.053792521092888, "grad_norm": 1.218809962272644, "learning_rate": 0.00018941013921368227, "loss": 4.4226, "step": 564700 }, { "epoch": 1.0538858266522666, "grad_norm": 1.1503347158432007, "learning_rate": 0.00018940825766415676, "loss": 4.5507, "step": 564750 }, { "epoch": 1.053979132211645, "grad_norm": 1.1058427095413208, "learning_rate": 0.0001894063759568414, "loss": 4.5025, "step": 564800 }, { "epoch": 1.0540724377710236, "grad_norm": 1.0686490535736084, "learning_rate": 0.00018940449409173956, "loss": 4.6626, "step": 564850 }, { "epoch": 1.054165743330402, "grad_norm": 0.9782431125640869, "learning_rate": 0.00018940261206885448, "loss": 4.5496, "step": 564900 }, { "epoch": 1.0542590488897805, "grad_norm": 1.3484246730804443, "learning_rate": 0.00018940072988818954, "loss": 4.5361, "step": 564950 }, { "epoch": 1.0543523544491589, "grad_norm": 0.9627152681350708, "learning_rate": 0.00018939884754974804, "loss": 4.5018, "step": 565000 }, { "epoch": 1.0544456600085375, "grad_norm": 1.2971142530441284, "learning_rate": 0.00018939696505353328, "loss": 4.6426, "step": 565050 }, { "epoch": 1.054538965567916, "grad_norm": 0.9498640298843384, "learning_rate": 0.00018939508239954864, "loss": 4.5059, "step": 565100 }, { "epoch": 1.0546322711272944, "grad_norm": 1.2256988286972046, "learning_rate": 0.00018939319958779737, "loss": 4.6414, "step": 565150 }, { "epoch": 1.054725576686673, "grad_norm": 1.24186110496521, "learning_rate": 0.00018939131661828288, "loss": 4.5398, "step": 565200 }, { "epoch": 1.0548188822460514, "grad_norm": 1.1320310831069946, "learning_rate": 0.00018938943349100843, "loss": 4.3676, "step": 565250 }, { "epoch": 1.05491218780543, "grad_norm": 1.145412564277649, "learning_rate": 0.00018938755020597733, "loss": 4.4686, "step": 565300 }, { "epoch": 1.0550054933648083, "grad_norm": 0.9532806873321533, "learning_rate": 0.00018938566676319295, "loss": 4.5023, "step": 565350 }, { "epoch": 1.055098798924187, "grad_norm": 1.1288763284683228, "learning_rate": 0.00018938378316265858, "loss": 4.6466, "step": 565400 }, { "epoch": 1.0551921044835655, "grad_norm": 1.0762965679168701, "learning_rate": 0.0001893818994043776, "loss": 4.4394, "step": 565450 }, { "epoch": 1.0552854100429438, "grad_norm": 1.1796870231628418, "learning_rate": 0.00018938001548835328, "loss": 4.4954, "step": 565500 }, { "epoch": 1.0553787156023224, "grad_norm": 0.989150881767273, "learning_rate": 0.00018937813141458896, "loss": 4.5215, "step": 565550 }, { "epoch": 1.0554720211617008, "grad_norm": 1.2503886222839355, "learning_rate": 0.00018937624718308796, "loss": 4.7107, "step": 565600 }, { "epoch": 1.0555653267210794, "grad_norm": 1.2224088907241821, "learning_rate": 0.0001893743627938536, "loss": 4.6226, "step": 565650 }, { "epoch": 1.0556586322804578, "grad_norm": 0.9587731957435608, "learning_rate": 0.0001893724782468892, "loss": 4.4826, "step": 565700 }, { "epoch": 1.0557519378398363, "grad_norm": 0.9316986799240112, "learning_rate": 0.00018937059354219812, "loss": 4.4279, "step": 565750 }, { "epoch": 1.055845243399215, "grad_norm": 1.5047273635864258, "learning_rate": 0.0001893687086797837, "loss": 4.4956, "step": 565800 }, { "epoch": 1.0559385489585933, "grad_norm": 1.2201826572418213, "learning_rate": 0.0001893668236596492, "loss": 4.2174, "step": 565850 }, { "epoch": 1.0560318545179719, "grad_norm": 1.4923267364501953, "learning_rate": 0.00018936493848179797, "loss": 4.5561, "step": 565900 }, { "epoch": 1.0561251600773502, "grad_norm": 0.969534695148468, "learning_rate": 0.00018936305314623338, "loss": 4.6218, "step": 565950 }, { "epoch": 1.0562184656367288, "grad_norm": 0.879870593547821, "learning_rate": 0.00018936116765295871, "loss": 4.5009, "step": 566000 }, { "epoch": 1.0563117711961072, "grad_norm": 1.0609794855117798, "learning_rate": 0.0001893592820019773, "loss": 4.6316, "step": 566050 }, { "epoch": 1.0564050767554858, "grad_norm": 1.0965336561203003, "learning_rate": 0.0001893573961932925, "loss": 4.5241, "step": 566100 }, { "epoch": 1.0564983823148644, "grad_norm": 1.243441104888916, "learning_rate": 0.00018935551022690756, "loss": 4.4447, "step": 566150 }, { "epoch": 1.0565916878742427, "grad_norm": 0.9869188070297241, "learning_rate": 0.00018935362410282594, "loss": 4.6915, "step": 566200 }, { "epoch": 1.0566849934336213, "grad_norm": 1.1112807989120483, "learning_rate": 0.00018935173782105085, "loss": 4.2601, "step": 566250 }, { "epoch": 1.0567782989929997, "grad_norm": 1.0017385482788086, "learning_rate": 0.00018934985138158568, "loss": 4.4507, "step": 566300 }, { "epoch": 1.0568716045523783, "grad_norm": 1.1273045539855957, "learning_rate": 0.00018934796478443374, "loss": 4.5472, "step": 566350 }, { "epoch": 1.0569649101117566, "grad_norm": 1.3804188966751099, "learning_rate": 0.00018934607802959837, "loss": 4.659, "step": 566400 }, { "epoch": 1.0570582156711352, "grad_norm": 1.1311862468719482, "learning_rate": 0.0001893441911170829, "loss": 4.6027, "step": 566450 }, { "epoch": 1.0571515212305138, "grad_norm": 0.6543911099433899, "learning_rate": 0.00018934230404689064, "loss": 4.6203, "step": 566500 }, { "epoch": 1.0572448267898922, "grad_norm": 0.809384822845459, "learning_rate": 0.00018934041681902493, "loss": 4.221, "step": 566550 }, { "epoch": 1.0573381323492708, "grad_norm": 1.2838523387908936, "learning_rate": 0.0001893385294334891, "loss": 4.5211, "step": 566600 }, { "epoch": 1.0574314379086491, "grad_norm": 1.2145520448684692, "learning_rate": 0.0001893366418902865, "loss": 4.5017, "step": 566650 }, { "epoch": 1.0575247434680277, "grad_norm": 1.218761920928955, "learning_rate": 0.00018933475418942043, "loss": 4.5196, "step": 566700 }, { "epoch": 1.057618049027406, "grad_norm": 1.276841163635254, "learning_rate": 0.00018933286633089427, "loss": 4.6093, "step": 566750 }, { "epoch": 1.0577113545867847, "grad_norm": 0.9276838898658752, "learning_rate": 0.00018933097831471128, "loss": 4.5849, "step": 566800 }, { "epoch": 1.0578046601461633, "grad_norm": 1.2034715414047241, "learning_rate": 0.00018932909014087484, "loss": 4.6211, "step": 566850 }, { "epoch": 1.0578979657055416, "grad_norm": 1.273783564567566, "learning_rate": 0.0001893272018093883, "loss": 4.3408, "step": 566900 }, { "epoch": 1.0579912712649202, "grad_norm": 1.0312716960906982, "learning_rate": 0.00018932531332025495, "loss": 4.3679, "step": 566950 }, { "epoch": 1.0580845768242986, "grad_norm": 1.303149700164795, "learning_rate": 0.00018932342467347814, "loss": 4.4394, "step": 567000 }, { "epoch": 1.0580845768242986, "eval_loss": 4.698708534240723, "eval_runtime": 230.5526, "eval_samples_per_second": 11.312, "eval_steps_per_second": 11.312, "eval_tts_loss": 7.657847614784616, "step": 567000 }, { "epoch": 1.0581778823836772, "grad_norm": 1.153511643409729, "learning_rate": 0.0001893215358690612, "loss": 4.5542, "step": 567050 }, { "epoch": 1.0582711879430555, "grad_norm": 1.3226401805877686, "learning_rate": 0.00018931964690700748, "loss": 4.5609, "step": 567100 }, { "epoch": 1.0583644935024341, "grad_norm": 1.2994920015335083, "learning_rate": 0.00018931775778732027, "loss": 4.5782, "step": 567150 }, { "epoch": 1.0584577990618127, "grad_norm": 1.181994080543518, "learning_rate": 0.00018931586851000297, "loss": 4.4875, "step": 567200 }, { "epoch": 1.058551104621191, "grad_norm": 1.1682064533233643, "learning_rate": 0.00018931397907505886, "loss": 4.4848, "step": 567250 }, { "epoch": 1.0586444101805697, "grad_norm": 1.275238037109375, "learning_rate": 0.0001893120894824913, "loss": 4.562, "step": 567300 }, { "epoch": 1.058737715739948, "grad_norm": 1.1859787702560425, "learning_rate": 0.0001893101997323036, "loss": 4.3556, "step": 567350 }, { "epoch": 1.0588310212993266, "grad_norm": 0.8065037727355957, "learning_rate": 0.00018930830982449913, "loss": 4.5912, "step": 567400 }, { "epoch": 1.058924326858705, "grad_norm": 0.8727437257766724, "learning_rate": 0.00018930641975908118, "loss": 4.4405, "step": 567450 }, { "epoch": 1.0590176324180836, "grad_norm": 0.9602881073951721, "learning_rate": 0.00018930452953605314, "loss": 4.4808, "step": 567500 }, { "epoch": 1.0591109379774621, "grad_norm": 1.3725870847702026, "learning_rate": 0.00018930263915541831, "loss": 4.5599, "step": 567550 }, { "epoch": 1.0592042435368405, "grad_norm": 1.0245697498321533, "learning_rate": 0.00018930074861718, "loss": 4.4973, "step": 567600 }, { "epoch": 1.059297549096219, "grad_norm": 1.4609979391098022, "learning_rate": 0.00018929885792134162, "loss": 4.578, "step": 567650 }, { "epoch": 1.0593908546555975, "grad_norm": 1.2206724882125854, "learning_rate": 0.00018929696706790648, "loss": 4.6017, "step": 567700 }, { "epoch": 1.059484160214976, "grad_norm": 1.2589075565338135, "learning_rate": 0.00018929507605687791, "loss": 4.413, "step": 567750 }, { "epoch": 1.0595774657743544, "grad_norm": 1.227514624595642, "learning_rate": 0.00018929318488825924, "loss": 4.4581, "step": 567800 }, { "epoch": 1.059670771333733, "grad_norm": 1.1442242860794067, "learning_rate": 0.00018929129356205375, "loss": 4.6822, "step": 567850 }, { "epoch": 1.0597640768931116, "grad_norm": 1.113810658454895, "learning_rate": 0.0001892894020782649, "loss": 4.4874, "step": 567900 }, { "epoch": 1.05985738245249, "grad_norm": 1.0505949258804321, "learning_rate": 0.00018928751043689594, "loss": 4.4293, "step": 567950 }, { "epoch": 1.0599506880118685, "grad_norm": 1.2758307456970215, "learning_rate": 0.00018928561863795025, "loss": 4.5354, "step": 568000 }, { "epoch": 1.060043993571247, "grad_norm": 1.05423104763031, "learning_rate": 0.00018928372668143112, "loss": 4.6212, "step": 568050 }, { "epoch": 1.0601372991306255, "grad_norm": 1.3656326532363892, "learning_rate": 0.00018928183456734196, "loss": 4.7716, "step": 568100 }, { "epoch": 1.0602306046900039, "grad_norm": 0.8524179458618164, "learning_rate": 0.00018927994229568607, "loss": 4.4383, "step": 568150 }, { "epoch": 1.0603239102493824, "grad_norm": 1.1355934143066406, "learning_rate": 0.00018927804986646678, "loss": 4.4317, "step": 568200 }, { "epoch": 1.060417215808761, "grad_norm": 1.0285577774047852, "learning_rate": 0.00018927615727968745, "loss": 4.7166, "step": 568250 }, { "epoch": 1.0605105213681394, "grad_norm": 1.4209928512573242, "learning_rate": 0.0001892742645353514, "loss": 4.4352, "step": 568300 }, { "epoch": 1.060603826927518, "grad_norm": 1.2423133850097656, "learning_rate": 0.00018927237163346197, "loss": 4.6223, "step": 568350 }, { "epoch": 1.0606971324868963, "grad_norm": 1.1761606931686401, "learning_rate": 0.00018927047857402254, "loss": 4.4365, "step": 568400 }, { "epoch": 1.060790438046275, "grad_norm": 0.6626602411270142, "learning_rate": 0.0001892685853570364, "loss": 4.546, "step": 568450 }, { "epoch": 1.0608837436056533, "grad_norm": 1.3304082155227661, "learning_rate": 0.00018926669198250694, "loss": 4.6591, "step": 568500 }, { "epoch": 1.0609770491650319, "grad_norm": 0.7672298550605774, "learning_rate": 0.00018926479845043746, "loss": 4.7461, "step": 568550 }, { "epoch": 1.0610703547244105, "grad_norm": 1.1396422386169434, "learning_rate": 0.0001892629047608313, "loss": 4.5578, "step": 568600 }, { "epoch": 1.0611636602837888, "grad_norm": 0.8128007054328918, "learning_rate": 0.00018926101091369183, "loss": 4.5099, "step": 568650 }, { "epoch": 1.0612569658431674, "grad_norm": 0.9942769408226013, "learning_rate": 0.0001892591169090224, "loss": 4.3801, "step": 568700 }, { "epoch": 1.0613502714025458, "grad_norm": 1.1350595951080322, "learning_rate": 0.00018925722274682632, "loss": 4.6313, "step": 568750 }, { "epoch": 1.0614435769619244, "grad_norm": 1.2154285907745361, "learning_rate": 0.00018925532842710692, "loss": 4.492, "step": 568800 }, { "epoch": 1.0615368825213027, "grad_norm": 0.9769437313079834, "learning_rate": 0.00018925343394986764, "loss": 4.5581, "step": 568850 }, { "epoch": 1.0616301880806813, "grad_norm": 1.6354825496673584, "learning_rate": 0.0001892515393151117, "loss": 4.4714, "step": 568900 }, { "epoch": 1.06172349364006, "grad_norm": 1.0088893175125122, "learning_rate": 0.00018924964452284252, "loss": 4.5976, "step": 568950 }, { "epoch": 1.0618167991994383, "grad_norm": 1.0912436246871948, "learning_rate": 0.0001892477495730634, "loss": 4.4957, "step": 569000 }, { "epoch": 1.0619101047588169, "grad_norm": 1.1202386617660522, "learning_rate": 0.00018924585446577772, "loss": 4.6302, "step": 569050 }, { "epoch": 1.0620034103181952, "grad_norm": 0.908039927482605, "learning_rate": 0.0001892439592009888, "loss": 4.4556, "step": 569100 }, { "epoch": 1.0620967158775738, "grad_norm": 1.0390937328338623, "learning_rate": 0.00018924206377870002, "loss": 4.4428, "step": 569150 }, { "epoch": 1.0621900214369522, "grad_norm": 0.9893091320991516, "learning_rate": 0.00018924016819891465, "loss": 4.3538, "step": 569200 }, { "epoch": 1.0622833269963308, "grad_norm": 1.2612234354019165, "learning_rate": 0.00018923827246163615, "loss": 4.579, "step": 569250 }, { "epoch": 1.0623766325557094, "grad_norm": 1.0196985006332397, "learning_rate": 0.00018923637656686774, "loss": 4.7088, "step": 569300 }, { "epoch": 1.0624699381150877, "grad_norm": 1.2372381687164307, "learning_rate": 0.00018923448051461287, "loss": 4.5659, "step": 569350 }, { "epoch": 1.0625632436744663, "grad_norm": 0.9859127998352051, "learning_rate": 0.00018923258430487484, "loss": 4.6062, "step": 569400 }, { "epoch": 1.0626565492338447, "grad_norm": 1.2294291257858276, "learning_rate": 0.00018923068793765698, "loss": 4.614, "step": 569450 }, { "epoch": 1.0627498547932233, "grad_norm": 1.1495294570922852, "learning_rate": 0.00018922879141296267, "loss": 4.5285, "step": 569500 }, { "epoch": 1.0628431603526016, "grad_norm": 1.0423833131790161, "learning_rate": 0.00018922689473079523, "loss": 4.5193, "step": 569550 }, { "epoch": 1.0629364659119802, "grad_norm": 1.2409894466400146, "learning_rate": 0.00018922499789115804, "loss": 4.5605, "step": 569600 }, { "epoch": 1.0630297714713586, "grad_norm": 1.1442854404449463, "learning_rate": 0.00018922310089405442, "loss": 4.495, "step": 569650 }, { "epoch": 1.0631230770307372, "grad_norm": 0.9948447942733765, "learning_rate": 0.00018922120373948773, "loss": 4.5828, "step": 569700 }, { "epoch": 1.0632163825901157, "grad_norm": 0.9279969930648804, "learning_rate": 0.0001892193064274613, "loss": 4.4418, "step": 569750 }, { "epoch": 1.0633096881494941, "grad_norm": 1.1630713939666748, "learning_rate": 0.00018921740895797847, "loss": 4.4979, "step": 569800 }, { "epoch": 1.0634029937088727, "grad_norm": 1.1626756191253662, "learning_rate": 0.00018921551133104268, "loss": 4.7807, "step": 569850 }, { "epoch": 1.063496299268251, "grad_norm": 1.1936137676239014, "learning_rate": 0.00018921361354665717, "loss": 4.2746, "step": 569900 }, { "epoch": 1.0635896048276297, "grad_norm": 1.2861826419830322, "learning_rate": 0.00018921171560482534, "loss": 4.49, "step": 569950 }, { "epoch": 1.063682910387008, "grad_norm": 0.7344893217086792, "learning_rate": 0.0001892098175055505, "loss": 4.4423, "step": 570000 }, { "epoch": 1.063682910387008, "eval_loss": 4.701704978942871, "eval_runtime": 233.4607, "eval_samples_per_second": 11.171, "eval_steps_per_second": 11.171, "eval_tts_loss": 7.596651152143415, "step": 570000 }, { "epoch": 1.0637762159463866, "grad_norm": 1.3398991823196411, "learning_rate": 0.00018920791924883606, "loss": 4.4902, "step": 570050 }, { "epoch": 1.0638695215057652, "grad_norm": 1.1734261512756348, "learning_rate": 0.00018920602083468533, "loss": 4.6038, "step": 570100 }, { "epoch": 1.0639628270651436, "grad_norm": 1.1737910509109497, "learning_rate": 0.00018920412226310166, "loss": 4.7536, "step": 570150 }, { "epoch": 1.0640561326245221, "grad_norm": 1.2287696599960327, "learning_rate": 0.00018920222353408842, "loss": 4.4979, "step": 570200 }, { "epoch": 1.0641494381839005, "grad_norm": 1.4696321487426758, "learning_rate": 0.00018920032464764895, "loss": 4.49, "step": 570250 }, { "epoch": 1.064242743743279, "grad_norm": 1.2224761247634888, "learning_rate": 0.0001891984256037866, "loss": 4.4263, "step": 570300 }, { "epoch": 1.0643360493026575, "grad_norm": 1.2485299110412598, "learning_rate": 0.00018919652640250474, "loss": 4.6266, "step": 570350 }, { "epoch": 1.064429354862036, "grad_norm": 1.2701716423034668, "learning_rate": 0.0001891946270438067, "loss": 4.4801, "step": 570400 }, { "epoch": 1.0645226604214146, "grad_norm": 1.1892791986465454, "learning_rate": 0.0001891927275276958, "loss": 4.4958, "step": 570450 }, { "epoch": 1.064615965980793, "grad_norm": 1.1193329095840454, "learning_rate": 0.00018919082785417548, "loss": 4.7243, "step": 570500 }, { "epoch": 1.0647092715401716, "grad_norm": 1.059034824371338, "learning_rate": 0.000189188928023249, "loss": 4.5876, "step": 570550 }, { "epoch": 1.06480257709955, "grad_norm": 1.4603407382965088, "learning_rate": 0.00018918702803491982, "loss": 4.5248, "step": 570600 }, { "epoch": 1.0648958826589285, "grad_norm": 1.036028265953064, "learning_rate": 0.00018918512788919118, "loss": 4.4827, "step": 570650 }, { "epoch": 1.064989188218307, "grad_norm": 1.2693250179290771, "learning_rate": 0.00018918322758606647, "loss": 4.856, "step": 570700 }, { "epoch": 1.0650824937776855, "grad_norm": 1.2520971298217773, "learning_rate": 0.00018918132712554908, "loss": 4.6144, "step": 570750 }, { "epoch": 1.065175799337064, "grad_norm": 1.2817808389663696, "learning_rate": 0.00018917942650764234, "loss": 4.3359, "step": 570800 }, { "epoch": 1.0652691048964424, "grad_norm": 0.8007941246032715, "learning_rate": 0.0001891775257323496, "loss": 4.7657, "step": 570850 }, { "epoch": 1.065362410455821, "grad_norm": 1.0290018320083618, "learning_rate": 0.00018917562479967421, "loss": 4.3544, "step": 570900 }, { "epoch": 1.0654557160151994, "grad_norm": 1.0568417310714722, "learning_rate": 0.00018917372370961952, "loss": 4.4777, "step": 570950 }, { "epoch": 1.065549021574578, "grad_norm": 1.193867802619934, "learning_rate": 0.00018917182246218893, "loss": 4.5453, "step": 571000 }, { "epoch": 1.0656423271339563, "grad_norm": 1.4166945219039917, "learning_rate": 0.00018916992105738576, "loss": 4.516, "step": 571050 }, { "epoch": 1.065735632693335, "grad_norm": 1.0618138313293457, "learning_rate": 0.00018916801949521336, "loss": 4.4146, "step": 571100 }, { "epoch": 1.0658289382527135, "grad_norm": 1.2633510828018188, "learning_rate": 0.0001891661177756751, "loss": 4.4727, "step": 571150 }, { "epoch": 1.0659222438120919, "grad_norm": 0.8626559376716614, "learning_rate": 0.0001891642158987743, "loss": 4.6934, "step": 571200 }, { "epoch": 1.0660155493714705, "grad_norm": 1.1052021980285645, "learning_rate": 0.0001891623138645144, "loss": 4.4912, "step": 571250 }, { "epoch": 1.0661088549308488, "grad_norm": 1.2448097467422485, "learning_rate": 0.00018916041167289866, "loss": 4.4474, "step": 571300 }, { "epoch": 1.0662021604902274, "grad_norm": 1.1931850910186768, "learning_rate": 0.00018915850932393048, "loss": 4.4658, "step": 571350 }, { "epoch": 1.0662954660496058, "grad_norm": 1.0170724391937256, "learning_rate": 0.00018915660681761328, "loss": 4.4849, "step": 571400 }, { "epoch": 1.0663887716089844, "grad_norm": 0.9955647587776184, "learning_rate": 0.00018915470415395032, "loss": 4.588, "step": 571450 }, { "epoch": 1.066482077168363, "grad_norm": 1.3579157590866089, "learning_rate": 0.00018915280133294498, "loss": 4.4125, "step": 571500 }, { "epoch": 1.0665753827277413, "grad_norm": 1.0479494333267212, "learning_rate": 0.00018915089835460066, "loss": 4.3795, "step": 571550 }, { "epoch": 1.06666868828712, "grad_norm": 0.9850384593009949, "learning_rate": 0.00018914899521892066, "loss": 4.5749, "step": 571600 }, { "epoch": 1.0667619938464983, "grad_norm": 1.0844939947128296, "learning_rate": 0.00018914709192590837, "loss": 4.4821, "step": 571650 }, { "epoch": 1.0668552994058769, "grad_norm": 1.1848816871643066, "learning_rate": 0.00018914518847556718, "loss": 4.5817, "step": 571700 }, { "epoch": 1.0669486049652552, "grad_norm": 1.259734034538269, "learning_rate": 0.0001891432848679004, "loss": 4.6987, "step": 571750 }, { "epoch": 1.0670419105246338, "grad_norm": 0.9914228320121765, "learning_rate": 0.0001891413811029114, "loss": 4.6188, "step": 571800 }, { "epoch": 1.0671352160840124, "grad_norm": 1.1027096509933472, "learning_rate": 0.00018913947718060358, "loss": 4.6113, "step": 571850 }, { "epoch": 1.0672285216433908, "grad_norm": 1.0420706272125244, "learning_rate": 0.00018913757310098022, "loss": 4.6269, "step": 571900 }, { "epoch": 1.0673218272027694, "grad_norm": 1.0619767904281616, "learning_rate": 0.00018913566886404474, "loss": 4.4794, "step": 571950 }, { "epoch": 1.0674151327621477, "grad_norm": 1.2577030658721924, "learning_rate": 0.0001891337644698005, "loss": 4.606, "step": 572000 }, { "epoch": 1.0675084383215263, "grad_norm": 1.3373936414718628, "learning_rate": 0.00018913185991825085, "loss": 4.4829, "step": 572050 }, { "epoch": 1.0676017438809047, "grad_norm": 1.0902354717254639, "learning_rate": 0.00018912995520939917, "loss": 4.667, "step": 572100 }, { "epoch": 1.0676950494402833, "grad_norm": 1.014407753944397, "learning_rate": 0.00018912805034324874, "loss": 4.324, "step": 572150 }, { "epoch": 1.0677883549996618, "grad_norm": 1.355505108833313, "learning_rate": 0.00018912614531980303, "loss": 4.4543, "step": 572200 }, { "epoch": 1.0678816605590402, "grad_norm": 0.9984519481658936, "learning_rate": 0.00018912424013906536, "loss": 4.6067, "step": 572250 }, { "epoch": 1.0679749661184188, "grad_norm": 1.3497934341430664, "learning_rate": 0.0001891223348010391, "loss": 4.467, "step": 572300 }, { "epoch": 1.0680682716777972, "grad_norm": 1.2496603727340698, "learning_rate": 0.00018912042930572756, "loss": 4.6784, "step": 572350 }, { "epoch": 1.0681615772371758, "grad_norm": 1.0776684284210205, "learning_rate": 0.00018911852365313415, "loss": 4.6097, "step": 572400 }, { "epoch": 1.0682548827965541, "grad_norm": 1.0048322677612305, "learning_rate": 0.00018911661784326225, "loss": 4.5947, "step": 572450 }, { "epoch": 1.0683481883559327, "grad_norm": 1.1588525772094727, "learning_rate": 0.00018911471187611518, "loss": 4.5053, "step": 572500 }, { "epoch": 1.0684414939153113, "grad_norm": 0.9867302179336548, "learning_rate": 0.00018911280575169634, "loss": 4.6019, "step": 572550 }, { "epoch": 1.0685347994746897, "grad_norm": 1.3045872449874878, "learning_rate": 0.00018911089947000906, "loss": 4.359, "step": 572600 }, { "epoch": 1.0686281050340682, "grad_norm": 1.028043270111084, "learning_rate": 0.00018910899303105672, "loss": 4.4614, "step": 572650 }, { "epoch": 1.0687214105934466, "grad_norm": 0.9268425703048706, "learning_rate": 0.00018910708643484273, "loss": 4.6624, "step": 572700 }, { "epoch": 1.0688147161528252, "grad_norm": 0.9128994345664978, "learning_rate": 0.00018910517968137035, "loss": 4.5273, "step": 572750 }, { "epoch": 1.0689080217122036, "grad_norm": 1.3179854154586792, "learning_rate": 0.00018910327277064305, "loss": 4.7219, "step": 572800 }, { "epoch": 1.0690013272715821, "grad_norm": 1.1312317848205566, "learning_rate": 0.00018910136570266412, "loss": 4.5295, "step": 572850 }, { "epoch": 1.0690946328309607, "grad_norm": 1.0971311330795288, "learning_rate": 0.000189099458477437, "loss": 4.4596, "step": 572900 }, { "epoch": 1.069187938390339, "grad_norm": 1.044753074645996, "learning_rate": 0.000189097551094965, "loss": 4.4974, "step": 572950 }, { "epoch": 1.0692812439497177, "grad_norm": 1.0512080192565918, "learning_rate": 0.00018909564355525148, "loss": 4.673, "step": 573000 }, { "epoch": 1.0692812439497177, "eval_loss": 4.688205242156982, "eval_runtime": 231.8606, "eval_samples_per_second": 11.248, "eval_steps_per_second": 11.248, "eval_tts_loss": 7.55570395143297, "step": 573000 }, { "epoch": 1.069374549509096, "grad_norm": 1.2581086158752441, "learning_rate": 0.00018909373585829982, "loss": 4.3785, "step": 573050 }, { "epoch": 1.0694678550684746, "grad_norm": 1.0723134279251099, "learning_rate": 0.0001890918280041134, "loss": 4.3869, "step": 573100 }, { "epoch": 1.069561160627853, "grad_norm": 1.1119645833969116, "learning_rate": 0.0001890899199926956, "loss": 4.4969, "step": 573150 }, { "epoch": 1.0696544661872316, "grad_norm": 1.1463125944137573, "learning_rate": 0.00018908801182404976, "loss": 4.4336, "step": 573200 }, { "epoch": 1.0697477717466102, "grad_norm": 1.3880419731140137, "learning_rate": 0.00018908610349817922, "loss": 4.6439, "step": 573250 }, { "epoch": 1.0698410773059885, "grad_norm": 1.1052860021591187, "learning_rate": 0.00018908419501508742, "loss": 4.5096, "step": 573300 }, { "epoch": 1.0699343828653671, "grad_norm": 1.0410033464431763, "learning_rate": 0.00018908228637477768, "loss": 4.421, "step": 573350 }, { "epoch": 1.0700276884247455, "grad_norm": 1.3432059288024902, "learning_rate": 0.0001890803775772534, "loss": 4.6174, "step": 573400 }, { "epoch": 1.070120993984124, "grad_norm": 1.0269262790679932, "learning_rate": 0.0001890784686225179, "loss": 4.3738, "step": 573450 }, { "epoch": 1.0702142995435024, "grad_norm": 1.1109803915023804, "learning_rate": 0.00018907655951057458, "loss": 4.646, "step": 573500 }, { "epoch": 1.070307605102881, "grad_norm": 1.2947088479995728, "learning_rate": 0.0001890746502414268, "loss": 4.5743, "step": 573550 }, { "epoch": 1.0704009106622596, "grad_norm": 1.2845351696014404, "learning_rate": 0.00018907274081507796, "loss": 4.555, "step": 573600 }, { "epoch": 1.070494216221638, "grad_norm": 1.3687711954116821, "learning_rate": 0.00018907083123153136, "loss": 4.8198, "step": 573650 }, { "epoch": 1.0705875217810166, "grad_norm": 0.9556741118431091, "learning_rate": 0.0001890689214907905, "loss": 4.7267, "step": 573700 }, { "epoch": 1.070680827340395, "grad_norm": 1.3474568128585815, "learning_rate": 0.0001890670115928586, "loss": 4.5759, "step": 573750 }, { "epoch": 1.0707741328997735, "grad_norm": 1.1721830368041992, "learning_rate": 0.0001890651015377391, "loss": 4.591, "step": 573800 }, { "epoch": 1.0708674384591519, "grad_norm": 1.140185832977295, "learning_rate": 0.0001890631913254354, "loss": 4.5562, "step": 573850 }, { "epoch": 1.0709607440185305, "grad_norm": 1.4096794128417969, "learning_rate": 0.0001890612809559508, "loss": 4.5733, "step": 573900 }, { "epoch": 1.071054049577909, "grad_norm": 1.0645264387130737, "learning_rate": 0.00018905937042928872, "loss": 4.5545, "step": 573950 }, { "epoch": 1.0711473551372874, "grad_norm": 1.3177111148834229, "learning_rate": 0.00018905745974545253, "loss": 4.7448, "step": 574000 }, { "epoch": 1.071240660696666, "grad_norm": 0.9302532076835632, "learning_rate": 0.0001890555489044456, "loss": 4.4438, "step": 574050 }, { "epoch": 1.0713339662560444, "grad_norm": 1.2031216621398926, "learning_rate": 0.0001890536379062713, "loss": 4.4179, "step": 574100 }, { "epoch": 1.071427271815423, "grad_norm": 1.4610893726348877, "learning_rate": 0.00018905172675093296, "loss": 4.664, "step": 574150 }, { "epoch": 1.0715205773748013, "grad_norm": 0.8830183148384094, "learning_rate": 0.000189049815438434, "loss": 4.4097, "step": 574200 }, { "epoch": 1.07161388293418, "grad_norm": 0.9919520020484924, "learning_rate": 0.0001890479039687778, "loss": 4.5589, "step": 574250 }, { "epoch": 1.0717071884935585, "grad_norm": 0.9809280633926392, "learning_rate": 0.0001890459923419677, "loss": 4.3398, "step": 574300 }, { "epoch": 1.0718004940529369, "grad_norm": 0.8736907243728638, "learning_rate": 0.00018904408055800715, "loss": 4.2055, "step": 574350 }, { "epoch": 1.0718937996123155, "grad_norm": 1.0092837810516357, "learning_rate": 0.0001890421686168994, "loss": 4.3762, "step": 574400 }, { "epoch": 1.0719871051716938, "grad_norm": 1.091582179069519, "learning_rate": 0.00018904025651864795, "loss": 4.4051, "step": 574450 }, { "epoch": 1.0720804107310724, "grad_norm": 1.1294649839401245, "learning_rate": 0.00018903834426325607, "loss": 4.3804, "step": 574500 }, { "epoch": 1.0721737162904508, "grad_norm": 1.0316120386123657, "learning_rate": 0.00018903643185072717, "loss": 4.6378, "step": 574550 }, { "epoch": 1.0722670218498294, "grad_norm": 1.1106230020523071, "learning_rate": 0.00018903451928106466, "loss": 4.6192, "step": 574600 }, { "epoch": 1.072360327409208, "grad_norm": 1.3156578540802002, "learning_rate": 0.00018903260655427187, "loss": 4.5855, "step": 574650 }, { "epoch": 1.0724536329685863, "grad_norm": 1.1000555753707886, "learning_rate": 0.00018903069367035222, "loss": 4.5062, "step": 574700 }, { "epoch": 1.072546938527965, "grad_norm": 1.1954295635223389, "learning_rate": 0.00018902878062930904, "loss": 4.5308, "step": 574750 }, { "epoch": 1.0726402440873433, "grad_norm": 1.2300238609313965, "learning_rate": 0.00018902686743114578, "loss": 4.3831, "step": 574800 }, { "epoch": 1.0727335496467219, "grad_norm": 1.0065537691116333, "learning_rate": 0.00018902495407586569, "loss": 4.5403, "step": 574850 }, { "epoch": 1.0728268552061002, "grad_norm": 1.3890169858932495, "learning_rate": 0.00018902304056347225, "loss": 4.516, "step": 574900 }, { "epoch": 1.0729201607654788, "grad_norm": 1.1417715549468994, "learning_rate": 0.00018902112689396882, "loss": 4.5275, "step": 574950 }, { "epoch": 1.0730134663248574, "grad_norm": 1.0484029054641724, "learning_rate": 0.00018901921306735875, "loss": 4.666, "step": 575000 }, { "epoch": 1.0731067718842358, "grad_norm": 1.1242634057998657, "learning_rate": 0.00018901729908364542, "loss": 4.5851, "step": 575050 }, { "epoch": 1.0732000774436143, "grad_norm": 0.9719246029853821, "learning_rate": 0.00018901538494283225, "loss": 4.6635, "step": 575100 }, { "epoch": 1.0732933830029927, "grad_norm": 1.0486211776733398, "learning_rate": 0.00018901347064492256, "loss": 4.6035, "step": 575150 }, { "epoch": 1.0733866885623713, "grad_norm": 1.3306796550750732, "learning_rate": 0.00018901155618991977, "loss": 4.3639, "step": 575200 }, { "epoch": 1.0734799941217497, "grad_norm": 1.140561819076538, "learning_rate": 0.00018900964157782726, "loss": 4.57, "step": 575250 }, { "epoch": 1.0735732996811282, "grad_norm": 0.9842890501022339, "learning_rate": 0.00018900772680864838, "loss": 4.3422, "step": 575300 }, { "epoch": 1.0736666052405068, "grad_norm": 0.8395258784294128, "learning_rate": 0.00018900581188238653, "loss": 4.464, "step": 575350 }, { "epoch": 1.0737599107998852, "grad_norm": 0.8899513483047485, "learning_rate": 0.00018900389679904508, "loss": 4.5142, "step": 575400 }, { "epoch": 1.0738532163592638, "grad_norm": 1.1523138284683228, "learning_rate": 0.0001890019815586274, "loss": 4.409, "step": 575450 }, { "epoch": 1.0739465219186421, "grad_norm": 1.2616117000579834, "learning_rate": 0.0001890000661611369, "loss": 4.5631, "step": 575500 }, { "epoch": 1.0740398274780207, "grad_norm": 1.0021190643310547, "learning_rate": 0.00018899815060657694, "loss": 4.4737, "step": 575550 }, { "epoch": 1.074133133037399, "grad_norm": 1.4968012571334839, "learning_rate": 0.0001889962348949509, "loss": 4.528, "step": 575600 }, { "epoch": 1.0742264385967777, "grad_norm": 1.259392261505127, "learning_rate": 0.00018899431902626214, "loss": 4.5494, "step": 575650 }, { "epoch": 1.0743197441561563, "grad_norm": 0.9002834558486938, "learning_rate": 0.0001889924030005141, "loss": 4.3641, "step": 575700 }, { "epoch": 1.0744130497155346, "grad_norm": 1.4466644525527954, "learning_rate": 0.00018899048681771013, "loss": 4.385, "step": 575750 }, { "epoch": 1.0745063552749132, "grad_norm": 1.3530876636505127, "learning_rate": 0.00018898857047785358, "loss": 4.5285, "step": 575800 }, { "epoch": 1.0745996608342916, "grad_norm": 1.1255598068237305, "learning_rate": 0.0001889866539809479, "loss": 4.5639, "step": 575850 }, { "epoch": 1.0746929663936702, "grad_norm": 1.0639228820800781, "learning_rate": 0.0001889847373269964, "loss": 4.6339, "step": 575900 }, { "epoch": 1.0747862719530485, "grad_norm": 1.073298454284668, "learning_rate": 0.0001889828205160025, "loss": 4.735, "step": 575950 }, { "epoch": 1.0748795775124271, "grad_norm": 1.2492073774337769, "learning_rate": 0.0001889809035479696, "loss": 4.4766, "step": 576000 }, { "epoch": 1.0748795775124271, "eval_loss": 4.686487674713135, "eval_runtime": 231.1995, "eval_samples_per_second": 11.28, "eval_steps_per_second": 11.28, "eval_tts_loss": 7.614909810104058, "step": 576000 }, { "epoch": 1.0749728830718057, "grad_norm": 0.8950192332267761, "learning_rate": 0.00018897898642290106, "loss": 4.532, "step": 576050 }, { "epoch": 1.075066188631184, "grad_norm": 1.031599521636963, "learning_rate": 0.00018897706914080024, "loss": 4.5835, "step": 576100 }, { "epoch": 1.0751594941905627, "grad_norm": 1.243776559829712, "learning_rate": 0.00018897515170167054, "loss": 4.5253, "step": 576150 }, { "epoch": 1.075252799749941, "grad_norm": 1.1623725891113281, "learning_rate": 0.0001889732341055154, "loss": 4.4269, "step": 576200 }, { "epoch": 1.0753461053093196, "grad_norm": 0.8873974680900574, "learning_rate": 0.00018897131635233814, "loss": 4.4357, "step": 576250 }, { "epoch": 1.075439410868698, "grad_norm": 1.0815998315811157, "learning_rate": 0.00018896939844214214, "loss": 4.6452, "step": 576300 }, { "epoch": 1.0755327164280766, "grad_norm": 1.1520164012908936, "learning_rate": 0.00018896748037493084, "loss": 4.5987, "step": 576350 }, { "epoch": 1.0756260219874552, "grad_norm": 1.2606395483016968, "learning_rate": 0.00018896556215070755, "loss": 4.6381, "step": 576400 }, { "epoch": 1.0757193275468335, "grad_norm": 1.1047056913375854, "learning_rate": 0.00018896364376947572, "loss": 4.4805, "step": 576450 }, { "epoch": 1.0758126331062121, "grad_norm": 1.3458294868469238, "learning_rate": 0.00018896172523123872, "loss": 4.6052, "step": 576500 }, { "epoch": 1.0759059386655905, "grad_norm": 1.0969464778900146, "learning_rate": 0.00018895980653599992, "loss": 4.4189, "step": 576550 }, { "epoch": 1.075999244224969, "grad_norm": 1.0623421669006348, "learning_rate": 0.00018895788768376272, "loss": 4.6172, "step": 576600 }, { "epoch": 1.0760925497843474, "grad_norm": 1.1569186449050903, "learning_rate": 0.0001889559686745305, "loss": 4.5737, "step": 576650 }, { "epoch": 1.076185855343726, "grad_norm": 1.067020058631897, "learning_rate": 0.00018895404950830663, "loss": 4.5853, "step": 576700 }, { "epoch": 1.0762791609031046, "grad_norm": 1.2207595109939575, "learning_rate": 0.00018895213018509455, "loss": 4.4203, "step": 576750 }, { "epoch": 1.076372466462483, "grad_norm": 1.1723705530166626, "learning_rate": 0.00018895021070489758, "loss": 4.4906, "step": 576800 }, { "epoch": 1.0764657720218616, "grad_norm": 1.1059699058532715, "learning_rate": 0.00018894829106771914, "loss": 4.4407, "step": 576850 }, { "epoch": 1.07655907758124, "grad_norm": 1.095132827758789, "learning_rate": 0.00018894637127356262, "loss": 4.5012, "step": 576900 }, { "epoch": 1.0766523831406185, "grad_norm": 0.9308245182037354, "learning_rate": 0.0001889444513224314, "loss": 4.5184, "step": 576950 }, { "epoch": 1.0767456886999969, "grad_norm": 1.1809437274932861, "learning_rate": 0.0001889425312143289, "loss": 4.2487, "step": 577000 }, { "epoch": 1.0768389942593755, "grad_norm": 1.3206501007080078, "learning_rate": 0.00018894061094925846, "loss": 4.5241, "step": 577050 }, { "epoch": 1.076932299818754, "grad_norm": 1.1436349153518677, "learning_rate": 0.0001889386905272235, "loss": 4.515, "step": 577100 }, { "epoch": 1.0770256053781324, "grad_norm": 1.2456576824188232, "learning_rate": 0.0001889367699482274, "loss": 4.5799, "step": 577150 }, { "epoch": 1.077118910937511, "grad_norm": 1.0375847816467285, "learning_rate": 0.00018893484921227356, "loss": 4.5224, "step": 577200 }, { "epoch": 1.0772122164968894, "grad_norm": 1.290520191192627, "learning_rate": 0.00018893292831936536, "loss": 4.6518, "step": 577250 }, { "epoch": 1.077305522056268, "grad_norm": 1.0556186437606812, "learning_rate": 0.00018893100726950616, "loss": 4.5456, "step": 577300 }, { "epoch": 1.0773988276156463, "grad_norm": 1.2314338684082031, "learning_rate": 0.0001889290860626994, "loss": 4.4664, "step": 577350 }, { "epoch": 1.077492133175025, "grad_norm": 1.1229864358901978, "learning_rate": 0.00018892716469894844, "loss": 4.5107, "step": 577400 }, { "epoch": 1.0775854387344035, "grad_norm": 1.2485672235488892, "learning_rate": 0.00018892524317825672, "loss": 4.6186, "step": 577450 }, { "epoch": 1.0776787442937819, "grad_norm": 1.964287519454956, "learning_rate": 0.00018892332150062755, "loss": 4.3853, "step": 577500 }, { "epoch": 1.0777720498531604, "grad_norm": 1.2745345830917358, "learning_rate": 0.00018892139966606437, "loss": 4.5592, "step": 577550 }, { "epoch": 1.0778653554125388, "grad_norm": 0.9507763385772705, "learning_rate": 0.00018891947767457056, "loss": 4.3154, "step": 577600 }, { "epoch": 1.0779586609719174, "grad_norm": 1.2392348051071167, "learning_rate": 0.00018891755552614953, "loss": 4.3664, "step": 577650 }, { "epoch": 1.0780519665312958, "grad_norm": 1.0969069004058838, "learning_rate": 0.00018891563322080468, "loss": 4.4944, "step": 577700 }, { "epoch": 1.0781452720906743, "grad_norm": 1.100913166999817, "learning_rate": 0.00018891371075853934, "loss": 4.6486, "step": 577750 }, { "epoch": 1.078238577650053, "grad_norm": 0.9712374806404114, "learning_rate": 0.000188911788139357, "loss": 4.3712, "step": 577800 }, { "epoch": 1.0783318832094313, "grad_norm": 1.3886131048202515, "learning_rate": 0.00018890986536326097, "loss": 4.602, "step": 577850 }, { "epoch": 1.0784251887688099, "grad_norm": 1.078465223312378, "learning_rate": 0.00018890794243025464, "loss": 4.3031, "step": 577900 }, { "epoch": 1.0785184943281882, "grad_norm": 1.2014867067337036, "learning_rate": 0.00018890601934034147, "loss": 4.6088, "step": 577950 }, { "epoch": 1.0786117998875668, "grad_norm": 1.033301591873169, "learning_rate": 0.0001889040960935248, "loss": 4.3398, "step": 578000 }, { "epoch": 1.0787051054469452, "grad_norm": 1.4282863140106201, "learning_rate": 0.00018890217268980808, "loss": 4.6616, "step": 578050 }, { "epoch": 1.0787984110063238, "grad_norm": 0.9226025938987732, "learning_rate": 0.00018890024912919463, "loss": 4.5651, "step": 578100 }, { "epoch": 1.0788917165657024, "grad_norm": 0.9606470465660095, "learning_rate": 0.0001888983254116879, "loss": 4.4273, "step": 578150 }, { "epoch": 1.0789850221250807, "grad_norm": 1.0527331829071045, "learning_rate": 0.00018889640153729127, "loss": 4.5268, "step": 578200 }, { "epoch": 1.0790783276844593, "grad_norm": 1.258779525756836, "learning_rate": 0.00018889447750600812, "loss": 4.4745, "step": 578250 }, { "epoch": 1.0791716332438377, "grad_norm": 1.2030781507492065, "learning_rate": 0.00018889255331784187, "loss": 4.4093, "step": 578300 }, { "epoch": 1.0792649388032163, "grad_norm": 1.0498098134994507, "learning_rate": 0.0001888906289727959, "loss": 4.3862, "step": 578350 }, { "epoch": 1.0793582443625946, "grad_norm": 1.1391030550003052, "learning_rate": 0.00018888870447087363, "loss": 4.4565, "step": 578400 }, { "epoch": 1.0794515499219732, "grad_norm": 0.912508487701416, "learning_rate": 0.00018888677981207842, "loss": 4.3698, "step": 578450 }, { "epoch": 1.0795448554813518, "grad_norm": 1.2204711437225342, "learning_rate": 0.00018888485499641368, "loss": 4.5013, "step": 578500 }, { "epoch": 1.0796381610407302, "grad_norm": 1.3718433380126953, "learning_rate": 0.0001888829300238828, "loss": 4.5139, "step": 578550 }, { "epoch": 1.0797314666001088, "grad_norm": 1.3143681287765503, "learning_rate": 0.0001888810048944892, "loss": 4.3366, "step": 578600 }, { "epoch": 1.0798247721594871, "grad_norm": 1.336580514907837, "learning_rate": 0.0001888790796082363, "loss": 4.5896, "step": 578650 }, { "epoch": 1.0799180777188657, "grad_norm": 1.2106866836547852, "learning_rate": 0.00018887715416512742, "loss": 4.507, "step": 578700 }, { "epoch": 1.080011383278244, "grad_norm": 0.9447749853134155, "learning_rate": 0.000188875228565166, "loss": 4.3295, "step": 578750 }, { "epoch": 1.0801046888376227, "grad_norm": 1.3038015365600586, "learning_rate": 0.00018887330280835545, "loss": 4.6924, "step": 578800 }, { "epoch": 1.0801979943970013, "grad_norm": 1.040339469909668, "learning_rate": 0.00018887137689469917, "loss": 4.5064, "step": 578850 }, { "epoch": 1.0802912999563796, "grad_norm": 0.9808902144432068, "learning_rate": 0.00018886945082420055, "loss": 4.4172, "step": 578900 }, { "epoch": 1.0803846055157582, "grad_norm": 1.1392594575881958, "learning_rate": 0.00018886752459686297, "loss": 4.3766, "step": 578950 }, { "epoch": 1.0804779110751366, "grad_norm": 1.2057980298995972, "learning_rate": 0.00018886559821268984, "loss": 4.5364, "step": 579000 }, { "epoch": 1.0804779110751366, "eval_loss": 4.689462184906006, "eval_runtime": 233.0051, "eval_samples_per_second": 11.193, "eval_steps_per_second": 11.193, "eval_tts_loss": 7.593260817983948, "step": 579000 }, { "epoch": 1.0805712166345152, "grad_norm": 1.1113775968551636, "learning_rate": 0.00018886367167168458, "loss": 4.3671, "step": 579050 }, { "epoch": 1.0806645221938935, "grad_norm": 0.9571734666824341, "learning_rate": 0.00018886174497385056, "loss": 4.7656, "step": 579100 }, { "epoch": 1.0807578277532721, "grad_norm": 0.9764923453330994, "learning_rate": 0.00018885981811919122, "loss": 4.6536, "step": 579150 }, { "epoch": 1.0808511333126507, "grad_norm": 1.3090908527374268, "learning_rate": 0.0001888578911077099, "loss": 4.5806, "step": 579200 }, { "epoch": 1.080944438872029, "grad_norm": 1.5141230821609497, "learning_rate": 0.00018885596393941007, "loss": 4.6405, "step": 579250 }, { "epoch": 1.0810377444314077, "grad_norm": 1.00600266456604, "learning_rate": 0.0001888540366142951, "loss": 4.5985, "step": 579300 }, { "epoch": 1.081131049990786, "grad_norm": 1.4517863988876343, "learning_rate": 0.00018885210913236836, "loss": 4.588, "step": 579350 }, { "epoch": 1.0812243555501646, "grad_norm": 1.2888449430465698, "learning_rate": 0.00018885018149363328, "loss": 4.6057, "step": 579400 }, { "epoch": 1.081317661109543, "grad_norm": 0.8514277935028076, "learning_rate": 0.0001888482536980933, "loss": 4.5477, "step": 579450 }, { "epoch": 1.0814109666689216, "grad_norm": 1.1739991903305054, "learning_rate": 0.00018884632574575177, "loss": 4.5784, "step": 579500 }, { "epoch": 1.0815042722283001, "grad_norm": 0.6899278163909912, "learning_rate": 0.0001888443976366121, "loss": 4.2747, "step": 579550 }, { "epoch": 1.0815975777876785, "grad_norm": 1.122205138206482, "learning_rate": 0.0001888424693706777, "loss": 4.4123, "step": 579600 }, { "epoch": 1.081690883347057, "grad_norm": 1.095961332321167, "learning_rate": 0.000188840540947952, "loss": 4.5998, "step": 579650 }, { "epoch": 1.0817841889064355, "grad_norm": 0.8658921718597412, "learning_rate": 0.00018883861236843836, "loss": 4.5311, "step": 579700 }, { "epoch": 1.081877494465814, "grad_norm": 0.9651498198509216, "learning_rate": 0.0001888366836321402, "loss": 4.8198, "step": 579750 }, { "epoch": 1.0819708000251924, "grad_norm": 1.4795972108840942, "learning_rate": 0.00018883475473906094, "loss": 4.6644, "step": 579800 }, { "epoch": 1.082064105584571, "grad_norm": 1.226870059967041, "learning_rate": 0.00018883282568920394, "loss": 4.5217, "step": 579850 }, { "epoch": 1.0821574111439496, "grad_norm": 1.4307820796966553, "learning_rate": 0.00018883089648257263, "loss": 4.5324, "step": 579900 }, { "epoch": 1.082250716703328, "grad_norm": 1.3067591190338135, "learning_rate": 0.00018882896711917043, "loss": 4.5646, "step": 579950 }, { "epoch": 1.0823440222627065, "grad_norm": 1.2040256261825562, "learning_rate": 0.00018882703759900076, "loss": 4.5576, "step": 580000 }, { "epoch": 1.082437327822085, "grad_norm": 1.4318382740020752, "learning_rate": 0.00018882510792206695, "loss": 4.4172, "step": 580050 }, { "epoch": 1.0825306333814635, "grad_norm": 0.9998859167098999, "learning_rate": 0.00018882317808837252, "loss": 4.4954, "step": 580100 }, { "epoch": 1.0826239389408419, "grad_norm": 1.1021161079406738, "learning_rate": 0.00018882124809792075, "loss": 4.4788, "step": 580150 }, { "epoch": 1.0827172445002204, "grad_norm": 1.070255160331726, "learning_rate": 0.00018881931795071514, "loss": 4.5148, "step": 580200 }, { "epoch": 1.082810550059599, "grad_norm": 1.1689507961273193, "learning_rate": 0.00018881738764675902, "loss": 4.6248, "step": 580250 }, { "epoch": 1.0829038556189774, "grad_norm": 1.0038528442382812, "learning_rate": 0.00018881545718605586, "loss": 4.619, "step": 580300 }, { "epoch": 1.082997161178356, "grad_norm": 1.0667141675949097, "learning_rate": 0.00018881352656860908, "loss": 4.6567, "step": 580350 }, { "epoch": 1.0830904667377343, "grad_norm": 0.9516568183898926, "learning_rate": 0.000188811595794422, "loss": 4.5147, "step": 580400 }, { "epoch": 1.083183772297113, "grad_norm": 0.8918102979660034, "learning_rate": 0.00018880966486349812, "loss": 4.7025, "step": 580450 }, { "epoch": 1.0832770778564913, "grad_norm": 1.2027040719985962, "learning_rate": 0.00018880773377584077, "loss": 4.2311, "step": 580500 }, { "epoch": 1.0833703834158699, "grad_norm": 1.007306456565857, "learning_rate": 0.0001888058025314534, "loss": 4.3828, "step": 580550 }, { "epoch": 1.0834636889752485, "grad_norm": 1.1865348815917969, "learning_rate": 0.00018880387113033947, "loss": 4.5679, "step": 580600 }, { "epoch": 1.0835569945346268, "grad_norm": 1.19545316696167, "learning_rate": 0.0001888019395725023, "loss": 4.6476, "step": 580650 }, { "epoch": 1.0836503000940054, "grad_norm": 1.2413691282272339, "learning_rate": 0.00018880000785794531, "loss": 4.4187, "step": 580700 }, { "epoch": 1.0837436056533838, "grad_norm": 1.0268831253051758, "learning_rate": 0.00018879807598667196, "loss": 4.3972, "step": 580750 }, { "epoch": 1.0838369112127624, "grad_norm": 1.0779285430908203, "learning_rate": 0.0001887961439586856, "loss": 4.4792, "step": 580800 }, { "epoch": 1.0839302167721407, "grad_norm": 0.6737936735153198, "learning_rate": 0.0001887942117739897, "loss": 4.293, "step": 580850 }, { "epoch": 1.0840235223315193, "grad_norm": 1.1180171966552734, "learning_rate": 0.00018879227943258763, "loss": 4.547, "step": 580900 }, { "epoch": 1.084116827890898, "grad_norm": 0.926271915435791, "learning_rate": 0.0001887903469344828, "loss": 4.6679, "step": 580950 }, { "epoch": 1.0842101334502763, "grad_norm": 1.0714268684387207, "learning_rate": 0.00018878841427967866, "loss": 4.6272, "step": 581000 }, { "epoch": 1.0843034390096549, "grad_norm": 1.6488090753555298, "learning_rate": 0.00018878648146817856, "loss": 4.4383, "step": 581050 }, { "epoch": 1.0843967445690332, "grad_norm": 0.8951486945152283, "learning_rate": 0.00018878454849998594, "loss": 4.4908, "step": 581100 }, { "epoch": 1.0844900501284118, "grad_norm": 1.033374547958374, "learning_rate": 0.00018878261537510425, "loss": 4.542, "step": 581150 }, { "epoch": 1.0845833556877902, "grad_norm": 1.0562711954116821, "learning_rate": 0.00018878068209353684, "loss": 4.5151, "step": 581200 }, { "epoch": 1.0846766612471688, "grad_norm": 1.549751877784729, "learning_rate": 0.00018877874865528714, "loss": 4.5325, "step": 581250 }, { "epoch": 1.0847699668065474, "grad_norm": 1.0369826555252075, "learning_rate": 0.0001887768150603586, "loss": 4.5153, "step": 581300 }, { "epoch": 1.0848632723659257, "grad_norm": 1.4468578100204468, "learning_rate": 0.00018877488130875455, "loss": 4.6094, "step": 581350 }, { "epoch": 1.0849565779253043, "grad_norm": 1.1167536973953247, "learning_rate": 0.00018877294740047852, "loss": 4.468, "step": 581400 }, { "epoch": 1.0850498834846827, "grad_norm": 1.2444230318069458, "learning_rate": 0.0001887710133355338, "loss": 4.5043, "step": 581450 }, { "epoch": 1.0851431890440613, "grad_norm": 1.1114223003387451, "learning_rate": 0.00018876907911392392, "loss": 4.5614, "step": 581500 }, { "epoch": 1.0852364946034396, "grad_norm": 1.0241403579711914, "learning_rate": 0.00018876714473565218, "loss": 4.3578, "step": 581550 }, { "epoch": 1.0853298001628182, "grad_norm": 0.48926210403442383, "learning_rate": 0.00018876521020072206, "loss": 4.4279, "step": 581600 }, { "epoch": 1.0854231057221968, "grad_norm": 0.9913998246192932, "learning_rate": 0.00018876327550913698, "loss": 4.6899, "step": 581650 }, { "epoch": 1.0855164112815752, "grad_norm": 1.1818612813949585, "learning_rate": 0.00018876134066090033, "loss": 4.5659, "step": 581700 }, { "epoch": 1.0856097168409538, "grad_norm": 0.9715335369110107, "learning_rate": 0.00018875940565601552, "loss": 4.6827, "step": 581750 }, { "epoch": 1.0857030224003321, "grad_norm": 1.0407130718231201, "learning_rate": 0.00018875747049448602, "loss": 4.6666, "step": 581800 }, { "epoch": 1.0857963279597107, "grad_norm": 1.3684943914413452, "learning_rate": 0.00018875553517631513, "loss": 4.5265, "step": 581850 }, { "epoch": 1.085889633519089, "grad_norm": 1.063700795173645, "learning_rate": 0.00018875359970150642, "loss": 4.3624, "step": 581900 }, { "epoch": 1.0859829390784677, "grad_norm": 0.9608325362205505, "learning_rate": 0.00018875166407006318, "loss": 4.3468, "step": 581950 }, { "epoch": 1.0860762446378462, "grad_norm": 1.1114795207977295, "learning_rate": 0.00018874972828198888, "loss": 4.6251, "step": 582000 }, { "epoch": 1.0860762446378462, "eval_loss": 4.693373680114746, "eval_runtime": 230.1869, "eval_samples_per_second": 11.33, "eval_steps_per_second": 11.33, "eval_tts_loss": 7.558666932240894, "step": 582000 }, { "epoch": 1.0861695501972246, "grad_norm": 0.8923165202140808, "learning_rate": 0.0001887477923372869, "loss": 4.3114, "step": 582050 }, { "epoch": 1.0862628557566032, "grad_norm": 0.9820541739463806, "learning_rate": 0.0001887458562359607, "loss": 4.6128, "step": 582100 }, { "epoch": 1.0863561613159816, "grad_norm": 0.9334239363670349, "learning_rate": 0.00018874391997801367, "loss": 4.4906, "step": 582150 }, { "epoch": 1.0864494668753601, "grad_norm": 0.7899730205535889, "learning_rate": 0.00018874198356344922, "loss": 4.6337, "step": 582200 }, { "epoch": 1.0865427724347385, "grad_norm": 0.9432328939437866, "learning_rate": 0.00018874004699227085, "loss": 4.4722, "step": 582250 }, { "epoch": 1.086636077994117, "grad_norm": 1.2875365018844604, "learning_rate": 0.00018873811026448185, "loss": 4.4548, "step": 582300 }, { "epoch": 1.0867293835534957, "grad_norm": 1.1109517812728882, "learning_rate": 0.00018873617338008574, "loss": 4.4669, "step": 582350 }, { "epoch": 1.086822689112874, "grad_norm": 1.1328788995742798, "learning_rate": 0.00018873423633908586, "loss": 4.7073, "step": 582400 }, { "epoch": 1.0869159946722526, "grad_norm": 1.3516120910644531, "learning_rate": 0.00018873229914148567, "loss": 4.3557, "step": 582450 }, { "epoch": 1.087009300231631, "grad_norm": 1.293311595916748, "learning_rate": 0.0001887303617872886, "loss": 4.4948, "step": 582500 }, { "epoch": 1.0871026057910096, "grad_norm": 1.3256378173828125, "learning_rate": 0.00018872842427649804, "loss": 4.5425, "step": 582550 }, { "epoch": 1.087195911350388, "grad_norm": 1.2524538040161133, "learning_rate": 0.00018872648660911746, "loss": 4.669, "step": 582600 }, { "epoch": 1.0872892169097665, "grad_norm": 1.0085126161575317, "learning_rate": 0.0001887245487851502, "loss": 4.631, "step": 582650 }, { "epoch": 1.0873825224691451, "grad_norm": 1.1164655685424805, "learning_rate": 0.00018872261080459974, "loss": 4.3568, "step": 582700 }, { "epoch": 1.0874758280285235, "grad_norm": 0.9210225343704224, "learning_rate": 0.0001887206726674695, "loss": 4.4645, "step": 582750 }, { "epoch": 1.087569133587902, "grad_norm": 1.3512184619903564, "learning_rate": 0.00018871873437376286, "loss": 4.6282, "step": 582800 }, { "epoch": 1.0876624391472804, "grad_norm": 0.9630610942840576, "learning_rate": 0.00018871679592348328, "loss": 4.3637, "step": 582850 }, { "epoch": 1.087755744706659, "grad_norm": 1.1503225564956665, "learning_rate": 0.00018871485731663414, "loss": 4.738, "step": 582900 }, { "epoch": 1.0878490502660374, "grad_norm": 1.0013177394866943, "learning_rate": 0.0001887129185532189, "loss": 4.4143, "step": 582950 }, { "epoch": 1.087942355825416, "grad_norm": 0.9119142293930054, "learning_rate": 0.00018871097963324098, "loss": 4.7743, "step": 583000 }, { "epoch": 1.0880356613847946, "grad_norm": 1.2202430963516235, "learning_rate": 0.0001887090405567038, "loss": 4.5442, "step": 583050 }, { "epoch": 1.088128966944173, "grad_norm": 1.178678274154663, "learning_rate": 0.00018870710132361072, "loss": 4.6334, "step": 583100 }, { "epoch": 1.0882222725035515, "grad_norm": 1.207959532737732, "learning_rate": 0.00018870516193396526, "loss": 4.5701, "step": 583150 }, { "epoch": 1.08831557806293, "grad_norm": 1.0675325393676758, "learning_rate": 0.0001887032223877708, "loss": 4.5585, "step": 583200 }, { "epoch": 1.0884088836223085, "grad_norm": 1.0169501304626465, "learning_rate": 0.00018870128268503073, "loss": 4.7295, "step": 583250 }, { "epoch": 1.0885021891816868, "grad_norm": 1.26615309715271, "learning_rate": 0.00018869934282574852, "loss": 4.5731, "step": 583300 }, { "epoch": 1.0885954947410654, "grad_norm": 0.7592546343803406, "learning_rate": 0.00018869740280992757, "loss": 4.5289, "step": 583350 }, { "epoch": 1.088688800300444, "grad_norm": 1.3602420091629028, "learning_rate": 0.00018869546263757133, "loss": 4.4298, "step": 583400 }, { "epoch": 1.0887821058598224, "grad_norm": 0.957187831401825, "learning_rate": 0.0001886935223086832, "loss": 4.6372, "step": 583450 }, { "epoch": 1.088875411419201, "grad_norm": 1.0717426538467407, "learning_rate": 0.0001886915818232666, "loss": 4.3322, "step": 583500 }, { "epoch": 1.0889687169785793, "grad_norm": 0.8310880661010742, "learning_rate": 0.00018868964118132496, "loss": 4.4749, "step": 583550 }, { "epoch": 1.089062022537958, "grad_norm": 1.0796977281570435, "learning_rate": 0.00018868770038286172, "loss": 4.3962, "step": 583600 }, { "epoch": 1.0891553280973363, "grad_norm": 1.1135965585708618, "learning_rate": 0.00018868575942788027, "loss": 4.5643, "step": 583650 }, { "epoch": 1.0892486336567149, "grad_norm": 1.2277497053146362, "learning_rate": 0.00018868381831638412, "loss": 4.4888, "step": 583700 }, { "epoch": 1.0893419392160935, "grad_norm": 1.264382004737854, "learning_rate": 0.00018868187704837656, "loss": 4.7007, "step": 583750 }, { "epoch": 1.0894352447754718, "grad_norm": 0.8704171180725098, "learning_rate": 0.00018867993562386113, "loss": 4.6002, "step": 583800 }, { "epoch": 1.0895285503348504, "grad_norm": 1.047865390777588, "learning_rate": 0.0001886779940428412, "loss": 4.4613, "step": 583850 }, { "epoch": 1.0896218558942288, "grad_norm": 1.2342171669006348, "learning_rate": 0.00018867605230532024, "loss": 4.2313, "step": 583900 }, { "epoch": 1.0897151614536074, "grad_norm": 1.0370007753372192, "learning_rate": 0.00018867411041130163, "loss": 4.4816, "step": 583950 }, { "epoch": 1.0898084670129857, "grad_norm": 1.1774909496307373, "learning_rate": 0.00018867216836078882, "loss": 4.5652, "step": 584000 }, { "epoch": 1.0899017725723643, "grad_norm": 1.3429521322250366, "learning_rate": 0.00018867022615378525, "loss": 4.7698, "step": 584050 }, { "epoch": 1.089995078131743, "grad_norm": 1.1255038976669312, "learning_rate": 0.00018866828379029431, "loss": 4.4205, "step": 584100 }, { "epoch": 1.0900883836911213, "grad_norm": 1.3047699928283691, "learning_rate": 0.00018866634127031947, "loss": 4.6273, "step": 584150 }, { "epoch": 1.0901816892504999, "grad_norm": 1.1736973524093628, "learning_rate": 0.00018866439859386414, "loss": 4.5509, "step": 584200 }, { "epoch": 1.0902749948098782, "grad_norm": 1.1735835075378418, "learning_rate": 0.00018866245576093174, "loss": 4.6637, "step": 584250 }, { "epoch": 1.0903683003692568, "grad_norm": 0.8955191969871521, "learning_rate": 0.0001886605127715257, "loss": 4.6566, "step": 584300 }, { "epoch": 1.0904616059286352, "grad_norm": 1.6329035758972168, "learning_rate": 0.00018865856962564946, "loss": 4.5303, "step": 584350 }, { "epoch": 1.0905549114880138, "grad_norm": 1.41972017288208, "learning_rate": 0.00018865662632330644, "loss": 4.7551, "step": 584400 }, { "epoch": 1.0906482170473923, "grad_norm": 1.4575518369674683, "learning_rate": 0.00018865468286450007, "loss": 4.5314, "step": 584450 }, { "epoch": 1.0907415226067707, "grad_norm": 1.2344341278076172, "learning_rate": 0.00018865273924923378, "loss": 4.481, "step": 584500 }, { "epoch": 1.0908348281661493, "grad_norm": 1.168646216392517, "learning_rate": 0.00018865079547751104, "loss": 4.5357, "step": 584550 }, { "epoch": 1.0909281337255277, "grad_norm": 1.1109296083450317, "learning_rate": 0.00018864885154933522, "loss": 4.4895, "step": 584600 }, { "epoch": 1.0910214392849062, "grad_norm": 1.0106289386749268, "learning_rate": 0.0001886469074647098, "loss": 4.5116, "step": 584650 }, { "epoch": 1.0911147448442846, "grad_norm": 1.3297501802444458, "learning_rate": 0.00018864496322363813, "loss": 4.5624, "step": 584700 }, { "epoch": 1.0912080504036632, "grad_norm": 1.0867538452148438, "learning_rate": 0.00018864301882612374, "loss": 4.6268, "step": 584750 }, { "epoch": 1.0913013559630416, "grad_norm": 0.8482877612113953, "learning_rate": 0.00018864107427217, "loss": 4.363, "step": 584800 }, { "epoch": 1.0913946615224202, "grad_norm": 1.2616050243377686, "learning_rate": 0.00018863912956178038, "loss": 4.4878, "step": 584850 }, { "epoch": 1.0914879670817987, "grad_norm": 1.1939424276351929, "learning_rate": 0.0001886371846949583, "loss": 4.582, "step": 584900 }, { "epoch": 1.091581272641177, "grad_norm": 1.0855286121368408, "learning_rate": 0.00018863523967170716, "loss": 4.715, "step": 584950 }, { "epoch": 1.0916745782005557, "grad_norm": 1.0112704038619995, "learning_rate": 0.00018863329449203043, "loss": 4.4758, "step": 585000 }, { "epoch": 1.0916745782005557, "eval_loss": 4.697065830230713, "eval_runtime": 233.0089, "eval_samples_per_second": 11.193, "eval_steps_per_second": 11.193, "eval_tts_loss": 7.564821515949386, "step": 585000 }, { "epoch": 1.091767883759934, "grad_norm": 1.2775155305862427, "learning_rate": 0.0001886313491559315, "loss": 4.4099, "step": 585050 }, { "epoch": 1.0918611893193126, "grad_norm": 1.074365496635437, "learning_rate": 0.0001886294036634139, "loss": 4.3246, "step": 585100 }, { "epoch": 1.091954494878691, "grad_norm": 0.9782500863075256, "learning_rate": 0.00018862745801448094, "loss": 4.4717, "step": 585150 }, { "epoch": 1.0920478004380696, "grad_norm": 1.382952332496643, "learning_rate": 0.00018862551220913613, "loss": 4.7916, "step": 585200 }, { "epoch": 1.0921411059974482, "grad_norm": 1.1335961818695068, "learning_rate": 0.0001886235662473829, "loss": 4.5453, "step": 585250 }, { "epoch": 1.0922344115568265, "grad_norm": 1.0161381959915161, "learning_rate": 0.00018862162012922465, "loss": 4.3456, "step": 585300 }, { "epoch": 1.0923277171162051, "grad_norm": 1.0690479278564453, "learning_rate": 0.00018861967385466483, "loss": 4.5245, "step": 585350 }, { "epoch": 1.0924210226755835, "grad_norm": 0.910042941570282, "learning_rate": 0.00018861772742370688, "loss": 4.4669, "step": 585400 }, { "epoch": 1.092514328234962, "grad_norm": 0.9803144335746765, "learning_rate": 0.00018861578083635423, "loss": 4.6377, "step": 585450 }, { "epoch": 1.0926076337943404, "grad_norm": 1.0522170066833496, "learning_rate": 0.00018861383409261032, "loss": 4.6179, "step": 585500 }, { "epoch": 1.092700939353719, "grad_norm": 1.011804461479187, "learning_rate": 0.00018861188719247862, "loss": 4.573, "step": 585550 }, { "epoch": 1.0927942449130976, "grad_norm": 0.9452503323554993, "learning_rate": 0.0001886099401359625, "loss": 4.4762, "step": 585600 }, { "epoch": 1.092887550472476, "grad_norm": 1.1122565269470215, "learning_rate": 0.00018860799292306544, "loss": 4.6818, "step": 585650 }, { "epoch": 1.0929808560318546, "grad_norm": 1.224002480506897, "learning_rate": 0.00018860604555379085, "loss": 4.7558, "step": 585700 }, { "epoch": 1.093074161591233, "grad_norm": 1.038970947265625, "learning_rate": 0.00018860409802814218, "loss": 4.4772, "step": 585750 }, { "epoch": 1.0931674671506115, "grad_norm": 1.2083722352981567, "learning_rate": 0.00018860215034612287, "loss": 4.639, "step": 585800 }, { "epoch": 1.09326077270999, "grad_norm": 0.8522143363952637, "learning_rate": 0.00018860020250773639, "loss": 4.5711, "step": 585850 }, { "epoch": 1.0933540782693685, "grad_norm": 1.3871517181396484, "learning_rate": 0.00018859825451298607, "loss": 4.4683, "step": 585900 }, { "epoch": 1.093447383828747, "grad_norm": 1.2222959995269775, "learning_rate": 0.00018859630636187545, "loss": 4.6149, "step": 585950 }, { "epoch": 1.0935406893881254, "grad_norm": 1.2566468715667725, "learning_rate": 0.00018859435805440797, "loss": 4.4855, "step": 586000 }, { "epoch": 1.093633994947504, "grad_norm": 1.704668641090393, "learning_rate": 0.00018859240959058702, "loss": 4.3871, "step": 586050 }, { "epoch": 1.0937273005068824, "grad_norm": 0.8975063562393188, "learning_rate": 0.00018859046097041606, "loss": 4.4956, "step": 586100 }, { "epoch": 1.093820606066261, "grad_norm": 0.8307956457138062, "learning_rate": 0.00018858851219389852, "loss": 4.2073, "step": 586150 }, { "epoch": 1.0939139116256393, "grad_norm": 1.1458641290664673, "learning_rate": 0.00018858656326103783, "loss": 4.3462, "step": 586200 }, { "epoch": 1.094007217185018, "grad_norm": 0.8480004072189331, "learning_rate": 0.00018858461417183746, "loss": 4.5675, "step": 586250 }, { "epoch": 1.0941005227443965, "grad_norm": 1.3548483848571777, "learning_rate": 0.0001885826649263008, "loss": 4.4286, "step": 586300 }, { "epoch": 1.0941938283037749, "grad_norm": 1.3063297271728516, "learning_rate": 0.00018858071552443139, "loss": 4.5541, "step": 586350 }, { "epoch": 1.0942871338631535, "grad_norm": 1.1184899806976318, "learning_rate": 0.00018857876596623253, "loss": 4.3658, "step": 586400 }, { "epoch": 1.0943804394225318, "grad_norm": 1.1189826726913452, "learning_rate": 0.00018857681625170778, "loss": 4.6489, "step": 586450 }, { "epoch": 1.0944737449819104, "grad_norm": 1.2237969636917114, "learning_rate": 0.0001885748663808605, "loss": 4.7196, "step": 586500 }, { "epoch": 1.0945670505412888, "grad_norm": 1.159197449684143, "learning_rate": 0.00018857291635369417, "loss": 4.6825, "step": 586550 }, { "epoch": 1.0946603561006674, "grad_norm": 1.0210381746292114, "learning_rate": 0.0001885709661702123, "loss": 4.5572, "step": 586600 }, { "epoch": 1.094753661660046, "grad_norm": 1.385698914527893, "learning_rate": 0.00018856901583041817, "loss": 4.5936, "step": 586650 }, { "epoch": 1.0948469672194243, "grad_norm": 1.1756649017333984, "learning_rate": 0.00018856706533431535, "loss": 4.4785, "step": 586700 }, { "epoch": 1.094940272778803, "grad_norm": 1.2283968925476074, "learning_rate": 0.00018856511468190722, "loss": 4.536, "step": 586750 }, { "epoch": 1.0950335783381813, "grad_norm": 1.2564771175384521, "learning_rate": 0.00018856316387319728, "loss": 4.5004, "step": 586800 }, { "epoch": 1.0951268838975599, "grad_norm": 0.9570348858833313, "learning_rate": 0.00018856121290818892, "loss": 4.5217, "step": 586850 }, { "epoch": 1.0952201894569382, "grad_norm": 1.4707893133163452, "learning_rate": 0.0001885592617868856, "loss": 4.671, "step": 586900 }, { "epoch": 1.0953134950163168, "grad_norm": 1.218386173248291, "learning_rate": 0.00018855731050929078, "loss": 4.4667, "step": 586950 }, { "epoch": 1.0954068005756954, "grad_norm": 1.0560225248336792, "learning_rate": 0.0001885553590754079, "loss": 4.4964, "step": 587000 }, { "epoch": 1.0955001061350738, "grad_norm": 0.9806395173072815, "learning_rate": 0.00018855340748524037, "loss": 4.3211, "step": 587050 }, { "epoch": 1.0955934116944523, "grad_norm": 1.2609374523162842, "learning_rate": 0.00018855145573879164, "loss": 4.4346, "step": 587100 }, { "epoch": 1.0956867172538307, "grad_norm": 1.1711134910583496, "learning_rate": 0.0001885495038360652, "loss": 4.5198, "step": 587150 }, { "epoch": 1.0957800228132093, "grad_norm": 1.2640150785446167, "learning_rate": 0.00018854755177706448, "loss": 4.7565, "step": 587200 }, { "epoch": 1.0958733283725877, "grad_norm": 1.1614716053009033, "learning_rate": 0.00018854559956179285, "loss": 4.6088, "step": 587250 }, { "epoch": 1.0959666339319662, "grad_norm": 1.0889352560043335, "learning_rate": 0.0001885436471902539, "loss": 4.4624, "step": 587300 }, { "epoch": 1.0960599394913448, "grad_norm": 1.3929716348648071, "learning_rate": 0.00018854169466245093, "loss": 4.5496, "step": 587350 }, { "epoch": 1.0961532450507232, "grad_norm": 1.1938459873199463, "learning_rate": 0.00018853974197838747, "loss": 4.4909, "step": 587400 }, { "epoch": 1.0962465506101018, "grad_norm": 1.1609278917312622, "learning_rate": 0.00018853778913806695, "loss": 4.5501, "step": 587450 }, { "epoch": 1.0963398561694802, "grad_norm": 1.0980507135391235, "learning_rate": 0.0001885358361414928, "loss": 4.5469, "step": 587500 }, { "epoch": 1.0964331617288587, "grad_norm": 1.3295481204986572, "learning_rate": 0.00018853388298866846, "loss": 4.4308, "step": 587550 }, { "epoch": 1.096526467288237, "grad_norm": 1.0381256341934204, "learning_rate": 0.0001885319296795974, "loss": 4.3905, "step": 587600 }, { "epoch": 1.0966197728476157, "grad_norm": 1.318153977394104, "learning_rate": 0.0001885299762142831, "loss": 4.6108, "step": 587650 }, { "epoch": 1.0967130784069943, "grad_norm": 1.1040711402893066, "learning_rate": 0.00018852802259272894, "loss": 4.3809, "step": 587700 }, { "epoch": 1.0968063839663726, "grad_norm": 1.2281560897827148, "learning_rate": 0.0001885260688149384, "loss": 4.587, "step": 587750 }, { "epoch": 1.0968996895257512, "grad_norm": 0.8822270035743713, "learning_rate": 0.00018852411488091492, "loss": 4.5063, "step": 587800 }, { "epoch": 1.0969929950851296, "grad_norm": 1.032925009727478, "learning_rate": 0.00018852216079066198, "loss": 4.5421, "step": 587850 }, { "epoch": 1.0970863006445082, "grad_norm": 1.2351397275924683, "learning_rate": 0.00018852020654418298, "loss": 4.5586, "step": 587900 }, { "epoch": 1.0971796062038865, "grad_norm": 1.292371392250061, "learning_rate": 0.0001885182521414814, "loss": 4.4969, "step": 587950 }, { "epoch": 1.0972729117632651, "grad_norm": 1.1131508350372314, "learning_rate": 0.00018851629758256068, "loss": 4.4917, "step": 588000 }, { "epoch": 1.0972729117632651, "eval_loss": 4.695961952209473, "eval_runtime": 232.8664, "eval_samples_per_second": 11.2, "eval_steps_per_second": 11.2, "eval_tts_loss": 7.565352336992081, "step": 588000 }, { "epoch": 1.0973662173226437, "grad_norm": 0.9609708189964294, "learning_rate": 0.00018851434286742426, "loss": 4.5145, "step": 588050 }, { "epoch": 1.097459522882022, "grad_norm": 1.0569230318069458, "learning_rate": 0.0001885123879960756, "loss": 4.5441, "step": 588100 }, { "epoch": 1.0975528284414007, "grad_norm": 0.9755051732063293, "learning_rate": 0.00018851043296851816, "loss": 4.4334, "step": 588150 }, { "epoch": 1.097646134000779, "grad_norm": 0.9799525737762451, "learning_rate": 0.00018850847778475535, "loss": 4.6348, "step": 588200 }, { "epoch": 1.0977394395601576, "grad_norm": 1.3120450973510742, "learning_rate": 0.0001885065224447907, "loss": 4.7886, "step": 588250 }, { "epoch": 1.097832745119536, "grad_norm": 1.1959748268127441, "learning_rate": 0.00018850456694862757, "loss": 4.6349, "step": 588300 }, { "epoch": 1.0979260506789146, "grad_norm": 1.049827218055725, "learning_rate": 0.00018850261129626947, "loss": 4.5265, "step": 588350 }, { "epoch": 1.0980193562382932, "grad_norm": 0.9035036563873291, "learning_rate": 0.00018850065548771982, "loss": 4.4249, "step": 588400 }, { "epoch": 1.0981126617976715, "grad_norm": 0.937915027141571, "learning_rate": 0.00018849869952298213, "loss": 4.3879, "step": 588450 }, { "epoch": 1.0982059673570501, "grad_norm": 1.1521128416061401, "learning_rate": 0.00018849674340205974, "loss": 4.5201, "step": 588500 }, { "epoch": 1.0982992729164285, "grad_norm": 1.1332216262817383, "learning_rate": 0.00018849478712495624, "loss": 4.607, "step": 588550 }, { "epoch": 1.098392578475807, "grad_norm": 0.962303638458252, "learning_rate": 0.00018849283069167494, "loss": 4.1662, "step": 588600 }, { "epoch": 1.0984858840351854, "grad_norm": 1.3432306051254272, "learning_rate": 0.0001884908741022194, "loss": 4.7836, "step": 588650 }, { "epoch": 1.098579189594564, "grad_norm": 1.0719109773635864, "learning_rate": 0.00018848891735659303, "loss": 4.4936, "step": 588700 }, { "epoch": 1.0986724951539426, "grad_norm": 1.2913343906402588, "learning_rate": 0.00018848696045479932, "loss": 4.4337, "step": 588750 }, { "epoch": 1.098765800713321, "grad_norm": 1.0295428037643433, "learning_rate": 0.00018848500339684168, "loss": 4.5423, "step": 588800 }, { "epoch": 1.0988591062726996, "grad_norm": 1.0688557624816895, "learning_rate": 0.00018848304618272356, "loss": 4.4943, "step": 588850 }, { "epoch": 1.098952411832078, "grad_norm": 1.0393550395965576, "learning_rate": 0.00018848108881244844, "loss": 4.4499, "step": 588900 }, { "epoch": 1.0990457173914565, "grad_norm": 1.3869493007659912, "learning_rate": 0.00018847913128601978, "loss": 4.4441, "step": 588950 }, { "epoch": 1.0991390229508349, "grad_norm": 1.0497158765792847, "learning_rate": 0.00018847717360344102, "loss": 4.5802, "step": 589000 }, { "epoch": 1.0992323285102135, "grad_norm": 1.3216798305511475, "learning_rate": 0.0001884752157647156, "loss": 4.3897, "step": 589050 }, { "epoch": 1.099325634069592, "grad_norm": 1.2395027875900269, "learning_rate": 0.00018847325776984698, "loss": 4.6756, "step": 589100 }, { "epoch": 1.0994189396289704, "grad_norm": 1.2498888969421387, "learning_rate": 0.00018847129961883864, "loss": 4.7413, "step": 589150 }, { "epoch": 1.099512245188349, "grad_norm": 1.0591673851013184, "learning_rate": 0.00018846934131169405, "loss": 4.4872, "step": 589200 }, { "epoch": 1.0996055507477274, "grad_norm": 1.0532019138336182, "learning_rate": 0.0001884673828484166, "loss": 4.5171, "step": 589250 }, { "epoch": 1.099698856307106, "grad_norm": 1.1950281858444214, "learning_rate": 0.00018846542422900982, "loss": 4.6478, "step": 589300 }, { "epoch": 1.0997921618664843, "grad_norm": 1.071777105331421, "learning_rate": 0.0001884634654534771, "loss": 4.4075, "step": 589350 }, { "epoch": 1.099885467425863, "grad_norm": 1.0904139280319214, "learning_rate": 0.00018846150652182195, "loss": 4.5509, "step": 589400 }, { "epoch": 1.0999787729852415, "grad_norm": 1.3796217441558838, "learning_rate": 0.0001884595474340478, "loss": 4.5019, "step": 589450 }, { "epoch": 1.1000720785446199, "grad_norm": 1.2729767560958862, "learning_rate": 0.00018845758819015815, "loss": 4.467, "step": 589500 }, { "epoch": 1.1001653841039984, "grad_norm": 1.3358039855957031, "learning_rate": 0.0001884556287901564, "loss": 4.4538, "step": 589550 }, { "epoch": 1.1002586896633768, "grad_norm": 1.1373659372329712, "learning_rate": 0.000188453669234046, "loss": 4.6206, "step": 589600 }, { "epoch": 1.1003519952227554, "grad_norm": 1.2427608966827393, "learning_rate": 0.00018845170952183044, "loss": 4.598, "step": 589650 }, { "epoch": 1.1004453007821338, "grad_norm": 1.359113335609436, "learning_rate": 0.00018844974965351319, "loss": 4.5275, "step": 589700 }, { "epoch": 1.1005386063415123, "grad_norm": 1.03851318359375, "learning_rate": 0.0001884477896290977, "loss": 4.4704, "step": 589750 }, { "epoch": 1.100631911900891, "grad_norm": 1.1576240062713623, "learning_rate": 0.00018844582944858745, "loss": 4.4868, "step": 589800 }, { "epoch": 1.1007252174602693, "grad_norm": 1.2408305406570435, "learning_rate": 0.00018844386911198583, "loss": 4.347, "step": 589850 }, { "epoch": 1.1008185230196479, "grad_norm": 0.7613019943237305, "learning_rate": 0.00018844190861929636, "loss": 4.3952, "step": 589900 }, { "epoch": 1.1009118285790263, "grad_norm": 1.2647994756698608, "learning_rate": 0.00018843994797052246, "loss": 4.6639, "step": 589950 }, { "epoch": 1.1010051341384048, "grad_norm": 1.0777170658111572, "learning_rate": 0.00018843798716566766, "loss": 4.6694, "step": 590000 }, { "epoch": 1.1010984396977832, "grad_norm": 1.1666394472122192, "learning_rate": 0.00018843602620473532, "loss": 4.3254, "step": 590050 }, { "epoch": 1.1011917452571618, "grad_norm": 0.955937385559082, "learning_rate": 0.00018843406508772898, "loss": 4.7742, "step": 590100 }, { "epoch": 1.1012850508165404, "grad_norm": 1.2898039817810059, "learning_rate": 0.0001884321038146521, "loss": 4.5316, "step": 590150 }, { "epoch": 1.1013783563759187, "grad_norm": 0.8135204911231995, "learning_rate": 0.0001884301423855081, "loss": 4.6568, "step": 590200 }, { "epoch": 1.1014716619352973, "grad_norm": 0.787746787071228, "learning_rate": 0.00018842818080030044, "loss": 4.4621, "step": 590250 }, { "epoch": 1.1015649674946757, "grad_norm": 0.9644063115119934, "learning_rate": 0.0001884262190590326, "loss": 4.4485, "step": 590300 }, { "epoch": 1.1016582730540543, "grad_norm": 1.113714337348938, "learning_rate": 0.00018842425716170808, "loss": 4.4943, "step": 590350 }, { "epoch": 1.1017515786134326, "grad_norm": 0.8823115825653076, "learning_rate": 0.00018842229510833026, "loss": 4.5165, "step": 590400 }, { "epoch": 1.1018448841728112, "grad_norm": 1.0359735488891602, "learning_rate": 0.0001884203328989027, "loss": 4.4835, "step": 590450 }, { "epoch": 1.1019381897321898, "grad_norm": 0.9581382870674133, "learning_rate": 0.00018841837053342877, "loss": 4.5604, "step": 590500 }, { "epoch": 1.1020314952915682, "grad_norm": 1.2359892129898071, "learning_rate": 0.00018841640801191197, "loss": 4.5445, "step": 590550 }, { "epoch": 1.1021248008509468, "grad_norm": 1.1109261512756348, "learning_rate": 0.0001884144453343558, "loss": 4.5455, "step": 590600 }, { "epoch": 1.1022181064103251, "grad_norm": 1.1797904968261719, "learning_rate": 0.00018841248250076365, "loss": 4.3875, "step": 590650 }, { "epoch": 1.1023114119697037, "grad_norm": 1.4320733547210693, "learning_rate": 0.00018841051951113907, "loss": 4.5436, "step": 590700 }, { "epoch": 1.102404717529082, "grad_norm": 1.2240647077560425, "learning_rate": 0.0001884085563654854, "loss": 4.4282, "step": 590750 }, { "epoch": 1.1024980230884607, "grad_norm": 0.886985182762146, "learning_rate": 0.00018840659306380627, "loss": 4.69, "step": 590800 }, { "epoch": 1.1025913286478393, "grad_norm": 1.1928297281265259, "learning_rate": 0.00018840462960610502, "loss": 4.4399, "step": 590850 }, { "epoch": 1.1026846342072176, "grad_norm": 1.0897561311721802, "learning_rate": 0.00018840266599238516, "loss": 4.7127, "step": 590900 }, { "epoch": 1.1027779397665962, "grad_norm": 1.146062970161438, "learning_rate": 0.00018840070222265017, "loss": 4.5147, "step": 590950 }, { "epoch": 1.1028712453259746, "grad_norm": 1.1875423192977905, "learning_rate": 0.00018839873829690345, "loss": 4.6225, "step": 591000 }, { "epoch": 1.1028712453259746, "eval_loss": 4.688204765319824, "eval_runtime": 229.7093, "eval_samples_per_second": 11.353, "eval_steps_per_second": 11.353, "eval_tts_loss": 7.642004450007508, "step": 591000 }, { "epoch": 1.1029645508853532, "grad_norm": 1.3133832216262817, "learning_rate": 0.00018839677421514855, "loss": 4.452, "step": 591050 }, { "epoch": 1.1030578564447315, "grad_norm": 0.7373575568199158, "learning_rate": 0.0001883948099773889, "loss": 4.5302, "step": 591100 }, { "epoch": 1.1031511620041101, "grad_norm": 1.1289856433868408, "learning_rate": 0.00018839284558362795, "loss": 4.7038, "step": 591150 }, { "epoch": 1.1032444675634885, "grad_norm": 1.519997000694275, "learning_rate": 0.00018839088103386918, "loss": 4.5821, "step": 591200 }, { "epoch": 1.103337773122867, "grad_norm": 1.252805471420288, "learning_rate": 0.00018838891632811606, "loss": 4.5147, "step": 591250 }, { "epoch": 1.1034310786822457, "grad_norm": 1.1697067022323608, "learning_rate": 0.00018838695146637205, "loss": 4.6269, "step": 591300 }, { "epoch": 1.103524384241624, "grad_norm": 1.2370368242263794, "learning_rate": 0.00018838498644864062, "loss": 4.4272, "step": 591350 }, { "epoch": 1.1036176898010026, "grad_norm": 1.2437517642974854, "learning_rate": 0.00018838302127492526, "loss": 4.7371, "step": 591400 }, { "epoch": 1.103710995360381, "grad_norm": 0.8420889973640442, "learning_rate": 0.00018838105594522938, "loss": 4.5757, "step": 591450 }, { "epoch": 1.1038043009197596, "grad_norm": 1.1960790157318115, "learning_rate": 0.00018837909045955653, "loss": 4.624, "step": 591500 }, { "epoch": 1.103897606479138, "grad_norm": 1.1207102537155151, "learning_rate": 0.00018837712481791012, "loss": 4.5122, "step": 591550 }, { "epoch": 1.1039909120385165, "grad_norm": 1.195600152015686, "learning_rate": 0.00018837515902029364, "loss": 4.4816, "step": 591600 }, { "epoch": 1.104084217597895, "grad_norm": 1.1802098751068115, "learning_rate": 0.00018837319306671052, "loss": 4.6365, "step": 591650 }, { "epoch": 1.1041775231572735, "grad_norm": 0.9305461049079895, "learning_rate": 0.00018837122695716433, "loss": 4.3938, "step": 591700 }, { "epoch": 1.104270828716652, "grad_norm": 1.0263454914093018, "learning_rate": 0.00018836926069165842, "loss": 4.5527, "step": 591750 }, { "epoch": 1.1043641342760304, "grad_norm": 1.2402276992797852, "learning_rate": 0.00018836729427019634, "loss": 4.4879, "step": 591800 }, { "epoch": 1.104457439835409, "grad_norm": 1.3399511575698853, "learning_rate": 0.00018836532769278154, "loss": 4.576, "step": 591850 }, { "epoch": 1.1045507453947874, "grad_norm": 1.1612963676452637, "learning_rate": 0.00018836336095941745, "loss": 4.5653, "step": 591900 }, { "epoch": 1.104644050954166, "grad_norm": 0.8741897940635681, "learning_rate": 0.00018836139407010758, "loss": 4.4721, "step": 591950 }, { "epoch": 1.1047373565135445, "grad_norm": 1.459704041481018, "learning_rate": 0.00018835942702485542, "loss": 4.446, "step": 592000 }, { "epoch": 1.104830662072923, "grad_norm": 1.3179489374160767, "learning_rate": 0.00018835745982366441, "loss": 4.6468, "step": 592050 }, { "epoch": 1.1049239676323015, "grad_norm": 1.0199700593948364, "learning_rate": 0.00018835549246653803, "loss": 4.4771, "step": 592100 }, { "epoch": 1.1050172731916799, "grad_norm": 0.9646822810173035, "learning_rate": 0.00018835352495347976, "loss": 4.5319, "step": 592150 }, { "epoch": 1.1051105787510584, "grad_norm": 1.0381524562835693, "learning_rate": 0.00018835155728449306, "loss": 4.5173, "step": 592200 }, { "epoch": 1.1052038843104368, "grad_norm": 0.9354884624481201, "learning_rate": 0.00018834958945958142, "loss": 4.4006, "step": 592250 }, { "epoch": 1.1052971898698154, "grad_norm": 1.1767635345458984, "learning_rate": 0.00018834762147874827, "loss": 4.5102, "step": 592300 }, { "epoch": 1.105390495429194, "grad_norm": 1.1324275732040405, "learning_rate": 0.00018834565334199712, "loss": 4.4489, "step": 592350 }, { "epoch": 1.1054838009885724, "grad_norm": 0.8641080856323242, "learning_rate": 0.00018834368504933147, "loss": 4.5176, "step": 592400 }, { "epoch": 1.105577106547951, "grad_norm": 1.0544503927230835, "learning_rate": 0.00018834171660075472, "loss": 4.4795, "step": 592450 }, { "epoch": 1.1056704121073293, "grad_norm": 0.975696325302124, "learning_rate": 0.0001883397479962704, "loss": 4.6393, "step": 592500 }, { "epoch": 1.105763717666708, "grad_norm": 1.1034462451934814, "learning_rate": 0.000188337779235882, "loss": 4.4893, "step": 592550 }, { "epoch": 1.1058570232260863, "grad_norm": 0.8972487449645996, "learning_rate": 0.0001883358103195929, "loss": 4.3922, "step": 592600 }, { "epoch": 1.1059503287854648, "grad_norm": 0.8479695916175842, "learning_rate": 0.00018833384124740667, "loss": 4.443, "step": 592650 }, { "epoch": 1.1060436343448434, "grad_norm": 1.3983571529388428, "learning_rate": 0.00018833187201932675, "loss": 4.4517, "step": 592700 }, { "epoch": 1.1061369399042218, "grad_norm": 1.114574670791626, "learning_rate": 0.00018832990263535664, "loss": 4.4131, "step": 592750 }, { "epoch": 1.1062302454636004, "grad_norm": 1.3605413436889648, "learning_rate": 0.00018832793309549975, "loss": 4.7454, "step": 592800 }, { "epoch": 1.1063235510229787, "grad_norm": 0.9035298824310303, "learning_rate": 0.00018832596339975966, "loss": 4.5161, "step": 592850 }, { "epoch": 1.1064168565823573, "grad_norm": 1.1796672344207764, "learning_rate": 0.00018832399354813974, "loss": 4.6018, "step": 592900 }, { "epoch": 1.1065101621417357, "grad_norm": 0.959598183631897, "learning_rate": 0.00018832202354064355, "loss": 4.6289, "step": 592950 }, { "epoch": 1.1066034677011143, "grad_norm": 1.2783652544021606, "learning_rate": 0.00018832005337727448, "loss": 4.71, "step": 593000 }, { "epoch": 1.1066967732604929, "grad_norm": 1.1815698146820068, "learning_rate": 0.0001883180830580361, "loss": 4.5255, "step": 593050 }, { "epoch": 1.1067900788198712, "grad_norm": 1.2025513648986816, "learning_rate": 0.0001883161125829318, "loss": 4.3967, "step": 593100 }, { "epoch": 1.1068833843792498, "grad_norm": 1.4194482564926147, "learning_rate": 0.00018831414195196513, "loss": 4.4261, "step": 593150 }, { "epoch": 1.1069766899386282, "grad_norm": 1.2385368347167969, "learning_rate": 0.00018831217116513954, "loss": 4.465, "step": 593200 }, { "epoch": 1.1070699954980068, "grad_norm": 0.9674620628356934, "learning_rate": 0.00018831020022245848, "loss": 4.3559, "step": 593250 }, { "epoch": 1.1071633010573851, "grad_norm": 1.4435473680496216, "learning_rate": 0.0001883082291239255, "loss": 4.6962, "step": 593300 }, { "epoch": 1.1072566066167637, "grad_norm": 1.2257434129714966, "learning_rate": 0.00018830625786954398, "loss": 4.4436, "step": 593350 }, { "epoch": 1.1073499121761423, "grad_norm": 1.1100847721099854, "learning_rate": 0.0001883042864593175, "loss": 4.5089, "step": 593400 }, { "epoch": 1.1074432177355207, "grad_norm": 1.1324775218963623, "learning_rate": 0.00018830231489324947, "loss": 4.4831, "step": 593450 }, { "epoch": 1.1075365232948993, "grad_norm": 1.3248528242111206, "learning_rate": 0.0001883003431713434, "loss": 4.3459, "step": 593500 }, { "epoch": 1.1076298288542776, "grad_norm": 0.9576318860054016, "learning_rate": 0.00018829837129360276, "loss": 4.4169, "step": 593550 }, { "epoch": 1.1077231344136562, "grad_norm": 1.1515893936157227, "learning_rate": 0.00018829639926003103, "loss": 4.5031, "step": 593600 }, { "epoch": 1.1078164399730346, "grad_norm": 1.069819450378418, "learning_rate": 0.00018829442707063166, "loss": 4.5484, "step": 593650 }, { "epoch": 1.1079097455324132, "grad_norm": 0.9660447239875793, "learning_rate": 0.0001882924547254082, "loss": 4.4963, "step": 593700 }, { "epoch": 1.1080030510917918, "grad_norm": 1.1475399732589722, "learning_rate": 0.00018829048222436412, "loss": 4.6599, "step": 593750 }, { "epoch": 1.1080963566511701, "grad_norm": 1.0066817998886108, "learning_rate": 0.00018828850956750283, "loss": 4.6875, "step": 593800 }, { "epoch": 1.1081896622105487, "grad_norm": 1.1797147989273071, "learning_rate": 0.00018828653675482786, "loss": 4.4712, "step": 593850 }, { "epoch": 1.108282967769927, "grad_norm": 1.1434378623962402, "learning_rate": 0.00018828456378634266, "loss": 4.6425, "step": 593900 }, { "epoch": 1.1083762733293057, "grad_norm": 0.9012452363967896, "learning_rate": 0.0001882825906620508, "loss": 4.57, "step": 593950 }, { "epoch": 1.108469578888684, "grad_norm": 0.962982177734375, "learning_rate": 0.00018828061738195565, "loss": 4.3032, "step": 594000 }, { "epoch": 1.108469578888684, "eval_loss": 4.695082664489746, "eval_runtime": 233.128, "eval_samples_per_second": 11.187, "eval_steps_per_second": 11.187, "eval_tts_loss": 7.616777886176608, "step": 594000 }, { "epoch": 1.1085628844480626, "grad_norm": 1.3535645008087158, "learning_rate": 0.00018827864394606078, "loss": 4.417, "step": 594050 }, { "epoch": 1.1086561900074412, "grad_norm": 1.191959023475647, "learning_rate": 0.00018827667035436961, "loss": 4.5385, "step": 594100 }, { "epoch": 1.1087494955668196, "grad_norm": 1.0976771116256714, "learning_rate": 0.0001882746966068857, "loss": 4.6683, "step": 594150 }, { "epoch": 1.1088428011261982, "grad_norm": 1.1928831338882446, "learning_rate": 0.00018827272270361244, "loss": 4.2423, "step": 594200 }, { "epoch": 1.1089361066855765, "grad_norm": 1.0553604364395142, "learning_rate": 0.00018827074864455334, "loss": 4.5694, "step": 594250 }, { "epoch": 1.109029412244955, "grad_norm": 0.8366163969039917, "learning_rate": 0.00018826877442971193, "loss": 4.4774, "step": 594300 }, { "epoch": 1.1091227178043335, "grad_norm": 1.29400634765625, "learning_rate": 0.00018826680005909165, "loss": 4.4817, "step": 594350 }, { "epoch": 1.109216023363712, "grad_norm": 1.0770092010498047, "learning_rate": 0.00018826482553269602, "loss": 4.5013, "step": 594400 }, { "epoch": 1.1093093289230906, "grad_norm": 0.9708148837089539, "learning_rate": 0.00018826285085052846, "loss": 4.5623, "step": 594450 }, { "epoch": 1.109402634482469, "grad_norm": 1.1294338703155518, "learning_rate": 0.00018826087601259253, "loss": 4.3233, "step": 594500 }, { "epoch": 1.1094959400418476, "grad_norm": 1.594743251800537, "learning_rate": 0.0001882589010188917, "loss": 4.5384, "step": 594550 }, { "epoch": 1.109589245601226, "grad_norm": 1.3100579977035522, "learning_rate": 0.00018825692586942942, "loss": 4.6106, "step": 594600 }, { "epoch": 1.1096825511606045, "grad_norm": 1.0811430215835571, "learning_rate": 0.0001882549505642092, "loss": 4.3901, "step": 594650 }, { "epoch": 1.109775856719983, "grad_norm": 0.7895335555076599, "learning_rate": 0.0001882529751032345, "loss": 4.4685, "step": 594700 }, { "epoch": 1.1098691622793615, "grad_norm": 1.1325701475143433, "learning_rate": 0.00018825099948650885, "loss": 4.4619, "step": 594750 }, { "epoch": 1.10996246783874, "grad_norm": 1.1729422807693481, "learning_rate": 0.0001882490237140357, "loss": 4.61, "step": 594800 }, { "epoch": 1.1100557733981185, "grad_norm": 1.0960415601730347, "learning_rate": 0.00018824704778581857, "loss": 4.5233, "step": 594850 }, { "epoch": 1.110149078957497, "grad_norm": 1.1578800678253174, "learning_rate": 0.00018824507170186092, "loss": 4.6045, "step": 594900 }, { "epoch": 1.1102423845168754, "grad_norm": 0.7354874610900879, "learning_rate": 0.00018824309546216624, "loss": 4.4359, "step": 594950 }, { "epoch": 1.110335690076254, "grad_norm": 1.1750355958938599, "learning_rate": 0.00018824111906673804, "loss": 4.4497, "step": 595000 }, { "epoch": 1.1104289956356324, "grad_norm": 1.2037566900253296, "learning_rate": 0.00018823914251557975, "loss": 4.4902, "step": 595050 }, { "epoch": 1.110522301195011, "grad_norm": 1.0697766542434692, "learning_rate": 0.00018823716580869496, "loss": 4.6528, "step": 595100 }, { "epoch": 1.1106156067543895, "grad_norm": 1.0923959016799927, "learning_rate": 0.00018823518894608705, "loss": 4.6313, "step": 595150 }, { "epoch": 1.110708912313768, "grad_norm": 1.1106702089309692, "learning_rate": 0.00018823321192775957, "loss": 4.5976, "step": 595200 }, { "epoch": 1.1108022178731465, "grad_norm": 1.0829216241836548, "learning_rate": 0.000188231234753716, "loss": 4.5805, "step": 595250 }, { "epoch": 1.1108955234325248, "grad_norm": 0.769853949546814, "learning_rate": 0.00018822925742395982, "loss": 4.4895, "step": 595300 }, { "epoch": 1.1109888289919034, "grad_norm": 1.5613293647766113, "learning_rate": 0.00018822727993849454, "loss": 4.5717, "step": 595350 }, { "epoch": 1.1110821345512818, "grad_norm": 1.0454961061477661, "learning_rate": 0.00018822530229732363, "loss": 4.5171, "step": 595400 }, { "epoch": 1.1111754401106604, "grad_norm": 0.9541016221046448, "learning_rate": 0.00018822332450045055, "loss": 4.5785, "step": 595450 }, { "epoch": 1.111268745670039, "grad_norm": 1.332036018371582, "learning_rate": 0.00018822134654787885, "loss": 4.4404, "step": 595500 }, { "epoch": 1.1113620512294173, "grad_norm": 1.2711645364761353, "learning_rate": 0.00018821936843961204, "loss": 4.5687, "step": 595550 }, { "epoch": 1.111455356788796, "grad_norm": 1.2488696575164795, "learning_rate": 0.00018821739017565352, "loss": 4.5183, "step": 595600 }, { "epoch": 1.1115486623481743, "grad_norm": 1.4064321517944336, "learning_rate": 0.00018821541175600683, "loss": 4.5743, "step": 595650 }, { "epoch": 1.1116419679075529, "grad_norm": 1.3288668394088745, "learning_rate": 0.00018821343318067545, "loss": 4.5752, "step": 595700 }, { "epoch": 1.1117352734669312, "grad_norm": 1.1184446811676025, "learning_rate": 0.00018821145444966294, "loss": 4.6127, "step": 595750 }, { "epoch": 1.1118285790263098, "grad_norm": 1.0871092081069946, "learning_rate": 0.00018820947556297268, "loss": 4.3822, "step": 595800 }, { "epoch": 1.1119218845856884, "grad_norm": 1.2087931632995605, "learning_rate": 0.0001882074965206082, "loss": 4.5746, "step": 595850 }, { "epoch": 1.1120151901450668, "grad_norm": 1.121762752532959, "learning_rate": 0.00018820551732257306, "loss": 4.3995, "step": 595900 }, { "epoch": 1.1121084957044454, "grad_norm": 1.0212441682815552, "learning_rate": 0.0001882035379688707, "loss": 4.5218, "step": 595950 }, { "epoch": 1.1122018012638237, "grad_norm": 1.138856053352356, "learning_rate": 0.0001882015584595046, "loss": 4.7315, "step": 596000 }, { "epoch": 1.1122951068232023, "grad_norm": 1.0418362617492676, "learning_rate": 0.00018819957879447825, "loss": 4.5214, "step": 596050 }, { "epoch": 1.1123884123825807, "grad_norm": 0.9491592049598694, "learning_rate": 0.00018819759897379518, "loss": 4.3786, "step": 596100 }, { "epoch": 1.1124817179419593, "grad_norm": 1.097137212753296, "learning_rate": 0.00018819561899745886, "loss": 4.4817, "step": 596150 }, { "epoch": 1.1125750235013379, "grad_norm": 1.1851975917816162, "learning_rate": 0.0001881936388654728, "loss": 4.4861, "step": 596200 }, { "epoch": 1.1126683290607162, "grad_norm": 1.4834537506103516, "learning_rate": 0.0001881916585778405, "loss": 4.4722, "step": 596250 }, { "epoch": 1.1127616346200948, "grad_norm": 1.0558851957321167, "learning_rate": 0.00018818967813456538, "loss": 4.3444, "step": 596300 }, { "epoch": 1.1128549401794732, "grad_norm": 1.359838604927063, "learning_rate": 0.00018818769753565104, "loss": 4.5459, "step": 596350 }, { "epoch": 1.1129482457388518, "grad_norm": 1.0155283212661743, "learning_rate": 0.00018818571678110092, "loss": 4.6541, "step": 596400 }, { "epoch": 1.1130415512982301, "grad_norm": 1.0982826948165894, "learning_rate": 0.00018818373587091856, "loss": 4.4475, "step": 596450 }, { "epoch": 1.1131348568576087, "grad_norm": 1.2411000728607178, "learning_rate": 0.00018818175480510738, "loss": 4.6041, "step": 596500 }, { "epoch": 1.1132281624169873, "grad_norm": 1.505938172340393, "learning_rate": 0.00018817977358367096, "loss": 4.5909, "step": 596550 }, { "epoch": 1.1133214679763657, "grad_norm": 1.2255579233169556, "learning_rate": 0.0001881777922066127, "loss": 4.3938, "step": 596600 }, { "epoch": 1.1134147735357443, "grad_norm": 1.3656517267227173, "learning_rate": 0.0001881758106739362, "loss": 4.4384, "step": 596650 }, { "epoch": 1.1135080790951226, "grad_norm": 1.1881074905395508, "learning_rate": 0.00018817382898564492, "loss": 4.4438, "step": 596700 }, { "epoch": 1.1136013846545012, "grad_norm": 0.6375342607498169, "learning_rate": 0.0001881718471417423, "loss": 4.5557, "step": 596750 }, { "epoch": 1.1136946902138796, "grad_norm": 0.9763136506080627, "learning_rate": 0.00018816986514223188, "loss": 4.3716, "step": 596800 }, { "epoch": 1.1137879957732582, "grad_norm": 0.9448479413986206, "learning_rate": 0.00018816788298711722, "loss": 4.438, "step": 596850 }, { "epoch": 1.1138813013326367, "grad_norm": 1.2400981187820435, "learning_rate": 0.00018816590067640172, "loss": 4.5736, "step": 596900 }, { "epoch": 1.113974606892015, "grad_norm": 1.076710820198059, "learning_rate": 0.00018816391821008894, "loss": 4.5135, "step": 596950 }, { "epoch": 1.1140679124513937, "grad_norm": 0.9921693801879883, "learning_rate": 0.00018816193558818234, "loss": 4.4545, "step": 597000 }, { "epoch": 1.1140679124513937, "eval_loss": 4.688167095184326, "eval_runtime": 230.4078, "eval_samples_per_second": 11.319, "eval_steps_per_second": 11.319, "eval_tts_loss": 7.633871327572862, "step": 597000 }, { "epoch": 1.114161218010772, "grad_norm": 1.0689194202423096, "learning_rate": 0.00018815995281068544, "loss": 4.3994, "step": 597050 }, { "epoch": 1.1142545235701506, "grad_norm": 1.2134331464767456, "learning_rate": 0.0001881579698776017, "loss": 4.5245, "step": 597100 }, { "epoch": 1.114347829129529, "grad_norm": 1.167988657951355, "learning_rate": 0.0001881559867889347, "loss": 4.6921, "step": 597150 }, { "epoch": 1.1144411346889076, "grad_norm": 0.8030964732170105, "learning_rate": 0.0001881540035446879, "loss": 4.5543, "step": 597200 }, { "epoch": 1.1145344402482862, "grad_norm": 1.1267341375350952, "learning_rate": 0.00018815202014486478, "loss": 4.4414, "step": 597250 }, { "epoch": 1.1146277458076645, "grad_norm": 0.9987810850143433, "learning_rate": 0.00018815003658946883, "loss": 4.536, "step": 597300 }, { "epoch": 1.1147210513670431, "grad_norm": 1.2188377380371094, "learning_rate": 0.00018814805287850362, "loss": 4.5928, "step": 597350 }, { "epoch": 1.1148143569264215, "grad_norm": 0.789045512676239, "learning_rate": 0.00018814606901197258, "loss": 4.7242, "step": 597400 }, { "epoch": 1.1149076624858, "grad_norm": 1.2680144309997559, "learning_rate": 0.00018814408498987924, "loss": 4.656, "step": 597450 }, { "epoch": 1.1150009680451785, "grad_norm": 1.2104791402816772, "learning_rate": 0.00018814210081222709, "loss": 4.5512, "step": 597500 }, { "epoch": 1.115094273604557, "grad_norm": 1.1773818731307983, "learning_rate": 0.00018814011647901968, "loss": 4.7193, "step": 597550 }, { "epoch": 1.1151875791639356, "grad_norm": 1.1549407243728638, "learning_rate": 0.00018813813199026046, "loss": 4.4162, "step": 597600 }, { "epoch": 1.115280884723314, "grad_norm": 1.1996943950653076, "learning_rate": 0.00018813614734595292, "loss": 4.563, "step": 597650 }, { "epoch": 1.1153741902826926, "grad_norm": 1.0241234302520752, "learning_rate": 0.00018813416254610061, "loss": 4.6639, "step": 597700 }, { "epoch": 1.115467495842071, "grad_norm": 0.9135685563087463, "learning_rate": 0.000188132177590707, "loss": 4.461, "step": 597750 }, { "epoch": 1.1155608014014495, "grad_norm": 1.1399866342544556, "learning_rate": 0.00018813019247977558, "loss": 4.3676, "step": 597800 }, { "epoch": 1.115654106960828, "grad_norm": 1.1060309410095215, "learning_rate": 0.00018812820721330993, "loss": 4.5985, "step": 597850 }, { "epoch": 1.1157474125202065, "grad_norm": 1.219186782836914, "learning_rate": 0.00018812622179131347, "loss": 4.3329, "step": 597900 }, { "epoch": 1.115840718079585, "grad_norm": 1.2959764003753662, "learning_rate": 0.00018812423621378973, "loss": 4.5617, "step": 597950 }, { "epoch": 1.1159340236389634, "grad_norm": 1.234645128250122, "learning_rate": 0.00018812225048074223, "loss": 4.5403, "step": 598000 }, { "epoch": 1.116027329198342, "grad_norm": 1.3577868938446045, "learning_rate": 0.00018812026459217446, "loss": 4.5747, "step": 598050 }, { "epoch": 1.1161206347577204, "grad_norm": 1.1316485404968262, "learning_rate": 0.00018811827854808995, "loss": 4.4017, "step": 598100 }, { "epoch": 1.116213940317099, "grad_norm": 1.311916470527649, "learning_rate": 0.00018811629234849215, "loss": 4.7297, "step": 598150 }, { "epoch": 1.1163072458764773, "grad_norm": 1.0299351215362549, "learning_rate": 0.00018811430599338462, "loss": 4.5107, "step": 598200 }, { "epoch": 1.116400551435856, "grad_norm": 1.272498607635498, "learning_rate": 0.00018811231948277085, "loss": 4.6277, "step": 598250 }, { "epoch": 1.1164938569952345, "grad_norm": 1.2242295742034912, "learning_rate": 0.0001881103328166543, "loss": 4.6499, "step": 598300 }, { "epoch": 1.1165871625546129, "grad_norm": 0.9778114557266235, "learning_rate": 0.00018810834599503854, "loss": 4.6464, "step": 598350 }, { "epoch": 1.1166804681139915, "grad_norm": 1.154250979423523, "learning_rate": 0.00018810635901792705, "loss": 4.5742, "step": 598400 }, { "epoch": 1.1167737736733698, "grad_norm": 1.2944416999816895, "learning_rate": 0.00018810437188532334, "loss": 4.6487, "step": 598450 }, { "epoch": 1.1168670792327484, "grad_norm": 1.34421968460083, "learning_rate": 0.00018810238459723093, "loss": 4.4566, "step": 598500 }, { "epoch": 1.1169603847921268, "grad_norm": 1.150518536567688, "learning_rate": 0.0001881003971536533, "loss": 4.4433, "step": 598550 }, { "epoch": 1.1170536903515054, "grad_norm": 1.2697851657867432, "learning_rate": 0.000188098409554594, "loss": 4.4547, "step": 598600 }, { "epoch": 1.117146995910884, "grad_norm": 1.338557481765747, "learning_rate": 0.00018809642180005645, "loss": 4.3729, "step": 598650 }, { "epoch": 1.1172403014702623, "grad_norm": 0.9992930889129639, "learning_rate": 0.00018809443389004426, "loss": 4.4697, "step": 598700 }, { "epoch": 1.117333607029641, "grad_norm": 1.189880132675171, "learning_rate": 0.0001880924458245609, "loss": 4.7272, "step": 598750 }, { "epoch": 1.1174269125890193, "grad_norm": 0.9383851885795593, "learning_rate": 0.00018809045760360984, "loss": 4.367, "step": 598800 }, { "epoch": 1.1175202181483979, "grad_norm": 1.2745143175125122, "learning_rate": 0.00018808846922719466, "loss": 4.4057, "step": 598850 }, { "epoch": 1.1176135237077762, "grad_norm": 1.193938136100769, "learning_rate": 0.00018808648069531877, "loss": 4.6325, "step": 598900 }, { "epoch": 1.1177068292671548, "grad_norm": 1.063791275024414, "learning_rate": 0.0001880844920079858, "loss": 4.3902, "step": 598950 }, { "epoch": 1.1178001348265334, "grad_norm": 0.9823865294456482, "learning_rate": 0.0001880825031651992, "loss": 4.6632, "step": 599000 }, { "epoch": 1.1178934403859118, "grad_norm": 1.256320834159851, "learning_rate": 0.00018808051416696245, "loss": 4.5409, "step": 599050 }, { "epoch": 1.1179867459452903, "grad_norm": 1.1148594617843628, "learning_rate": 0.0001880785250132791, "loss": 4.4913, "step": 599100 }, { "epoch": 1.1180800515046687, "grad_norm": 0.9170576930046082, "learning_rate": 0.00018807653570415263, "loss": 4.3689, "step": 599150 }, { "epoch": 1.1181733570640473, "grad_norm": 1.5635290145874023, "learning_rate": 0.00018807454623958662, "loss": 4.6915, "step": 599200 }, { "epoch": 1.1182666626234257, "grad_norm": 1.1275635957717896, "learning_rate": 0.00018807255661958453, "loss": 4.5702, "step": 599250 }, { "epoch": 1.1183599681828043, "grad_norm": 1.2036988735198975, "learning_rate": 0.00018807056684414986, "loss": 4.5606, "step": 599300 }, { "epoch": 1.1184532737421828, "grad_norm": 1.086449384689331, "learning_rate": 0.0001880685769132861, "loss": 4.5415, "step": 599350 }, { "epoch": 1.1185465793015612, "grad_norm": 1.3360416889190674, "learning_rate": 0.00018806658682699685, "loss": 4.6246, "step": 599400 }, { "epoch": 1.1186398848609398, "grad_norm": 1.1793767213821411, "learning_rate": 0.00018806459658528553, "loss": 4.3689, "step": 599450 }, { "epoch": 1.1187331904203182, "grad_norm": 1.3979791402816772, "learning_rate": 0.00018806260618815573, "loss": 4.6344, "step": 599500 }, { "epoch": 1.1188264959796967, "grad_norm": 1.0470740795135498, "learning_rate": 0.0001880606156356109, "loss": 4.437, "step": 599550 }, { "epoch": 1.118919801539075, "grad_norm": 1.167197585105896, "learning_rate": 0.00018805862492765458, "loss": 4.4238, "step": 599600 }, { "epoch": 1.1190131070984537, "grad_norm": 1.2411152124404907, "learning_rate": 0.00018805663406429028, "loss": 4.3657, "step": 599650 }, { "epoch": 1.1191064126578323, "grad_norm": 1.107071876525879, "learning_rate": 0.0001880546430455215, "loss": 4.597, "step": 599700 }, { "epoch": 1.1191997182172106, "grad_norm": 1.1010178327560425, "learning_rate": 0.00018805265187135181, "loss": 4.4926, "step": 599750 }, { "epoch": 1.1192930237765892, "grad_norm": 1.4962952136993408, "learning_rate": 0.00018805066054178467, "loss": 4.5005, "step": 599800 }, { "epoch": 1.1193863293359676, "grad_norm": 1.1885900497436523, "learning_rate": 0.0001880486690568236, "loss": 4.6571, "step": 599850 }, { "epoch": 1.1194796348953462, "grad_norm": 1.0619927644729614, "learning_rate": 0.0001880466774164721, "loss": 4.6692, "step": 599900 }, { "epoch": 1.1195729404547246, "grad_norm": 1.264176845550537, "learning_rate": 0.00018804468562073375, "loss": 4.7805, "step": 599950 }, { "epoch": 1.1196662460141031, "grad_norm": 1.1927601099014282, "learning_rate": 0.00018804269366961198, "loss": 4.561, "step": 600000 }, { "epoch": 1.1196662460141031, "eval_loss": 4.690563201904297, "eval_runtime": 229.5919, "eval_samples_per_second": 11.359, "eval_steps_per_second": 11.359, "eval_tts_loss": 7.596550817037161, "step": 600000 }, { "epoch": 1.1197595515734817, "grad_norm": 1.0309759378433228, "learning_rate": 0.0001880407015631104, "loss": 4.4301, "step": 600050 }, { "epoch": 1.11985285713286, "grad_norm": 0.7386344075202942, "learning_rate": 0.00018803870930123243, "loss": 4.4495, "step": 600100 }, { "epoch": 1.1199461626922387, "grad_norm": 1.0129657983779907, "learning_rate": 0.0001880367168839816, "loss": 4.4903, "step": 600150 }, { "epoch": 1.120039468251617, "grad_norm": 1.2976124286651611, "learning_rate": 0.00018803472431136152, "loss": 4.4285, "step": 600200 }, { "epoch": 1.1201327738109956, "grad_norm": 0.9000319838523865, "learning_rate": 0.0001880327315833756, "loss": 4.4435, "step": 600250 }, { "epoch": 1.120226079370374, "grad_norm": 1.0231966972351074, "learning_rate": 0.00018803073870002744, "loss": 4.6099, "step": 600300 }, { "epoch": 1.1203193849297526, "grad_norm": 1.2984495162963867, "learning_rate": 0.00018802874566132047, "loss": 4.5297, "step": 600350 }, { "epoch": 1.1204126904891312, "grad_norm": 1.5801889896392822, "learning_rate": 0.00018802675246725828, "loss": 4.5318, "step": 600400 }, { "epoch": 1.1205059960485095, "grad_norm": 1.5039234161376953, "learning_rate": 0.00018802475911784434, "loss": 4.7027, "step": 600450 }, { "epoch": 1.1205993016078881, "grad_norm": 0.9449006915092468, "learning_rate": 0.00018802276561308222, "loss": 4.452, "step": 600500 }, { "epoch": 1.1206926071672665, "grad_norm": 1.2167737483978271, "learning_rate": 0.00018802077195297537, "loss": 4.6185, "step": 600550 }, { "epoch": 1.120785912726645, "grad_norm": 1.080098271369934, "learning_rate": 0.00018801877813752737, "loss": 4.4906, "step": 600600 }, { "epoch": 1.1208792182860234, "grad_norm": 1.3584978580474854, "learning_rate": 0.00018801678416674174, "loss": 4.5395, "step": 600650 }, { "epoch": 1.120972523845402, "grad_norm": 1.1792688369750977, "learning_rate": 0.00018801479004062192, "loss": 4.4794, "step": 600700 }, { "epoch": 1.1210658294047806, "grad_norm": 1.1479394435882568, "learning_rate": 0.0001880127957591715, "loss": 4.4981, "step": 600750 }, { "epoch": 1.121159134964159, "grad_norm": 0.8136838674545288, "learning_rate": 0.000188010801322394, "loss": 4.51, "step": 600800 }, { "epoch": 1.1212524405235376, "grad_norm": 1.1819379329681396, "learning_rate": 0.0001880088067302929, "loss": 4.3647, "step": 600850 }, { "epoch": 1.121345746082916, "grad_norm": 1.033268928527832, "learning_rate": 0.00018800681198287173, "loss": 4.4832, "step": 600900 }, { "epoch": 1.1214390516422945, "grad_norm": 0.8835206627845764, "learning_rate": 0.00018800481708013407, "loss": 4.3872, "step": 600950 }, { "epoch": 1.1215323572016729, "grad_norm": 1.414609432220459, "learning_rate": 0.00018800282202208334, "loss": 4.5763, "step": 601000 }, { "epoch": 1.1216256627610515, "grad_norm": 0.7735211253166199, "learning_rate": 0.00018800082680872312, "loss": 4.6528, "step": 601050 }, { "epoch": 1.12171896832043, "grad_norm": 0.9446475505828857, "learning_rate": 0.00018799883144005698, "loss": 4.5431, "step": 601100 }, { "epoch": 1.1218122738798084, "grad_norm": 1.1820610761642456, "learning_rate": 0.0001879968359160883, "loss": 4.5384, "step": 601150 }, { "epoch": 1.121905579439187, "grad_norm": 1.108776569366455, "learning_rate": 0.00018799484023682077, "loss": 4.5309, "step": 601200 }, { "epoch": 1.1219988849985654, "grad_norm": 1.0821504592895508, "learning_rate": 0.0001879928444022578, "loss": 4.5612, "step": 601250 }, { "epoch": 1.122092190557944, "grad_norm": 1.0957658290863037, "learning_rate": 0.00018799084841240293, "loss": 4.3388, "step": 601300 }, { "epoch": 1.1221854961173223, "grad_norm": 0.8595952391624451, "learning_rate": 0.0001879888522672597, "loss": 4.5195, "step": 601350 }, { "epoch": 1.122278801676701, "grad_norm": 1.2193304300308228, "learning_rate": 0.00018798685596683163, "loss": 4.6044, "step": 601400 }, { "epoch": 1.1223721072360795, "grad_norm": 1.131418228149414, "learning_rate": 0.00018798485951112225, "loss": 4.6415, "step": 601450 }, { "epoch": 1.1224654127954579, "grad_norm": 1.317354440689087, "learning_rate": 0.00018798286290013508, "loss": 4.5777, "step": 601500 }, { "epoch": 1.1225587183548364, "grad_norm": 0.9986324310302734, "learning_rate": 0.00018798086613387363, "loss": 4.6315, "step": 601550 }, { "epoch": 1.1226520239142148, "grad_norm": 0.8964020609855652, "learning_rate": 0.00018797886921234142, "loss": 4.6396, "step": 601600 }, { "epoch": 1.1227453294735934, "grad_norm": 1.1442105770111084, "learning_rate": 0.00018797687213554197, "loss": 4.6561, "step": 601650 }, { "epoch": 1.1228386350329718, "grad_norm": 1.2588227987289429, "learning_rate": 0.00018797487490347886, "loss": 4.5885, "step": 601700 }, { "epoch": 1.1229319405923504, "grad_norm": 0.852618396282196, "learning_rate": 0.00018797287751615556, "loss": 4.5893, "step": 601750 }, { "epoch": 1.123025246151729, "grad_norm": 1.0891003608703613, "learning_rate": 0.00018797087997357562, "loss": 4.3912, "step": 601800 }, { "epoch": 1.1231185517111073, "grad_norm": 1.2330260276794434, "learning_rate": 0.00018796888227574254, "loss": 4.272, "step": 601850 }, { "epoch": 1.123211857270486, "grad_norm": 1.0581685304641724, "learning_rate": 0.00018796688442265986, "loss": 4.3786, "step": 601900 }, { "epoch": 1.1233051628298643, "grad_norm": 1.2142200469970703, "learning_rate": 0.00018796488641433111, "loss": 4.5511, "step": 601950 }, { "epoch": 1.1233984683892428, "grad_norm": 1.1155216693878174, "learning_rate": 0.00018796288825075982, "loss": 4.7314, "step": 602000 }, { "epoch": 1.1234917739486212, "grad_norm": 1.2222517728805542, "learning_rate": 0.0001879608899319495, "loss": 4.5012, "step": 602050 }, { "epoch": 1.1235850795079998, "grad_norm": 1.2458221912384033, "learning_rate": 0.0001879588914579037, "loss": 4.5024, "step": 602100 }, { "epoch": 1.1236783850673784, "grad_norm": 1.045380711555481, "learning_rate": 0.00018795689282862595, "loss": 4.6545, "step": 602150 }, { "epoch": 1.1237716906267567, "grad_norm": 1.0781384706497192, "learning_rate": 0.0001879548940441197, "loss": 4.3878, "step": 602200 }, { "epoch": 1.1238649961861353, "grad_norm": 1.5653189420700073, "learning_rate": 0.0001879528951043886, "loss": 4.6361, "step": 602250 }, { "epoch": 1.1239583017455137, "grad_norm": 0.795764684677124, "learning_rate": 0.0001879508960094361, "loss": 4.4479, "step": 602300 }, { "epoch": 1.1240516073048923, "grad_norm": 0.9439929127693176, "learning_rate": 0.00018794889675926576, "loss": 4.4479, "step": 602350 }, { "epoch": 1.1241449128642707, "grad_norm": 1.0757577419281006, "learning_rate": 0.00018794689735388104, "loss": 4.5229, "step": 602400 }, { "epoch": 1.1242382184236492, "grad_norm": 1.243285894393921, "learning_rate": 0.00018794489779328557, "loss": 4.4234, "step": 602450 }, { "epoch": 1.1243315239830278, "grad_norm": 1.187574863433838, "learning_rate": 0.0001879428980774828, "loss": 4.6186, "step": 602500 }, { "epoch": 1.1244248295424062, "grad_norm": 1.2882564067840576, "learning_rate": 0.00018794089820647628, "loss": 4.2644, "step": 602550 }, { "epoch": 1.1245181351017848, "grad_norm": 1.268721580505371, "learning_rate": 0.00018793889818026957, "loss": 4.6223, "step": 602600 }, { "epoch": 1.1246114406611631, "grad_norm": 1.19908607006073, "learning_rate": 0.00018793689799886617, "loss": 4.6475, "step": 602650 }, { "epoch": 1.1247047462205417, "grad_norm": 1.266588568687439, "learning_rate": 0.00018793489766226964, "loss": 4.4792, "step": 602700 }, { "epoch": 1.12479805177992, "grad_norm": 1.1467400789260864, "learning_rate": 0.00018793289717048347, "loss": 4.5008, "step": 602750 }, { "epoch": 1.1248913573392987, "grad_norm": 1.2207145690917969, "learning_rate": 0.00018793089652351122, "loss": 4.8235, "step": 602800 }, { "epoch": 1.1249846628986773, "grad_norm": 1.2437217235565186, "learning_rate": 0.0001879288957213564, "loss": 4.4484, "step": 602850 }, { "epoch": 1.1250779684580556, "grad_norm": 1.3227612972259521, "learning_rate": 0.00018792689476402252, "loss": 4.4368, "step": 602900 }, { "epoch": 1.1251712740174342, "grad_norm": 1.0564016103744507, "learning_rate": 0.00018792489365151318, "loss": 4.6003, "step": 602950 }, { "epoch": 1.1252645795768126, "grad_norm": 0.6650514006614685, "learning_rate": 0.00018792289238383188, "loss": 4.2191, "step": 603000 }, { "epoch": 1.1252645795768126, "eval_loss": 4.697524547576904, "eval_runtime": 230.1653, "eval_samples_per_second": 11.331, "eval_steps_per_second": 11.331, "eval_tts_loss": 7.5680320383238096, "step": 603000 }, { "epoch": 1.1253578851361912, "grad_norm": 1.1871076822280884, "learning_rate": 0.00018792089096098213, "loss": 4.4266, "step": 603050 }, { "epoch": 1.1254511906955695, "grad_norm": 0.8059503436088562, "learning_rate": 0.0001879188893829675, "loss": 4.6624, "step": 603100 }, { "epoch": 1.1255444962549481, "grad_norm": 1.1675211191177368, "learning_rate": 0.00018791688764979145, "loss": 4.518, "step": 603150 }, { "epoch": 1.1256378018143267, "grad_norm": 1.4043402671813965, "learning_rate": 0.0001879148857614576, "loss": 4.4969, "step": 603200 }, { "epoch": 1.125731107373705, "grad_norm": 0.9664667248725891, "learning_rate": 0.00018791288371796942, "loss": 4.4882, "step": 603250 }, { "epoch": 1.1258244129330837, "grad_norm": 1.3521616458892822, "learning_rate": 0.0001879108815193305, "loss": 4.6226, "step": 603300 }, { "epoch": 1.125917718492462, "grad_norm": 1.201263666152954, "learning_rate": 0.00018790887916554435, "loss": 4.3833, "step": 603350 }, { "epoch": 1.1260110240518406, "grad_norm": 1.4198589324951172, "learning_rate": 0.00018790687665661446, "loss": 4.4994, "step": 603400 }, { "epoch": 1.126104329611219, "grad_norm": 1.2315673828125, "learning_rate": 0.00018790487399254443, "loss": 4.5994, "step": 603450 }, { "epoch": 1.1261976351705976, "grad_norm": 1.3423243761062622, "learning_rate": 0.00018790287117333772, "loss": 4.5662, "step": 603500 }, { "epoch": 1.1262909407299762, "grad_norm": 1.07157301902771, "learning_rate": 0.00018790086819899795, "loss": 4.4291, "step": 603550 }, { "epoch": 1.1263842462893545, "grad_norm": 1.057149887084961, "learning_rate": 0.00018789886506952858, "loss": 4.6188, "step": 603600 }, { "epoch": 1.126477551848733, "grad_norm": 1.1015348434448242, "learning_rate": 0.00018789686178493321, "loss": 4.5378, "step": 603650 }, { "epoch": 1.1265708574081115, "grad_norm": 1.2191746234893799, "learning_rate": 0.00018789485834521535, "loss": 4.4141, "step": 603700 }, { "epoch": 1.12666416296749, "grad_norm": 1.1586687564849854, "learning_rate": 0.00018789285475037852, "loss": 4.4799, "step": 603750 }, { "epoch": 1.1267574685268684, "grad_norm": 1.5176801681518555, "learning_rate": 0.00018789085100042626, "loss": 4.4691, "step": 603800 }, { "epoch": 1.126850774086247, "grad_norm": 1.1869370937347412, "learning_rate": 0.00018788884709536213, "loss": 4.5518, "step": 603850 }, { "epoch": 1.1269440796456256, "grad_norm": 1.3327937126159668, "learning_rate": 0.00018788684303518963, "loss": 4.7133, "step": 603900 }, { "epoch": 1.127037385205004, "grad_norm": 1.2550445795059204, "learning_rate": 0.00018788483881991233, "loss": 4.5715, "step": 603950 }, { "epoch": 1.1271306907643825, "grad_norm": 1.0853204727172852, "learning_rate": 0.00018788283444953375, "loss": 4.6343, "step": 604000 }, { "epoch": 1.127223996323761, "grad_norm": 1.0555534362792969, "learning_rate": 0.00018788082992405743, "loss": 4.4894, "step": 604050 }, { "epoch": 1.1273173018831395, "grad_norm": 1.1535853147506714, "learning_rate": 0.0001878788252434869, "loss": 4.6251, "step": 604100 }, { "epoch": 1.1274106074425179, "grad_norm": 1.3375940322875977, "learning_rate": 0.0001878768204078257, "loss": 4.5846, "step": 604150 }, { "epoch": 1.1275039130018965, "grad_norm": 0.9539817571640015, "learning_rate": 0.00018787481541707738, "loss": 4.4051, "step": 604200 }, { "epoch": 1.127597218561275, "grad_norm": 1.1581308841705322, "learning_rate": 0.00018787281027124548, "loss": 4.5347, "step": 604250 }, { "epoch": 1.1276905241206534, "grad_norm": 1.3022631406784058, "learning_rate": 0.00018787080497033353, "loss": 4.4421, "step": 604300 }, { "epoch": 1.127783829680032, "grad_norm": 0.9808488488197327, "learning_rate": 0.00018786879951434508, "loss": 4.5964, "step": 604350 }, { "epoch": 1.1278771352394104, "grad_norm": 1.185690999031067, "learning_rate": 0.00018786679390328366, "loss": 4.5449, "step": 604400 }, { "epoch": 1.127970440798789, "grad_norm": 1.1865053176879883, "learning_rate": 0.00018786478813715278, "loss": 4.4667, "step": 604450 }, { "epoch": 1.1280637463581673, "grad_norm": 1.0674970149993896, "learning_rate": 0.00018786278221595603, "loss": 4.383, "step": 604500 }, { "epoch": 1.128157051917546, "grad_norm": 0.667429506778717, "learning_rate": 0.00018786077613969695, "loss": 4.6612, "step": 604550 }, { "epoch": 1.1282503574769245, "grad_norm": 1.325022578239441, "learning_rate": 0.00018785876990837905, "loss": 4.6758, "step": 604600 }, { "epoch": 1.1283436630363028, "grad_norm": 0.9522653818130493, "learning_rate": 0.00018785676352200587, "loss": 4.5021, "step": 604650 }, { "epoch": 1.1284369685956814, "grad_norm": 1.129472017288208, "learning_rate": 0.00018785475698058094, "loss": 4.5263, "step": 604700 }, { "epoch": 1.1285302741550598, "grad_norm": 1.3725894689559937, "learning_rate": 0.00018785275028410784, "loss": 4.4162, "step": 604750 }, { "epoch": 1.1286235797144384, "grad_norm": 1.0831230878829956, "learning_rate": 0.0001878507434325901, "loss": 4.5329, "step": 604800 }, { "epoch": 1.1287168852738168, "grad_norm": 0.8433416485786438, "learning_rate": 0.00018784873642603125, "loss": 4.3509, "step": 604850 }, { "epoch": 1.1288101908331953, "grad_norm": 1.0095195770263672, "learning_rate": 0.00018784672926443484, "loss": 4.2798, "step": 604900 }, { "epoch": 1.128903496392574, "grad_norm": 1.1606768369674683, "learning_rate": 0.00018784472194780443, "loss": 4.6567, "step": 604950 }, { "epoch": 1.1289968019519523, "grad_norm": 1.2486916780471802, "learning_rate": 0.00018784271447614353, "loss": 4.499, "step": 605000 }, { "epoch": 1.1290901075113309, "grad_norm": 1.2631617784500122, "learning_rate": 0.00018784070684945567, "loss": 4.4072, "step": 605050 }, { "epoch": 1.1291834130707092, "grad_norm": 1.0100871324539185, "learning_rate": 0.00018783869906774446, "loss": 4.3494, "step": 605100 }, { "epoch": 1.1292767186300878, "grad_norm": 1.217388391494751, "learning_rate": 0.00018783669113101335, "loss": 4.6678, "step": 605150 }, { "epoch": 1.1293700241894662, "grad_norm": 0.6118002533912659, "learning_rate": 0.00018783468303926597, "loss": 4.4931, "step": 605200 }, { "epoch": 1.1294633297488448, "grad_norm": 1.3230414390563965, "learning_rate": 0.00018783267479250584, "loss": 4.4835, "step": 605250 }, { "epoch": 1.1295566353082234, "grad_norm": 1.373229742050171, "learning_rate": 0.00018783066639073645, "loss": 4.4997, "step": 605300 }, { "epoch": 1.1296499408676017, "grad_norm": 1.2163255214691162, "learning_rate": 0.00018782865783396142, "loss": 4.5161, "step": 605350 }, { "epoch": 1.1297432464269803, "grad_norm": 1.075547456741333, "learning_rate": 0.00018782664912218427, "loss": 4.3836, "step": 605400 }, { "epoch": 1.1298365519863587, "grad_norm": 1.7416740655899048, "learning_rate": 0.00018782464025540851, "loss": 4.5948, "step": 605450 }, { "epoch": 1.1299298575457373, "grad_norm": 1.199902057647705, "learning_rate": 0.00018782263123363774, "loss": 4.4909, "step": 605500 }, { "epoch": 1.1300231631051156, "grad_norm": 1.2784595489501953, "learning_rate": 0.00018782062205687544, "loss": 4.5692, "step": 605550 }, { "epoch": 1.1301164686644942, "grad_norm": 1.0424398183822632, "learning_rate": 0.00018781861272512527, "loss": 4.5967, "step": 605600 }, { "epoch": 1.1302097742238728, "grad_norm": 1.3738667964935303, "learning_rate": 0.00018781660323839063, "loss": 4.5317, "step": 605650 }, { "epoch": 1.1303030797832512, "grad_norm": 1.1562663316726685, "learning_rate": 0.00018781459359667516, "loss": 4.5801, "step": 605700 }, { "epoch": 1.1303963853426298, "grad_norm": 1.133294939994812, "learning_rate": 0.00018781258379998238, "loss": 4.2804, "step": 605750 }, { "epoch": 1.1304896909020081, "grad_norm": 1.061258316040039, "learning_rate": 0.00018781057384831585, "loss": 4.5971, "step": 605800 }, { "epoch": 1.1305829964613867, "grad_norm": 0.999945342540741, "learning_rate": 0.0001878085637416791, "loss": 4.4398, "step": 605850 }, { "epoch": 1.130676302020765, "grad_norm": 1.0730555057525635, "learning_rate": 0.0001878065534800757, "loss": 4.6836, "step": 605900 }, { "epoch": 1.1307696075801437, "grad_norm": 1.0507534742355347, "learning_rate": 0.00018780454306350917, "loss": 4.5301, "step": 605950 }, { "epoch": 1.1308629131395223, "grad_norm": 1.183223009109497, "learning_rate": 0.00018780253249198309, "loss": 4.5501, "step": 606000 }, { "epoch": 1.1308629131395223, "eval_loss": 4.7015275955200195, "eval_runtime": 230.356, "eval_samples_per_second": 11.322, "eval_steps_per_second": 11.322, "eval_tts_loss": 7.672846266821289, "step": 606000 }, { "epoch": 1.1309562186989006, "grad_norm": 0.7479442358016968, "learning_rate": 0.00018780052176550096, "loss": 4.5553, "step": 606050 }, { "epoch": 1.1310495242582792, "grad_norm": 1.082808017730713, "learning_rate": 0.00018779851088406634, "loss": 4.5574, "step": 606100 }, { "epoch": 1.1311428298176576, "grad_norm": 1.083471655845642, "learning_rate": 0.00018779649984768283, "loss": 4.5221, "step": 606150 }, { "epoch": 1.1312361353770362, "grad_norm": 1.1183828115463257, "learning_rate": 0.00018779448865635393, "loss": 4.6795, "step": 606200 }, { "epoch": 1.1313294409364145, "grad_norm": 1.1436219215393066, "learning_rate": 0.00018779247731008324, "loss": 4.5957, "step": 606250 }, { "epoch": 1.131422746495793, "grad_norm": 0.8406942486763, "learning_rate": 0.00018779046580887426, "loss": 4.2859, "step": 606300 }, { "epoch": 1.1315160520551717, "grad_norm": 0.7894384860992432, "learning_rate": 0.00018778845415273054, "loss": 4.2681, "step": 606350 }, { "epoch": 1.13160935761455, "grad_norm": 1.0419232845306396, "learning_rate": 0.00018778644234165566, "loss": 4.4949, "step": 606400 }, { "epoch": 1.1317026631739286, "grad_norm": 1.1284624338150024, "learning_rate": 0.00018778443037565314, "loss": 4.4958, "step": 606450 }, { "epoch": 1.131795968733307, "grad_norm": 1.4120094776153564, "learning_rate": 0.0001877824182547266, "loss": 4.5959, "step": 606500 }, { "epoch": 1.1318892742926856, "grad_norm": 1.1258330345153809, "learning_rate": 0.00018778040597887947, "loss": 4.7431, "step": 606550 }, { "epoch": 1.131982579852064, "grad_norm": 1.2629408836364746, "learning_rate": 0.0001877783935481154, "loss": 4.469, "step": 606600 }, { "epoch": 1.1320758854114426, "grad_norm": 1.3828574419021606, "learning_rate": 0.00018777638096243788, "loss": 4.5966, "step": 606650 }, { "epoch": 1.1321691909708211, "grad_norm": 1.0701236724853516, "learning_rate": 0.0001877743682218505, "loss": 4.5184, "step": 606700 }, { "epoch": 1.1322624965301995, "grad_norm": 1.3393992185592651, "learning_rate": 0.00018777235532635683, "loss": 4.4886, "step": 606750 }, { "epoch": 1.132355802089578, "grad_norm": 1.0942800045013428, "learning_rate": 0.0001877703422759604, "loss": 4.435, "step": 606800 }, { "epoch": 1.1324491076489565, "grad_norm": 1.2101401090621948, "learning_rate": 0.00018776832907066476, "loss": 4.7042, "step": 606850 }, { "epoch": 1.132542413208335, "grad_norm": 1.0085785388946533, "learning_rate": 0.00018776631571047343, "loss": 4.3839, "step": 606900 }, { "epoch": 1.1326357187677134, "grad_norm": 0.9800994396209717, "learning_rate": 0.00018776430219539, "loss": 4.4621, "step": 606950 }, { "epoch": 1.132729024327092, "grad_norm": 1.103370189666748, "learning_rate": 0.00018776228852541803, "loss": 4.4833, "step": 607000 }, { "epoch": 1.1328223298864706, "grad_norm": 0.9238476157188416, "learning_rate": 0.00018776027470056108, "loss": 4.4295, "step": 607050 }, { "epoch": 1.132915635445849, "grad_norm": 1.0744343996047974, "learning_rate": 0.00018775826072082267, "loss": 4.6098, "step": 607100 }, { "epoch": 1.1330089410052275, "grad_norm": 1.361956238746643, "learning_rate": 0.0001877562465862064, "loss": 4.4254, "step": 607150 }, { "epoch": 1.133102246564606, "grad_norm": 0.9821152091026306, "learning_rate": 0.00018775423229671576, "loss": 4.2684, "step": 607200 }, { "epoch": 1.1331955521239845, "grad_norm": 0.8999354243278503, "learning_rate": 0.00018775221785235436, "loss": 4.1703, "step": 607250 }, { "epoch": 1.1332888576833628, "grad_norm": 1.100506067276001, "learning_rate": 0.0001877502032531257, "loss": 4.5967, "step": 607300 }, { "epoch": 1.1333821632427414, "grad_norm": 0.9767203330993652, "learning_rate": 0.0001877481884990334, "loss": 4.5797, "step": 607350 }, { "epoch": 1.13347546880212, "grad_norm": 1.0538393259048462, "learning_rate": 0.000187746173590081, "loss": 4.5605, "step": 607400 }, { "epoch": 1.1335687743614984, "grad_norm": 1.153192162513733, "learning_rate": 0.00018774415852627202, "loss": 4.6027, "step": 607450 }, { "epoch": 1.133662079920877, "grad_norm": 0.9711982011795044, "learning_rate": 0.00018774214330761005, "loss": 4.4158, "step": 607500 }, { "epoch": 1.1337553854802553, "grad_norm": 1.2521356344223022, "learning_rate": 0.00018774012793409863, "loss": 4.6466, "step": 607550 }, { "epoch": 1.133848691039634, "grad_norm": 0.9963209629058838, "learning_rate": 0.00018773811240574133, "loss": 4.5334, "step": 607600 }, { "epoch": 1.1339419965990123, "grad_norm": 0.9207013845443726, "learning_rate": 0.0001877360967225417, "loss": 4.5883, "step": 607650 }, { "epoch": 1.1340353021583909, "grad_norm": 1.182856798171997, "learning_rate": 0.0001877340808845033, "loss": 4.5714, "step": 607700 }, { "epoch": 1.1341286077177695, "grad_norm": 1.0200783014297485, "learning_rate": 0.0001877320648916297, "loss": 4.4807, "step": 607750 }, { "epoch": 1.1342219132771478, "grad_norm": 0.987350344657898, "learning_rate": 0.0001877300487439244, "loss": 4.6787, "step": 607800 }, { "epoch": 1.1343152188365264, "grad_norm": 1.281598448753357, "learning_rate": 0.000187728032441391, "loss": 4.5009, "step": 607850 }, { "epoch": 1.1344085243959048, "grad_norm": 1.2667522430419922, "learning_rate": 0.0001877260159840331, "loss": 4.5124, "step": 607900 }, { "epoch": 1.1345018299552834, "grad_norm": 1.485671877861023, "learning_rate": 0.0001877239993718542, "loss": 4.7627, "step": 607950 }, { "epoch": 1.1345951355146617, "grad_norm": 1.1045420169830322, "learning_rate": 0.00018772198260485787, "loss": 4.5722, "step": 608000 }, { "epoch": 1.1346884410740403, "grad_norm": 1.4284780025482178, "learning_rate": 0.00018771996568304767, "loss": 4.5157, "step": 608050 }, { "epoch": 1.134781746633419, "grad_norm": 1.0907540321350098, "learning_rate": 0.0001877179486064272, "loss": 4.3916, "step": 608100 }, { "epoch": 1.1348750521927973, "grad_norm": 1.129516363143921, "learning_rate": 0.0001877159313749999, "loss": 4.5008, "step": 608150 }, { "epoch": 1.1349683577521759, "grad_norm": 1.247053623199463, "learning_rate": 0.0001877139139887695, "loss": 4.5083, "step": 608200 }, { "epoch": 1.1350616633115542, "grad_norm": 1.36463463306427, "learning_rate": 0.00018771189644773946, "loss": 4.5478, "step": 608250 }, { "epoch": 1.1351549688709328, "grad_norm": 1.0695486068725586, "learning_rate": 0.00018770987875191335, "loss": 4.4663, "step": 608300 }, { "epoch": 1.1352482744303112, "grad_norm": 1.54143226146698, "learning_rate": 0.00018770786090129473, "loss": 4.5112, "step": 608350 }, { "epoch": 1.1353415799896898, "grad_norm": 1.2490921020507812, "learning_rate": 0.0001877058428958872, "loss": 4.4487, "step": 608400 }, { "epoch": 1.1354348855490684, "grad_norm": 0.9497748613357544, "learning_rate": 0.00018770382473569424, "loss": 4.4389, "step": 608450 }, { "epoch": 1.1355281911084467, "grad_norm": 1.1101946830749512, "learning_rate": 0.00018770180642071948, "loss": 4.6246, "step": 608500 }, { "epoch": 1.1356214966678253, "grad_norm": 1.055654764175415, "learning_rate": 0.00018769978795096648, "loss": 4.4374, "step": 608550 }, { "epoch": 1.1357148022272037, "grad_norm": 1.1176707744598389, "learning_rate": 0.00018769776932643878, "loss": 4.3885, "step": 608600 }, { "epoch": 1.1358081077865823, "grad_norm": 0.8324061632156372, "learning_rate": 0.00018769575054713994, "loss": 4.5738, "step": 608650 }, { "epoch": 1.1359014133459606, "grad_norm": 1.320004940032959, "learning_rate": 0.00018769373161307354, "loss": 4.3325, "step": 608700 }, { "epoch": 1.1359947189053392, "grad_norm": 1.3298802375793457, "learning_rate": 0.0001876917125242431, "loss": 4.4437, "step": 608750 }, { "epoch": 1.1360880244647178, "grad_norm": 0.9249049425125122, "learning_rate": 0.00018768969328065226, "loss": 4.6955, "step": 608800 }, { "epoch": 1.1361813300240962, "grad_norm": 1.1302558183670044, "learning_rate": 0.0001876876738823045, "loss": 4.4165, "step": 608850 }, { "epoch": 1.1362746355834747, "grad_norm": 1.1291462182998657, "learning_rate": 0.00018768565432920347, "loss": 4.3581, "step": 608900 }, { "epoch": 1.136367941142853, "grad_norm": 1.2697051763534546, "learning_rate": 0.00018768363462135266, "loss": 4.3262, "step": 608950 }, { "epoch": 1.1364612467022317, "grad_norm": 0.7152708172798157, "learning_rate": 0.00018768161475875568, "loss": 4.6664, "step": 609000 }, { "epoch": 1.1364612467022317, "eval_loss": 4.691771507263184, "eval_runtime": 230.4972, "eval_samples_per_second": 11.315, "eval_steps_per_second": 11.315, "eval_tts_loss": 7.634240982872634, "step": 609000 }, { "epoch": 1.13655455226161, "grad_norm": 0.9137928485870361, "learning_rate": 0.00018767959474141606, "loss": 4.355, "step": 609050 }, { "epoch": 1.1366478578209886, "grad_norm": 1.2737878561019897, "learning_rate": 0.00018767757456933737, "loss": 4.7051, "step": 609100 }, { "epoch": 1.136741163380367, "grad_norm": 0.8820251226425171, "learning_rate": 0.0001876755542425232, "loss": 4.6893, "step": 609150 }, { "epoch": 1.1368344689397456, "grad_norm": 1.217132806777954, "learning_rate": 0.00018767353376097714, "loss": 4.5205, "step": 609200 }, { "epoch": 1.1369277744991242, "grad_norm": 1.2005654573440552, "learning_rate": 0.0001876715131247027, "loss": 4.5415, "step": 609250 }, { "epoch": 1.1370210800585026, "grad_norm": 1.1363874673843384, "learning_rate": 0.00018766949233370346, "loss": 4.7458, "step": 609300 }, { "epoch": 1.1371143856178811, "grad_norm": 0.9925332069396973, "learning_rate": 0.000187667471387983, "loss": 4.3167, "step": 609350 }, { "epoch": 1.1372076911772595, "grad_norm": 1.1521447896957397, "learning_rate": 0.00018766545028754483, "loss": 4.6657, "step": 609400 }, { "epoch": 1.137300996736638, "grad_norm": 1.1831419467926025, "learning_rate": 0.0001876634290323926, "loss": 4.6055, "step": 609450 }, { "epoch": 1.1373943022960165, "grad_norm": 1.3204360008239746, "learning_rate": 0.00018766140762252985, "loss": 4.7332, "step": 609500 }, { "epoch": 1.137487607855395, "grad_norm": 1.2714821100234985, "learning_rate": 0.00018765938605796016, "loss": 4.6429, "step": 609550 }, { "epoch": 1.1375809134147736, "grad_norm": 0.9901278018951416, "learning_rate": 0.00018765736433868705, "loss": 4.5265, "step": 609600 }, { "epoch": 1.137674218974152, "grad_norm": 1.1576309204101562, "learning_rate": 0.0001876553424647141, "loss": 4.5103, "step": 609650 }, { "epoch": 1.1377675245335306, "grad_norm": 1.2979158163070679, "learning_rate": 0.00018765332043604494, "loss": 4.4914, "step": 609700 }, { "epoch": 1.137860830092909, "grad_norm": 1.2533447742462158, "learning_rate": 0.00018765129825268307, "loss": 4.5368, "step": 609750 }, { "epoch": 1.1379541356522875, "grad_norm": 1.0590882301330566, "learning_rate": 0.00018764927591463206, "loss": 4.4487, "step": 609800 }, { "epoch": 1.138047441211666, "grad_norm": 1.1768226623535156, "learning_rate": 0.00018764725342189552, "loss": 4.5124, "step": 609850 }, { "epoch": 1.1381407467710445, "grad_norm": 0.8428975939750671, "learning_rate": 0.000187645230774477, "loss": 4.5174, "step": 609900 }, { "epoch": 1.138234052330423, "grad_norm": 1.176612377166748, "learning_rate": 0.00018764320797238007, "loss": 4.5414, "step": 609950 }, { "epoch": 1.1383273578898014, "grad_norm": 0.9303444027900696, "learning_rate": 0.0001876411850156083, "loss": 4.5409, "step": 610000 }, { "epoch": 1.13842066344918, "grad_norm": 1.0724202394485474, "learning_rate": 0.00018763916190416526, "loss": 4.5961, "step": 610050 }, { "epoch": 1.1385139690085584, "grad_norm": 1.035103678703308, "learning_rate": 0.00018763713863805453, "loss": 4.6885, "step": 610100 }, { "epoch": 1.138607274567937, "grad_norm": 0.8363282680511475, "learning_rate": 0.00018763511521727963, "loss": 4.656, "step": 610150 }, { "epoch": 1.1387005801273153, "grad_norm": 0.9723330736160278, "learning_rate": 0.0001876330916418442, "loss": 4.6266, "step": 610200 }, { "epoch": 1.138793885686694, "grad_norm": 1.2526624202728271, "learning_rate": 0.00018763106791175178, "loss": 4.6739, "step": 610250 }, { "epoch": 1.1388871912460723, "grad_norm": 1.126080870628357, "learning_rate": 0.00018762904402700595, "loss": 4.4145, "step": 610300 }, { "epoch": 1.1389804968054509, "grad_norm": 1.210460901260376, "learning_rate": 0.00018762701998761028, "loss": 4.6973, "step": 610350 }, { "epoch": 1.1390738023648295, "grad_norm": 1.036097526550293, "learning_rate": 0.00018762499579356832, "loss": 4.5181, "step": 610400 }, { "epoch": 1.1391671079242078, "grad_norm": 1.2147551774978638, "learning_rate": 0.00018762297144488366, "loss": 4.504, "step": 610450 }, { "epoch": 1.1392604134835864, "grad_norm": 1.11159086227417, "learning_rate": 0.00018762094694155988, "loss": 4.4098, "step": 610500 }, { "epoch": 1.1393537190429648, "grad_norm": 1.263816237449646, "learning_rate": 0.00018761892228360053, "loss": 4.6489, "step": 610550 }, { "epoch": 1.1394470246023434, "grad_norm": 1.04669189453125, "learning_rate": 0.00018761689747100925, "loss": 4.6829, "step": 610600 }, { "epoch": 1.1395403301617217, "grad_norm": 1.0361238718032837, "learning_rate": 0.00018761487250378952, "loss": 4.6186, "step": 610650 }, { "epoch": 1.1396336357211003, "grad_norm": 0.9696475267410278, "learning_rate": 0.00018761284738194497, "loss": 4.45, "step": 610700 }, { "epoch": 1.139726941280479, "grad_norm": 1.4422982931137085, "learning_rate": 0.00018761082210547918, "loss": 4.5495, "step": 610750 }, { "epoch": 1.1398202468398573, "grad_norm": 0.8988963961601257, "learning_rate": 0.00018760879667439567, "loss": 4.5918, "step": 610800 }, { "epoch": 1.1399135523992359, "grad_norm": 1.176245093345642, "learning_rate": 0.00018760677108869804, "loss": 4.446, "step": 610850 }, { "epoch": 1.1400068579586142, "grad_norm": 1.2040241956710815, "learning_rate": 0.00018760474534838993, "loss": 4.5575, "step": 610900 }, { "epoch": 1.1401001635179928, "grad_norm": 1.147659420967102, "learning_rate": 0.0001876027194534748, "loss": 4.5506, "step": 610950 }, { "epoch": 1.1401934690773712, "grad_norm": 0.8435444235801697, "learning_rate": 0.0001876006934039563, "loss": 4.4131, "step": 611000 }, { "epoch": 1.1402867746367498, "grad_norm": 0.9867866039276123, "learning_rate": 0.000187598667199838, "loss": 4.4497, "step": 611050 }, { "epoch": 1.1403800801961284, "grad_norm": 1.194301962852478, "learning_rate": 0.00018759664084112346, "loss": 4.5459, "step": 611100 }, { "epoch": 1.1404733857555067, "grad_norm": 0.7353188991546631, "learning_rate": 0.00018759461432781629, "loss": 4.3809, "step": 611150 }, { "epoch": 1.1405666913148853, "grad_norm": 1.3131449222564697, "learning_rate": 0.00018759258765992, "loss": 4.7183, "step": 611200 }, { "epoch": 1.1406599968742637, "grad_norm": 0.8323267102241516, "learning_rate": 0.00018759056083743822, "loss": 4.6744, "step": 611250 }, { "epoch": 1.1407533024336423, "grad_norm": 1.152233600616455, "learning_rate": 0.0001875885338603745, "loss": 4.5232, "step": 611300 }, { "epoch": 1.1408466079930206, "grad_norm": 1.0876924991607666, "learning_rate": 0.00018758650672873246, "loss": 4.5585, "step": 611350 }, { "epoch": 1.1409399135523992, "grad_norm": 1.0472588539123535, "learning_rate": 0.00018758447944251563, "loss": 4.6169, "step": 611400 }, { "epoch": 1.1410332191117778, "grad_norm": 1.2931296825408936, "learning_rate": 0.00018758245200172763, "loss": 4.5532, "step": 611450 }, { "epoch": 1.1411265246711562, "grad_norm": 1.2079931497573853, "learning_rate": 0.00018758042440637196, "loss": 4.3029, "step": 611500 }, { "epoch": 1.1412198302305347, "grad_norm": 0.7780976295471191, "learning_rate": 0.0001875783966564523, "loss": 4.5165, "step": 611550 }, { "epoch": 1.1413131357899131, "grad_norm": 1.109825849533081, "learning_rate": 0.00018757636875197213, "loss": 4.5339, "step": 611600 }, { "epoch": 1.1414064413492917, "grad_norm": 1.2638359069824219, "learning_rate": 0.0001875743406929351, "loss": 4.5771, "step": 611650 }, { "epoch": 1.14149974690867, "grad_norm": 1.1832927465438843, "learning_rate": 0.00018757231247934478, "loss": 4.6137, "step": 611700 }, { "epoch": 1.1415930524680487, "grad_norm": 1.265371561050415, "learning_rate": 0.00018757028411120472, "loss": 4.6302, "step": 611750 }, { "epoch": 1.1416863580274272, "grad_norm": 1.2222152948379517, "learning_rate": 0.00018756825558851855, "loss": 4.557, "step": 611800 }, { "epoch": 1.1417796635868056, "grad_norm": 1.111236572265625, "learning_rate": 0.0001875662269112898, "loss": 4.5128, "step": 611850 }, { "epoch": 1.1418729691461842, "grad_norm": 1.1465693712234497, "learning_rate": 0.00018756419807952206, "loss": 4.5941, "step": 611900 }, { "epoch": 1.1419662747055626, "grad_norm": 1.2291172742843628, "learning_rate": 0.0001875621690932189, "loss": 4.4845, "step": 611950 }, { "epoch": 1.1420595802649411, "grad_norm": 0.7344144582748413, "learning_rate": 0.00018756013995238394, "loss": 4.6223, "step": 612000 }, { "epoch": 1.1420595802649411, "eval_loss": 4.682921409606934, "eval_runtime": 230.25, "eval_samples_per_second": 11.327, "eval_steps_per_second": 11.327, "eval_tts_loss": 7.65676561401917, "step": 612000 }, { "epoch": 1.1421528858243195, "grad_norm": 1.2309528589248657, "learning_rate": 0.00018755811065702073, "loss": 4.3926, "step": 612050 }, { "epoch": 1.142246191383698, "grad_norm": 0.9850509166717529, "learning_rate": 0.0001875560812071329, "loss": 4.3799, "step": 612100 }, { "epoch": 1.1423394969430767, "grad_norm": 1.2319365739822388, "learning_rate": 0.00018755405160272393, "loss": 4.5618, "step": 612150 }, { "epoch": 1.142432802502455, "grad_norm": 0.9990773797035217, "learning_rate": 0.00018755202184379753, "loss": 4.8161, "step": 612200 }, { "epoch": 1.1425261080618336, "grad_norm": 1.155137062072754, "learning_rate": 0.00018754999193035715, "loss": 4.3676, "step": 612250 }, { "epoch": 1.142619413621212, "grad_norm": 1.0722136497497559, "learning_rate": 0.0001875479618624065, "loss": 4.4475, "step": 612300 }, { "epoch": 1.1427127191805906, "grad_norm": 1.0073786973953247, "learning_rate": 0.00018754593163994906, "loss": 4.5232, "step": 612350 }, { "epoch": 1.142806024739969, "grad_norm": 1.2036980390548706, "learning_rate": 0.00018754390126298845, "loss": 4.5771, "step": 612400 }, { "epoch": 1.1428993302993475, "grad_norm": 1.1916866302490234, "learning_rate": 0.0001875418707315283, "loss": 4.4606, "step": 612450 }, { "epoch": 1.1429926358587261, "grad_norm": 1.1823457479476929, "learning_rate": 0.00018753984004557213, "loss": 4.6319, "step": 612500 }, { "epoch": 1.1430859414181045, "grad_norm": 1.0949862003326416, "learning_rate": 0.00018753780920512353, "loss": 4.3279, "step": 612550 }, { "epoch": 1.143179246977483, "grad_norm": 1.0443819761276245, "learning_rate": 0.00018753577821018614, "loss": 4.5286, "step": 612600 }, { "epoch": 1.1432725525368614, "grad_norm": 1.1115772724151611, "learning_rate": 0.00018753374706076347, "loss": 4.6439, "step": 612650 }, { "epoch": 1.14336585809624, "grad_norm": 1.2833280563354492, "learning_rate": 0.00018753171575685914, "loss": 4.37, "step": 612700 }, { "epoch": 1.1434591636556184, "grad_norm": 1.302943229675293, "learning_rate": 0.00018752968429847674, "loss": 4.276, "step": 612750 }, { "epoch": 1.143552469214997, "grad_norm": 1.233963966369629, "learning_rate": 0.00018752765268561985, "loss": 4.718, "step": 612800 }, { "epoch": 1.1436457747743756, "grad_norm": 1.103427767753601, "learning_rate": 0.00018752562091829203, "loss": 4.694, "step": 612850 }, { "epoch": 1.143739080333754, "grad_norm": 1.1772546768188477, "learning_rate": 0.00018752358899649693, "loss": 4.4813, "step": 612900 }, { "epoch": 1.1438323858931325, "grad_norm": 1.2482569217681885, "learning_rate": 0.00018752155692023807, "loss": 4.4484, "step": 612950 }, { "epoch": 1.1439256914525109, "grad_norm": 0.869523823261261, "learning_rate": 0.00018751952468951908, "loss": 4.5069, "step": 613000 }, { "epoch": 1.1440189970118895, "grad_norm": 1.2256423234939575, "learning_rate": 0.0001875174923043435, "loss": 4.6591, "step": 613050 }, { "epoch": 1.1441123025712678, "grad_norm": 1.0441380739212036, "learning_rate": 0.00018751545976471498, "loss": 4.5097, "step": 613100 }, { "epoch": 1.1442056081306464, "grad_norm": 1.0771477222442627, "learning_rate": 0.00018751342707063704, "loss": 4.5987, "step": 613150 }, { "epoch": 1.144298913690025, "grad_norm": 1.2491371631622314, "learning_rate": 0.00018751139422211333, "loss": 4.6635, "step": 613200 }, { "epoch": 1.1443922192494034, "grad_norm": 1.2369304895401, "learning_rate": 0.00018750936121914738, "loss": 4.4821, "step": 613250 }, { "epoch": 1.144485524808782, "grad_norm": 1.2457172870635986, "learning_rate": 0.00018750732806174282, "loss": 4.5836, "step": 613300 }, { "epoch": 1.1445788303681603, "grad_norm": 1.1058460474014282, "learning_rate": 0.00018750529474990322, "loss": 4.5343, "step": 613350 }, { "epoch": 1.144672135927539, "grad_norm": 1.413244605064392, "learning_rate": 0.00018750326128363218, "loss": 4.5556, "step": 613400 }, { "epoch": 1.1447654414869173, "grad_norm": 1.1277343034744263, "learning_rate": 0.00018750122766293328, "loss": 4.6488, "step": 613450 }, { "epoch": 1.1448587470462959, "grad_norm": 0.872635006904602, "learning_rate": 0.00018749919388781007, "loss": 4.5788, "step": 613500 }, { "epoch": 1.1449520526056745, "grad_norm": 1.0446231365203857, "learning_rate": 0.00018749715995826623, "loss": 4.478, "step": 613550 }, { "epoch": 1.1450453581650528, "grad_norm": 1.1410351991653442, "learning_rate": 0.00018749512587430526, "loss": 4.7243, "step": 613600 }, { "epoch": 1.1451386637244314, "grad_norm": 1.2358026504516602, "learning_rate": 0.00018749309163593083, "loss": 4.4188, "step": 613650 }, { "epoch": 1.1452319692838098, "grad_norm": 1.231947660446167, "learning_rate": 0.00018749105724314647, "loss": 4.5984, "step": 613700 }, { "epoch": 1.1453252748431884, "grad_norm": 1.465236783027649, "learning_rate": 0.00018748902269595576, "loss": 4.4458, "step": 613750 }, { "epoch": 1.1454185804025667, "grad_norm": 1.1547354459762573, "learning_rate": 0.00018748698799436238, "loss": 4.5298, "step": 613800 }, { "epoch": 1.1455118859619453, "grad_norm": 1.1867533922195435, "learning_rate": 0.0001874849531383698, "loss": 4.4545, "step": 613850 }, { "epoch": 1.145605191521324, "grad_norm": 1.329681634902954, "learning_rate": 0.0001874829181279817, "loss": 4.5342, "step": 613900 }, { "epoch": 1.1456984970807023, "grad_norm": 1.1748394966125488, "learning_rate": 0.00018748088296320162, "loss": 4.5282, "step": 613950 }, { "epoch": 1.1457918026400808, "grad_norm": 1.260264277458191, "learning_rate": 0.0001874788476440332, "loss": 4.5135, "step": 614000 }, { "epoch": 1.1458851081994592, "grad_norm": 1.0510722398757935, "learning_rate": 0.00018747681217048002, "loss": 4.5601, "step": 614050 }, { "epoch": 1.1459784137588378, "grad_norm": 1.2642810344696045, "learning_rate": 0.00018747477654254563, "loss": 4.4109, "step": 614100 }, { "epoch": 1.1460717193182162, "grad_norm": 0.9258463978767395, "learning_rate": 0.00018747274076023364, "loss": 4.5162, "step": 614150 }, { "epoch": 1.1461650248775948, "grad_norm": 1.2207305431365967, "learning_rate": 0.0001874707048235477, "loss": 4.5183, "step": 614200 }, { "epoch": 1.1462583304369733, "grad_norm": 1.0644514560699463, "learning_rate": 0.00018746866873249133, "loss": 4.5486, "step": 614250 }, { "epoch": 1.1463516359963517, "grad_norm": 0.9622443318367004, "learning_rate": 0.0001874666324870681, "loss": 4.4741, "step": 614300 }, { "epoch": 1.1464449415557303, "grad_norm": 1.6566842794418335, "learning_rate": 0.00018746459608728173, "loss": 4.5511, "step": 614350 }, { "epoch": 1.1465382471151087, "grad_norm": 1.1152957677841187, "learning_rate": 0.00018746255953313573, "loss": 4.5737, "step": 614400 }, { "epoch": 1.1466315526744872, "grad_norm": 1.380327820777893, "learning_rate": 0.00018746052282463365, "loss": 4.4215, "step": 614450 }, { "epoch": 1.1467248582338656, "grad_norm": 1.1992549896240234, "learning_rate": 0.00018745848596177917, "loss": 4.6259, "step": 614500 }, { "epoch": 1.1468181637932442, "grad_norm": 1.192407250404358, "learning_rate": 0.00018745644894457587, "loss": 4.4321, "step": 614550 }, { "epoch": 1.1469114693526228, "grad_norm": 1.1134976148605347, "learning_rate": 0.00018745441177302728, "loss": 4.7195, "step": 614600 }, { "epoch": 1.1470047749120011, "grad_norm": 1.190924882888794, "learning_rate": 0.00018745237444713707, "loss": 4.512, "step": 614650 }, { "epoch": 1.1470980804713797, "grad_norm": 1.0396488904953003, "learning_rate": 0.0001874503369669088, "loss": 4.464, "step": 614700 }, { "epoch": 1.147191386030758, "grad_norm": 0.9273954629898071, "learning_rate": 0.0001874482993323461, "loss": 4.5788, "step": 614750 }, { "epoch": 1.1472846915901367, "grad_norm": 0.9697921276092529, "learning_rate": 0.0001874462615434525, "loss": 4.4131, "step": 614800 }, { "epoch": 1.147377997149515, "grad_norm": 1.1421858072280884, "learning_rate": 0.00018744422360023166, "loss": 4.4895, "step": 614850 }, { "epoch": 1.1474713027088936, "grad_norm": 1.1067836284637451, "learning_rate": 0.00018744218550268714, "loss": 4.6036, "step": 614900 }, { "epoch": 1.1475646082682722, "grad_norm": 1.4476970434188843, "learning_rate": 0.00018744014725082255, "loss": 4.5407, "step": 614950 }, { "epoch": 1.1476579138276506, "grad_norm": 0.961185872554779, "learning_rate": 0.00018743810884464147, "loss": 4.6155, "step": 615000 }, { "epoch": 1.1476579138276506, "eval_loss": 4.700410842895508, "eval_runtime": 229.0385, "eval_samples_per_second": 11.387, "eval_steps_per_second": 11.387, "eval_tts_loss": 7.655812186961765, "step": 615000 }, { "epoch": 1.1477512193870292, "grad_norm": 1.2455925941467285, "learning_rate": 0.00018743607028414753, "loss": 4.5082, "step": 615050 }, { "epoch": 1.1478445249464075, "grad_norm": 1.3034168481826782, "learning_rate": 0.00018743403156934433, "loss": 4.3946, "step": 615100 }, { "epoch": 1.1479378305057861, "grad_norm": 1.3057416677474976, "learning_rate": 0.0001874319927002354, "loss": 4.5326, "step": 615150 }, { "epoch": 1.1480311360651645, "grad_norm": 1.1331218481063843, "learning_rate": 0.00018742995367682444, "loss": 4.5062, "step": 615200 }, { "epoch": 1.148124441624543, "grad_norm": 1.2566500902175903, "learning_rate": 0.00018742791449911494, "loss": 4.4944, "step": 615250 }, { "epoch": 1.1482177471839217, "grad_norm": 1.1495364904403687, "learning_rate": 0.00018742587516711062, "loss": 4.4748, "step": 615300 }, { "epoch": 1.1483110527433, "grad_norm": 0.7561204433441162, "learning_rate": 0.00018742383568081498, "loss": 4.5674, "step": 615350 }, { "epoch": 1.1484043583026786, "grad_norm": 1.2406260967254639, "learning_rate": 0.00018742179604023165, "loss": 4.5427, "step": 615400 }, { "epoch": 1.148497663862057, "grad_norm": 1.210116982460022, "learning_rate": 0.0001874197562453642, "loss": 4.476, "step": 615450 }, { "epoch": 1.1485909694214356, "grad_norm": 1.2253526449203491, "learning_rate": 0.00018741771629621632, "loss": 4.5029, "step": 615500 }, { "epoch": 1.148684274980814, "grad_norm": 1.2410143613815308, "learning_rate": 0.00018741567619279153, "loss": 4.5873, "step": 615550 }, { "epoch": 1.1487775805401925, "grad_norm": 1.3258795738220215, "learning_rate": 0.00018741363593509345, "loss": 4.5782, "step": 615600 }, { "epoch": 1.148870886099571, "grad_norm": 0.8600603938102722, "learning_rate": 0.00018741159552312567, "loss": 4.5101, "step": 615650 }, { "epoch": 1.1489641916589495, "grad_norm": 1.1454845666885376, "learning_rate": 0.00018740955495689184, "loss": 4.6981, "step": 615700 }, { "epoch": 1.149057497218328, "grad_norm": 1.0831043720245361, "learning_rate": 0.0001874075142363955, "loss": 4.5874, "step": 615750 }, { "epoch": 1.1491508027777064, "grad_norm": 1.2245137691497803, "learning_rate": 0.00018740547336164028, "loss": 4.4553, "step": 615800 }, { "epoch": 1.149244108337085, "grad_norm": 1.0997858047485352, "learning_rate": 0.0001874034323326298, "loss": 4.5846, "step": 615850 }, { "epoch": 1.1493374138964634, "grad_norm": 1.2813262939453125, "learning_rate": 0.00018740139114936758, "loss": 4.5771, "step": 615900 }, { "epoch": 1.149430719455842, "grad_norm": 1.1390681266784668, "learning_rate": 0.00018739934981185734, "loss": 4.5825, "step": 615950 }, { "epoch": 1.1495240250152206, "grad_norm": 1.0331727266311646, "learning_rate": 0.0001873973083201026, "loss": 4.3818, "step": 616000 }, { "epoch": 1.149617330574599, "grad_norm": 1.009640097618103, "learning_rate": 0.000187395266674107, "loss": 4.6526, "step": 616050 }, { "epoch": 1.1497106361339775, "grad_norm": 1.185707688331604, "learning_rate": 0.00018739322487387416, "loss": 4.4797, "step": 616100 }, { "epoch": 1.1498039416933559, "grad_norm": 1.1602838039398193, "learning_rate": 0.00018739118291940762, "loss": 4.5968, "step": 616150 }, { "epoch": 1.1498972472527345, "grad_norm": 1.028673768043518, "learning_rate": 0.000187389140810711, "loss": 4.4654, "step": 616200 }, { "epoch": 1.1499905528121128, "grad_norm": 1.0361686944961548, "learning_rate": 0.00018738709854778797, "loss": 4.4471, "step": 616250 }, { "epoch": 1.1500838583714914, "grad_norm": 1.3441355228424072, "learning_rate": 0.00018738505613064205, "loss": 4.6134, "step": 616300 }, { "epoch": 1.15017716393087, "grad_norm": 1.0358151197433472, "learning_rate": 0.00018738301355927692, "loss": 4.5793, "step": 616350 }, { "epoch": 1.1502704694902484, "grad_norm": 0.9130428433418274, "learning_rate": 0.00018738097083369613, "loss": 4.5526, "step": 616400 }, { "epoch": 1.150363775049627, "grad_norm": 1.0744229555130005, "learning_rate": 0.0001873789279539033, "loss": 4.6722, "step": 616450 }, { "epoch": 1.1504570806090053, "grad_norm": 1.2532386779785156, "learning_rate": 0.000187376884919902, "loss": 4.6318, "step": 616500 }, { "epoch": 1.150550386168384, "grad_norm": 0.8471574783325195, "learning_rate": 0.00018737484173169593, "loss": 4.6325, "step": 616550 }, { "epoch": 1.1506436917277623, "grad_norm": 1.0985628366470337, "learning_rate": 0.0001873727983892886, "loss": 4.4899, "step": 616600 }, { "epoch": 1.1507369972871409, "grad_norm": 1.3741086721420288, "learning_rate": 0.00018737075489268367, "loss": 4.562, "step": 616650 }, { "epoch": 1.1508303028465194, "grad_norm": 1.0937474966049194, "learning_rate": 0.00018736871124188476, "loss": 4.486, "step": 616700 }, { "epoch": 1.1509236084058978, "grad_norm": 1.0235892534255981, "learning_rate": 0.0001873666674368954, "loss": 4.5033, "step": 616750 }, { "epoch": 1.1510169139652764, "grad_norm": 1.0768550634384155, "learning_rate": 0.00018736462347771927, "loss": 4.5396, "step": 616800 }, { "epoch": 1.1511102195246548, "grad_norm": 1.1266000270843506, "learning_rate": 0.00018736257936435995, "loss": 4.5323, "step": 616850 }, { "epoch": 1.1512035250840333, "grad_norm": 1.0581296682357788, "learning_rate": 0.00018736053509682104, "loss": 4.501, "step": 616900 }, { "epoch": 1.1512968306434117, "grad_norm": 1.297607183456421, "learning_rate": 0.00018735849067510617, "loss": 4.6711, "step": 616950 }, { "epoch": 1.1513901362027903, "grad_norm": 1.1087689399719238, "learning_rate": 0.00018735644609921888, "loss": 4.6925, "step": 617000 }, { "epoch": 1.1514834417621689, "grad_norm": 1.1448590755462646, "learning_rate": 0.0001873544013691629, "loss": 4.4013, "step": 617050 }, { "epoch": 1.1515767473215472, "grad_norm": 1.0479661226272583, "learning_rate": 0.00018735235648494175, "loss": 4.7487, "step": 617100 }, { "epoch": 1.1516700528809258, "grad_norm": 1.1140984296798706, "learning_rate": 0.00018735031144655908, "loss": 4.4289, "step": 617150 }, { "epoch": 1.1517633584403042, "grad_norm": 1.3007330894470215, "learning_rate": 0.00018734826625401848, "loss": 4.4346, "step": 617200 }, { "epoch": 1.1518566639996828, "grad_norm": 1.0906962156295776, "learning_rate": 0.0001873462209073235, "loss": 4.4657, "step": 617250 }, { "epoch": 1.1519499695590611, "grad_norm": 1.1663942337036133, "learning_rate": 0.00018734417540647788, "loss": 4.5744, "step": 617300 }, { "epoch": 1.1520432751184397, "grad_norm": 1.3014216423034668, "learning_rate": 0.00018734212975148512, "loss": 4.4258, "step": 617350 }, { "epoch": 1.1521365806778183, "grad_norm": 0.8901997804641724, "learning_rate": 0.00018734008394234887, "loss": 4.39, "step": 617400 }, { "epoch": 1.1522298862371967, "grad_norm": 1.1264897584915161, "learning_rate": 0.00018733803797907274, "loss": 4.6062, "step": 617450 }, { "epoch": 1.1523231917965753, "grad_norm": 1.4481980800628662, "learning_rate": 0.00018733599186166036, "loss": 4.4793, "step": 617500 }, { "epoch": 1.1524164973559536, "grad_norm": 1.3156102895736694, "learning_rate": 0.00018733394559011533, "loss": 4.5204, "step": 617550 }, { "epoch": 1.1525098029153322, "grad_norm": 0.753998875617981, "learning_rate": 0.00018733189916444123, "loss": 4.4092, "step": 617600 }, { "epoch": 1.1526031084747106, "grad_norm": 1.187789797782898, "learning_rate": 0.0001873298525846417, "loss": 4.5852, "step": 617650 }, { "epoch": 1.1526964140340892, "grad_norm": 1.6532102823257446, "learning_rate": 0.00018732780585072035, "loss": 4.5256, "step": 617700 }, { "epoch": 1.1527897195934678, "grad_norm": 0.7438343167304993, "learning_rate": 0.00018732575896268077, "loss": 4.538, "step": 617750 }, { "epoch": 1.1528830251528461, "grad_norm": 1.3406038284301758, "learning_rate": 0.0001873237119205266, "loss": 4.4097, "step": 617800 }, { "epoch": 1.1529763307122247, "grad_norm": 1.277105689048767, "learning_rate": 0.00018732166472426145, "loss": 4.4144, "step": 617850 }, { "epoch": 1.153069636271603, "grad_norm": 1.0939826965332031, "learning_rate": 0.0001873196173738889, "loss": 4.5798, "step": 617900 }, { "epoch": 1.1531629418309817, "grad_norm": 1.2747819423675537, "learning_rate": 0.00018731756986941264, "loss": 4.6472, "step": 617950 }, { "epoch": 1.15325624739036, "grad_norm": 1.2140191793441772, "learning_rate": 0.00018731552221083622, "loss": 4.3948, "step": 618000 }, { "epoch": 1.15325624739036, "eval_loss": 4.6872944831848145, "eval_runtime": 232.8558, "eval_samples_per_second": 11.2, "eval_steps_per_second": 11.2, "eval_tts_loss": 7.647557245534587, "step": 618000 }, { "epoch": 1.1533495529497386, "grad_norm": 0.8708065152168274, "learning_rate": 0.00018731347439816324, "loss": 4.5055, "step": 618050 }, { "epoch": 1.1534428585091172, "grad_norm": 0.7688286304473877, "learning_rate": 0.00018731142643139736, "loss": 4.4107, "step": 618100 }, { "epoch": 1.1535361640684956, "grad_norm": 1.0506742000579834, "learning_rate": 0.00018730937831054219, "loss": 4.5292, "step": 618150 }, { "epoch": 1.1536294696278742, "grad_norm": 1.2807648181915283, "learning_rate": 0.00018730733003560128, "loss": 4.3593, "step": 618200 }, { "epoch": 1.1537227751872525, "grad_norm": 1.2053073644638062, "learning_rate": 0.00018730528160657833, "loss": 4.4835, "step": 618250 }, { "epoch": 1.1538160807466311, "grad_norm": 1.1898099184036255, "learning_rate": 0.00018730323302347692, "loss": 4.7487, "step": 618300 }, { "epoch": 1.1539093863060095, "grad_norm": 0.9637723565101624, "learning_rate": 0.00018730118428630064, "loss": 4.5794, "step": 618350 }, { "epoch": 1.154002691865388, "grad_norm": 0.7701144218444824, "learning_rate": 0.0001872991353950532, "loss": 4.4799, "step": 618400 }, { "epoch": 1.1540959974247667, "grad_norm": 1.1989245414733887, "learning_rate": 0.00018729708634973808, "loss": 4.4217, "step": 618450 }, { "epoch": 1.154189302984145, "grad_norm": 1.0814086198806763, "learning_rate": 0.00018729503715035896, "loss": 4.5488, "step": 618500 }, { "epoch": 1.1542826085435236, "grad_norm": 0.6153669953346252, "learning_rate": 0.0001872929877969195, "loss": 4.467, "step": 618550 }, { "epoch": 1.154375914102902, "grad_norm": 1.294966459274292, "learning_rate": 0.00018729093828942327, "loss": 4.3907, "step": 618600 }, { "epoch": 1.1544692196622806, "grad_norm": 1.0802162885665894, "learning_rate": 0.00018728888862787387, "loss": 4.4779, "step": 618650 }, { "epoch": 1.154562525221659, "grad_norm": 0.9705920219421387, "learning_rate": 0.00018728683881227495, "loss": 4.564, "step": 618700 }, { "epoch": 1.1546558307810375, "grad_norm": 0.9712350964546204, "learning_rate": 0.00018728478884263012, "loss": 4.5078, "step": 618750 }, { "epoch": 1.154749136340416, "grad_norm": 1.391241431236267, "learning_rate": 0.000187282738718943, "loss": 4.5766, "step": 618800 }, { "epoch": 1.1548424418997945, "grad_norm": 1.1224496364593506, "learning_rate": 0.00018728068844121718, "loss": 4.6898, "step": 618850 }, { "epoch": 1.154935747459173, "grad_norm": 0.9610186219215393, "learning_rate": 0.0001872786380094563, "loss": 4.455, "step": 618900 }, { "epoch": 1.1550290530185514, "grad_norm": 1.0971403121948242, "learning_rate": 0.00018727658742366402, "loss": 4.6606, "step": 618950 }, { "epoch": 1.15512235857793, "grad_norm": 1.1370238065719604, "learning_rate": 0.0001872745366838439, "loss": 4.715, "step": 619000 }, { "epoch": 1.1552156641373084, "grad_norm": 0.7937226295471191, "learning_rate": 0.00018727248578999958, "loss": 4.6, "step": 619050 }, { "epoch": 1.155308969696687, "grad_norm": 1.1171377897262573, "learning_rate": 0.0001872704347421347, "loss": 4.4462, "step": 619100 }, { "epoch": 1.1554022752560655, "grad_norm": 1.0800731182098389, "learning_rate": 0.00018726838354025284, "loss": 4.783, "step": 619150 }, { "epoch": 1.155495580815444, "grad_norm": 1.150365948677063, "learning_rate": 0.0001872663321843576, "loss": 4.4522, "step": 619200 }, { "epoch": 1.1555888863748225, "grad_norm": 1.2535557746887207, "learning_rate": 0.0001872642806744527, "loss": 4.4852, "step": 619250 }, { "epoch": 1.1556821919342009, "grad_norm": 0.9320636987686157, "learning_rate": 0.0001872622290105417, "loss": 4.3424, "step": 619300 }, { "epoch": 1.1557754974935794, "grad_norm": 0.9277698993682861, "learning_rate": 0.00018726017719262816, "loss": 4.5906, "step": 619350 }, { "epoch": 1.1558688030529578, "grad_norm": 1.1980524063110352, "learning_rate": 0.0001872581252207158, "loss": 4.4793, "step": 619400 }, { "epoch": 1.1559621086123364, "grad_norm": 0.6821892261505127, "learning_rate": 0.0001872560730948082, "loss": 4.5399, "step": 619450 }, { "epoch": 1.156055414171715, "grad_norm": 1.203525185585022, "learning_rate": 0.000187254020814909, "loss": 4.5032, "step": 619500 }, { "epoch": 1.1561487197310933, "grad_norm": 1.0589244365692139, "learning_rate": 0.00018725196838102176, "loss": 4.7374, "step": 619550 }, { "epoch": 1.156242025290472, "grad_norm": 1.1813032627105713, "learning_rate": 0.0001872499157931502, "loss": 4.5772, "step": 619600 }, { "epoch": 1.1563353308498503, "grad_norm": 1.217543601989746, "learning_rate": 0.00018724786305129783, "loss": 4.2616, "step": 619650 }, { "epoch": 1.1564286364092289, "grad_norm": 1.097296953201294, "learning_rate": 0.00018724581015546837, "loss": 4.5373, "step": 619700 }, { "epoch": 1.1565219419686072, "grad_norm": 0.9272122979164124, "learning_rate": 0.00018724375710566543, "loss": 4.2882, "step": 619750 }, { "epoch": 1.1566152475279858, "grad_norm": 1.1909784078598022, "learning_rate": 0.0001872417039018926, "loss": 4.6835, "step": 619800 }, { "epoch": 1.1567085530873644, "grad_norm": 1.3474748134613037, "learning_rate": 0.0001872396505441535, "loss": 4.4458, "step": 619850 }, { "epoch": 1.1568018586467428, "grad_norm": 1.4020636081695557, "learning_rate": 0.00018723759703245175, "loss": 4.5763, "step": 619900 }, { "epoch": 1.1568951642061214, "grad_norm": 1.1713743209838867, "learning_rate": 0.00018723554336679102, "loss": 4.501, "step": 619950 }, { "epoch": 1.1569884697654997, "grad_norm": 0.8767468333244324, "learning_rate": 0.00018723348954717488, "loss": 4.3698, "step": 620000 }, { "epoch": 1.1570817753248783, "grad_norm": 1.1057740449905396, "learning_rate": 0.00018723143557360702, "loss": 4.5402, "step": 620050 }, { "epoch": 1.1571750808842567, "grad_norm": 1.2701382637023926, "learning_rate": 0.000187229381446091, "loss": 4.5495, "step": 620100 }, { "epoch": 1.1572683864436353, "grad_norm": 1.0158929824829102, "learning_rate": 0.00018722732716463044, "loss": 4.3744, "step": 620150 }, { "epoch": 1.1573616920030139, "grad_norm": 1.192915916442871, "learning_rate": 0.00018722527272922905, "loss": 4.433, "step": 620200 }, { "epoch": 1.1574549975623922, "grad_norm": 1.166266679763794, "learning_rate": 0.00018722321813989035, "loss": 4.4683, "step": 620250 }, { "epoch": 1.1575483031217708, "grad_norm": 1.3814746141433716, "learning_rate": 0.00018722116339661807, "loss": 4.5925, "step": 620300 }, { "epoch": 1.1576416086811492, "grad_norm": 1.0989406108856201, "learning_rate": 0.00018721910849941575, "loss": 4.537, "step": 620350 }, { "epoch": 1.1577349142405278, "grad_norm": 1.4005439281463623, "learning_rate": 0.00018721705344828705, "loss": 4.3779, "step": 620400 }, { "epoch": 1.1578282197999061, "grad_norm": 1.0557955503463745, "learning_rate": 0.0001872149982432356, "loss": 4.3824, "step": 620450 }, { "epoch": 1.1579215253592847, "grad_norm": 1.0352810621261597, "learning_rate": 0.00018721294288426505, "loss": 4.5471, "step": 620500 }, { "epoch": 1.1580148309186633, "grad_norm": 1.5270755290985107, "learning_rate": 0.00018721088737137898, "loss": 4.6031, "step": 620550 }, { "epoch": 1.1581081364780417, "grad_norm": 1.1993657350540161, "learning_rate": 0.00018720883170458103, "loss": 4.5757, "step": 620600 }, { "epoch": 1.1582014420374203, "grad_norm": 1.0891492366790771, "learning_rate": 0.00018720677588387482, "loss": 4.5928, "step": 620650 }, { "epoch": 1.1582947475967986, "grad_norm": 0.9441148638725281, "learning_rate": 0.00018720471990926404, "loss": 4.1535, "step": 620700 }, { "epoch": 1.1583880531561772, "grad_norm": 0.8947464227676392, "learning_rate": 0.00018720266378075226, "loss": 4.7544, "step": 620750 }, { "epoch": 1.1584813587155556, "grad_norm": 1.0551327466964722, "learning_rate": 0.0001872006074983431, "loss": 4.3539, "step": 620800 }, { "epoch": 1.1585746642749342, "grad_norm": 1.2830688953399658, "learning_rate": 0.00018719855106204026, "loss": 4.6862, "step": 620850 }, { "epoch": 1.1586679698343127, "grad_norm": 0.9703192710876465, "learning_rate": 0.00018719649447184727, "loss": 4.5609, "step": 620900 }, { "epoch": 1.1587612753936911, "grad_norm": 1.1477733850479126, "learning_rate": 0.00018719443772776783, "loss": 4.5064, "step": 620950 }, { "epoch": 1.1588545809530697, "grad_norm": 0.9930626749992371, "learning_rate": 0.00018719238082980555, "loss": 4.5044, "step": 621000 }, { "epoch": 1.1588545809530697, "eval_loss": 4.68816614151001, "eval_runtime": 228.4536, "eval_samples_per_second": 11.416, "eval_steps_per_second": 11.416, "eval_tts_loss": 7.595208236413431, "step": 621000 }, { "epoch": 1.158947886512448, "grad_norm": 1.1157398223876953, "learning_rate": 0.00018719032377796405, "loss": 4.357, "step": 621050 }, { "epoch": 1.1590411920718267, "grad_norm": 1.2488956451416016, "learning_rate": 0.000187188266572247, "loss": 4.5547, "step": 621100 }, { "epoch": 1.159134497631205, "grad_norm": 1.1114578247070312, "learning_rate": 0.00018718620921265796, "loss": 4.6136, "step": 621150 }, { "epoch": 1.1592278031905836, "grad_norm": 1.1520808935165405, "learning_rate": 0.00018718415169920063, "loss": 4.7051, "step": 621200 }, { "epoch": 1.1593211087499622, "grad_norm": 1.5872066020965576, "learning_rate": 0.0001871820940318786, "loss": 4.6092, "step": 621250 }, { "epoch": 1.1594144143093406, "grad_norm": 1.0834479331970215, "learning_rate": 0.0001871800362106955, "loss": 4.4966, "step": 621300 }, { "epoch": 1.1595077198687191, "grad_norm": 1.176805019378662, "learning_rate": 0.00018717797823565498, "loss": 4.39, "step": 621350 }, { "epoch": 1.1596010254280975, "grad_norm": 0.9698250889778137, "learning_rate": 0.0001871759201067607, "loss": 4.5033, "step": 621400 }, { "epoch": 1.159694330987476, "grad_norm": 1.105061650276184, "learning_rate": 0.00018717386182401626, "loss": 4.5323, "step": 621450 }, { "epoch": 1.1597876365468545, "grad_norm": 1.2306482791900635, "learning_rate": 0.00018717180338742527, "loss": 4.559, "step": 621500 }, { "epoch": 1.159880942106233, "grad_norm": 1.346102237701416, "learning_rate": 0.00018716974479699138, "loss": 4.5958, "step": 621550 }, { "epoch": 1.1599742476656116, "grad_norm": 1.09315824508667, "learning_rate": 0.00018716768605271823, "loss": 4.47, "step": 621600 }, { "epoch": 1.16006755322499, "grad_norm": 0.9878866076469421, "learning_rate": 0.00018716562715460948, "loss": 4.5646, "step": 621650 }, { "epoch": 1.1601608587843686, "grad_norm": 1.2461152076721191, "learning_rate": 0.0001871635681026687, "loss": 4.5935, "step": 621700 }, { "epoch": 1.160254164343747, "grad_norm": 1.1942822933197021, "learning_rate": 0.0001871615088968996, "loss": 4.6145, "step": 621750 }, { "epoch": 1.1603474699031255, "grad_norm": 1.075040340423584, "learning_rate": 0.00018715944953730574, "loss": 4.5345, "step": 621800 }, { "epoch": 1.160440775462504, "grad_norm": 1.099146842956543, "learning_rate": 0.00018715739002389079, "loss": 4.685, "step": 621850 }, { "epoch": 1.1605340810218825, "grad_norm": 1.1391459703445435, "learning_rate": 0.0001871553303566584, "loss": 4.4449, "step": 621900 }, { "epoch": 1.160627386581261, "grad_norm": 1.0743235349655151, "learning_rate": 0.00018715327053561216, "loss": 4.4983, "step": 621950 }, { "epoch": 1.1607206921406394, "grad_norm": 1.1742624044418335, "learning_rate": 0.00018715121056075575, "loss": 4.7664, "step": 622000 }, { "epoch": 1.160813997700018, "grad_norm": 1.204901933670044, "learning_rate": 0.0001871491504320928, "loss": 4.5458, "step": 622050 }, { "epoch": 1.1609073032593964, "grad_norm": 1.1302874088287354, "learning_rate": 0.00018714709014962692, "loss": 4.538, "step": 622100 }, { "epoch": 1.161000608818775, "grad_norm": 0.9886438250541687, "learning_rate": 0.00018714502971336173, "loss": 4.5124, "step": 622150 }, { "epoch": 1.1610939143781533, "grad_norm": 1.0302006006240845, "learning_rate": 0.00018714296912330093, "loss": 4.4282, "step": 622200 }, { "epoch": 1.161187219937532, "grad_norm": 1.0567829608917236, "learning_rate": 0.00018714090837944813, "loss": 4.4903, "step": 622250 }, { "epoch": 1.1612805254969105, "grad_norm": 1.0260131359100342, "learning_rate": 0.00018713884748180695, "loss": 4.5158, "step": 622300 }, { "epoch": 1.1613738310562889, "grad_norm": 1.4492963552474976, "learning_rate": 0.00018713678643038101, "loss": 4.5406, "step": 622350 }, { "epoch": 1.1614671366156675, "grad_norm": 1.1444125175476074, "learning_rate": 0.000187134725225174, "loss": 4.3248, "step": 622400 }, { "epoch": 1.1615604421750458, "grad_norm": 1.0207113027572632, "learning_rate": 0.00018713266386618954, "loss": 4.5114, "step": 622450 }, { "epoch": 1.1616537477344244, "grad_norm": 0.9496476650238037, "learning_rate": 0.00018713060235343125, "loss": 4.4146, "step": 622500 }, { "epoch": 1.1617470532938028, "grad_norm": 1.2605677843093872, "learning_rate": 0.00018712854068690278, "loss": 4.3308, "step": 622550 }, { "epoch": 1.1618403588531814, "grad_norm": 0.964164674282074, "learning_rate": 0.00018712647886660777, "loss": 4.5822, "step": 622600 }, { "epoch": 1.16193366441256, "grad_norm": 1.0551470518112183, "learning_rate": 0.00018712441689254985, "loss": 4.6844, "step": 622650 }, { "epoch": 1.1620269699719383, "grad_norm": 1.090302586555481, "learning_rate": 0.0001871223547647327, "loss": 4.3383, "step": 622700 }, { "epoch": 1.162120275531317, "grad_norm": 0.9000625610351562, "learning_rate": 0.00018712029248315985, "loss": 4.6081, "step": 622750 }, { "epoch": 1.1622135810906953, "grad_norm": 0.8118919730186462, "learning_rate": 0.00018711823004783504, "loss": 4.6599, "step": 622800 }, { "epoch": 1.1623068866500739, "grad_norm": 1.025738000869751, "learning_rate": 0.0001871161674587619, "loss": 4.5141, "step": 622850 }, { "epoch": 1.1624001922094522, "grad_norm": 0.9654110670089722, "learning_rate": 0.00018711410471594406, "loss": 4.3628, "step": 622900 }, { "epoch": 1.1624934977688308, "grad_norm": 1.1826362609863281, "learning_rate": 0.00018711204181938513, "loss": 4.5561, "step": 622950 }, { "epoch": 1.1625868033282094, "grad_norm": 1.1639569997787476, "learning_rate": 0.00018710997876908878, "loss": 4.336, "step": 623000 }, { "epoch": 1.1626801088875878, "grad_norm": 1.692498803138733, "learning_rate": 0.00018710791556505866, "loss": 4.6043, "step": 623050 }, { "epoch": 1.1627734144469664, "grad_norm": 0.8245764374732971, "learning_rate": 0.00018710585220729837, "loss": 4.4769, "step": 623100 }, { "epoch": 1.1628667200063447, "grad_norm": 1.1571229696273804, "learning_rate": 0.0001871037886958116, "loss": 4.6883, "step": 623150 }, { "epoch": 1.1629600255657233, "grad_norm": 1.4256987571716309, "learning_rate": 0.00018710172503060197, "loss": 4.566, "step": 623200 }, { "epoch": 1.1630533311251017, "grad_norm": 0.9102997779846191, "learning_rate": 0.0001870996612116731, "loss": 4.3955, "step": 623250 }, { "epoch": 1.1631466366844803, "grad_norm": 1.1691938638687134, "learning_rate": 0.00018709759723902866, "loss": 4.5567, "step": 623300 }, { "epoch": 1.1632399422438588, "grad_norm": 1.0917346477508545, "learning_rate": 0.0001870955331126723, "loss": 4.5901, "step": 623350 }, { "epoch": 1.1633332478032372, "grad_norm": 1.2647641897201538, "learning_rate": 0.00018709346883260766, "loss": 4.3629, "step": 623400 }, { "epoch": 1.1634265533626158, "grad_norm": 1.3919912576675415, "learning_rate": 0.00018709140439883835, "loss": 4.5054, "step": 623450 }, { "epoch": 1.1635198589219942, "grad_norm": 1.3540605306625366, "learning_rate": 0.00018708933981136805, "loss": 4.5296, "step": 623500 }, { "epoch": 1.1636131644813728, "grad_norm": 1.1864969730377197, "learning_rate": 0.00018708727507020037, "loss": 4.3624, "step": 623550 }, { "epoch": 1.1637064700407511, "grad_norm": 1.2075598239898682, "learning_rate": 0.00018708521017533897, "loss": 4.3916, "step": 623600 }, { "epoch": 1.1637997756001297, "grad_norm": 1.0634722709655762, "learning_rate": 0.0001870831451267875, "loss": 4.5213, "step": 623650 }, { "epoch": 1.1638930811595083, "grad_norm": 0.9352198243141174, "learning_rate": 0.0001870810799245496, "loss": 4.3323, "step": 623700 }, { "epoch": 1.1639863867188867, "grad_norm": 1.1010615825653076, "learning_rate": 0.00018707901456862895, "loss": 4.5294, "step": 623750 }, { "epoch": 1.1640796922782652, "grad_norm": 1.1252425909042358, "learning_rate": 0.00018707694905902911, "loss": 4.5528, "step": 623800 }, { "epoch": 1.1641729978376436, "grad_norm": 1.1164824962615967, "learning_rate": 0.00018707488339575386, "loss": 4.6339, "step": 623850 }, { "epoch": 1.1642663033970222, "grad_norm": 1.2948801517486572, "learning_rate": 0.0001870728175788067, "loss": 4.3919, "step": 623900 }, { "epoch": 1.1643596089564006, "grad_norm": 1.2088079452514648, "learning_rate": 0.00018707075160819133, "loss": 4.3922, "step": 623950 }, { "epoch": 1.1644529145157791, "grad_norm": 1.19192373752594, "learning_rate": 0.00018706868548391142, "loss": 4.5254, "step": 624000 }, { "epoch": 1.1644529145157791, "eval_loss": 4.688456058502197, "eval_runtime": 228.3667, "eval_samples_per_second": 11.42, "eval_steps_per_second": 11.42, "eval_tts_loss": 7.671764042354161, "step": 624000 }, { "epoch": 1.1645462200751577, "grad_norm": 1.2204796075820923, "learning_rate": 0.0001870666192059706, "loss": 4.4955, "step": 624050 }, { "epoch": 1.164639525634536, "grad_norm": 0.8462492227554321, "learning_rate": 0.00018706455277437252, "loss": 4.5664, "step": 624100 }, { "epoch": 1.1647328311939147, "grad_norm": 1.3917512893676758, "learning_rate": 0.0001870624861891208, "loss": 4.5525, "step": 624150 }, { "epoch": 1.164826136753293, "grad_norm": 1.2731354236602783, "learning_rate": 0.00018706041945021915, "loss": 4.5055, "step": 624200 }, { "epoch": 1.1649194423126716, "grad_norm": 1.3041512966156006, "learning_rate": 0.00018705835255767114, "loss": 4.4452, "step": 624250 }, { "epoch": 1.16501274787205, "grad_norm": 1.3252314329147339, "learning_rate": 0.00018705628551148049, "loss": 4.5028, "step": 624300 }, { "epoch": 1.1651060534314286, "grad_norm": 1.1419275999069214, "learning_rate": 0.0001870542183116508, "loss": 4.6232, "step": 624350 }, { "epoch": 1.1651993589908072, "grad_norm": 1.4463404417037964, "learning_rate": 0.00018705215095818573, "loss": 4.4989, "step": 624400 }, { "epoch": 1.1652926645501855, "grad_norm": 1.3596315383911133, "learning_rate": 0.00018705008345108891, "loss": 4.4491, "step": 624450 }, { "epoch": 1.1653859701095641, "grad_norm": 1.133570909500122, "learning_rate": 0.00018704801579036407, "loss": 4.4008, "step": 624500 }, { "epoch": 1.1654792756689425, "grad_norm": 1.3333323001861572, "learning_rate": 0.00018704594797601477, "loss": 4.6438, "step": 624550 }, { "epoch": 1.165572581228321, "grad_norm": 1.2853883504867554, "learning_rate": 0.00018704388000804464, "loss": 4.592, "step": 624600 }, { "epoch": 1.1656658867876994, "grad_norm": 1.4410090446472168, "learning_rate": 0.0001870418118864574, "loss": 4.5424, "step": 624650 }, { "epoch": 1.165759192347078, "grad_norm": 1.1484107971191406, "learning_rate": 0.0001870397436112567, "loss": 4.6069, "step": 624700 }, { "epoch": 1.1658524979064566, "grad_norm": 1.0502270460128784, "learning_rate": 0.00018703767518244617, "loss": 4.5837, "step": 624750 }, { "epoch": 1.165945803465835, "grad_norm": 1.309407353401184, "learning_rate": 0.00018703560660002948, "loss": 4.4612, "step": 624800 }, { "epoch": 1.1660391090252136, "grad_norm": 1.2566345930099487, "learning_rate": 0.00018703353786401024, "loss": 4.5168, "step": 624850 }, { "epoch": 1.166132414584592, "grad_norm": 0.8360883593559265, "learning_rate": 0.00018703146897439208, "loss": 4.5809, "step": 624900 }, { "epoch": 1.1662257201439705, "grad_norm": 1.4270579814910889, "learning_rate": 0.0001870293999311787, "loss": 4.4183, "step": 624950 }, { "epoch": 1.1663190257033489, "grad_norm": 1.0912750959396362, "learning_rate": 0.00018702733073437376, "loss": 4.5339, "step": 625000 }, { "epoch": 1.1664123312627275, "grad_norm": 1.1434093713760376, "learning_rate": 0.0001870252613839809, "loss": 4.5117, "step": 625050 }, { "epoch": 1.166505636822106, "grad_norm": 1.3263338804244995, "learning_rate": 0.00018702319188000378, "loss": 4.5541, "step": 625100 }, { "epoch": 1.1665989423814844, "grad_norm": 2.091458797454834, "learning_rate": 0.000187021122222446, "loss": 4.3786, "step": 625150 }, { "epoch": 1.166692247940863, "grad_norm": 1.0456491708755493, "learning_rate": 0.0001870190524113113, "loss": 4.6097, "step": 625200 }, { "epoch": 1.1667855535002414, "grad_norm": 1.325098991394043, "learning_rate": 0.00018701698244660326, "loss": 4.7787, "step": 625250 }, { "epoch": 1.16687885905962, "grad_norm": 1.396916151046753, "learning_rate": 0.00018701491232832554, "loss": 4.533, "step": 625300 }, { "epoch": 1.1669721646189983, "grad_norm": 1.046447992324829, "learning_rate": 0.00018701284205648185, "loss": 4.3675, "step": 625350 }, { "epoch": 1.167065470178377, "grad_norm": 0.7528923153877258, "learning_rate": 0.00018701077163107575, "loss": 4.5232, "step": 625400 }, { "epoch": 1.1671587757377555, "grad_norm": 0.7341768741607666, "learning_rate": 0.00018700870105211096, "loss": 4.4043, "step": 625450 }, { "epoch": 1.1672520812971339, "grad_norm": 1.2267897129058838, "learning_rate": 0.00018700663031959115, "loss": 4.5051, "step": 625500 }, { "epoch": 1.1673453868565125, "grad_norm": 1.3222310543060303, "learning_rate": 0.00018700455943351996, "loss": 4.4049, "step": 625550 }, { "epoch": 1.1674386924158908, "grad_norm": 1.395042061805725, "learning_rate": 0.000187002488393901, "loss": 4.541, "step": 625600 }, { "epoch": 1.1675319979752694, "grad_norm": 1.1642966270446777, "learning_rate": 0.00018700041720073792, "loss": 4.48, "step": 625650 }, { "epoch": 1.1676253035346478, "grad_norm": 1.1171382665634155, "learning_rate": 0.00018699834585403447, "loss": 4.635, "step": 625700 }, { "epoch": 1.1677186090940264, "grad_norm": 1.267793893814087, "learning_rate": 0.00018699627435379425, "loss": 4.5138, "step": 625750 }, { "epoch": 1.167811914653405, "grad_norm": 1.0610812902450562, "learning_rate": 0.0001869942027000209, "loss": 4.5238, "step": 625800 }, { "epoch": 1.1679052202127833, "grad_norm": 1.4428682327270508, "learning_rate": 0.00018699213089271808, "loss": 4.3922, "step": 625850 }, { "epoch": 1.167998525772162, "grad_norm": 0.9477778077125549, "learning_rate": 0.00018699005893188945, "loss": 4.6994, "step": 625900 }, { "epoch": 1.1680918313315403, "grad_norm": 1.1446776390075684, "learning_rate": 0.00018698798681753869, "loss": 4.5182, "step": 625950 }, { "epoch": 1.1681851368909189, "grad_norm": 1.4497040510177612, "learning_rate": 0.00018698591454966942, "loss": 4.4964, "step": 626000 }, { "epoch": 1.1682784424502972, "grad_norm": 1.3777141571044922, "learning_rate": 0.0001869838421282853, "loss": 4.4848, "step": 626050 }, { "epoch": 1.1683717480096758, "grad_norm": 1.1912305355072021, "learning_rate": 0.00018698176955339005, "loss": 4.5482, "step": 626100 }, { "epoch": 1.1684650535690544, "grad_norm": 1.2888257503509521, "learning_rate": 0.0001869796968249873, "loss": 4.4472, "step": 626150 }, { "epoch": 1.1685583591284328, "grad_norm": 1.1667802333831787, "learning_rate": 0.00018697762394308062, "loss": 4.6764, "step": 626200 }, { "epoch": 1.1686516646878113, "grad_norm": 1.3154703378677368, "learning_rate": 0.0001869755509076738, "loss": 4.5282, "step": 626250 }, { "epoch": 1.1687449702471897, "grad_norm": 1.3021610975265503, "learning_rate": 0.00018697347771877039, "loss": 4.591, "step": 626300 }, { "epoch": 1.1688382758065683, "grad_norm": 0.9950264096260071, "learning_rate": 0.00018697140437637412, "loss": 4.4462, "step": 626350 }, { "epoch": 1.1689315813659467, "grad_norm": 1.0097334384918213, "learning_rate": 0.0001869693308804886, "loss": 4.4311, "step": 626400 }, { "epoch": 1.1690248869253252, "grad_norm": 1.38573157787323, "learning_rate": 0.00018696725723111755, "loss": 4.3897, "step": 626450 }, { "epoch": 1.1691181924847038, "grad_norm": 1.2172220945358276, "learning_rate": 0.00018696518342826458, "loss": 4.6829, "step": 626500 }, { "epoch": 1.1692114980440822, "grad_norm": 1.5864307880401611, "learning_rate": 0.00018696310947193334, "loss": 4.3908, "step": 626550 }, { "epoch": 1.1693048036034608, "grad_norm": 1.0372898578643799, "learning_rate": 0.00018696103536212753, "loss": 4.3715, "step": 626600 }, { "epoch": 1.1693981091628391, "grad_norm": 1.2939163446426392, "learning_rate": 0.0001869589610988508, "loss": 4.404, "step": 626650 }, { "epoch": 1.1694914147222177, "grad_norm": 1.1738375425338745, "learning_rate": 0.0001869568866821068, "loss": 4.586, "step": 626700 }, { "epoch": 1.169584720281596, "grad_norm": 1.04043447971344, "learning_rate": 0.0001869548121118992, "loss": 4.5291, "step": 626750 }, { "epoch": 1.1696780258409747, "grad_norm": 1.0743727684020996, "learning_rate": 0.00018695273738823163, "loss": 4.506, "step": 626800 }, { "epoch": 1.1697713314003533, "grad_norm": 1.1688910722732544, "learning_rate": 0.0001869506625111078, "loss": 4.5984, "step": 626850 }, { "epoch": 1.1698646369597316, "grad_norm": 1.1010264158248901, "learning_rate": 0.00018694858748053136, "loss": 4.5093, "step": 626900 }, { "epoch": 1.1699579425191102, "grad_norm": 1.2346636056900024, "learning_rate": 0.00018694651229650598, "loss": 4.4325, "step": 626950 }, { "epoch": 1.1700512480784886, "grad_norm": 1.0903149843215942, "learning_rate": 0.00018694443695903528, "loss": 4.6639, "step": 627000 }, { "epoch": 1.1700512480784886, "eval_loss": 4.685917854309082, "eval_runtime": 230.3495, "eval_samples_per_second": 11.322, "eval_steps_per_second": 11.322, "eval_tts_loss": 7.638569358382031, "step": 627000 }, { "epoch": 1.1701445536378672, "grad_norm": 0.8650082349777222, "learning_rate": 0.00018694236146812293, "loss": 4.3599, "step": 627050 }, { "epoch": 1.1702378591972455, "grad_norm": 1.3698625564575195, "learning_rate": 0.00018694028582377263, "loss": 4.2801, "step": 627100 }, { "epoch": 1.1703311647566241, "grad_norm": 1.1924580335617065, "learning_rate": 0.00018693821002598802, "loss": 4.5067, "step": 627150 }, { "epoch": 1.1704244703160027, "grad_norm": 0.9330627918243408, "learning_rate": 0.00018693613407477276, "loss": 4.5711, "step": 627200 }, { "epoch": 1.170517775875381, "grad_norm": 1.1635091304779053, "learning_rate": 0.0001869340579701305, "loss": 4.5354, "step": 627250 }, { "epoch": 1.1706110814347597, "grad_norm": 1.188436508178711, "learning_rate": 0.00018693198171206496, "loss": 4.3329, "step": 627300 }, { "epoch": 1.170704386994138, "grad_norm": 1.1061711311340332, "learning_rate": 0.00018692990530057977, "loss": 4.5564, "step": 627350 }, { "epoch": 1.1707976925535166, "grad_norm": 1.0507713556289673, "learning_rate": 0.0001869278287356786, "loss": 4.513, "step": 627400 }, { "epoch": 1.170890998112895, "grad_norm": 0.7093117833137512, "learning_rate": 0.00018692575201736505, "loss": 4.3519, "step": 627450 }, { "epoch": 1.1709843036722736, "grad_norm": 1.127143383026123, "learning_rate": 0.0001869236751456429, "loss": 4.2411, "step": 627500 }, { "epoch": 1.1710776092316522, "grad_norm": 1.3525944948196411, "learning_rate": 0.00018692159812051573, "loss": 4.4058, "step": 627550 }, { "epoch": 1.1711709147910305, "grad_norm": 0.9845049977302551, "learning_rate": 0.00018691952094198726, "loss": 4.6401, "step": 627600 }, { "epoch": 1.1712642203504091, "grad_norm": 0.8313885927200317, "learning_rate": 0.0001869174436100611, "loss": 4.4736, "step": 627650 }, { "epoch": 1.1713575259097875, "grad_norm": 0.9910776615142822, "learning_rate": 0.00018691536612474097, "loss": 4.5049, "step": 627700 }, { "epoch": 1.171450831469166, "grad_norm": 1.2265032529830933, "learning_rate": 0.00018691328848603048, "loss": 4.4648, "step": 627750 }, { "epoch": 1.1715441370285444, "grad_norm": 1.4783027172088623, "learning_rate": 0.00018691121069393333, "loss": 4.5658, "step": 627800 }, { "epoch": 1.171637442587923, "grad_norm": 1.0806165933609009, "learning_rate": 0.0001869091327484532, "loss": 4.6448, "step": 627850 }, { "epoch": 1.1717307481473016, "grad_norm": 0.8357411026954651, "learning_rate": 0.00018690705464959376, "loss": 4.412, "step": 627900 }, { "epoch": 1.17182405370668, "grad_norm": 1.17483651638031, "learning_rate": 0.00018690497639735866, "loss": 4.6788, "step": 627950 }, { "epoch": 1.1719173592660586, "grad_norm": 1.0239256620407104, "learning_rate": 0.00018690289799175154, "loss": 4.7138, "step": 628000 }, { "epoch": 1.172010664825437, "grad_norm": 1.1558183431625366, "learning_rate": 0.0001869008194327761, "loss": 4.6246, "step": 628050 }, { "epoch": 1.1721039703848155, "grad_norm": 0.933174729347229, "learning_rate": 0.00018689874072043604, "loss": 4.433, "step": 628100 }, { "epoch": 1.1721972759441939, "grad_norm": 1.216267466545105, "learning_rate": 0.00018689666185473498, "loss": 4.4468, "step": 628150 }, { "epoch": 1.1722905815035725, "grad_norm": 1.3252307176589966, "learning_rate": 0.00018689458283567656, "loss": 4.6281, "step": 628200 }, { "epoch": 1.172383887062951, "grad_norm": 1.3511385917663574, "learning_rate": 0.00018689250366326455, "loss": 4.6627, "step": 628250 }, { "epoch": 1.1724771926223294, "grad_norm": 1.330969214439392, "learning_rate": 0.00018689042433750252, "loss": 4.5171, "step": 628300 }, { "epoch": 1.172570498181708, "grad_norm": 1.0245919227600098, "learning_rate": 0.0001868883448583942, "loss": 4.2841, "step": 628350 }, { "epoch": 1.1726638037410864, "grad_norm": 0.9434064030647278, "learning_rate": 0.00018688626522594322, "loss": 4.5311, "step": 628400 }, { "epoch": 1.172757109300465, "grad_norm": 1.1506779193878174, "learning_rate": 0.00018688418544015326, "loss": 4.5433, "step": 628450 }, { "epoch": 1.1728504148598433, "grad_norm": 0.8704435229301453, "learning_rate": 0.00018688210550102806, "loss": 4.5021, "step": 628500 }, { "epoch": 1.172943720419222, "grad_norm": 1.164217233657837, "learning_rate": 0.00018688002540857116, "loss": 4.4987, "step": 628550 }, { "epoch": 1.1730370259786005, "grad_norm": 1.182116150856018, "learning_rate": 0.00018687794516278634, "loss": 4.5154, "step": 628600 }, { "epoch": 1.1731303315379789, "grad_norm": 1.655911922454834, "learning_rate": 0.00018687586476367725, "loss": 4.6261, "step": 628650 }, { "epoch": 1.1732236370973574, "grad_norm": 1.534780502319336, "learning_rate": 0.0001868737842112475, "loss": 4.5485, "step": 628700 }, { "epoch": 1.1733169426567358, "grad_norm": 1.1519715785980225, "learning_rate": 0.00018687170350550084, "loss": 4.6777, "step": 628750 }, { "epoch": 1.1734102482161144, "grad_norm": 1.2488718032836914, "learning_rate": 0.0001868696226464409, "loss": 4.3286, "step": 628800 }, { "epoch": 1.1735035537754928, "grad_norm": 1.3434399366378784, "learning_rate": 0.00018686754163407137, "loss": 4.5028, "step": 628850 }, { "epoch": 1.1735968593348713, "grad_norm": 1.0529520511627197, "learning_rate": 0.00018686546046839587, "loss": 4.5296, "step": 628900 }, { "epoch": 1.17369016489425, "grad_norm": 0.8629308938980103, "learning_rate": 0.00018686337914941816, "loss": 4.591, "step": 628950 }, { "epoch": 1.1737834704536283, "grad_norm": 1.2502009868621826, "learning_rate": 0.00018686129767714186, "loss": 4.4413, "step": 629000 }, { "epoch": 1.1738767760130069, "grad_norm": 1.0293396711349487, "learning_rate": 0.00018685921605157067, "loss": 4.4677, "step": 629050 }, { "epoch": 1.1739700815723852, "grad_norm": 1.0789506435394287, "learning_rate": 0.0001868571342727082, "loss": 4.4874, "step": 629100 }, { "epoch": 1.1740633871317638, "grad_norm": 0.9702855348587036, "learning_rate": 0.0001868550523405582, "loss": 4.4831, "step": 629150 }, { "epoch": 1.1741566926911422, "grad_norm": 0.9111902117729187, "learning_rate": 0.00018685297025512434, "loss": 4.4737, "step": 629200 }, { "epoch": 1.1742499982505208, "grad_norm": 1.263879656791687, "learning_rate": 0.00018685088801641023, "loss": 4.6413, "step": 629250 }, { "epoch": 1.1743433038098994, "grad_norm": 1.296815037727356, "learning_rate": 0.00018684880562441962, "loss": 4.4261, "step": 629300 }, { "epoch": 1.1744366093692777, "grad_norm": 1.3881855010986328, "learning_rate": 0.00018684672307915613, "loss": 4.5561, "step": 629350 }, { "epoch": 1.1745299149286563, "grad_norm": 1.094893455505371, "learning_rate": 0.00018684464038062345, "loss": 4.4806, "step": 629400 }, { "epoch": 1.1746232204880347, "grad_norm": 1.5944145917892456, "learning_rate": 0.00018684255752882524, "loss": 4.5644, "step": 629450 }, { "epoch": 1.1747165260474133, "grad_norm": 1.2475653886795044, "learning_rate": 0.00018684047452376525, "loss": 4.6203, "step": 629500 }, { "epoch": 1.1748098316067916, "grad_norm": 1.1808891296386719, "learning_rate": 0.00018683839136544706, "loss": 4.6237, "step": 629550 }, { "epoch": 1.1749031371661702, "grad_norm": 1.1535227298736572, "learning_rate": 0.0001868363080538744, "loss": 4.535, "step": 629600 }, { "epoch": 1.1749964427255488, "grad_norm": 0.9833834171295166, "learning_rate": 0.00018683422458905095, "loss": 4.4969, "step": 629650 }, { "epoch": 1.1750897482849272, "grad_norm": 1.2283682823181152, "learning_rate": 0.00018683214097098036, "loss": 4.4745, "step": 629700 }, { "epoch": 1.1751830538443058, "grad_norm": 1.1768029928207397, "learning_rate": 0.0001868300571996663, "loss": 4.5676, "step": 629750 }, { "epoch": 1.1752763594036841, "grad_norm": 1.3637018203735352, "learning_rate": 0.0001868279732751125, "loss": 4.5489, "step": 629800 }, { "epoch": 1.1753696649630627, "grad_norm": 1.48409903049469, "learning_rate": 0.0001868258891973226, "loss": 4.5711, "step": 629850 }, { "epoch": 1.175462970522441, "grad_norm": 1.0877732038497925, "learning_rate": 0.0001868238049663003, "loss": 4.5317, "step": 629900 }, { "epoch": 1.1755562760818197, "grad_norm": 0.9844371676445007, "learning_rate": 0.00018682172058204924, "loss": 4.4755, "step": 629950 }, { "epoch": 1.1756495816411983, "grad_norm": 1.1016099452972412, "learning_rate": 0.00018681963604457312, "loss": 4.6641, "step": 630000 }, { "epoch": 1.1756495816411983, "eval_loss": 4.687828063964844, "eval_runtime": 230.4308, "eval_samples_per_second": 11.318, "eval_steps_per_second": 11.318, "eval_tts_loss": 7.659826005504758, "step": 630000 }, { "epoch": 1.1757428872005766, "grad_norm": 1.32335364818573, "learning_rate": 0.0001868175513538756, "loss": 4.5414, "step": 630050 }, { "epoch": 1.1758361927599552, "grad_norm": 1.280318260192871, "learning_rate": 0.00018681546650996042, "loss": 4.4751, "step": 630100 }, { "epoch": 1.1759294983193336, "grad_norm": 0.9461749196052551, "learning_rate": 0.00018681338151283118, "loss": 4.5255, "step": 630150 }, { "epoch": 1.1760228038787122, "grad_norm": 1.4637399911880493, "learning_rate": 0.0001868112963624916, "loss": 4.5148, "step": 630200 }, { "epoch": 1.1761161094380905, "grad_norm": 1.3530263900756836, "learning_rate": 0.00018680921105894535, "loss": 4.4165, "step": 630250 }, { "epoch": 1.1762094149974691, "grad_norm": 0.9361532330513, "learning_rate": 0.00018680712560219614, "loss": 4.6259, "step": 630300 }, { "epoch": 1.1763027205568477, "grad_norm": 0.8507053852081299, "learning_rate": 0.00018680503999224764, "loss": 4.3507, "step": 630350 }, { "epoch": 1.176396026116226, "grad_norm": 1.6341888904571533, "learning_rate": 0.00018680295422910347, "loss": 4.4632, "step": 630400 }, { "epoch": 1.1764893316756047, "grad_norm": 1.2350000143051147, "learning_rate": 0.0001868008683127674, "loss": 4.5247, "step": 630450 }, { "epoch": 1.176582637234983, "grad_norm": 1.122847557067871, "learning_rate": 0.00018679878224324306, "loss": 4.747, "step": 630500 }, { "epoch": 1.1766759427943616, "grad_norm": 0.9010679721832275, "learning_rate": 0.00018679669602053415, "loss": 4.6002, "step": 630550 }, { "epoch": 1.17676924835374, "grad_norm": 1.077505111694336, "learning_rate": 0.00018679460964464434, "loss": 4.6072, "step": 630600 }, { "epoch": 1.1768625539131186, "grad_norm": 1.306441307067871, "learning_rate": 0.0001867925231155773, "loss": 4.3944, "step": 630650 }, { "epoch": 1.176955859472497, "grad_norm": 1.3105803728103638, "learning_rate": 0.00018679043643333673, "loss": 4.4313, "step": 630700 }, { "epoch": 1.1770491650318755, "grad_norm": 1.147951602935791, "learning_rate": 0.00018678834959792633, "loss": 4.5791, "step": 630750 }, { "epoch": 1.177142470591254, "grad_norm": 1.1021790504455566, "learning_rate": 0.00018678626260934975, "loss": 4.4727, "step": 630800 }, { "epoch": 1.1772357761506325, "grad_norm": 1.1899380683898926, "learning_rate": 0.0001867841754676107, "loss": 4.6577, "step": 630850 }, { "epoch": 1.177329081710011, "grad_norm": 1.2117353677749634, "learning_rate": 0.00018678208817271284, "loss": 4.5184, "step": 630900 }, { "epoch": 1.1774223872693894, "grad_norm": 1.0587164163589478, "learning_rate": 0.00018678000072465983, "loss": 4.5673, "step": 630950 }, { "epoch": 1.177515692828768, "grad_norm": 1.0936379432678223, "learning_rate": 0.00018677791312345546, "loss": 4.4174, "step": 631000 }, { "epoch": 1.1776089983881464, "grad_norm": 1.2785255908966064, "learning_rate": 0.0001867758253691033, "loss": 4.534, "step": 631050 }, { "epoch": 1.177702303947525, "grad_norm": 1.451784610748291, "learning_rate": 0.00018677373746160708, "loss": 4.5272, "step": 631100 }, { "epoch": 1.1777956095069035, "grad_norm": 1.236657977104187, "learning_rate": 0.00018677164940097046, "loss": 4.5478, "step": 631150 }, { "epoch": 1.177888915066282, "grad_norm": 1.5764262676239014, "learning_rate": 0.00018676956118719718, "loss": 4.5279, "step": 631200 }, { "epoch": 1.1779822206256605, "grad_norm": 1.2897465229034424, "learning_rate": 0.00018676747282029088, "loss": 4.3739, "step": 631250 }, { "epoch": 1.1780755261850389, "grad_norm": 1.0805246829986572, "learning_rate": 0.00018676538430025525, "loss": 4.6521, "step": 631300 }, { "epoch": 1.1781688317444174, "grad_norm": 1.0308562517166138, "learning_rate": 0.000186763295627094, "loss": 4.4967, "step": 631350 }, { "epoch": 1.1782621373037958, "grad_norm": 1.011112093925476, "learning_rate": 0.0001867612068008108, "loss": 4.4429, "step": 631400 }, { "epoch": 1.1783554428631744, "grad_norm": 1.2910566329956055, "learning_rate": 0.00018675911782140933, "loss": 4.4184, "step": 631450 }, { "epoch": 1.178448748422553, "grad_norm": 1.271154522895813, "learning_rate": 0.0001867570286888933, "loss": 4.2867, "step": 631500 }, { "epoch": 1.1785420539819313, "grad_norm": 1.1910946369171143, "learning_rate": 0.00018675493940326635, "loss": 4.6765, "step": 631550 }, { "epoch": 1.17863535954131, "grad_norm": 1.13923978805542, "learning_rate": 0.0001867528499645322, "loss": 4.5915, "step": 631600 }, { "epoch": 1.1787286651006883, "grad_norm": 1.2161822319030762, "learning_rate": 0.00018675076037269456, "loss": 4.4745, "step": 631650 }, { "epoch": 1.1788219706600669, "grad_norm": 1.1378246545791626, "learning_rate": 0.00018674867062775708, "loss": 4.3783, "step": 631700 }, { "epoch": 1.1789152762194453, "grad_norm": 1.13596773147583, "learning_rate": 0.00018674658072972346, "loss": 4.4104, "step": 631750 }, { "epoch": 1.1790085817788238, "grad_norm": 1.0933655500411987, "learning_rate": 0.0001867444906785974, "loss": 4.3698, "step": 631800 }, { "epoch": 1.1791018873382022, "grad_norm": 1.3416017293930054, "learning_rate": 0.00018674240047438255, "loss": 4.486, "step": 631850 }, { "epoch": 1.1791951928975808, "grad_norm": 1.3839190006256104, "learning_rate": 0.00018674031011708264, "loss": 4.6027, "step": 631900 }, { "epoch": 1.1792884984569594, "grad_norm": 1.4472228288650513, "learning_rate": 0.00018673821960670133, "loss": 4.6182, "step": 631950 }, { "epoch": 1.1793818040163377, "grad_norm": 1.1989108324050903, "learning_rate": 0.00018673612894324232, "loss": 4.3415, "step": 632000 }, { "epoch": 1.1794751095757163, "grad_norm": 1.2308205366134644, "learning_rate": 0.00018673403812670934, "loss": 4.5, "step": 632050 }, { "epoch": 1.1795684151350947, "grad_norm": 0.9890864491462708, "learning_rate": 0.000186731947157106, "loss": 4.5438, "step": 632100 }, { "epoch": 1.1796617206944733, "grad_norm": 0.8498798608779907, "learning_rate": 0.00018672985603443608, "loss": 4.5242, "step": 632150 }, { "epoch": 1.1797550262538516, "grad_norm": 0.9306506514549255, "learning_rate": 0.00018672776475870323, "loss": 4.37, "step": 632200 }, { "epoch": 1.1798483318132302, "grad_norm": 1.3427760601043701, "learning_rate": 0.00018672567332991109, "loss": 4.598, "step": 632250 }, { "epoch": 1.1799416373726088, "grad_norm": 1.2254829406738281, "learning_rate": 0.00018672358174806343, "loss": 4.474, "step": 632300 }, { "epoch": 1.1800349429319872, "grad_norm": 1.0360194444656372, "learning_rate": 0.00018672149001316388, "loss": 4.6945, "step": 632350 }, { "epoch": 1.1801282484913658, "grad_norm": 1.3377501964569092, "learning_rate": 0.00018671939812521618, "loss": 4.579, "step": 632400 }, { "epoch": 1.1802215540507441, "grad_norm": 0.9272412657737732, "learning_rate": 0.000186717306084224, "loss": 4.6824, "step": 632450 }, { "epoch": 1.1803148596101227, "grad_norm": 1.043984055519104, "learning_rate": 0.000186715213890191, "loss": 4.4595, "step": 632500 }, { "epoch": 1.180408165169501, "grad_norm": 0.9672336578369141, "learning_rate": 0.00018671312154312094, "loss": 4.6217, "step": 632550 }, { "epoch": 1.1805014707288797, "grad_norm": 1.2290335893630981, "learning_rate": 0.0001867110290430175, "loss": 4.6481, "step": 632600 }, { "epoch": 1.1805947762882583, "grad_norm": 1.1284291744232178, "learning_rate": 0.0001867089363898843, "loss": 4.6126, "step": 632650 }, { "epoch": 1.1806880818476366, "grad_norm": 1.1016309261322021, "learning_rate": 0.0001867068435837251, "loss": 4.6133, "step": 632700 }, { "epoch": 1.1807813874070152, "grad_norm": 1.1490650177001953, "learning_rate": 0.00018670475062454358, "loss": 4.46, "step": 632750 }, { "epoch": 1.1808746929663936, "grad_norm": 1.3210769891738892, "learning_rate": 0.00018670265751234344, "loss": 4.4896, "step": 632800 }, { "epoch": 1.1809679985257722, "grad_norm": 1.390998363494873, "learning_rate": 0.00018670056424712836, "loss": 4.4098, "step": 632850 }, { "epoch": 1.1810613040851505, "grad_norm": 1.4368407726287842, "learning_rate": 0.00018669847082890205, "loss": 4.3665, "step": 632900 }, { "epoch": 1.1811546096445291, "grad_norm": 1.0392221212387085, "learning_rate": 0.00018669637725766817, "loss": 4.5518, "step": 632950 }, { "epoch": 1.1812479152039077, "grad_norm": 1.1916841268539429, "learning_rate": 0.00018669428353343048, "loss": 4.4368, "step": 633000 }, { "epoch": 1.1812479152039077, "eval_loss": 4.688902854919434, "eval_runtime": 231.0769, "eval_samples_per_second": 11.286, "eval_steps_per_second": 11.286, "eval_tts_loss": 7.64127273015975, "step": 633000 }, { "epoch": 1.181341220763286, "grad_norm": 0.9333406090736389, "learning_rate": 0.0001866921896561926, "loss": 4.3848, "step": 633050 }, { "epoch": 1.1814345263226647, "grad_norm": 0.9498289227485657, "learning_rate": 0.00018669009562595826, "loss": 4.4564, "step": 633100 }, { "epoch": 1.181527831882043, "grad_norm": 1.1435285806655884, "learning_rate": 0.00018668800144273115, "loss": 4.4738, "step": 633150 }, { "epoch": 1.1816211374414216, "grad_norm": 0.9243069887161255, "learning_rate": 0.000186685907106515, "loss": 4.4057, "step": 633200 }, { "epoch": 1.1817144430008, "grad_norm": 1.1272633075714111, "learning_rate": 0.00018668381261731343, "loss": 4.4557, "step": 633250 }, { "epoch": 1.1818077485601786, "grad_norm": 1.2550169229507446, "learning_rate": 0.00018668171797513023, "loss": 4.3959, "step": 633300 }, { "epoch": 1.1819010541195571, "grad_norm": 1.0256906747817993, "learning_rate": 0.00018667962317996903, "loss": 4.5042, "step": 633350 }, { "epoch": 1.1819943596789355, "grad_norm": 1.1795967817306519, "learning_rate": 0.00018667752823183355, "loss": 4.5363, "step": 633400 }, { "epoch": 1.182087665238314, "grad_norm": 0.8414153456687927, "learning_rate": 0.0001866754331307275, "loss": 4.5539, "step": 633450 }, { "epoch": 1.1821809707976925, "grad_norm": 0.9817190766334534, "learning_rate": 0.00018667333787665452, "loss": 4.4225, "step": 633500 }, { "epoch": 1.182274276357071, "grad_norm": 1.2204008102416992, "learning_rate": 0.00018667124246961837, "loss": 4.4295, "step": 633550 }, { "epoch": 1.1823675819164494, "grad_norm": 1.092909812927246, "learning_rate": 0.00018666914690962274, "loss": 4.3945, "step": 633600 }, { "epoch": 1.182460887475828, "grad_norm": 1.214686632156372, "learning_rate": 0.0001866670511966713, "loss": 4.443, "step": 633650 }, { "epoch": 1.1825541930352066, "grad_norm": 1.0990686416625977, "learning_rate": 0.00018666495533076777, "loss": 4.6563, "step": 633700 }, { "epoch": 1.182647498594585, "grad_norm": 1.1440353393554688, "learning_rate": 0.00018666285931191585, "loss": 4.7976, "step": 633750 }, { "epoch": 1.1827408041539635, "grad_norm": 1.0357285737991333, "learning_rate": 0.00018666076314011925, "loss": 4.6582, "step": 633800 }, { "epoch": 1.182834109713342, "grad_norm": 1.0413166284561157, "learning_rate": 0.0001866586668153816, "loss": 4.6646, "step": 633850 }, { "epoch": 1.1829274152727205, "grad_norm": 0.98824143409729, "learning_rate": 0.0001866565703377067, "loss": 4.5218, "step": 633900 }, { "epoch": 1.1830207208320989, "grad_norm": 1.2242239713668823, "learning_rate": 0.00018665447370709817, "loss": 4.6304, "step": 633950 }, { "epoch": 1.1831140263914774, "grad_norm": 0.8873030543327332, "learning_rate": 0.00018665237692355973, "loss": 4.2708, "step": 634000 }, { "epoch": 1.183207331950856, "grad_norm": 1.1717530488967896, "learning_rate": 0.00018665027998709512, "loss": 4.4982, "step": 634050 }, { "epoch": 1.1833006375102344, "grad_norm": 1.2255396842956543, "learning_rate": 0.000186648182897708, "loss": 4.488, "step": 634100 }, { "epoch": 1.183393943069613, "grad_norm": 1.449304223060608, "learning_rate": 0.0001866460856554021, "loss": 4.5819, "step": 634150 }, { "epoch": 1.1834872486289914, "grad_norm": 1.0509963035583496, "learning_rate": 0.0001866439882601811, "loss": 4.4509, "step": 634200 }, { "epoch": 1.18358055418837, "grad_norm": 1.2923243045806885, "learning_rate": 0.00018664189071204866, "loss": 4.5695, "step": 634250 }, { "epoch": 1.1836738597477483, "grad_norm": 1.1941941976547241, "learning_rate": 0.0001866397930110086, "loss": 4.4078, "step": 634300 }, { "epoch": 1.183767165307127, "grad_norm": 1.059219241142273, "learning_rate": 0.0001866376951570645, "loss": 4.6571, "step": 634350 }, { "epoch": 1.1838604708665055, "grad_norm": 1.1892679929733276, "learning_rate": 0.00018663559715022013, "loss": 4.4474, "step": 634400 }, { "epoch": 1.1839537764258838, "grad_norm": 1.2145761251449585, "learning_rate": 0.00018663349899047916, "loss": 4.5541, "step": 634450 }, { "epoch": 1.1840470819852624, "grad_norm": 1.4189577102661133, "learning_rate": 0.00018663140067784535, "loss": 4.5149, "step": 634500 }, { "epoch": 1.1841403875446408, "grad_norm": 1.3423880338668823, "learning_rate": 0.00018662930221232234, "loss": 4.4842, "step": 634550 }, { "epoch": 1.1842336931040194, "grad_norm": 1.2143076658248901, "learning_rate": 0.00018662720359391384, "loss": 4.6325, "step": 634600 }, { "epoch": 1.1843269986633977, "grad_norm": 1.1192318201065063, "learning_rate": 0.00018662510482262356, "loss": 4.4928, "step": 634650 }, { "epoch": 1.1844203042227763, "grad_norm": 1.1620713472366333, "learning_rate": 0.00018662300589845522, "loss": 4.3818, "step": 634700 }, { "epoch": 1.184513609782155, "grad_norm": 1.1683677434921265, "learning_rate": 0.00018662090682141256, "loss": 4.3924, "step": 634750 }, { "epoch": 1.1846069153415333, "grad_norm": 1.271966576576233, "learning_rate": 0.0001866188075914992, "loss": 4.5939, "step": 634800 }, { "epoch": 1.1847002209009119, "grad_norm": 1.0893933773040771, "learning_rate": 0.00018661670820871887, "loss": 4.3896, "step": 634850 }, { "epoch": 1.1847935264602902, "grad_norm": 1.0636085271835327, "learning_rate": 0.0001866146086730753, "loss": 4.2989, "step": 634900 }, { "epoch": 1.1848868320196688, "grad_norm": 0.9339725971221924, "learning_rate": 0.00018661250898457224, "loss": 4.4232, "step": 634950 }, { "epoch": 1.1849801375790472, "grad_norm": 0.9015090465545654, "learning_rate": 0.0001866104091432133, "loss": 4.2969, "step": 635000 }, { "epoch": 1.1850734431384258, "grad_norm": 1.2927316427230835, "learning_rate": 0.00018660830914900221, "loss": 4.5015, "step": 635050 }, { "epoch": 1.1851667486978044, "grad_norm": 1.1513205766677856, "learning_rate": 0.0001866062090019427, "loss": 4.5025, "step": 635100 }, { "epoch": 1.1852600542571827, "grad_norm": 1.4362471103668213, "learning_rate": 0.0001866041087020385, "loss": 4.4194, "step": 635150 }, { "epoch": 1.1853533598165613, "grad_norm": 0.992199182510376, "learning_rate": 0.00018660200824929327, "loss": 4.6202, "step": 635200 }, { "epoch": 1.1854466653759397, "grad_norm": 0.836097776889801, "learning_rate": 0.00018659990764371073, "loss": 4.485, "step": 635250 }, { "epoch": 1.1855399709353183, "grad_norm": 0.9875657558441162, "learning_rate": 0.00018659780688529457, "loss": 4.4638, "step": 635300 }, { "epoch": 1.1856332764946966, "grad_norm": 0.8600675463676453, "learning_rate": 0.00018659570597404855, "loss": 4.4585, "step": 635350 }, { "epoch": 1.1857265820540752, "grad_norm": 0.938561737537384, "learning_rate": 0.00018659360490997634, "loss": 4.5203, "step": 635400 }, { "epoch": 1.1858198876134538, "grad_norm": 1.1716737747192383, "learning_rate": 0.00018659150369308166, "loss": 4.6032, "step": 635450 }, { "epoch": 1.1859131931728322, "grad_norm": 0.7677475214004517, "learning_rate": 0.00018658940232336817, "loss": 4.649, "step": 635500 }, { "epoch": 1.1860064987322108, "grad_norm": 1.020861029624939, "learning_rate": 0.00018658730080083966, "loss": 4.491, "step": 635550 }, { "epoch": 1.1860998042915891, "grad_norm": 1.3872333765029907, "learning_rate": 0.0001865851991254998, "loss": 4.7247, "step": 635600 }, { "epoch": 1.1861931098509677, "grad_norm": 1.2099345922470093, "learning_rate": 0.0001865830972973523, "loss": 4.5397, "step": 635650 }, { "epoch": 1.186286415410346, "grad_norm": 0.9144419431686401, "learning_rate": 0.00018658099531640085, "loss": 4.4519, "step": 635700 }, { "epoch": 1.1863797209697247, "grad_norm": 0.8962127566337585, "learning_rate": 0.0001865788931826492, "loss": 4.4397, "step": 635750 }, { "epoch": 1.1864730265291032, "grad_norm": 1.0530954599380493, "learning_rate": 0.000186576790896101, "loss": 4.5929, "step": 635800 }, { "epoch": 1.1865663320884816, "grad_norm": 1.3043595552444458, "learning_rate": 0.00018657468845676003, "loss": 4.3913, "step": 635850 }, { "epoch": 1.1866596376478602, "grad_norm": 1.742296576499939, "learning_rate": 0.00018657258586462995, "loss": 4.5339, "step": 635900 }, { "epoch": 1.1867529432072386, "grad_norm": 1.2760881185531616, "learning_rate": 0.0001865704831197145, "loss": 4.5408, "step": 635950 }, { "epoch": 1.1868462487666172, "grad_norm": 1.0376005172729492, "learning_rate": 0.00018656838022201736, "loss": 4.5421, "step": 636000 }, { "epoch": 1.1868462487666172, "eval_loss": 4.683443069458008, "eval_runtime": 229.2732, "eval_samples_per_second": 11.375, "eval_steps_per_second": 11.375, "eval_tts_loss": 7.618551781845273, "step": 636000 }, { "epoch": 1.1869395543259955, "grad_norm": 0.9431971907615662, "learning_rate": 0.00018656627717154233, "loss": 4.495, "step": 636050 }, { "epoch": 1.187032859885374, "grad_norm": 0.9043135643005371, "learning_rate": 0.000186564173968293, "loss": 4.592, "step": 636100 }, { "epoch": 1.1871261654447527, "grad_norm": 0.8410301804542542, "learning_rate": 0.00018656207061227312, "loss": 4.5263, "step": 636150 }, { "epoch": 1.187219471004131, "grad_norm": 1.2121303081512451, "learning_rate": 0.00018655996710348642, "loss": 4.5675, "step": 636200 }, { "epoch": 1.1873127765635096, "grad_norm": 1.0419446229934692, "learning_rate": 0.0001865578634419366, "loss": 4.4579, "step": 636250 }, { "epoch": 1.187406082122888, "grad_norm": 1.323534369468689, "learning_rate": 0.0001865557596276274, "loss": 4.5428, "step": 636300 }, { "epoch": 1.1874993876822666, "grad_norm": 1.2432265281677246, "learning_rate": 0.00018655365566056252, "loss": 4.7112, "step": 636350 }, { "epoch": 1.187592693241645, "grad_norm": 1.028665542602539, "learning_rate": 0.00018655155154074572, "loss": 4.3339, "step": 636400 }, { "epoch": 1.1876859988010235, "grad_norm": 1.3145458698272705, "learning_rate": 0.00018654944726818057, "loss": 4.5174, "step": 636450 }, { "epoch": 1.1877793043604021, "grad_norm": 0.6324059963226318, "learning_rate": 0.00018654734284287092, "loss": 4.5107, "step": 636500 }, { "epoch": 1.1878726099197805, "grad_norm": 0.8802583813667297, "learning_rate": 0.00018654523826482043, "loss": 4.4472, "step": 636550 }, { "epoch": 1.187965915479159, "grad_norm": 0.7566893696784973, "learning_rate": 0.00018654313353403282, "loss": 4.4874, "step": 636600 }, { "epoch": 1.1880592210385374, "grad_norm": 0.9633519053459167, "learning_rate": 0.00018654102865051181, "loss": 4.4236, "step": 636650 }, { "epoch": 1.188152526597916, "grad_norm": 0.7699225544929504, "learning_rate": 0.00018653892361426112, "loss": 4.4859, "step": 636700 }, { "epoch": 1.1882458321572944, "grad_norm": 1.015519380569458, "learning_rate": 0.00018653681842528444, "loss": 4.4494, "step": 636750 }, { "epoch": 1.188339137716673, "grad_norm": 1.1785295009613037, "learning_rate": 0.0001865347130835855, "loss": 4.6189, "step": 636800 }, { "epoch": 1.1884324432760516, "grad_norm": 0.9489648938179016, "learning_rate": 0.00018653260758916802, "loss": 4.63, "step": 636850 }, { "epoch": 1.18852574883543, "grad_norm": 1.11617112159729, "learning_rate": 0.00018653050194203572, "loss": 4.7874, "step": 636900 }, { "epoch": 1.1886190543948085, "grad_norm": 1.1796563863754272, "learning_rate": 0.0001865283961421923, "loss": 4.6092, "step": 636950 }, { "epoch": 1.188712359954187, "grad_norm": 1.2295002937316895, "learning_rate": 0.0001865262901896415, "loss": 4.5414, "step": 637000 }, { "epoch": 1.1888056655135655, "grad_norm": 1.0654597282409668, "learning_rate": 0.00018652418408438703, "loss": 4.6988, "step": 637050 }, { "epoch": 1.1888989710729438, "grad_norm": 1.0731580257415771, "learning_rate": 0.00018652207782643258, "loss": 4.4798, "step": 637100 }, { "epoch": 1.1889922766323224, "grad_norm": 1.036722183227539, "learning_rate": 0.00018651997141578191, "loss": 4.4102, "step": 637150 }, { "epoch": 1.189085582191701, "grad_norm": 1.326063632965088, "learning_rate": 0.0001865178648524387, "loss": 4.3574, "step": 637200 }, { "epoch": 1.1891788877510794, "grad_norm": 1.2792494297027588, "learning_rate": 0.00018651575813640668, "loss": 4.4723, "step": 637250 }, { "epoch": 1.189272193310458, "grad_norm": 1.1374564170837402, "learning_rate": 0.00018651365126768954, "loss": 4.4523, "step": 637300 }, { "epoch": 1.1893654988698363, "grad_norm": 1.108860969543457, "learning_rate": 0.00018651154424629106, "loss": 4.4734, "step": 637350 }, { "epoch": 1.189458804429215, "grad_norm": 1.0970964431762695, "learning_rate": 0.00018650943707221494, "loss": 4.4354, "step": 637400 }, { "epoch": 1.1895521099885933, "grad_norm": 1.0115246772766113, "learning_rate": 0.00018650732974546488, "loss": 4.6248, "step": 637450 }, { "epoch": 1.1896454155479719, "grad_norm": 1.340340495109558, "learning_rate": 0.00018650522226604458, "loss": 4.6512, "step": 637500 }, { "epoch": 1.1897387211073505, "grad_norm": 1.3017529249191284, "learning_rate": 0.0001865031146339578, "loss": 4.3768, "step": 637550 }, { "epoch": 1.1898320266667288, "grad_norm": 1.0981175899505615, "learning_rate": 0.00018650100684920823, "loss": 4.4427, "step": 637600 }, { "epoch": 1.1899253322261074, "grad_norm": 0.799454927444458, "learning_rate": 0.00018649889891179963, "loss": 4.6329, "step": 637650 }, { "epoch": 1.1900186377854858, "grad_norm": 1.2660908699035645, "learning_rate": 0.00018649679082173565, "loss": 4.4005, "step": 637700 }, { "epoch": 1.1901119433448644, "grad_norm": 0.7469118237495422, "learning_rate": 0.0001864946825790201, "loss": 4.7372, "step": 637750 }, { "epoch": 1.1902052489042427, "grad_norm": 1.0295900106430054, "learning_rate": 0.00018649257418365664, "loss": 4.5837, "step": 637800 }, { "epoch": 1.1902985544636213, "grad_norm": 1.9771125316619873, "learning_rate": 0.000186490465635649, "loss": 4.6203, "step": 637850 }, { "epoch": 1.190391860023, "grad_norm": 1.094752550125122, "learning_rate": 0.0001864883569350009, "loss": 4.4023, "step": 637900 }, { "epoch": 1.1904851655823783, "grad_norm": 1.2486549615859985, "learning_rate": 0.0001864862480817161, "loss": 4.548, "step": 637950 }, { "epoch": 1.1905784711417569, "grad_norm": 0.9376929402351379, "learning_rate": 0.00018648413907579824, "loss": 4.51, "step": 638000 }, { "epoch": 1.1906717767011352, "grad_norm": 0.9893847703933716, "learning_rate": 0.00018648202991725113, "loss": 4.4798, "step": 638050 }, { "epoch": 1.1907650822605138, "grad_norm": 1.2601311206817627, "learning_rate": 0.00018647992060607843, "loss": 4.3883, "step": 638100 }, { "epoch": 1.1908583878198922, "grad_norm": 1.0857207775115967, "learning_rate": 0.0001864778111422839, "loss": 4.1466, "step": 638150 }, { "epoch": 1.1909516933792708, "grad_norm": 0.9587647318840027, "learning_rate": 0.00018647570152587127, "loss": 4.4097, "step": 638200 }, { "epoch": 1.1910449989386493, "grad_norm": 1.3716044425964355, "learning_rate": 0.0001864735917568442, "loss": 4.6783, "step": 638250 }, { "epoch": 1.1911383044980277, "grad_norm": 1.2490928173065186, "learning_rate": 0.00018647148183520648, "loss": 4.5904, "step": 638300 }, { "epoch": 1.1912316100574063, "grad_norm": 1.3140032291412354, "learning_rate": 0.0001864693717609618, "loss": 4.5045, "step": 638350 }, { "epoch": 1.1913249156167847, "grad_norm": 0.7622527480125427, "learning_rate": 0.00018646726153411392, "loss": 4.5038, "step": 638400 }, { "epoch": 1.1914182211761633, "grad_norm": 1.2634466886520386, "learning_rate": 0.0001864651511546665, "loss": 4.6212, "step": 638450 }, { "epoch": 1.1915115267355416, "grad_norm": 1.2206952571868896, "learning_rate": 0.00018646304062262331, "loss": 4.4857, "step": 638500 }, { "epoch": 1.1916048322949202, "grad_norm": 1.2637317180633545, "learning_rate": 0.0001864609299379881, "loss": 4.6022, "step": 638550 }, { "epoch": 1.1916981378542988, "grad_norm": 1.7297202348709106, "learning_rate": 0.00018645881910076453, "loss": 4.4121, "step": 638600 }, { "epoch": 1.1917914434136772, "grad_norm": 1.345077633857727, "learning_rate": 0.00018645670811095636, "loss": 4.3847, "step": 638650 }, { "epoch": 1.1918847489730557, "grad_norm": 0.9718067646026611, "learning_rate": 0.00018645459696856734, "loss": 4.2353, "step": 638700 }, { "epoch": 1.191978054532434, "grad_norm": 1.462079644203186, "learning_rate": 0.00018645248567360115, "loss": 4.3518, "step": 638750 }, { "epoch": 1.1920713600918127, "grad_norm": 0.9675108194351196, "learning_rate": 0.00018645037422606152, "loss": 4.513, "step": 638800 }, { "epoch": 1.192164665651191, "grad_norm": 1.374213695526123, "learning_rate": 0.00018644826262595218, "loss": 4.3991, "step": 638850 }, { "epoch": 1.1922579712105696, "grad_norm": 1.3180235624313354, "learning_rate": 0.00018644615087327692, "loss": 4.5309, "step": 638900 }, { "epoch": 1.1923512767699482, "grad_norm": 1.3403595685958862, "learning_rate": 0.0001864440389680394, "loss": 4.7198, "step": 638950 }, { "epoch": 1.1924445823293266, "grad_norm": 0.9950474500656128, "learning_rate": 0.0001864419269102433, "loss": 4.6096, "step": 639000 }, { "epoch": 1.1924445823293266, "eval_loss": 4.674089431762695, "eval_runtime": 229.1198, "eval_samples_per_second": 11.383, "eval_steps_per_second": 11.383, "eval_tts_loss": 7.627736899253278, "step": 639000 }, { "epoch": 1.1925378878887052, "grad_norm": 1.2997337579727173, "learning_rate": 0.00018643981469989248, "loss": 4.6818, "step": 639050 }, { "epoch": 1.1926311934480835, "grad_norm": 0.8451749682426453, "learning_rate": 0.00018643770233699059, "loss": 4.4085, "step": 639100 }, { "epoch": 1.1927244990074621, "grad_norm": 0.9879077076911926, "learning_rate": 0.00018643558982154134, "loss": 4.4177, "step": 639150 }, { "epoch": 1.1928178045668405, "grad_norm": 1.2931820154190063, "learning_rate": 0.00018643347715354848, "loss": 4.4915, "step": 639200 }, { "epoch": 1.192911110126219, "grad_norm": 0.9837812185287476, "learning_rate": 0.00018643136433301575, "loss": 4.6233, "step": 639250 }, { "epoch": 1.1930044156855977, "grad_norm": 1.0005444288253784, "learning_rate": 0.00018642925135994685, "loss": 4.4586, "step": 639300 }, { "epoch": 1.193097721244976, "grad_norm": 0.9264402985572815, "learning_rate": 0.00018642713823434557, "loss": 4.5646, "step": 639350 }, { "epoch": 1.1931910268043546, "grad_norm": 1.3419628143310547, "learning_rate": 0.00018642502495621558, "loss": 4.7517, "step": 639400 }, { "epoch": 1.193284332363733, "grad_norm": 1.182094931602478, "learning_rate": 0.00018642291152556065, "loss": 4.3886, "step": 639450 }, { "epoch": 1.1933776379231116, "grad_norm": 1.251510500907898, "learning_rate": 0.00018642079794238442, "loss": 4.5857, "step": 639500 }, { "epoch": 1.19347094348249, "grad_norm": 1.1486066579818726, "learning_rate": 0.00018641868420669075, "loss": 4.4044, "step": 639550 }, { "epoch": 1.1935642490418685, "grad_norm": 0.9741039276123047, "learning_rate": 0.0001864165703184833, "loss": 4.6049, "step": 639600 }, { "epoch": 1.1936575546012471, "grad_norm": 1.1230623722076416, "learning_rate": 0.0001864144562777658, "loss": 4.5726, "step": 639650 }, { "epoch": 1.1937508601606255, "grad_norm": 1.189998745918274, "learning_rate": 0.00018641234208454196, "loss": 4.6446, "step": 639700 }, { "epoch": 1.193844165720004, "grad_norm": 1.1524838209152222, "learning_rate": 0.0001864102277388156, "loss": 4.498, "step": 639750 }, { "epoch": 1.1939374712793824, "grad_norm": 1.23847496509552, "learning_rate": 0.00018640811324059032, "loss": 4.2032, "step": 639800 }, { "epoch": 1.194030776838761, "grad_norm": 1.0417327880859375, "learning_rate": 0.00018640599858986998, "loss": 4.5617, "step": 639850 }, { "epoch": 1.1941240823981394, "grad_norm": 1.3418307304382324, "learning_rate": 0.00018640388378665827, "loss": 4.2573, "step": 639900 }, { "epoch": 1.194217387957518, "grad_norm": 0.8882644176483154, "learning_rate": 0.00018640176883095885, "loss": 4.3302, "step": 639950 }, { "epoch": 1.1943106935168966, "grad_norm": 1.0943028926849365, "learning_rate": 0.00018639965372277555, "loss": 4.4704, "step": 640000 }, { "epoch": 1.194403999076275, "grad_norm": 1.2041810750961304, "learning_rate": 0.00018639753846211207, "loss": 4.4815, "step": 640050 }, { "epoch": 1.1944973046356535, "grad_norm": 1.4019402265548706, "learning_rate": 0.00018639542304897214, "loss": 4.7007, "step": 640100 }, { "epoch": 1.1945906101950319, "grad_norm": 1.3780162334442139, "learning_rate": 0.00018639330748335944, "loss": 4.7519, "step": 640150 }, { "epoch": 1.1946839157544105, "grad_norm": 1.2929978370666504, "learning_rate": 0.00018639119176527778, "loss": 4.5335, "step": 640200 }, { "epoch": 1.1947772213137888, "grad_norm": 1.2567002773284912, "learning_rate": 0.00018638907589473087, "loss": 4.3567, "step": 640250 }, { "epoch": 1.1948705268731674, "grad_norm": 0.9701261520385742, "learning_rate": 0.00018638695987172244, "loss": 4.4479, "step": 640300 }, { "epoch": 1.194963832432546, "grad_norm": 1.2003233432769775, "learning_rate": 0.00018638484369625627, "loss": 4.6338, "step": 640350 }, { "epoch": 1.1950571379919244, "grad_norm": 0.9812152981758118, "learning_rate": 0.00018638272736833596, "loss": 4.5899, "step": 640400 }, { "epoch": 1.195150443551303, "grad_norm": 0.7968065738677979, "learning_rate": 0.00018638061088796542, "loss": 4.4692, "step": 640450 }, { "epoch": 1.1952437491106813, "grad_norm": 0.9107528328895569, "learning_rate": 0.00018637849425514827, "loss": 4.4474, "step": 640500 }, { "epoch": 1.19533705467006, "grad_norm": 1.1837598085403442, "learning_rate": 0.00018637637746988825, "loss": 4.5779, "step": 640550 }, { "epoch": 1.1954303602294383, "grad_norm": 1.031278371810913, "learning_rate": 0.00018637426053218917, "loss": 4.5573, "step": 640600 }, { "epoch": 1.1955236657888169, "grad_norm": 1.2726558446884155, "learning_rate": 0.0001863721434420547, "loss": 4.4426, "step": 640650 }, { "epoch": 1.1956169713481954, "grad_norm": 1.048248052597046, "learning_rate": 0.0001863700261994886, "loss": 4.3857, "step": 640700 }, { "epoch": 1.1957102769075738, "grad_norm": 1.132348656654358, "learning_rate": 0.00018636790880449458, "loss": 4.5417, "step": 640750 }, { "epoch": 1.1958035824669524, "grad_norm": 0.9132658839225769, "learning_rate": 0.00018636579125707638, "loss": 4.4994, "step": 640800 }, { "epoch": 1.1958968880263308, "grad_norm": 1.1901758909225464, "learning_rate": 0.0001863636735572378, "loss": 4.4676, "step": 640850 }, { "epoch": 1.1959901935857093, "grad_norm": 1.4709508419036865, "learning_rate": 0.00018636155570498253, "loss": 4.5034, "step": 640900 }, { "epoch": 1.1960834991450877, "grad_norm": 1.206506609916687, "learning_rate": 0.0001863594377003143, "loss": 4.7253, "step": 640950 }, { "epoch": 1.1961768047044663, "grad_norm": 0.9764949083328247, "learning_rate": 0.00018635731954323682, "loss": 4.4844, "step": 641000 }, { "epoch": 1.1962701102638449, "grad_norm": 1.4703900814056396, "learning_rate": 0.00018635520123375392, "loss": 4.6105, "step": 641050 }, { "epoch": 1.1963634158232233, "grad_norm": 0.7416175603866577, "learning_rate": 0.00018635308277186927, "loss": 4.4959, "step": 641100 }, { "epoch": 1.1964567213826018, "grad_norm": 0.9862844347953796, "learning_rate": 0.0001863509641575866, "loss": 4.6102, "step": 641150 }, { "epoch": 1.1965500269419802, "grad_norm": 1.1369844675064087, "learning_rate": 0.0001863488453909097, "loss": 4.412, "step": 641200 }, { "epoch": 1.1966433325013588, "grad_norm": 1.3420132398605347, "learning_rate": 0.00018634672647184226, "loss": 4.5957, "step": 641250 }, { "epoch": 1.1967366380607372, "grad_norm": 1.0192725658416748, "learning_rate": 0.00018634460740038806, "loss": 4.6453, "step": 641300 }, { "epoch": 1.1968299436201157, "grad_norm": 0.8492221832275391, "learning_rate": 0.0001863424881765508, "loss": 4.4904, "step": 641350 }, { "epoch": 1.1969232491794943, "grad_norm": 0.8357344269752502, "learning_rate": 0.00018634036880033427, "loss": 4.4341, "step": 641400 }, { "epoch": 1.1970165547388727, "grad_norm": 1.281644582748413, "learning_rate": 0.00018633824927174217, "loss": 4.5818, "step": 641450 }, { "epoch": 1.1971098602982513, "grad_norm": 1.169800877571106, "learning_rate": 0.0001863361295907782, "loss": 4.4623, "step": 641500 }, { "epoch": 1.1972031658576296, "grad_norm": 1.312618374824524, "learning_rate": 0.00018633400975744623, "loss": 4.5136, "step": 641550 }, { "epoch": 1.1972964714170082, "grad_norm": 1.2998690605163574, "learning_rate": 0.00018633188977174988, "loss": 4.5519, "step": 641600 }, { "epoch": 1.1973897769763866, "grad_norm": 0.6514567732810974, "learning_rate": 0.00018632976963369296, "loss": 4.414, "step": 641650 }, { "epoch": 1.1974830825357652, "grad_norm": 1.059252381324768, "learning_rate": 0.00018632764934327917, "loss": 4.6347, "step": 641700 }, { "epoch": 1.1975763880951438, "grad_norm": 1.2648838758468628, "learning_rate": 0.00018632552890051227, "loss": 4.3812, "step": 641750 }, { "epoch": 1.1976696936545221, "grad_norm": 1.9365733861923218, "learning_rate": 0.00018632340830539602, "loss": 4.4074, "step": 641800 }, { "epoch": 1.1977629992139007, "grad_norm": 1.0791640281677246, "learning_rate": 0.00018632128755793412, "loss": 4.6533, "step": 641850 }, { "epoch": 1.197856304773279, "grad_norm": 1.3329881429672241, "learning_rate": 0.00018631916665813033, "loss": 4.5867, "step": 641900 }, { "epoch": 1.1979496103326577, "grad_norm": 1.289642095565796, "learning_rate": 0.0001863170456059884, "loss": 4.4435, "step": 641950 }, { "epoch": 1.198042915892036, "grad_norm": 1.0639313459396362, "learning_rate": 0.0001863149244015121, "loss": 4.5762, "step": 642000 }, { "epoch": 1.198042915892036, "eval_loss": 4.68009614944458, "eval_runtime": 230.2124, "eval_samples_per_second": 11.329, "eval_steps_per_second": 11.329, "eval_tts_loss": 7.625009246201945, "step": 642000 }, { "epoch": 1.1981362214514146, "grad_norm": 1.3434603214263916, "learning_rate": 0.00018631280304470514, "loss": 4.6693, "step": 642050 }, { "epoch": 1.1982295270107932, "grad_norm": 1.0747425556182861, "learning_rate": 0.00018631068153557125, "loss": 4.4763, "step": 642100 }, { "epoch": 1.1983228325701716, "grad_norm": 1.0418719053268433, "learning_rate": 0.0001863085598741142, "loss": 4.4915, "step": 642150 }, { "epoch": 1.1984161381295502, "grad_norm": 1.1566179990768433, "learning_rate": 0.00018630643806033775, "loss": 4.3854, "step": 642200 }, { "epoch": 1.1985094436889285, "grad_norm": 0.8421035408973694, "learning_rate": 0.00018630431609424562, "loss": 4.5536, "step": 642250 }, { "epoch": 1.1986027492483071, "grad_norm": 1.4412914514541626, "learning_rate": 0.00018630219397584155, "loss": 4.5226, "step": 642300 }, { "epoch": 1.1986960548076855, "grad_norm": 1.2333420515060425, "learning_rate": 0.00018630007170512928, "loss": 4.597, "step": 642350 }, { "epoch": 1.198789360367064, "grad_norm": 0.8859036564826965, "learning_rate": 0.00018629794928211256, "loss": 4.6658, "step": 642400 }, { "epoch": 1.1988826659264427, "grad_norm": 0.6754058003425598, "learning_rate": 0.00018629582670679515, "loss": 4.3599, "step": 642450 }, { "epoch": 1.198975971485821, "grad_norm": 1.1753071546554565, "learning_rate": 0.00018629370397918083, "loss": 4.438, "step": 642500 }, { "epoch": 1.1990692770451996, "grad_norm": 1.2624220848083496, "learning_rate": 0.00018629158109927324, "loss": 4.6811, "step": 642550 }, { "epoch": 1.199162582604578, "grad_norm": 1.4596682786941528, "learning_rate": 0.00018628945806707626, "loss": 4.5964, "step": 642600 }, { "epoch": 1.1992558881639566, "grad_norm": 1.2606921195983887, "learning_rate": 0.0001862873348825935, "loss": 4.723, "step": 642650 }, { "epoch": 1.199349193723335, "grad_norm": 1.0022921562194824, "learning_rate": 0.00018628521154582882, "loss": 4.4996, "step": 642700 }, { "epoch": 1.1994424992827135, "grad_norm": 1.5848878622055054, "learning_rate": 0.00018628308805678593, "loss": 4.7207, "step": 642750 }, { "epoch": 1.199535804842092, "grad_norm": 1.357489824295044, "learning_rate": 0.00018628096441546856, "loss": 4.2993, "step": 642800 }, { "epoch": 1.1996291104014705, "grad_norm": 1.1110410690307617, "learning_rate": 0.00018627884062188046, "loss": 4.4176, "step": 642850 }, { "epoch": 1.199722415960849, "grad_norm": 0.7941170334815979, "learning_rate": 0.00018627671667602539, "loss": 4.3954, "step": 642900 }, { "epoch": 1.1998157215202274, "grad_norm": 1.289516806602478, "learning_rate": 0.00018627459257790713, "loss": 4.4728, "step": 642950 }, { "epoch": 1.199909027079606, "grad_norm": 1.2845627069473267, "learning_rate": 0.00018627246832752937, "loss": 4.4889, "step": 643000 }, { "epoch": 1.2000023326389844, "grad_norm": 1.3066474199295044, "learning_rate": 0.00018627034392489588, "loss": 4.4359, "step": 643050 }, { "epoch": 1.200095638198363, "grad_norm": 0.9735157489776611, "learning_rate": 0.0001862682193700104, "loss": 4.4697, "step": 643100 }, { "epoch": 1.2001889437577415, "grad_norm": 1.25910222530365, "learning_rate": 0.0001862660946628767, "loss": 4.557, "step": 643150 }, { "epoch": 1.20028224931712, "grad_norm": 1.2104603052139282, "learning_rate": 0.00018626396980349854, "loss": 4.5888, "step": 643200 }, { "epoch": 1.2003755548764985, "grad_norm": 1.0589752197265625, "learning_rate": 0.00018626184479187964, "loss": 4.385, "step": 643250 }, { "epoch": 1.2004688604358769, "grad_norm": 0.9369708895683289, "learning_rate": 0.00018625971962802376, "loss": 4.5823, "step": 643300 }, { "epoch": 1.2005621659952554, "grad_norm": 1.4143023490905762, "learning_rate": 0.00018625759431193466, "loss": 4.5659, "step": 643350 }, { "epoch": 1.2006554715546338, "grad_norm": 0.9500061869621277, "learning_rate": 0.00018625546884361608, "loss": 4.6306, "step": 643400 }, { "epoch": 1.2007487771140124, "grad_norm": 1.4214580059051514, "learning_rate": 0.00018625334322307176, "loss": 4.5515, "step": 643450 }, { "epoch": 1.200842082673391, "grad_norm": 1.002866268157959, "learning_rate": 0.0001862512174503055, "loss": 4.5969, "step": 643500 }, { "epoch": 1.2009353882327694, "grad_norm": 1.0495461225509644, "learning_rate": 0.000186249091525321, "loss": 4.5478, "step": 643550 }, { "epoch": 1.201028693792148, "grad_norm": 0.9577365517616272, "learning_rate": 0.000186246965448122, "loss": 4.4971, "step": 643600 }, { "epoch": 1.2011219993515263, "grad_norm": 1.3258568048477173, "learning_rate": 0.0001862448392187123, "loss": 4.5503, "step": 643650 }, { "epoch": 1.201215304910905, "grad_norm": 1.2185635566711426, "learning_rate": 0.00018624271283709565, "loss": 4.6723, "step": 643700 }, { "epoch": 1.2013086104702833, "grad_norm": 0.6784278154373169, "learning_rate": 0.0001862405863032758, "loss": 4.5678, "step": 643750 }, { "epoch": 1.2014019160296618, "grad_norm": 1.4083774089813232, "learning_rate": 0.00018623845961725645, "loss": 4.7337, "step": 643800 }, { "epoch": 1.2014952215890404, "grad_norm": 1.197752833366394, "learning_rate": 0.00018623633277904143, "loss": 4.7021, "step": 643850 }, { "epoch": 1.2015885271484188, "grad_norm": 1.1499849557876587, "learning_rate": 0.00018623420578863442, "loss": 4.7168, "step": 643900 }, { "epoch": 1.2016818327077974, "grad_norm": 1.3370985984802246, "learning_rate": 0.00018623207864603924, "loss": 4.4574, "step": 643950 }, { "epoch": 1.2017751382671757, "grad_norm": 1.2331434488296509, "learning_rate": 0.0001862299513512596, "loss": 4.5365, "step": 644000 }, { "epoch": 1.2018684438265543, "grad_norm": 0.8760834336280823, "learning_rate": 0.00018622782390429924, "loss": 4.3921, "step": 644050 }, { "epoch": 1.2019617493859327, "grad_norm": 1.1950548887252808, "learning_rate": 0.00018622569630516198, "loss": 4.6323, "step": 644100 }, { "epoch": 1.2020550549453113, "grad_norm": 1.064375877380371, "learning_rate": 0.00018622356855385155, "loss": 4.6436, "step": 644150 }, { "epoch": 1.2021483605046899, "grad_norm": 1.1184276342391968, "learning_rate": 0.0001862214406503717, "loss": 4.7602, "step": 644200 }, { "epoch": 1.2022416660640682, "grad_norm": 1.3811519145965576, "learning_rate": 0.00018621931259472612, "loss": 4.5759, "step": 644250 }, { "epoch": 1.2023349716234468, "grad_norm": 1.3148369789123535, "learning_rate": 0.00018621718438691865, "loss": 4.5415, "step": 644300 }, { "epoch": 1.2024282771828252, "grad_norm": 0.9589272141456604, "learning_rate": 0.000186215056026953, "loss": 4.5169, "step": 644350 }, { "epoch": 1.2025215827422038, "grad_norm": 1.1583423614501953, "learning_rate": 0.000186212927514833, "loss": 4.3962, "step": 644400 }, { "epoch": 1.2026148883015821, "grad_norm": 1.1043435335159302, "learning_rate": 0.00018621079885056233, "loss": 4.563, "step": 644450 }, { "epoch": 1.2027081938609607, "grad_norm": 0.8919975161552429, "learning_rate": 0.00018620867003414478, "loss": 4.4573, "step": 644500 }, { "epoch": 1.2028014994203393, "grad_norm": 1.100671410560608, "learning_rate": 0.00018620654106558409, "loss": 4.493, "step": 644550 }, { "epoch": 1.2028948049797177, "grad_norm": 0.9695348739624023, "learning_rate": 0.00018620441194488397, "loss": 4.5369, "step": 644600 }, { "epoch": 1.2029881105390963, "grad_norm": 1.6761195659637451, "learning_rate": 0.0001862022826720483, "loss": 4.7408, "step": 644650 }, { "epoch": 1.2030814160984746, "grad_norm": 1.0705463886260986, "learning_rate": 0.00018620015324708074, "loss": 4.4934, "step": 644700 }, { "epoch": 1.2031747216578532, "grad_norm": 0.9039011597633362, "learning_rate": 0.0001861980236699851, "loss": 4.3919, "step": 644750 }, { "epoch": 1.2032680272172316, "grad_norm": 1.176025152206421, "learning_rate": 0.00018619589394076508, "loss": 4.5739, "step": 644800 }, { "epoch": 1.2033613327766102, "grad_norm": 1.1442394256591797, "learning_rate": 0.0001861937640594245, "loss": 4.5566, "step": 644850 }, { "epoch": 1.2034546383359888, "grad_norm": 1.2151931524276733, "learning_rate": 0.00018619163402596706, "loss": 4.5912, "step": 644900 }, { "epoch": 1.2035479438953671, "grad_norm": 1.046729326248169, "learning_rate": 0.00018618950384039658, "loss": 4.5823, "step": 644950 }, { "epoch": 1.2036412494547457, "grad_norm": 0.9931709170341492, "learning_rate": 0.00018618737350271678, "loss": 4.5241, "step": 645000 }, { "epoch": 1.2036412494547457, "eval_loss": 4.687857627868652, "eval_runtime": 231.666, "eval_samples_per_second": 11.258, "eval_steps_per_second": 11.258, "eval_tts_loss": 7.561708564698651, "step": 645000 }, { "epoch": 1.203734555014124, "grad_norm": 1.1435506343841553, "learning_rate": 0.00018618524301293143, "loss": 4.3869, "step": 645050 }, { "epoch": 1.2038278605735027, "grad_norm": 1.269706130027771, "learning_rate": 0.0001861831123710443, "loss": 4.5233, "step": 645100 }, { "epoch": 1.203921166132881, "grad_norm": 1.2369558811187744, "learning_rate": 0.00018618098157705917, "loss": 4.6636, "step": 645150 }, { "epoch": 1.2040144716922596, "grad_norm": 1.0601272583007812, "learning_rate": 0.00018617885063097972, "loss": 4.4793, "step": 645200 }, { "epoch": 1.2041077772516382, "grad_norm": 1.060555338859558, "learning_rate": 0.00018617671953280977, "loss": 4.5582, "step": 645250 }, { "epoch": 1.2042010828110166, "grad_norm": 1.2559823989868164, "learning_rate": 0.0001861745882825531, "loss": 4.5391, "step": 645300 }, { "epoch": 1.2042943883703952, "grad_norm": 1.3860992193222046, "learning_rate": 0.00018617245688021342, "loss": 4.5236, "step": 645350 }, { "epoch": 1.2043876939297735, "grad_norm": 1.1392579078674316, "learning_rate": 0.00018617032532579456, "loss": 4.5494, "step": 645400 }, { "epoch": 1.204480999489152, "grad_norm": 1.3880369663238525, "learning_rate": 0.00018616819361930018, "loss": 4.415, "step": 645450 }, { "epoch": 1.2045743050485305, "grad_norm": 1.227962851524353, "learning_rate": 0.00018616606176073412, "loss": 4.6117, "step": 645500 }, { "epoch": 1.204667610607909, "grad_norm": 1.173925518989563, "learning_rate": 0.00018616392975010014, "loss": 4.4349, "step": 645550 }, { "epoch": 1.2047609161672876, "grad_norm": 1.0447715520858765, "learning_rate": 0.00018616179758740198, "loss": 4.4566, "step": 645600 }, { "epoch": 1.204854221726666, "grad_norm": 0.8806125521659851, "learning_rate": 0.0001861596652726434, "loss": 4.2078, "step": 645650 }, { "epoch": 1.2049475272860446, "grad_norm": 1.0493006706237793, "learning_rate": 0.00018615753280582816, "loss": 4.6754, "step": 645700 }, { "epoch": 1.205040832845423, "grad_norm": 1.0297471284866333, "learning_rate": 0.00018615540018696004, "loss": 4.4924, "step": 645750 }, { "epoch": 1.2051341384048015, "grad_norm": 0.9174513220787048, "learning_rate": 0.0001861532674160428, "loss": 4.5307, "step": 645800 }, { "epoch": 1.20522744396418, "grad_norm": 1.1964694261550903, "learning_rate": 0.0001861511344930802, "loss": 4.4739, "step": 645850 }, { "epoch": 1.2053207495235585, "grad_norm": 0.773618757724762, "learning_rate": 0.000186149001418076, "loss": 4.4253, "step": 645900 }, { "epoch": 1.205414055082937, "grad_norm": 1.0280283689498901, "learning_rate": 0.000186146868191034, "loss": 4.3649, "step": 645950 }, { "epoch": 1.2055073606423155, "grad_norm": 1.2932860851287842, "learning_rate": 0.0001861447348119579, "loss": 4.7017, "step": 646000 }, { "epoch": 1.205600666201694, "grad_norm": 1.3683998584747314, "learning_rate": 0.00018614260128085156, "loss": 4.5167, "step": 646050 }, { "epoch": 1.2056939717610724, "grad_norm": 1.2443735599517822, "learning_rate": 0.00018614046759771863, "loss": 4.4311, "step": 646100 }, { "epoch": 1.205787277320451, "grad_norm": 0.9961922764778137, "learning_rate": 0.00018613833376256294, "loss": 4.4805, "step": 646150 }, { "epoch": 1.2058805828798294, "grad_norm": 1.2461585998535156, "learning_rate": 0.00018613619977538828, "loss": 4.4868, "step": 646200 }, { "epoch": 1.205973888439208, "grad_norm": 1.1577589511871338, "learning_rate": 0.00018613406563619832, "loss": 4.6189, "step": 646250 }, { "epoch": 1.2060671939985865, "grad_norm": 1.440785527229309, "learning_rate": 0.00018613193134499694, "loss": 4.4141, "step": 646300 }, { "epoch": 1.206160499557965, "grad_norm": 0.828487753868103, "learning_rate": 0.00018612979690178785, "loss": 4.5963, "step": 646350 }, { "epoch": 1.2062538051173435, "grad_norm": 1.3241865634918213, "learning_rate": 0.0001861276623065748, "loss": 4.519, "step": 646400 }, { "epoch": 1.2063471106767218, "grad_norm": 1.2797560691833496, "learning_rate": 0.0001861255275593616, "loss": 4.4877, "step": 646450 }, { "epoch": 1.2064404162361004, "grad_norm": 1.1558012962341309, "learning_rate": 0.00018612339266015199, "loss": 4.587, "step": 646500 }, { "epoch": 1.2065337217954788, "grad_norm": 1.1986855268478394, "learning_rate": 0.00018612125760894977, "loss": 4.4407, "step": 646550 }, { "epoch": 1.2066270273548574, "grad_norm": 0.9267823696136475, "learning_rate": 0.00018611912240575865, "loss": 4.5292, "step": 646600 }, { "epoch": 1.206720332914236, "grad_norm": 1.1973270177841187, "learning_rate": 0.00018611698705058247, "loss": 4.6015, "step": 646650 }, { "epoch": 1.2068136384736143, "grad_norm": 0.8140276074409485, "learning_rate": 0.0001861148515434249, "loss": 4.5005, "step": 646700 }, { "epoch": 1.206906944032993, "grad_norm": 1.195936679840088, "learning_rate": 0.00018611271588428983, "loss": 4.5822, "step": 646750 }, { "epoch": 1.2070002495923713, "grad_norm": 0.8443145751953125, "learning_rate": 0.00018611058007318097, "loss": 4.4749, "step": 646800 }, { "epoch": 1.2070935551517499, "grad_norm": 1.1901483535766602, "learning_rate": 0.00018610844411010206, "loss": 4.348, "step": 646850 }, { "epoch": 1.2071868607111282, "grad_norm": 1.4227650165557861, "learning_rate": 0.0001861063079950569, "loss": 4.623, "step": 646900 }, { "epoch": 1.2072801662705068, "grad_norm": 1.3114646673202515, "learning_rate": 0.00018610417172804927, "loss": 4.5334, "step": 646950 }, { "epoch": 1.2073734718298854, "grad_norm": 1.0649514198303223, "learning_rate": 0.0001861020353090829, "loss": 4.669, "step": 647000 }, { "epoch": 1.2074667773892638, "grad_norm": 0.7447493672370911, "learning_rate": 0.0001860998987381616, "loss": 4.4688, "step": 647050 }, { "epoch": 1.2075600829486424, "grad_norm": 1.472016453742981, "learning_rate": 0.00018609776201528914, "loss": 4.4444, "step": 647100 }, { "epoch": 1.2076533885080207, "grad_norm": 1.3249789476394653, "learning_rate": 0.00018609562514046926, "loss": 4.4261, "step": 647150 }, { "epoch": 1.2077466940673993, "grad_norm": 1.0039921998977661, "learning_rate": 0.00018609348811370578, "loss": 4.37, "step": 647200 }, { "epoch": 1.2078399996267777, "grad_norm": 1.2435988187789917, "learning_rate": 0.0001860913509350024, "loss": 4.3972, "step": 647250 }, { "epoch": 1.2079333051861563, "grad_norm": 1.1655999422073364, "learning_rate": 0.00018608921360436297, "loss": 4.5535, "step": 647300 }, { "epoch": 1.2080266107455349, "grad_norm": 1.1999410390853882, "learning_rate": 0.00018608707612179123, "loss": 4.6046, "step": 647350 }, { "epoch": 1.2081199163049132, "grad_norm": 1.3475697040557861, "learning_rate": 0.0001860849384872909, "loss": 4.5672, "step": 647400 }, { "epoch": 1.2082132218642918, "grad_norm": 0.9314871430397034, "learning_rate": 0.00018608280070086587, "loss": 4.2898, "step": 647450 }, { "epoch": 1.2083065274236702, "grad_norm": 1.3992780447006226, "learning_rate": 0.00018608066276251979, "loss": 4.3272, "step": 647500 }, { "epoch": 1.2083998329830488, "grad_norm": 0.9836651086807251, "learning_rate": 0.00018607852467225652, "loss": 4.572, "step": 647550 }, { "epoch": 1.2084931385424271, "grad_norm": 1.2360426187515259, "learning_rate": 0.00018607638643007977, "loss": 4.6652, "step": 647600 }, { "epoch": 1.2085864441018057, "grad_norm": 1.0882500410079956, "learning_rate": 0.00018607424803599335, "loss": 4.6988, "step": 647650 }, { "epoch": 1.2086797496611843, "grad_norm": 1.3163599967956543, "learning_rate": 0.00018607210949000104, "loss": 4.4021, "step": 647700 }, { "epoch": 1.2087730552205627, "grad_norm": 1.1852872371673584, "learning_rate": 0.00018606997079210658, "loss": 4.5199, "step": 647750 }, { "epoch": 1.2088663607799413, "grad_norm": 0.9970261454582214, "learning_rate": 0.00018606783194231378, "loss": 4.6231, "step": 647800 }, { "epoch": 1.2089596663393196, "grad_norm": 1.1398433446884155, "learning_rate": 0.00018606569294062642, "loss": 4.4279, "step": 647850 }, { "epoch": 1.2090529718986982, "grad_norm": 0.8866944313049316, "learning_rate": 0.00018606355378704824, "loss": 4.661, "step": 647900 }, { "epoch": 1.2091462774580766, "grad_norm": 1.1583337783813477, "learning_rate": 0.00018606141448158304, "loss": 4.6527, "step": 647950 }, { "epoch": 1.2092395830174552, "grad_norm": 1.221325397491455, "learning_rate": 0.00018605927502423456, "loss": 4.6492, "step": 648000 }, { "epoch": 1.2092395830174552, "eval_loss": 4.688560485839844, "eval_runtime": 231.0617, "eval_samples_per_second": 11.287, "eval_steps_per_second": 11.287, "eval_tts_loss": 7.650846162253842, "step": 648000 }, { "epoch": 1.2093328885768337, "grad_norm": 1.2243255376815796, "learning_rate": 0.00018605713541500663, "loss": 4.7158, "step": 648050 }, { "epoch": 1.209426194136212, "grad_norm": 1.1913467645645142, "learning_rate": 0.00018605499565390297, "loss": 4.4162, "step": 648100 }, { "epoch": 1.2095194996955907, "grad_norm": 1.2128405570983887, "learning_rate": 0.00018605285574092742, "loss": 4.4916, "step": 648150 }, { "epoch": 1.209612805254969, "grad_norm": 0.9441714882850647, "learning_rate": 0.00018605071567608369, "loss": 4.4108, "step": 648200 }, { "epoch": 1.2097061108143476, "grad_norm": 1.041772723197937, "learning_rate": 0.00018604857545937562, "loss": 4.56, "step": 648250 }, { "epoch": 1.209799416373726, "grad_norm": 1.32791268825531, "learning_rate": 0.00018604643509080693, "loss": 4.6671, "step": 648300 }, { "epoch": 1.2098927219331046, "grad_norm": 1.2649025917053223, "learning_rate": 0.0001860442945703814, "loss": 4.4932, "step": 648350 }, { "epoch": 1.2099860274924832, "grad_norm": 1.1391412019729614, "learning_rate": 0.0001860421538981029, "loss": 4.705, "step": 648400 }, { "epoch": 1.2100793330518615, "grad_norm": 0.8975533843040466, "learning_rate": 0.00018604001307397506, "loss": 4.5114, "step": 648450 }, { "epoch": 1.2101726386112401, "grad_norm": 0.747383177280426, "learning_rate": 0.00018603787209800179, "loss": 4.6376, "step": 648500 }, { "epoch": 1.2102659441706185, "grad_norm": 1.0694948434829712, "learning_rate": 0.00018603573097018678, "loss": 4.5253, "step": 648550 }, { "epoch": 1.210359249729997, "grad_norm": 1.139189600944519, "learning_rate": 0.00018603358969053387, "loss": 4.5601, "step": 648600 }, { "epoch": 1.2104525552893755, "grad_norm": 0.7038496136665344, "learning_rate": 0.0001860314482590468, "loss": 4.5085, "step": 648650 }, { "epoch": 1.210545860848754, "grad_norm": 1.1939607858657837, "learning_rate": 0.00018602930667572934, "loss": 4.6403, "step": 648700 }, { "epoch": 1.2106391664081326, "grad_norm": 1.1003414392471313, "learning_rate": 0.00018602716494058532, "loss": 4.4792, "step": 648750 }, { "epoch": 1.210732471967511, "grad_norm": 1.4457433223724365, "learning_rate": 0.0001860250230536185, "loss": 4.5545, "step": 648800 }, { "epoch": 1.2108257775268896, "grad_norm": 1.226444125175476, "learning_rate": 0.00018602288101483264, "loss": 4.5012, "step": 648850 }, { "epoch": 1.210919083086268, "grad_norm": 1.2194348573684692, "learning_rate": 0.00018602073882423151, "loss": 4.5283, "step": 648900 }, { "epoch": 1.2110123886456465, "grad_norm": 1.2050790786743164, "learning_rate": 0.00018601859648181894, "loss": 4.5471, "step": 648950 }, { "epoch": 1.211105694205025, "grad_norm": 1.0387877225875854, "learning_rate": 0.00018601645398759868, "loss": 4.5006, "step": 649000 }, { "epoch": 1.2111989997644035, "grad_norm": 1.1765788793563843, "learning_rate": 0.00018601431134157448, "loss": 4.4117, "step": 649050 }, { "epoch": 1.211292305323782, "grad_norm": 1.2536404132843018, "learning_rate": 0.00018601216854375018, "loss": 4.5417, "step": 649100 }, { "epoch": 1.2113856108831604, "grad_norm": 1.0867823362350464, "learning_rate": 0.00018601002559412953, "loss": 4.6397, "step": 649150 }, { "epoch": 1.211478916442539, "grad_norm": 1.2369378805160522, "learning_rate": 0.00018600788249271635, "loss": 4.6646, "step": 649200 }, { "epoch": 1.2115722220019174, "grad_norm": 1.1416916847229004, "learning_rate": 0.00018600573923951438, "loss": 4.4328, "step": 649250 }, { "epoch": 1.211665527561296, "grad_norm": 0.9973885416984558, "learning_rate": 0.00018600359583452738, "loss": 4.556, "step": 649300 }, { "epoch": 1.2117588331206743, "grad_norm": 1.065839409828186, "learning_rate": 0.0001860014522777592, "loss": 4.3081, "step": 649350 }, { "epoch": 1.211852138680053, "grad_norm": 1.0967504978179932, "learning_rate": 0.00018599930856921356, "loss": 4.3679, "step": 649400 }, { "epoch": 1.2119454442394315, "grad_norm": 1.3924851417541504, "learning_rate": 0.0001859971647088943, "loss": 4.6242, "step": 649450 }, { "epoch": 1.2120387497988099, "grad_norm": 0.8641680479049683, "learning_rate": 0.00018599502069680516, "loss": 4.5735, "step": 649500 }, { "epoch": 1.2121320553581885, "grad_norm": 1.4534165859222412, "learning_rate": 0.00018599287653294993, "loss": 4.3937, "step": 649550 }, { "epoch": 1.2122253609175668, "grad_norm": 1.3730312585830688, "learning_rate": 0.00018599073221733242, "loss": 4.5947, "step": 649600 }, { "epoch": 1.2123186664769454, "grad_norm": 1.0847413539886475, "learning_rate": 0.0001859885877499564, "loss": 4.6544, "step": 649650 }, { "epoch": 1.2124119720363238, "grad_norm": 1.2497204542160034, "learning_rate": 0.00018598644313082564, "loss": 4.3585, "step": 649700 }, { "epoch": 1.2125052775957024, "grad_norm": 1.0109866857528687, "learning_rate": 0.00018598429835994394, "loss": 4.402, "step": 649750 }, { "epoch": 1.212598583155081, "grad_norm": 1.2405818700790405, "learning_rate": 0.00018598215343731508, "loss": 4.2943, "step": 649800 }, { "epoch": 1.2126918887144593, "grad_norm": 1.1172492504119873, "learning_rate": 0.00018598000836294286, "loss": 4.6422, "step": 649850 }, { "epoch": 1.212785194273838, "grad_norm": 1.1117241382598877, "learning_rate": 0.00018597786313683106, "loss": 4.6061, "step": 649900 }, { "epoch": 1.2128784998332163, "grad_norm": 0.6190176606178284, "learning_rate": 0.00018597571775898347, "loss": 4.3142, "step": 649950 }, { "epoch": 1.2129718053925949, "grad_norm": 1.1342167854309082, "learning_rate": 0.00018597357222940387, "loss": 4.4993, "step": 650000 }, { "epoch": 1.2130651109519732, "grad_norm": 0.9634416699409485, "learning_rate": 0.000185971426548096, "loss": 4.375, "step": 650050 }, { "epoch": 1.2131584165113518, "grad_norm": 1.0090088844299316, "learning_rate": 0.0001859692807150637, "loss": 4.5171, "step": 650100 }, { "epoch": 1.2132517220707304, "grad_norm": 1.2310750484466553, "learning_rate": 0.00018596713473031075, "loss": 4.5737, "step": 650150 }, { "epoch": 1.2133450276301088, "grad_norm": 1.0660909414291382, "learning_rate": 0.00018596498859384096, "loss": 4.635, "step": 650200 }, { "epoch": 1.2134383331894874, "grad_norm": 1.2995070219039917, "learning_rate": 0.00018596284230565805, "loss": 4.6533, "step": 650250 }, { "epoch": 1.2135316387488657, "grad_norm": 1.0401726961135864, "learning_rate": 0.00018596069586576588, "loss": 4.6687, "step": 650300 }, { "epoch": 1.2136249443082443, "grad_norm": 1.1432294845581055, "learning_rate": 0.00018595854927416818, "loss": 4.4262, "step": 650350 }, { "epoch": 1.2137182498676227, "grad_norm": 1.1811769008636475, "learning_rate": 0.00018595640253086878, "loss": 4.4564, "step": 650400 }, { "epoch": 1.2138115554270013, "grad_norm": 1.1119364500045776, "learning_rate": 0.00018595425563587144, "loss": 4.5125, "step": 650450 }, { "epoch": 1.2139048609863798, "grad_norm": 1.06651771068573, "learning_rate": 0.00018595210858918, "loss": 4.5853, "step": 650500 }, { "epoch": 1.2139981665457582, "grad_norm": 1.2298578023910522, "learning_rate": 0.0001859499613907982, "loss": 4.6223, "step": 650550 }, { "epoch": 1.2140914721051368, "grad_norm": 1.1350921392440796, "learning_rate": 0.00018594781404072982, "loss": 4.4882, "step": 650600 }, { "epoch": 1.2141847776645152, "grad_norm": 0.8742543458938599, "learning_rate": 0.00018594566653897868, "loss": 4.5383, "step": 650650 }, { "epoch": 1.2142780832238937, "grad_norm": 1.1101077795028687, "learning_rate": 0.00018594351888554856, "loss": 4.6519, "step": 650700 }, { "epoch": 1.214371388783272, "grad_norm": 0.906525194644928, "learning_rate": 0.00018594137108044327, "loss": 4.4676, "step": 650750 }, { "epoch": 1.2144646943426507, "grad_norm": 0.9489380717277527, "learning_rate": 0.00018593922312366655, "loss": 4.624, "step": 650800 }, { "epoch": 1.2145579999020293, "grad_norm": 1.3058985471725464, "learning_rate": 0.00018593707501522226, "loss": 4.415, "step": 650850 }, { "epoch": 1.2146513054614076, "grad_norm": 1.2478697299957275, "learning_rate": 0.00018593492675511412, "loss": 4.5698, "step": 650900 }, { "epoch": 1.2147446110207862, "grad_norm": 0.9196845889091492, "learning_rate": 0.00018593277834334597, "loss": 4.4845, "step": 650950 }, { "epoch": 1.2148379165801646, "grad_norm": 1.2844328880310059, "learning_rate": 0.0001859306297799216, "loss": 4.4746, "step": 651000 }, { "epoch": 1.2148379165801646, "eval_loss": 4.684513568878174, "eval_runtime": 230.4422, "eval_samples_per_second": 11.317, "eval_steps_per_second": 11.317, "eval_tts_loss": 7.638542191154021, "step": 651000 }, { "epoch": 1.2149312221395432, "grad_norm": 1.5438870191574097, "learning_rate": 0.00018592848106484478, "loss": 4.5826, "step": 651050 }, { "epoch": 1.2150245276989216, "grad_norm": 1.4784736633300781, "learning_rate": 0.00018592633219811927, "loss": 4.5583, "step": 651100 }, { "epoch": 1.2151178332583001, "grad_norm": 1.206885576248169, "learning_rate": 0.00018592418317974895, "loss": 4.5438, "step": 651150 }, { "epoch": 1.2152111388176787, "grad_norm": 1.232595443725586, "learning_rate": 0.00018592203400973756, "loss": 4.4081, "step": 651200 }, { "epoch": 1.215304444377057, "grad_norm": 0.977764368057251, "learning_rate": 0.00018591988468808887, "loss": 4.54, "step": 651250 }, { "epoch": 1.2153977499364357, "grad_norm": 1.088091254234314, "learning_rate": 0.00018591773521480675, "loss": 4.4373, "step": 651300 }, { "epoch": 1.215491055495814, "grad_norm": 0.9590242505073547, "learning_rate": 0.0001859155855898949, "loss": 4.6211, "step": 651350 }, { "epoch": 1.2155843610551926, "grad_norm": 1.067743182182312, "learning_rate": 0.00018591343581335717, "loss": 4.6496, "step": 651400 }, { "epoch": 1.215677666614571, "grad_norm": 1.1695927381515503, "learning_rate": 0.00018591128588519738, "loss": 4.498, "step": 651450 }, { "epoch": 1.2157709721739496, "grad_norm": 1.3539774417877197, "learning_rate": 0.00018590913580541924, "loss": 4.5903, "step": 651500 }, { "epoch": 1.2158642777333282, "grad_norm": 1.1211614608764648, "learning_rate": 0.0001859069855740266, "loss": 4.3407, "step": 651550 }, { "epoch": 1.2159575832927065, "grad_norm": 0.8773329854011536, "learning_rate": 0.00018590483519102328, "loss": 4.6642, "step": 651600 }, { "epoch": 1.2160508888520851, "grad_norm": 1.0817570686340332, "learning_rate": 0.000185902684656413, "loss": 4.5162, "step": 651650 }, { "epoch": 1.2161441944114635, "grad_norm": 1.3566313982009888, "learning_rate": 0.00018590053397019962, "loss": 4.3746, "step": 651700 }, { "epoch": 1.216237499970842, "grad_norm": 1.5036829710006714, "learning_rate": 0.0001858983831323869, "loss": 4.2581, "step": 651750 }, { "epoch": 1.2163308055302204, "grad_norm": 1.1579976081848145, "learning_rate": 0.00018589623214297865, "loss": 4.647, "step": 651800 }, { "epoch": 1.216424111089599, "grad_norm": 1.216587781906128, "learning_rate": 0.00018589408100197865, "loss": 4.5081, "step": 651850 }, { "epoch": 1.2165174166489776, "grad_norm": 1.356290578842163, "learning_rate": 0.00018589192970939073, "loss": 4.4446, "step": 651900 }, { "epoch": 1.216610722208356, "grad_norm": 1.2447288036346436, "learning_rate": 0.00018588977826521866, "loss": 4.4328, "step": 651950 }, { "epoch": 1.2167040277677346, "grad_norm": 1.1046193838119507, "learning_rate": 0.00018588762666946627, "loss": 4.6452, "step": 652000 }, { "epoch": 1.216797333327113, "grad_norm": 1.1710786819458008, "learning_rate": 0.00018588547492213732, "loss": 4.4226, "step": 652050 }, { "epoch": 1.2168906388864915, "grad_norm": 0.9529326558113098, "learning_rate": 0.00018588332302323557, "loss": 4.2487, "step": 652100 }, { "epoch": 1.2169839444458699, "grad_norm": 0.8778839707374573, "learning_rate": 0.0001858811709727649, "loss": 4.3973, "step": 652150 }, { "epoch": 1.2170772500052485, "grad_norm": 1.1988921165466309, "learning_rate": 0.0001858790187707291, "loss": 4.4347, "step": 652200 }, { "epoch": 1.2171705555646268, "grad_norm": 1.1277692317962646, "learning_rate": 0.00018587686641713194, "loss": 4.6114, "step": 652250 }, { "epoch": 1.2172638611240054, "grad_norm": 0.8896984457969666, "learning_rate": 0.00018587471391197719, "loss": 4.3317, "step": 652300 }, { "epoch": 1.217357166683384, "grad_norm": 0.948337972164154, "learning_rate": 0.00018587256125526867, "loss": 4.5793, "step": 652350 }, { "epoch": 1.2174504722427624, "grad_norm": 1.0679608583450317, "learning_rate": 0.00018587040844701024, "loss": 4.7073, "step": 652400 }, { "epoch": 1.217543777802141, "grad_norm": 1.2565737962722778, "learning_rate": 0.0001858682554872056, "loss": 4.7003, "step": 652450 }, { "epoch": 1.2176370833615193, "grad_norm": 0.967158854007721, "learning_rate": 0.00018586610237585863, "loss": 4.5998, "step": 652500 }, { "epoch": 1.217730388920898, "grad_norm": 1.4818975925445557, "learning_rate": 0.0001858639491129731, "loss": 4.6259, "step": 652550 }, { "epoch": 1.2178236944802763, "grad_norm": 1.2174160480499268, "learning_rate": 0.00018586179569855278, "loss": 4.1898, "step": 652600 }, { "epoch": 1.2179170000396549, "grad_norm": 1.0967553853988647, "learning_rate": 0.0001858596421326015, "loss": 4.7101, "step": 652650 }, { "epoch": 1.2180103055990334, "grad_norm": 1.195443868637085, "learning_rate": 0.0001858574884151231, "loss": 4.4711, "step": 652700 }, { "epoch": 1.2181036111584118, "grad_norm": 0.9037935733795166, "learning_rate": 0.00018585533454612132, "loss": 4.4077, "step": 652750 }, { "epoch": 1.2181969167177904, "grad_norm": 1.2329943180084229, "learning_rate": 0.00018585318052559996, "loss": 4.5081, "step": 652800 }, { "epoch": 1.2182902222771688, "grad_norm": 0.9731119275093079, "learning_rate": 0.00018585102635356284, "loss": 4.4946, "step": 652850 }, { "epoch": 1.2183835278365474, "grad_norm": 1.1737048625946045, "learning_rate": 0.00018584887203001378, "loss": 4.3942, "step": 652900 }, { "epoch": 1.2184768333959257, "grad_norm": 1.0261882543563843, "learning_rate": 0.00018584671755495657, "loss": 4.668, "step": 652950 }, { "epoch": 1.2185701389553043, "grad_norm": 1.2983951568603516, "learning_rate": 0.000185844562928395, "loss": 4.5351, "step": 653000 }, { "epoch": 1.218663444514683, "grad_norm": 1.2656253576278687, "learning_rate": 0.0001858424081503329, "loss": 4.6091, "step": 653050 }, { "epoch": 1.2187567500740613, "grad_norm": 1.0067027807235718, "learning_rate": 0.00018584025322077404, "loss": 4.3494, "step": 653100 }, { "epoch": 1.2188500556334398, "grad_norm": 1.0989404916763306, "learning_rate": 0.00018583809813972225, "loss": 4.6894, "step": 653150 }, { "epoch": 1.2189433611928182, "grad_norm": 1.1562669277191162, "learning_rate": 0.0001858359429071813, "loss": 4.6663, "step": 653200 }, { "epoch": 1.2190366667521968, "grad_norm": 0.9721519947052002, "learning_rate": 0.000185833787523155, "loss": 4.6572, "step": 653250 }, { "epoch": 1.2191299723115752, "grad_norm": 1.258521318435669, "learning_rate": 0.0001858316319876472, "loss": 4.5741, "step": 653300 }, { "epoch": 1.2192232778709537, "grad_norm": 1.2050127983093262, "learning_rate": 0.00018582947630066168, "loss": 4.5549, "step": 653350 }, { "epoch": 1.2193165834303323, "grad_norm": 1.0551210641860962, "learning_rate": 0.00018582732046220222, "loss": 4.4424, "step": 653400 }, { "epoch": 1.2194098889897107, "grad_norm": 1.171412467956543, "learning_rate": 0.00018582516447227264, "loss": 4.5456, "step": 653450 }, { "epoch": 1.2195031945490893, "grad_norm": 0.951436460018158, "learning_rate": 0.00018582300833087673, "loss": 4.4318, "step": 653500 }, { "epoch": 1.2195965001084677, "grad_norm": 1.305679202079773, "learning_rate": 0.00018582085203801833, "loss": 4.2834, "step": 653550 }, { "epoch": 1.2196898056678462, "grad_norm": 1.0619028806686401, "learning_rate": 0.0001858186955937012, "loss": 4.5413, "step": 653600 }, { "epoch": 1.2197831112272246, "grad_norm": 1.3350852727890015, "learning_rate": 0.00018581653899792924, "loss": 4.5901, "step": 653650 }, { "epoch": 1.2198764167866032, "grad_norm": 1.0492174625396729, "learning_rate": 0.00018581438225070612, "loss": 4.6343, "step": 653700 }, { "epoch": 1.2199697223459816, "grad_norm": 1.4955275058746338, "learning_rate": 0.00018581222535203574, "loss": 4.487, "step": 653750 }, { "epoch": 1.2200630279053601, "grad_norm": 0.7797152400016785, "learning_rate": 0.00018581006830192188, "loss": 4.3904, "step": 653800 }, { "epoch": 1.2201563334647387, "grad_norm": 1.2893060445785522, "learning_rate": 0.00018580791110036837, "loss": 4.5048, "step": 653850 }, { "epoch": 1.220249639024117, "grad_norm": 0.8679825663566589, "learning_rate": 0.00018580575374737898, "loss": 4.5631, "step": 653900 }, { "epoch": 1.2203429445834957, "grad_norm": 0.8895425200462341, "learning_rate": 0.00018580359624295752, "loss": 4.3866, "step": 653950 }, { "epoch": 1.220436250142874, "grad_norm": 1.2855700254440308, "learning_rate": 0.0001858014385871078, "loss": 4.4431, "step": 654000 }, { "epoch": 1.220436250142874, "eval_loss": 4.675308704376221, "eval_runtime": 228.45, "eval_samples_per_second": 11.416, "eval_steps_per_second": 11.416, "eval_tts_loss": 7.636172809412394, "step": 654000 }, { "epoch": 1.2205295557022526, "grad_norm": 1.1376595497131348, "learning_rate": 0.00018579928077983368, "loss": 4.4923, "step": 654050 }, { "epoch": 1.220622861261631, "grad_norm": 0.6174269318580627, "learning_rate": 0.0001857971228211389, "loss": 4.4561, "step": 654100 }, { "epoch": 1.2207161668210096, "grad_norm": 1.25332772731781, "learning_rate": 0.00018579496471102732, "loss": 4.6978, "step": 654150 }, { "epoch": 1.2208094723803882, "grad_norm": 1.17857825756073, "learning_rate": 0.0001857928064495027, "loss": 4.339, "step": 654200 }, { "epoch": 1.2209027779397665, "grad_norm": 0.9567538499832153, "learning_rate": 0.00018579064803656885, "loss": 4.2598, "step": 654250 }, { "epoch": 1.2209960834991451, "grad_norm": 1.2542535066604614, "learning_rate": 0.0001857884894722296, "loss": 4.609, "step": 654300 }, { "epoch": 1.2210893890585235, "grad_norm": 1.349761962890625, "learning_rate": 0.0001857863307564888, "loss": 4.6573, "step": 654350 }, { "epoch": 1.221182694617902, "grad_norm": 1.1513382196426392, "learning_rate": 0.00018578417188935023, "loss": 4.5105, "step": 654400 }, { "epoch": 1.2212760001772804, "grad_norm": 1.267492413520813, "learning_rate": 0.00018578201287081764, "loss": 4.6446, "step": 654450 }, { "epoch": 1.221369305736659, "grad_norm": 0.8505188226699829, "learning_rate": 0.00018577985370089493, "loss": 4.6584, "step": 654500 }, { "epoch": 1.2214626112960376, "grad_norm": 1.3523976802825928, "learning_rate": 0.00018577769437958587, "loss": 4.565, "step": 654550 }, { "epoch": 1.221555916855416, "grad_norm": 0.9807010293006897, "learning_rate": 0.00018577553490689428, "loss": 4.6904, "step": 654600 }, { "epoch": 1.2216492224147946, "grad_norm": 0.9061399102210999, "learning_rate": 0.0001857733752828239, "loss": 4.5075, "step": 654650 }, { "epoch": 1.221742527974173, "grad_norm": 1.1481069326400757, "learning_rate": 0.00018577121550737867, "loss": 4.6225, "step": 654700 }, { "epoch": 1.2218358335335515, "grad_norm": 1.027021050453186, "learning_rate": 0.00018576905558056233, "loss": 4.6249, "step": 654750 }, { "epoch": 1.2219291390929299, "grad_norm": 1.0878961086273193, "learning_rate": 0.00018576689550237868, "loss": 4.6642, "step": 654800 }, { "epoch": 1.2220224446523085, "grad_norm": 1.01434326171875, "learning_rate": 0.00018576473527283157, "loss": 4.5421, "step": 654850 }, { "epoch": 1.222115750211687, "grad_norm": 1.2235907316207886, "learning_rate": 0.00018576257489192477, "loss": 4.5277, "step": 654900 }, { "epoch": 1.2222090557710654, "grad_norm": 1.1861783266067505, "learning_rate": 0.00018576041435966213, "loss": 4.5558, "step": 654950 }, { "epoch": 1.222302361330444, "grad_norm": 1.050596833229065, "learning_rate": 0.00018575825367604743, "loss": 4.5125, "step": 655000 }, { "epoch": 1.2223956668898224, "grad_norm": 0.9791225790977478, "learning_rate": 0.00018575609284108454, "loss": 4.5215, "step": 655050 }, { "epoch": 1.222488972449201, "grad_norm": 1.2039810419082642, "learning_rate": 0.0001857539318547772, "loss": 4.4815, "step": 655100 }, { "epoch": 1.2225822780085793, "grad_norm": 1.3799984455108643, "learning_rate": 0.00018575177071712925, "loss": 4.5096, "step": 655150 }, { "epoch": 1.222675583567958, "grad_norm": 1.2410647869110107, "learning_rate": 0.00018574960942814453, "loss": 4.6625, "step": 655200 }, { "epoch": 1.2227688891273365, "grad_norm": 1.1157286167144775, "learning_rate": 0.00018574744798782687, "loss": 4.651, "step": 655250 }, { "epoch": 1.2228621946867149, "grad_norm": 1.2746057510375977, "learning_rate": 0.00018574528639618004, "loss": 4.646, "step": 655300 }, { "epoch": 1.2229555002460935, "grad_norm": 0.9694011211395264, "learning_rate": 0.00018574312465320783, "loss": 4.4529, "step": 655350 }, { "epoch": 1.2230488058054718, "grad_norm": 0.9683193564414978, "learning_rate": 0.00018574096275891412, "loss": 4.4148, "step": 655400 }, { "epoch": 1.2231421113648504, "grad_norm": 1.179792881011963, "learning_rate": 0.00018573880071330267, "loss": 4.4745, "step": 655450 }, { "epoch": 1.2232354169242288, "grad_norm": 1.1245102882385254, "learning_rate": 0.00018573663851637734, "loss": 4.5265, "step": 655500 }, { "epoch": 1.2233287224836074, "grad_norm": 1.4471015930175781, "learning_rate": 0.00018573447616814195, "loss": 4.4113, "step": 655550 }, { "epoch": 1.223422028042986, "grad_norm": 1.4259343147277832, "learning_rate": 0.00018573231366860027, "loss": 4.6226, "step": 655600 }, { "epoch": 1.2235153336023643, "grad_norm": 0.9361169934272766, "learning_rate": 0.00018573015101775613, "loss": 4.3275, "step": 655650 }, { "epoch": 1.223608639161743, "grad_norm": 1.1099869012832642, "learning_rate": 0.0001857279882156134, "loss": 4.7338, "step": 655700 }, { "epoch": 1.2237019447211213, "grad_norm": 0.9362122416496277, "learning_rate": 0.00018572582526217582, "loss": 4.5656, "step": 655750 }, { "epoch": 1.2237952502804998, "grad_norm": 1.4260603189468384, "learning_rate": 0.00018572366215744725, "loss": 4.3907, "step": 655800 }, { "epoch": 1.2238885558398782, "grad_norm": 1.204066514968872, "learning_rate": 0.0001857214989014315, "loss": 4.5542, "step": 655850 }, { "epoch": 1.2239818613992568, "grad_norm": 1.0756545066833496, "learning_rate": 0.0001857193354941324, "loss": 4.4462, "step": 655900 }, { "epoch": 1.2240751669586354, "grad_norm": 1.3867384195327759, "learning_rate": 0.00018571717193555375, "loss": 4.5708, "step": 655950 }, { "epoch": 1.2241684725180138, "grad_norm": 1.0212152004241943, "learning_rate": 0.00018571500822569937, "loss": 4.6663, "step": 656000 }, { "epoch": 1.2242617780773923, "grad_norm": 1.3445554971694946, "learning_rate": 0.0001857128443645731, "loss": 4.6488, "step": 656050 }, { "epoch": 1.2243550836367707, "grad_norm": 1.1284013986587524, "learning_rate": 0.00018571068035217868, "loss": 4.5629, "step": 656100 }, { "epoch": 1.2244483891961493, "grad_norm": 2.0593338012695312, "learning_rate": 0.00018570851618852005, "loss": 4.5134, "step": 656150 }, { "epoch": 1.2245416947555277, "grad_norm": 1.370063066482544, "learning_rate": 0.00018570635187360095, "loss": 4.5228, "step": 656200 }, { "epoch": 1.2246350003149062, "grad_norm": 1.111016869544983, "learning_rate": 0.00018570418740742524, "loss": 4.5185, "step": 656250 }, { "epoch": 1.2247283058742848, "grad_norm": 0.9248209595680237, "learning_rate": 0.0001857020227899967, "loss": 4.5068, "step": 656300 }, { "epoch": 1.2248216114336632, "grad_norm": 1.1682900190353394, "learning_rate": 0.00018569985802131917, "loss": 4.548, "step": 656350 }, { "epoch": 1.2249149169930418, "grad_norm": 1.1871378421783447, "learning_rate": 0.00018569769310139647, "loss": 4.4139, "step": 656400 }, { "epoch": 1.2250082225524201, "grad_norm": 1.269714117050171, "learning_rate": 0.00018569552803023243, "loss": 4.5577, "step": 656450 }, { "epoch": 1.2251015281117987, "grad_norm": 1.2309085130691528, "learning_rate": 0.00018569336280783086, "loss": 4.6694, "step": 656500 }, { "epoch": 1.225194833671177, "grad_norm": 1.1276769638061523, "learning_rate": 0.00018569119743419557, "loss": 4.4728, "step": 656550 }, { "epoch": 1.2252881392305557, "grad_norm": 1.021859884262085, "learning_rate": 0.00018568903190933038, "loss": 4.3616, "step": 656600 }, { "epoch": 1.2253814447899343, "grad_norm": 1.4429341554641724, "learning_rate": 0.00018568686623323917, "loss": 4.4557, "step": 656650 }, { "epoch": 1.2254747503493126, "grad_norm": 1.1370371580123901, "learning_rate": 0.0001856847004059257, "loss": 4.6082, "step": 656700 }, { "epoch": 1.2255680559086912, "grad_norm": 0.8863230347633362, "learning_rate": 0.0001856825344273938, "loss": 4.5538, "step": 656750 }, { "epoch": 1.2256613614680696, "grad_norm": 1.1482950448989868, "learning_rate": 0.0001856803682976473, "loss": 4.4583, "step": 656800 }, { "epoch": 1.2257546670274482, "grad_norm": 0.9905821084976196, "learning_rate": 0.00018567820201669007, "loss": 4.5111, "step": 656850 }, { "epoch": 1.2258479725868265, "grad_norm": 1.1899008750915527, "learning_rate": 0.00018567603558452582, "loss": 4.5269, "step": 656900 }, { "epoch": 1.2259412781462051, "grad_norm": 1.1617501974105835, "learning_rate": 0.0001856738690011585, "loss": 4.5918, "step": 656950 }, { "epoch": 1.2260345837055837, "grad_norm": 1.4769470691680908, "learning_rate": 0.00018567170226659182, "loss": 4.6391, "step": 657000 }, { "epoch": 1.2260345837055837, "eval_loss": 4.673868656158447, "eval_runtime": 230.6831, "eval_samples_per_second": 11.306, "eval_steps_per_second": 11.306, "eval_tts_loss": 7.607360904090724, "step": 657000 }, { "epoch": 1.226127889264962, "grad_norm": 0.9327487945556641, "learning_rate": 0.00018566953538082972, "loss": 4.5075, "step": 657050 }, { "epoch": 1.2262211948243407, "grad_norm": 1.1986678838729858, "learning_rate": 0.00018566736834387592, "loss": 4.652, "step": 657100 }, { "epoch": 1.226314500383719, "grad_norm": 1.1232380867004395, "learning_rate": 0.00018566520115573433, "loss": 4.3729, "step": 657150 }, { "epoch": 1.2264078059430976, "grad_norm": 1.1070131063461304, "learning_rate": 0.0001856630338164087, "loss": 4.4638, "step": 657200 }, { "epoch": 1.226501111502476, "grad_norm": 0.9490290880203247, "learning_rate": 0.0001856608663259029, "loss": 4.4233, "step": 657250 }, { "epoch": 1.2265944170618546, "grad_norm": 1.1436617374420166, "learning_rate": 0.00018565869868422073, "loss": 4.616, "step": 657300 }, { "epoch": 1.2266877226212332, "grad_norm": 0.9338767528533936, "learning_rate": 0.00018565653089136602, "loss": 4.5685, "step": 657350 }, { "epoch": 1.2267810281806115, "grad_norm": 1.7048883438110352, "learning_rate": 0.00018565436294734264, "loss": 4.6295, "step": 657400 }, { "epoch": 1.22687433373999, "grad_norm": 1.2740790843963623, "learning_rate": 0.00018565219485215437, "loss": 4.5277, "step": 657450 }, { "epoch": 1.2269676392993685, "grad_norm": 0.9780893325805664, "learning_rate": 0.00018565002660580503, "loss": 4.641, "step": 657500 }, { "epoch": 1.227060944858747, "grad_norm": 1.2198879718780518, "learning_rate": 0.00018564785820829847, "loss": 4.4706, "step": 657550 }, { "epoch": 1.2271542504181254, "grad_norm": 1.2191646099090576, "learning_rate": 0.00018564568965963852, "loss": 4.5765, "step": 657600 }, { "epoch": 1.227247555977504, "grad_norm": 1.35011887550354, "learning_rate": 0.00018564352095982903, "loss": 4.4634, "step": 657650 }, { "epoch": 1.2273408615368826, "grad_norm": 1.2260960340499878, "learning_rate": 0.00018564135210887375, "loss": 4.6728, "step": 657700 }, { "epoch": 1.227434167096261, "grad_norm": 1.4104201793670654, "learning_rate": 0.00018563918310677654, "loss": 4.4217, "step": 657750 }, { "epoch": 1.2275274726556396, "grad_norm": 1.0541867017745972, "learning_rate": 0.00018563701395354128, "loss": 4.5793, "step": 657800 }, { "epoch": 1.227620778215018, "grad_norm": 1.4122992753982544, "learning_rate": 0.0001856348446491717, "loss": 4.3917, "step": 657850 }, { "epoch": 1.2277140837743965, "grad_norm": 1.0537043809890747, "learning_rate": 0.00018563267519367175, "loss": 4.6952, "step": 657900 }, { "epoch": 1.2278073893337749, "grad_norm": 1.2942408323287964, "learning_rate": 0.00018563050558704518, "loss": 4.5092, "step": 657950 }, { "epoch": 1.2279006948931535, "grad_norm": 1.245284914970398, "learning_rate": 0.00018562833582929581, "loss": 4.5383, "step": 658000 }, { "epoch": 1.227994000452532, "grad_norm": 1.30874764919281, "learning_rate": 0.00018562616592042753, "loss": 4.5857, "step": 658050 }, { "epoch": 1.2280873060119104, "grad_norm": 1.2272487878799438, "learning_rate": 0.00018562399586044412, "loss": 4.6049, "step": 658100 }, { "epoch": 1.228180611571289, "grad_norm": 1.2049269676208496, "learning_rate": 0.00018562182564934942, "loss": 4.6773, "step": 658150 }, { "epoch": 1.2282739171306674, "grad_norm": 1.2600430250167847, "learning_rate": 0.00018561965528714723, "loss": 4.5084, "step": 658200 }, { "epoch": 1.228367222690046, "grad_norm": 1.0235731601715088, "learning_rate": 0.00018561748477384147, "loss": 4.5598, "step": 658250 }, { "epoch": 1.2284605282494243, "grad_norm": 1.069442629814148, "learning_rate": 0.0001856153141094359, "loss": 4.5118, "step": 658300 }, { "epoch": 1.228553833808803, "grad_norm": 1.3327456712722778, "learning_rate": 0.00018561314329393436, "loss": 4.4678, "step": 658350 }, { "epoch": 1.2286471393681815, "grad_norm": 1.475574016571045, "learning_rate": 0.00018561097232734067, "loss": 4.6259, "step": 658400 }, { "epoch": 1.2287404449275598, "grad_norm": 1.0195128917694092, "learning_rate": 0.00018560880120965868, "loss": 4.4051, "step": 658450 }, { "epoch": 1.2288337504869384, "grad_norm": 1.2307562828063965, "learning_rate": 0.00018560662994089225, "loss": 4.6016, "step": 658500 }, { "epoch": 1.2289270560463168, "grad_norm": 1.2197580337524414, "learning_rate": 0.00018560445852104515, "loss": 4.298, "step": 658550 }, { "epoch": 1.2290203616056954, "grad_norm": 1.4353586435317993, "learning_rate": 0.00018560228695012123, "loss": 4.6024, "step": 658600 }, { "epoch": 1.2291136671650738, "grad_norm": 1.2093204259872437, "learning_rate": 0.00018560011522812436, "loss": 4.5311, "step": 658650 }, { "epoch": 1.2292069727244523, "grad_norm": 1.4406057596206665, "learning_rate": 0.00018559794335505837, "loss": 4.5362, "step": 658700 }, { "epoch": 1.229300278283831, "grad_norm": 1.292359471321106, "learning_rate": 0.00018559577133092704, "loss": 4.3435, "step": 658750 }, { "epoch": 1.2293935838432093, "grad_norm": 1.0128477811813354, "learning_rate": 0.00018559359915573423, "loss": 4.6366, "step": 658800 }, { "epoch": 1.2294868894025879, "grad_norm": 1.1675090789794922, "learning_rate": 0.00018559142682948382, "loss": 4.5122, "step": 658850 }, { "epoch": 1.2295801949619662, "grad_norm": 1.2270092964172363, "learning_rate": 0.00018558925435217956, "loss": 4.5052, "step": 658900 }, { "epoch": 1.2296735005213448, "grad_norm": 1.2603273391723633, "learning_rate": 0.00018558708172382532, "loss": 4.4644, "step": 658950 }, { "epoch": 1.2297668060807232, "grad_norm": 0.8732103705406189, "learning_rate": 0.00018558490894442494, "loss": 4.6125, "step": 659000 }, { "epoch": 1.2298601116401018, "grad_norm": 1.0404551029205322, "learning_rate": 0.00018558273601398232, "loss": 4.4613, "step": 659050 }, { "epoch": 1.2299534171994804, "grad_norm": 1.311153769493103, "learning_rate": 0.00018558056293250117, "loss": 4.474, "step": 659100 }, { "epoch": 1.2300467227588587, "grad_norm": 1.0443525314331055, "learning_rate": 0.0001855783896999854, "loss": 4.2828, "step": 659150 }, { "epoch": 1.2301400283182373, "grad_norm": 1.2783119678497314, "learning_rate": 0.00018557621631643881, "loss": 4.3991, "step": 659200 }, { "epoch": 1.2302333338776157, "grad_norm": 1.2335021495819092, "learning_rate": 0.0001855740427818653, "loss": 4.4388, "step": 659250 }, { "epoch": 1.2303266394369943, "grad_norm": 0.9851624965667725, "learning_rate": 0.00018557186909626863, "loss": 4.4678, "step": 659300 }, { "epoch": 1.2304199449963726, "grad_norm": 0.9839292764663696, "learning_rate": 0.00018556969525965268, "loss": 4.5258, "step": 659350 }, { "epoch": 1.2305132505557512, "grad_norm": 1.2246633768081665, "learning_rate": 0.00018556752127202124, "loss": 4.5571, "step": 659400 }, { "epoch": 1.2306065561151298, "grad_norm": 1.185158133506775, "learning_rate": 0.0001855653471333782, "loss": 4.4545, "step": 659450 }, { "epoch": 1.2306998616745082, "grad_norm": 1.0281490087509155, "learning_rate": 0.00018556317284372742, "loss": 4.4965, "step": 659500 }, { "epoch": 1.2307931672338868, "grad_norm": 0.7727369070053101, "learning_rate": 0.00018556099840307265, "loss": 4.6875, "step": 659550 }, { "epoch": 1.2308864727932651, "grad_norm": 0.9159561991691589, "learning_rate": 0.0001855588238114178, "loss": 4.7235, "step": 659600 }, { "epoch": 1.2309797783526437, "grad_norm": 1.384789228439331, "learning_rate": 0.00018555664906876667, "loss": 4.4981, "step": 659650 }, { "epoch": 1.231073083912022, "grad_norm": 1.3203697204589844, "learning_rate": 0.0001855544741751231, "loss": 4.5147, "step": 659700 }, { "epoch": 1.2311663894714007, "grad_norm": 1.1435706615447998, "learning_rate": 0.00018555229913049093, "loss": 4.4938, "step": 659750 }, { "epoch": 1.2312596950307793, "grad_norm": 1.0226768255233765, "learning_rate": 0.00018555012393487404, "loss": 4.5408, "step": 659800 }, { "epoch": 1.2313530005901576, "grad_norm": 1.2715861797332764, "learning_rate": 0.0001855479485882762, "loss": 4.54, "step": 659850 }, { "epoch": 1.2314463061495362, "grad_norm": 1.1990790367126465, "learning_rate": 0.00018554577309070128, "loss": 4.4021, "step": 659900 }, { "epoch": 1.2315396117089146, "grad_norm": 0.8381608128547668, "learning_rate": 0.00018554359744215316, "loss": 4.4031, "step": 659950 }, { "epoch": 1.2316329172682932, "grad_norm": 0.8825395703315735, "learning_rate": 0.0001855414216426356, "loss": 4.4788, "step": 660000 }, { "epoch": 1.2316329172682932, "eval_loss": 4.681206226348877, "eval_runtime": 228.3359, "eval_samples_per_second": 11.422, "eval_steps_per_second": 11.422, "eval_tts_loss": 7.6083566470963, "step": 660000 }, { "epoch": 1.2317262228276715, "grad_norm": 1.0836704969406128, "learning_rate": 0.0001855392456921525, "loss": 4.6507, "step": 660050 }, { "epoch": 1.2318195283870501, "grad_norm": 1.1111444234848022, "learning_rate": 0.00018553706959070768, "loss": 4.4236, "step": 660100 }, { "epoch": 1.2319128339464287, "grad_norm": 0.9190298914909363, "learning_rate": 0.00018553489333830497, "loss": 4.5405, "step": 660150 }, { "epoch": 1.232006139505807, "grad_norm": 1.3368139266967773, "learning_rate": 0.00018553271693494826, "loss": 4.5246, "step": 660200 }, { "epoch": 1.2320994450651856, "grad_norm": 1.3500257730484009, "learning_rate": 0.00018553054038064132, "loss": 4.375, "step": 660250 }, { "epoch": 1.232192750624564, "grad_norm": 1.3166524171829224, "learning_rate": 0.00018552836367538805, "loss": 4.5543, "step": 660300 }, { "epoch": 1.2322860561839426, "grad_norm": 1.031599998474121, "learning_rate": 0.00018552618681919226, "loss": 4.5154, "step": 660350 }, { "epoch": 1.232379361743321, "grad_norm": 1.0453590154647827, "learning_rate": 0.0001855240098120578, "loss": 4.5732, "step": 660400 }, { "epoch": 1.2324726673026996, "grad_norm": 1.3358734846115112, "learning_rate": 0.0001855218326539885, "loss": 4.3343, "step": 660450 }, { "epoch": 1.2325659728620781, "grad_norm": 1.061242699623108, "learning_rate": 0.00018551965534498822, "loss": 4.57, "step": 660500 }, { "epoch": 1.2326592784214565, "grad_norm": 1.2574350833892822, "learning_rate": 0.0001855174778850608, "loss": 4.5846, "step": 660550 }, { "epoch": 1.232752583980835, "grad_norm": 1.0444380044937134, "learning_rate": 0.0001855153002742101, "loss": 4.6142, "step": 660600 }, { "epoch": 1.2328458895402135, "grad_norm": 1.2533833980560303, "learning_rate": 0.0001855131225124399, "loss": 4.5963, "step": 660650 }, { "epoch": 1.232939195099592, "grad_norm": 1.1825668811798096, "learning_rate": 0.0001855109445997541, "loss": 4.6917, "step": 660700 }, { "epoch": 1.2330325006589704, "grad_norm": 1.0909925699234009, "learning_rate": 0.00018550876653615654, "loss": 4.4599, "step": 660750 }, { "epoch": 1.233125806218349, "grad_norm": 1.0530390739440918, "learning_rate": 0.00018550658832165105, "loss": 4.5377, "step": 660800 }, { "epoch": 1.2332191117777276, "grad_norm": 1.0855281352996826, "learning_rate": 0.00018550440995624147, "loss": 4.4866, "step": 660850 }, { "epoch": 1.233312417337106, "grad_norm": 1.2069910764694214, "learning_rate": 0.00018550223143993163, "loss": 4.643, "step": 660900 }, { "epoch": 1.2334057228964845, "grad_norm": 1.1195517778396606, "learning_rate": 0.00018550005277272544, "loss": 4.713, "step": 660950 }, { "epoch": 1.233499028455863, "grad_norm": 1.1591451168060303, "learning_rate": 0.00018549787395462667, "loss": 4.6303, "step": 661000 }, { "epoch": 1.2335923340152415, "grad_norm": 1.1353427171707153, "learning_rate": 0.0001854956949856392, "loss": 4.4559, "step": 661050 }, { "epoch": 1.2336856395746199, "grad_norm": 0.9605555534362793, "learning_rate": 0.0001854935158657669, "loss": 4.6434, "step": 661100 }, { "epoch": 1.2337789451339984, "grad_norm": 1.2010184526443481, "learning_rate": 0.00018549133659501356, "loss": 4.4734, "step": 661150 }, { "epoch": 1.233872250693377, "grad_norm": 1.3283385038375854, "learning_rate": 0.00018548915717338306, "loss": 4.3589, "step": 661200 }, { "epoch": 1.2339655562527554, "grad_norm": 0.8993031978607178, "learning_rate": 0.00018548697760087924, "loss": 4.6712, "step": 661250 }, { "epoch": 1.234058861812134, "grad_norm": 0.9349741339683533, "learning_rate": 0.00018548479787750596, "loss": 4.5182, "step": 661300 }, { "epoch": 1.2341521673715123, "grad_norm": 0.8097958564758301, "learning_rate": 0.00018548261800326702, "loss": 4.5736, "step": 661350 }, { "epoch": 1.234245472930891, "grad_norm": 0.9240305423736572, "learning_rate": 0.00018548043797816631, "loss": 4.4926, "step": 661400 }, { "epoch": 1.2343387784902693, "grad_norm": 1.2127889394760132, "learning_rate": 0.00018547825780220768, "loss": 4.6095, "step": 661450 }, { "epoch": 1.2344320840496479, "grad_norm": 1.709245204925537, "learning_rate": 0.00018547607747539498, "loss": 4.6997, "step": 661500 }, { "epoch": 1.2345253896090265, "grad_norm": 1.0050301551818848, "learning_rate": 0.00018547389699773202, "loss": 4.5731, "step": 661550 }, { "epoch": 1.2346186951684048, "grad_norm": 1.0924513339996338, "learning_rate": 0.0001854717163692227, "loss": 4.3601, "step": 661600 }, { "epoch": 1.2347120007277834, "grad_norm": 1.3142215013504028, "learning_rate": 0.0001854695355898708, "loss": 4.4539, "step": 661650 }, { "epoch": 1.2348053062871618, "grad_norm": 0.9761170744895935, "learning_rate": 0.0001854673546596802, "loss": 4.4799, "step": 661700 }, { "epoch": 1.2348986118465404, "grad_norm": 1.236376166343689, "learning_rate": 0.00018546517357865481, "loss": 4.5129, "step": 661750 }, { "epoch": 1.2349919174059187, "grad_norm": 1.5054585933685303, "learning_rate": 0.0001854629923467984, "loss": 4.4417, "step": 661800 }, { "epoch": 1.2350852229652973, "grad_norm": 1.3685975074768066, "learning_rate": 0.00018546081096411485, "loss": 4.6723, "step": 661850 }, { "epoch": 1.235178528524676, "grad_norm": 1.1883097887039185, "learning_rate": 0.00018545862943060798, "loss": 4.6264, "step": 661900 }, { "epoch": 1.2352718340840543, "grad_norm": 1.0290864706039429, "learning_rate": 0.0001854564477462817, "loss": 4.6057, "step": 661950 }, { "epoch": 1.2353651396434329, "grad_norm": 1.29996919631958, "learning_rate": 0.0001854542659111398, "loss": 4.633, "step": 662000 }, { "epoch": 1.2354584452028112, "grad_norm": 0.9306638836860657, "learning_rate": 0.00018545208392518618, "loss": 4.3363, "step": 662050 }, { "epoch": 1.2355517507621898, "grad_norm": 1.1411551237106323, "learning_rate": 0.00018544990178842466, "loss": 4.539, "step": 662100 }, { "epoch": 1.2356450563215682, "grad_norm": 0.9166069030761719, "learning_rate": 0.0001854477195008591, "loss": 4.3672, "step": 662150 }, { "epoch": 1.2357383618809468, "grad_norm": 1.2213295698165894, "learning_rate": 0.00018544553706249337, "loss": 4.6478, "step": 662200 }, { "epoch": 1.2358316674403254, "grad_norm": 0.8786745667457581, "learning_rate": 0.00018544335447333126, "loss": 4.5591, "step": 662250 }, { "epoch": 1.2359249729997037, "grad_norm": 1.3257813453674316, "learning_rate": 0.0001854411717333767, "loss": 4.489, "step": 662300 }, { "epoch": 1.2360182785590823, "grad_norm": 1.2323122024536133, "learning_rate": 0.00018543898884263346, "loss": 4.672, "step": 662350 }, { "epoch": 1.2361115841184607, "grad_norm": 0.8901577591896057, "learning_rate": 0.00018543680580110548, "loss": 4.6395, "step": 662400 }, { "epoch": 1.2362048896778393, "grad_norm": 1.1469279527664185, "learning_rate": 0.00018543462260879655, "loss": 4.4586, "step": 662450 }, { "epoch": 1.2362981952372176, "grad_norm": 1.034690499305725, "learning_rate": 0.00018543243926571058, "loss": 4.626, "step": 662500 }, { "epoch": 1.2363915007965962, "grad_norm": 0.8178995847702026, "learning_rate": 0.00018543025577185134, "loss": 4.5387, "step": 662550 }, { "epoch": 1.2364848063559748, "grad_norm": 0.9269349575042725, "learning_rate": 0.00018542807212722275, "loss": 4.5762, "step": 662600 }, { "epoch": 1.2365781119153532, "grad_norm": 1.38555908203125, "learning_rate": 0.00018542588833182864, "loss": 4.4334, "step": 662650 }, { "epoch": 1.2366714174747317, "grad_norm": 1.044996738433838, "learning_rate": 0.00018542370438567288, "loss": 4.5315, "step": 662700 }, { "epoch": 1.2367647230341101, "grad_norm": 1.2824044227600098, "learning_rate": 0.0001854215202887593, "loss": 4.3515, "step": 662750 }, { "epoch": 1.2368580285934887, "grad_norm": 0.9962798357009888, "learning_rate": 0.00018541933604109177, "loss": 4.5556, "step": 662800 }, { "epoch": 1.236951334152867, "grad_norm": 1.0329440832138062, "learning_rate": 0.00018541715164267414, "loss": 4.394, "step": 662850 }, { "epoch": 1.2370446397122457, "grad_norm": 1.4065485000610352, "learning_rate": 0.00018541496709351026, "loss": 4.4734, "step": 662900 }, { "epoch": 1.2371379452716242, "grad_norm": 1.4453318119049072, "learning_rate": 0.000185412782393604, "loss": 4.541, "step": 662950 }, { "epoch": 1.2372312508310026, "grad_norm": 0.9167221188545227, "learning_rate": 0.0001854105975429592, "loss": 4.3507, "step": 663000 }, { "epoch": 1.2372312508310026, "eval_loss": 4.685970306396484, "eval_runtime": 230.4176, "eval_samples_per_second": 11.319, "eval_steps_per_second": 11.319, "eval_tts_loss": 7.6499996333514435, "step": 663000 }, { "epoch": 1.2373245563903812, "grad_norm": 1.1270875930786133, "learning_rate": 0.00018540841254157978, "loss": 4.4613, "step": 663050 }, { "epoch": 1.2374178619497596, "grad_norm": 1.2081797122955322, "learning_rate": 0.0001854062273894695, "loss": 4.5367, "step": 663100 }, { "epoch": 1.2375111675091381, "grad_norm": 1.2133448123931885, "learning_rate": 0.00018540404208663224, "loss": 4.3578, "step": 663150 }, { "epoch": 1.2376044730685165, "grad_norm": 1.1448124647140503, "learning_rate": 0.00018540185663307185, "loss": 4.4426, "step": 663200 }, { "epoch": 1.237697778627895, "grad_norm": 1.2775667905807495, "learning_rate": 0.00018539967102879226, "loss": 4.7017, "step": 663250 }, { "epoch": 1.2377910841872737, "grad_norm": 1.0883375406265259, "learning_rate": 0.00018539748527379726, "loss": 4.5337, "step": 663300 }, { "epoch": 1.237884389746652, "grad_norm": 1.4175406694412231, "learning_rate": 0.00018539529936809074, "loss": 4.6681, "step": 663350 }, { "epoch": 1.2379776953060306, "grad_norm": 1.3226066827774048, "learning_rate": 0.0001853931133116765, "loss": 4.6225, "step": 663400 }, { "epoch": 1.238071000865409, "grad_norm": 0.7805131077766418, "learning_rate": 0.00018539092710455847, "loss": 4.4008, "step": 663450 }, { "epoch": 1.2381643064247876, "grad_norm": 1.051641583442688, "learning_rate": 0.00018538874074674046, "loss": 4.592, "step": 663500 }, { "epoch": 1.238257611984166, "grad_norm": 1.2522155046463013, "learning_rate": 0.0001853865542382264, "loss": 4.5176, "step": 663550 }, { "epoch": 1.2383509175435445, "grad_norm": 1.2126011848449707, "learning_rate": 0.00018538436757902004, "loss": 4.4055, "step": 663600 }, { "epoch": 1.2384442231029231, "grad_norm": 0.866830050945282, "learning_rate": 0.00018538218076912533, "loss": 4.6982, "step": 663650 }, { "epoch": 1.2385375286623015, "grad_norm": 0.8090375065803528, "learning_rate": 0.00018537999380854604, "loss": 4.2251, "step": 663700 }, { "epoch": 1.23863083422168, "grad_norm": 1.2263990640640259, "learning_rate": 0.00018537780669728614, "loss": 4.5191, "step": 663750 }, { "epoch": 1.2387241397810584, "grad_norm": 1.06594979763031, "learning_rate": 0.00018537561943534944, "loss": 4.2853, "step": 663800 }, { "epoch": 1.238817445340437, "grad_norm": 1.0782897472381592, "learning_rate": 0.00018537343202273978, "loss": 4.3953, "step": 663850 }, { "epoch": 1.2389107508998154, "grad_norm": 1.1910455226898193, "learning_rate": 0.00018537124445946099, "loss": 4.714, "step": 663900 }, { "epoch": 1.239004056459194, "grad_norm": 1.0369423627853394, "learning_rate": 0.00018536905674551704, "loss": 4.4912, "step": 663950 }, { "epoch": 1.2390973620185726, "grad_norm": 1.0372705459594727, "learning_rate": 0.00018536686888091168, "loss": 4.491, "step": 664000 }, { "epoch": 1.239190667577951, "grad_norm": 0.944460928440094, "learning_rate": 0.00018536468086564884, "loss": 4.4999, "step": 664050 }, { "epoch": 1.2392839731373295, "grad_norm": 1.1760995388031006, "learning_rate": 0.00018536249269973234, "loss": 4.4715, "step": 664100 }, { "epoch": 1.2393772786967079, "grad_norm": 1.0524054765701294, "learning_rate": 0.00018536030438316608, "loss": 4.5084, "step": 664150 }, { "epoch": 1.2394705842560865, "grad_norm": 1.1889894008636475, "learning_rate": 0.0001853581159159539, "loss": 4.4631, "step": 664200 }, { "epoch": 1.2395638898154648, "grad_norm": 1.1519922018051147, "learning_rate": 0.00018535592729809967, "loss": 4.4583, "step": 664250 }, { "epoch": 1.2396571953748434, "grad_norm": 0.7938538789749146, "learning_rate": 0.00018535373852960723, "loss": 4.3554, "step": 664300 }, { "epoch": 1.239750500934222, "grad_norm": 1.2883696556091309, "learning_rate": 0.00018535154961048047, "loss": 4.6737, "step": 664350 }, { "epoch": 1.2398438064936004, "grad_norm": 0.9200964570045471, "learning_rate": 0.00018534936054072322, "loss": 4.3024, "step": 664400 }, { "epoch": 1.239937112052979, "grad_norm": 0.9933300018310547, "learning_rate": 0.00018534717132033938, "loss": 4.445, "step": 664450 }, { "epoch": 1.2400304176123573, "grad_norm": 1.3174785375595093, "learning_rate": 0.00018534498194933282, "loss": 4.5894, "step": 664500 }, { "epoch": 1.240123723171736, "grad_norm": 0.8731040358543396, "learning_rate": 0.00018534279242770734, "loss": 4.5632, "step": 664550 }, { "epoch": 1.2402170287311143, "grad_norm": 1.3123462200164795, "learning_rate": 0.0001853406027554669, "loss": 4.3876, "step": 664600 }, { "epoch": 1.2403103342904929, "grad_norm": 1.386484980583191, "learning_rate": 0.0001853384129326153, "loss": 4.6027, "step": 664650 }, { "epoch": 1.2404036398498715, "grad_norm": 1.3694813251495361, "learning_rate": 0.0001853362229591564, "loss": 4.5094, "step": 664700 }, { "epoch": 1.2404969454092498, "grad_norm": 1.014398455619812, "learning_rate": 0.0001853340328350941, "loss": 4.5536, "step": 664750 }, { "epoch": 1.2405902509686284, "grad_norm": 1.162325143814087, "learning_rate": 0.00018533184256043224, "loss": 4.5192, "step": 664800 }, { "epoch": 1.2406835565280068, "grad_norm": 1.1127030849456787, "learning_rate": 0.00018532965213517466, "loss": 4.4858, "step": 664850 }, { "epoch": 1.2407768620873854, "grad_norm": 1.3457621335983276, "learning_rate": 0.0001853274615593253, "loss": 4.5377, "step": 664900 }, { "epoch": 1.2408701676467637, "grad_norm": 1.1787924766540527, "learning_rate": 0.00018532527083288795, "loss": 4.453, "step": 664950 }, { "epoch": 1.2409634732061423, "grad_norm": 0.9719934463500977, "learning_rate": 0.00018532307995586654, "loss": 4.2891, "step": 665000 }, { "epoch": 1.241056778765521, "grad_norm": 1.1712827682495117, "learning_rate": 0.00018532088892826492, "loss": 4.4445, "step": 665050 }, { "epoch": 1.2411500843248993, "grad_norm": 1.285674810409546, "learning_rate": 0.00018531869775008692, "loss": 4.5468, "step": 665100 }, { "epoch": 1.2412433898842778, "grad_norm": 1.077782392501831, "learning_rate": 0.0001853165064213364, "loss": 4.2012, "step": 665150 }, { "epoch": 1.2413366954436562, "grad_norm": 1.181934118270874, "learning_rate": 0.0001853143149420173, "loss": 4.3056, "step": 665200 }, { "epoch": 1.2414300010030348, "grad_norm": 1.2469463348388672, "learning_rate": 0.00018531212331213346, "loss": 4.6928, "step": 665250 }, { "epoch": 1.2415233065624132, "grad_norm": 1.0450129508972168, "learning_rate": 0.00018530993153168868, "loss": 4.6111, "step": 665300 }, { "epoch": 1.2416166121217918, "grad_norm": 1.1256777048110962, "learning_rate": 0.00018530773960068693, "loss": 4.5711, "step": 665350 }, { "epoch": 1.2417099176811703, "grad_norm": 1.0177929401397705, "learning_rate": 0.000185305547519132, "loss": 4.3905, "step": 665400 }, { "epoch": 1.2418032232405487, "grad_norm": 0.7534099817276001, "learning_rate": 0.0001853033552870278, "loss": 4.4262, "step": 665450 }, { "epoch": 1.2418965287999273, "grad_norm": 1.326778769493103, "learning_rate": 0.0001853011629043782, "loss": 4.6255, "step": 665500 }, { "epoch": 1.2419898343593057, "grad_norm": 1.5219345092773438, "learning_rate": 0.00018529897037118705, "loss": 4.5337, "step": 665550 }, { "epoch": 1.2420831399186842, "grad_norm": 1.1565732955932617, "learning_rate": 0.0001852967776874582, "loss": 4.4327, "step": 665600 }, { "epoch": 1.2421764454780626, "grad_norm": 1.2199112176895142, "learning_rate": 0.00018529458485319557, "loss": 4.4243, "step": 665650 }, { "epoch": 1.2422697510374412, "grad_norm": 0.8797560334205627, "learning_rate": 0.000185292391868403, "loss": 4.3249, "step": 665700 }, { "epoch": 1.2423630565968198, "grad_norm": 0.8820335268974304, "learning_rate": 0.0001852901987330844, "loss": 4.1931, "step": 665750 }, { "epoch": 1.2424563621561981, "grad_norm": 1.0334835052490234, "learning_rate": 0.00018528800544724355, "loss": 4.7528, "step": 665800 }, { "epoch": 1.2425496677155767, "grad_norm": 1.3034424781799316, "learning_rate": 0.0001852858120108844, "loss": 4.512, "step": 665850 }, { "epoch": 1.242642973274955, "grad_norm": 1.2331149578094482, "learning_rate": 0.00018528361842401082, "loss": 4.4708, "step": 665900 }, { "epoch": 1.2427362788343337, "grad_norm": 0.8885931372642517, "learning_rate": 0.0001852814246866266, "loss": 4.2552, "step": 665950 }, { "epoch": 1.242829584393712, "grad_norm": 0.9997462630271912, "learning_rate": 0.00018527923079873575, "loss": 4.661, "step": 666000 }, { "epoch": 1.242829584393712, "eval_loss": 4.680281162261963, "eval_runtime": 229.0107, "eval_samples_per_second": 11.388, "eval_steps_per_second": 11.388, "eval_tts_loss": 7.6560429558200775, "step": 666000 }, { "epoch": 1.2429228899530906, "grad_norm": 1.2796733379364014, "learning_rate": 0.000185277036760342, "loss": 4.5106, "step": 666050 }, { "epoch": 1.2430161955124692, "grad_norm": 1.3298333883285522, "learning_rate": 0.00018527484257144933, "loss": 4.4034, "step": 666100 }, { "epoch": 1.2431095010718476, "grad_norm": 1.0224436521530151, "learning_rate": 0.00018527264823206153, "loss": 4.4186, "step": 666150 }, { "epoch": 1.2432028066312262, "grad_norm": 1.1939806938171387, "learning_rate": 0.00018527045374218252, "loss": 4.6442, "step": 666200 }, { "epoch": 1.2432961121906045, "grad_norm": 1.140907883644104, "learning_rate": 0.0001852682591018162, "loss": 4.4497, "step": 666250 }, { "epoch": 1.2433894177499831, "grad_norm": 1.1890062093734741, "learning_rate": 0.00018526606431096637, "loss": 4.4313, "step": 666300 }, { "epoch": 1.2434827233093615, "grad_norm": 1.293433427810669, "learning_rate": 0.0001852638693696369, "loss": 4.5425, "step": 666350 }, { "epoch": 1.24357602886874, "grad_norm": 1.1131811141967773, "learning_rate": 0.0001852616742778318, "loss": 4.4821, "step": 666400 }, { "epoch": 1.2436693344281187, "grad_norm": 0.7742553353309631, "learning_rate": 0.00018525947903555475, "loss": 4.6167, "step": 666450 }, { "epoch": 1.243762639987497, "grad_norm": 1.2441455125808716, "learning_rate": 0.0001852572836428098, "loss": 4.5646, "step": 666500 }, { "epoch": 1.2438559455468756, "grad_norm": 1.2040443420410156, "learning_rate": 0.0001852550880996007, "loss": 4.3645, "step": 666550 }, { "epoch": 1.243949251106254, "grad_norm": 1.2166908979415894, "learning_rate": 0.00018525289240593139, "loss": 4.4965, "step": 666600 }, { "epoch": 1.2440425566656326, "grad_norm": 1.302786946296692, "learning_rate": 0.00018525069656180572, "loss": 4.6496, "step": 666650 }, { "epoch": 1.244135862225011, "grad_norm": 1.1337661743164062, "learning_rate": 0.00018524850056722756, "loss": 4.3786, "step": 666700 }, { "epoch": 1.2442291677843895, "grad_norm": 1.06217360496521, "learning_rate": 0.00018524630442220082, "loss": 4.696, "step": 666750 }, { "epoch": 1.244322473343768, "grad_norm": 1.071700096130371, "learning_rate": 0.00018524410812672932, "loss": 4.515, "step": 666800 }, { "epoch": 1.2444157789031465, "grad_norm": 0.9893403053283691, "learning_rate": 0.000185241911680817, "loss": 4.4643, "step": 666850 }, { "epoch": 1.244509084462525, "grad_norm": 1.1475359201431274, "learning_rate": 0.00018523971508446768, "loss": 4.4122, "step": 666900 }, { "epoch": 1.2446023900219034, "grad_norm": 1.1715385913848877, "learning_rate": 0.0001852375183376853, "loss": 4.6125, "step": 666950 }, { "epoch": 1.244695695581282, "grad_norm": 1.2944334745407104, "learning_rate": 0.00018523532144047364, "loss": 4.4897, "step": 667000 }, { "epoch": 1.2447890011406604, "grad_norm": 1.3338533639907837, "learning_rate": 0.00018523312439283667, "loss": 4.4523, "step": 667050 }, { "epoch": 1.244882306700039, "grad_norm": 1.1022392511367798, "learning_rate": 0.00018523092719477822, "loss": 4.5925, "step": 667100 }, { "epoch": 1.2449756122594176, "grad_norm": 1.4297833442687988, "learning_rate": 0.0001852287298463022, "loss": 4.5999, "step": 667150 }, { "epoch": 1.245068917818796, "grad_norm": 1.15416419506073, "learning_rate": 0.00018522653234741244, "loss": 4.5205, "step": 667200 }, { "epoch": 1.2451622233781745, "grad_norm": 1.3173967599868774, "learning_rate": 0.00018522433469811287, "loss": 4.5857, "step": 667250 }, { "epoch": 1.2452555289375529, "grad_norm": 1.1590454578399658, "learning_rate": 0.00018522213689840732, "loss": 4.4749, "step": 667300 }, { "epoch": 1.2453488344969315, "grad_norm": 1.3843194246292114, "learning_rate": 0.00018521993894829973, "loss": 4.6125, "step": 667350 }, { "epoch": 1.2454421400563098, "grad_norm": 0.83402019739151, "learning_rate": 0.00018521774084779392, "loss": 4.4224, "step": 667400 }, { "epoch": 1.2455354456156884, "grad_norm": 1.1673972606658936, "learning_rate": 0.0001852155425968938, "loss": 4.5124, "step": 667450 }, { "epoch": 1.245628751175067, "grad_norm": 1.0598748922348022, "learning_rate": 0.00018521334419560324, "loss": 4.6377, "step": 667500 }, { "epoch": 1.2457220567344454, "grad_norm": 1.1959640979766846, "learning_rate": 0.00018521114564392613, "loss": 4.4743, "step": 667550 }, { "epoch": 1.245815362293824, "grad_norm": 1.2341710329055786, "learning_rate": 0.00018520894694186633, "loss": 4.4577, "step": 667600 }, { "epoch": 1.2459086678532023, "grad_norm": 1.1307636499404907, "learning_rate": 0.00018520674808942774, "loss": 4.6055, "step": 667650 }, { "epoch": 1.246001973412581, "grad_norm": 0.7109047174453735, "learning_rate": 0.0001852045490866142, "loss": 4.522, "step": 667700 }, { "epoch": 1.2460952789719593, "grad_norm": 1.2398737668991089, "learning_rate": 0.00018520234993342965, "loss": 4.4651, "step": 667750 }, { "epoch": 1.2461885845313379, "grad_norm": 1.1565322875976562, "learning_rate": 0.00018520015062987796, "loss": 4.4868, "step": 667800 }, { "epoch": 1.2462818900907164, "grad_norm": 1.3676830530166626, "learning_rate": 0.00018519795117596298, "loss": 4.6424, "step": 667850 }, { "epoch": 1.2463751956500948, "grad_norm": 1.2071691751480103, "learning_rate": 0.00018519575157168857, "loss": 4.4907, "step": 667900 }, { "epoch": 1.2464685012094734, "grad_norm": 1.131850004196167, "learning_rate": 0.0001851935518170587, "loss": 4.3589, "step": 667950 }, { "epoch": 1.2465618067688518, "grad_norm": 1.213726282119751, "learning_rate": 0.0001851913519120772, "loss": 4.507, "step": 668000 }, { "epoch": 1.2466551123282303, "grad_norm": 1.1079111099243164, "learning_rate": 0.00018518915185674793, "loss": 4.6338, "step": 668050 }, { "epoch": 1.2467484178876087, "grad_norm": 1.3508548736572266, "learning_rate": 0.0001851869516510748, "loss": 4.5465, "step": 668100 }, { "epoch": 1.2468417234469873, "grad_norm": 1.2168718576431274, "learning_rate": 0.0001851847512950617, "loss": 4.5883, "step": 668150 }, { "epoch": 1.2469350290063659, "grad_norm": 1.2854657173156738, "learning_rate": 0.00018518255078871248, "loss": 4.5707, "step": 668200 }, { "epoch": 1.2470283345657442, "grad_norm": 1.5036282539367676, "learning_rate": 0.00018518035013203107, "loss": 4.6087, "step": 668250 }, { "epoch": 1.2471216401251228, "grad_norm": 0.926520049571991, "learning_rate": 0.00018517814932502132, "loss": 4.5382, "step": 668300 }, { "epoch": 1.2472149456845012, "grad_norm": 1.1330171823501587, "learning_rate": 0.00018517594836768716, "loss": 4.5463, "step": 668350 }, { "epoch": 1.2473082512438798, "grad_norm": 1.069250226020813, "learning_rate": 0.00018517374726003237, "loss": 4.3579, "step": 668400 }, { "epoch": 1.2474015568032581, "grad_norm": 1.4110642671585083, "learning_rate": 0.00018517154600206095, "loss": 4.6466, "step": 668450 }, { "epoch": 1.2474948623626367, "grad_norm": 1.086166262626648, "learning_rate": 0.00018516934459377672, "loss": 4.385, "step": 668500 }, { "epoch": 1.2475881679220153, "grad_norm": 1.3381248712539673, "learning_rate": 0.0001851671430351836, "loss": 4.4985, "step": 668550 }, { "epoch": 1.2476814734813937, "grad_norm": 1.149175763130188, "learning_rate": 0.00018516494132628542, "loss": 4.5983, "step": 668600 }, { "epoch": 1.2477747790407723, "grad_norm": 1.3954811096191406, "learning_rate": 0.00018516273946708614, "loss": 4.5819, "step": 668650 }, { "epoch": 1.2478680846001506, "grad_norm": 0.9269979596138, "learning_rate": 0.0001851605374575896, "loss": 4.5264, "step": 668700 }, { "epoch": 1.2479613901595292, "grad_norm": 1.207831859588623, "learning_rate": 0.00018515833529779972, "loss": 4.5048, "step": 668750 }, { "epoch": 1.2480546957189076, "grad_norm": 1.3795820474624634, "learning_rate": 0.00018515613298772032, "loss": 4.3842, "step": 668800 }, { "epoch": 1.2481480012782862, "grad_norm": 1.0055474042892456, "learning_rate": 0.00018515393052735536, "loss": 4.3682, "step": 668850 }, { "epoch": 1.2482413068376648, "grad_norm": 1.110858678817749, "learning_rate": 0.00018515172791670867, "loss": 4.4439, "step": 668900 }, { "epoch": 1.2483346123970431, "grad_norm": 1.048248052597046, "learning_rate": 0.0001851495251557842, "loss": 4.445, "step": 668950 }, { "epoch": 1.2484279179564217, "grad_norm": 1.146277666091919, "learning_rate": 0.0001851473222445858, "loss": 4.2989, "step": 669000 }, { "epoch": 1.2484279179564217, "eval_loss": 4.6813764572143555, "eval_runtime": 228.674, "eval_samples_per_second": 11.405, "eval_steps_per_second": 11.405, "eval_tts_loss": 7.6151670969799845, "step": 669000 }, { "epoch": 1.2485212235158, "grad_norm": 0.9680073857307434, "learning_rate": 0.0001851451191831173, "loss": 4.465, "step": 669050 }, { "epoch": 1.2486145290751787, "grad_norm": 1.1596018075942993, "learning_rate": 0.00018514291597138266, "loss": 4.5504, "step": 669100 }, { "epoch": 1.248707834634557, "grad_norm": 1.0267274379730225, "learning_rate": 0.00018514071260938579, "loss": 4.3805, "step": 669150 }, { "epoch": 1.2488011401939356, "grad_norm": 0.9124011993408203, "learning_rate": 0.00018513850909713053, "loss": 4.6286, "step": 669200 }, { "epoch": 1.2488944457533142, "grad_norm": 1.0320978164672852, "learning_rate": 0.0001851363054346208, "loss": 4.4821, "step": 669250 }, { "epoch": 1.2489877513126926, "grad_norm": 1.0879652500152588, "learning_rate": 0.00018513410162186044, "loss": 4.5722, "step": 669300 }, { "epoch": 1.2490810568720712, "grad_norm": 1.3285393714904785, "learning_rate": 0.0001851318976588534, "loss": 4.6115, "step": 669350 }, { "epoch": 1.2491743624314495, "grad_norm": 1.1578034162521362, "learning_rate": 0.0001851296935456035, "loss": 4.4201, "step": 669400 }, { "epoch": 1.2492676679908281, "grad_norm": 1.4635430574417114, "learning_rate": 0.0001851274892821147, "loss": 4.5421, "step": 669450 }, { "epoch": 1.2493609735502065, "grad_norm": 1.1676958799362183, "learning_rate": 0.00018512528486839088, "loss": 4.4596, "step": 669500 }, { "epoch": 1.249454279109585, "grad_norm": 1.4455692768096924, "learning_rate": 0.00018512308030443585, "loss": 4.4797, "step": 669550 }, { "epoch": 1.2495475846689637, "grad_norm": 1.4037418365478516, "learning_rate": 0.00018512087559025358, "loss": 4.3518, "step": 669600 }, { "epoch": 1.249640890228342, "grad_norm": 1.290556788444519, "learning_rate": 0.00018511867072584798, "loss": 4.3673, "step": 669650 }, { "epoch": 1.2497341957877206, "grad_norm": 1.2086806297302246, "learning_rate": 0.0001851164657112229, "loss": 4.5122, "step": 669700 }, { "epoch": 1.249827501347099, "grad_norm": 1.1142642498016357, "learning_rate": 0.00018511426054638222, "loss": 4.6527, "step": 669750 }, { "epoch": 1.2499208069064776, "grad_norm": 1.308208703994751, "learning_rate": 0.00018511205523132982, "loss": 4.5061, "step": 669800 }, { "epoch": 1.250014112465856, "grad_norm": 1.0574862957000732, "learning_rate": 0.00018510984976606964, "loss": 4.6974, "step": 669850 }, { "epoch": 1.2501074180252345, "grad_norm": 0.9494866132736206, "learning_rate": 0.00018510764415060555, "loss": 4.7215, "step": 669900 }, { "epoch": 1.250200723584613, "grad_norm": 1.2254480123519897, "learning_rate": 0.00018510543838494146, "loss": 4.6163, "step": 669950 }, { "epoch": 1.2502940291439915, "grad_norm": 1.312201738357544, "learning_rate": 0.00018510323246908122, "loss": 4.4732, "step": 670000 }, { "epoch": 1.25038733470337, "grad_norm": 1.1795334815979004, "learning_rate": 0.00018510102640302876, "loss": 4.2927, "step": 670050 }, { "epoch": 1.2504806402627484, "grad_norm": 1.2151035070419312, "learning_rate": 0.00018509882018678795, "loss": 4.6788, "step": 670100 }, { "epoch": 1.250573945822127, "grad_norm": 1.0702143907546997, "learning_rate": 0.00018509661382036272, "loss": 4.4339, "step": 670150 }, { "epoch": 1.2506672513815054, "grad_norm": 1.2892051935195923, "learning_rate": 0.00018509440730375692, "loss": 4.5838, "step": 670200 }, { "epoch": 1.250760556940884, "grad_norm": 1.2257487773895264, "learning_rate": 0.00018509220063697446, "loss": 4.5494, "step": 670250 }, { "epoch": 1.2508538625002625, "grad_norm": 1.3180806636810303, "learning_rate": 0.00018508999382001927, "loss": 4.3731, "step": 670300 }, { "epoch": 1.250947168059641, "grad_norm": 1.0213433504104614, "learning_rate": 0.0001850877868528952, "loss": 4.4378, "step": 670350 }, { "epoch": 1.2510404736190193, "grad_norm": 0.6996897459030151, "learning_rate": 0.00018508557973560615, "loss": 4.4691, "step": 670400 }, { "epoch": 1.2511337791783979, "grad_norm": 0.8439481854438782, "learning_rate": 0.000185083372468156, "loss": 4.4479, "step": 670450 }, { "epoch": 1.2512270847377764, "grad_norm": 1.1673742532730103, "learning_rate": 0.00018508116505054872, "loss": 4.3795, "step": 670500 }, { "epoch": 1.2513203902971548, "grad_norm": 0.9821051955223083, "learning_rate": 0.00018507895748278813, "loss": 4.1294, "step": 670550 }, { "epoch": 1.2514136958565334, "grad_norm": 1.3237957954406738, "learning_rate": 0.00018507674976487811, "loss": 4.393, "step": 670600 }, { "epoch": 1.251507001415912, "grad_norm": 1.2885621786117554, "learning_rate": 0.00018507454189682266, "loss": 4.6591, "step": 670650 }, { "epoch": 1.2516003069752903, "grad_norm": 1.5363093614578247, "learning_rate": 0.00018507233387862559, "loss": 4.5089, "step": 670700 }, { "epoch": 1.2516936125346687, "grad_norm": 1.3501886129379272, "learning_rate": 0.0001850701257102908, "loss": 4.6479, "step": 670750 }, { "epoch": 1.2517869180940473, "grad_norm": 1.211461067199707, "learning_rate": 0.0001850679173918222, "loss": 4.2822, "step": 670800 }, { "epoch": 1.2518802236534259, "grad_norm": 1.1773744821548462, "learning_rate": 0.00018506570892322374, "loss": 4.5344, "step": 670850 }, { "epoch": 1.2519735292128042, "grad_norm": 0.8482151031494141, "learning_rate": 0.00018506350030449923, "loss": 4.4572, "step": 670900 }, { "epoch": 1.2520668347721828, "grad_norm": 1.131448745727539, "learning_rate": 0.00018506129153565267, "loss": 4.5269, "step": 670950 }, { "epoch": 1.2521601403315614, "grad_norm": 1.0034500360488892, "learning_rate": 0.00018505908261668783, "loss": 4.4769, "step": 671000 }, { "epoch": 1.2522534458909398, "grad_norm": 1.139027714729309, "learning_rate": 0.0001850568735476087, "loss": 4.4549, "step": 671050 }, { "epoch": 1.2523467514503182, "grad_norm": 1.1971385478973389, "learning_rate": 0.00018505466432841916, "loss": 4.5622, "step": 671100 }, { "epoch": 1.2524400570096967, "grad_norm": 1.4359465837478638, "learning_rate": 0.00018505245495912309, "loss": 4.5284, "step": 671150 }, { "epoch": 1.2525333625690753, "grad_norm": 1.1884537935256958, "learning_rate": 0.0001850502454397244, "loss": 4.5403, "step": 671200 }, { "epoch": 1.2526266681284537, "grad_norm": 0.9961506724357605, "learning_rate": 0.00018504803577022703, "loss": 4.4841, "step": 671250 }, { "epoch": 1.2527199736878323, "grad_norm": 1.3668924570083618, "learning_rate": 0.0001850458259506348, "loss": 4.6834, "step": 671300 }, { "epoch": 1.2528132792472109, "grad_norm": 1.1029890775680542, "learning_rate": 0.00018504361598095165, "loss": 4.6311, "step": 671350 }, { "epoch": 1.2529065848065892, "grad_norm": 1.1455581188201904, "learning_rate": 0.00018504140586118151, "loss": 4.4947, "step": 671400 }, { "epoch": 1.2529998903659676, "grad_norm": 1.201797604560852, "learning_rate": 0.00018503919559132825, "loss": 4.4551, "step": 671450 }, { "epoch": 1.2530931959253462, "grad_norm": 1.0861570835113525, "learning_rate": 0.00018503698517139574, "loss": 4.4583, "step": 671500 }, { "epoch": 1.2531865014847248, "grad_norm": 1.2867714166641235, "learning_rate": 0.00018503477460138793, "loss": 4.3901, "step": 671550 }, { "epoch": 1.2532798070441031, "grad_norm": 1.06697678565979, "learning_rate": 0.0001850325638813087, "loss": 4.615, "step": 671600 }, { "epoch": 1.2533731126034817, "grad_norm": 1.0974059104919434, "learning_rate": 0.00018503035301116198, "loss": 4.4989, "step": 671650 }, { "epoch": 1.2534664181628603, "grad_norm": 1.1045030355453491, "learning_rate": 0.00018502814199095163, "loss": 4.6234, "step": 671700 }, { "epoch": 1.2535597237222387, "grad_norm": 1.117690920829773, "learning_rate": 0.0001850259308206816, "loss": 4.4146, "step": 671750 }, { "epoch": 1.253653029281617, "grad_norm": 1.2372121810913086, "learning_rate": 0.00018502371950035573, "loss": 4.6804, "step": 671800 }, { "epoch": 1.2537463348409956, "grad_norm": 0.941888153553009, "learning_rate": 0.00018502150802997798, "loss": 4.2729, "step": 671850 }, { "epoch": 1.2538396404003742, "grad_norm": 1.2882981300354004, "learning_rate": 0.0001850192964095522, "loss": 4.3906, "step": 671900 }, { "epoch": 1.2539329459597526, "grad_norm": 1.0780928134918213, "learning_rate": 0.00018501708463908233, "loss": 4.4662, "step": 671950 }, { "epoch": 1.2540262515191312, "grad_norm": 1.2527164220809937, "learning_rate": 0.00018501487271857227, "loss": 4.7036, "step": 672000 }, { "epoch": 1.2540262515191312, "eval_loss": 4.684379577636719, "eval_runtime": 229.2992, "eval_samples_per_second": 11.374, "eval_steps_per_second": 11.374, "eval_tts_loss": 7.657498355492081, "step": 672000 }, { "epoch": 1.2541195570785098, "grad_norm": 1.1852953433990479, "learning_rate": 0.00018501266064802595, "loss": 4.7287, "step": 672050 }, { "epoch": 1.2542128626378881, "grad_norm": 1.0847089290618896, "learning_rate": 0.00018501044842744716, "loss": 4.6054, "step": 672100 }, { "epoch": 1.2543061681972665, "grad_norm": 1.5235379934310913, "learning_rate": 0.00018500823605683997, "loss": 4.5976, "step": 672150 }, { "epoch": 1.254399473756645, "grad_norm": 1.2535310983657837, "learning_rate": 0.00018500602353620815, "loss": 4.4337, "step": 672200 }, { "epoch": 1.2544927793160237, "grad_norm": 1.5305616855621338, "learning_rate": 0.00018500381086555568, "loss": 4.5659, "step": 672250 }, { "epoch": 1.254586084875402, "grad_norm": 1.3183680772781372, "learning_rate": 0.00018500159804488644, "loss": 4.6223, "step": 672300 }, { "epoch": 1.2546793904347806, "grad_norm": 0.9783040285110474, "learning_rate": 0.00018499938507420433, "loss": 4.521, "step": 672350 }, { "epoch": 1.2547726959941592, "grad_norm": 1.173769235610962, "learning_rate": 0.00018499717195351325, "loss": 4.5007, "step": 672400 }, { "epoch": 1.2548660015535376, "grad_norm": 0.967626690864563, "learning_rate": 0.00018499495868281714, "loss": 4.524, "step": 672450 }, { "epoch": 1.254959307112916, "grad_norm": 1.3498642444610596, "learning_rate": 0.00018499274526211987, "loss": 4.6583, "step": 672500 }, { "epoch": 1.2550526126722945, "grad_norm": 1.1174567937850952, "learning_rate": 0.00018499053169142537, "loss": 4.4864, "step": 672550 }, { "epoch": 1.255145918231673, "grad_norm": 1.2687735557556152, "learning_rate": 0.00018498831797073752, "loss": 4.511, "step": 672600 }, { "epoch": 1.2552392237910515, "grad_norm": 0.887858510017395, "learning_rate": 0.00018498610410006024, "loss": 4.4724, "step": 672650 }, { "epoch": 1.25533252935043, "grad_norm": 1.019060492515564, "learning_rate": 0.00018498389007939748, "loss": 4.4233, "step": 672700 }, { "epoch": 1.2554258349098086, "grad_norm": 1.0394374132156372, "learning_rate": 0.00018498167590875306, "loss": 4.5728, "step": 672750 }, { "epoch": 1.255519140469187, "grad_norm": 1.1815487146377563, "learning_rate": 0.00018497946158813096, "loss": 4.3551, "step": 672800 }, { "epoch": 1.2556124460285654, "grad_norm": 1.2263134717941284, "learning_rate": 0.00018497724711753505, "loss": 4.5886, "step": 672850 }, { "epoch": 1.255705751587944, "grad_norm": 1.3408105373382568, "learning_rate": 0.00018497503249696924, "loss": 4.5301, "step": 672900 }, { "epoch": 1.2557990571473225, "grad_norm": 1.0370341539382935, "learning_rate": 0.00018497281772643747, "loss": 4.5171, "step": 672950 }, { "epoch": 1.255892362706701, "grad_norm": 0.8629051446914673, "learning_rate": 0.00018497060280594366, "loss": 4.4509, "step": 673000 }, { "epoch": 1.2559856682660795, "grad_norm": 1.2424461841583252, "learning_rate": 0.00018496838773549164, "loss": 4.6154, "step": 673050 }, { "epoch": 1.256078973825458, "grad_norm": 1.390733003616333, "learning_rate": 0.00018496617251508536, "loss": 4.521, "step": 673100 }, { "epoch": 1.2561722793848364, "grad_norm": 1.179004430770874, "learning_rate": 0.00018496395714472877, "loss": 4.5865, "step": 673150 }, { "epoch": 1.2562655849442148, "grad_norm": 1.1466048955917358, "learning_rate": 0.00018496174162442572, "loss": 4.6098, "step": 673200 }, { "epoch": 1.2563588905035934, "grad_norm": 1.0782053470611572, "learning_rate": 0.00018495952595418017, "loss": 4.6322, "step": 673250 }, { "epoch": 1.256452196062972, "grad_norm": 1.038752555847168, "learning_rate": 0.00018495731013399598, "loss": 4.4134, "step": 673300 }, { "epoch": 1.2565455016223503, "grad_norm": 1.1292320489883423, "learning_rate": 0.00018495509416387707, "loss": 4.5333, "step": 673350 }, { "epoch": 1.256638807181729, "grad_norm": 1.2808153629302979, "learning_rate": 0.00018495287804382743, "loss": 4.3112, "step": 673400 }, { "epoch": 1.2567321127411075, "grad_norm": 1.1191731691360474, "learning_rate": 0.00018495066177385088, "loss": 4.368, "step": 673450 }, { "epoch": 1.2568254183004859, "grad_norm": 1.212639331817627, "learning_rate": 0.00018494844535395133, "loss": 4.6388, "step": 673500 }, { "epoch": 1.2569187238598643, "grad_norm": 1.0033625364303589, "learning_rate": 0.00018494622878413274, "loss": 4.4537, "step": 673550 }, { "epoch": 1.2570120294192428, "grad_norm": 0.6983006596565247, "learning_rate": 0.000184944012064399, "loss": 4.491, "step": 673600 }, { "epoch": 1.2571053349786214, "grad_norm": 1.231045126914978, "learning_rate": 0.00018494179519475403, "loss": 4.5754, "step": 673650 }, { "epoch": 1.2571986405379998, "grad_norm": 1.012865424156189, "learning_rate": 0.00018493957817520174, "loss": 4.4743, "step": 673700 }, { "epoch": 1.2572919460973784, "grad_norm": 1.382179856300354, "learning_rate": 0.00018493736100574606, "loss": 4.5743, "step": 673750 }, { "epoch": 1.257385251656757, "grad_norm": 1.2653876543045044, "learning_rate": 0.00018493514368639086, "loss": 4.3686, "step": 673800 }, { "epoch": 1.2574785572161353, "grad_norm": 1.277330994606018, "learning_rate": 0.00018493292621714008, "loss": 4.3728, "step": 673850 }, { "epoch": 1.2575718627755137, "grad_norm": 0.9845053553581238, "learning_rate": 0.0001849307085979976, "loss": 4.5918, "step": 673900 }, { "epoch": 1.2576651683348923, "grad_norm": 1.2767815589904785, "learning_rate": 0.00018492849082896742, "loss": 4.3833, "step": 673950 }, { "epoch": 1.2577584738942709, "grad_norm": 1.18427574634552, "learning_rate": 0.00018492627291005336, "loss": 4.5669, "step": 674000 }, { "epoch": 1.2578517794536492, "grad_norm": 1.1793664693832397, "learning_rate": 0.00018492405484125938, "loss": 4.5591, "step": 674050 }, { "epoch": 1.2579450850130278, "grad_norm": 1.1636524200439453, "learning_rate": 0.00018492183662258938, "loss": 4.4713, "step": 674100 }, { "epoch": 1.2580383905724064, "grad_norm": 1.2880363464355469, "learning_rate": 0.0001849196182540473, "loss": 4.5325, "step": 674150 }, { "epoch": 1.2581316961317848, "grad_norm": 1.2224889993667603, "learning_rate": 0.000184917399735637, "loss": 4.2867, "step": 674200 }, { "epoch": 1.2582250016911631, "grad_norm": 1.2206158638000488, "learning_rate": 0.00018491518106736248, "loss": 4.5703, "step": 674250 }, { "epoch": 1.2583183072505417, "grad_norm": 0.8942020535469055, "learning_rate": 0.00018491296224922756, "loss": 4.6112, "step": 674300 }, { "epoch": 1.2584116128099203, "grad_norm": 1.2274607419967651, "learning_rate": 0.00018491074328123622, "loss": 4.4569, "step": 674350 }, { "epoch": 1.2585049183692987, "grad_norm": 1.3399341106414795, "learning_rate": 0.00018490852416339236, "loss": 4.6385, "step": 674400 }, { "epoch": 1.2585982239286773, "grad_norm": 0.9390766620635986, "learning_rate": 0.00018490630489569992, "loss": 4.6198, "step": 674450 }, { "epoch": 1.2586915294880558, "grad_norm": 1.2388652563095093, "learning_rate": 0.00018490408547816275, "loss": 4.5398, "step": 674500 }, { "epoch": 1.2587848350474342, "grad_norm": 1.276523232460022, "learning_rate": 0.00018490186591078482, "loss": 4.5708, "step": 674550 }, { "epoch": 1.2588781406068126, "grad_norm": 1.3922771215438843, "learning_rate": 0.00018489964619357004, "loss": 4.6924, "step": 674600 }, { "epoch": 1.2589714461661912, "grad_norm": 1.2666330337524414, "learning_rate": 0.00018489742632652235, "loss": 4.5944, "step": 674650 }, { "epoch": 1.2590647517255698, "grad_norm": 1.4954805374145508, "learning_rate": 0.00018489520630964558, "loss": 4.5831, "step": 674700 }, { "epoch": 1.2591580572849481, "grad_norm": 1.1129313707351685, "learning_rate": 0.00018489298614294376, "loss": 4.5139, "step": 674750 }, { "epoch": 1.2592513628443267, "grad_norm": 1.0135051012039185, "learning_rate": 0.00018489076582642073, "loss": 4.5336, "step": 674800 }, { "epoch": 1.2593446684037053, "grad_norm": 1.3396539688110352, "learning_rate": 0.00018488854536008047, "loss": 4.4814, "step": 674850 }, { "epoch": 1.2594379739630837, "grad_norm": 1.3152085542678833, "learning_rate": 0.00018488632474392684, "loss": 4.4553, "step": 674900 }, { "epoch": 1.259531279522462, "grad_norm": 0.9369372129440308, "learning_rate": 0.00018488410397796376, "loss": 4.5128, "step": 674950 }, { "epoch": 1.2596245850818406, "grad_norm": 1.447873830795288, "learning_rate": 0.0001848818830621952, "loss": 4.602, "step": 675000 }, { "epoch": 1.2596245850818406, "eval_loss": 4.678734302520752, "eval_runtime": 230.3996, "eval_samples_per_second": 11.319, "eval_steps_per_second": 11.319, "eval_tts_loss": 7.643828824011072, "step": 675000 }, { "epoch": 1.2597178906412192, "grad_norm": 1.050093650817871, "learning_rate": 0.00018487966199662504, "loss": 4.433, "step": 675050 }, { "epoch": 1.2598111962005976, "grad_norm": 1.5931801795959473, "learning_rate": 0.00018487744078125724, "loss": 4.4395, "step": 675100 }, { "epoch": 1.2599045017599761, "grad_norm": 0.9175257086753845, "learning_rate": 0.00018487521941609568, "loss": 4.6679, "step": 675150 }, { "epoch": 1.2599978073193547, "grad_norm": 1.1838937997817993, "learning_rate": 0.00018487299790114427, "loss": 4.5956, "step": 675200 }, { "epoch": 1.260091112878733, "grad_norm": 1.3313302993774414, "learning_rate": 0.00018487077623640695, "loss": 4.4877, "step": 675250 }, { "epoch": 1.2601844184381115, "grad_norm": 1.1364085674285889, "learning_rate": 0.0001848685544218877, "loss": 4.4865, "step": 675300 }, { "epoch": 1.26027772399749, "grad_norm": 1.0744816064834595, "learning_rate": 0.00018486633245759032, "loss": 4.7256, "step": 675350 }, { "epoch": 1.2603710295568686, "grad_norm": 0.9846819639205933, "learning_rate": 0.00018486411034351882, "loss": 4.4788, "step": 675400 }, { "epoch": 1.260464335116247, "grad_norm": 1.231880784034729, "learning_rate": 0.00018486188807967708, "loss": 4.4343, "step": 675450 }, { "epoch": 1.2605576406756256, "grad_norm": 0.9274787902832031, "learning_rate": 0.00018485966566606906, "loss": 4.5516, "step": 675500 }, { "epoch": 1.2606509462350042, "grad_norm": 1.3950375318527222, "learning_rate": 0.00018485744310269868, "loss": 4.4496, "step": 675550 }, { "epoch": 1.2607442517943825, "grad_norm": 0.9788838028907776, "learning_rate": 0.00018485522038956978, "loss": 4.5771, "step": 675600 }, { "epoch": 1.260837557353761, "grad_norm": 1.1778128147125244, "learning_rate": 0.00018485299752668644, "loss": 4.3818, "step": 675650 }, { "epoch": 1.2609308629131395, "grad_norm": 1.3202683925628662, "learning_rate": 0.00018485077451405241, "loss": 4.5194, "step": 675700 }, { "epoch": 1.261024168472518, "grad_norm": 1.1114870309829712, "learning_rate": 0.00018484855135167175, "loss": 4.5484, "step": 675750 }, { "epoch": 1.2611174740318964, "grad_norm": 1.2369747161865234, "learning_rate": 0.00018484632803954826, "loss": 4.6511, "step": 675800 }, { "epoch": 1.261210779591275, "grad_norm": 1.1194487810134888, "learning_rate": 0.00018484410457768598, "loss": 4.4106, "step": 675850 }, { "epoch": 1.2613040851506536, "grad_norm": 1.4283881187438965, "learning_rate": 0.00018484188096608878, "loss": 4.5447, "step": 675900 }, { "epoch": 1.261397390710032, "grad_norm": 1.2632185220718384, "learning_rate": 0.00018483965720476057, "loss": 4.367, "step": 675950 }, { "epoch": 1.2614906962694103, "grad_norm": 1.240343689918518, "learning_rate": 0.0001848374332937053, "loss": 4.4623, "step": 676000 }, { "epoch": 1.261584001828789, "grad_norm": 0.9438549876213074, "learning_rate": 0.0001848352092329269, "loss": 4.691, "step": 676050 }, { "epoch": 1.2616773073881675, "grad_norm": 0.922001302242279, "learning_rate": 0.0001848329850224293, "loss": 4.6595, "step": 676100 }, { "epoch": 1.2617706129475459, "grad_norm": 1.0436054468154907, "learning_rate": 0.00018483076066221635, "loss": 4.5351, "step": 676150 }, { "epoch": 1.2618639185069245, "grad_norm": 1.253005862236023, "learning_rate": 0.00018482853615229208, "loss": 4.3632, "step": 676200 }, { "epoch": 1.261957224066303, "grad_norm": 0.9166373610496521, "learning_rate": 0.00018482631149266036, "loss": 4.47, "step": 676250 }, { "epoch": 1.2620505296256814, "grad_norm": 0.6935456395149231, "learning_rate": 0.00018482408668332511, "loss": 4.4874, "step": 676300 }, { "epoch": 1.2621438351850598, "grad_norm": 1.0185860395431519, "learning_rate": 0.00018482186172429028, "loss": 4.4305, "step": 676350 }, { "epoch": 1.2622371407444384, "grad_norm": 1.2536160945892334, "learning_rate": 0.00018481963661555982, "loss": 4.359, "step": 676400 }, { "epoch": 1.262330446303817, "grad_norm": 1.3654159307479858, "learning_rate": 0.00018481741135713758, "loss": 4.3703, "step": 676450 }, { "epoch": 1.2624237518631953, "grad_norm": 1.1900140047073364, "learning_rate": 0.00018481518594902756, "loss": 4.4941, "step": 676500 }, { "epoch": 1.262517057422574, "grad_norm": 1.0242934226989746, "learning_rate": 0.00018481296039123363, "loss": 4.2698, "step": 676550 }, { "epoch": 1.2626103629819525, "grad_norm": 0.8850924372673035, "learning_rate": 0.0001848107346837598, "loss": 4.5733, "step": 676600 }, { "epoch": 1.2627036685413309, "grad_norm": 1.1098710298538208, "learning_rate": 0.00018480850882660993, "loss": 4.468, "step": 676650 }, { "epoch": 1.2627969741007092, "grad_norm": 0.8750025033950806, "learning_rate": 0.00018480628281978794, "loss": 4.6362, "step": 676700 }, { "epoch": 1.2628902796600878, "grad_norm": 1.285516619682312, "learning_rate": 0.00018480405666329775, "loss": 4.543, "step": 676750 }, { "epoch": 1.2629835852194664, "grad_norm": 1.3653290271759033, "learning_rate": 0.0001848018303571434, "loss": 4.4823, "step": 676800 }, { "epoch": 1.2630768907788448, "grad_norm": 1.1246299743652344, "learning_rate": 0.00018479960390132871, "loss": 4.5651, "step": 676850 }, { "epoch": 1.2631701963382234, "grad_norm": 1.1523019075393677, "learning_rate": 0.00018479737729585764, "loss": 4.6117, "step": 676900 }, { "epoch": 1.263263501897602, "grad_norm": 1.3792613744735718, "learning_rate": 0.00018479515054073412, "loss": 4.297, "step": 676950 }, { "epoch": 1.2633568074569803, "grad_norm": 1.0924694538116455, "learning_rate": 0.00018479292363596207, "loss": 4.5771, "step": 677000 }, { "epoch": 1.2634501130163587, "grad_norm": 0.8439170122146606, "learning_rate": 0.00018479069658154545, "loss": 4.3884, "step": 677050 }, { "epoch": 1.2635434185757373, "grad_norm": 1.2244864702224731, "learning_rate": 0.00018478846937748812, "loss": 4.589, "step": 677100 }, { "epoch": 1.2636367241351159, "grad_norm": 1.5647135972976685, "learning_rate": 0.00018478624202379412, "loss": 4.4372, "step": 677150 }, { "epoch": 1.2637300296944942, "grad_norm": 1.2246736288070679, "learning_rate": 0.0001847840145204673, "loss": 4.4617, "step": 677200 }, { "epoch": 1.2638233352538728, "grad_norm": 1.3606178760528564, "learning_rate": 0.0001847817868675116, "loss": 4.6354, "step": 677250 }, { "epoch": 1.2639166408132514, "grad_norm": 1.2321351766586304, "learning_rate": 0.00018477955906493097, "loss": 4.5144, "step": 677300 }, { "epoch": 1.2640099463726298, "grad_norm": 1.1658254861831665, "learning_rate": 0.00018477733111272934, "loss": 4.4422, "step": 677350 }, { "epoch": 1.2641032519320081, "grad_norm": 1.3232557773590088, "learning_rate": 0.00018477510301091063, "loss": 4.4344, "step": 677400 }, { "epoch": 1.2641965574913867, "grad_norm": 0.6871310472488403, "learning_rate": 0.0001847728747594788, "loss": 4.4044, "step": 677450 }, { "epoch": 1.2642898630507653, "grad_norm": 1.3084734678268433, "learning_rate": 0.00018477064635843775, "loss": 4.3462, "step": 677500 }, { "epoch": 1.2643831686101437, "grad_norm": 1.03783118724823, "learning_rate": 0.0001847684178077914, "loss": 4.5971, "step": 677550 }, { "epoch": 1.2644764741695222, "grad_norm": 1.368334174156189, "learning_rate": 0.00018476618910754375, "loss": 4.6386, "step": 677600 }, { "epoch": 1.2645697797289008, "grad_norm": 1.1412273645401, "learning_rate": 0.00018476396025769867, "loss": 4.5559, "step": 677650 }, { "epoch": 1.2646630852882792, "grad_norm": 1.0761845111846924, "learning_rate": 0.00018476173125826014, "loss": 4.463, "step": 677700 }, { "epoch": 1.2647563908476576, "grad_norm": 1.289672613143921, "learning_rate": 0.00018475950210923207, "loss": 4.5546, "step": 677750 }, { "epoch": 1.2648496964070362, "grad_norm": 1.2862247228622437, "learning_rate": 0.00018475727281061835, "loss": 4.7272, "step": 677800 }, { "epoch": 1.2649430019664147, "grad_norm": 1.2224860191345215, "learning_rate": 0.00018475504336242296, "loss": 4.5749, "step": 677850 }, { "epoch": 1.265036307525793, "grad_norm": 1.1127855777740479, "learning_rate": 0.00018475281376464984, "loss": 4.4237, "step": 677900 }, { "epoch": 1.2651296130851717, "grad_norm": 0.9493654370307922, "learning_rate": 0.00018475058401730293, "loss": 4.5772, "step": 677950 }, { "epoch": 1.2652229186445503, "grad_norm": 1.298736333847046, "learning_rate": 0.00018474835412038616, "loss": 4.5362, "step": 678000 }, { "epoch": 1.2652229186445503, "eval_loss": 4.6840386390686035, "eval_runtime": 229.1677, "eval_samples_per_second": 11.38, "eval_steps_per_second": 11.38, "eval_tts_loss": 7.709944981450295, "step": 678000 }, { "epoch": 1.2653162242039286, "grad_norm": 1.2786900997161865, "learning_rate": 0.00018474612407390345, "loss": 4.5715, "step": 678050 }, { "epoch": 1.265409529763307, "grad_norm": 0.7049222588539124, "learning_rate": 0.00018474389387785874, "loss": 4.4446, "step": 678100 }, { "epoch": 1.2655028353226856, "grad_norm": 1.2928568124771118, "learning_rate": 0.00018474166353225595, "loss": 4.6319, "step": 678150 }, { "epoch": 1.2655961408820642, "grad_norm": 1.0299468040466309, "learning_rate": 0.00018473943303709905, "loss": 4.6873, "step": 678200 }, { "epoch": 1.2656894464414425, "grad_norm": 0.9450460076332092, "learning_rate": 0.00018473720239239194, "loss": 4.4598, "step": 678250 }, { "epoch": 1.2657827520008211, "grad_norm": 1.0672329664230347, "learning_rate": 0.00018473497159813862, "loss": 4.576, "step": 678300 }, { "epoch": 1.2658760575601997, "grad_norm": 1.4044899940490723, "learning_rate": 0.00018473274065434296, "loss": 4.6208, "step": 678350 }, { "epoch": 1.265969363119578, "grad_norm": 1.2460038661956787, "learning_rate": 0.00018473050956100892, "loss": 4.5309, "step": 678400 }, { "epoch": 1.2660626686789564, "grad_norm": 1.2784771919250488, "learning_rate": 0.00018472827831814045, "loss": 4.6483, "step": 678450 }, { "epoch": 1.266155974238335, "grad_norm": 1.261493444442749, "learning_rate": 0.00018472604692574147, "loss": 4.5232, "step": 678500 }, { "epoch": 1.2662492797977136, "grad_norm": 0.9764151573181152, "learning_rate": 0.0001847238153838159, "loss": 4.5425, "step": 678550 }, { "epoch": 1.266342585357092, "grad_norm": 1.42178475856781, "learning_rate": 0.00018472158369236774, "loss": 4.6436, "step": 678600 }, { "epoch": 1.2664358909164706, "grad_norm": 1.102962851524353, "learning_rate": 0.0001847193518514009, "loss": 4.4034, "step": 678650 }, { "epoch": 1.266529196475849, "grad_norm": 1.1221246719360352, "learning_rate": 0.00018471711986091927, "loss": 4.4232, "step": 678700 }, { "epoch": 1.2666225020352275, "grad_norm": 1.1222387552261353, "learning_rate": 0.00018471488772092687, "loss": 4.3838, "step": 678750 }, { "epoch": 1.266715807594606, "grad_norm": 1.120452880859375, "learning_rate": 0.00018471265543142757, "loss": 4.4854, "step": 678800 }, { "epoch": 1.2668091131539845, "grad_norm": 1.262215256690979, "learning_rate": 0.00018471042299242536, "loss": 4.7756, "step": 678850 }, { "epoch": 1.266902418713363, "grad_norm": 1.0459620952606201, "learning_rate": 0.00018470819040392416, "loss": 4.4738, "step": 678900 }, { "epoch": 1.2669957242727414, "grad_norm": 1.3806945085525513, "learning_rate": 0.00018470595766592788, "loss": 4.6013, "step": 678950 }, { "epoch": 1.26708902983212, "grad_norm": 1.3304641246795654, "learning_rate": 0.0001847037247784405, "loss": 4.4349, "step": 679000 }, { "epoch": 1.2671823353914984, "grad_norm": 1.5261013507843018, "learning_rate": 0.00018470149174146597, "loss": 4.6264, "step": 679050 }, { "epoch": 1.267275640950877, "grad_norm": 1.3177268505096436, "learning_rate": 0.00018469925855500822, "loss": 4.2942, "step": 679100 }, { "epoch": 1.2673689465102553, "grad_norm": 1.3208204507827759, "learning_rate": 0.00018469702521907113, "loss": 4.551, "step": 679150 }, { "epoch": 1.267462252069634, "grad_norm": 1.1189994812011719, "learning_rate": 0.00018469479173365873, "loss": 4.49, "step": 679200 }, { "epoch": 1.2675555576290125, "grad_norm": 1.361460566520691, "learning_rate": 0.00018469255809877493, "loss": 4.6709, "step": 679250 }, { "epoch": 1.2676488631883909, "grad_norm": 1.2242043018341064, "learning_rate": 0.00018469032431442365, "loss": 4.5082, "step": 679300 }, { "epoch": 1.2677421687477695, "grad_norm": 1.2202632427215576, "learning_rate": 0.00018468809038060886, "loss": 4.5202, "step": 679350 }, { "epoch": 1.2678354743071478, "grad_norm": 1.2023708820343018, "learning_rate": 0.00018468585629733447, "loss": 4.4774, "step": 679400 }, { "epoch": 1.2679287798665264, "grad_norm": 1.2995531558990479, "learning_rate": 0.00018468362206460446, "loss": 4.566, "step": 679450 }, { "epoch": 1.2680220854259048, "grad_norm": 1.2138429880142212, "learning_rate": 0.00018468138768242275, "loss": 4.6744, "step": 679500 }, { "epoch": 1.2681153909852834, "grad_norm": 1.1316689252853394, "learning_rate": 0.00018467915315079333, "loss": 4.4412, "step": 679550 }, { "epoch": 1.268208696544662, "grad_norm": 1.6544005870819092, "learning_rate": 0.00018467691846972006, "loss": 4.5066, "step": 679600 }, { "epoch": 1.2683020021040403, "grad_norm": 1.381272792816162, "learning_rate": 0.00018467468363920694, "loss": 4.536, "step": 679650 }, { "epoch": 1.268395307663419, "grad_norm": 1.0533658266067505, "learning_rate": 0.00018467244865925792, "loss": 4.4223, "step": 679700 }, { "epoch": 1.2684886132227973, "grad_norm": 1.4160455465316772, "learning_rate": 0.0001846702135298769, "loss": 4.3915, "step": 679750 }, { "epoch": 1.2685819187821759, "grad_norm": 1.128787875175476, "learning_rate": 0.00018466797825106786, "loss": 4.5845, "step": 679800 }, { "epoch": 1.2686752243415542, "grad_norm": 1.339247465133667, "learning_rate": 0.00018466574282283475, "loss": 4.5902, "step": 679850 }, { "epoch": 1.2687685299009328, "grad_norm": 1.153434157371521, "learning_rate": 0.0001846635072451815, "loss": 4.6805, "step": 679900 }, { "epoch": 1.2688618354603114, "grad_norm": 0.923381507396698, "learning_rate": 0.000184661271518112, "loss": 4.3372, "step": 679950 }, { "epoch": 1.2689551410196898, "grad_norm": 0.959805965423584, "learning_rate": 0.00018465903564163032, "loss": 4.3198, "step": 680000 }, { "epoch": 1.2690484465790683, "grad_norm": 1.249977946281433, "learning_rate": 0.00018465679961574032, "loss": 4.555, "step": 680050 }, { "epoch": 1.2691417521384467, "grad_norm": 0.9630039930343628, "learning_rate": 0.00018465456344044596, "loss": 4.5726, "step": 680100 }, { "epoch": 1.2692350576978253, "grad_norm": 1.578234314918518, "learning_rate": 0.0001846523271157512, "loss": 4.5609, "step": 680150 }, { "epoch": 1.2693283632572037, "grad_norm": 1.216994285583496, "learning_rate": 0.00018465009064165997, "loss": 4.5672, "step": 680200 }, { "epoch": 1.2694216688165822, "grad_norm": 0.977780818939209, "learning_rate": 0.00018464785401817624, "loss": 4.5357, "step": 680250 }, { "epoch": 1.2695149743759608, "grad_norm": 1.3047444820404053, "learning_rate": 0.0001846456172453039, "loss": 4.6025, "step": 680300 }, { "epoch": 1.2696082799353392, "grad_norm": 0.958842396736145, "learning_rate": 0.00018464338032304697, "loss": 4.5976, "step": 680350 }, { "epoch": 1.2697015854947178, "grad_norm": 1.0226985216140747, "learning_rate": 0.0001846411432514094, "loss": 4.69, "step": 680400 }, { "epoch": 1.2697948910540962, "grad_norm": 1.2814065217971802, "learning_rate": 0.00018463890603039504, "loss": 4.772, "step": 680450 }, { "epoch": 1.2698881966134747, "grad_norm": 0.9521793723106384, "learning_rate": 0.00018463666866000794, "loss": 4.4945, "step": 680500 }, { "epoch": 1.269981502172853, "grad_norm": 1.1713703870773315, "learning_rate": 0.000184634431140252, "loss": 4.5229, "step": 680550 }, { "epoch": 1.2700748077322317, "grad_norm": 1.1780476570129395, "learning_rate": 0.00018463219347113118, "loss": 4.3864, "step": 680600 }, { "epoch": 1.2701681132916103, "grad_norm": 1.1456300020217896, "learning_rate": 0.00018462995565264945, "loss": 4.4831, "step": 680650 }, { "epoch": 1.2702614188509886, "grad_norm": 1.4831351041793823, "learning_rate": 0.0001846277176848107, "loss": 4.4291, "step": 680700 }, { "epoch": 1.2703547244103672, "grad_norm": 1.0602538585662842, "learning_rate": 0.00018462547956761896, "loss": 4.6966, "step": 680750 }, { "epoch": 1.2704480299697456, "grad_norm": 1.1563079357147217, "learning_rate": 0.00018462324130107813, "loss": 4.4865, "step": 680800 }, { "epoch": 1.2705413355291242, "grad_norm": 1.0178416967391968, "learning_rate": 0.00018462100288519216, "loss": 4.4987, "step": 680850 }, { "epoch": 1.2706346410885025, "grad_norm": 1.3962833881378174, "learning_rate": 0.000184618764319965, "loss": 4.3381, "step": 680900 }, { "epoch": 1.2707279466478811, "grad_norm": 0.9716721773147583, "learning_rate": 0.00018461652560540062, "loss": 4.4575, "step": 680950 }, { "epoch": 1.2708212522072597, "grad_norm": 1.4356253147125244, "learning_rate": 0.00018461428674150296, "loss": 4.6026, "step": 681000 }, { "epoch": 1.2708212522072597, "eval_loss": 4.67944860458374, "eval_runtime": 229.277, "eval_samples_per_second": 11.375, "eval_steps_per_second": 11.375, "eval_tts_loss": 7.571945258099129, "step": 681000 }, { "epoch": 1.270914557766638, "grad_norm": 1.1949741840362549, "learning_rate": 0.000184612047728276, "loss": 4.4544, "step": 681050 }, { "epoch": 1.2710078633260167, "grad_norm": 1.0949891805648804, "learning_rate": 0.00018460980856572362, "loss": 4.5532, "step": 681100 }, { "epoch": 1.271101168885395, "grad_norm": 1.1831259727478027, "learning_rate": 0.00018460756925384983, "loss": 4.356, "step": 681150 }, { "epoch": 1.2711944744447736, "grad_norm": 0.8143863677978516, "learning_rate": 0.00018460532979265855, "loss": 4.587, "step": 681200 }, { "epoch": 1.271287780004152, "grad_norm": 1.1938937902450562, "learning_rate": 0.00018460309018215377, "loss": 4.5514, "step": 681250 }, { "epoch": 1.2713810855635306, "grad_norm": 1.0087896585464478, "learning_rate": 0.00018460085042233944, "loss": 4.4322, "step": 681300 }, { "epoch": 1.2714743911229092, "grad_norm": 1.1534209251403809, "learning_rate": 0.0001845986105132195, "loss": 4.5268, "step": 681350 }, { "epoch": 1.2715676966822875, "grad_norm": 1.331303596496582, "learning_rate": 0.00018459637045479784, "loss": 4.4814, "step": 681400 }, { "epoch": 1.2716610022416661, "grad_norm": 1.254723310470581, "learning_rate": 0.00018459413024707853, "loss": 4.5311, "step": 681450 }, { "epoch": 1.2717543078010445, "grad_norm": 1.0136011838912964, "learning_rate": 0.00018459188989006545, "loss": 4.3483, "step": 681500 }, { "epoch": 1.271847613360423, "grad_norm": 0.9966015219688416, "learning_rate": 0.0001845896493837626, "loss": 4.4064, "step": 681550 }, { "epoch": 1.2719409189198014, "grad_norm": 1.3059234619140625, "learning_rate": 0.0001845874087281738, "loss": 4.5629, "step": 681600 }, { "epoch": 1.27203422447918, "grad_norm": 1.0130586624145508, "learning_rate": 0.0001845851679233032, "loss": 4.4852, "step": 681650 }, { "epoch": 1.2721275300385586, "grad_norm": 1.1917186975479126, "learning_rate": 0.00018458292696915462, "loss": 4.519, "step": 681700 }, { "epoch": 1.272220835597937, "grad_norm": 0.90953528881073, "learning_rate": 0.00018458068586573207, "loss": 4.4987, "step": 681750 }, { "epoch": 1.2723141411573156, "grad_norm": 0.9469759464263916, "learning_rate": 0.0001845784446130395, "loss": 4.6033, "step": 681800 }, { "epoch": 1.272407446716694, "grad_norm": 1.2046890258789062, "learning_rate": 0.00018457620321108085, "loss": 4.4837, "step": 681850 }, { "epoch": 1.2725007522760725, "grad_norm": 1.2973721027374268, "learning_rate": 0.00018457396165986008, "loss": 4.435, "step": 681900 }, { "epoch": 1.2725940578354509, "grad_norm": 1.2963703870773315, "learning_rate": 0.00018457171995938114, "loss": 4.5558, "step": 681950 }, { "epoch": 1.2726873633948295, "grad_norm": 0.9790825843811035, "learning_rate": 0.00018456947810964802, "loss": 4.3331, "step": 682000 }, { "epoch": 1.272780668954208, "grad_norm": 1.0612525939941406, "learning_rate": 0.00018456723611066464, "loss": 4.545, "step": 682050 }, { "epoch": 1.2728739745135864, "grad_norm": 1.018945336341858, "learning_rate": 0.00018456499396243495, "loss": 4.5841, "step": 682100 }, { "epoch": 1.272967280072965, "grad_norm": 1.2650353908538818, "learning_rate": 0.00018456275166496292, "loss": 4.4841, "step": 682150 }, { "epoch": 1.2730605856323434, "grad_norm": 1.0791069269180298, "learning_rate": 0.00018456050921825256, "loss": 4.4247, "step": 682200 }, { "epoch": 1.273153891191722, "grad_norm": 1.1520460844039917, "learning_rate": 0.00018455826662230776, "loss": 4.5687, "step": 682250 }, { "epoch": 1.2732471967511003, "grad_norm": 1.1916439533233643, "learning_rate": 0.00018455602387713248, "loss": 4.663, "step": 682300 }, { "epoch": 1.273340502310479, "grad_norm": 1.1835590600967407, "learning_rate": 0.00018455378098273072, "loss": 4.5419, "step": 682350 }, { "epoch": 1.2734338078698575, "grad_norm": 1.1858782768249512, "learning_rate": 0.0001845515379391064, "loss": 4.5992, "step": 682400 }, { "epoch": 1.2735271134292359, "grad_norm": 1.6548361778259277, "learning_rate": 0.00018454929474626347, "loss": 4.4689, "step": 682450 }, { "epoch": 1.2736204189886144, "grad_norm": 1.1629254817962646, "learning_rate": 0.00018454705140420597, "loss": 4.6945, "step": 682500 }, { "epoch": 1.2737137245479928, "grad_norm": 1.1832695007324219, "learning_rate": 0.00018454480791293775, "loss": 4.5669, "step": 682550 }, { "epoch": 1.2738070301073714, "grad_norm": 1.1979652643203735, "learning_rate": 0.00018454256427246284, "loss": 4.59, "step": 682600 }, { "epoch": 1.2739003356667498, "grad_norm": 1.2776238918304443, "learning_rate": 0.00018454032048278519, "loss": 4.4969, "step": 682650 }, { "epoch": 1.2739936412261283, "grad_norm": 1.0656739473342896, "learning_rate": 0.00018453807654390873, "loss": 4.6035, "step": 682700 }, { "epoch": 1.274086946785507, "grad_norm": 1.0953510999679565, "learning_rate": 0.00018453583245583747, "loss": 4.3117, "step": 682750 }, { "epoch": 1.2741802523448853, "grad_norm": 1.2329922914505005, "learning_rate": 0.0001845335882185753, "loss": 4.544, "step": 682800 }, { "epoch": 1.2742735579042639, "grad_norm": 1.2543703317642212, "learning_rate": 0.00018453134383212622, "loss": 4.4473, "step": 682850 }, { "epoch": 1.2743668634636423, "grad_norm": 1.0144957304000854, "learning_rate": 0.0001845290992964942, "loss": 4.462, "step": 682900 }, { "epoch": 1.2744601690230208, "grad_norm": 1.2518835067749023, "learning_rate": 0.00018452685461168322, "loss": 4.5156, "step": 682950 }, { "epoch": 1.2745534745823992, "grad_norm": 1.2375555038452148, "learning_rate": 0.00018452460977769718, "loss": 4.5045, "step": 683000 }, { "epoch": 1.2746467801417778, "grad_norm": 0.9784048199653625, "learning_rate": 0.00018452236479454012, "loss": 4.2923, "step": 683050 }, { "epoch": 1.2747400857011564, "grad_norm": 1.3766660690307617, "learning_rate": 0.00018452011966221593, "loss": 4.4397, "step": 683100 }, { "epoch": 1.2748333912605347, "grad_norm": 1.4794206619262695, "learning_rate": 0.0001845178743807286, "loss": 4.5424, "step": 683150 }, { "epoch": 1.2749266968199133, "grad_norm": 1.0107816457748413, "learning_rate": 0.0001845156289500821, "loss": 4.3714, "step": 683200 }, { "epoch": 1.2750200023792917, "grad_norm": 1.1785094738006592, "learning_rate": 0.0001845133833702804, "loss": 4.533, "step": 683250 }, { "epoch": 1.2751133079386703, "grad_norm": 1.3257737159729004, "learning_rate": 0.00018451113764132742, "loss": 4.5366, "step": 683300 }, { "epoch": 1.2752066134980486, "grad_norm": 1.2592499256134033, "learning_rate": 0.00018450889176322718, "loss": 4.4216, "step": 683350 }, { "epoch": 1.2752999190574272, "grad_norm": 1.3016138076782227, "learning_rate": 0.00018450664573598356, "loss": 4.5185, "step": 683400 }, { "epoch": 1.2753932246168058, "grad_norm": 1.3459314107894897, "learning_rate": 0.00018450439955960062, "loss": 4.5037, "step": 683450 }, { "epoch": 1.2754865301761842, "grad_norm": 1.1414835453033447, "learning_rate": 0.0001845021532340823, "loss": 4.5177, "step": 683500 }, { "epoch": 1.2755798357355628, "grad_norm": 1.3457715511322021, "learning_rate": 0.0001844999067594325, "loss": 4.5765, "step": 683550 }, { "epoch": 1.2756731412949411, "grad_norm": 1.0881266593933105, "learning_rate": 0.00018449766013565528, "loss": 4.6363, "step": 683600 }, { "epoch": 1.2757664468543197, "grad_norm": 0.9685719609260559, "learning_rate": 0.00018449541336275455, "loss": 4.4632, "step": 683650 }, { "epoch": 1.275859752413698, "grad_norm": 0.9479784965515137, "learning_rate": 0.00018449316644073428, "loss": 4.4635, "step": 683700 }, { "epoch": 1.2759530579730767, "grad_norm": 1.1909276247024536, "learning_rate": 0.00018449091936959843, "loss": 4.5727, "step": 683750 }, { "epoch": 1.2760463635324553, "grad_norm": 1.2693300247192383, "learning_rate": 0.00018448867214935097, "loss": 4.4067, "step": 683800 }, { "epoch": 1.2761396690918336, "grad_norm": 0.7554545998573303, "learning_rate": 0.00018448642477999588, "loss": 4.4708, "step": 683850 }, { "epoch": 1.2762329746512122, "grad_norm": 1.1346781253814697, "learning_rate": 0.0001844841772615371, "loss": 4.4822, "step": 683900 }, { "epoch": 1.2763262802105906, "grad_norm": 1.4270949363708496, "learning_rate": 0.00018448192959397863, "loss": 4.5484, "step": 683950 }, { "epoch": 1.2764195857699692, "grad_norm": 0.8210241198539734, "learning_rate": 0.0001844796817773244, "loss": 4.3571, "step": 684000 }, { "epoch": 1.2764195857699692, "eval_loss": 4.698482513427734, "eval_runtime": 229.1356, "eval_samples_per_second": 11.382, "eval_steps_per_second": 11.382, "eval_tts_loss": 7.578603884160941, "step": 684000 }, { "epoch": 1.2765128913293475, "grad_norm": 1.2036874294281006, "learning_rate": 0.00018447743381157843, "loss": 4.5618, "step": 684050 }, { "epoch": 1.2766061968887261, "grad_norm": 0.9324430227279663, "learning_rate": 0.00018447518569674463, "loss": 4.5278, "step": 684100 }, { "epoch": 1.2766995024481047, "grad_norm": 1.2076671123504639, "learning_rate": 0.00018447293743282698, "loss": 4.4777, "step": 684150 }, { "epoch": 1.276792808007483, "grad_norm": 1.2023981809616089, "learning_rate": 0.0001844706890198295, "loss": 4.6117, "step": 684200 }, { "epoch": 1.2768861135668617, "grad_norm": 1.0647075176239014, "learning_rate": 0.0001844684404577561, "loss": 4.4089, "step": 684250 }, { "epoch": 1.27697941912624, "grad_norm": 1.0800987482070923, "learning_rate": 0.00018446619174661074, "loss": 4.5235, "step": 684300 }, { "epoch": 1.2770727246856186, "grad_norm": 1.0828824043273926, "learning_rate": 0.0001844639428863974, "loss": 4.5436, "step": 684350 }, { "epoch": 1.277166030244997, "grad_norm": 1.453656792640686, "learning_rate": 0.00018446169387712012, "loss": 4.4641, "step": 684400 }, { "epoch": 1.2772593358043756, "grad_norm": 0.681578516960144, "learning_rate": 0.00018445944471878279, "loss": 4.5508, "step": 684450 }, { "epoch": 1.2773526413637541, "grad_norm": 1.0871641635894775, "learning_rate": 0.0001844571954113894, "loss": 4.1813, "step": 684500 }, { "epoch": 1.2774459469231325, "grad_norm": 1.0098086595535278, "learning_rate": 0.00018445494595494387, "loss": 4.6045, "step": 684550 }, { "epoch": 1.277539252482511, "grad_norm": 0.9367932081222534, "learning_rate": 0.0001844526963494503, "loss": 4.4097, "step": 684600 }, { "epoch": 1.2776325580418895, "grad_norm": 1.018019199371338, "learning_rate": 0.00018445044659491256, "loss": 4.5787, "step": 684650 }, { "epoch": 1.277725863601268, "grad_norm": 1.1098785400390625, "learning_rate": 0.0001844481966913346, "loss": 4.3769, "step": 684700 }, { "epoch": 1.2778191691606464, "grad_norm": 1.2084110975265503, "learning_rate": 0.00018444594663872047, "loss": 4.4362, "step": 684750 }, { "epoch": 1.277912474720025, "grad_norm": 1.5332664251327515, "learning_rate": 0.00018444369643707412, "loss": 4.4093, "step": 684800 }, { "epoch": 1.2780057802794036, "grad_norm": 0.9569141268730164, "learning_rate": 0.00018444144608639946, "loss": 4.6464, "step": 684850 }, { "epoch": 1.278099085838782, "grad_norm": 1.3428438901901245, "learning_rate": 0.0001844391955867005, "loss": 4.302, "step": 684900 }, { "epoch": 1.2781923913981605, "grad_norm": 1.1949706077575684, "learning_rate": 0.00018443694493798125, "loss": 4.732, "step": 684950 }, { "epoch": 1.278285696957539, "grad_norm": 1.0948492288589478, "learning_rate": 0.00018443469414024564, "loss": 4.4861, "step": 685000 }, { "epoch": 1.2783790025169175, "grad_norm": 1.3561238050460815, "learning_rate": 0.00018443244319349766, "loss": 4.6217, "step": 685050 }, { "epoch": 1.2784723080762959, "grad_norm": 0.8994597792625427, "learning_rate": 0.00018443019209774127, "loss": 4.5928, "step": 685100 }, { "epoch": 1.2785656136356744, "grad_norm": 1.278558611869812, "learning_rate": 0.00018442794085298046, "loss": 4.5473, "step": 685150 }, { "epoch": 1.278658919195053, "grad_norm": 1.0130949020385742, "learning_rate": 0.00018442568945921915, "loss": 4.5175, "step": 685200 }, { "epoch": 1.2787522247544314, "grad_norm": 1.2430975437164307, "learning_rate": 0.00018442343791646136, "loss": 4.663, "step": 685250 }, { "epoch": 1.27884553031381, "grad_norm": 1.4214489459991455, "learning_rate": 0.0001844211862247111, "loss": 4.6527, "step": 685300 }, { "epoch": 1.2789388358731884, "grad_norm": 1.063011646270752, "learning_rate": 0.00018441893438397226, "loss": 4.6807, "step": 685350 }, { "epoch": 1.279032141432567, "grad_norm": 1.5450149774551392, "learning_rate": 0.00018441668239424887, "loss": 4.4185, "step": 685400 }, { "epoch": 1.2791254469919453, "grad_norm": 1.0038058757781982, "learning_rate": 0.00018441443025554488, "loss": 4.5942, "step": 685450 }, { "epoch": 1.279218752551324, "grad_norm": 1.350097894668579, "learning_rate": 0.0001844121779678643, "loss": 4.3794, "step": 685500 }, { "epoch": 1.2793120581107025, "grad_norm": 1.231081247329712, "learning_rate": 0.00018440992553121107, "loss": 4.5688, "step": 685550 }, { "epoch": 1.2794053636700808, "grad_norm": 1.396668553352356, "learning_rate": 0.00018440767294558916, "loss": 4.546, "step": 685600 }, { "epoch": 1.2794986692294594, "grad_norm": 0.9601945877075195, "learning_rate": 0.00018440542021100256, "loss": 4.5186, "step": 685650 }, { "epoch": 1.2795919747888378, "grad_norm": 1.6573967933654785, "learning_rate": 0.00018440316732745525, "loss": 4.4952, "step": 685700 }, { "epoch": 1.2796852803482164, "grad_norm": 1.048809289932251, "learning_rate": 0.0001844009142949512, "loss": 4.4966, "step": 685750 }, { "epoch": 1.2797785859075947, "grad_norm": 1.2405446767807007, "learning_rate": 0.00018439866111349439, "loss": 4.7196, "step": 685800 }, { "epoch": 1.2798718914669733, "grad_norm": 1.3093500137329102, "learning_rate": 0.00018439640778308876, "loss": 4.4279, "step": 685850 }, { "epoch": 1.279965197026352, "grad_norm": 1.2604055404663086, "learning_rate": 0.00018439415430373838, "loss": 4.5294, "step": 685900 }, { "epoch": 1.2800585025857303, "grad_norm": 1.0512810945510864, "learning_rate": 0.00018439190067544712, "loss": 4.3145, "step": 685950 }, { "epoch": 1.2801518081451089, "grad_norm": 1.0449227094650269, "learning_rate": 0.00018438964689821902, "loss": 4.6729, "step": 686000 }, { "epoch": 1.2802451137044872, "grad_norm": 1.0509430170059204, "learning_rate": 0.00018438739297205805, "loss": 4.4621, "step": 686050 }, { "epoch": 1.2803384192638658, "grad_norm": 0.5579020977020264, "learning_rate": 0.0001843851388969682, "loss": 4.3435, "step": 686100 }, { "epoch": 1.2804317248232442, "grad_norm": 1.0548137426376343, "learning_rate": 0.00018438288467295338, "loss": 4.5255, "step": 686150 }, { "epoch": 1.2805250303826228, "grad_norm": 1.2272684574127197, "learning_rate": 0.00018438063030001763, "loss": 4.5465, "step": 686200 }, { "epoch": 1.2806183359420014, "grad_norm": 0.9534603953361511, "learning_rate": 0.00018437837577816492, "loss": 4.5464, "step": 686250 }, { "epoch": 1.2807116415013797, "grad_norm": 0.9632763862609863, "learning_rate": 0.00018437612110739924, "loss": 4.374, "step": 686300 }, { "epoch": 1.2808049470607583, "grad_norm": 1.2491899728775024, "learning_rate": 0.00018437386628772452, "loss": 4.5609, "step": 686350 }, { "epoch": 1.2808982526201367, "grad_norm": 1.0648865699768066, "learning_rate": 0.0001843716113191448, "loss": 4.5942, "step": 686400 }, { "epoch": 1.2809915581795153, "grad_norm": 1.2533990144729614, "learning_rate": 0.000184369356201664, "loss": 4.5534, "step": 686450 }, { "epoch": 1.2810848637388936, "grad_norm": 0.9580297470092773, "learning_rate": 0.00018436710093528618, "loss": 4.4751, "step": 686500 }, { "epoch": 1.2811781692982722, "grad_norm": 1.145330786705017, "learning_rate": 0.00018436484552001523, "loss": 4.4101, "step": 686550 }, { "epoch": 1.2812714748576508, "grad_norm": 1.3001517057418823, "learning_rate": 0.0001843625899558552, "loss": 4.5902, "step": 686600 }, { "epoch": 1.2813647804170292, "grad_norm": 1.3478891849517822, "learning_rate": 0.00018436033424281, "loss": 4.7241, "step": 686650 }, { "epoch": 1.2814580859764078, "grad_norm": 1.2944649457931519, "learning_rate": 0.0001843580783808837, "loss": 4.505, "step": 686700 }, { "epoch": 1.2815513915357861, "grad_norm": 1.024078607559204, "learning_rate": 0.00018435582237008022, "loss": 4.5902, "step": 686750 }, { "epoch": 1.2816446970951647, "grad_norm": 1.0610401630401611, "learning_rate": 0.00018435356621040357, "loss": 4.4799, "step": 686800 }, { "epoch": 1.281738002654543, "grad_norm": 0.9353392720222473, "learning_rate": 0.0001843513099018577, "loss": 4.5923, "step": 686850 }, { "epoch": 1.2818313082139217, "grad_norm": 0.9111230969429016, "learning_rate": 0.00018434905344444662, "loss": 4.5735, "step": 686900 }, { "epoch": 1.2819246137733002, "grad_norm": 1.2439987659454346, "learning_rate": 0.00018434679683817427, "loss": 4.5376, "step": 686950 }, { "epoch": 1.2820179193326786, "grad_norm": 1.285531997680664, "learning_rate": 0.00018434454008304472, "loss": 4.3985, "step": 687000 }, { "epoch": 1.2820179193326786, "eval_loss": 4.674289226531982, "eval_runtime": 233.1501, "eval_samples_per_second": 11.186, "eval_steps_per_second": 11.186, "eval_tts_loss": 7.635929417767865, "step": 687000 }, { "epoch": 1.2821112248920572, "grad_norm": 1.1917221546173096, "learning_rate": 0.00018434228317906187, "loss": 4.479, "step": 687050 }, { "epoch": 1.2822045304514356, "grad_norm": 1.394010305404663, "learning_rate": 0.00018434002612622972, "loss": 4.7206, "step": 687100 }, { "epoch": 1.2822978360108142, "grad_norm": 0.8601192235946655, "learning_rate": 0.00018433776892455228, "loss": 4.5149, "step": 687150 }, { "epoch": 1.2823911415701925, "grad_norm": 1.123416781425476, "learning_rate": 0.0001843355115740335, "loss": 4.3671, "step": 687200 }, { "epoch": 1.282484447129571, "grad_norm": 1.2890983819961548, "learning_rate": 0.00018433325407467738, "loss": 4.6271, "step": 687250 }, { "epoch": 1.2825777526889497, "grad_norm": 0.8121176958084106, "learning_rate": 0.00018433099642648792, "loss": 4.4446, "step": 687300 }, { "epoch": 1.282671058248328, "grad_norm": 1.1560871601104736, "learning_rate": 0.0001843287386294691, "loss": 4.3736, "step": 687350 }, { "epoch": 1.2827643638077066, "grad_norm": 1.148853063583374, "learning_rate": 0.00018432648068362486, "loss": 4.4874, "step": 687400 }, { "epoch": 1.282857669367085, "grad_norm": 1.0627683401107788, "learning_rate": 0.00018432422258895926, "loss": 4.5386, "step": 687450 }, { "epoch": 1.2829509749264636, "grad_norm": 1.5874539613723755, "learning_rate": 0.00018432196434547623, "loss": 4.4212, "step": 687500 }, { "epoch": 1.283044280485842, "grad_norm": 0.9959418773651123, "learning_rate": 0.00018431970595317977, "loss": 4.5191, "step": 687550 }, { "epoch": 1.2831375860452205, "grad_norm": 1.0339405536651611, "learning_rate": 0.00018431744741207384, "loss": 4.3719, "step": 687600 }, { "epoch": 1.2832308916045991, "grad_norm": 1.36300528049469, "learning_rate": 0.0001843151887221625, "loss": 4.4957, "step": 687650 }, { "epoch": 1.2833241971639775, "grad_norm": 1.1878827810287476, "learning_rate": 0.00018431292988344965, "loss": 4.5431, "step": 687700 }, { "epoch": 1.283417502723356, "grad_norm": 0.759378969669342, "learning_rate": 0.0001843106708959393, "loss": 4.5971, "step": 687750 }, { "epoch": 1.2835108082827345, "grad_norm": 1.2821844816207886, "learning_rate": 0.0001843084117596355, "loss": 4.5573, "step": 687800 }, { "epoch": 1.283604113842113, "grad_norm": 1.0793043375015259, "learning_rate": 0.00018430615247454215, "loss": 4.6274, "step": 687850 }, { "epoch": 1.2836974194014914, "grad_norm": 0.9859644174575806, "learning_rate": 0.00018430389304066328, "loss": 4.3562, "step": 687900 }, { "epoch": 1.28379072496087, "grad_norm": 0.9679508209228516, "learning_rate": 0.0001843016334580029, "loss": 4.3474, "step": 687950 }, { "epoch": 1.2838840305202486, "grad_norm": 1.0614757537841797, "learning_rate": 0.00018429937372656493, "loss": 4.4651, "step": 688000 }, { "epoch": 1.283977336079627, "grad_norm": 1.3508762121200562, "learning_rate": 0.00018429711384635342, "loss": 4.4153, "step": 688050 }, { "epoch": 1.2840706416390055, "grad_norm": 1.0053499937057495, "learning_rate": 0.00018429485381737232, "loss": 4.3141, "step": 688100 }, { "epoch": 1.284163947198384, "grad_norm": 0.9558650255203247, "learning_rate": 0.00018429259363962567, "loss": 4.7093, "step": 688150 }, { "epoch": 1.2842572527577625, "grad_norm": 1.0553396940231323, "learning_rate": 0.0001842903333131174, "loss": 4.6149, "step": 688200 }, { "epoch": 1.2843505583171408, "grad_norm": 1.0472782850265503, "learning_rate": 0.00018428807283785154, "loss": 4.5918, "step": 688250 }, { "epoch": 1.2844438638765194, "grad_norm": 0.9262356162071228, "learning_rate": 0.00018428581221383207, "loss": 4.3696, "step": 688300 }, { "epoch": 1.284537169435898, "grad_norm": 1.0140273571014404, "learning_rate": 0.00018428355144106294, "loss": 4.6094, "step": 688350 }, { "epoch": 1.2846304749952764, "grad_norm": 1.1281394958496094, "learning_rate": 0.00018428129051954818, "loss": 4.6163, "step": 688400 }, { "epoch": 1.284723780554655, "grad_norm": 1.1518676280975342, "learning_rate": 0.00018427902944929178, "loss": 4.5044, "step": 688450 }, { "epoch": 1.2848170861140333, "grad_norm": 0.8620916604995728, "learning_rate": 0.00018427676823029772, "loss": 4.2609, "step": 688500 }, { "epoch": 1.284910391673412, "grad_norm": 1.1241519451141357, "learning_rate": 0.00018427450686257, "loss": 4.3235, "step": 688550 }, { "epoch": 1.2850036972327903, "grad_norm": 1.0311418771743774, "learning_rate": 0.0001842722453461126, "loss": 4.5816, "step": 688600 }, { "epoch": 1.2850970027921689, "grad_norm": 1.0403538942337036, "learning_rate": 0.0001842699836809295, "loss": 4.5807, "step": 688650 }, { "epoch": 1.2851903083515475, "grad_norm": 0.9353972673416138, "learning_rate": 0.0001842677218670247, "loss": 4.5879, "step": 688700 }, { "epoch": 1.2852836139109258, "grad_norm": 1.202971339225769, "learning_rate": 0.0001842654599044022, "loss": 4.6513, "step": 688750 }, { "epoch": 1.2853769194703044, "grad_norm": 1.1938093900680542, "learning_rate": 0.000184263197793066, "loss": 4.4479, "step": 688800 }, { "epoch": 1.2854702250296828, "grad_norm": 1.5217254161834717, "learning_rate": 0.00018426093553302009, "loss": 4.4204, "step": 688850 }, { "epoch": 1.2855635305890614, "grad_norm": 0.873814582824707, "learning_rate": 0.00018425867312426845, "loss": 4.3885, "step": 688900 }, { "epoch": 1.2856568361484397, "grad_norm": 1.159995198249817, "learning_rate": 0.00018425641056681508, "loss": 4.4758, "step": 688950 }, { "epoch": 1.2857501417078183, "grad_norm": 1.2675409317016602, "learning_rate": 0.000184254147860664, "loss": 4.6563, "step": 689000 }, { "epoch": 1.285843447267197, "grad_norm": 0.8674375414848328, "learning_rate": 0.0001842518850058191, "loss": 4.4234, "step": 689050 }, { "epoch": 1.2859367528265753, "grad_norm": 1.3953529596328735, "learning_rate": 0.0001842496220022845, "loss": 4.6421, "step": 689100 }, { "epoch": 1.2860300583859539, "grad_norm": 1.1026346683502197, "learning_rate": 0.00018424735885006418, "loss": 4.7101, "step": 689150 }, { "epoch": 1.2861233639453322, "grad_norm": 1.0520707368850708, "learning_rate": 0.00018424509554916202, "loss": 4.5083, "step": 689200 }, { "epoch": 1.2862166695047108, "grad_norm": 1.173195242881775, "learning_rate": 0.0001842428320995821, "loss": 4.6624, "step": 689250 }, { "epoch": 1.2863099750640892, "grad_norm": 0.8486551642417908, "learning_rate": 0.00018424056850132844, "loss": 4.5736, "step": 689300 }, { "epoch": 1.2864032806234678, "grad_norm": 1.281187653541565, "learning_rate": 0.000184238304754405, "loss": 4.5911, "step": 689350 }, { "epoch": 1.2864965861828463, "grad_norm": 0.8941449522972107, "learning_rate": 0.00018423604085881574, "loss": 4.5964, "step": 689400 }, { "epoch": 1.2865898917422247, "grad_norm": 0.8673259019851685, "learning_rate": 0.0001842337768145647, "loss": 4.3949, "step": 689450 }, { "epoch": 1.2866831973016033, "grad_norm": 1.2519278526306152, "learning_rate": 0.00018423151262165588, "loss": 4.631, "step": 689500 }, { "epoch": 1.2867765028609817, "grad_norm": 1.132042646408081, "learning_rate": 0.00018422924828009326, "loss": 4.4582, "step": 689550 }, { "epoch": 1.2868698084203603, "grad_norm": 1.3662277460098267, "learning_rate": 0.0001842269837898808, "loss": 4.4246, "step": 689600 }, { "epoch": 1.2869631139797386, "grad_norm": 1.4120208024978638, "learning_rate": 0.0001842247191510226, "loss": 4.6338, "step": 689650 }, { "epoch": 1.2870564195391172, "grad_norm": 1.3141610622406006, "learning_rate": 0.00018422245436352255, "loss": 4.5251, "step": 689700 }, { "epoch": 1.2871497250984958, "grad_norm": 1.3096686601638794, "learning_rate": 0.0001842201894273847, "loss": 4.6418, "step": 689750 }, { "epoch": 1.2872430306578742, "grad_norm": 1.3683671951293945, "learning_rate": 0.00018421792434261303, "loss": 4.6784, "step": 689800 }, { "epoch": 1.2873363362172527, "grad_norm": 0.9529086351394653, "learning_rate": 0.00018421565910921152, "loss": 4.5799, "step": 689850 }, { "epoch": 1.287429641776631, "grad_norm": 1.014106035232544, "learning_rate": 0.0001842133937271842, "loss": 4.5265, "step": 689900 }, { "epoch": 1.2875229473360097, "grad_norm": 1.2727725505828857, "learning_rate": 0.00018421112819653508, "loss": 4.4312, "step": 689950 }, { "epoch": 1.287616252895388, "grad_norm": 1.1307765245437622, "learning_rate": 0.00018420886251726813, "loss": 4.5029, "step": 690000 }, { "epoch": 1.287616252895388, "eval_loss": 4.669147491455078, "eval_runtime": 229.4881, "eval_samples_per_second": 11.364, "eval_steps_per_second": 11.364, "eval_tts_loss": 7.616201093436099, "step": 690000 }, { "epoch": 1.2877095584547666, "grad_norm": 0.9264296293258667, "learning_rate": 0.00018420659668938734, "loss": 4.5725, "step": 690050 }, { "epoch": 1.2878028640141452, "grad_norm": 1.2746309041976929, "learning_rate": 0.00018420433071289672, "loss": 4.6201, "step": 690100 }, { "epoch": 1.2878961695735236, "grad_norm": 1.1417357921600342, "learning_rate": 0.0001842020645878003, "loss": 4.4495, "step": 690150 }, { "epoch": 1.2879894751329022, "grad_norm": 1.1852238178253174, "learning_rate": 0.00018419979831410204, "loss": 4.7611, "step": 690200 }, { "epoch": 1.2880827806922805, "grad_norm": 1.1793030500411987, "learning_rate": 0.00018419753189180592, "loss": 4.371, "step": 690250 }, { "epoch": 1.2881760862516591, "grad_norm": 1.12693190574646, "learning_rate": 0.00018419526532091602, "loss": 4.6073, "step": 690300 }, { "epoch": 1.2882693918110375, "grad_norm": 1.124153733253479, "learning_rate": 0.00018419299860143626, "loss": 4.6322, "step": 690350 }, { "epoch": 1.288362697370416, "grad_norm": 1.1916775703430176, "learning_rate": 0.00018419073173337068, "loss": 4.6684, "step": 690400 }, { "epoch": 1.2884560029297947, "grad_norm": 1.302439570426941, "learning_rate": 0.00018418846471672325, "loss": 4.6191, "step": 690450 }, { "epoch": 1.288549308489173, "grad_norm": 1.7275673151016235, "learning_rate": 0.00018418619755149803, "loss": 4.4394, "step": 690500 }, { "epoch": 1.2886426140485516, "grad_norm": 1.0587892532348633, "learning_rate": 0.00018418393023769897, "loss": 4.4967, "step": 690550 }, { "epoch": 1.28873591960793, "grad_norm": 1.5676963329315186, "learning_rate": 0.00018418166277533012, "loss": 4.3104, "step": 690600 }, { "epoch": 1.2888292251673086, "grad_norm": 0.9393081665039062, "learning_rate": 0.0001841793951643954, "loss": 4.4666, "step": 690650 }, { "epoch": 1.288922530726687, "grad_norm": 1.0360493659973145, "learning_rate": 0.00018417712740489885, "loss": 4.6519, "step": 690700 }, { "epoch": 1.2890158362860655, "grad_norm": 0.9734581708908081, "learning_rate": 0.00018417485949684452, "loss": 4.2668, "step": 690750 }, { "epoch": 1.2891091418454441, "grad_norm": 1.2650303840637207, "learning_rate": 0.00018417259144023636, "loss": 4.5115, "step": 690800 }, { "epoch": 1.2892024474048225, "grad_norm": 0.9618446826934814, "learning_rate": 0.0001841703232350784, "loss": 4.6256, "step": 690850 }, { "epoch": 1.289295752964201, "grad_norm": 1.1916064023971558, "learning_rate": 0.00018416805488137458, "loss": 4.4181, "step": 690900 }, { "epoch": 1.2893890585235794, "grad_norm": 1.1117289066314697, "learning_rate": 0.000184165786379129, "loss": 4.2692, "step": 690950 }, { "epoch": 1.289482364082958, "grad_norm": 1.2509772777557373, "learning_rate": 0.00018416351772834562, "loss": 4.4127, "step": 691000 }, { "epoch": 1.2895756696423364, "grad_norm": 1.2009600400924683, "learning_rate": 0.00018416124892902842, "loss": 4.2983, "step": 691050 }, { "epoch": 1.289668975201715, "grad_norm": 0.9763325452804565, "learning_rate": 0.0001841589799811814, "loss": 4.4246, "step": 691100 }, { "epoch": 1.2897622807610936, "grad_norm": 1.2052439451217651, "learning_rate": 0.00018415671088480864, "loss": 4.4702, "step": 691150 }, { "epoch": 1.289855586320472, "grad_norm": 1.0581713914871216, "learning_rate": 0.00018415444163991408, "loss": 4.7176, "step": 691200 }, { "epoch": 1.2899488918798505, "grad_norm": 1.2135562896728516, "learning_rate": 0.0001841521722465017, "loss": 4.4477, "step": 691250 }, { "epoch": 1.2900421974392289, "grad_norm": 1.1380517482757568, "learning_rate": 0.0001841499027045756, "loss": 4.4838, "step": 691300 }, { "epoch": 1.2901355029986075, "grad_norm": 1.119773030281067, "learning_rate": 0.00018414763301413968, "loss": 4.4631, "step": 691350 }, { "epoch": 1.2902288085579858, "grad_norm": 1.3427542448043823, "learning_rate": 0.000184145363175198, "loss": 4.5033, "step": 691400 }, { "epoch": 1.2903221141173644, "grad_norm": 1.0488735437393188, "learning_rate": 0.0001841430931877546, "loss": 4.5875, "step": 691450 }, { "epoch": 1.290415419676743, "grad_norm": 0.9796529412269592, "learning_rate": 0.00018414082305181343, "loss": 4.4388, "step": 691500 }, { "epoch": 1.2905087252361214, "grad_norm": 1.0404736995697021, "learning_rate": 0.0001841385527673785, "loss": 4.5732, "step": 691550 }, { "epoch": 1.2906020307955, "grad_norm": 1.2868335247039795, "learning_rate": 0.0001841362823344538, "loss": 4.6321, "step": 691600 }, { "epoch": 1.2906953363548783, "grad_norm": 1.1727312803268433, "learning_rate": 0.0001841340117530434, "loss": 4.2843, "step": 691650 }, { "epoch": 1.290788641914257, "grad_norm": 1.156617283821106, "learning_rate": 0.0001841317410231513, "loss": 4.5149, "step": 691700 }, { "epoch": 1.2908819474736353, "grad_norm": 1.0426771640777588, "learning_rate": 0.00018412947014478145, "loss": 4.4736, "step": 691750 }, { "epoch": 1.2909752530330139, "grad_norm": 1.0665901899337769, "learning_rate": 0.00018412719911793787, "loss": 4.6209, "step": 691800 }, { "epoch": 1.2910685585923924, "grad_norm": 1.2331914901733398, "learning_rate": 0.00018412492794262462, "loss": 4.3237, "step": 691850 }, { "epoch": 1.2911618641517708, "grad_norm": 1.2286107540130615, "learning_rate": 0.00018412265661884564, "loss": 4.6099, "step": 691900 }, { "epoch": 1.2912551697111494, "grad_norm": 1.1824464797973633, "learning_rate": 0.000184120385146605, "loss": 4.6875, "step": 691950 }, { "epoch": 1.2913484752705278, "grad_norm": 1.287689447402954, "learning_rate": 0.00018411811352590668, "loss": 4.3017, "step": 692000 }, { "epoch": 1.2914417808299063, "grad_norm": 1.2916274070739746, "learning_rate": 0.00018411584175675468, "loss": 4.4524, "step": 692050 }, { "epoch": 1.2915350863892847, "grad_norm": 1.2580164670944214, "learning_rate": 0.00018411356983915304, "loss": 4.4634, "step": 692100 }, { "epoch": 1.2916283919486633, "grad_norm": 1.3331435918807983, "learning_rate": 0.00018411129777310572, "loss": 4.5975, "step": 692150 }, { "epoch": 1.2917216975080419, "grad_norm": 1.2346582412719727, "learning_rate": 0.00018410902555861678, "loss": 4.503, "step": 692200 }, { "epoch": 1.2918150030674203, "grad_norm": 1.078025221824646, "learning_rate": 0.00018410675319569019, "loss": 4.6092, "step": 692250 }, { "epoch": 1.2919083086267986, "grad_norm": 1.5245380401611328, "learning_rate": 0.00018410448068433, "loss": 4.4573, "step": 692300 }, { "epoch": 1.2920016141861772, "grad_norm": 1.312732458114624, "learning_rate": 0.0001841022080245402, "loss": 4.6368, "step": 692350 }, { "epoch": 1.2920949197455558, "grad_norm": 1.3264482021331787, "learning_rate": 0.0001840999352163248, "loss": 4.45, "step": 692400 }, { "epoch": 1.2921882253049342, "grad_norm": 0.9829143285751343, "learning_rate": 0.0001840976622596878, "loss": 4.4885, "step": 692450 }, { "epoch": 1.2922815308643127, "grad_norm": 1.1931672096252441, "learning_rate": 0.00018409538915463322, "loss": 4.4189, "step": 692500 }, { "epoch": 1.2923748364236913, "grad_norm": 0.9485961198806763, "learning_rate": 0.0001840931159011651, "loss": 4.4196, "step": 692550 }, { "epoch": 1.2924681419830697, "grad_norm": 1.0094338655471802, "learning_rate": 0.0001840908424992874, "loss": 4.4463, "step": 692600 }, { "epoch": 1.292561447542448, "grad_norm": 1.2581331729888916, "learning_rate": 0.00018408856894900418, "loss": 4.6238, "step": 692650 }, { "epoch": 1.2926547531018266, "grad_norm": 1.0806974172592163, "learning_rate": 0.00018408629525031942, "loss": 4.4475, "step": 692700 }, { "epoch": 1.2927480586612052, "grad_norm": 1.1520966291427612, "learning_rate": 0.00018408402140323714, "loss": 4.4392, "step": 692750 }, { "epoch": 1.2928413642205836, "grad_norm": 1.0598182678222656, "learning_rate": 0.00018408174740776136, "loss": 4.4781, "step": 692800 }, { "epoch": 1.2929346697799622, "grad_norm": 1.0515919923782349, "learning_rate": 0.0001840794732638961, "loss": 4.4821, "step": 692850 }, { "epoch": 1.2930279753393408, "grad_norm": 1.2541158199310303, "learning_rate": 0.00018407719897164536, "loss": 4.4149, "step": 692900 }, { "epoch": 1.2931212808987191, "grad_norm": 1.29762864112854, "learning_rate": 0.00018407492453101316, "loss": 4.6765, "step": 692950 }, { "epoch": 1.2932145864580975, "grad_norm": 1.0467809438705444, "learning_rate": 0.00018407264994200348, "loss": 4.406, "step": 693000 }, { "epoch": 1.2932145864580975, "eval_loss": 4.668984413146973, "eval_runtime": 231.7815, "eval_samples_per_second": 11.252, "eval_steps_per_second": 11.252, "eval_tts_loss": 7.654490709127687, "step": 693000 }, { "epoch": 1.293307892017476, "grad_norm": 1.0571093559265137, "learning_rate": 0.0001840703752046204, "loss": 4.6538, "step": 693050 }, { "epoch": 1.2934011975768547, "grad_norm": 1.1144204139709473, "learning_rate": 0.00018406810031886792, "loss": 4.4381, "step": 693100 }, { "epoch": 1.293494503136233, "grad_norm": 1.2274638414382935, "learning_rate": 0.00018406582528475, "loss": 4.5327, "step": 693150 }, { "epoch": 1.2935878086956116, "grad_norm": 1.1898003816604614, "learning_rate": 0.0001840635501022707, "loss": 4.4332, "step": 693200 }, { "epoch": 1.2936811142549902, "grad_norm": 1.3296303749084473, "learning_rate": 0.00018406127477143402, "loss": 4.4751, "step": 693250 }, { "epoch": 1.2937744198143686, "grad_norm": 0.9184523820877075, "learning_rate": 0.000184058999292244, "loss": 4.6081, "step": 693300 }, { "epoch": 1.293867725373747, "grad_norm": 0.956450879573822, "learning_rate": 0.00018405672366470462, "loss": 4.5734, "step": 693350 }, { "epoch": 1.2939610309331255, "grad_norm": 1.1786452531814575, "learning_rate": 0.0001840544478888199, "loss": 4.4576, "step": 693400 }, { "epoch": 1.2940543364925041, "grad_norm": 1.194350242614746, "learning_rate": 0.00018405217196459388, "loss": 4.3885, "step": 693450 }, { "epoch": 1.2941476420518825, "grad_norm": 1.1313588619232178, "learning_rate": 0.00018404989589203058, "loss": 4.4969, "step": 693500 }, { "epoch": 1.294240947611261, "grad_norm": 1.0978504419326782, "learning_rate": 0.000184047619671134, "loss": 4.4256, "step": 693550 }, { "epoch": 1.2943342531706397, "grad_norm": 1.1216918230056763, "learning_rate": 0.00018404534330190817, "loss": 4.3544, "step": 693600 }, { "epoch": 1.294427558730018, "grad_norm": 1.2641496658325195, "learning_rate": 0.00018404306678435707, "loss": 4.4988, "step": 693650 }, { "epoch": 1.2945208642893964, "grad_norm": 1.3575563430786133, "learning_rate": 0.00018404079011848475, "loss": 4.5988, "step": 693700 }, { "epoch": 1.294614169848775, "grad_norm": 1.0623162984848022, "learning_rate": 0.00018403851330429523, "loss": 4.4431, "step": 693750 }, { "epoch": 1.2947074754081536, "grad_norm": 1.193487524986267, "learning_rate": 0.00018403623634179252, "loss": 4.4187, "step": 693800 }, { "epoch": 1.294800780967532, "grad_norm": 1.0469638109207153, "learning_rate": 0.00018403395923098065, "loss": 4.2971, "step": 693850 }, { "epoch": 1.2948940865269105, "grad_norm": 1.1027700901031494, "learning_rate": 0.0001840316819718636, "loss": 4.2777, "step": 693900 }, { "epoch": 1.294987392086289, "grad_norm": 1.0908170938491821, "learning_rate": 0.00018402940456444544, "loss": 4.6224, "step": 693950 }, { "epoch": 1.2950806976456675, "grad_norm": 1.121140718460083, "learning_rate": 0.00018402712700873018, "loss": 4.7133, "step": 694000 }, { "epoch": 1.2951740032050458, "grad_norm": 1.0808812379837036, "learning_rate": 0.0001840248493047218, "loss": 4.6966, "step": 694050 }, { "epoch": 1.2952673087644244, "grad_norm": 1.2994263172149658, "learning_rate": 0.00018402257145242436, "loss": 4.289, "step": 694100 }, { "epoch": 1.295360614323803, "grad_norm": 1.2827099561691284, "learning_rate": 0.00018402029345184184, "loss": 4.4986, "step": 694150 }, { "epoch": 1.2954539198831814, "grad_norm": 1.4906426668167114, "learning_rate": 0.0001840180153029783, "loss": 4.498, "step": 694200 }, { "epoch": 1.29554722544256, "grad_norm": 1.354494333267212, "learning_rate": 0.00018401573700583775, "loss": 4.7159, "step": 694250 }, { "epoch": 1.2956405310019385, "grad_norm": 1.1773203611373901, "learning_rate": 0.0001840134585604242, "loss": 4.3995, "step": 694300 }, { "epoch": 1.295733836561317, "grad_norm": 0.8671444654464722, "learning_rate": 0.00018401117996674172, "loss": 4.4673, "step": 694350 }, { "epoch": 1.2958271421206953, "grad_norm": 1.2426869869232178, "learning_rate": 0.00018400890122479425, "loss": 4.3981, "step": 694400 }, { "epoch": 1.2959204476800739, "grad_norm": 1.217600703239441, "learning_rate": 0.00018400662233458584, "loss": 4.4476, "step": 694450 }, { "epoch": 1.2960137532394524, "grad_norm": 0.9457398653030396, "learning_rate": 0.00018400434329612052, "loss": 4.3659, "step": 694500 }, { "epoch": 1.2961070587988308, "grad_norm": 1.4713127613067627, "learning_rate": 0.00018400206410940236, "loss": 4.5404, "step": 694550 }, { "epoch": 1.2962003643582094, "grad_norm": 0.9574200510978699, "learning_rate": 0.0001839997847744353, "loss": 4.6218, "step": 694600 }, { "epoch": 1.296293669917588, "grad_norm": 1.7349281311035156, "learning_rate": 0.00018399750529122342, "loss": 4.696, "step": 694650 }, { "epoch": 1.2963869754769664, "grad_norm": 0.9953287243843079, "learning_rate": 0.0001839952256597707, "loss": 4.5447, "step": 694700 }, { "epoch": 1.2964802810363447, "grad_norm": 1.3735957145690918, "learning_rate": 0.00018399294588008122, "loss": 4.3754, "step": 694750 }, { "epoch": 1.2965735865957233, "grad_norm": 0.9704420566558838, "learning_rate": 0.00018399066595215896, "loss": 4.4901, "step": 694800 }, { "epoch": 1.296666892155102, "grad_norm": 1.2315129041671753, "learning_rate": 0.00018398838587600796, "loss": 4.5406, "step": 694850 }, { "epoch": 1.2967601977144803, "grad_norm": 1.3021641969680786, "learning_rate": 0.0001839861056516322, "loss": 4.431, "step": 694900 }, { "epoch": 1.2968535032738588, "grad_norm": 0.9682461619377136, "learning_rate": 0.00018398382527903576, "loss": 4.5267, "step": 694950 }, { "epoch": 1.2969468088332374, "grad_norm": 1.0316259860992432, "learning_rate": 0.00018398154475822266, "loss": 4.4554, "step": 695000 }, { "epoch": 1.2970401143926158, "grad_norm": 1.3954741954803467, "learning_rate": 0.0001839792640891969, "loss": 4.6639, "step": 695050 }, { "epoch": 1.2971334199519942, "grad_norm": 1.2502602338790894, "learning_rate": 0.0001839769832719625, "loss": 4.4814, "step": 695100 }, { "epoch": 1.2972267255113727, "grad_norm": 1.0411535501480103, "learning_rate": 0.0001839747023065235, "loss": 4.486, "step": 695150 }, { "epoch": 1.2973200310707513, "grad_norm": 1.0347245931625366, "learning_rate": 0.00018397242119288395, "loss": 4.3502, "step": 695200 }, { "epoch": 1.2974133366301297, "grad_norm": 1.1431283950805664, "learning_rate": 0.00018397013993104786, "loss": 4.4183, "step": 695250 }, { "epoch": 1.2975066421895083, "grad_norm": 1.309677243232727, "learning_rate": 0.00018396785852101925, "loss": 4.5232, "step": 695300 }, { "epoch": 1.2975999477488869, "grad_norm": 1.129940390586853, "learning_rate": 0.0001839655769628021, "loss": 4.5776, "step": 695350 }, { "epoch": 1.2976932533082652, "grad_norm": 1.0474542379379272, "learning_rate": 0.0001839632952564005, "loss": 4.3542, "step": 695400 }, { "epoch": 1.2977865588676436, "grad_norm": 0.7993698120117188, "learning_rate": 0.00018396101340181848, "loss": 4.5316, "step": 695450 }, { "epoch": 1.2978798644270222, "grad_norm": 1.3087067604064941, "learning_rate": 0.00018395873139906, "loss": 4.6229, "step": 695500 }, { "epoch": 1.2979731699864008, "grad_norm": 1.4193729162216187, "learning_rate": 0.0001839564492481292, "loss": 4.4114, "step": 695550 }, { "epoch": 1.2980664755457791, "grad_norm": 1.2514451742172241, "learning_rate": 0.00018395416694903, "loss": 4.5626, "step": 695600 }, { "epoch": 1.2981597811051577, "grad_norm": 1.107145071029663, "learning_rate": 0.00018395188450176647, "loss": 4.6267, "step": 695650 }, { "epoch": 1.2982530866645363, "grad_norm": 1.1795315742492676, "learning_rate": 0.0001839496019063426, "loss": 4.4446, "step": 695700 }, { "epoch": 1.2983463922239147, "grad_norm": 1.2412095069885254, "learning_rate": 0.00018394731916276252, "loss": 4.4615, "step": 695750 }, { "epoch": 1.298439697783293, "grad_norm": 1.4747081995010376, "learning_rate": 0.00018394503627103015, "loss": 4.5203, "step": 695800 }, { "epoch": 1.2985330033426716, "grad_norm": 1.0873076915740967, "learning_rate": 0.00018394275323114955, "loss": 4.4776, "step": 695850 }, { "epoch": 1.2986263089020502, "grad_norm": 1.1709339618682861, "learning_rate": 0.0001839404700431248, "loss": 4.4772, "step": 695900 }, { "epoch": 1.2987196144614286, "grad_norm": 1.1638633012771606, "learning_rate": 0.00018393818670695987, "loss": 4.372, "step": 695950 }, { "epoch": 1.2988129200208072, "grad_norm": 1.3940825462341309, "learning_rate": 0.0001839359032226588, "loss": 4.5602, "step": 696000 }, { "epoch": 1.2988129200208072, "eval_loss": 4.6858038902282715, "eval_runtime": 229.1169, "eval_samples_per_second": 11.383, "eval_steps_per_second": 11.383, "eval_tts_loss": 7.62773142548198, "step": 696000 }, { "epoch": 1.2989062255801858, "grad_norm": 0.9808474779129028, "learning_rate": 0.00018393361959022567, "loss": 4.6445, "step": 696050 }, { "epoch": 1.2989995311395641, "grad_norm": 0.9523788690567017, "learning_rate": 0.00018393133580966442, "loss": 4.6166, "step": 696100 }, { "epoch": 1.2990928366989425, "grad_norm": 1.3033477067947388, "learning_rate": 0.0001839290518809792, "loss": 4.4118, "step": 696150 }, { "epoch": 1.299186142258321, "grad_norm": 1.0133358240127563, "learning_rate": 0.0001839267678041739, "loss": 4.6656, "step": 696200 }, { "epoch": 1.2992794478176997, "grad_norm": 1.4119865894317627, "learning_rate": 0.00018392448357925262, "loss": 4.6032, "step": 696250 }, { "epoch": 1.299372753377078, "grad_norm": 1.3954150676727295, "learning_rate": 0.00018392219920621944, "loss": 4.4086, "step": 696300 }, { "epoch": 1.2994660589364566, "grad_norm": 1.183457851409912, "learning_rate": 0.00018391991468507836, "loss": 4.6114, "step": 696350 }, { "epoch": 1.2995593644958352, "grad_norm": 1.1394522190093994, "learning_rate": 0.00018391763001583333, "loss": 4.4969, "step": 696400 }, { "epoch": 1.2996526700552136, "grad_norm": 1.2038110494613647, "learning_rate": 0.00018391534519848846, "loss": 4.5134, "step": 696450 }, { "epoch": 1.299745975614592, "grad_norm": 1.3777105808258057, "learning_rate": 0.00018391306023304778, "loss": 4.4386, "step": 696500 }, { "epoch": 1.2998392811739705, "grad_norm": 1.0079373121261597, "learning_rate": 0.00018391077511951536, "loss": 4.572, "step": 696550 }, { "epoch": 1.299932586733349, "grad_norm": 0.9195224046707153, "learning_rate": 0.00018390848985789515, "loss": 4.6629, "step": 696600 }, { "epoch": 1.3000258922927275, "grad_norm": 0.6199584007263184, "learning_rate": 0.00018390620444819122, "loss": 4.4455, "step": 696650 }, { "epoch": 1.300119197852106, "grad_norm": 1.1942825317382812, "learning_rate": 0.0001839039188904076, "loss": 4.4816, "step": 696700 }, { "epoch": 1.3002125034114846, "grad_norm": 1.3428884744644165, "learning_rate": 0.00018390163318454832, "loss": 4.6392, "step": 696750 }, { "epoch": 1.300305808970863, "grad_norm": 1.0196369886398315, "learning_rate": 0.00018389934733061746, "loss": 4.5901, "step": 696800 }, { "epoch": 1.3003991145302414, "grad_norm": 1.068696141242981, "learning_rate": 0.00018389706132861897, "loss": 4.5304, "step": 696850 }, { "epoch": 1.30049242008962, "grad_norm": 1.0438507795333862, "learning_rate": 0.00018389477517855696, "loss": 4.5761, "step": 696900 }, { "epoch": 1.3005857256489985, "grad_norm": 1.3663272857666016, "learning_rate": 0.0001838924888804354, "loss": 4.2944, "step": 696950 }, { "epoch": 1.300679031208377, "grad_norm": 1.288014531135559, "learning_rate": 0.00018389020243425838, "loss": 4.4282, "step": 697000 }, { "epoch": 1.3007723367677555, "grad_norm": 1.201066493988037, "learning_rate": 0.0001838879158400299, "loss": 4.6315, "step": 697050 }, { "epoch": 1.300865642327134, "grad_norm": 1.3677390813827515, "learning_rate": 0.000183885629097754, "loss": 4.7349, "step": 697100 }, { "epoch": 1.3009589478865125, "grad_norm": 1.1553943157196045, "learning_rate": 0.00018388334220743478, "loss": 4.4573, "step": 697150 }, { "epoch": 1.3010522534458908, "grad_norm": 1.1024162769317627, "learning_rate": 0.00018388105516907617, "loss": 4.2347, "step": 697200 }, { "epoch": 1.3011455590052694, "grad_norm": 0.8770056366920471, "learning_rate": 0.00018387876798268228, "loss": 4.322, "step": 697250 }, { "epoch": 1.301238864564648, "grad_norm": 1.1204779148101807, "learning_rate": 0.00018387648064825714, "loss": 4.4999, "step": 697300 }, { "epoch": 1.3013321701240264, "grad_norm": 1.3154140710830688, "learning_rate": 0.00018387419316580474, "loss": 4.3978, "step": 697350 }, { "epoch": 1.301425475683405, "grad_norm": 1.216508388519287, "learning_rate": 0.00018387190553532912, "loss": 4.4765, "step": 697400 }, { "epoch": 1.3015187812427835, "grad_norm": 1.0255389213562012, "learning_rate": 0.0001838696177568344, "loss": 4.1716, "step": 697450 }, { "epoch": 1.301612086802162, "grad_norm": 1.0601043701171875, "learning_rate": 0.00018386732983032453, "loss": 4.7382, "step": 697500 }, { "epoch": 1.3017053923615403, "grad_norm": 0.8397194147109985, "learning_rate": 0.00018386504175580358, "loss": 4.459, "step": 697550 }, { "epoch": 1.3017986979209188, "grad_norm": 1.2517287731170654, "learning_rate": 0.00018386275353327562, "loss": 4.4934, "step": 697600 }, { "epoch": 1.3018920034802974, "grad_norm": 1.257463812828064, "learning_rate": 0.00018386046516274463, "loss": 4.5153, "step": 697650 }, { "epoch": 1.3019853090396758, "grad_norm": 1.1605104207992554, "learning_rate": 0.00018385817664421468, "loss": 4.645, "step": 697700 }, { "epoch": 1.3020786145990544, "grad_norm": 1.5905795097351074, "learning_rate": 0.0001838558879776898, "loss": 4.5452, "step": 697750 }, { "epoch": 1.302171920158433, "grad_norm": 1.121583104133606, "learning_rate": 0.000183853599163174, "loss": 4.4333, "step": 697800 }, { "epoch": 1.3022652257178113, "grad_norm": 1.1360876560211182, "learning_rate": 0.00018385131020067142, "loss": 4.5972, "step": 697850 }, { "epoch": 1.3023585312771897, "grad_norm": 1.0242228507995605, "learning_rate": 0.00018384902109018597, "loss": 4.4489, "step": 697900 }, { "epoch": 1.3024518368365683, "grad_norm": 1.5653820037841797, "learning_rate": 0.0001838467318317218, "loss": 4.4414, "step": 697950 }, { "epoch": 1.3025451423959469, "grad_norm": 0.9875494241714478, "learning_rate": 0.0001838444424252829, "loss": 4.5838, "step": 698000 }, { "epoch": 1.3026384479553252, "grad_norm": 1.098537802696228, "learning_rate": 0.00018384215287087327, "loss": 4.4558, "step": 698050 }, { "epoch": 1.3027317535147038, "grad_norm": 1.0364642143249512, "learning_rate": 0.000183839863168497, "loss": 4.4341, "step": 698100 }, { "epoch": 1.3028250590740824, "grad_norm": 1.0796170234680176, "learning_rate": 0.00018383757331815815, "loss": 4.4401, "step": 698150 }, { "epoch": 1.3029183646334608, "grad_norm": 1.3138214349746704, "learning_rate": 0.00018383528331986074, "loss": 4.5168, "step": 698200 }, { "epoch": 1.3030116701928391, "grad_norm": 1.3367422819137573, "learning_rate": 0.00018383299317360877, "loss": 4.5079, "step": 698250 }, { "epoch": 1.3031049757522177, "grad_norm": 1.2113454341888428, "learning_rate": 0.00018383070287940635, "loss": 4.5929, "step": 698300 }, { "epoch": 1.3031982813115963, "grad_norm": 1.1787830591201782, "learning_rate": 0.00018382841243725747, "loss": 4.6495, "step": 698350 }, { "epoch": 1.3032915868709747, "grad_norm": 1.3863154649734497, "learning_rate": 0.0001838261218471662, "loss": 4.342, "step": 698400 }, { "epoch": 1.3033848924303533, "grad_norm": 0.9779516458511353, "learning_rate": 0.00018382383110913657, "loss": 4.5543, "step": 698450 }, { "epoch": 1.3034781979897319, "grad_norm": 1.1980119943618774, "learning_rate": 0.00018382154022317265, "loss": 4.5686, "step": 698500 }, { "epoch": 1.3035715035491102, "grad_norm": 1.3223329782485962, "learning_rate": 0.00018381924918927842, "loss": 4.4138, "step": 698550 }, { "epoch": 1.3036648091084886, "grad_norm": 1.1538584232330322, "learning_rate": 0.000183816958007458, "loss": 4.5258, "step": 698600 }, { "epoch": 1.3037581146678672, "grad_norm": 0.9887345433235168, "learning_rate": 0.00018381466667771536, "loss": 4.592, "step": 698650 }, { "epoch": 1.3038514202272458, "grad_norm": 1.1716288328170776, "learning_rate": 0.00018381237520005464, "loss": 4.6159, "step": 698700 }, { "epoch": 1.3039447257866241, "grad_norm": 1.5807160139083862, "learning_rate": 0.00018381008357447976, "loss": 4.5569, "step": 698750 }, { "epoch": 1.3040380313460027, "grad_norm": 1.2708077430725098, "learning_rate": 0.00018380779180099488, "loss": 4.4394, "step": 698800 }, { "epoch": 1.3041313369053813, "grad_norm": 1.083714485168457, "learning_rate": 0.00018380549987960396, "loss": 4.5864, "step": 698850 }, { "epoch": 1.3042246424647597, "grad_norm": 1.442801833152771, "learning_rate": 0.00018380320781031108, "loss": 4.4571, "step": 698900 }, { "epoch": 1.304317948024138, "grad_norm": 1.2007122039794922, "learning_rate": 0.0001838009155931203, "loss": 4.6546, "step": 698950 }, { "epoch": 1.3044112535835166, "grad_norm": 1.1808751821517944, "learning_rate": 0.00018379862322803566, "loss": 4.3456, "step": 699000 }, { "epoch": 1.3044112535835166, "eval_loss": 4.666496276855469, "eval_runtime": 228.7059, "eval_samples_per_second": 11.403, "eval_steps_per_second": 11.403, "eval_tts_loss": 7.691604443522959, "step": 699000 }, { "epoch": 1.3045045591428952, "grad_norm": 1.1014106273651123, "learning_rate": 0.00018379633071506114, "loss": 4.4503, "step": 699050 }, { "epoch": 1.3045978647022736, "grad_norm": 1.2599925994873047, "learning_rate": 0.0001837940380542009, "loss": 4.2471, "step": 699100 }, { "epoch": 1.3046911702616522, "grad_norm": 1.2553775310516357, "learning_rate": 0.0001837917452454589, "loss": 4.3038, "step": 699150 }, { "epoch": 1.3047844758210307, "grad_norm": 0.9492983818054199, "learning_rate": 0.00018378945228883923, "loss": 4.5576, "step": 699200 }, { "epoch": 1.304877781380409, "grad_norm": 1.4133830070495605, "learning_rate": 0.00018378715918434592, "loss": 4.6098, "step": 699250 }, { "epoch": 1.3049710869397875, "grad_norm": 1.3064991235733032, "learning_rate": 0.000183784865931983, "loss": 4.3122, "step": 699300 }, { "epoch": 1.305064392499166, "grad_norm": 1.2116867303848267, "learning_rate": 0.00018378257253175452, "loss": 4.6236, "step": 699350 }, { "epoch": 1.3051576980585446, "grad_norm": 1.0481200218200684, "learning_rate": 0.0001837802789836646, "loss": 4.4912, "step": 699400 }, { "epoch": 1.305251003617923, "grad_norm": 1.2031511068344116, "learning_rate": 0.00018377798528771718, "loss": 4.5028, "step": 699450 }, { "epoch": 1.3053443091773016, "grad_norm": 1.2981420755386353, "learning_rate": 0.00018377569144391635, "loss": 4.6183, "step": 699500 }, { "epoch": 1.3054376147366802, "grad_norm": 1.6936641931533813, "learning_rate": 0.00018377339745226617, "loss": 4.516, "step": 699550 }, { "epoch": 1.3055309202960586, "grad_norm": 1.0564312934875488, "learning_rate": 0.0001837711033127707, "loss": 4.5881, "step": 699600 }, { "epoch": 1.305624225855437, "grad_norm": 1.0515471696853638, "learning_rate": 0.00018376880902543397, "loss": 4.5364, "step": 699650 }, { "epoch": 1.3057175314148155, "grad_norm": 1.2713841199874878, "learning_rate": 0.00018376651459026005, "loss": 4.7295, "step": 699700 }, { "epoch": 1.305810836974194, "grad_norm": 1.2217315435409546, "learning_rate": 0.00018376422000725295, "loss": 4.7183, "step": 699750 }, { "epoch": 1.3059041425335725, "grad_norm": 1.374277949333191, "learning_rate": 0.00018376192527641677, "loss": 4.3464, "step": 699800 }, { "epoch": 1.305997448092951, "grad_norm": 1.2962636947631836, "learning_rate": 0.00018375963039775548, "loss": 4.3889, "step": 699850 }, { "epoch": 1.3060907536523296, "grad_norm": 1.2797019481658936, "learning_rate": 0.0001837573353712732, "loss": 4.5152, "step": 699900 }, { "epoch": 1.306184059211708, "grad_norm": 1.1680121421813965, "learning_rate": 0.00018375504019697397, "loss": 4.4337, "step": 699950 }, { "epoch": 1.3062773647710864, "grad_norm": 1.151685118675232, "learning_rate": 0.00018375274487486182, "loss": 4.5078, "step": 700000 }, { "epoch": 1.306370670330465, "grad_norm": 1.1212929487228394, "learning_rate": 0.00018375044940494083, "loss": 4.4509, "step": 700050 }, { "epoch": 1.3064639758898435, "grad_norm": 1.2027311325073242, "learning_rate": 0.000183748153787215, "loss": 4.6338, "step": 700100 }, { "epoch": 1.306557281449222, "grad_norm": 0.8819425106048584, "learning_rate": 0.00018374585802168845, "loss": 4.4602, "step": 700150 }, { "epoch": 1.3066505870086005, "grad_norm": 1.1399351358413696, "learning_rate": 0.0001837435621083652, "loss": 4.6214, "step": 700200 }, { "epoch": 1.3067438925679788, "grad_norm": 1.5317254066467285, "learning_rate": 0.00018374126604724924, "loss": 4.5221, "step": 700250 }, { "epoch": 1.3068371981273574, "grad_norm": 1.0293326377868652, "learning_rate": 0.00018373896983834472, "loss": 4.4302, "step": 700300 }, { "epoch": 1.3069305036867358, "grad_norm": 1.3144042491912842, "learning_rate": 0.00018373667348165566, "loss": 4.5176, "step": 700350 }, { "epoch": 1.3070238092461144, "grad_norm": 1.2248711585998535, "learning_rate": 0.0001837343769771861, "loss": 4.4373, "step": 700400 }, { "epoch": 1.307117114805493, "grad_norm": 1.5845885276794434, "learning_rate": 0.0001837320803249401, "loss": 4.3219, "step": 700450 }, { "epoch": 1.3072104203648713, "grad_norm": 0.9750577211380005, "learning_rate": 0.0001837297835249217, "loss": 4.4792, "step": 700500 }, { "epoch": 1.30730372592425, "grad_norm": 1.375092625617981, "learning_rate": 0.00018372748657713496, "loss": 4.6274, "step": 700550 }, { "epoch": 1.3073970314836283, "grad_norm": 0.8594903945922852, "learning_rate": 0.00018372518948158394, "loss": 4.3388, "step": 700600 }, { "epoch": 1.3074903370430069, "grad_norm": 0.7430895566940308, "learning_rate": 0.00018372289223827273, "loss": 4.4169, "step": 700650 }, { "epoch": 1.3075836426023852, "grad_norm": 1.320732593536377, "learning_rate": 0.0001837205948472053, "loss": 4.7113, "step": 700700 }, { "epoch": 1.3076769481617638, "grad_norm": 1.300972819328308, "learning_rate": 0.0001837182973083858, "loss": 4.5001, "step": 700750 }, { "epoch": 1.3077702537211424, "grad_norm": 1.059051275253296, "learning_rate": 0.0001837159996218182, "loss": 4.5889, "step": 700800 }, { "epoch": 1.3078635592805208, "grad_norm": 1.491579294204712, "learning_rate": 0.0001837137017875066, "loss": 4.3665, "step": 700850 }, { "epoch": 1.3079568648398994, "grad_norm": 1.1959495544433594, "learning_rate": 0.00018371140380545503, "loss": 4.4831, "step": 700900 }, { "epoch": 1.3080501703992777, "grad_norm": 1.254097580909729, "learning_rate": 0.0001837091056756676, "loss": 4.5284, "step": 700950 }, { "epoch": 1.3081434759586563, "grad_norm": 0.8977311253547668, "learning_rate": 0.00018370680739814826, "loss": 4.3261, "step": 701000 }, { "epoch": 1.3082367815180347, "grad_norm": 1.4227238893508911, "learning_rate": 0.00018370450897290118, "loss": 4.6062, "step": 701050 }, { "epoch": 1.3083300870774133, "grad_norm": 0.9942972660064697, "learning_rate": 0.00018370221039993036, "loss": 4.2714, "step": 701100 }, { "epoch": 1.3084233926367919, "grad_norm": 1.4382866621017456, "learning_rate": 0.0001836999116792399, "loss": 4.5691, "step": 701150 }, { "epoch": 1.3085166981961702, "grad_norm": 0.7618278861045837, "learning_rate": 0.00018369761281083376, "loss": 4.2617, "step": 701200 }, { "epoch": 1.3086100037555488, "grad_norm": 1.1584378480911255, "learning_rate": 0.00018369531379471612, "loss": 4.6778, "step": 701250 }, { "epoch": 1.3087033093149272, "grad_norm": 0.9558416604995728, "learning_rate": 0.00018369301463089097, "loss": 4.5068, "step": 701300 }, { "epoch": 1.3087966148743058, "grad_norm": 1.3652429580688477, "learning_rate": 0.00018369071531936236, "loss": 4.447, "step": 701350 }, { "epoch": 1.3088899204336841, "grad_norm": 1.3041274547576904, "learning_rate": 0.00018368841586013435, "loss": 4.613, "step": 701400 }, { "epoch": 1.3089832259930627, "grad_norm": 1.297996997833252, "learning_rate": 0.00018368611625321104, "loss": 4.3868, "step": 701450 }, { "epoch": 1.3090765315524413, "grad_norm": 0.9443566799163818, "learning_rate": 0.00018368381649859643, "loss": 4.4373, "step": 701500 }, { "epoch": 1.3091698371118197, "grad_norm": 1.059092402458191, "learning_rate": 0.00018368151659629462, "loss": 4.2855, "step": 701550 }, { "epoch": 1.3092631426711983, "grad_norm": 1.1253266334533691, "learning_rate": 0.00018367921654630965, "loss": 4.4992, "step": 701600 }, { "epoch": 1.3093564482305766, "grad_norm": 1.1886119842529297, "learning_rate": 0.0001836769163486456, "loss": 4.4922, "step": 701650 }, { "epoch": 1.3094497537899552, "grad_norm": 1.295644760131836, "learning_rate": 0.00018367461600330654, "loss": 4.4512, "step": 701700 }, { "epoch": 1.3095430593493336, "grad_norm": 1.2292804718017578, "learning_rate": 0.00018367231551029645, "loss": 4.4446, "step": 701750 }, { "epoch": 1.3096363649087122, "grad_norm": 1.3638666868209839, "learning_rate": 0.0001836700148696195, "loss": 4.5246, "step": 701800 }, { "epoch": 1.3097296704680907, "grad_norm": 1.1936604976654053, "learning_rate": 0.00018366771408127966, "loss": 4.429, "step": 701850 }, { "epoch": 1.309822976027469, "grad_norm": 1.2178692817687988, "learning_rate": 0.00018366541314528102, "loss": 4.5548, "step": 701900 }, { "epoch": 1.3099162815868477, "grad_norm": 1.2458739280700684, "learning_rate": 0.00018366311206162767, "loss": 4.795, "step": 701950 }, { "epoch": 1.310009587146226, "grad_norm": 1.393414855003357, "learning_rate": 0.00018366081083032366, "loss": 4.5446, "step": 702000 }, { "epoch": 1.310009587146226, "eval_loss": 4.669854164123535, "eval_runtime": 230.4522, "eval_samples_per_second": 11.317, "eval_steps_per_second": 11.317, "eval_tts_loss": 7.62797422892774, "step": 702000 }, { "epoch": 1.3101028927056046, "grad_norm": 1.1992928981781006, "learning_rate": 0.00018365850945137302, "loss": 4.402, "step": 702050 }, { "epoch": 1.310196198264983, "grad_norm": 1.1770150661468506, "learning_rate": 0.00018365620792477983, "loss": 4.5771, "step": 702100 }, { "epoch": 1.3102895038243616, "grad_norm": 1.009460210800171, "learning_rate": 0.00018365390625054814, "loss": 4.5697, "step": 702150 }, { "epoch": 1.3103828093837402, "grad_norm": 1.1731947660446167, "learning_rate": 0.00018365160442868205, "loss": 4.6491, "step": 702200 }, { "epoch": 1.3104761149431186, "grad_norm": 1.2416298389434814, "learning_rate": 0.00018364930245918558, "loss": 4.5035, "step": 702250 }, { "epoch": 1.3105694205024971, "grad_norm": 1.1008800268173218, "learning_rate": 0.0001836470003420628, "loss": 4.6072, "step": 702300 }, { "epoch": 1.3106627260618755, "grad_norm": 1.0923075675964355, "learning_rate": 0.00018364469807731783, "loss": 4.3833, "step": 702350 }, { "epoch": 1.310756031621254, "grad_norm": 0.8377133011817932, "learning_rate": 0.00018364239566495466, "loss": 4.4105, "step": 702400 }, { "epoch": 1.3108493371806325, "grad_norm": 1.677588939666748, "learning_rate": 0.00018364009310497737, "loss": 4.5481, "step": 702450 }, { "epoch": 1.310942642740011, "grad_norm": 1.344097375869751, "learning_rate": 0.00018363779039739004, "loss": 4.4274, "step": 702500 }, { "epoch": 1.3110359482993896, "grad_norm": 1.249723196029663, "learning_rate": 0.00018363548754219672, "loss": 4.5977, "step": 702550 }, { "epoch": 1.311129253858768, "grad_norm": 1.3707104921340942, "learning_rate": 0.00018363318453940147, "loss": 4.3924, "step": 702600 }, { "epoch": 1.3112225594181466, "grad_norm": 1.2216304540634155, "learning_rate": 0.00018363088138900836, "loss": 4.5959, "step": 702650 }, { "epoch": 1.311315864977525, "grad_norm": 1.1417005062103271, "learning_rate": 0.0001836285780910215, "loss": 4.5589, "step": 702700 }, { "epoch": 1.3114091705369035, "grad_norm": 1.4705288410186768, "learning_rate": 0.00018362627464544492, "loss": 4.4673, "step": 702750 }, { "epoch": 1.311502476096282, "grad_norm": 0.7284272313117981, "learning_rate": 0.00018362397105228264, "loss": 4.4996, "step": 702800 }, { "epoch": 1.3115957816556605, "grad_norm": 1.2868388891220093, "learning_rate": 0.00018362166731153879, "loss": 4.29, "step": 702850 }, { "epoch": 1.311689087215039, "grad_norm": 1.5436631441116333, "learning_rate": 0.00018361936342321737, "loss": 4.4185, "step": 702900 }, { "epoch": 1.3117823927744174, "grad_norm": 1.3750163316726685, "learning_rate": 0.0001836170593873225, "loss": 4.4099, "step": 702950 }, { "epoch": 1.311875698333796, "grad_norm": 1.1987541913986206, "learning_rate": 0.00018361475520385826, "loss": 4.3919, "step": 703000 }, { "epoch": 1.3119690038931744, "grad_norm": 1.429570198059082, "learning_rate": 0.0001836124508728287, "loss": 4.5209, "step": 703050 }, { "epoch": 1.312062309452553, "grad_norm": 1.1087148189544678, "learning_rate": 0.00018361014639423784, "loss": 4.3595, "step": 703100 }, { "epoch": 1.3121556150119313, "grad_norm": 0.892978310585022, "learning_rate": 0.0001836078417680898, "loss": 4.5552, "step": 703150 }, { "epoch": 1.31224892057131, "grad_norm": 1.3177238702774048, "learning_rate": 0.00018360553699438861, "loss": 4.5958, "step": 703200 }, { "epoch": 1.3123422261306885, "grad_norm": 1.3796517848968506, "learning_rate": 0.0001836032320731384, "loss": 4.3505, "step": 703250 }, { "epoch": 1.3124355316900669, "grad_norm": 1.4100067615509033, "learning_rate": 0.00018360092700434318, "loss": 4.3201, "step": 703300 }, { "epoch": 1.3125288372494455, "grad_norm": 0.9839244484901428, "learning_rate": 0.00018359862178800703, "loss": 4.3682, "step": 703350 }, { "epoch": 1.3126221428088238, "grad_norm": 0.9262556433677673, "learning_rate": 0.000183596316424134, "loss": 4.595, "step": 703400 }, { "epoch": 1.3127154483682024, "grad_norm": 1.1527810096740723, "learning_rate": 0.0001835940109127282, "loss": 4.6175, "step": 703450 }, { "epoch": 1.3128087539275808, "grad_norm": 1.0052317380905151, "learning_rate": 0.0001835917052537937, "loss": 4.7002, "step": 703500 }, { "epoch": 1.3129020594869594, "grad_norm": 1.3903570175170898, "learning_rate": 0.0001835893994473345, "loss": 4.4687, "step": 703550 }, { "epoch": 1.312995365046338, "grad_norm": 0.980562686920166, "learning_rate": 0.00018358709349335474, "loss": 4.3578, "step": 703600 }, { "epoch": 1.3130886706057163, "grad_norm": 1.1490875482559204, "learning_rate": 0.00018358478739185848, "loss": 4.4042, "step": 703650 }, { "epoch": 1.313181976165095, "grad_norm": 1.3070005178451538, "learning_rate": 0.00018358248114284977, "loss": 4.7038, "step": 703700 }, { "epoch": 1.3132752817244733, "grad_norm": 1.1509292125701904, "learning_rate": 0.0001835801747463327, "loss": 4.5231, "step": 703750 }, { "epoch": 1.3133685872838519, "grad_norm": 1.2558419704437256, "learning_rate": 0.00018357786820231132, "loss": 4.3749, "step": 703800 }, { "epoch": 1.3134618928432302, "grad_norm": 1.274100422859192, "learning_rate": 0.0001835755615107897, "loss": 4.6642, "step": 703850 }, { "epoch": 1.3135551984026088, "grad_norm": 1.3975058794021606, "learning_rate": 0.0001835732546717719, "loss": 4.3798, "step": 703900 }, { "epoch": 1.3136485039619874, "grad_norm": 1.3530360460281372, "learning_rate": 0.00018357094768526206, "loss": 4.4188, "step": 703950 }, { "epoch": 1.3137418095213658, "grad_norm": 1.2386243343353271, "learning_rate": 0.00018356864055126416, "loss": 4.42, "step": 704000 }, { "epoch": 1.3138351150807444, "grad_norm": 1.0247175693511963, "learning_rate": 0.00018356633326978235, "loss": 4.2713, "step": 704050 }, { "epoch": 1.3139284206401227, "grad_norm": 1.2610719203948975, "learning_rate": 0.00018356402584082063, "loss": 4.6014, "step": 704100 }, { "epoch": 1.3140217261995013, "grad_norm": 1.75786292552948, "learning_rate": 0.00018356171826438314, "loss": 4.3775, "step": 704150 }, { "epoch": 1.3141150317588797, "grad_norm": 1.5015391111373901, "learning_rate": 0.00018355941054047387, "loss": 4.5667, "step": 704200 }, { "epoch": 1.3142083373182583, "grad_norm": 1.4037730693817139, "learning_rate": 0.000183557102669097, "loss": 4.4112, "step": 704250 }, { "epoch": 1.3143016428776368, "grad_norm": 1.1276054382324219, "learning_rate": 0.00018355479465025654, "loss": 4.3903, "step": 704300 }, { "epoch": 1.3143949484370152, "grad_norm": 1.2118586301803589, "learning_rate": 0.00018355248648395652, "loss": 4.4551, "step": 704350 }, { "epoch": 1.3144882539963938, "grad_norm": 1.3275623321533203, "learning_rate": 0.0001835501781702011, "loss": 4.5742, "step": 704400 }, { "epoch": 1.3145815595557722, "grad_norm": 1.19938325881958, "learning_rate": 0.00018354786970899433, "loss": 4.4523, "step": 704450 }, { "epoch": 1.3146748651151507, "grad_norm": 1.149420142173767, "learning_rate": 0.00018354556110034025, "loss": 4.6101, "step": 704500 }, { "epoch": 1.3147681706745291, "grad_norm": 1.3785301446914673, "learning_rate": 0.00018354325234424296, "loss": 4.4836, "step": 704550 }, { "epoch": 1.3148614762339077, "grad_norm": 1.1128681898117065, "learning_rate": 0.00018354094344070655, "loss": 4.455, "step": 704600 }, { "epoch": 1.3149547817932863, "grad_norm": 1.229477882385254, "learning_rate": 0.00018353863438973504, "loss": 4.4371, "step": 704650 }, { "epoch": 1.3150480873526647, "grad_norm": 1.0847150087356567, "learning_rate": 0.00018353632519133256, "loss": 4.5149, "step": 704700 }, { "epoch": 1.3151413929120432, "grad_norm": 1.5701794624328613, "learning_rate": 0.00018353401584550314, "loss": 4.3967, "step": 704750 }, { "epoch": 1.3152346984714216, "grad_norm": 1.192399501800537, "learning_rate": 0.0001835317063522509, "loss": 4.7055, "step": 704800 }, { "epoch": 1.3153280040308002, "grad_norm": 1.1114749908447266, "learning_rate": 0.00018352939671157988, "loss": 4.5112, "step": 704850 }, { "epoch": 1.3154213095901786, "grad_norm": 1.5972880125045776, "learning_rate": 0.00018352708692349418, "loss": 4.5551, "step": 704900 }, { "epoch": 1.3155146151495571, "grad_norm": 1.1549891233444214, "learning_rate": 0.00018352477698799788, "loss": 4.488, "step": 704950 }, { "epoch": 1.3156079207089357, "grad_norm": 1.094204306602478, "learning_rate": 0.00018352246690509505, "loss": 4.683, "step": 705000 }, { "epoch": 1.3156079207089357, "eval_loss": 4.672201633453369, "eval_runtime": 229.2311, "eval_samples_per_second": 11.377, "eval_steps_per_second": 11.377, "eval_tts_loss": 7.576802273299488, "step": 705000 }, { "epoch": 1.315701226268314, "grad_norm": 0.9960780739784241, "learning_rate": 0.0001835201566747897, "loss": 4.4931, "step": 705050 }, { "epoch": 1.3157945318276927, "grad_norm": 1.3106698989868164, "learning_rate": 0.00018351784629708605, "loss": 4.6493, "step": 705100 }, { "epoch": 1.315887837387071, "grad_norm": 1.1256202459335327, "learning_rate": 0.0001835155357719881, "loss": 4.344, "step": 705150 }, { "epoch": 1.3159811429464496, "grad_norm": 1.0360794067382812, "learning_rate": 0.00018351322509949986, "loss": 4.484, "step": 705200 }, { "epoch": 1.316074448505828, "grad_norm": 1.4605445861816406, "learning_rate": 0.0001835109142796255, "loss": 4.3801, "step": 705250 }, { "epoch": 1.3161677540652066, "grad_norm": 1.0742769241333008, "learning_rate": 0.0001835086033123691, "loss": 4.5787, "step": 705300 }, { "epoch": 1.3162610596245852, "grad_norm": 1.4498381614685059, "learning_rate": 0.00018350629219773468, "loss": 4.4376, "step": 705350 }, { "epoch": 1.3163543651839635, "grad_norm": 1.5302695035934448, "learning_rate": 0.00018350398093572635, "loss": 4.5585, "step": 705400 }, { "epoch": 1.3164476707433421, "grad_norm": 0.9497649669647217, "learning_rate": 0.00018350166952634818, "loss": 4.4878, "step": 705450 }, { "epoch": 1.3165409763027205, "grad_norm": 0.9207112193107605, "learning_rate": 0.00018349935796960426, "loss": 4.5145, "step": 705500 }, { "epoch": 1.316634281862099, "grad_norm": 0.9004650115966797, "learning_rate": 0.00018349704626549867, "loss": 4.6239, "step": 705550 }, { "epoch": 1.3167275874214774, "grad_norm": 1.035599946975708, "learning_rate": 0.00018349473441403547, "loss": 4.6526, "step": 705600 }, { "epoch": 1.316820892980856, "grad_norm": 1.1780034303665161, "learning_rate": 0.00018349242241521877, "loss": 4.4895, "step": 705650 }, { "epoch": 1.3169141985402346, "grad_norm": 0.9396145343780518, "learning_rate": 0.00018349011026905262, "loss": 4.5346, "step": 705700 }, { "epoch": 1.317007504099613, "grad_norm": 1.2970777750015259, "learning_rate": 0.00018348779797554117, "loss": 4.7358, "step": 705750 }, { "epoch": 1.3171008096589916, "grad_norm": 0.9447508454322815, "learning_rate": 0.00018348548553468838, "loss": 4.3624, "step": 705800 }, { "epoch": 1.31719411521837, "grad_norm": 1.1495122909545898, "learning_rate": 0.00018348317294649845, "loss": 4.5335, "step": 705850 }, { "epoch": 1.3172874207777485, "grad_norm": 1.2257719039916992, "learning_rate": 0.00018348086021097537, "loss": 4.3429, "step": 705900 }, { "epoch": 1.3173807263371269, "grad_norm": 1.407866358757019, "learning_rate": 0.00018347854732812327, "loss": 4.6733, "step": 705950 }, { "epoch": 1.3174740318965055, "grad_norm": 1.209058403968811, "learning_rate": 0.00018347623429794621, "loss": 4.3762, "step": 706000 }, { "epoch": 1.317567337455884, "grad_norm": 1.1408312320709229, "learning_rate": 0.00018347392112044833, "loss": 4.513, "step": 706050 }, { "epoch": 1.3176606430152624, "grad_norm": 1.4903448820114136, "learning_rate": 0.00018347160779563364, "loss": 4.4844, "step": 706100 }, { "epoch": 1.317753948574641, "grad_norm": 1.1722826957702637, "learning_rate": 0.00018346929432350623, "loss": 4.6587, "step": 706150 }, { "epoch": 1.3178472541340194, "grad_norm": 1.552720069885254, "learning_rate": 0.00018346698070407022, "loss": 4.391, "step": 706200 }, { "epoch": 1.317940559693398, "grad_norm": 0.8814159035682678, "learning_rate": 0.00018346466693732965, "loss": 4.4371, "step": 706250 }, { "epoch": 1.3180338652527763, "grad_norm": 1.2607420682907104, "learning_rate": 0.00018346235302328867, "loss": 4.5765, "step": 706300 }, { "epoch": 1.318127170812155, "grad_norm": 1.2193397283554077, "learning_rate": 0.00018346003896195132, "loss": 4.5331, "step": 706350 }, { "epoch": 1.3182204763715335, "grad_norm": 1.0754103660583496, "learning_rate": 0.0001834577247533217, "loss": 4.5745, "step": 706400 }, { "epoch": 1.3183137819309119, "grad_norm": 1.0212899446487427, "learning_rate": 0.00018345541039740382, "loss": 4.3935, "step": 706450 }, { "epoch": 1.3184070874902905, "grad_norm": 0.8962958455085754, "learning_rate": 0.00018345309589420187, "loss": 4.4306, "step": 706500 }, { "epoch": 1.3185003930496688, "grad_norm": 1.2989037036895752, "learning_rate": 0.00018345078124371988, "loss": 4.5653, "step": 706550 }, { "epoch": 1.3185936986090474, "grad_norm": 1.2930138111114502, "learning_rate": 0.00018344846644596194, "loss": 4.5325, "step": 706600 }, { "epoch": 1.3186870041684258, "grad_norm": 1.3549121618270874, "learning_rate": 0.00018344615150093217, "loss": 4.5409, "step": 706650 }, { "epoch": 1.3187803097278044, "grad_norm": 1.3048454523086548, "learning_rate": 0.00018344383640863456, "loss": 4.4501, "step": 706700 }, { "epoch": 1.318873615287183, "grad_norm": 1.0230778455734253, "learning_rate": 0.00018344152116907332, "loss": 4.4358, "step": 706750 }, { "epoch": 1.3189669208465613, "grad_norm": 1.2162796258926392, "learning_rate": 0.00018343920578225243, "loss": 4.3039, "step": 706800 }, { "epoch": 1.31906022640594, "grad_norm": 0.9299620985984802, "learning_rate": 0.00018343689024817608, "loss": 4.4538, "step": 706850 }, { "epoch": 1.3191535319653183, "grad_norm": 1.14543879032135, "learning_rate": 0.00018343457456684826, "loss": 4.5722, "step": 706900 }, { "epoch": 1.3192468375246968, "grad_norm": 0.8599335551261902, "learning_rate": 0.00018343225873827313, "loss": 4.4904, "step": 706950 }, { "epoch": 1.3193401430840752, "grad_norm": 1.4482200145721436, "learning_rate": 0.0001834299427624547, "loss": 4.5449, "step": 707000 }, { "epoch": 1.3194334486434538, "grad_norm": 1.5558801889419556, "learning_rate": 0.00018342762663939711, "loss": 4.4983, "step": 707050 }, { "epoch": 1.3195267542028324, "grad_norm": 1.210713267326355, "learning_rate": 0.00018342531036910447, "loss": 4.5692, "step": 707100 }, { "epoch": 1.3196200597622108, "grad_norm": 1.0205093622207642, "learning_rate": 0.0001834229939515808, "loss": 4.4111, "step": 707150 }, { "epoch": 1.3197133653215893, "grad_norm": 1.0352411270141602, "learning_rate": 0.00018342067738683025, "loss": 4.5983, "step": 707200 }, { "epoch": 1.3198066708809677, "grad_norm": 1.2423447370529175, "learning_rate": 0.00018341836067485687, "loss": 4.5948, "step": 707250 }, { "epoch": 1.3198999764403463, "grad_norm": 1.0278270244598389, "learning_rate": 0.00018341604381566477, "loss": 4.6663, "step": 707300 }, { "epoch": 1.3199932819997247, "grad_norm": 1.248896598815918, "learning_rate": 0.00018341372680925804, "loss": 4.5043, "step": 707350 }, { "epoch": 1.3200865875591032, "grad_norm": 1.1909626722335815, "learning_rate": 0.00018341140965564073, "loss": 4.442, "step": 707400 }, { "epoch": 1.3201798931184818, "grad_norm": 1.030617356300354, "learning_rate": 0.00018340909235481698, "loss": 4.5472, "step": 707450 }, { "epoch": 1.3202731986778602, "grad_norm": 1.1060172319412231, "learning_rate": 0.00018340677490679083, "loss": 4.493, "step": 707500 }, { "epoch": 1.3203665042372388, "grad_norm": 1.0220701694488525, "learning_rate": 0.00018340445731156643, "loss": 4.4153, "step": 707550 }, { "epoch": 1.3204598097966171, "grad_norm": 0.9419045448303223, "learning_rate": 0.00018340213956914781, "loss": 4.612, "step": 707600 }, { "epoch": 1.3205531153559957, "grad_norm": 1.0996309518814087, "learning_rate": 0.00018339982167953913, "loss": 4.5462, "step": 707650 }, { "epoch": 1.320646420915374, "grad_norm": 1.024959683418274, "learning_rate": 0.00018339750364274442, "loss": 4.6799, "step": 707700 }, { "epoch": 1.3207397264747527, "grad_norm": 1.1082286834716797, "learning_rate": 0.0001833951854587678, "loss": 4.4898, "step": 707750 }, { "epoch": 1.3208330320341313, "grad_norm": 1.0347074270248413, "learning_rate": 0.00018339286712761333, "loss": 4.6675, "step": 707800 }, { "epoch": 1.3209263375935096, "grad_norm": 1.197447419166565, "learning_rate": 0.00018339054864928512, "loss": 4.356, "step": 707850 }, { "epoch": 1.3210196431528882, "grad_norm": 1.0320606231689453, "learning_rate": 0.00018338823002378728, "loss": 4.5209, "step": 707900 }, { "epoch": 1.3211129487122666, "grad_norm": 1.26980721950531, "learning_rate": 0.00018338591125112387, "loss": 4.5425, "step": 707950 }, { "epoch": 1.3212062542716452, "grad_norm": 0.8710963726043701, "learning_rate": 0.00018338359233129902, "loss": 4.349, "step": 708000 }, { "epoch": 1.3212062542716452, "eval_loss": 4.677615642547607, "eval_runtime": 230.518, "eval_samples_per_second": 11.314, "eval_steps_per_second": 11.314, "eval_tts_loss": 7.653953886082385, "step": 708000 }, { "epoch": 1.3212995598310235, "grad_norm": 0.9818108081817627, "learning_rate": 0.0001833812732643168, "loss": 4.494, "step": 708050 }, { "epoch": 1.3213928653904021, "grad_norm": 1.4680957794189453, "learning_rate": 0.0001833789540501813, "loss": 4.4136, "step": 708100 }, { "epoch": 1.3214861709497807, "grad_norm": 1.1758538484573364, "learning_rate": 0.0001833766346888966, "loss": 4.4609, "step": 708150 }, { "epoch": 1.321579476509159, "grad_norm": 1.4567863941192627, "learning_rate": 0.0001833743151804668, "loss": 4.5754, "step": 708200 }, { "epoch": 1.3216727820685377, "grad_norm": 1.1099510192871094, "learning_rate": 0.00018337199552489604, "loss": 4.5592, "step": 708250 }, { "epoch": 1.321766087627916, "grad_norm": 1.1618531942367554, "learning_rate": 0.00018336967572218837, "loss": 4.5588, "step": 708300 }, { "epoch": 1.3218593931872946, "grad_norm": 1.0667684078216553, "learning_rate": 0.0001833673557723479, "loss": 4.3837, "step": 708350 }, { "epoch": 1.321952698746673, "grad_norm": 1.2158434391021729, "learning_rate": 0.0001833650356753787, "loss": 4.4636, "step": 708400 }, { "epoch": 1.3220460043060516, "grad_norm": 1.0587146282196045, "learning_rate": 0.00018336271543128486, "loss": 4.6221, "step": 708450 }, { "epoch": 1.3221393098654302, "grad_norm": 1.0799835920333862, "learning_rate": 0.00018336039504007054, "loss": 4.5618, "step": 708500 }, { "epoch": 1.3222326154248085, "grad_norm": 1.2482560873031616, "learning_rate": 0.00018335807450173977, "loss": 4.5814, "step": 708550 }, { "epoch": 1.322325920984187, "grad_norm": 1.2263017892837524, "learning_rate": 0.00018335575381629666, "loss": 4.5521, "step": 708600 }, { "epoch": 1.3224192265435655, "grad_norm": 1.045095443725586, "learning_rate": 0.00018335343298374532, "loss": 4.5181, "step": 708650 }, { "epoch": 1.322512532102944, "grad_norm": 1.2474509477615356, "learning_rate": 0.00018335111200408984, "loss": 4.316, "step": 708700 }, { "epoch": 1.3226058376623224, "grad_norm": 1.1526609659194946, "learning_rate": 0.00018334879087733433, "loss": 4.5755, "step": 708750 }, { "epoch": 1.322699143221701, "grad_norm": 1.0286951065063477, "learning_rate": 0.00018334646960348284, "loss": 4.5376, "step": 708800 }, { "epoch": 1.3227924487810796, "grad_norm": 1.4859594106674194, "learning_rate": 0.00018334414818253948, "loss": 4.4853, "step": 708850 }, { "epoch": 1.322885754340458, "grad_norm": 1.2087652683258057, "learning_rate": 0.00018334182661450837, "loss": 4.2955, "step": 708900 }, { "epoch": 1.3229790598998366, "grad_norm": 1.2679734230041504, "learning_rate": 0.00018333950489939364, "loss": 4.5035, "step": 708950 }, { "epoch": 1.323072365459215, "grad_norm": 1.2197210788726807, "learning_rate": 0.0001833371830371993, "loss": 4.2759, "step": 709000 }, { "epoch": 1.3231656710185935, "grad_norm": 1.1909102201461792, "learning_rate": 0.00018333486102792954, "loss": 4.7405, "step": 709050 }, { "epoch": 1.3232589765779719, "grad_norm": 1.0849087238311768, "learning_rate": 0.00018333253887158838, "loss": 4.4862, "step": 709100 }, { "epoch": 1.3233522821373505, "grad_norm": 1.0201200246810913, "learning_rate": 0.00018333021656817997, "loss": 4.6408, "step": 709150 }, { "epoch": 1.323445587696729, "grad_norm": 0.921275794506073, "learning_rate": 0.0001833278941177084, "loss": 4.4822, "step": 709200 }, { "epoch": 1.3235388932561074, "grad_norm": 1.1400314569473267, "learning_rate": 0.00018332557152017775, "loss": 4.4365, "step": 709250 }, { "epoch": 1.323632198815486, "grad_norm": 0.889840304851532, "learning_rate": 0.00018332324877559213, "loss": 4.5257, "step": 709300 }, { "epoch": 1.3237255043748644, "grad_norm": 1.1283223628997803, "learning_rate": 0.00018332092588395563, "loss": 4.6439, "step": 709350 }, { "epoch": 1.323818809934243, "grad_norm": 1.3155548572540283, "learning_rate": 0.00018331860284527234, "loss": 4.3779, "step": 709400 }, { "epoch": 1.3239121154936213, "grad_norm": 1.3005857467651367, "learning_rate": 0.0001833162796595464, "loss": 4.1377, "step": 709450 }, { "epoch": 1.324005421053, "grad_norm": 1.1399154663085938, "learning_rate": 0.00018331395632678187, "loss": 4.5035, "step": 709500 }, { "epoch": 1.3240987266123785, "grad_norm": 1.1868959665298462, "learning_rate": 0.00018331163284698287, "loss": 4.6812, "step": 709550 }, { "epoch": 1.3241920321717568, "grad_norm": 1.2002218961715698, "learning_rate": 0.00018330930922015351, "loss": 4.699, "step": 709600 }, { "epoch": 1.3242853377311354, "grad_norm": 0.9125586748123169, "learning_rate": 0.00018330698544629788, "loss": 4.5165, "step": 709650 }, { "epoch": 1.3243786432905138, "grad_norm": 0.8916546702384949, "learning_rate": 0.00018330466152542007, "loss": 4.5999, "step": 709700 }, { "epoch": 1.3244719488498924, "grad_norm": 1.0152018070220947, "learning_rate": 0.00018330233745752418, "loss": 4.4334, "step": 709750 }, { "epoch": 1.3245652544092708, "grad_norm": 1.2260816097259521, "learning_rate": 0.00018330001324261434, "loss": 4.6999, "step": 709800 }, { "epoch": 1.3246585599686493, "grad_norm": 0.9767405986785889, "learning_rate": 0.00018329768888069464, "loss": 4.6042, "step": 709850 }, { "epoch": 1.324751865528028, "grad_norm": 1.2037492990493774, "learning_rate": 0.00018329536437176915, "loss": 4.419, "step": 709900 }, { "epoch": 1.3248451710874063, "grad_norm": 1.1466888189315796, "learning_rate": 0.000183293039715842, "loss": 4.3826, "step": 709950 }, { "epoch": 1.3249384766467849, "grad_norm": 1.3488885164260864, "learning_rate": 0.00018329071491291732, "loss": 4.4739, "step": 710000 }, { "epoch": 1.3250317822061632, "grad_norm": 0.9292241930961609, "learning_rate": 0.00018328838996299916, "loss": 4.6583, "step": 710050 }, { "epoch": 1.3251250877655418, "grad_norm": 0.889931321144104, "learning_rate": 0.00018328606486609164, "loss": 4.4744, "step": 710100 }, { "epoch": 1.3252183933249202, "grad_norm": 1.3071229457855225, "learning_rate": 0.0001832837396221989, "loss": 4.5169, "step": 710150 }, { "epoch": 1.3253116988842988, "grad_norm": 1.1417741775512695, "learning_rate": 0.00018328141423132498, "loss": 4.6228, "step": 710200 }, { "epoch": 1.3254050044436774, "grad_norm": 1.1649006605148315, "learning_rate": 0.00018327908869347404, "loss": 4.3167, "step": 710250 }, { "epoch": 1.3254983100030557, "grad_norm": 1.0937727689743042, "learning_rate": 0.00018327676300865014, "loss": 4.5867, "step": 710300 }, { "epoch": 1.3255916155624343, "grad_norm": 1.3280470371246338, "learning_rate": 0.0001832744371768574, "loss": 4.5841, "step": 710350 }, { "epoch": 1.3256849211218127, "grad_norm": 1.4030792713165283, "learning_rate": 0.0001832721111981, "loss": 4.5804, "step": 710400 }, { "epoch": 1.3257782266811913, "grad_norm": 0.9621208310127258, "learning_rate": 0.0001832697850723819, "loss": 4.6117, "step": 710450 }, { "epoch": 1.3258715322405696, "grad_norm": 1.2396498918533325, "learning_rate": 0.00018326745879970732, "loss": 4.5282, "step": 710500 }, { "epoch": 1.3259648377999482, "grad_norm": 1.3512295484542847, "learning_rate": 0.00018326513238008031, "loss": 4.5979, "step": 710550 }, { "epoch": 1.3260581433593268, "grad_norm": 1.4631006717681885, "learning_rate": 0.000183262805813505, "loss": 4.5862, "step": 710600 }, { "epoch": 1.3261514489187052, "grad_norm": 1.1794764995574951, "learning_rate": 0.0001832604790999855, "loss": 4.4212, "step": 710650 }, { "epoch": 1.3262447544780838, "grad_norm": 1.1507667303085327, "learning_rate": 0.00018325815223952586, "loss": 4.3061, "step": 710700 }, { "epoch": 1.3263380600374621, "grad_norm": 1.1797529458999634, "learning_rate": 0.00018325582523213028, "loss": 4.4211, "step": 710750 }, { "epoch": 1.3264313655968407, "grad_norm": 1.078227162361145, "learning_rate": 0.0001832534980778028, "loss": 4.4337, "step": 710800 }, { "epoch": 1.326524671156219, "grad_norm": 1.1896167993545532, "learning_rate": 0.00018325117077654755, "loss": 4.3495, "step": 710850 }, { "epoch": 1.3266179767155977, "grad_norm": 1.4299726486206055, "learning_rate": 0.00018324884332836862, "loss": 4.455, "step": 710900 }, { "epoch": 1.3267112822749763, "grad_norm": 1.3845069408416748, "learning_rate": 0.00018324651573327014, "loss": 4.6879, "step": 710950 }, { "epoch": 1.3268045878343546, "grad_norm": 1.2045581340789795, "learning_rate": 0.0001832441879912562, "loss": 4.4093, "step": 711000 }, { "epoch": 1.3268045878343546, "eval_loss": 4.674876689910889, "eval_runtime": 228.7573, "eval_samples_per_second": 11.401, "eval_steps_per_second": 11.401, "eval_tts_loss": 7.629302071711688, "step": 711000 }, { "epoch": 1.3268978933937332, "grad_norm": 0.9563471078872681, "learning_rate": 0.0001832418601023309, "loss": 4.5215, "step": 711050 }, { "epoch": 1.3269911989531116, "grad_norm": 1.4422688484191895, "learning_rate": 0.0001832395320664984, "loss": 4.5395, "step": 711100 }, { "epoch": 1.3270845045124902, "grad_norm": 1.31195867061615, "learning_rate": 0.00018323720388376279, "loss": 4.4986, "step": 711150 }, { "epoch": 1.3271778100718685, "grad_norm": 1.410945177078247, "learning_rate": 0.0001832348755541281, "loss": 4.6783, "step": 711200 }, { "epoch": 1.3272711156312471, "grad_norm": 1.4838124513626099, "learning_rate": 0.00018323254707759852, "loss": 4.5338, "step": 711250 }, { "epoch": 1.3273644211906257, "grad_norm": 1.186169981956482, "learning_rate": 0.00018323021845417813, "loss": 4.4373, "step": 711300 }, { "epoch": 1.327457726750004, "grad_norm": 1.2047319412231445, "learning_rate": 0.0001832278896838711, "loss": 4.5023, "step": 711350 }, { "epoch": 1.3275510323093827, "grad_norm": 1.3560255765914917, "learning_rate": 0.00018322556076668143, "loss": 4.5138, "step": 711400 }, { "epoch": 1.327644337868761, "grad_norm": 1.4438098669052124, "learning_rate": 0.00018322323170261333, "loss": 4.4321, "step": 711450 }, { "epoch": 1.3277376434281396, "grad_norm": 1.1492985486984253, "learning_rate": 0.00018322090249167086, "loss": 4.6421, "step": 711500 }, { "epoch": 1.327830948987518, "grad_norm": 1.0783565044403076, "learning_rate": 0.00018321857313385813, "loss": 4.6736, "step": 711550 }, { "epoch": 1.3279242545468966, "grad_norm": 1.3469622135162354, "learning_rate": 0.00018321624362917924, "loss": 4.3304, "step": 711600 }, { "epoch": 1.3280175601062751, "grad_norm": 0.8506263494491577, "learning_rate": 0.00018321391397763836, "loss": 4.5641, "step": 711650 }, { "epoch": 1.3281108656656535, "grad_norm": 1.256977915763855, "learning_rate": 0.00018321158417923953, "loss": 4.5854, "step": 711700 }, { "epoch": 1.328204171225032, "grad_norm": 1.2218027114868164, "learning_rate": 0.00018320925423398693, "loss": 4.4487, "step": 711750 }, { "epoch": 1.3282974767844105, "grad_norm": 1.757312536239624, "learning_rate": 0.00018320692414188462, "loss": 4.5267, "step": 711800 }, { "epoch": 1.328390782343789, "grad_norm": 1.1718025207519531, "learning_rate": 0.00018320459390293675, "loss": 4.5628, "step": 711850 }, { "epoch": 1.3284840879031674, "grad_norm": 0.9904388189315796, "learning_rate": 0.0001832022635171474, "loss": 4.372, "step": 711900 }, { "epoch": 1.328577393462546, "grad_norm": 0.8558657765388489, "learning_rate": 0.00018319993298452066, "loss": 4.4227, "step": 711950 }, { "epoch": 1.3286706990219246, "grad_norm": 1.3153644800186157, "learning_rate": 0.00018319760230506074, "loss": 4.5546, "step": 712000 }, { "epoch": 1.328764004581303, "grad_norm": 1.846069097518921, "learning_rate": 0.00018319527147877166, "loss": 4.4709, "step": 712050 }, { "epoch": 1.3288573101406815, "grad_norm": 1.1760563850402832, "learning_rate": 0.00018319294050565756, "loss": 4.2434, "step": 712100 }, { "epoch": 1.32895061570006, "grad_norm": 1.2539209127426147, "learning_rate": 0.00018319060938572253, "loss": 4.2261, "step": 712150 }, { "epoch": 1.3290439212594385, "grad_norm": 1.1570241451263428, "learning_rate": 0.00018318827811897073, "loss": 4.4202, "step": 712200 }, { "epoch": 1.3291372268188169, "grad_norm": 2.024951934814453, "learning_rate": 0.0001831859467054063, "loss": 4.5653, "step": 712250 }, { "epoch": 1.3292305323781954, "grad_norm": 1.0586397647857666, "learning_rate": 0.00018318361514503328, "loss": 4.4298, "step": 712300 }, { "epoch": 1.329323837937574, "grad_norm": 1.367389440536499, "learning_rate": 0.00018318128343785582, "loss": 4.5519, "step": 712350 }, { "epoch": 1.3294171434969524, "grad_norm": 1.147118330001831, "learning_rate": 0.00018317895158387803, "loss": 4.5075, "step": 712400 }, { "epoch": 1.329510449056331, "grad_norm": 0.9498062133789062, "learning_rate": 0.00018317661958310404, "loss": 4.5423, "step": 712450 }, { "epoch": 1.3296037546157093, "grad_norm": 1.1654185056686401, "learning_rate": 0.00018317428743553791, "loss": 4.615, "step": 712500 }, { "epoch": 1.329697060175088, "grad_norm": 1.1282330751419067, "learning_rate": 0.00018317195514118385, "loss": 4.4299, "step": 712550 }, { "epoch": 1.3297903657344663, "grad_norm": 1.111141562461853, "learning_rate": 0.00018316962270004589, "loss": 4.5702, "step": 712600 }, { "epoch": 1.3298836712938449, "grad_norm": 1.1217598915100098, "learning_rate": 0.0001831672901121282, "loss": 4.5651, "step": 712650 }, { "epoch": 1.3299769768532235, "grad_norm": 1.2727150917053223, "learning_rate": 0.00018316495737743485, "loss": 4.6071, "step": 712700 }, { "epoch": 1.3300702824126018, "grad_norm": 1.3248602151870728, "learning_rate": 0.00018316262449597, "loss": 4.3789, "step": 712750 }, { "epoch": 1.3301635879719804, "grad_norm": 1.1639046669006348, "learning_rate": 0.00018316029146773775, "loss": 4.3977, "step": 712800 }, { "epoch": 1.3302568935313588, "grad_norm": 1.1361329555511475, "learning_rate": 0.0001831579582927422, "loss": 4.689, "step": 712850 }, { "epoch": 1.3303501990907374, "grad_norm": 1.4606677293777466, "learning_rate": 0.0001831556249709875, "loss": 4.2491, "step": 712900 }, { "epoch": 1.3304435046501157, "grad_norm": 0.7569972276687622, "learning_rate": 0.00018315329150247775, "loss": 4.3325, "step": 712950 }, { "epoch": 1.3305368102094943, "grad_norm": 1.1583762168884277, "learning_rate": 0.0001831509578872171, "loss": 4.4996, "step": 713000 }, { "epoch": 1.330630115768873, "grad_norm": 1.2499489784240723, "learning_rate": 0.0001831486241252096, "loss": 4.4171, "step": 713050 }, { "epoch": 1.3307234213282513, "grad_norm": 1.384984016418457, "learning_rate": 0.00018314629021645945, "loss": 4.5041, "step": 713100 }, { "epoch": 1.3308167268876299, "grad_norm": 1.2906252145767212, "learning_rate": 0.0001831439561609707, "loss": 4.5696, "step": 713150 }, { "epoch": 1.3309100324470082, "grad_norm": 1.2806222438812256, "learning_rate": 0.0001831416219587475, "loss": 4.3741, "step": 713200 }, { "epoch": 1.3310033380063868, "grad_norm": 1.2693309783935547, "learning_rate": 0.00018313928760979397, "loss": 4.7026, "step": 713250 }, { "epoch": 1.3310966435657652, "grad_norm": 1.2903976440429688, "learning_rate": 0.00018313695311411422, "loss": 4.5857, "step": 713300 }, { "epoch": 1.3311899491251438, "grad_norm": 1.0674537420272827, "learning_rate": 0.00018313461847171237, "loss": 4.4207, "step": 713350 }, { "epoch": 1.3312832546845224, "grad_norm": 1.393294095993042, "learning_rate": 0.00018313228368259257, "loss": 4.3272, "step": 713400 }, { "epoch": 1.3313765602439007, "grad_norm": 1.1906671524047852, "learning_rate": 0.00018312994874675892, "loss": 4.4818, "step": 713450 }, { "epoch": 1.3314698658032793, "grad_norm": 1.5869470834732056, "learning_rate": 0.0001831276136642155, "loss": 4.482, "step": 713500 }, { "epoch": 1.3315631713626577, "grad_norm": 1.1162480115890503, "learning_rate": 0.0001831252784349665, "loss": 4.5035, "step": 713550 }, { "epoch": 1.3316564769220363, "grad_norm": 1.5096560716629028, "learning_rate": 0.000183122943059016, "loss": 4.6385, "step": 713600 }, { "epoch": 1.3317497824814146, "grad_norm": 0.9520151615142822, "learning_rate": 0.0001831206075363681, "loss": 4.5745, "step": 713650 }, { "epoch": 1.3318430880407932, "grad_norm": 1.0274381637573242, "learning_rate": 0.000183118271867027, "loss": 4.3361, "step": 713700 }, { "epoch": 1.3319363936001718, "grad_norm": 1.1635626554489136, "learning_rate": 0.00018311593605099678, "loss": 4.6955, "step": 713750 }, { "epoch": 1.3320296991595502, "grad_norm": 1.503702163696289, "learning_rate": 0.00018311360008828152, "loss": 4.5462, "step": 713800 }, { "epoch": 1.3321230047189285, "grad_norm": 1.1301119327545166, "learning_rate": 0.0001831112639788854, "loss": 4.3589, "step": 713850 }, { "epoch": 1.3322163102783071, "grad_norm": 1.3967399597167969, "learning_rate": 0.00018310892772281255, "loss": 4.4551, "step": 713900 }, { "epoch": 1.3323096158376857, "grad_norm": 1.2648407220840454, "learning_rate": 0.000183106591320067, "loss": 4.4431, "step": 713950 }, { "epoch": 1.332402921397064, "grad_norm": 1.1140605211257935, "learning_rate": 0.000183104254770653, "loss": 4.3267, "step": 714000 }, { "epoch": 1.332402921397064, "eval_loss": 4.680948734283447, "eval_runtime": 232.0978, "eval_samples_per_second": 11.237, "eval_steps_per_second": 11.237, "eval_tts_loss": 7.650018445366102, "step": 714000 }, { "epoch": 1.3324962269564427, "grad_norm": 1.2127209901809692, "learning_rate": 0.00018310191807457458, "loss": 4.5099, "step": 714050 }, { "epoch": 1.3325895325158212, "grad_norm": 0.9641920328140259, "learning_rate": 0.0001830995812318359, "loss": 4.6686, "step": 714100 }, { "epoch": 1.3326828380751996, "grad_norm": 0.9805702567100525, "learning_rate": 0.0001830972442424411, "loss": 4.4953, "step": 714150 }, { "epoch": 1.332776143634578, "grad_norm": 1.3272759914398193, "learning_rate": 0.00018309490710639428, "loss": 4.5965, "step": 714200 }, { "epoch": 1.3328694491939566, "grad_norm": 1.2475680112838745, "learning_rate": 0.00018309256982369953, "loss": 4.4485, "step": 714250 }, { "epoch": 1.3329627547533351, "grad_norm": 1.2206214666366577, "learning_rate": 0.00018309023239436106, "loss": 4.3924, "step": 714300 }, { "epoch": 1.3330560603127135, "grad_norm": 1.2247333526611328, "learning_rate": 0.00018308789481838293, "loss": 4.3942, "step": 714350 }, { "epoch": 1.333149365872092, "grad_norm": 1.0631234645843506, "learning_rate": 0.0001830855570957693, "loss": 4.5271, "step": 714400 }, { "epoch": 1.3332426714314707, "grad_norm": 0.9795876145362854, "learning_rate": 0.00018308321922652426, "loss": 4.3129, "step": 714450 }, { "epoch": 1.333335976990849, "grad_norm": 1.071608066558838, "learning_rate": 0.00018308088121065196, "loss": 4.5944, "step": 714500 }, { "epoch": 1.3334292825502274, "grad_norm": 1.1208454370498657, "learning_rate": 0.00018307854304815656, "loss": 4.6442, "step": 714550 }, { "epoch": 1.333522588109606, "grad_norm": 1.2317123413085938, "learning_rate": 0.00018307620473904213, "loss": 4.446, "step": 714600 }, { "epoch": 1.3336158936689846, "grad_norm": 1.3131434917449951, "learning_rate": 0.0001830738662833128, "loss": 4.5161, "step": 714650 }, { "epoch": 1.333709199228363, "grad_norm": 1.0751861333847046, "learning_rate": 0.00018307152768097274, "loss": 4.5007, "step": 714700 }, { "epoch": 1.3338025047877415, "grad_norm": 1.0153485536575317, "learning_rate": 0.000183069188932026, "loss": 4.4324, "step": 714750 }, { "epoch": 1.3338958103471201, "grad_norm": 1.5131027698516846, "learning_rate": 0.00018306685003647682, "loss": 4.7297, "step": 714800 }, { "epoch": 1.3339891159064985, "grad_norm": 1.7317752838134766, "learning_rate": 0.00018306451099432921, "loss": 4.5798, "step": 714850 }, { "epoch": 1.3340824214658769, "grad_norm": 1.0991414785385132, "learning_rate": 0.0001830621718055874, "loss": 4.3164, "step": 714900 }, { "epoch": 1.3341757270252554, "grad_norm": 1.1223978996276855, "learning_rate": 0.00018305983247025545, "loss": 4.4456, "step": 714950 }, { "epoch": 1.334269032584634, "grad_norm": 1.2451069355010986, "learning_rate": 0.00018305749298833752, "loss": 4.5026, "step": 715000 }, { "epoch": 1.3343623381440124, "grad_norm": 1.2074612379074097, "learning_rate": 0.0001830551533598377, "loss": 4.5803, "step": 715050 }, { "epoch": 1.334455643703391, "grad_norm": 1.010645866394043, "learning_rate": 0.0001830528135847602, "loss": 4.4605, "step": 715100 }, { "epoch": 1.3345489492627696, "grad_norm": 1.2807483673095703, "learning_rate": 0.00018305047366310906, "loss": 4.6097, "step": 715150 }, { "epoch": 1.334642254822148, "grad_norm": 0.992500901222229, "learning_rate": 0.00018304813359488843, "loss": 4.564, "step": 715200 }, { "epoch": 1.3347355603815263, "grad_norm": 0.9755114316940308, "learning_rate": 0.0001830457933801025, "loss": 4.3783, "step": 715250 }, { "epoch": 1.3348288659409049, "grad_norm": 1.2022271156311035, "learning_rate": 0.00018304345301875535, "loss": 4.5742, "step": 715300 }, { "epoch": 1.3349221715002835, "grad_norm": 1.3685028553009033, "learning_rate": 0.0001830411125108511, "loss": 4.5864, "step": 715350 }, { "epoch": 1.3350154770596618, "grad_norm": 1.5301676988601685, "learning_rate": 0.0001830387718563939, "loss": 4.4001, "step": 715400 }, { "epoch": 1.3351087826190404, "grad_norm": 1.0757914781570435, "learning_rate": 0.0001830364310553879, "loss": 4.3573, "step": 715450 }, { "epoch": 1.335202088178419, "grad_norm": 1.0385708808898926, "learning_rate": 0.00018303409010783718, "loss": 4.5908, "step": 715500 }, { "epoch": 1.3352953937377974, "grad_norm": 1.0516430139541626, "learning_rate": 0.00018303174901374591, "loss": 4.6019, "step": 715550 }, { "epoch": 1.3353886992971757, "grad_norm": 1.264967679977417, "learning_rate": 0.00018302940777311822, "loss": 4.385, "step": 715600 }, { "epoch": 1.3354820048565543, "grad_norm": 0.8779059648513794, "learning_rate": 0.00018302706638595822, "loss": 4.36, "step": 715650 }, { "epoch": 1.335575310415933, "grad_norm": 1.0630375146865845, "learning_rate": 0.00018302472485227006, "loss": 4.4937, "step": 715700 }, { "epoch": 1.3356686159753113, "grad_norm": 1.3747918605804443, "learning_rate": 0.00018302238317205788, "loss": 4.5884, "step": 715750 }, { "epoch": 1.3357619215346899, "grad_norm": 1.0915942192077637, "learning_rate": 0.0001830200413453258, "loss": 4.4274, "step": 715800 }, { "epoch": 1.3358552270940685, "grad_norm": 1.308139681816101, "learning_rate": 0.000183017699372078, "loss": 4.5477, "step": 715850 }, { "epoch": 1.3359485326534468, "grad_norm": 0.9784302115440369, "learning_rate": 0.00018301535725231852, "loss": 4.5156, "step": 715900 }, { "epoch": 1.3360418382128252, "grad_norm": 1.2524387836456299, "learning_rate": 0.00018301301498605153, "loss": 4.5817, "step": 715950 }, { "epoch": 1.3361351437722038, "grad_norm": 1.397838830947876, "learning_rate": 0.0001830106725732812, "loss": 4.274, "step": 716000 }, { "epoch": 1.3362284493315824, "grad_norm": 1.2730604410171509, "learning_rate": 0.0001830083300140116, "loss": 4.6848, "step": 716050 }, { "epoch": 1.3363217548909607, "grad_norm": 0.8434470891952515, "learning_rate": 0.00018300598730824693, "loss": 4.3548, "step": 716100 }, { "epoch": 1.3364150604503393, "grad_norm": 0.8471757173538208, "learning_rate": 0.0001830036444559913, "loss": 4.5411, "step": 716150 }, { "epoch": 1.336508366009718, "grad_norm": 1.4233442544937134, "learning_rate": 0.00018300130145724888, "loss": 4.4247, "step": 716200 }, { "epoch": 1.3366016715690963, "grad_norm": 1.3700203895568848, "learning_rate": 0.00018299895831202373, "loss": 4.5736, "step": 716250 }, { "epoch": 1.3366949771284746, "grad_norm": 1.3264968395233154, "learning_rate": 0.00018299661502032, "loss": 4.5197, "step": 716300 }, { "epoch": 1.3367882826878532, "grad_norm": 1.2423739433288574, "learning_rate": 0.00018299427158214186, "loss": 4.4232, "step": 716350 }, { "epoch": 1.3368815882472318, "grad_norm": 1.1304372549057007, "learning_rate": 0.00018299192799749346, "loss": 4.5662, "step": 716400 }, { "epoch": 1.3369748938066102, "grad_norm": 0.9997304081916809, "learning_rate": 0.00018298958426637892, "loss": 4.5918, "step": 716450 }, { "epoch": 1.3370681993659888, "grad_norm": 1.0660480260849, "learning_rate": 0.00018298724038880233, "loss": 4.5911, "step": 716500 }, { "epoch": 1.3371615049253673, "grad_norm": 0.9851765632629395, "learning_rate": 0.0001829848963647679, "loss": 4.6158, "step": 716550 }, { "epoch": 1.3372548104847457, "grad_norm": 1.1552637815475464, "learning_rate": 0.00018298255219427966, "loss": 4.2918, "step": 716600 }, { "epoch": 1.337348116044124, "grad_norm": 1.1929959058761597, "learning_rate": 0.00018298020787734187, "loss": 4.5655, "step": 716650 }, { "epoch": 1.3374414216035027, "grad_norm": 1.0340602397918701, "learning_rate": 0.0001829778634139586, "loss": 4.5534, "step": 716700 }, { "epoch": 1.3375347271628812, "grad_norm": 1.3561204671859741, "learning_rate": 0.000182975518804134, "loss": 4.6325, "step": 716750 }, { "epoch": 1.3376280327222596, "grad_norm": 1.3070539236068726, "learning_rate": 0.0001829731740478722, "loss": 4.3743, "step": 716800 }, { "epoch": 1.3377213382816382, "grad_norm": 0.829119861125946, "learning_rate": 0.00018297082914517737, "loss": 4.4302, "step": 716850 }, { "epoch": 1.3378146438410168, "grad_norm": 1.239051103591919, "learning_rate": 0.00018296848409605362, "loss": 4.5148, "step": 716900 }, { "epoch": 1.3379079494003951, "grad_norm": 1.3790678977966309, "learning_rate": 0.00018296613890050512, "loss": 4.6812, "step": 716950 }, { "epoch": 1.3380012549597735, "grad_norm": 1.1659432649612427, "learning_rate": 0.00018296379355853596, "loss": 4.4055, "step": 717000 }, { "epoch": 1.3380012549597735, "eval_loss": 4.671762466430664, "eval_runtime": 231.7475, "eval_samples_per_second": 11.254, "eval_steps_per_second": 11.254, "eval_tts_loss": 7.659415488994103, "step": 717000 }, { "epoch": 1.338094560519152, "grad_norm": 1.1596704721450806, "learning_rate": 0.00018296144807015028, "loss": 4.5869, "step": 717050 }, { "epoch": 1.3381878660785307, "grad_norm": 1.227131724357605, "learning_rate": 0.00018295910243535225, "loss": 4.5339, "step": 717100 }, { "epoch": 1.338281171637909, "grad_norm": 0.713901937007904, "learning_rate": 0.000182956756654146, "loss": 4.5864, "step": 717150 }, { "epoch": 1.3383744771972876, "grad_norm": 1.2227914333343506, "learning_rate": 0.0001829544107265357, "loss": 4.6724, "step": 717200 }, { "epoch": 1.3384677827566662, "grad_norm": 1.193021535873413, "learning_rate": 0.00018295206465252545, "loss": 4.4334, "step": 717250 }, { "epoch": 1.3385610883160446, "grad_norm": 1.1111620664596558, "learning_rate": 0.0001829497184321194, "loss": 4.6258, "step": 717300 }, { "epoch": 1.338654393875423, "grad_norm": 1.0398434400558472, "learning_rate": 0.00018294737206532168, "loss": 4.4811, "step": 717350 }, { "epoch": 1.3387476994348015, "grad_norm": 1.3475229740142822, "learning_rate": 0.00018294502555213647, "loss": 4.5691, "step": 717400 }, { "epoch": 1.3388410049941801, "grad_norm": 1.1407712697982788, "learning_rate": 0.00018294267889256785, "loss": 4.4998, "step": 717450 }, { "epoch": 1.3389343105535585, "grad_norm": 1.3098093271255493, "learning_rate": 0.00018294033208662, "loss": 4.3172, "step": 717500 }, { "epoch": 1.339027616112937, "grad_norm": 1.0240371227264404, "learning_rate": 0.00018293798513429707, "loss": 4.5602, "step": 717550 }, { "epoch": 1.3391209216723157, "grad_norm": 1.1010318994522095, "learning_rate": 0.00018293563803560317, "loss": 4.5525, "step": 717600 }, { "epoch": 1.339214227231694, "grad_norm": 1.4469845294952393, "learning_rate": 0.0001829332907905425, "loss": 4.5206, "step": 717650 }, { "epoch": 1.3393075327910724, "grad_norm": 0.8986558318138123, "learning_rate": 0.00018293094339911917, "loss": 4.5654, "step": 717700 }, { "epoch": 1.339400838350451, "grad_norm": 1.1487009525299072, "learning_rate": 0.00018292859586133728, "loss": 4.7203, "step": 717750 }, { "epoch": 1.3394941439098296, "grad_norm": 1.0499553680419922, "learning_rate": 0.00018292624817720104, "loss": 4.3443, "step": 717800 }, { "epoch": 1.339587449469208, "grad_norm": 1.184159278869629, "learning_rate": 0.0001829239003467145, "loss": 4.5216, "step": 717850 }, { "epoch": 1.3396807550285865, "grad_norm": 1.4192224740982056, "learning_rate": 0.00018292155236988196, "loss": 4.5835, "step": 717900 }, { "epoch": 1.339774060587965, "grad_norm": 1.0665255784988403, "learning_rate": 0.00018291920424670738, "loss": 4.3514, "step": 717950 }, { "epoch": 1.3398673661473435, "grad_norm": 1.054736852645874, "learning_rate": 0.00018291685597719506, "loss": 4.5171, "step": 718000 }, { "epoch": 1.3399606717067218, "grad_norm": 1.2817853689193726, "learning_rate": 0.00018291450756134905, "loss": 4.5274, "step": 718050 }, { "epoch": 1.3400539772661004, "grad_norm": 1.4034074544906616, "learning_rate": 0.00018291215899917353, "loss": 4.601, "step": 718100 }, { "epoch": 1.340147282825479, "grad_norm": 1.3095685243606567, "learning_rate": 0.00018290981029067265, "loss": 4.5725, "step": 718150 }, { "epoch": 1.3402405883848574, "grad_norm": 1.1568747758865356, "learning_rate": 0.00018290746143585052, "loss": 4.6331, "step": 718200 }, { "epoch": 1.340333893944236, "grad_norm": 1.2875946760177612, "learning_rate": 0.00018290511243471133, "loss": 4.4888, "step": 718250 }, { "epoch": 1.3404271995036146, "grad_norm": 1.082026481628418, "learning_rate": 0.00018290276328725917, "loss": 4.666, "step": 718300 }, { "epoch": 1.340520505062993, "grad_norm": 1.2711042165756226, "learning_rate": 0.00018290041399349826, "loss": 4.6829, "step": 718350 }, { "epoch": 1.3406138106223713, "grad_norm": 1.2334351539611816, "learning_rate": 0.0001828980645534327, "loss": 4.4064, "step": 718400 }, { "epoch": 1.3407071161817499, "grad_norm": 1.0347411632537842, "learning_rate": 0.00018289571496706662, "loss": 4.524, "step": 718450 }, { "epoch": 1.3408004217411285, "grad_norm": 1.2218667268753052, "learning_rate": 0.00018289336523440422, "loss": 4.6565, "step": 718500 }, { "epoch": 1.3408937273005068, "grad_norm": 1.177647352218628, "learning_rate": 0.0001828910153554496, "loss": 4.4274, "step": 718550 }, { "epoch": 1.3409870328598854, "grad_norm": 1.2965525388717651, "learning_rate": 0.0001828886653302069, "loss": 4.4754, "step": 718600 }, { "epoch": 1.341080338419264, "grad_norm": 1.2305800914764404, "learning_rate": 0.00018288631515868032, "loss": 4.5299, "step": 718650 }, { "epoch": 1.3411736439786424, "grad_norm": 1.1197649240493774, "learning_rate": 0.00018288396484087395, "loss": 4.4672, "step": 718700 }, { "epoch": 1.3412669495380207, "grad_norm": 1.057196855545044, "learning_rate": 0.00018288161437679198, "loss": 4.5023, "step": 718750 }, { "epoch": 1.3413602550973993, "grad_norm": 0.7750287055969238, "learning_rate": 0.00018287926376643856, "loss": 4.5933, "step": 718800 }, { "epoch": 1.341453560656778, "grad_norm": 1.3602368831634521, "learning_rate": 0.00018287691300981782, "loss": 4.6775, "step": 718850 }, { "epoch": 1.3415468662161563, "grad_norm": 1.060402750968933, "learning_rate": 0.0001828745621069339, "loss": 4.4825, "step": 718900 }, { "epoch": 1.3416401717755349, "grad_norm": 0.9970949292182922, "learning_rate": 0.00018287221105779095, "loss": 4.441, "step": 718950 }, { "epoch": 1.3417334773349134, "grad_norm": 1.3610543012619019, "learning_rate": 0.00018286985986239315, "loss": 4.4369, "step": 719000 }, { "epoch": 1.3418267828942918, "grad_norm": 1.1624280214309692, "learning_rate": 0.0001828675085207446, "loss": 4.489, "step": 719050 }, { "epoch": 1.3419200884536702, "grad_norm": 1.0609585046768188, "learning_rate": 0.00018286515703284952, "loss": 4.3462, "step": 719100 }, { "epoch": 1.3420133940130488, "grad_norm": 0.9756843447685242, "learning_rate": 0.000182862805398712, "loss": 4.626, "step": 719150 }, { "epoch": 1.3421066995724273, "grad_norm": 1.1462373733520508, "learning_rate": 0.00018286045361833622, "loss": 4.4521, "step": 719200 }, { "epoch": 1.3422000051318057, "grad_norm": 1.082505464553833, "learning_rate": 0.0001828581016917263, "loss": 4.403, "step": 719250 }, { "epoch": 1.3422933106911843, "grad_norm": 1.0132452249526978, "learning_rate": 0.0001828557496188864, "loss": 4.3168, "step": 719300 }, { "epoch": 1.3423866162505629, "grad_norm": 1.698133945465088, "learning_rate": 0.0001828533973998207, "loss": 4.4735, "step": 719350 }, { "epoch": 1.3424799218099412, "grad_norm": 1.1703957319259644, "learning_rate": 0.00018285104503453333, "loss": 4.3498, "step": 719400 }, { "epoch": 1.3425732273693196, "grad_norm": 0.9553424119949341, "learning_rate": 0.00018284869252302842, "loss": 4.5713, "step": 719450 }, { "epoch": 1.3426665329286982, "grad_norm": 1.0753251314163208, "learning_rate": 0.0001828463398653102, "loss": 4.6325, "step": 719500 }, { "epoch": 1.3427598384880768, "grad_norm": 1.1915663480758667, "learning_rate": 0.0001828439870613827, "loss": 4.6515, "step": 719550 }, { "epoch": 1.3428531440474551, "grad_norm": 1.0035454034805298, "learning_rate": 0.00018284163411125018, "loss": 4.4136, "step": 719600 }, { "epoch": 1.3429464496068337, "grad_norm": 0.8618941307067871, "learning_rate": 0.00018283928101491676, "loss": 4.5718, "step": 719650 }, { "epoch": 1.3430397551662123, "grad_norm": 1.1903536319732666, "learning_rate": 0.00018283692777238658, "loss": 4.5023, "step": 719700 }, { "epoch": 1.3431330607255907, "grad_norm": 1.329116940498352, "learning_rate": 0.00018283457438366378, "loss": 4.5024, "step": 719750 }, { "epoch": 1.343226366284969, "grad_norm": 1.0333548784255981, "learning_rate": 0.00018283222084875255, "loss": 4.4538, "step": 719800 }, { "epoch": 1.3433196718443476, "grad_norm": 1.1603901386260986, "learning_rate": 0.000182829867167657, "loss": 4.4832, "step": 719850 }, { "epoch": 1.3434129774037262, "grad_norm": 1.251969575881958, "learning_rate": 0.00018282751334038134, "loss": 4.6521, "step": 719900 }, { "epoch": 1.3435062829631046, "grad_norm": 1.1947156190872192, "learning_rate": 0.00018282515936692969, "loss": 4.7215, "step": 719950 }, { "epoch": 1.3435995885224832, "grad_norm": 1.212303876876831, "learning_rate": 0.00018282280524730616, "loss": 4.4079, "step": 720000 }, { "epoch": 1.3435995885224832, "eval_loss": 4.668597221374512, "eval_runtime": 230.5607, "eval_samples_per_second": 11.312, "eval_steps_per_second": 11.312, "eval_tts_loss": 7.671031935878978, "step": 720000 }, { "epoch": 1.3436928940818618, "grad_norm": 1.5442335605621338, "learning_rate": 0.000182820450981515, "loss": 4.3604, "step": 720050 }, { "epoch": 1.3437861996412401, "grad_norm": 1.2628570795059204, "learning_rate": 0.00018281809656956033, "loss": 4.3788, "step": 720100 }, { "epoch": 1.3438795052006185, "grad_norm": 1.2394814491271973, "learning_rate": 0.00018281574201144625, "loss": 4.5317, "step": 720150 }, { "epoch": 1.343972810759997, "grad_norm": 1.1410938501358032, "learning_rate": 0.00018281338730717696, "loss": 4.5303, "step": 720200 }, { "epoch": 1.3440661163193757, "grad_norm": 1.3568419218063354, "learning_rate": 0.00018281103245675664, "loss": 4.5913, "step": 720250 }, { "epoch": 1.344159421878754, "grad_norm": 1.2986372709274292, "learning_rate": 0.0001828086774601894, "loss": 4.4262, "step": 720300 }, { "epoch": 1.3442527274381326, "grad_norm": 1.078208088874817, "learning_rate": 0.0001828063223174794, "loss": 4.4558, "step": 720350 }, { "epoch": 1.3443460329975112, "grad_norm": 1.3531571626663208, "learning_rate": 0.00018280396702863083, "loss": 4.64, "step": 720400 }, { "epoch": 1.3444393385568896, "grad_norm": 1.1682240962982178, "learning_rate": 0.00018280161159364784, "loss": 4.4816, "step": 720450 }, { "epoch": 1.344532644116268, "grad_norm": 1.0927633047103882, "learning_rate": 0.00018279925601253459, "loss": 4.5772, "step": 720500 }, { "epoch": 1.3446259496756465, "grad_norm": 1.3668309450149536, "learning_rate": 0.00018279690028529517, "loss": 4.5118, "step": 720550 }, { "epoch": 1.3447192552350251, "grad_norm": 0.9260604381561279, "learning_rate": 0.0001827945444119338, "loss": 4.3798, "step": 720600 }, { "epoch": 1.3448125607944035, "grad_norm": 1.333821177482605, "learning_rate": 0.00018279218839245464, "loss": 4.4423, "step": 720650 }, { "epoch": 1.344905866353782, "grad_norm": 1.0726733207702637, "learning_rate": 0.00018278983222686185, "loss": 4.5717, "step": 720700 }, { "epoch": 1.3449991719131607, "grad_norm": 1.1527931690216064, "learning_rate": 0.00018278747591515956, "loss": 4.4919, "step": 720750 }, { "epoch": 1.345092477472539, "grad_norm": 1.3017957210540771, "learning_rate": 0.00018278511945735195, "loss": 4.5585, "step": 720800 }, { "epoch": 1.3451857830319174, "grad_norm": 1.3663607835769653, "learning_rate": 0.0001827827628534432, "loss": 4.4815, "step": 720850 }, { "epoch": 1.345279088591296, "grad_norm": 1.243022084236145, "learning_rate": 0.00018278040610343737, "loss": 4.6305, "step": 720900 }, { "epoch": 1.3453723941506746, "grad_norm": 1.451054573059082, "learning_rate": 0.00018277804920733873, "loss": 4.5912, "step": 720950 }, { "epoch": 1.345465699710053, "grad_norm": 1.346252679824829, "learning_rate": 0.00018277569216515138, "loss": 4.4717, "step": 721000 }, { "epoch": 1.3455590052694315, "grad_norm": 1.0227022171020508, "learning_rate": 0.00018277333497687952, "loss": 4.6317, "step": 721050 }, { "epoch": 1.34565231082881, "grad_norm": 1.3935433626174927, "learning_rate": 0.00018277097764252728, "loss": 4.6504, "step": 721100 }, { "epoch": 1.3457456163881885, "grad_norm": 1.0730392932891846, "learning_rate": 0.0001827686201620988, "loss": 4.3844, "step": 721150 }, { "epoch": 1.3458389219475668, "grad_norm": 1.316937804222107, "learning_rate": 0.0001827662625355983, "loss": 4.5691, "step": 721200 }, { "epoch": 1.3459322275069454, "grad_norm": 1.662070631980896, "learning_rate": 0.0001827639047630299, "loss": 4.6441, "step": 721250 }, { "epoch": 1.346025533066324, "grad_norm": 1.1710325479507446, "learning_rate": 0.00018276154684439776, "loss": 4.4571, "step": 721300 }, { "epoch": 1.3461188386257024, "grad_norm": 1.300044298171997, "learning_rate": 0.0001827591887797061, "loss": 4.4586, "step": 721350 }, { "epoch": 1.346212144185081, "grad_norm": 1.280794382095337, "learning_rate": 0.000182756830568959, "loss": 4.3692, "step": 721400 }, { "epoch": 1.3463054497444595, "grad_norm": 1.3169435262680054, "learning_rate": 0.00018275447221216066, "loss": 4.6516, "step": 721450 }, { "epoch": 1.346398755303838, "grad_norm": 1.0449484586715698, "learning_rate": 0.0001827521137093152, "loss": 4.3382, "step": 721500 }, { "epoch": 1.3464920608632163, "grad_norm": 0.9752941727638245, "learning_rate": 0.00018274975506042687, "loss": 4.5821, "step": 721550 }, { "epoch": 1.3465853664225949, "grad_norm": 1.2695879936218262, "learning_rate": 0.00018274739626549977, "loss": 4.403, "step": 721600 }, { "epoch": 1.3466786719819734, "grad_norm": 1.35903799533844, "learning_rate": 0.00018274503732453807, "loss": 4.3418, "step": 721650 }, { "epoch": 1.3467719775413518, "grad_norm": 1.5047234296798706, "learning_rate": 0.00018274267823754597, "loss": 4.4901, "step": 721700 }, { "epoch": 1.3468652831007304, "grad_norm": 1.065405011177063, "learning_rate": 0.00018274031900452755, "loss": 4.5053, "step": 721750 }, { "epoch": 1.346958588660109, "grad_norm": 1.144824743270874, "learning_rate": 0.00018273795962548705, "loss": 4.6431, "step": 721800 }, { "epoch": 1.3470518942194873, "grad_norm": 1.1677567958831787, "learning_rate": 0.0001827356001004286, "loss": 4.2934, "step": 721850 }, { "epoch": 1.3471451997788657, "grad_norm": 1.5927538871765137, "learning_rate": 0.00018273324042935638, "loss": 4.4666, "step": 721900 }, { "epoch": 1.3472385053382443, "grad_norm": 1.3598376512527466, "learning_rate": 0.00018273088061227454, "loss": 4.6821, "step": 721950 }, { "epoch": 1.3473318108976229, "grad_norm": 1.086592197418213, "learning_rate": 0.00018272852064918729, "loss": 4.5574, "step": 722000 }, { "epoch": 1.3474251164570012, "grad_norm": 1.0438820123672485, "learning_rate": 0.0001827261605400987, "loss": 4.4655, "step": 722050 }, { "epoch": 1.3475184220163798, "grad_norm": 1.3114039897918701, "learning_rate": 0.00018272380028501304, "loss": 4.418, "step": 722100 }, { "epoch": 1.3476117275757582, "grad_norm": 1.335744857788086, "learning_rate": 0.0001827214398839344, "loss": 4.5327, "step": 722150 }, { "epoch": 1.3477050331351368, "grad_norm": 1.1496973037719727, "learning_rate": 0.000182719079336867, "loss": 4.5893, "step": 722200 }, { "epoch": 1.3477983386945152, "grad_norm": 1.025964617729187, "learning_rate": 0.00018271671864381498, "loss": 4.601, "step": 722250 }, { "epoch": 1.3478916442538937, "grad_norm": 0.898279070854187, "learning_rate": 0.00018271435780478247, "loss": 4.5502, "step": 722300 }, { "epoch": 1.3479849498132723, "grad_norm": 1.2886511087417603, "learning_rate": 0.0001827119968197737, "loss": 4.5249, "step": 722350 }, { "epoch": 1.3480782553726507, "grad_norm": 1.2058513164520264, "learning_rate": 0.00018270963568879279, "loss": 4.5389, "step": 722400 }, { "epoch": 1.3481715609320293, "grad_norm": 1.46294367313385, "learning_rate": 0.00018270727441184394, "loss": 4.7874, "step": 722450 }, { "epoch": 1.3482648664914076, "grad_norm": 1.1921461820602417, "learning_rate": 0.0001827049129889313, "loss": 4.7026, "step": 722500 }, { "epoch": 1.3483581720507862, "grad_norm": 1.0609911680221558, "learning_rate": 0.00018270255142005906, "loss": 4.5512, "step": 722550 }, { "epoch": 1.3484514776101646, "grad_norm": 1.2498924732208252, "learning_rate": 0.00018270018970523134, "loss": 4.4938, "step": 722600 }, { "epoch": 1.3485447831695432, "grad_norm": 1.26821768283844, "learning_rate": 0.00018269782784445236, "loss": 4.4233, "step": 722650 }, { "epoch": 1.3486380887289218, "grad_norm": 1.300122618675232, "learning_rate": 0.00018269546583772622, "loss": 4.4757, "step": 722700 }, { "epoch": 1.3487313942883001, "grad_norm": 1.2033895254135132, "learning_rate": 0.00018269310368505717, "loss": 4.5827, "step": 722750 }, { "epoch": 1.3488246998476787, "grad_norm": 1.0281646251678467, "learning_rate": 0.00018269074138644936, "loss": 4.3585, "step": 722800 }, { "epoch": 1.348918005407057, "grad_norm": 1.0636831521987915, "learning_rate": 0.0001826883789419069, "loss": 4.5264, "step": 722850 }, { "epoch": 1.3490113109664357, "grad_norm": 0.9558767676353455, "learning_rate": 0.00018268601635143403, "loss": 4.5066, "step": 722900 }, { "epoch": 1.349104616525814, "grad_norm": 1.1011196374893188, "learning_rate": 0.00018268365361503487, "loss": 4.5949, "step": 722950 }, { "epoch": 1.3491979220851926, "grad_norm": 1.5195274353027344, "learning_rate": 0.00018268129073271364, "loss": 4.4197, "step": 723000 }, { "epoch": 1.3491979220851926, "eval_loss": 4.670181751251221, "eval_runtime": 230.3031, "eval_samples_per_second": 11.324, "eval_steps_per_second": 11.324, "eval_tts_loss": 7.645632944474042, "step": 723000 }, { "epoch": 1.3492912276445712, "grad_norm": 1.313236951828003, "learning_rate": 0.00018267892770447444, "loss": 4.699, "step": 723050 }, { "epoch": 1.3493845332039496, "grad_norm": 1.4705567359924316, "learning_rate": 0.00018267656453032152, "loss": 4.7368, "step": 723100 }, { "epoch": 1.3494778387633282, "grad_norm": 1.2146289348602295, "learning_rate": 0.000182674201210259, "loss": 4.5518, "step": 723150 }, { "epoch": 1.3495711443227065, "grad_norm": 1.2895835638046265, "learning_rate": 0.00018267183774429104, "loss": 4.5123, "step": 723200 }, { "epoch": 1.3496644498820851, "grad_norm": 1.190589189529419, "learning_rate": 0.00018266947413242182, "loss": 4.4879, "step": 723250 }, { "epoch": 1.3497577554414635, "grad_norm": 1.2966797351837158, "learning_rate": 0.0001826671103746556, "loss": 4.61, "step": 723300 }, { "epoch": 1.349851061000842, "grad_norm": 1.308138370513916, "learning_rate": 0.0001826647464709964, "loss": 4.5361, "step": 723350 }, { "epoch": 1.3499443665602207, "grad_norm": 1.0537230968475342, "learning_rate": 0.0001826623824214485, "loss": 4.6055, "step": 723400 }, { "epoch": 1.350037672119599, "grad_norm": 0.8806184530258179, "learning_rate": 0.00018266001822601604, "loss": 4.4843, "step": 723450 }, { "epoch": 1.3501309776789776, "grad_norm": 1.0151857137680054, "learning_rate": 0.00018265765388470316, "loss": 4.3862, "step": 723500 }, { "epoch": 1.350224283238356, "grad_norm": 2.278646230697632, "learning_rate": 0.0001826552893975141, "loss": 4.5827, "step": 723550 }, { "epoch": 1.3503175887977346, "grad_norm": 0.8076476454734802, "learning_rate": 0.00018265292476445303, "loss": 4.2966, "step": 723600 }, { "epoch": 1.350410894357113, "grad_norm": 1.1485521793365479, "learning_rate": 0.00018265055998552404, "loss": 4.5599, "step": 723650 }, { "epoch": 1.3505041999164915, "grad_norm": 1.187851071357727, "learning_rate": 0.00018264819506073136, "loss": 4.4713, "step": 723700 }, { "epoch": 1.35059750547587, "grad_norm": 1.336297869682312, "learning_rate": 0.00018264582999007916, "loss": 4.4377, "step": 723750 }, { "epoch": 1.3506908110352485, "grad_norm": 1.2064104080200195, "learning_rate": 0.00018264346477357162, "loss": 4.557, "step": 723800 }, { "epoch": 1.350784116594627, "grad_norm": 1.4891018867492676, "learning_rate": 0.0001826410994112129, "loss": 4.611, "step": 723850 }, { "epoch": 1.3508774221540054, "grad_norm": 0.8795645833015442, "learning_rate": 0.00018263873390300715, "loss": 4.7068, "step": 723900 }, { "epoch": 1.350970727713384, "grad_norm": 1.338773250579834, "learning_rate": 0.00018263636824895863, "loss": 4.4928, "step": 723950 }, { "epoch": 1.3510640332727624, "grad_norm": 1.2526360750198364, "learning_rate": 0.00018263400244907145, "loss": 4.4438, "step": 724000 }, { "epoch": 1.351157338832141, "grad_norm": 1.066449761390686, "learning_rate": 0.00018263163650334978, "loss": 4.54, "step": 724050 }, { "epoch": 1.3512506443915195, "grad_norm": 0.9963470697402954, "learning_rate": 0.00018262927041179781, "loss": 4.4403, "step": 724100 }, { "epoch": 1.351343949950898, "grad_norm": 1.1210556030273438, "learning_rate": 0.0001826269041744197, "loss": 4.5798, "step": 724150 }, { "epoch": 1.3514372555102765, "grad_norm": 1.3287009000778198, "learning_rate": 0.0001826245377912197, "loss": 4.6243, "step": 724200 }, { "epoch": 1.3515305610696549, "grad_norm": 1.0786824226379395, "learning_rate": 0.00018262217126220186, "loss": 4.4271, "step": 724250 }, { "epoch": 1.3516238666290334, "grad_norm": 1.0540883541107178, "learning_rate": 0.00018261980458737046, "loss": 4.6066, "step": 724300 }, { "epoch": 1.3517171721884118, "grad_norm": 1.4414379596710205, "learning_rate": 0.00018261743776672965, "loss": 4.5434, "step": 724350 }, { "epoch": 1.3518104777477904, "grad_norm": 1.1740529537200928, "learning_rate": 0.00018261507080028362, "loss": 4.4147, "step": 724400 }, { "epoch": 1.351903783307169, "grad_norm": 1.262338399887085, "learning_rate": 0.00018261270368803646, "loss": 4.3652, "step": 724450 }, { "epoch": 1.3519970888665473, "grad_norm": 1.1080259084701538, "learning_rate": 0.00018261033642999247, "loss": 4.4489, "step": 724500 }, { "epoch": 1.352090394425926, "grad_norm": 1.2746515274047852, "learning_rate": 0.00018260796902615575, "loss": 4.4096, "step": 724550 }, { "epoch": 1.3521836999853043, "grad_norm": 1.2748832702636719, "learning_rate": 0.0001826056014765305, "loss": 4.6738, "step": 724600 }, { "epoch": 1.3522770055446829, "grad_norm": 1.1504278182983398, "learning_rate": 0.0001826032337811209, "loss": 4.6309, "step": 724650 }, { "epoch": 1.3523703111040613, "grad_norm": 1.3853473663330078, "learning_rate": 0.00018260086593993116, "loss": 4.4725, "step": 724700 }, { "epoch": 1.3524636166634398, "grad_norm": 1.5016512870788574, "learning_rate": 0.00018259849795296536, "loss": 4.4781, "step": 724750 }, { "epoch": 1.3525569222228184, "grad_norm": 0.7780452370643616, "learning_rate": 0.0001825961298202278, "loss": 4.4677, "step": 724800 }, { "epoch": 1.3526502277821968, "grad_norm": 1.732835292816162, "learning_rate": 0.00018259376154172258, "loss": 4.6435, "step": 724850 }, { "epoch": 1.3527435333415754, "grad_norm": 1.5641696453094482, "learning_rate": 0.00018259139311745391, "loss": 4.6399, "step": 724900 }, { "epoch": 1.3528368389009537, "grad_norm": 1.2362728118896484, "learning_rate": 0.00018258902454742595, "loss": 4.6044, "step": 724950 }, { "epoch": 1.3529301444603323, "grad_norm": 1.128363013267517, "learning_rate": 0.00018258665583164292, "loss": 4.4526, "step": 725000 }, { "epoch": 1.3530234500197107, "grad_norm": 1.210057020187378, "learning_rate": 0.00018258428697010896, "loss": 4.4197, "step": 725050 }, { "epoch": 1.3531167555790893, "grad_norm": 0.9356469511985779, "learning_rate": 0.00018258191796282827, "loss": 4.5634, "step": 725100 }, { "epoch": 1.3532100611384679, "grad_norm": 1.1539859771728516, "learning_rate": 0.000182579548809805, "loss": 4.3039, "step": 725150 }, { "epoch": 1.3533033666978462, "grad_norm": 1.0428646802902222, "learning_rate": 0.00018257717951104336, "loss": 4.418, "step": 725200 }, { "epoch": 1.3533966722572248, "grad_norm": 1.0285353660583496, "learning_rate": 0.00018257481006654755, "loss": 4.5983, "step": 725250 }, { "epoch": 1.3534899778166032, "grad_norm": 1.6851294040679932, "learning_rate": 0.00018257244047632174, "loss": 4.622, "step": 725300 }, { "epoch": 1.3535832833759818, "grad_norm": 0.9363089799880981, "learning_rate": 0.00018257007074037009, "loss": 4.4463, "step": 725350 }, { "epoch": 1.3536765889353601, "grad_norm": 1.290359377861023, "learning_rate": 0.00018256770085869678, "loss": 4.4239, "step": 725400 }, { "epoch": 1.3537698944947387, "grad_norm": 1.1298034191131592, "learning_rate": 0.000182565330831306, "loss": 4.4891, "step": 725450 }, { "epoch": 1.3538632000541173, "grad_norm": 1.170775294303894, "learning_rate": 0.00018256296065820196, "loss": 4.62, "step": 725500 }, { "epoch": 1.3539565056134957, "grad_norm": 1.1958184242248535, "learning_rate": 0.00018256059033938882, "loss": 4.3451, "step": 725550 }, { "epoch": 1.3540498111728743, "grad_norm": 0.8711984753608704, "learning_rate": 0.00018255821987487076, "loss": 4.5208, "step": 725600 }, { "epoch": 1.3541431167322526, "grad_norm": 1.2064021825790405, "learning_rate": 0.00018255584926465197, "loss": 4.5517, "step": 725650 }, { "epoch": 1.3542364222916312, "grad_norm": 1.156962275505066, "learning_rate": 0.00018255347850873663, "loss": 4.5327, "step": 725700 }, { "epoch": 1.3543297278510096, "grad_norm": 1.2961362600326538, "learning_rate": 0.00018255110760712892, "loss": 4.4689, "step": 725750 }, { "epoch": 1.3544230334103882, "grad_norm": 1.1533552408218384, "learning_rate": 0.00018254873655983305, "loss": 4.6812, "step": 725800 }, { "epoch": 1.3545163389697668, "grad_norm": 1.0160434246063232, "learning_rate": 0.00018254636536685316, "loss": 4.4978, "step": 725850 }, { "epoch": 1.3546096445291451, "grad_norm": 1.137358546257019, "learning_rate": 0.0001825439940281935, "loss": 4.5476, "step": 725900 }, { "epoch": 1.3547029500885237, "grad_norm": 0.9772936701774597, "learning_rate": 0.00018254162254385817, "loss": 4.4017, "step": 725950 }, { "epoch": 1.354796255647902, "grad_norm": 1.0919733047485352, "learning_rate": 0.0001825392509138514, "loss": 4.5805, "step": 726000 }, { "epoch": 1.354796255647902, "eval_loss": 4.686704635620117, "eval_runtime": 230.5119, "eval_samples_per_second": 11.314, "eval_steps_per_second": 11.314, "eval_tts_loss": 7.649469767691442, "step": 726000 }, { "epoch": 1.3548895612072807, "grad_norm": 1.0449957847595215, "learning_rate": 0.00018253687913817742, "loss": 4.5025, "step": 726050 }, { "epoch": 1.354982866766659, "grad_norm": 1.2512603998184204, "learning_rate": 0.00018253450721684034, "loss": 4.5342, "step": 726100 }, { "epoch": 1.3550761723260376, "grad_norm": 1.2292732000350952, "learning_rate": 0.00018253213514984437, "loss": 4.5542, "step": 726150 }, { "epoch": 1.3551694778854162, "grad_norm": 1.4540189504623413, "learning_rate": 0.00018252976293719371, "loss": 4.5101, "step": 726200 }, { "epoch": 1.3552627834447946, "grad_norm": 1.1732397079467773, "learning_rate": 0.00018252739057889253, "loss": 4.3427, "step": 726250 }, { "epoch": 1.3553560890041731, "grad_norm": 1.1799020767211914, "learning_rate": 0.00018252501807494506, "loss": 4.6075, "step": 726300 }, { "epoch": 1.3554493945635515, "grad_norm": 1.3092387914657593, "learning_rate": 0.00018252264542535543, "loss": 4.5032, "step": 726350 }, { "epoch": 1.35554270012293, "grad_norm": 1.3461332321166992, "learning_rate": 0.00018252027263012784, "loss": 4.4391, "step": 726400 }, { "epoch": 1.3556360056823085, "grad_norm": 1.0088399648666382, "learning_rate": 0.00018251789968926652, "loss": 4.4977, "step": 726450 }, { "epoch": 1.355729311241687, "grad_norm": 1.0882623195648193, "learning_rate": 0.0001825155266027756, "loss": 4.3636, "step": 726500 }, { "epoch": 1.3558226168010656, "grad_norm": 1.3299607038497925, "learning_rate": 0.0001825131533706593, "loss": 4.5698, "step": 726550 }, { "epoch": 1.355915922360444, "grad_norm": 1.2337397336959839, "learning_rate": 0.0001825107799929218, "loss": 4.475, "step": 726600 }, { "epoch": 1.3560092279198226, "grad_norm": 1.2088146209716797, "learning_rate": 0.00018250840646956732, "loss": 4.4241, "step": 726650 }, { "epoch": 1.356102533479201, "grad_norm": 1.0898922681808472, "learning_rate": 0.0001825060328006, "loss": 4.5702, "step": 726700 }, { "epoch": 1.3561958390385795, "grad_norm": 1.1384861469268799, "learning_rate": 0.00018250365898602407, "loss": 4.4543, "step": 726750 }, { "epoch": 1.356289144597958, "grad_norm": 0.747077465057373, "learning_rate": 0.00018250128502584367, "loss": 4.4654, "step": 726800 }, { "epoch": 1.3563824501573365, "grad_norm": 1.4521085023880005, "learning_rate": 0.00018249891092006303, "loss": 4.4471, "step": 726850 }, { "epoch": 1.356475755716715, "grad_norm": 1.1638509035110474, "learning_rate": 0.0001824965366686863, "loss": 4.3767, "step": 726900 }, { "epoch": 1.3565690612760934, "grad_norm": 1.1984643936157227, "learning_rate": 0.00018249416227171774, "loss": 4.6619, "step": 726950 }, { "epoch": 1.356662366835472, "grad_norm": 1.318793773651123, "learning_rate": 0.00018249178772916148, "loss": 4.5775, "step": 727000 }, { "epoch": 1.3567556723948504, "grad_norm": 1.3447638750076294, "learning_rate": 0.00018248941304102175, "loss": 4.5667, "step": 727050 }, { "epoch": 1.356848977954229, "grad_norm": 1.132716178894043, "learning_rate": 0.0001824870382073027, "loss": 4.5949, "step": 727100 }, { "epoch": 1.3569422835136074, "grad_norm": 1.0300718545913696, "learning_rate": 0.00018248466322800852, "loss": 4.6024, "step": 727150 }, { "epoch": 1.357035589072986, "grad_norm": 1.219099521636963, "learning_rate": 0.00018248228810314348, "loss": 4.3783, "step": 727200 }, { "epoch": 1.3571288946323645, "grad_norm": 0.9341549873352051, "learning_rate": 0.00018247991283271167, "loss": 4.4717, "step": 727250 }, { "epoch": 1.357222200191743, "grad_norm": 1.1375728845596313, "learning_rate": 0.0001824775374167173, "loss": 4.3835, "step": 727300 }, { "epoch": 1.3573155057511215, "grad_norm": 1.2401444911956787, "learning_rate": 0.00018247516185516464, "loss": 4.5951, "step": 727350 }, { "epoch": 1.3574088113104998, "grad_norm": 1.0372564792633057, "learning_rate": 0.00018247278614805782, "loss": 4.5962, "step": 727400 }, { "epoch": 1.3575021168698784, "grad_norm": 0.8566879630088806, "learning_rate": 0.00018247041029540104, "loss": 4.6095, "step": 727450 }, { "epoch": 1.3575954224292568, "grad_norm": 1.5325331687927246, "learning_rate": 0.0001824680342971985, "loss": 4.3354, "step": 727500 }, { "epoch": 1.3576887279886354, "grad_norm": 0.9921556115150452, "learning_rate": 0.0001824656581534544, "loss": 4.5277, "step": 727550 }, { "epoch": 1.357782033548014, "grad_norm": 0.844407320022583, "learning_rate": 0.00018246328186417293, "loss": 4.4516, "step": 727600 }, { "epoch": 1.3578753391073923, "grad_norm": 1.2577937841415405, "learning_rate": 0.00018246090542935824, "loss": 4.6595, "step": 727650 }, { "epoch": 1.357968644666771, "grad_norm": 1.1597188711166382, "learning_rate": 0.0001824585288490146, "loss": 4.5986, "step": 727700 }, { "epoch": 1.3580619502261493, "grad_norm": 1.2254786491394043, "learning_rate": 0.00018245615212314614, "loss": 4.444, "step": 727750 }, { "epoch": 1.3581552557855279, "grad_norm": 1.0827590227127075, "learning_rate": 0.0001824537752517571, "loss": 4.6735, "step": 727800 }, { "epoch": 1.3582485613449062, "grad_norm": 1.02262282371521, "learning_rate": 0.00018245139823485165, "loss": 4.4334, "step": 727850 }, { "epoch": 1.3583418669042848, "grad_norm": 1.3388478755950928, "learning_rate": 0.00018244902107243397, "loss": 4.6956, "step": 727900 }, { "epoch": 1.3584351724636634, "grad_norm": 1.1750307083129883, "learning_rate": 0.00018244664376450826, "loss": 4.6754, "step": 727950 }, { "epoch": 1.3585284780230418, "grad_norm": 1.2880897521972656, "learning_rate": 0.00018244426631107877, "loss": 4.3589, "step": 728000 }, { "epoch": 1.3586217835824204, "grad_norm": 1.1455796957015991, "learning_rate": 0.00018244188871214964, "loss": 4.5224, "step": 728050 }, { "epoch": 1.3587150891417987, "grad_norm": 1.129857063293457, "learning_rate": 0.0001824395109677251, "loss": 4.4286, "step": 728100 }, { "epoch": 1.3588083947011773, "grad_norm": 1.101718544960022, "learning_rate": 0.00018243713307780933, "loss": 4.5436, "step": 728150 }, { "epoch": 1.3589017002605557, "grad_norm": 1.2696237564086914, "learning_rate": 0.0001824347550424065, "loss": 4.4254, "step": 728200 }, { "epoch": 1.3589950058199343, "grad_norm": 1.2697737216949463, "learning_rate": 0.00018243237686152084, "loss": 4.62, "step": 728250 }, { "epoch": 1.3590883113793129, "grad_norm": 1.2402477264404297, "learning_rate": 0.00018242999853515654, "loss": 4.5548, "step": 728300 }, { "epoch": 1.3591816169386912, "grad_norm": 0.9545737504959106, "learning_rate": 0.00018242762006331778, "loss": 4.3565, "step": 728350 }, { "epoch": 1.3592749224980698, "grad_norm": 1.2260619401931763, "learning_rate": 0.0001824252414460088, "loss": 4.4508, "step": 728400 }, { "epoch": 1.3593682280574482, "grad_norm": 1.2439419031143188, "learning_rate": 0.00018242286268323378, "loss": 4.5783, "step": 728450 }, { "epoch": 1.3594615336168268, "grad_norm": 1.2133325338363647, "learning_rate": 0.0001824204837749969, "loss": 4.5589, "step": 728500 }, { "epoch": 1.3595548391762051, "grad_norm": 1.157065510749817, "learning_rate": 0.00018241810472130235, "loss": 4.5201, "step": 728550 }, { "epoch": 1.3596481447355837, "grad_norm": 1.2336939573287964, "learning_rate": 0.00018241572552215438, "loss": 4.4243, "step": 728600 }, { "epoch": 1.3597414502949623, "grad_norm": 1.316603183746338, "learning_rate": 0.00018241334617755714, "loss": 4.6525, "step": 728650 }, { "epoch": 1.3598347558543407, "grad_norm": 1.0523053407669067, "learning_rate": 0.00018241096668751485, "loss": 4.5497, "step": 728700 }, { "epoch": 1.3599280614137192, "grad_norm": 1.0533334016799927, "learning_rate": 0.0001824085870520317, "loss": 4.392, "step": 728750 }, { "epoch": 1.3600213669730976, "grad_norm": 1.3165284395217896, "learning_rate": 0.00018240620727111187, "loss": 4.4828, "step": 728800 }, { "epoch": 1.3601146725324762, "grad_norm": 1.009934663772583, "learning_rate": 0.0001824038273447596, "loss": 4.6086, "step": 728850 }, { "epoch": 1.3602079780918546, "grad_norm": 1.2932511568069458, "learning_rate": 0.0001824014472729791, "loss": 4.5521, "step": 728900 }, { "epoch": 1.3603012836512332, "grad_norm": 1.2168806791305542, "learning_rate": 0.0001823990670557745, "loss": 4.437, "step": 728950 }, { "epoch": 1.3603945892106117, "grad_norm": 1.1743770837783813, "learning_rate": 0.00018239668669315008, "loss": 4.3349, "step": 729000 }, { "epoch": 1.3603945892106117, "eval_loss": 4.66847562789917, "eval_runtime": 230.6383, "eval_samples_per_second": 11.308, "eval_steps_per_second": 11.308, "eval_tts_loss": 7.626812305747092, "step": 729000 }, { "epoch": 1.36048789476999, "grad_norm": 1.1642770767211914, "learning_rate": 0.00018239430618511, "loss": 4.3456, "step": 729050 }, { "epoch": 1.3605812003293687, "grad_norm": 1.300085425376892, "learning_rate": 0.00018239192553165846, "loss": 4.6234, "step": 729100 }, { "epoch": 1.360674505888747, "grad_norm": 1.2443251609802246, "learning_rate": 0.00018238954473279967, "loss": 4.3937, "step": 729150 }, { "epoch": 1.3607678114481256, "grad_norm": 1.1274253129959106, "learning_rate": 0.00018238716378853783, "loss": 4.4831, "step": 729200 }, { "epoch": 1.360861117007504, "grad_norm": 1.3472120761871338, "learning_rate": 0.00018238478269887715, "loss": 4.6206, "step": 729250 }, { "epoch": 1.3609544225668826, "grad_norm": 1.6040093898773193, "learning_rate": 0.00018238240146382182, "loss": 4.4765, "step": 729300 }, { "epoch": 1.3610477281262612, "grad_norm": 1.0231380462646484, "learning_rate": 0.00018238002008337601, "loss": 4.4035, "step": 729350 }, { "epoch": 1.3611410336856395, "grad_norm": 1.0706677436828613, "learning_rate": 0.000182377638557544, "loss": 4.5563, "step": 729400 }, { "epoch": 1.3612343392450181, "grad_norm": 1.1838470697402954, "learning_rate": 0.00018237525688632997, "loss": 4.723, "step": 729450 }, { "epoch": 1.3613276448043965, "grad_norm": 1.0312621593475342, "learning_rate": 0.0001823728750697381, "loss": 4.5489, "step": 729500 }, { "epoch": 1.361420950363775, "grad_norm": 1.2169641256332397, "learning_rate": 0.00018237049310777258, "loss": 4.4475, "step": 729550 }, { "epoch": 1.3615142559231534, "grad_norm": 1.3004813194274902, "learning_rate": 0.00018236811100043763, "loss": 4.5784, "step": 729600 }, { "epoch": 1.361607561482532, "grad_norm": 1.1826603412628174, "learning_rate": 0.00018236572874773748, "loss": 4.6416, "step": 729650 }, { "epoch": 1.3617008670419106, "grad_norm": 1.1159125566482544, "learning_rate": 0.0001823633463496763, "loss": 4.4653, "step": 729700 }, { "epoch": 1.361794172601289, "grad_norm": 1.274246096611023, "learning_rate": 0.00018236096380625825, "loss": 4.6362, "step": 729750 }, { "epoch": 1.3618874781606676, "grad_norm": 1.308957576751709, "learning_rate": 0.00018235858111748766, "loss": 4.632, "step": 729800 }, { "epoch": 1.361980783720046, "grad_norm": 0.82074373960495, "learning_rate": 0.00018235619828336865, "loss": 4.537, "step": 729850 }, { "epoch": 1.3620740892794245, "grad_norm": 1.4565922021865845, "learning_rate": 0.00018235381530390545, "loss": 4.4238, "step": 729900 }, { "epoch": 1.362167394838803, "grad_norm": 1.220731258392334, "learning_rate": 0.00018235143217910224, "loss": 4.3162, "step": 729950 }, { "epoch": 1.3622607003981815, "grad_norm": 1.1985918283462524, "learning_rate": 0.00018234904890896325, "loss": 4.6271, "step": 730000 }, { "epoch": 1.36235400595756, "grad_norm": 1.13936185836792, "learning_rate": 0.00018234666549349268, "loss": 4.4818, "step": 730050 }, { "epoch": 1.3624473115169384, "grad_norm": 1.2406173944473267, "learning_rate": 0.00018234428193269472, "loss": 4.3713, "step": 730100 }, { "epoch": 1.362540617076317, "grad_norm": 1.3027818202972412, "learning_rate": 0.0001823418982265736, "loss": 4.6688, "step": 730150 }, { "epoch": 1.3626339226356954, "grad_norm": 1.2941128015518188, "learning_rate": 0.00018233951437513355, "loss": 4.5257, "step": 730200 }, { "epoch": 1.362727228195074, "grad_norm": 0.8767145872116089, "learning_rate": 0.0001823371303783787, "loss": 4.5372, "step": 730250 }, { "epoch": 1.3628205337544523, "grad_norm": 1.0703469514846802, "learning_rate": 0.00018233474623631334, "loss": 4.4613, "step": 730300 }, { "epoch": 1.362913839313831, "grad_norm": 1.3578933477401733, "learning_rate": 0.00018233236194894158, "loss": 4.2865, "step": 730350 }, { "epoch": 1.3630071448732095, "grad_norm": 1.4024401903152466, "learning_rate": 0.00018232997751626773, "loss": 4.7249, "step": 730400 }, { "epoch": 1.3631004504325879, "grad_norm": 1.3123254776000977, "learning_rate": 0.000182327592938296, "loss": 4.5191, "step": 730450 }, { "epoch": 1.3631937559919665, "grad_norm": 1.2045743465423584, "learning_rate": 0.00018232520821503048, "loss": 4.3855, "step": 730500 }, { "epoch": 1.3632870615513448, "grad_norm": 1.1473578214645386, "learning_rate": 0.00018232282334647545, "loss": 4.5056, "step": 730550 }, { "epoch": 1.3633803671107234, "grad_norm": 1.0342416763305664, "learning_rate": 0.00018232043833263517, "loss": 4.5435, "step": 730600 }, { "epoch": 1.3634736726701018, "grad_norm": 0.7952665090560913, "learning_rate": 0.00018231805317351377, "loss": 4.3825, "step": 730650 }, { "epoch": 1.3635669782294804, "grad_norm": 1.1602782011032104, "learning_rate": 0.0001823156678691155, "loss": 4.4385, "step": 730700 }, { "epoch": 1.363660283788859, "grad_norm": 0.9990757703781128, "learning_rate": 0.00018231328241944458, "loss": 4.3257, "step": 730750 }, { "epoch": 1.3637535893482373, "grad_norm": 1.1286808252334595, "learning_rate": 0.00018231089682450518, "loss": 4.5095, "step": 730800 }, { "epoch": 1.363846894907616, "grad_norm": 1.281816005706787, "learning_rate": 0.00018230851108430154, "loss": 4.7156, "step": 730850 }, { "epoch": 1.3639402004669943, "grad_norm": 1.0494389533996582, "learning_rate": 0.00018230612519883784, "loss": 4.7279, "step": 730900 }, { "epoch": 1.3640335060263729, "grad_norm": 1.154935598373413, "learning_rate": 0.00018230373916811835, "loss": 4.4464, "step": 730950 }, { "epoch": 1.3641268115857512, "grad_norm": 1.23104989528656, "learning_rate": 0.0001823013529921472, "loss": 4.5386, "step": 731000 }, { "epoch": 1.3642201171451298, "grad_norm": 1.140464186668396, "learning_rate": 0.00018229896667092867, "loss": 4.4968, "step": 731050 }, { "epoch": 1.3643134227045084, "grad_norm": 1.199709415435791, "learning_rate": 0.00018229658020446694, "loss": 4.4738, "step": 731100 }, { "epoch": 1.3644067282638868, "grad_norm": 0.9797897338867188, "learning_rate": 0.00018229419359276622, "loss": 4.227, "step": 731150 }, { "epoch": 1.3645000338232653, "grad_norm": 1.1550041437149048, "learning_rate": 0.0001822918068358307, "loss": 4.4992, "step": 731200 }, { "epoch": 1.3645933393826437, "grad_norm": 1.1756529808044434, "learning_rate": 0.00018228941993366467, "loss": 4.4081, "step": 731250 }, { "epoch": 1.3646866449420223, "grad_norm": 1.0971152782440186, "learning_rate": 0.0001822870328862723, "loss": 4.42, "step": 731300 }, { "epoch": 1.3647799505014007, "grad_norm": 1.4491357803344727, "learning_rate": 0.00018228464569365775, "loss": 4.4607, "step": 731350 }, { "epoch": 1.3648732560607792, "grad_norm": 1.3413618803024292, "learning_rate": 0.0001822822583558253, "loss": 4.5274, "step": 731400 }, { "epoch": 1.3649665616201578, "grad_norm": 1.4833455085754395, "learning_rate": 0.00018227987087277914, "loss": 4.5961, "step": 731450 }, { "epoch": 1.3650598671795362, "grad_norm": 1.1869001388549805, "learning_rate": 0.0001822774832445235, "loss": 4.5128, "step": 731500 }, { "epoch": 1.3651531727389148, "grad_norm": 1.0004557371139526, "learning_rate": 0.00018227509547106256, "loss": 4.3903, "step": 731550 }, { "epoch": 1.3652464782982932, "grad_norm": 1.7397109270095825, "learning_rate": 0.00018227270755240055, "loss": 4.5219, "step": 731600 }, { "epoch": 1.3653397838576717, "grad_norm": 1.199029564857483, "learning_rate": 0.0001822703194885417, "loss": 4.3, "step": 731650 }, { "epoch": 1.36543308941705, "grad_norm": 0.9370952844619751, "learning_rate": 0.0001822679312794902, "loss": 4.5006, "step": 731700 }, { "epoch": 1.3655263949764287, "grad_norm": 1.0914407968521118, "learning_rate": 0.0001822655429252503, "loss": 4.5979, "step": 731750 }, { "epoch": 1.3656197005358073, "grad_norm": 1.2877815961837769, "learning_rate": 0.00018226315442582618, "loss": 4.3686, "step": 731800 }, { "epoch": 1.3657130060951856, "grad_norm": 1.27789306640625, "learning_rate": 0.00018226076578122205, "loss": 4.5239, "step": 731850 }, { "epoch": 1.3658063116545642, "grad_norm": 1.2306227684020996, "learning_rate": 0.00018225837699144216, "loss": 4.6439, "step": 731900 }, { "epoch": 1.3658996172139426, "grad_norm": 1.1950677633285522, "learning_rate": 0.00018225598805649072, "loss": 4.3584, "step": 731950 }, { "epoch": 1.3659929227733212, "grad_norm": 1.2591156959533691, "learning_rate": 0.0001822535989763719, "loss": 4.5271, "step": 732000 }, { "epoch": 1.3659929227733212, "eval_loss": 4.668415069580078, "eval_runtime": 233.2707, "eval_samples_per_second": 11.18, "eval_steps_per_second": 11.18, "eval_tts_loss": 7.644703588868417, "step": 732000 }, { "epoch": 1.3660862283326995, "grad_norm": 1.42213773727417, "learning_rate": 0.00018225120975108998, "loss": 4.4851, "step": 732050 }, { "epoch": 1.3661795338920781, "grad_norm": 1.1079912185668945, "learning_rate": 0.0001822488203806491, "loss": 4.5437, "step": 732100 }, { "epoch": 1.3662728394514567, "grad_norm": 1.290531873703003, "learning_rate": 0.0001822464308650536, "loss": 4.5875, "step": 732150 }, { "epoch": 1.366366145010835, "grad_norm": 1.308176040649414, "learning_rate": 0.00018224404120430757, "loss": 4.5711, "step": 732200 }, { "epoch": 1.3664594505702137, "grad_norm": 1.2245643138885498, "learning_rate": 0.0001822416513984153, "loss": 4.3742, "step": 732250 }, { "epoch": 1.366552756129592, "grad_norm": 0.9682570695877075, "learning_rate": 0.000182239261447381, "loss": 4.3468, "step": 732300 }, { "epoch": 1.3666460616889706, "grad_norm": 1.152635097503662, "learning_rate": 0.00018223687135120883, "loss": 4.5308, "step": 732350 }, { "epoch": 1.366739367248349, "grad_norm": 1.1383434534072876, "learning_rate": 0.0001822344811099031, "loss": 4.5845, "step": 732400 }, { "epoch": 1.3668326728077276, "grad_norm": 1.3489753007888794, "learning_rate": 0.00018223209072346794, "loss": 4.5073, "step": 732450 }, { "epoch": 1.3669259783671062, "grad_norm": 1.3288629055023193, "learning_rate": 0.0001822297001919076, "loss": 4.7206, "step": 732500 }, { "epoch": 1.3670192839264845, "grad_norm": 1.481525182723999, "learning_rate": 0.00018222730951522636, "loss": 4.5078, "step": 732550 }, { "epoch": 1.3671125894858631, "grad_norm": 0.9424459338188171, "learning_rate": 0.00018222491869342836, "loss": 4.3036, "step": 732600 }, { "epoch": 1.3672058950452415, "grad_norm": 1.1260565519332886, "learning_rate": 0.00018222252772651784, "loss": 4.5238, "step": 732650 }, { "epoch": 1.36729920060462, "grad_norm": 1.5328397750854492, "learning_rate": 0.000182220136614499, "loss": 4.49, "step": 732700 }, { "epoch": 1.3673925061639984, "grad_norm": 1.1881719827651978, "learning_rate": 0.00018221774535737611, "loss": 4.4513, "step": 732750 }, { "epoch": 1.367485811723377, "grad_norm": 1.2913823127746582, "learning_rate": 0.00018221535395515338, "loss": 4.3919, "step": 732800 }, { "epoch": 1.3675791172827556, "grad_norm": 1.3064284324645996, "learning_rate": 0.00018221296240783502, "loss": 4.6277, "step": 732850 }, { "epoch": 1.367672422842134, "grad_norm": 1.0377432107925415, "learning_rate": 0.00018221057071542526, "loss": 4.7398, "step": 732900 }, { "epoch": 1.3677657284015126, "grad_norm": 1.2785406112670898, "learning_rate": 0.00018220817887792826, "loss": 4.4639, "step": 732950 }, { "epoch": 1.367859033960891, "grad_norm": 1.0735493898391724, "learning_rate": 0.00018220578689534834, "loss": 4.375, "step": 733000 }, { "epoch": 1.3679523395202695, "grad_norm": 1.157127857208252, "learning_rate": 0.00018220339476768965, "loss": 4.4714, "step": 733050 }, { "epoch": 1.3680456450796479, "grad_norm": 1.3886470794677734, "learning_rate": 0.00018220100249495642, "loss": 4.4762, "step": 733100 }, { "epoch": 1.3681389506390265, "grad_norm": 1.205466389656067, "learning_rate": 0.0001821986100771529, "loss": 4.4403, "step": 733150 }, { "epoch": 1.368232256198405, "grad_norm": 1.5568933486938477, "learning_rate": 0.00018219621751428328, "loss": 4.5566, "step": 733200 }, { "epoch": 1.3683255617577834, "grad_norm": 1.115957260131836, "learning_rate": 0.0001821938248063518, "loss": 4.6631, "step": 733250 }, { "epoch": 1.368418867317162, "grad_norm": 1.0834838151931763, "learning_rate": 0.00018219143195336268, "loss": 4.5637, "step": 733300 }, { "epoch": 1.3685121728765404, "grad_norm": 1.4430235624313354, "learning_rate": 0.00018218903895532014, "loss": 4.4983, "step": 733350 }, { "epoch": 1.368605478435919, "grad_norm": 1.1073914766311646, "learning_rate": 0.00018218664581222844, "loss": 4.5102, "step": 733400 }, { "epoch": 1.3686987839952973, "grad_norm": 1.2442829608917236, "learning_rate": 0.00018218425252409175, "loss": 4.4915, "step": 733450 }, { "epoch": 1.368792089554676, "grad_norm": 1.0670374631881714, "learning_rate": 0.00018218185909091433, "loss": 4.3457, "step": 733500 }, { "epoch": 1.3688853951140545, "grad_norm": 1.258642315864563, "learning_rate": 0.00018217946551270036, "loss": 4.6007, "step": 733550 }, { "epoch": 1.3689787006734329, "grad_norm": 1.1165242195129395, "learning_rate": 0.0001821770717894541, "loss": 4.6578, "step": 733600 }, { "epoch": 1.3690720062328114, "grad_norm": 1.1714274883270264, "learning_rate": 0.00018217467792117979, "loss": 4.5056, "step": 733650 }, { "epoch": 1.3691653117921898, "grad_norm": 1.0473724603652954, "learning_rate": 0.0001821722839078816, "loss": 4.6229, "step": 733700 }, { "epoch": 1.3692586173515684, "grad_norm": 1.1497236490249634, "learning_rate": 0.00018216988974956383, "loss": 4.5943, "step": 733750 }, { "epoch": 1.3693519229109468, "grad_norm": 1.0766515731811523, "learning_rate": 0.00018216749544623065, "loss": 4.5195, "step": 733800 }, { "epoch": 1.3694452284703253, "grad_norm": 1.3346211910247803, "learning_rate": 0.00018216510099788627, "loss": 4.5131, "step": 733850 }, { "epoch": 1.369538534029704, "grad_norm": 1.099848985671997, "learning_rate": 0.000182162706404535, "loss": 4.539, "step": 733900 }, { "epoch": 1.3696318395890823, "grad_norm": 0.8915748596191406, "learning_rate": 0.00018216031166618096, "loss": 4.5554, "step": 733950 }, { "epoch": 1.3697251451484609, "grad_norm": 1.230568528175354, "learning_rate": 0.00018215791678282843, "loss": 4.7136, "step": 734000 }, { "epoch": 1.3698184507078393, "grad_norm": 1.334097981452942, "learning_rate": 0.00018215552175448164, "loss": 4.5789, "step": 734050 }, { "epoch": 1.3699117562672178, "grad_norm": 1.3228007555007935, "learning_rate": 0.00018215312658114482, "loss": 4.5074, "step": 734100 }, { "epoch": 1.3700050618265962, "grad_norm": 1.1261327266693115, "learning_rate": 0.00018215073126282217, "loss": 4.6178, "step": 734150 }, { "epoch": 1.3700983673859748, "grad_norm": 1.1997768878936768, "learning_rate": 0.00018214833579951794, "loss": 4.5374, "step": 734200 }, { "epoch": 1.3701916729453534, "grad_norm": 1.0242983102798462, "learning_rate": 0.00018214594019123637, "loss": 4.5537, "step": 734250 }, { "epoch": 1.3702849785047317, "grad_norm": 1.3346383571624756, "learning_rate": 0.00018214354443798166, "loss": 4.5482, "step": 734300 }, { "epoch": 1.3703782840641103, "grad_norm": 1.4824155569076538, "learning_rate": 0.00018214114853975803, "loss": 4.5199, "step": 734350 }, { "epoch": 1.3704715896234887, "grad_norm": 1.1359385251998901, "learning_rate": 0.00018213875249656973, "loss": 4.5174, "step": 734400 }, { "epoch": 1.3705648951828673, "grad_norm": 1.1856859922409058, "learning_rate": 0.00018213635630842098, "loss": 4.4918, "step": 734450 }, { "epoch": 1.3706582007422456, "grad_norm": 1.098907709121704, "learning_rate": 0.00018213395997531605, "loss": 4.5839, "step": 734500 }, { "epoch": 1.3707515063016242, "grad_norm": 1.1827560663223267, "learning_rate": 0.00018213156349725912, "loss": 4.5505, "step": 734550 }, { "epoch": 1.3708448118610028, "grad_norm": 1.0902764797210693, "learning_rate": 0.0001821291668742544, "loss": 4.5593, "step": 734600 }, { "epoch": 1.3709381174203812, "grad_norm": 1.1135292053222656, "learning_rate": 0.00018212677010630616, "loss": 4.6096, "step": 734650 }, { "epoch": 1.3710314229797598, "grad_norm": 1.1513612270355225, "learning_rate": 0.00018212437319341867, "loss": 4.2625, "step": 734700 }, { "epoch": 1.3711247285391381, "grad_norm": 1.0887086391448975, "learning_rate": 0.00018212197613559606, "loss": 4.5665, "step": 734750 }, { "epoch": 1.3712180340985167, "grad_norm": 1.2815632820129395, "learning_rate": 0.00018211957893284262, "loss": 4.6396, "step": 734800 }, { "epoch": 1.371311339657895, "grad_norm": 1.1421213150024414, "learning_rate": 0.0001821171815851626, "loss": 4.3789, "step": 734850 }, { "epoch": 1.3714046452172737, "grad_norm": 1.070897102355957, "learning_rate": 0.00018211478409256015, "loss": 4.4736, "step": 734900 }, { "epoch": 1.3714979507766523, "grad_norm": 1.0784715414047241, "learning_rate": 0.0001821123864550396, "loss": 4.3338, "step": 734950 }, { "epoch": 1.3715912563360306, "grad_norm": 1.1867337226867676, "learning_rate": 0.00018210998867260514, "loss": 4.4801, "step": 735000 }, { "epoch": 1.3715912563360306, "eval_loss": 4.672613620758057, "eval_runtime": 228.5612, "eval_samples_per_second": 11.411, "eval_steps_per_second": 11.411, "eval_tts_loss": 7.723074588286147, "step": 735000 }, { "epoch": 1.3716845618954092, "grad_norm": 1.2326065301895142, "learning_rate": 0.00018210759074526096, "loss": 4.4322, "step": 735050 }, { "epoch": 1.3717778674547876, "grad_norm": 1.2731330394744873, "learning_rate": 0.00018210519267301137, "loss": 4.4227, "step": 735100 }, { "epoch": 1.3718711730141662, "grad_norm": 1.1047251224517822, "learning_rate": 0.00018210279445586054, "loss": 4.5308, "step": 735150 }, { "epoch": 1.3719644785735445, "grad_norm": 1.8172887563705444, "learning_rate": 0.00018210039609381273, "loss": 4.3914, "step": 735200 }, { "epoch": 1.3720577841329231, "grad_norm": 1.0627243518829346, "learning_rate": 0.00018209799758687215, "loss": 4.5464, "step": 735250 }, { "epoch": 1.3721510896923017, "grad_norm": 1.0314337015151978, "learning_rate": 0.00018209559893504307, "loss": 4.5397, "step": 735300 }, { "epoch": 1.37224439525168, "grad_norm": 1.507953405380249, "learning_rate": 0.00018209320013832972, "loss": 4.5296, "step": 735350 }, { "epoch": 1.3723377008110584, "grad_norm": 1.578869342803955, "learning_rate": 0.0001820908011967363, "loss": 4.3408, "step": 735400 }, { "epoch": 1.372431006370437, "grad_norm": 1.3137719631195068, "learning_rate": 0.00018208840211026702, "loss": 4.471, "step": 735450 }, { "epoch": 1.3725243119298156, "grad_norm": 1.042088508605957, "learning_rate": 0.0001820860028789262, "loss": 4.3806, "step": 735500 }, { "epoch": 1.372617617489194, "grad_norm": 1.0040303468704224, "learning_rate": 0.00018208360350271803, "loss": 4.5326, "step": 735550 }, { "epoch": 1.3727109230485726, "grad_norm": 1.135806918144226, "learning_rate": 0.0001820812039816467, "loss": 4.5157, "step": 735600 }, { "epoch": 1.3728042286079511, "grad_norm": 1.1844542026519775, "learning_rate": 0.00018207880431571653, "loss": 4.5915, "step": 735650 }, { "epoch": 1.3728975341673295, "grad_norm": 1.1990412473678589, "learning_rate": 0.0001820764045049317, "loss": 4.5718, "step": 735700 }, { "epoch": 1.3729908397267079, "grad_norm": 1.0717333555221558, "learning_rate": 0.00018207400454929644, "loss": 4.479, "step": 735750 }, { "epoch": 1.3730841452860865, "grad_norm": 1.165457010269165, "learning_rate": 0.00018207160444881504, "loss": 4.4976, "step": 735800 }, { "epoch": 1.373177450845465, "grad_norm": 0.9509486556053162, "learning_rate": 0.00018206920420349166, "loss": 4.5539, "step": 735850 }, { "epoch": 1.3732707564048434, "grad_norm": 1.3006377220153809, "learning_rate": 0.0001820668038133306, "loss": 4.4772, "step": 735900 }, { "epoch": 1.373364061964222, "grad_norm": 1.2724429368972778, "learning_rate": 0.00018206440327833608, "loss": 4.4478, "step": 735950 }, { "epoch": 1.3734573675236006, "grad_norm": 0.8939981460571289, "learning_rate": 0.0001820620025985123, "loss": 4.4227, "step": 736000 }, { "epoch": 1.373550673082979, "grad_norm": 1.38505220413208, "learning_rate": 0.00018205960177386355, "loss": 4.4205, "step": 736050 }, { "epoch": 1.3736439786423573, "grad_norm": 1.3837196826934814, "learning_rate": 0.00018205720080439404, "loss": 4.4934, "step": 736100 }, { "epoch": 1.373737284201736, "grad_norm": 1.208355188369751, "learning_rate": 0.00018205479969010799, "loss": 4.5532, "step": 736150 }, { "epoch": 1.3738305897611145, "grad_norm": 0.999760091304779, "learning_rate": 0.00018205239843100969, "loss": 4.5214, "step": 736200 }, { "epoch": 1.3739238953204929, "grad_norm": 1.1248618364334106, "learning_rate": 0.0001820499970271033, "loss": 4.5712, "step": 736250 }, { "epoch": 1.3740172008798714, "grad_norm": 1.3618665933609009, "learning_rate": 0.00018204759547839312, "loss": 4.7246, "step": 736300 }, { "epoch": 1.37411050643925, "grad_norm": 1.2529972791671753, "learning_rate": 0.00018204519378488339, "loss": 4.556, "step": 736350 }, { "epoch": 1.3742038119986284, "grad_norm": 1.227560043334961, "learning_rate": 0.00018204279194657832, "loss": 4.6398, "step": 736400 }, { "epoch": 1.3742971175580068, "grad_norm": 1.0459915399551392, "learning_rate": 0.00018204038996348212, "loss": 4.6243, "step": 736450 }, { "epoch": 1.3743904231173854, "grad_norm": 1.2024273872375488, "learning_rate": 0.00018203798783559912, "loss": 4.4545, "step": 736500 }, { "epoch": 1.374483728676764, "grad_norm": 1.1361045837402344, "learning_rate": 0.00018203558556293347, "loss": 4.6727, "step": 736550 }, { "epoch": 1.3745770342361423, "grad_norm": 1.2022897005081177, "learning_rate": 0.00018203318314548948, "loss": 4.513, "step": 736600 }, { "epoch": 1.374670339795521, "grad_norm": 0.9374406933784485, "learning_rate": 0.00018203078058327132, "loss": 4.3369, "step": 736650 }, { "epoch": 1.3747636453548995, "grad_norm": 1.4106011390686035, "learning_rate": 0.00018202837787628328, "loss": 4.5315, "step": 736700 }, { "epoch": 1.3748569509142778, "grad_norm": 1.2017098665237427, "learning_rate": 0.00018202597502452959, "loss": 4.526, "step": 736750 }, { "epoch": 1.3749502564736562, "grad_norm": 1.1423646211624146, "learning_rate": 0.0001820235720280145, "loss": 4.669, "step": 736800 }, { "epoch": 1.3750435620330348, "grad_norm": 1.2309614419937134, "learning_rate": 0.00018202116888674222, "loss": 4.5364, "step": 736850 }, { "epoch": 1.3751368675924134, "grad_norm": 1.1816354990005493, "learning_rate": 0.000182018765600717, "loss": 4.5655, "step": 736900 }, { "epoch": 1.3752301731517917, "grad_norm": 1.133590817451477, "learning_rate": 0.0001820163621699431, "loss": 4.5068, "step": 736950 }, { "epoch": 1.3753234787111703, "grad_norm": 1.264103889465332, "learning_rate": 0.00018201395859442475, "loss": 4.4976, "step": 737000 }, { "epoch": 1.375416784270549, "grad_norm": 1.1916100978851318, "learning_rate": 0.00018201155487416622, "loss": 4.3885, "step": 737050 }, { "epoch": 1.3755100898299273, "grad_norm": 0.9239583015441895, "learning_rate": 0.0001820091510091717, "loss": 4.512, "step": 737100 }, { "epoch": 1.3756033953893056, "grad_norm": 1.1363614797592163, "learning_rate": 0.00018200674699944547, "loss": 4.6257, "step": 737150 }, { "epoch": 1.3756967009486842, "grad_norm": 1.2927193641662598, "learning_rate": 0.00018200434284499175, "loss": 4.3546, "step": 737200 }, { "epoch": 1.3757900065080628, "grad_norm": 1.1732802391052246, "learning_rate": 0.0001820019385458148, "loss": 4.5902, "step": 737250 }, { "epoch": 1.3758833120674412, "grad_norm": 1.626499056816101, "learning_rate": 0.00018199953410191885, "loss": 4.4242, "step": 737300 }, { "epoch": 1.3759766176268198, "grad_norm": 1.166311264038086, "learning_rate": 0.00018199712951330816, "loss": 4.4817, "step": 737350 }, { "epoch": 1.3760699231861984, "grad_norm": 1.1325236558914185, "learning_rate": 0.00018199472477998697, "loss": 4.3447, "step": 737400 }, { "epoch": 1.3761632287455767, "grad_norm": 1.3078035116195679, "learning_rate": 0.00018199231990195948, "loss": 4.5005, "step": 737450 }, { "epoch": 1.376256534304955, "grad_norm": 1.2113882303237915, "learning_rate": 0.00018198991487923004, "loss": 4.3993, "step": 737500 }, { "epoch": 1.3763498398643337, "grad_norm": 1.4464737176895142, "learning_rate": 0.00018198750971180276, "loss": 4.499, "step": 737550 }, { "epoch": 1.3764431454237123, "grad_norm": 1.1914018392562866, "learning_rate": 0.000181985104399682, "loss": 4.7073, "step": 737600 }, { "epoch": 1.3765364509830906, "grad_norm": 1.1310336589813232, "learning_rate": 0.00018198269894287192, "loss": 4.703, "step": 737650 }, { "epoch": 1.3766297565424692, "grad_norm": 1.2337299585342407, "learning_rate": 0.00018198029334137684, "loss": 4.3418, "step": 737700 }, { "epoch": 1.3767230621018478, "grad_norm": 1.2073575258255005, "learning_rate": 0.00018197788759520092, "loss": 4.5033, "step": 737750 }, { "epoch": 1.3768163676612262, "grad_norm": 1.1496751308441162, "learning_rate": 0.0001819754817043485, "loss": 4.4736, "step": 737800 }, { "epoch": 1.3769096732206045, "grad_norm": 0.9903450012207031, "learning_rate": 0.00018197307566882373, "loss": 4.6974, "step": 737850 }, { "epoch": 1.3770029787799831, "grad_norm": 0.9429528713226318, "learning_rate": 0.00018197066948863095, "loss": 4.6276, "step": 737900 }, { "epoch": 1.3770962843393617, "grad_norm": 1.3893846273422241, "learning_rate": 0.00018196826316377434, "loss": 4.4261, "step": 737950 }, { "epoch": 1.37718958989874, "grad_norm": 1.1294351816177368, "learning_rate": 0.00018196585669425817, "loss": 4.4018, "step": 738000 }, { "epoch": 1.37718958989874, "eval_loss": 4.688307285308838, "eval_runtime": 228.2864, "eval_samples_per_second": 11.424, "eval_steps_per_second": 11.424, "eval_tts_loss": 7.567780133093953, "step": 738000 }, { "epoch": 1.3772828954581187, "grad_norm": 1.0019197463989258, "learning_rate": 0.0001819634500800867, "loss": 4.4515, "step": 738050 }, { "epoch": 1.3773762010174972, "grad_norm": 1.2432219982147217, "learning_rate": 0.00018196104332126413, "loss": 4.5865, "step": 738100 }, { "epoch": 1.3774695065768756, "grad_norm": 1.1337308883666992, "learning_rate": 0.00018195863641779476, "loss": 4.5654, "step": 738150 }, { "epoch": 1.377562812136254, "grad_norm": 1.4000921249389648, "learning_rate": 0.00018195622936968282, "loss": 4.2588, "step": 738200 }, { "epoch": 1.3776561176956326, "grad_norm": 1.2461419105529785, "learning_rate": 0.00018195382217693254, "loss": 4.6449, "step": 738250 }, { "epoch": 1.3777494232550112, "grad_norm": 1.0295132398605347, "learning_rate": 0.0001819514148395482, "loss": 4.6541, "step": 738300 }, { "epoch": 1.3778427288143895, "grad_norm": 1.1963058710098267, "learning_rate": 0.00018194900735753403, "loss": 4.4549, "step": 738350 }, { "epoch": 1.377936034373768, "grad_norm": 1.1836023330688477, "learning_rate": 0.00018194659973089428, "loss": 4.6551, "step": 738400 }, { "epoch": 1.3780293399331467, "grad_norm": 1.12228262424469, "learning_rate": 0.00018194419195963318, "loss": 4.6553, "step": 738450 }, { "epoch": 1.378122645492525, "grad_norm": 1.1734741926193237, "learning_rate": 0.00018194178404375505, "loss": 4.3965, "step": 738500 }, { "epoch": 1.3782159510519034, "grad_norm": 0.9173708558082581, "learning_rate": 0.00018193937598326407, "loss": 4.3467, "step": 738550 }, { "epoch": 1.378309256611282, "grad_norm": 1.1589096784591675, "learning_rate": 0.0001819369677781645, "loss": 4.4438, "step": 738600 }, { "epoch": 1.3784025621706606, "grad_norm": 1.1795318126678467, "learning_rate": 0.00018193455942846058, "loss": 4.6004, "step": 738650 }, { "epoch": 1.378495867730039, "grad_norm": 1.0837074518203735, "learning_rate": 0.0001819321509341566, "loss": 4.3709, "step": 738700 }, { "epoch": 1.3785891732894175, "grad_norm": 1.107070803642273, "learning_rate": 0.00018192974229525678, "loss": 4.582, "step": 738750 }, { "epoch": 1.3786824788487961, "grad_norm": 1.2992217540740967, "learning_rate": 0.0001819273335117654, "loss": 4.6154, "step": 738800 }, { "epoch": 1.3787757844081745, "grad_norm": 0.6269069314002991, "learning_rate": 0.0001819249245836867, "loss": 4.4746, "step": 738850 }, { "epoch": 1.3788690899675529, "grad_norm": 1.120640754699707, "learning_rate": 0.00018192251551102487, "loss": 4.5528, "step": 738900 }, { "epoch": 1.3789623955269315, "grad_norm": 1.4784859418869019, "learning_rate": 0.00018192010629378426, "loss": 4.5772, "step": 738950 }, { "epoch": 1.37905570108631, "grad_norm": 1.2471284866333008, "learning_rate": 0.00018191769693196909, "loss": 4.4015, "step": 739000 }, { "epoch": 1.3791490066456884, "grad_norm": 1.0942065715789795, "learning_rate": 0.00018191528742558357, "loss": 4.7002, "step": 739050 }, { "epoch": 1.379242312205067, "grad_norm": 1.3342229127883911, "learning_rate": 0.00018191287777463195, "loss": 4.3671, "step": 739100 }, { "epoch": 1.3793356177644456, "grad_norm": 1.1442571878433228, "learning_rate": 0.00018191046797911856, "loss": 4.5094, "step": 739150 }, { "epoch": 1.379428923323824, "grad_norm": 1.2759796380996704, "learning_rate": 0.0001819080580390476, "loss": 4.3608, "step": 739200 }, { "epoch": 1.3795222288832023, "grad_norm": 1.1455414295196533, "learning_rate": 0.00018190564795442333, "loss": 4.4101, "step": 739250 }, { "epoch": 1.379615534442581, "grad_norm": 1.3588656187057495, "learning_rate": 0.00018190323772525, "loss": 4.5566, "step": 739300 }, { "epoch": 1.3797088400019595, "grad_norm": 0.7553815841674805, "learning_rate": 0.00018190082735153188, "loss": 4.3647, "step": 739350 }, { "epoch": 1.3798021455613378, "grad_norm": 1.0276620388031006, "learning_rate": 0.00018189841683327316, "loss": 4.6332, "step": 739400 }, { "epoch": 1.3798954511207164, "grad_norm": 1.1570426225662231, "learning_rate": 0.00018189600617047819, "loss": 4.6445, "step": 739450 }, { "epoch": 1.379988756680095, "grad_norm": 0.9749423265457153, "learning_rate": 0.00018189359536315117, "loss": 4.5376, "step": 739500 }, { "epoch": 1.3800820622394734, "grad_norm": 1.301543951034546, "learning_rate": 0.00018189118441129634, "loss": 4.5919, "step": 739550 }, { "epoch": 1.3801753677988517, "grad_norm": 1.320989727973938, "learning_rate": 0.00018188877331491802, "loss": 4.5888, "step": 739600 }, { "epoch": 1.3802686733582303, "grad_norm": 1.1201611757278442, "learning_rate": 0.00018188636207402042, "loss": 4.3135, "step": 739650 }, { "epoch": 1.380361978917609, "grad_norm": 1.200514316558838, "learning_rate": 0.00018188395068860779, "loss": 4.4166, "step": 739700 }, { "epoch": 1.3804552844769873, "grad_norm": 1.466609239578247, "learning_rate": 0.0001818815391586844, "loss": 4.4657, "step": 739750 }, { "epoch": 1.3805485900363659, "grad_norm": 1.2917295694351196, "learning_rate": 0.0001818791274842545, "loss": 4.5313, "step": 739800 }, { "epoch": 1.3806418955957445, "grad_norm": 1.294675350189209, "learning_rate": 0.0001818767156653223, "loss": 4.452, "step": 739850 }, { "epoch": 1.3807352011551228, "grad_norm": 1.311125636100769, "learning_rate": 0.00018187430370189217, "loss": 4.6292, "step": 739900 }, { "epoch": 1.3808285067145012, "grad_norm": 1.4174914360046387, "learning_rate": 0.00018187189159396826, "loss": 4.4697, "step": 739950 }, { "epoch": 1.3809218122738798, "grad_norm": 0.8422597646713257, "learning_rate": 0.00018186947934155487, "loss": 4.5705, "step": 740000 }, { "epoch": 1.3810151178332584, "grad_norm": 1.3974944353103638, "learning_rate": 0.00018186706694465628, "loss": 4.5413, "step": 740050 }, { "epoch": 1.3811084233926367, "grad_norm": 1.2775051593780518, "learning_rate": 0.00018186465440327672, "loss": 4.4324, "step": 740100 }, { "epoch": 1.3812017289520153, "grad_norm": 1.1954485177993774, "learning_rate": 0.0001818622417174204, "loss": 4.523, "step": 740150 }, { "epoch": 1.381295034511394, "grad_norm": 1.2812470197677612, "learning_rate": 0.00018185982888709168, "loss": 4.3411, "step": 740200 }, { "epoch": 1.3813883400707723, "grad_norm": 1.1083112955093384, "learning_rate": 0.00018185741591229476, "loss": 4.3836, "step": 740250 }, { "epoch": 1.3814816456301506, "grad_norm": 1.2739968299865723, "learning_rate": 0.00018185500279303387, "loss": 4.5441, "step": 740300 }, { "epoch": 1.3815749511895292, "grad_norm": 1.4455690383911133, "learning_rate": 0.0001818525895293133, "loss": 4.4546, "step": 740350 }, { "epoch": 1.3816682567489078, "grad_norm": 1.268727421760559, "learning_rate": 0.00018185017612113737, "loss": 4.6359, "step": 740400 }, { "epoch": 1.3817615623082862, "grad_norm": 1.3261535167694092, "learning_rate": 0.00018184776256851027, "loss": 4.3572, "step": 740450 }, { "epoch": 1.3818548678676648, "grad_norm": 1.1327229738235474, "learning_rate": 0.00018184534887143622, "loss": 4.3604, "step": 740500 }, { "epoch": 1.3819481734270433, "grad_norm": 0.8929272890090942, "learning_rate": 0.00018184293502991955, "loss": 4.7439, "step": 740550 }, { "epoch": 1.3820414789864217, "grad_norm": 1.4520336389541626, "learning_rate": 0.00018184052104396452, "loss": 4.8486, "step": 740600 }, { "epoch": 1.3821347845458, "grad_norm": 1.334388256072998, "learning_rate": 0.00018183810691357538, "loss": 4.6565, "step": 740650 }, { "epoch": 1.3822280901051787, "grad_norm": 1.2245125770568848, "learning_rate": 0.00018183569263875634, "loss": 4.5767, "step": 740700 }, { "epoch": 1.3823213956645573, "grad_norm": 1.1840416193008423, "learning_rate": 0.00018183327821951172, "loss": 4.6461, "step": 740750 }, { "epoch": 1.3824147012239356, "grad_norm": 1.4155598878860474, "learning_rate": 0.00018183086365584577, "loss": 4.464, "step": 740800 }, { "epoch": 1.3825080067833142, "grad_norm": 1.07320237159729, "learning_rate": 0.00018182844894776276, "loss": 4.5737, "step": 740850 }, { "epoch": 1.3826013123426928, "grad_norm": 1.2409127950668335, "learning_rate": 0.0001818260340952669, "loss": 4.5012, "step": 740900 }, { "epoch": 1.3826946179020712, "grad_norm": 1.2487311363220215, "learning_rate": 0.0001818236190983625, "loss": 4.4423, "step": 740950 }, { "epoch": 1.3827879234614495, "grad_norm": 1.1758610010147095, "learning_rate": 0.0001818212039570538, "loss": 4.5736, "step": 741000 }, { "epoch": 1.3827879234614495, "eval_loss": 4.670938491821289, "eval_runtime": 230.5561, "eval_samples_per_second": 11.312, "eval_steps_per_second": 11.312, "eval_tts_loss": 7.581402274141476, "step": 741000 }, { "epoch": 1.382881229020828, "grad_norm": 1.2267298698425293, "learning_rate": 0.00018181878867134512, "loss": 4.3385, "step": 741050 }, { "epoch": 1.3829745345802067, "grad_norm": 1.223245620727539, "learning_rate": 0.0001818163732412406, "loss": 4.531, "step": 741100 }, { "epoch": 1.383067840139585, "grad_norm": 1.3421828746795654, "learning_rate": 0.00018181395766674464, "loss": 4.3949, "step": 741150 }, { "epoch": 1.3831611456989636, "grad_norm": 0.8390029668807983, "learning_rate": 0.00018181154194786144, "loss": 4.3298, "step": 741200 }, { "epoch": 1.3832544512583422, "grad_norm": 1.120850920677185, "learning_rate": 0.00018180912608459522, "loss": 4.3406, "step": 741250 }, { "epoch": 1.3833477568177206, "grad_norm": 1.0253372192382812, "learning_rate": 0.00018180671007695032, "loss": 4.5536, "step": 741300 }, { "epoch": 1.383441062377099, "grad_norm": 1.0396504402160645, "learning_rate": 0.00018180429392493095, "loss": 4.5022, "step": 741350 }, { "epoch": 1.3835343679364775, "grad_norm": 0.9932730197906494, "learning_rate": 0.00018180187762854142, "loss": 4.6097, "step": 741400 }, { "epoch": 1.3836276734958561, "grad_norm": 1.2692090272903442, "learning_rate": 0.00018179946118778592, "loss": 4.3968, "step": 741450 }, { "epoch": 1.3837209790552345, "grad_norm": 1.0941412448883057, "learning_rate": 0.0001817970446026688, "loss": 4.531, "step": 741500 }, { "epoch": 1.383814284614613, "grad_norm": 1.3016831874847412, "learning_rate": 0.0001817946278731943, "loss": 4.5884, "step": 741550 }, { "epoch": 1.3839075901739917, "grad_norm": 1.0995057821273804, "learning_rate": 0.00018179221099936664, "loss": 4.6447, "step": 741600 }, { "epoch": 1.38400089573337, "grad_norm": 1.315546989440918, "learning_rate": 0.00018178979398119017, "loss": 4.4148, "step": 741650 }, { "epoch": 1.3840942012927484, "grad_norm": 1.0577149391174316, "learning_rate": 0.00018178737681866905, "loss": 4.4434, "step": 741700 }, { "epoch": 1.384187506852127, "grad_norm": 1.1359922885894775, "learning_rate": 0.00018178495951180765, "loss": 4.5552, "step": 741750 }, { "epoch": 1.3842808124115056, "grad_norm": 0.7878245711326599, "learning_rate": 0.00018178254206061017, "loss": 4.301, "step": 741800 }, { "epoch": 1.384374117970884, "grad_norm": 0.9029074907302856, "learning_rate": 0.00018178012446508089, "loss": 4.2025, "step": 741850 }, { "epoch": 1.3844674235302625, "grad_norm": 0.9319016933441162, "learning_rate": 0.00018177770672522404, "loss": 4.4661, "step": 741900 }, { "epoch": 1.3845607290896411, "grad_norm": 1.2459003925323486, "learning_rate": 0.00018177528884104398, "loss": 4.6916, "step": 741950 }, { "epoch": 1.3846540346490195, "grad_norm": 1.1567898988723755, "learning_rate": 0.0001817728708125449, "loss": 4.3946, "step": 742000 }, { "epoch": 1.3847473402083978, "grad_norm": 1.3948370218276978, "learning_rate": 0.0001817704526397311, "loss": 4.4825, "step": 742050 }, { "epoch": 1.3848406457677764, "grad_norm": 1.0539705753326416, "learning_rate": 0.00018176803432260686, "loss": 4.3778, "step": 742100 }, { "epoch": 1.384933951327155, "grad_norm": 0.8697056770324707, "learning_rate": 0.00018176561586117637, "loss": 4.6005, "step": 742150 }, { "epoch": 1.3850272568865334, "grad_norm": 1.2844717502593994, "learning_rate": 0.000181763197255444, "loss": 4.5695, "step": 742200 }, { "epoch": 1.385120562445912, "grad_norm": 1.1817405223846436, "learning_rate": 0.00018176077850541396, "loss": 4.5681, "step": 742250 }, { "epoch": 1.3852138680052906, "grad_norm": 1.107040286064148, "learning_rate": 0.00018175835961109054, "loss": 4.2164, "step": 742300 }, { "epoch": 1.385307173564669, "grad_norm": 1.1814676523208618, "learning_rate": 0.000181755940572478, "loss": 4.5459, "step": 742350 }, { "epoch": 1.3854004791240473, "grad_norm": 1.3418138027191162, "learning_rate": 0.0001817535213895806, "loss": 4.5556, "step": 742400 }, { "epoch": 1.3854937846834259, "grad_norm": 1.0078847408294678, "learning_rate": 0.00018175110206240263, "loss": 4.6295, "step": 742450 }, { "epoch": 1.3855870902428045, "grad_norm": 1.1013423204421997, "learning_rate": 0.00018174868259094836, "loss": 4.6485, "step": 742500 }, { "epoch": 1.3856803958021828, "grad_norm": 0.8448772430419922, "learning_rate": 0.00018174626297522202, "loss": 4.6937, "step": 742550 }, { "epoch": 1.3857737013615614, "grad_norm": 0.9531406760215759, "learning_rate": 0.00018174384321522792, "loss": 4.2781, "step": 742600 }, { "epoch": 1.38586700692094, "grad_norm": 1.0159580707550049, "learning_rate": 0.00018174142331097034, "loss": 4.4729, "step": 742650 }, { "epoch": 1.3859603124803184, "grad_norm": 1.2858434915542603, "learning_rate": 0.00018173900326245353, "loss": 4.6101, "step": 742700 }, { "epoch": 1.3860536180396967, "grad_norm": 1.0895453691482544, "learning_rate": 0.00018173658306968173, "loss": 4.4871, "step": 742750 }, { "epoch": 1.3861469235990753, "grad_norm": 1.1962652206420898, "learning_rate": 0.00018173416273265926, "loss": 4.5275, "step": 742800 }, { "epoch": 1.386240229158454, "grad_norm": 1.2174633741378784, "learning_rate": 0.00018173174225139037, "loss": 4.7684, "step": 742850 }, { "epoch": 1.3863335347178323, "grad_norm": 1.2196329832077026, "learning_rate": 0.00018172932162587935, "loss": 4.6329, "step": 742900 }, { "epoch": 1.3864268402772109, "grad_norm": 1.3134591579437256, "learning_rate": 0.00018172690085613044, "loss": 4.4364, "step": 742950 }, { "epoch": 1.3865201458365894, "grad_norm": 1.1753904819488525, "learning_rate": 0.00018172447994214795, "loss": 4.3735, "step": 743000 }, { "epoch": 1.3866134513959678, "grad_norm": 0.8776636719703674, "learning_rate": 0.00018172205888393613, "loss": 4.3503, "step": 743050 }, { "epoch": 1.3867067569553462, "grad_norm": 0.9745319485664368, "learning_rate": 0.0001817196376814992, "loss": 4.527, "step": 743100 }, { "epoch": 1.3868000625147248, "grad_norm": 1.390006184577942, "learning_rate": 0.00018171721633484156, "loss": 4.3717, "step": 743150 }, { "epoch": 1.3868933680741033, "grad_norm": 1.455784797668457, "learning_rate": 0.0001817147948439674, "loss": 4.6434, "step": 743200 }, { "epoch": 1.3869866736334817, "grad_norm": 1.1880831718444824, "learning_rate": 0.000181712373208881, "loss": 4.5959, "step": 743250 }, { "epoch": 1.3870799791928603, "grad_norm": 1.1660850048065186, "learning_rate": 0.00018170995142958662, "loss": 4.3931, "step": 743300 }, { "epoch": 1.3871732847522389, "grad_norm": 1.1215369701385498, "learning_rate": 0.00018170752950608855, "loss": 4.5012, "step": 743350 }, { "epoch": 1.3872665903116173, "grad_norm": 1.139059066772461, "learning_rate": 0.00018170510743839108, "loss": 4.4408, "step": 743400 }, { "epoch": 1.3873598958709956, "grad_norm": 1.228177785873413, "learning_rate": 0.00018170268522649848, "loss": 4.347, "step": 743450 }, { "epoch": 1.3874532014303742, "grad_norm": 1.1280699968338013, "learning_rate": 0.00018170026287041502, "loss": 4.2955, "step": 743500 }, { "epoch": 1.3875465069897528, "grad_norm": 1.1662507057189941, "learning_rate": 0.00018169784037014495, "loss": 4.4356, "step": 743550 }, { "epoch": 1.3876398125491312, "grad_norm": 1.4087531566619873, "learning_rate": 0.00018169541772569258, "loss": 4.6561, "step": 743600 }, { "epoch": 1.3877331181085097, "grad_norm": 0.9973386526107788, "learning_rate": 0.0001816929949370622, "loss": 4.4546, "step": 743650 }, { "epoch": 1.387826423667888, "grad_norm": 0.9776493906974792, "learning_rate": 0.000181690572004258, "loss": 4.7011, "step": 743700 }, { "epoch": 1.3879197292272667, "grad_norm": 1.5359793901443481, "learning_rate": 0.00018168814892728435, "loss": 4.4952, "step": 743750 }, { "epoch": 1.388013034786645, "grad_norm": 1.2765954732894897, "learning_rate": 0.0001816857257061455, "loss": 4.6882, "step": 743800 }, { "epoch": 1.3881063403460236, "grad_norm": 1.1062088012695312, "learning_rate": 0.0001816833023408457, "loss": 4.5426, "step": 743850 }, { "epoch": 1.3881996459054022, "grad_norm": 1.2386276721954346, "learning_rate": 0.00018168087883138924, "loss": 4.5787, "step": 743900 }, { "epoch": 1.3882929514647806, "grad_norm": 1.1232293844223022, "learning_rate": 0.00018167845517778043, "loss": 4.4551, "step": 743950 }, { "epoch": 1.3883862570241592, "grad_norm": 1.3388580083847046, "learning_rate": 0.00018167603138002347, "loss": 4.4248, "step": 744000 }, { "epoch": 1.3883862570241592, "eval_loss": 4.665449619293213, "eval_runtime": 232.6593, "eval_samples_per_second": 11.21, "eval_steps_per_second": 11.21, "eval_tts_loss": 7.548311103998106, "step": 744000 }, { "epoch": 1.3884795625835376, "grad_norm": 1.1443004608154297, "learning_rate": 0.0001816736074381227, "loss": 4.5333, "step": 744050 }, { "epoch": 1.3885728681429161, "grad_norm": 1.1931462287902832, "learning_rate": 0.0001816711833520824, "loss": 4.4598, "step": 744100 }, { "epoch": 1.3886661737022945, "grad_norm": 1.23349928855896, "learning_rate": 0.00018166875912190684, "loss": 4.6273, "step": 744150 }, { "epoch": 1.388759479261673, "grad_norm": 1.162979245185852, "learning_rate": 0.00018166633474760028, "loss": 4.5263, "step": 744200 }, { "epoch": 1.3888527848210517, "grad_norm": 1.4025274515151978, "learning_rate": 0.00018166391022916698, "loss": 4.5832, "step": 744250 }, { "epoch": 1.38894609038043, "grad_norm": 1.3977222442626953, "learning_rate": 0.0001816614855666113, "loss": 4.7893, "step": 744300 }, { "epoch": 1.3890393959398086, "grad_norm": 1.182018756866455, "learning_rate": 0.00018165906075993742, "loss": 4.3554, "step": 744350 }, { "epoch": 1.389132701499187, "grad_norm": 0.7959296107292175, "learning_rate": 0.0001816566358091497, "loss": 4.5722, "step": 744400 }, { "epoch": 1.3892260070585656, "grad_norm": 1.0364402532577515, "learning_rate": 0.00018165421071425237, "loss": 4.4124, "step": 744450 }, { "epoch": 1.389319312617944, "grad_norm": 1.0856233835220337, "learning_rate": 0.00018165178547524972, "loss": 4.3885, "step": 744500 }, { "epoch": 1.3894126181773225, "grad_norm": 1.0726815462112427, "learning_rate": 0.00018164936009214605, "loss": 4.5266, "step": 744550 }, { "epoch": 1.3895059237367011, "grad_norm": 0.9698043465614319, "learning_rate": 0.0001816469345649456, "loss": 4.5377, "step": 744600 }, { "epoch": 1.3895992292960795, "grad_norm": 0.9028865098953247, "learning_rate": 0.00018164450889365273, "loss": 4.6209, "step": 744650 }, { "epoch": 1.389692534855458, "grad_norm": 1.1449333429336548, "learning_rate": 0.00018164208307827162, "loss": 4.6014, "step": 744700 }, { "epoch": 1.3897858404148364, "grad_norm": 1.1962692737579346, "learning_rate": 0.0001816396571188066, "loss": 4.5034, "step": 744750 }, { "epoch": 1.389879145974215, "grad_norm": 1.293436050415039, "learning_rate": 0.00018163723101526197, "loss": 4.5959, "step": 744800 }, { "epoch": 1.3899724515335934, "grad_norm": 1.300286889076233, "learning_rate": 0.00018163480476764197, "loss": 4.7467, "step": 744850 }, { "epoch": 1.390065757092972, "grad_norm": 1.1174542903900146, "learning_rate": 0.00018163237837595094, "loss": 4.564, "step": 744900 }, { "epoch": 1.3901590626523506, "grad_norm": 1.178055763244629, "learning_rate": 0.0001816299518401931, "loss": 4.4959, "step": 744950 }, { "epoch": 1.390252368211729, "grad_norm": 1.1373776197433472, "learning_rate": 0.00018162752516037275, "loss": 4.4362, "step": 745000 }, { "epoch": 1.3903456737711075, "grad_norm": 1.296624779701233, "learning_rate": 0.0001816250983364942, "loss": 4.6398, "step": 745050 }, { "epoch": 1.3904389793304859, "grad_norm": 1.3729219436645508, "learning_rate": 0.00018162267136856173, "loss": 4.4137, "step": 745100 }, { "epoch": 1.3905322848898645, "grad_norm": 1.3173812627792358, "learning_rate": 0.00018162024425657957, "loss": 4.5347, "step": 745150 }, { "epoch": 1.3906255904492428, "grad_norm": 1.092727780342102, "learning_rate": 0.00018161781700055206, "loss": 4.5736, "step": 745200 }, { "epoch": 1.3907188960086214, "grad_norm": 0.9062284827232361, "learning_rate": 0.00018161538960048343, "loss": 4.4916, "step": 745250 }, { "epoch": 1.390812201568, "grad_norm": 1.3439369201660156, "learning_rate": 0.00018161296205637805, "loss": 4.5186, "step": 745300 }, { "epoch": 1.3909055071273784, "grad_norm": 1.3360991477966309, "learning_rate": 0.0001816105343682401, "loss": 4.4892, "step": 745350 }, { "epoch": 1.390998812686757, "grad_norm": 1.2580780982971191, "learning_rate": 0.00018160810653607398, "loss": 4.4984, "step": 745400 }, { "epoch": 1.3910921182461353, "grad_norm": 1.3013616800308228, "learning_rate": 0.00018160567855988387, "loss": 4.4311, "step": 745450 }, { "epoch": 1.391185423805514, "grad_norm": 1.1744035482406616, "learning_rate": 0.0001816032504396741, "loss": 4.3648, "step": 745500 }, { "epoch": 1.3912787293648923, "grad_norm": 1.3316484689712524, "learning_rate": 0.00018160082217544895, "loss": 4.6565, "step": 745550 }, { "epoch": 1.3913720349242709, "grad_norm": 1.2531282901763916, "learning_rate": 0.00018159839376721273, "loss": 4.417, "step": 745600 }, { "epoch": 1.3914653404836494, "grad_norm": 1.4662128686904907, "learning_rate": 0.0001815959652149697, "loss": 4.5968, "step": 745650 }, { "epoch": 1.3915586460430278, "grad_norm": 1.3194403648376465, "learning_rate": 0.0001815935365187241, "loss": 4.4418, "step": 745700 }, { "epoch": 1.3916519516024064, "grad_norm": 1.2246984243392944, "learning_rate": 0.0001815911076784803, "loss": 4.375, "step": 745750 }, { "epoch": 1.3917452571617848, "grad_norm": 1.0469554662704468, "learning_rate": 0.00018158867869424254, "loss": 4.3668, "step": 745800 }, { "epoch": 1.3918385627211634, "grad_norm": 0.8312752842903137, "learning_rate": 0.00018158624956601514, "loss": 4.4089, "step": 745850 }, { "epoch": 1.3919318682805417, "grad_norm": 1.115979552268982, "learning_rate": 0.00018158382029380237, "loss": 4.5188, "step": 745900 }, { "epoch": 1.3920251738399203, "grad_norm": 1.1936399936676025, "learning_rate": 0.00018158139087760846, "loss": 4.6216, "step": 745950 }, { "epoch": 1.392118479399299, "grad_norm": 1.2202603816986084, "learning_rate": 0.0001815789613174378, "loss": 4.3947, "step": 746000 }, { "epoch": 1.3922117849586773, "grad_norm": 1.0221086740493774, "learning_rate": 0.0001815765316132946, "loss": 4.5704, "step": 746050 }, { "epoch": 1.3923050905180558, "grad_norm": 1.1236578226089478, "learning_rate": 0.0001815741017651832, "loss": 4.4467, "step": 746100 }, { "epoch": 1.3923983960774342, "grad_norm": 1.0076820850372314, "learning_rate": 0.00018157167177310784, "loss": 4.3994, "step": 746150 }, { "epoch": 1.3924917016368128, "grad_norm": 0.9319136142730713, "learning_rate": 0.00018156924163707284, "loss": 4.5603, "step": 746200 }, { "epoch": 1.3925850071961912, "grad_norm": 0.78514164686203, "learning_rate": 0.0001815668113570825, "loss": 4.6471, "step": 746250 }, { "epoch": 1.3926783127555697, "grad_norm": 1.1521482467651367, "learning_rate": 0.00018156438093314107, "loss": 4.4884, "step": 746300 }, { "epoch": 1.3927716183149483, "grad_norm": 1.066415786743164, "learning_rate": 0.00018156195036525286, "loss": 4.5241, "step": 746350 }, { "epoch": 1.3928649238743267, "grad_norm": 1.1827502250671387, "learning_rate": 0.00018155951965342215, "loss": 4.5111, "step": 746400 }, { "epoch": 1.3929582294337053, "grad_norm": 0.7660015225410461, "learning_rate": 0.00018155708879765325, "loss": 4.1643, "step": 746450 }, { "epoch": 1.3930515349930837, "grad_norm": 1.2079882621765137, "learning_rate": 0.00018155465779795047, "loss": 4.4852, "step": 746500 }, { "epoch": 1.3931448405524622, "grad_norm": 1.2508578300476074, "learning_rate": 0.000181552226654318, "loss": 4.3099, "step": 746550 }, { "epoch": 1.3932381461118406, "grad_norm": 1.376076579093933, "learning_rate": 0.00018154979536676028, "loss": 4.5731, "step": 746600 }, { "epoch": 1.3933314516712192, "grad_norm": 0.9002833962440491, "learning_rate": 0.00018154736393528146, "loss": 4.286, "step": 746650 }, { "epoch": 1.3934247572305978, "grad_norm": 1.3027229309082031, "learning_rate": 0.0001815449323598859, "loss": 4.5449, "step": 746700 }, { "epoch": 1.3935180627899761, "grad_norm": 1.207923412322998, "learning_rate": 0.0001815425006405779, "loss": 4.4892, "step": 746750 }, { "epoch": 1.3936113683493547, "grad_norm": 1.0031918287277222, "learning_rate": 0.00018154006877736174, "loss": 4.6372, "step": 746800 }, { "epoch": 1.393704673908733, "grad_norm": 1.349355697631836, "learning_rate": 0.00018153763677024169, "loss": 4.3827, "step": 746850 }, { "epoch": 1.3937979794681117, "grad_norm": 1.1594687700271606, "learning_rate": 0.00018153520461922207, "loss": 4.4116, "step": 746900 }, { "epoch": 1.39389128502749, "grad_norm": 1.3418527841567993, "learning_rate": 0.00018153277232430714, "loss": 4.3375, "step": 746950 }, { "epoch": 1.3939845905868686, "grad_norm": 1.2757490873336792, "learning_rate": 0.00018153033988550123, "loss": 4.5746, "step": 747000 }, { "epoch": 1.3939845905868686, "eval_loss": 4.660811424255371, "eval_runtime": 233.4628, "eval_samples_per_second": 11.171, "eval_steps_per_second": 11.171, "eval_tts_loss": 7.634264121751981, "step": 747000 }, { "epoch": 1.3940778961462472, "grad_norm": 1.5538183450698853, "learning_rate": 0.0001815279073028086, "loss": 4.4649, "step": 747050 }, { "epoch": 1.3941712017056256, "grad_norm": 1.3139029741287231, "learning_rate": 0.00018152547457623357, "loss": 4.4767, "step": 747100 }, { "epoch": 1.3942645072650042, "grad_norm": 1.4131461381912231, "learning_rate": 0.00018152304170578043, "loss": 4.3864, "step": 747150 }, { "epoch": 1.3943578128243825, "grad_norm": 1.1941815614700317, "learning_rate": 0.00018152060869145345, "loss": 4.5516, "step": 747200 }, { "epoch": 1.3944511183837611, "grad_norm": 1.149739146232605, "learning_rate": 0.00018151817553325694, "loss": 4.4308, "step": 747250 }, { "epoch": 1.3945444239431395, "grad_norm": 0.9161651134490967, "learning_rate": 0.00018151574223119522, "loss": 4.5143, "step": 747300 }, { "epoch": 1.394637729502518, "grad_norm": 0.8368246555328369, "learning_rate": 0.00018151330878527253, "loss": 4.414, "step": 747350 }, { "epoch": 1.3947310350618967, "grad_norm": 1.3266220092773438, "learning_rate": 0.00018151087519549321, "loss": 4.395, "step": 747400 }, { "epoch": 1.394824340621275, "grad_norm": 1.1753978729248047, "learning_rate": 0.00018150844146186152, "loss": 4.5348, "step": 747450 }, { "epoch": 1.3949176461806536, "grad_norm": 1.209136724472046, "learning_rate": 0.00018150600758438178, "loss": 4.4855, "step": 747500 }, { "epoch": 1.395010951740032, "grad_norm": 1.1439342498779297, "learning_rate": 0.00018150357356305828, "loss": 4.4803, "step": 747550 }, { "epoch": 1.3951042572994106, "grad_norm": 1.2330849170684814, "learning_rate": 0.00018150113939789534, "loss": 4.393, "step": 747600 }, { "epoch": 1.395197562858789, "grad_norm": 1.0816789865493774, "learning_rate": 0.00018149870508889717, "loss": 4.4776, "step": 747650 }, { "epoch": 1.3952908684181675, "grad_norm": 1.3368231058120728, "learning_rate": 0.00018149627063606815, "loss": 4.6868, "step": 747700 }, { "epoch": 1.395384173977546, "grad_norm": 1.3570938110351562, "learning_rate": 0.00018149383603941258, "loss": 4.6495, "step": 747750 }, { "epoch": 1.3954774795369245, "grad_norm": 1.0068148374557495, "learning_rate": 0.0001814914012989347, "loss": 4.7285, "step": 747800 }, { "epoch": 1.395570785096303, "grad_norm": 1.410873532295227, "learning_rate": 0.00018148896641463887, "loss": 4.3233, "step": 747850 }, { "epoch": 1.3956640906556814, "grad_norm": 1.2484678030014038, "learning_rate": 0.0001814865313865293, "loss": 4.5296, "step": 747900 }, { "epoch": 1.39575739621506, "grad_norm": 1.2434500455856323, "learning_rate": 0.00018148409621461037, "loss": 4.4355, "step": 747950 }, { "epoch": 1.3958507017744384, "grad_norm": 0.9417231678962708, "learning_rate": 0.00018148166089888634, "loss": 4.4651, "step": 748000 }, { "epoch": 1.395944007333817, "grad_norm": 0.9733771681785583, "learning_rate": 0.00018147922543936153, "loss": 4.414, "step": 748050 }, { "epoch": 1.3960373128931955, "grad_norm": 1.2862821817398071, "learning_rate": 0.0001814767898360402, "loss": 4.5858, "step": 748100 }, { "epoch": 1.396130618452574, "grad_norm": 1.4232697486877441, "learning_rate": 0.0001814743540889267, "loss": 4.4794, "step": 748150 }, { "epoch": 1.3962239240119525, "grad_norm": 1.304499864578247, "learning_rate": 0.0001814719181980253, "loss": 4.4085, "step": 748200 }, { "epoch": 1.3963172295713309, "grad_norm": 1.1524606943130493, "learning_rate": 0.0001814694821633403, "loss": 4.5677, "step": 748250 }, { "epoch": 1.3964105351307095, "grad_norm": 1.3950220346450806, "learning_rate": 0.00018146704598487597, "loss": 4.6488, "step": 748300 }, { "epoch": 1.3965038406900878, "grad_norm": 1.0185641050338745, "learning_rate": 0.00018146460966263664, "loss": 4.5063, "step": 748350 }, { "epoch": 1.3965971462494664, "grad_norm": 1.425874948501587, "learning_rate": 0.00018146217319662663, "loss": 4.5938, "step": 748400 }, { "epoch": 1.396690451808845, "grad_norm": 1.1669811010360718, "learning_rate": 0.00018145973658685021, "loss": 4.6055, "step": 748450 }, { "epoch": 1.3967837573682234, "grad_norm": 1.270644187927246, "learning_rate": 0.0001814572998333117, "loss": 4.5307, "step": 748500 }, { "epoch": 1.396877062927602, "grad_norm": 1.3163431882858276, "learning_rate": 0.00018145486293601537, "loss": 4.4152, "step": 748550 }, { "epoch": 1.3969703684869803, "grad_norm": 1.2029614448547363, "learning_rate": 0.00018145242589496556, "loss": 4.5301, "step": 748600 }, { "epoch": 1.397063674046359, "grad_norm": 1.2176302671432495, "learning_rate": 0.00018144998871016653, "loss": 4.683, "step": 748650 }, { "epoch": 1.3971569796057373, "grad_norm": 1.4381041526794434, "learning_rate": 0.00018144755138162262, "loss": 4.5874, "step": 748700 }, { "epoch": 1.3972502851651158, "grad_norm": 1.0806533098220825, "learning_rate": 0.0001814451139093381, "loss": 4.5279, "step": 748750 }, { "epoch": 1.3973435907244944, "grad_norm": 1.1419132947921753, "learning_rate": 0.00018144267629331728, "loss": 4.445, "step": 748800 }, { "epoch": 1.3974368962838728, "grad_norm": 1.0192393064498901, "learning_rate": 0.00018144023853356445, "loss": 4.4629, "step": 748850 }, { "epoch": 1.3975302018432514, "grad_norm": 1.0064804553985596, "learning_rate": 0.00018143780063008396, "loss": 4.3196, "step": 748900 }, { "epoch": 1.3976235074026298, "grad_norm": 1.1373525857925415, "learning_rate": 0.0001814353625828801, "loss": 4.512, "step": 748950 }, { "epoch": 1.3977168129620083, "grad_norm": 0.7406309843063354, "learning_rate": 0.0001814329243919571, "loss": 4.5361, "step": 749000 }, { "epoch": 1.3978101185213867, "grad_norm": 1.1158356666564941, "learning_rate": 0.00018143048605731937, "loss": 4.47, "step": 749050 }, { "epoch": 1.3979034240807653, "grad_norm": 1.3140724897384644, "learning_rate": 0.00018142804757897114, "loss": 4.3818, "step": 749100 }, { "epoch": 1.3979967296401439, "grad_norm": 1.448439598083496, "learning_rate": 0.00018142560895691672, "loss": 4.3786, "step": 749150 }, { "epoch": 1.3980900351995222, "grad_norm": 1.1739212274551392, "learning_rate": 0.00018142317019116043, "loss": 4.5299, "step": 749200 }, { "epoch": 1.3981833407589008, "grad_norm": 1.1563185453414917, "learning_rate": 0.00018142073128170658, "loss": 4.5032, "step": 749250 }, { "epoch": 1.3982766463182792, "grad_norm": 1.6373839378356934, "learning_rate": 0.00018141829222855947, "loss": 4.6848, "step": 749300 }, { "epoch": 1.3983699518776578, "grad_norm": 0.9949313402175903, "learning_rate": 0.0001814158530317234, "loss": 4.3613, "step": 749350 }, { "epoch": 1.3984632574370361, "grad_norm": 1.1779184341430664, "learning_rate": 0.00018141341369120267, "loss": 4.6059, "step": 749400 }, { "epoch": 1.3985565629964147, "grad_norm": 1.4595201015472412, "learning_rate": 0.00018141097420700157, "loss": 4.8709, "step": 749450 }, { "epoch": 1.3986498685557933, "grad_norm": 1.2074248790740967, "learning_rate": 0.00018140853457912447, "loss": 4.43, "step": 749500 }, { "epoch": 1.3987431741151717, "grad_norm": 1.1824707984924316, "learning_rate": 0.00018140609480757559, "loss": 4.382, "step": 749550 }, { "epoch": 1.3988364796745503, "grad_norm": 1.041249394416809, "learning_rate": 0.0001814036548923593, "loss": 4.6087, "step": 749600 }, { "epoch": 1.3989297852339286, "grad_norm": 1.1472258567810059, "learning_rate": 0.00018140121483347987, "loss": 4.3556, "step": 749650 }, { "epoch": 1.3990230907933072, "grad_norm": 1.085038185119629, "learning_rate": 0.00018139877463094163, "loss": 4.6154, "step": 749700 }, { "epoch": 1.3991163963526856, "grad_norm": 1.3075592517852783, "learning_rate": 0.00018139633428474884, "loss": 4.3898, "step": 749750 }, { "epoch": 1.3992097019120642, "grad_norm": 1.3621082305908203, "learning_rate": 0.0001813938937949059, "loss": 4.6028, "step": 749800 }, { "epoch": 1.3993030074714428, "grad_norm": 1.0288785696029663, "learning_rate": 0.000181391453161417, "loss": 4.5036, "step": 749850 }, { "epoch": 1.3993963130308211, "grad_norm": 1.408133864402771, "learning_rate": 0.00018138901238428653, "loss": 4.5087, "step": 749900 }, { "epoch": 1.3994896185901997, "grad_norm": 1.3699218034744263, "learning_rate": 0.00018138657146351877, "loss": 4.5979, "step": 749950 }, { "epoch": 1.399582924149578, "grad_norm": 1.3071013689041138, "learning_rate": 0.00018138413039911805, "loss": 4.4914, "step": 750000 }, { "epoch": 1.399582924149578, "eval_loss": 4.6659064292907715, "eval_runtime": 229.1701, "eval_samples_per_second": 11.38, "eval_steps_per_second": 11.38, "eval_tts_loss": 7.646634515602016, "step": 750000 }, { "epoch": 1.3996762297089567, "grad_norm": 1.3810770511627197, "learning_rate": 0.00018138168919108866, "loss": 4.6663, "step": 750050 }, { "epoch": 1.399769535268335, "grad_norm": 1.0062932968139648, "learning_rate": 0.0001813792478394349, "loss": 4.5754, "step": 750100 }, { "epoch": 1.3998628408277136, "grad_norm": 1.303175449371338, "learning_rate": 0.00018137680634416109, "loss": 4.1712, "step": 750150 }, { "epoch": 1.3999561463870922, "grad_norm": 0.9073581099510193, "learning_rate": 0.0001813743647052715, "loss": 4.381, "step": 750200 }, { "epoch": 1.4000494519464706, "grad_norm": 0.6323612928390503, "learning_rate": 0.0001813719229227705, "loss": 4.2899, "step": 750250 }, { "epoch": 1.4001427575058492, "grad_norm": 1.0785611867904663, "learning_rate": 0.0001813694809966624, "loss": 4.7587, "step": 750300 }, { "epoch": 1.4002360630652275, "grad_norm": 1.104772925376892, "learning_rate": 0.0001813670389269515, "loss": 4.3802, "step": 750350 }, { "epoch": 1.400329368624606, "grad_norm": 1.345407485961914, "learning_rate": 0.00018136459671364203, "loss": 4.5418, "step": 750400 }, { "epoch": 1.4004226741839845, "grad_norm": 1.2825371026992798, "learning_rate": 0.0001813621543567384, "loss": 4.647, "step": 750450 }, { "epoch": 1.400515979743363, "grad_norm": 0.9695392847061157, "learning_rate": 0.00018135971185624485, "loss": 4.4187, "step": 750500 }, { "epoch": 1.4006092853027416, "grad_norm": 1.2502747774124146, "learning_rate": 0.00018135726921216576, "loss": 4.4868, "step": 750550 }, { "epoch": 1.40070259086212, "grad_norm": 0.6947107911109924, "learning_rate": 0.00018135482642450538, "loss": 4.4627, "step": 750600 }, { "epoch": 1.4007958964214986, "grad_norm": 1.0346921682357788, "learning_rate": 0.0001813523834932681, "loss": 4.4005, "step": 750650 }, { "epoch": 1.400889201980877, "grad_norm": 1.1918665170669556, "learning_rate": 0.00018134994041845811, "loss": 4.4218, "step": 750700 }, { "epoch": 1.4009825075402556, "grad_norm": 0.8686269521713257, "learning_rate": 0.00018134749720007982, "loss": 4.3636, "step": 750750 }, { "epoch": 1.401075813099634, "grad_norm": 1.2843174934387207, "learning_rate": 0.0001813450538381375, "loss": 4.4959, "step": 750800 }, { "epoch": 1.4011691186590125, "grad_norm": 1.0479570627212524, "learning_rate": 0.00018134261033263548, "loss": 4.4259, "step": 750850 }, { "epoch": 1.401262424218391, "grad_norm": 0.9469064474105835, "learning_rate": 0.00018134016668357807, "loss": 4.4923, "step": 750900 }, { "epoch": 1.4013557297777695, "grad_norm": 1.2337836027145386, "learning_rate": 0.00018133772289096958, "loss": 4.4663, "step": 750950 }, { "epoch": 1.401449035337148, "grad_norm": 1.1494117975234985, "learning_rate": 0.0001813352789548143, "loss": 4.3756, "step": 751000 }, { "epoch": 1.4015423408965264, "grad_norm": 1.3942409753799438, "learning_rate": 0.00018133283487511658, "loss": 4.5973, "step": 751050 }, { "epoch": 1.401635646455905, "grad_norm": 0.954532265663147, "learning_rate": 0.00018133039065188072, "loss": 4.1357, "step": 751100 }, { "epoch": 1.4017289520152834, "grad_norm": 1.3457560539245605, "learning_rate": 0.00018132794628511104, "loss": 4.5775, "step": 751150 }, { "epoch": 1.401822257574662, "grad_norm": 1.2417519092559814, "learning_rate": 0.00018132550177481183, "loss": 4.5677, "step": 751200 }, { "epoch": 1.4019155631340405, "grad_norm": 1.2547211647033691, "learning_rate": 0.0001813230571209874, "loss": 4.5298, "step": 751250 }, { "epoch": 1.402008868693419, "grad_norm": 1.1811082363128662, "learning_rate": 0.00018132061232364213, "loss": 4.7601, "step": 751300 }, { "epoch": 1.4021021742527975, "grad_norm": 1.0948911905288696, "learning_rate": 0.00018131816738278024, "loss": 4.5313, "step": 751350 }, { "epoch": 1.4021954798121758, "grad_norm": 1.3452109098434448, "learning_rate": 0.00018131572229840613, "loss": 4.6487, "step": 751400 }, { "epoch": 1.4022887853715544, "grad_norm": 1.0783995389938354, "learning_rate": 0.00018131327707052405, "loss": 4.5296, "step": 751450 }, { "epoch": 1.4023820909309328, "grad_norm": 1.4549293518066406, "learning_rate": 0.00018131083169913836, "loss": 4.5746, "step": 751500 }, { "epoch": 1.4024753964903114, "grad_norm": 0.9473280310630798, "learning_rate": 0.00018130838618425336, "loss": 4.6686, "step": 751550 }, { "epoch": 1.40256870204969, "grad_norm": 1.310096263885498, "learning_rate": 0.00018130594052587334, "loss": 4.4615, "step": 751600 }, { "epoch": 1.4026620076090683, "grad_norm": 1.3425230979919434, "learning_rate": 0.00018130349472400269, "loss": 4.647, "step": 751650 }, { "epoch": 1.402755313168447, "grad_norm": 1.322732925415039, "learning_rate": 0.00018130104877864565, "loss": 4.4952, "step": 751700 }, { "epoch": 1.4028486187278253, "grad_norm": 0.94557785987854, "learning_rate": 0.00018129860268980657, "loss": 4.5389, "step": 751750 }, { "epoch": 1.4029419242872039, "grad_norm": 1.3913589715957642, "learning_rate": 0.0001812961564574897, "loss": 4.6098, "step": 751800 }, { "epoch": 1.4030352298465822, "grad_norm": 1.1062239408493042, "learning_rate": 0.0001812937100816995, "loss": 4.4806, "step": 751850 }, { "epoch": 1.4031285354059608, "grad_norm": 0.892577052116394, "learning_rate": 0.0001812912635624402, "loss": 4.5152, "step": 751900 }, { "epoch": 1.4032218409653394, "grad_norm": 1.3997602462768555, "learning_rate": 0.00018128881689971607, "loss": 4.6063, "step": 751950 }, { "epoch": 1.4033151465247178, "grad_norm": 1.410657525062561, "learning_rate": 0.00018128637009353154, "loss": 4.5716, "step": 752000 }, { "epoch": 1.4034084520840964, "grad_norm": 1.2767425775527954, "learning_rate": 0.00018128392314389083, "loss": 4.3379, "step": 752050 }, { "epoch": 1.4035017576434747, "grad_norm": 1.3148198127746582, "learning_rate": 0.00018128147605079832, "loss": 4.5392, "step": 752100 }, { "epoch": 1.4035950632028533, "grad_norm": 0.9114159345626831, "learning_rate": 0.00018127902881425828, "loss": 4.461, "step": 752150 }, { "epoch": 1.4036883687622317, "grad_norm": 1.0078926086425781, "learning_rate": 0.0001812765814342751, "loss": 4.2428, "step": 752200 }, { "epoch": 1.4037816743216103, "grad_norm": 0.9964181780815125, "learning_rate": 0.00018127413391085301, "loss": 4.6167, "step": 752250 }, { "epoch": 1.4038749798809889, "grad_norm": 1.2222840785980225, "learning_rate": 0.0001812716862439964, "loss": 4.497, "step": 752300 }, { "epoch": 1.4039682854403672, "grad_norm": 1.3474773168563843, "learning_rate": 0.0001812692384337096, "loss": 4.3931, "step": 752350 }, { "epoch": 1.4040615909997458, "grad_norm": 1.2632108926773071, "learning_rate": 0.00018126679047999683, "loss": 4.4993, "step": 752400 }, { "epoch": 1.4041548965591242, "grad_norm": 1.0086350440979004, "learning_rate": 0.00018126434238286248, "loss": 4.3586, "step": 752450 }, { "epoch": 1.4042482021185028, "grad_norm": 1.1757704019546509, "learning_rate": 0.0001812618941423109, "loss": 4.6933, "step": 752500 }, { "epoch": 1.4043415076778811, "grad_norm": 1.4259330034255981, "learning_rate": 0.00018125944575834633, "loss": 4.3825, "step": 752550 }, { "epoch": 1.4044348132372597, "grad_norm": 0.984950840473175, "learning_rate": 0.00018125699723097315, "loss": 4.552, "step": 752600 }, { "epoch": 1.4045281187966383, "grad_norm": 1.090908408164978, "learning_rate": 0.0001812545485601957, "loss": 4.5659, "step": 752650 }, { "epoch": 1.4046214243560167, "grad_norm": 0.9569710493087769, "learning_rate": 0.0001812520997460182, "loss": 4.4475, "step": 752700 }, { "epoch": 1.4047147299153953, "grad_norm": 1.3750537633895874, "learning_rate": 0.00018124965078844513, "loss": 4.5148, "step": 752750 }, { "epoch": 1.4048080354747736, "grad_norm": 1.240437626838684, "learning_rate": 0.00018124720168748065, "loss": 4.5591, "step": 752800 }, { "epoch": 1.4049013410341522, "grad_norm": 1.185357689857483, "learning_rate": 0.0001812447524431292, "loss": 4.3475, "step": 752850 }, { "epoch": 1.4049946465935306, "grad_norm": 1.37898850440979, "learning_rate": 0.00018124230305539504, "loss": 4.7217, "step": 752900 }, { "epoch": 1.4050879521529092, "grad_norm": 1.04668390750885, "learning_rate": 0.00018123985352428253, "loss": 4.4683, "step": 752950 }, { "epoch": 1.4051812577122877, "grad_norm": 1.2033179998397827, "learning_rate": 0.00018123740384979597, "loss": 4.3967, "step": 753000 }, { "epoch": 1.4051812577122877, "eval_loss": 4.675203323364258, "eval_runtime": 230.3986, "eval_samples_per_second": 11.32, "eval_steps_per_second": 11.32, "eval_tts_loss": 7.627497483736213, "step": 753000 }, { "epoch": 1.405274563271666, "grad_norm": 1.5615344047546387, "learning_rate": 0.00018123495403193966, "loss": 4.5477, "step": 753050 }, { "epoch": 1.4053678688310447, "grad_norm": 1.4519710540771484, "learning_rate": 0.00018123250407071795, "loss": 4.663, "step": 753100 }, { "epoch": 1.405461174390423, "grad_norm": 1.363918662071228, "learning_rate": 0.0001812300539661352, "loss": 4.2627, "step": 753150 }, { "epoch": 1.4055544799498016, "grad_norm": 1.341680645942688, "learning_rate": 0.00018122760371819567, "loss": 4.4476, "step": 753200 }, { "epoch": 1.40564778550918, "grad_norm": 1.2611197233200073, "learning_rate": 0.00018122515332690372, "loss": 4.7653, "step": 753250 }, { "epoch": 1.4057410910685586, "grad_norm": 1.2973344326019287, "learning_rate": 0.00018122270279226363, "loss": 4.5566, "step": 753300 }, { "epoch": 1.4058343966279372, "grad_norm": 1.3508353233337402, "learning_rate": 0.0001812202521142798, "loss": 4.3369, "step": 753350 }, { "epoch": 1.4059277021873156, "grad_norm": 0.8899083137512207, "learning_rate": 0.00018121780129295653, "loss": 4.4786, "step": 753400 }, { "epoch": 1.4060210077466941, "grad_norm": 1.25014328956604, "learning_rate": 0.00018121535032829811, "loss": 4.6042, "step": 753450 }, { "epoch": 1.4061143133060725, "grad_norm": 1.2352862358093262, "learning_rate": 0.00018121289922030885, "loss": 4.3729, "step": 753500 }, { "epoch": 1.406207618865451, "grad_norm": 1.1526981592178345, "learning_rate": 0.00018121044796899318, "loss": 4.4085, "step": 753550 }, { "epoch": 1.4063009244248295, "grad_norm": 0.7834089994430542, "learning_rate": 0.00018120799657435533, "loss": 4.6874, "step": 753600 }, { "epoch": 1.406394229984208, "grad_norm": 1.2034504413604736, "learning_rate": 0.00018120554503639965, "loss": 4.5004, "step": 753650 }, { "epoch": 1.4064875355435866, "grad_norm": 0.9010719656944275, "learning_rate": 0.0001812030933551305, "loss": 4.4895, "step": 753700 }, { "epoch": 1.406580841102965, "grad_norm": 1.1982351541519165, "learning_rate": 0.00018120064153055214, "loss": 4.5176, "step": 753750 }, { "epoch": 1.4066741466623436, "grad_norm": 1.2482162714004517, "learning_rate": 0.00018119818956266893, "loss": 4.5567, "step": 753800 }, { "epoch": 1.406767452221722, "grad_norm": 1.3605908155441284, "learning_rate": 0.00018119573745148523, "loss": 4.6168, "step": 753850 }, { "epoch": 1.4068607577811005, "grad_norm": 1.2579305171966553, "learning_rate": 0.00018119328519700535, "loss": 4.4284, "step": 753900 }, { "epoch": 1.406954063340479, "grad_norm": 1.5076669454574585, "learning_rate": 0.0001811908327992336, "loss": 4.5176, "step": 753950 }, { "epoch": 1.4070473688998575, "grad_norm": 1.260831594467163, "learning_rate": 0.00018118838025817427, "loss": 4.298, "step": 754000 }, { "epoch": 1.407140674459236, "grad_norm": 1.3094662427902222, "learning_rate": 0.00018118592757383178, "loss": 4.4185, "step": 754050 }, { "epoch": 1.4072339800186144, "grad_norm": 1.2245591878890991, "learning_rate": 0.0001811834747462104, "loss": 4.4659, "step": 754100 }, { "epoch": 1.407327285577993, "grad_norm": 0.9593348503112793, "learning_rate": 0.00018118102177531447, "loss": 4.5286, "step": 754150 }, { "epoch": 1.4074205911373714, "grad_norm": 1.285040259361267, "learning_rate": 0.00018117856866114833, "loss": 4.3368, "step": 754200 }, { "epoch": 1.40751389669675, "grad_norm": 1.2288318872451782, "learning_rate": 0.0001811761154037163, "loss": 4.4018, "step": 754250 }, { "epoch": 1.4076072022561283, "grad_norm": 1.2103488445281982, "learning_rate": 0.0001811736620030227, "loss": 4.4103, "step": 754300 }, { "epoch": 1.407700507815507, "grad_norm": 1.0883028507232666, "learning_rate": 0.0001811712084590719, "loss": 4.455, "step": 754350 }, { "epoch": 1.4077938133748855, "grad_norm": 1.148790955543518, "learning_rate": 0.00018116875477186816, "loss": 4.554, "step": 754400 }, { "epoch": 1.4078871189342639, "grad_norm": 1.2996517419815063, "learning_rate": 0.00018116630094141588, "loss": 4.4089, "step": 754450 }, { "epoch": 1.4079804244936425, "grad_norm": 1.3423142433166504, "learning_rate": 0.00018116384696771935, "loss": 4.7384, "step": 754500 }, { "epoch": 1.4080737300530208, "grad_norm": 0.9766300320625305, "learning_rate": 0.00018116139285078288, "loss": 4.6815, "step": 754550 }, { "epoch": 1.4081670356123994, "grad_norm": 1.3591853380203247, "learning_rate": 0.00018115893859061086, "loss": 4.6411, "step": 754600 }, { "epoch": 1.4082603411717778, "grad_norm": 1.237546443939209, "learning_rate": 0.0001811564841872076, "loss": 4.4208, "step": 754650 }, { "epoch": 1.4083536467311564, "grad_norm": 0.9792125225067139, "learning_rate": 0.00018115402964057742, "loss": 4.5552, "step": 754700 }, { "epoch": 1.408446952290535, "grad_norm": 1.1960304975509644, "learning_rate": 0.00018115157495072464, "loss": 4.6028, "step": 754750 }, { "epoch": 1.4085402578499133, "grad_norm": 1.3594309091567993, "learning_rate": 0.00018114912011765367, "loss": 4.4861, "step": 754800 }, { "epoch": 1.408633563409292, "grad_norm": 1.2784836292266846, "learning_rate": 0.00018114666514136874, "loss": 4.5876, "step": 754850 }, { "epoch": 1.4087268689686703, "grad_norm": 1.0439603328704834, "learning_rate": 0.00018114421002187419, "loss": 4.4756, "step": 754900 }, { "epoch": 1.4088201745280489, "grad_norm": 1.1469647884368896, "learning_rate": 0.0001811417547591744, "loss": 4.4801, "step": 754950 }, { "epoch": 1.4089134800874272, "grad_norm": 1.1643606424331665, "learning_rate": 0.00018113929935327375, "loss": 4.6398, "step": 755000 }, { "epoch": 1.4090067856468058, "grad_norm": 1.4443286657333374, "learning_rate": 0.00018113684380417647, "loss": 4.5752, "step": 755050 }, { "epoch": 1.4091000912061844, "grad_norm": 0.9405973553657532, "learning_rate": 0.00018113438811188695, "loss": 4.405, "step": 755100 }, { "epoch": 1.4091933967655628, "grad_norm": 1.1024534702301025, "learning_rate": 0.00018113193227640948, "loss": 4.5452, "step": 755150 }, { "epoch": 1.4092867023249414, "grad_norm": 1.2613942623138428, "learning_rate": 0.00018112947629774843, "loss": 4.5033, "step": 755200 }, { "epoch": 1.4093800078843197, "grad_norm": 1.2096219062805176, "learning_rate": 0.00018112702017590815, "loss": 4.7483, "step": 755250 }, { "epoch": 1.4094733134436983, "grad_norm": 1.0992473363876343, "learning_rate": 0.00018112456391089295, "loss": 4.5493, "step": 755300 }, { "epoch": 1.4095666190030767, "grad_norm": 1.1370683908462524, "learning_rate": 0.00018112210750270718, "loss": 4.5967, "step": 755350 }, { "epoch": 1.4096599245624553, "grad_norm": 1.7298223972320557, "learning_rate": 0.00018111965095135511, "loss": 4.4333, "step": 755400 }, { "epoch": 1.4097532301218338, "grad_norm": 1.1736270189285278, "learning_rate": 0.00018111719425684122, "loss": 4.4096, "step": 755450 }, { "epoch": 1.4098465356812122, "grad_norm": 1.1672894954681396, "learning_rate": 0.0001811147374191697, "loss": 4.3299, "step": 755500 }, { "epoch": 1.4099398412405908, "grad_norm": 1.2445381879806519, "learning_rate": 0.00018111228043834495, "loss": 4.5271, "step": 755550 }, { "epoch": 1.4100331467999692, "grad_norm": 1.1843805313110352, "learning_rate": 0.00018110982331437129, "loss": 4.6365, "step": 755600 }, { "epoch": 1.4101264523593477, "grad_norm": 1.2262697219848633, "learning_rate": 0.00018110736604725304, "loss": 4.5409, "step": 755650 }, { "epoch": 1.4102197579187261, "grad_norm": 0.952824056148529, "learning_rate": 0.0001811049086369946, "loss": 4.3588, "step": 755700 }, { "epoch": 1.4103130634781047, "grad_norm": 1.1851266622543335, "learning_rate": 0.00018110245108360026, "loss": 4.6432, "step": 755750 }, { "epoch": 1.4104063690374833, "grad_norm": 1.2045528888702393, "learning_rate": 0.00018109999338707436, "loss": 4.5159, "step": 755800 }, { "epoch": 1.4104996745968617, "grad_norm": 1.3131206035614014, "learning_rate": 0.00018109753554742123, "loss": 4.5499, "step": 755850 }, { "epoch": 1.4105929801562402, "grad_norm": 1.2962650060653687, "learning_rate": 0.00018109507756464526, "loss": 4.5097, "step": 755900 }, { "epoch": 1.4106862857156186, "grad_norm": 1.3557630777359009, "learning_rate": 0.00018109261943875072, "loss": 4.4471, "step": 755950 }, { "epoch": 1.4107795912749972, "grad_norm": 1.1807628870010376, "learning_rate": 0.00018109016116974196, "loss": 4.3831, "step": 756000 }, { "epoch": 1.4107795912749972, "eval_loss": 4.66172981262207, "eval_runtime": 231.3729, "eval_samples_per_second": 11.272, "eval_steps_per_second": 11.272, "eval_tts_loss": 7.594197759917577, "step": 756000 }, { "epoch": 1.4108728968343756, "grad_norm": 1.265444040298462, "learning_rate": 0.00018108770275762336, "loss": 4.557, "step": 756050 }, { "epoch": 1.4109662023937541, "grad_norm": 1.0888968706130981, "learning_rate": 0.0001810852442023992, "loss": 4.5107, "step": 756100 }, { "epoch": 1.4110595079531327, "grad_norm": 1.4977471828460693, "learning_rate": 0.00018108278550407387, "loss": 4.5175, "step": 756150 }, { "epoch": 1.411152813512511, "grad_norm": 1.2672964334487915, "learning_rate": 0.0001810803266626517, "loss": 4.4174, "step": 756200 }, { "epoch": 1.4112461190718897, "grad_norm": 1.2657983303070068, "learning_rate": 0.00018107786767813702, "loss": 4.4075, "step": 756250 }, { "epoch": 1.411339424631268, "grad_norm": 1.7145322561264038, "learning_rate": 0.00018107540855053417, "loss": 4.3631, "step": 756300 }, { "epoch": 1.4114327301906466, "grad_norm": 1.205639123916626, "learning_rate": 0.00018107294927984747, "loss": 4.5054, "step": 756350 }, { "epoch": 1.411526035750025, "grad_norm": 1.1812183856964111, "learning_rate": 0.0001810704898660813, "loss": 4.342, "step": 756400 }, { "epoch": 1.4116193413094036, "grad_norm": 1.133000373840332, "learning_rate": 0.00018106803030923995, "loss": 4.5607, "step": 756450 }, { "epoch": 1.4117126468687822, "grad_norm": 1.0306590795516968, "learning_rate": 0.00018106557060932783, "loss": 4.6783, "step": 756500 }, { "epoch": 1.4118059524281605, "grad_norm": 1.2804341316223145, "learning_rate": 0.00018106311076634922, "loss": 4.514, "step": 756550 }, { "epoch": 1.4118992579875391, "grad_norm": 1.1785862445831299, "learning_rate": 0.00018106065078030847, "loss": 4.4776, "step": 756600 }, { "epoch": 1.4119925635469175, "grad_norm": 0.897530198097229, "learning_rate": 0.00018105819065120996, "loss": 4.4958, "step": 756650 }, { "epoch": 1.412085869106296, "grad_norm": 1.3041436672210693, "learning_rate": 0.000181055730379058, "loss": 4.4589, "step": 756700 }, { "epoch": 1.4121791746656744, "grad_norm": 0.9818947315216064, "learning_rate": 0.0001810532699638569, "loss": 4.5432, "step": 756750 }, { "epoch": 1.412272480225053, "grad_norm": 1.3876928091049194, "learning_rate": 0.0001810508094056111, "loss": 4.5669, "step": 756800 }, { "epoch": 1.4123657857844316, "grad_norm": 1.4193178415298462, "learning_rate": 0.00018104834870432486, "loss": 4.4775, "step": 756850 }, { "epoch": 1.41245909134381, "grad_norm": 1.2440530061721802, "learning_rate": 0.00018104588786000254, "loss": 4.4637, "step": 756900 }, { "epoch": 1.4125523969031886, "grad_norm": 1.3843438625335693, "learning_rate": 0.0001810434268726485, "loss": 4.5204, "step": 756950 }, { "epoch": 1.412645702462567, "grad_norm": 1.4365525245666504, "learning_rate": 0.00018104096574226704, "loss": 4.3995, "step": 757000 }, { "epoch": 1.4127390080219455, "grad_norm": 1.0366308689117432, "learning_rate": 0.00018103850446886258, "loss": 4.5314, "step": 757050 }, { "epoch": 1.4128323135813239, "grad_norm": 1.263674020767212, "learning_rate": 0.0001810360430524394, "loss": 4.4395, "step": 757100 }, { "epoch": 1.4129256191407025, "grad_norm": 1.2899723052978516, "learning_rate": 0.00018103358149300185, "loss": 4.3019, "step": 757150 }, { "epoch": 1.413018924700081, "grad_norm": 1.0365010499954224, "learning_rate": 0.0001810311197905543, "loss": 4.6298, "step": 757200 }, { "epoch": 1.4131122302594594, "grad_norm": 0.8417057991027832, "learning_rate": 0.00018102865794510109, "loss": 4.378, "step": 757250 }, { "epoch": 1.4132055358188378, "grad_norm": 1.1272296905517578, "learning_rate": 0.00018102619595664654, "loss": 4.4171, "step": 757300 }, { "epoch": 1.4132988413782164, "grad_norm": 1.3105295896530151, "learning_rate": 0.000181023733825195, "loss": 4.5674, "step": 757350 }, { "epoch": 1.413392146937595, "grad_norm": 1.341001272201538, "learning_rate": 0.00018102127155075087, "loss": 4.475, "step": 757400 }, { "epoch": 1.4134854524969733, "grad_norm": 1.200605034828186, "learning_rate": 0.00018101880913331842, "loss": 4.2851, "step": 757450 }, { "epoch": 1.413578758056352, "grad_norm": 0.8877935409545898, "learning_rate": 0.000181016346572902, "loss": 4.4687, "step": 757500 }, { "epoch": 1.4136720636157305, "grad_norm": 1.1278319358825684, "learning_rate": 0.00018101388386950604, "loss": 4.5652, "step": 757550 }, { "epoch": 1.4137653691751089, "grad_norm": 1.4971600770950317, "learning_rate": 0.0001810114210231348, "loss": 4.3938, "step": 757600 }, { "epoch": 1.4138586747344872, "grad_norm": 1.466628909111023, "learning_rate": 0.00018100895803379268, "loss": 4.5182, "step": 757650 }, { "epoch": 1.4139519802938658, "grad_norm": 1.05574369430542, "learning_rate": 0.000181006494901484, "loss": 4.6292, "step": 757700 }, { "epoch": 1.4140452858532444, "grad_norm": 1.229044795036316, "learning_rate": 0.00018100403162621306, "loss": 4.4597, "step": 757750 }, { "epoch": 1.4141385914126228, "grad_norm": 1.4380329847335815, "learning_rate": 0.0001810015682079843, "loss": 4.6666, "step": 757800 }, { "epoch": 1.4142318969720014, "grad_norm": 1.201385498046875, "learning_rate": 0.000180999104646802, "loss": 4.4932, "step": 757850 }, { "epoch": 1.41432520253138, "grad_norm": 1.0387357473373413, "learning_rate": 0.00018099664094267057, "loss": 4.7351, "step": 757900 }, { "epoch": 1.4144185080907583, "grad_norm": 1.2229691743850708, "learning_rate": 0.0001809941770955943, "loss": 4.3904, "step": 757950 }, { "epoch": 1.4145118136501367, "grad_norm": 1.2191208600997925, "learning_rate": 0.00018099171310557755, "loss": 4.3005, "step": 758000 }, { "epoch": 1.4146051192095153, "grad_norm": 1.0868217945098877, "learning_rate": 0.0001809892489726247, "loss": 4.4999, "step": 758050 }, { "epoch": 1.4146984247688938, "grad_norm": 1.104870319366455, "learning_rate": 0.00018098678469674005, "loss": 4.4982, "step": 758100 }, { "epoch": 1.4147917303282722, "grad_norm": 0.9389768838882446, "learning_rate": 0.00018098432027792799, "loss": 4.4006, "step": 758150 }, { "epoch": 1.4148850358876508, "grad_norm": 1.1148312091827393, "learning_rate": 0.00018098185571619284, "loss": 4.3955, "step": 758200 }, { "epoch": 1.4149783414470294, "grad_norm": 1.200258731842041, "learning_rate": 0.000180979391011539, "loss": 4.5528, "step": 758250 }, { "epoch": 1.4150716470064078, "grad_norm": 1.0380362272262573, "learning_rate": 0.00018097692616397072, "loss": 4.6396, "step": 758300 }, { "epoch": 1.4151649525657861, "grad_norm": 1.305518627166748, "learning_rate": 0.00018097446117349247, "loss": 4.7112, "step": 758350 }, { "epoch": 1.4152582581251647, "grad_norm": 1.1728357076644897, "learning_rate": 0.00018097199604010853, "loss": 4.4634, "step": 758400 }, { "epoch": 1.4153515636845433, "grad_norm": 1.1617236137390137, "learning_rate": 0.00018096953076382323, "loss": 4.4555, "step": 758450 }, { "epoch": 1.4154448692439217, "grad_norm": 1.1427037715911865, "learning_rate": 0.000180967065344641, "loss": 4.5507, "step": 758500 }, { "epoch": 1.4155381748033002, "grad_norm": 1.0990989208221436, "learning_rate": 0.00018096459978256615, "loss": 4.4423, "step": 758550 }, { "epoch": 1.4156314803626788, "grad_norm": 1.3851333856582642, "learning_rate": 0.000180962134077603, "loss": 4.4725, "step": 758600 }, { "epoch": 1.4157247859220572, "grad_norm": 1.2131006717681885, "learning_rate": 0.00018095966822975594, "loss": 4.5149, "step": 758650 }, { "epoch": 1.4158180914814356, "grad_norm": 1.1497539281845093, "learning_rate": 0.0001809572022390293, "loss": 4.3716, "step": 758700 }, { "epoch": 1.4159113970408141, "grad_norm": 1.1728429794311523, "learning_rate": 0.00018095473610542747, "loss": 4.2497, "step": 758750 }, { "epoch": 1.4160047026001927, "grad_norm": 1.4330480098724365, "learning_rate": 0.00018095226982895474, "loss": 4.3805, "step": 758800 }, { "epoch": 1.416098008159571, "grad_norm": 1.3654059171676636, "learning_rate": 0.00018094980340961555, "loss": 4.4335, "step": 758850 }, { "epoch": 1.4161913137189497, "grad_norm": 0.5898260474205017, "learning_rate": 0.00018094733684741413, "loss": 4.3659, "step": 758900 }, { "epoch": 1.4162846192783283, "grad_norm": 0.8158581852912903, "learning_rate": 0.00018094487014235492, "loss": 4.3783, "step": 758950 }, { "epoch": 1.4163779248377066, "grad_norm": 1.0536508560180664, "learning_rate": 0.0001809424032944423, "loss": 4.414, "step": 759000 }, { "epoch": 1.4163779248377066, "eval_loss": 4.666611671447754, "eval_runtime": 231.1499, "eval_samples_per_second": 11.283, "eval_steps_per_second": 11.283, "eval_tts_loss": 7.618569120786159, "step": 759000 }, { "epoch": 1.416471230397085, "grad_norm": 1.2087711095809937, "learning_rate": 0.00018093993630368055, "loss": 4.6816, "step": 759050 }, { "epoch": 1.4165645359564636, "grad_norm": 1.2877044677734375, "learning_rate": 0.0001809374691700741, "loss": 4.5132, "step": 759100 }, { "epoch": 1.4166578415158422, "grad_norm": 1.1033086776733398, "learning_rate": 0.0001809350018936272, "loss": 4.604, "step": 759150 }, { "epoch": 1.4167511470752205, "grad_norm": 0.9428020715713501, "learning_rate": 0.00018093253447434425, "loss": 4.7329, "step": 759200 }, { "epoch": 1.4168444526345991, "grad_norm": 0.9632729887962341, "learning_rate": 0.00018093006691222968, "loss": 4.3896, "step": 759250 }, { "epoch": 1.4169377581939777, "grad_norm": 0.7582852244377136, "learning_rate": 0.00018092759920728774, "loss": 4.2966, "step": 759300 }, { "epoch": 1.417031063753356, "grad_norm": 1.2549529075622559, "learning_rate": 0.00018092513135952283, "loss": 4.335, "step": 759350 }, { "epoch": 1.4171243693127344, "grad_norm": 1.2859209775924683, "learning_rate": 0.00018092266336893932, "loss": 4.5165, "step": 759400 }, { "epoch": 1.417217674872113, "grad_norm": 1.0312479734420776, "learning_rate": 0.00018092019523554154, "loss": 4.2101, "step": 759450 }, { "epoch": 1.4173109804314916, "grad_norm": 1.5824967622756958, "learning_rate": 0.00018091772695933385, "loss": 4.4625, "step": 759500 }, { "epoch": 1.41740428599087, "grad_norm": 1.3354967832565308, "learning_rate": 0.0001809152585403206, "loss": 4.5682, "step": 759550 }, { "epoch": 1.4174975915502486, "grad_norm": 1.3541855812072754, "learning_rate": 0.00018091278997850617, "loss": 4.4551, "step": 759600 }, { "epoch": 1.4175908971096272, "grad_norm": 1.0990486145019531, "learning_rate": 0.0001809103212738949, "loss": 4.5717, "step": 759650 }, { "epoch": 1.4176842026690055, "grad_norm": 0.7515814900398254, "learning_rate": 0.00018090785242649116, "loss": 4.4005, "step": 759700 }, { "epoch": 1.4177775082283839, "grad_norm": 1.3272322416305542, "learning_rate": 0.0001809053834362993, "loss": 4.3777, "step": 759750 }, { "epoch": 1.4178708137877625, "grad_norm": 1.3668854236602783, "learning_rate": 0.00018090291430332365, "loss": 4.4356, "step": 759800 }, { "epoch": 1.417964119347141, "grad_norm": 1.4488301277160645, "learning_rate": 0.0001809004450275686, "loss": 4.3518, "step": 759850 }, { "epoch": 1.4180574249065194, "grad_norm": 1.1320677995681763, "learning_rate": 0.0001808979756090385, "loss": 4.3, "step": 759900 }, { "epoch": 1.418150730465898, "grad_norm": 0.9284247756004333, "learning_rate": 0.0001808955060477377, "loss": 4.4808, "step": 759950 }, { "epoch": 1.4182440360252766, "grad_norm": 0.9685070514678955, "learning_rate": 0.0001808930363436706, "loss": 4.5439, "step": 760000 }, { "epoch": 1.418337341584655, "grad_norm": 1.3964598178863525, "learning_rate": 0.00018089056649684147, "loss": 4.8191, "step": 760050 }, { "epoch": 1.4184306471440333, "grad_norm": 1.291913390159607, "learning_rate": 0.00018088809650725475, "loss": 4.5334, "step": 760100 }, { "epoch": 1.418523952703412, "grad_norm": 1.0546672344207764, "learning_rate": 0.00018088562637491479, "loss": 4.5707, "step": 760150 }, { "epoch": 1.4186172582627905, "grad_norm": 0.9526347517967224, "learning_rate": 0.00018088315609982592, "loss": 4.6617, "step": 760200 }, { "epoch": 1.4187105638221689, "grad_norm": 0.9171462059020996, "learning_rate": 0.0001808806856819925, "loss": 4.6393, "step": 760250 }, { "epoch": 1.4188038693815475, "grad_norm": 1.194250226020813, "learning_rate": 0.00018087821512141893, "loss": 4.6884, "step": 760300 }, { "epoch": 1.418897174940926, "grad_norm": 1.2326525449752808, "learning_rate": 0.0001808757444181095, "loss": 4.5756, "step": 760350 }, { "epoch": 1.4189904805003044, "grad_norm": 1.2667412757873535, "learning_rate": 0.00018087327357206866, "loss": 4.5241, "step": 760400 }, { "epoch": 1.4190837860596828, "grad_norm": 1.1117559671401978, "learning_rate": 0.00018087080258330068, "loss": 4.4885, "step": 760450 }, { "epoch": 1.4191770916190614, "grad_norm": 1.0282741785049438, "learning_rate": 0.00018086833145181, "loss": 4.4563, "step": 760500 }, { "epoch": 1.41927039717844, "grad_norm": 1.4156287908554077, "learning_rate": 0.0001808658601776009, "loss": 4.4245, "step": 760550 }, { "epoch": 1.4193637027378183, "grad_norm": 1.0417289733886719, "learning_rate": 0.0001808633887606778, "loss": 4.6201, "step": 760600 }, { "epoch": 1.419457008297197, "grad_norm": 1.1251620054244995, "learning_rate": 0.00018086091720104507, "loss": 4.6166, "step": 760650 }, { "epoch": 1.4195503138565755, "grad_norm": 1.4621045589447021, "learning_rate": 0.00018085844549870704, "loss": 4.6135, "step": 760700 }, { "epoch": 1.4196436194159539, "grad_norm": 0.8972177505493164, "learning_rate": 0.00018085597365366806, "loss": 4.5295, "step": 760750 }, { "epoch": 1.4197369249753322, "grad_norm": 1.0228012800216675, "learning_rate": 0.00018085350166593253, "loss": 4.2787, "step": 760800 }, { "epoch": 1.4198302305347108, "grad_norm": 1.3379920721054077, "learning_rate": 0.0001808510295355048, "loss": 4.4451, "step": 760850 }, { "epoch": 1.4199235360940894, "grad_norm": 1.3583483695983887, "learning_rate": 0.0001808485572623892, "loss": 4.5436, "step": 760900 }, { "epoch": 1.4200168416534678, "grad_norm": 1.158334732055664, "learning_rate": 0.00018084608484659017, "loss": 4.2975, "step": 760950 }, { "epoch": 1.4201101472128463, "grad_norm": 1.079274296760559, "learning_rate": 0.000180843612288112, "loss": 4.4814, "step": 761000 }, { "epoch": 1.420203452772225, "grad_norm": 1.4939348697662354, "learning_rate": 0.00018084113958695907, "loss": 4.5123, "step": 761050 }, { "epoch": 1.4202967583316033, "grad_norm": 1.0729737281799316, "learning_rate": 0.00018083866674313578, "loss": 4.3506, "step": 761100 }, { "epoch": 1.4203900638909817, "grad_norm": 1.1155366897583008, "learning_rate": 0.00018083619375664644, "loss": 4.3275, "step": 761150 }, { "epoch": 1.4204833694503602, "grad_norm": 1.2313144207000732, "learning_rate": 0.00018083372062749546, "loss": 4.4626, "step": 761200 }, { "epoch": 1.4205766750097388, "grad_norm": 1.0425912141799927, "learning_rate": 0.00018083124735568718, "loss": 4.373, "step": 761250 }, { "epoch": 1.4206699805691172, "grad_norm": 1.119278907775879, "learning_rate": 0.00018082877394122595, "loss": 4.6009, "step": 761300 }, { "epoch": 1.4207632861284958, "grad_norm": 1.2457655668258667, "learning_rate": 0.0001808263003841162, "loss": 4.4311, "step": 761350 }, { "epoch": 1.4208565916878744, "grad_norm": 1.413806438446045, "learning_rate": 0.00018082382668436223, "loss": 4.4409, "step": 761400 }, { "epoch": 1.4209498972472527, "grad_norm": 0.9332594871520996, "learning_rate": 0.0001808213528419684, "loss": 4.4728, "step": 761450 }, { "epoch": 1.421043202806631, "grad_norm": 0.8530699610710144, "learning_rate": 0.00018081887885693914, "loss": 4.2883, "step": 761500 }, { "epoch": 1.4211365083660097, "grad_norm": 1.289154291152954, "learning_rate": 0.00018081640472927878, "loss": 4.5275, "step": 761550 }, { "epoch": 1.4212298139253883, "grad_norm": 1.0198618173599243, "learning_rate": 0.00018081393045899167, "loss": 4.2108, "step": 761600 }, { "epoch": 1.4213231194847666, "grad_norm": 1.3132344484329224, "learning_rate": 0.0001808114560460822, "loss": 4.4259, "step": 761650 }, { "epoch": 1.4214164250441452, "grad_norm": 1.7252916097640991, "learning_rate": 0.00018080898149055473, "loss": 4.5073, "step": 761700 }, { "epoch": 1.4215097306035238, "grad_norm": 1.0251563787460327, "learning_rate": 0.0001808065067924136, "loss": 4.3634, "step": 761750 }, { "epoch": 1.4216030361629022, "grad_norm": 1.2410342693328857, "learning_rate": 0.00018080403195166325, "loss": 4.553, "step": 761800 }, { "epoch": 1.4216963417222805, "grad_norm": 1.0780324935913086, "learning_rate": 0.00018080155696830797, "loss": 4.5026, "step": 761850 }, { "epoch": 1.4217896472816591, "grad_norm": 1.1388055086135864, "learning_rate": 0.00018079908184235216, "loss": 4.6911, "step": 761900 }, { "epoch": 1.4218829528410377, "grad_norm": 1.297707200050354, "learning_rate": 0.0001807966065738002, "loss": 4.3925, "step": 761950 }, { "epoch": 1.421976258400416, "grad_norm": 1.549561619758606, "learning_rate": 0.0001807941311626564, "loss": 4.5329, "step": 762000 }, { "epoch": 1.421976258400416, "eval_loss": 4.664612293243408, "eval_runtime": 231.0218, "eval_samples_per_second": 11.289, "eval_steps_per_second": 11.289, "eval_tts_loss": 7.6488346556213, "step": 762000 }, { "epoch": 1.4220695639597947, "grad_norm": 0.9171646237373352, "learning_rate": 0.00018079165560892525, "loss": 4.5311, "step": 762050 }, { "epoch": 1.4221628695191733, "grad_norm": 0.9633502960205078, "learning_rate": 0.000180789179912611, "loss": 4.3825, "step": 762100 }, { "epoch": 1.4222561750785516, "grad_norm": 1.1474465131759644, "learning_rate": 0.00018078670407371808, "loss": 4.5244, "step": 762150 }, { "epoch": 1.42234948063793, "grad_norm": 1.0521365404129028, "learning_rate": 0.00018078422809225084, "loss": 4.5793, "step": 762200 }, { "epoch": 1.4224427861973086, "grad_norm": 1.3407196998596191, "learning_rate": 0.0001807817519682136, "loss": 4.6409, "step": 762250 }, { "epoch": 1.4225360917566872, "grad_norm": 1.099604845046997, "learning_rate": 0.00018077927570161087, "loss": 4.4068, "step": 762300 }, { "epoch": 1.4226293973160655, "grad_norm": 1.3070287704467773, "learning_rate": 0.00018077679929244685, "loss": 4.3551, "step": 762350 }, { "epoch": 1.4227227028754441, "grad_norm": 1.3101091384887695, "learning_rate": 0.00018077432274072605, "loss": 4.3426, "step": 762400 }, { "epoch": 1.4228160084348227, "grad_norm": 0.723798930644989, "learning_rate": 0.00018077184604645277, "loss": 4.4612, "step": 762450 }, { "epoch": 1.422909313994201, "grad_norm": 1.2410943508148193, "learning_rate": 0.00018076936920963138, "loss": 4.7303, "step": 762500 }, { "epoch": 1.4230026195535794, "grad_norm": 1.1212464570999146, "learning_rate": 0.00018076689223026627, "loss": 4.4676, "step": 762550 }, { "epoch": 1.423095925112958, "grad_norm": 1.4662188291549683, "learning_rate": 0.0001807644151083618, "loss": 4.5136, "step": 762600 }, { "epoch": 1.4231892306723366, "grad_norm": 1.3476225137710571, "learning_rate": 0.00018076193784392238, "loss": 4.3283, "step": 762650 }, { "epoch": 1.423282536231715, "grad_norm": 0.9672389626502991, "learning_rate": 0.00018075946043695233, "loss": 4.481, "step": 762700 }, { "epoch": 1.4233758417910936, "grad_norm": 0.9978131055831909, "learning_rate": 0.00018075698288745604, "loss": 4.3956, "step": 762750 }, { "epoch": 1.4234691473504721, "grad_norm": 1.0551705360412598, "learning_rate": 0.00018075450519543788, "loss": 4.5978, "step": 762800 }, { "epoch": 1.4235624529098505, "grad_norm": 1.3335601091384888, "learning_rate": 0.00018075202736090223, "loss": 4.3818, "step": 762850 }, { "epoch": 1.4236557584692289, "grad_norm": 1.25326669216156, "learning_rate": 0.00018074954938385347, "loss": 4.5951, "step": 762900 }, { "epoch": 1.4237490640286075, "grad_norm": 1.051470398902893, "learning_rate": 0.00018074707126429596, "loss": 4.5171, "step": 762950 }, { "epoch": 1.423842369587986, "grad_norm": 1.1813299655914307, "learning_rate": 0.00018074459300223406, "loss": 4.4025, "step": 763000 }, { "epoch": 1.4239356751473644, "grad_norm": 1.0822112560272217, "learning_rate": 0.00018074211459767218, "loss": 4.4187, "step": 763050 }, { "epoch": 1.424028980706743, "grad_norm": 0.9034096598625183, "learning_rate": 0.00018073963605061465, "loss": 4.6138, "step": 763100 }, { "epoch": 1.4241222862661216, "grad_norm": 1.3311245441436768, "learning_rate": 0.00018073715736106592, "loss": 4.5384, "step": 763150 }, { "epoch": 1.4242155918255, "grad_norm": 0.9163987636566162, "learning_rate": 0.00018073467852903027, "loss": 4.4951, "step": 763200 }, { "epoch": 1.4243088973848783, "grad_norm": 0.9637299180030823, "learning_rate": 0.00018073219955451215, "loss": 4.6686, "step": 763250 }, { "epoch": 1.424402202944257, "grad_norm": 1.6130867004394531, "learning_rate": 0.00018072972043751588, "loss": 4.4463, "step": 763300 }, { "epoch": 1.4244955085036355, "grad_norm": 0.9424489736557007, "learning_rate": 0.00018072724117804584, "loss": 4.5847, "step": 763350 }, { "epoch": 1.4245888140630139, "grad_norm": 1.0344994068145752, "learning_rate": 0.00018072476177610644, "loss": 4.3394, "step": 763400 }, { "epoch": 1.4246821196223924, "grad_norm": 1.2635926008224487, "learning_rate": 0.00018072228223170206, "loss": 4.5103, "step": 763450 }, { "epoch": 1.424775425181771, "grad_norm": 0.9642047882080078, "learning_rate": 0.000180719802544837, "loss": 4.5467, "step": 763500 }, { "epoch": 1.4248687307411494, "grad_norm": 1.2532107830047607, "learning_rate": 0.00018071732271551577, "loss": 4.4316, "step": 763550 }, { "epoch": 1.4249620363005278, "grad_norm": 1.309352159500122, "learning_rate": 0.0001807148427437426, "loss": 4.5755, "step": 763600 }, { "epoch": 1.4250553418599063, "grad_norm": 0.9566655158996582, "learning_rate": 0.000180712362629522, "loss": 4.4264, "step": 763650 }, { "epoch": 1.425148647419285, "grad_norm": 1.3844624757766724, "learning_rate": 0.00018070988237285825, "loss": 4.5551, "step": 763700 }, { "epoch": 1.4252419529786633, "grad_norm": 1.1607639789581299, "learning_rate": 0.00018070740197375574, "loss": 4.3743, "step": 763750 }, { "epoch": 1.4253352585380419, "grad_norm": 1.1049363613128662, "learning_rate": 0.0001807049214322189, "loss": 4.5524, "step": 763800 }, { "epoch": 1.4254285640974205, "grad_norm": 1.5078009366989136, "learning_rate": 0.00018070244074825206, "loss": 4.7372, "step": 763850 }, { "epoch": 1.4255218696567988, "grad_norm": 1.3951081037521362, "learning_rate": 0.0001806999599218596, "loss": 4.5349, "step": 763900 }, { "epoch": 1.4256151752161772, "grad_norm": 1.4398036003112793, "learning_rate": 0.0001806974789530459, "loss": 4.5599, "step": 763950 }, { "epoch": 1.4257084807755558, "grad_norm": 1.030371069908142, "learning_rate": 0.00018069499784181538, "loss": 4.4646, "step": 764000 }, { "epoch": 1.4258017863349344, "grad_norm": 1.479935884475708, "learning_rate": 0.00018069251658817237, "loss": 4.4501, "step": 764050 }, { "epoch": 1.4258950918943127, "grad_norm": 1.3881566524505615, "learning_rate": 0.00018069003519212128, "loss": 4.6296, "step": 764100 }, { "epoch": 1.4259883974536913, "grad_norm": 1.1931768655776978, "learning_rate": 0.00018068755365366646, "loss": 4.3786, "step": 764150 }, { "epoch": 1.42608170301307, "grad_norm": 1.1732676029205322, "learning_rate": 0.00018068507197281232, "loss": 4.3886, "step": 764200 }, { "epoch": 1.4261750085724483, "grad_norm": 1.0092607736587524, "learning_rate": 0.0001806825901495632, "loss": 4.5329, "step": 764250 }, { "epoch": 1.4262683141318266, "grad_norm": 0.8263181447982788, "learning_rate": 0.00018068010818392353, "loss": 4.4257, "step": 764300 }, { "epoch": 1.4263616196912052, "grad_norm": 0.8855903148651123, "learning_rate": 0.00018067762607589763, "loss": 4.3887, "step": 764350 }, { "epoch": 1.4264549252505838, "grad_norm": 1.334502100944519, "learning_rate": 0.00018067514382548997, "loss": 4.5669, "step": 764400 }, { "epoch": 1.4265482308099622, "grad_norm": 1.0588082075119019, "learning_rate": 0.00018067266143270485, "loss": 4.5298, "step": 764450 }, { "epoch": 1.4266415363693408, "grad_norm": 1.2949501276016235, "learning_rate": 0.00018067017889754667, "loss": 4.4864, "step": 764500 }, { "epoch": 1.4267348419287194, "grad_norm": 1.3025978803634644, "learning_rate": 0.0001806676962200198, "loss": 4.5723, "step": 764550 }, { "epoch": 1.4268281474880977, "grad_norm": 1.3379372358322144, "learning_rate": 0.0001806652134001287, "loss": 4.5407, "step": 764600 }, { "epoch": 1.426921453047476, "grad_norm": 1.1428602933883667, "learning_rate": 0.00018066273043787764, "loss": 4.6068, "step": 764650 }, { "epoch": 1.4270147586068547, "grad_norm": 1.1264219284057617, "learning_rate": 0.0001806602473332711, "loss": 4.4868, "step": 764700 }, { "epoch": 1.4271080641662333, "grad_norm": 1.3247716426849365, "learning_rate": 0.00018065776408631337, "loss": 4.4458, "step": 764750 }, { "epoch": 1.4272013697256116, "grad_norm": 1.0201964378356934, "learning_rate": 0.0001806552806970089, "loss": 4.4968, "step": 764800 }, { "epoch": 1.4272946752849902, "grad_norm": 1.3889565467834473, "learning_rate": 0.00018065279716536202, "loss": 4.5756, "step": 764850 }, { "epoch": 1.4273879808443688, "grad_norm": 1.1165498495101929, "learning_rate": 0.0001806503134913772, "loss": 4.4646, "step": 764900 }, { "epoch": 1.4274812864037472, "grad_norm": 1.1695499420166016, "learning_rate": 0.00018064782967505872, "loss": 4.4879, "step": 764950 }, { "epoch": 1.4275745919631255, "grad_norm": 1.1555471420288086, "learning_rate": 0.00018064534571641104, "loss": 4.5095, "step": 765000 }, { "epoch": 1.4275745919631255, "eval_loss": 4.667006492614746, "eval_runtime": 233.511, "eval_samples_per_second": 11.169, "eval_steps_per_second": 11.169, "eval_tts_loss": 7.5989065228481465, "step": 765000 }, { "epoch": 1.4276678975225041, "grad_norm": 1.148463249206543, "learning_rate": 0.0001806428616154385, "loss": 4.5535, "step": 765050 }, { "epoch": 1.4277612030818827, "grad_norm": 1.0166361331939697, "learning_rate": 0.0001806403773721455, "loss": 4.3507, "step": 765100 }, { "epoch": 1.427854508641261, "grad_norm": 1.0395158529281616, "learning_rate": 0.00018063789298653643, "loss": 4.6513, "step": 765150 }, { "epoch": 1.4279478142006397, "grad_norm": 1.087009072303772, "learning_rate": 0.00018063540845861565, "loss": 4.5266, "step": 765200 }, { "epoch": 1.428041119760018, "grad_norm": 1.012269139289856, "learning_rate": 0.0001806329237883876, "loss": 4.4601, "step": 765250 }, { "epoch": 1.4281344253193966, "grad_norm": 1.5610164403915405, "learning_rate": 0.0001806304389758566, "loss": 4.4372, "step": 765300 }, { "epoch": 1.428227730878775, "grad_norm": 0.8427965044975281, "learning_rate": 0.00018062795402102704, "loss": 4.6555, "step": 765350 }, { "epoch": 1.4283210364381536, "grad_norm": 1.4934394359588623, "learning_rate": 0.00018062546892390338, "loss": 4.4164, "step": 765400 }, { "epoch": 1.4284143419975321, "grad_norm": 1.3681020736694336, "learning_rate": 0.00018062298368448992, "loss": 4.4399, "step": 765450 }, { "epoch": 1.4285076475569105, "grad_norm": 0.9862996935844421, "learning_rate": 0.00018062049830279108, "loss": 4.3632, "step": 765500 }, { "epoch": 1.428600953116289, "grad_norm": 1.1354358196258545, "learning_rate": 0.00018061801277881128, "loss": 4.6653, "step": 765550 }, { "epoch": 1.4286942586756675, "grad_norm": 1.2349942922592163, "learning_rate": 0.00018061552711255482, "loss": 4.4502, "step": 765600 }, { "epoch": 1.428787564235046, "grad_norm": 0.9900490045547485, "learning_rate": 0.00018061304130402619, "loss": 4.4656, "step": 765650 }, { "epoch": 1.4288808697944244, "grad_norm": 1.205300211906433, "learning_rate": 0.0001806105553532297, "loss": 4.4944, "step": 765700 }, { "epoch": 1.428974175353803, "grad_norm": 1.070824146270752, "learning_rate": 0.00018060806926016975, "loss": 4.3609, "step": 765750 }, { "epoch": 1.4290674809131816, "grad_norm": 0.9711747765541077, "learning_rate": 0.00018060558302485075, "loss": 4.6007, "step": 765800 }, { "epoch": 1.42916078647256, "grad_norm": 1.0082446336746216, "learning_rate": 0.00018060309664727711, "loss": 4.4006, "step": 765850 }, { "epoch": 1.4292540920319385, "grad_norm": 0.978020191192627, "learning_rate": 0.00018060061012745314, "loss": 4.4997, "step": 765900 }, { "epoch": 1.429347397591317, "grad_norm": 0.9886460304260254, "learning_rate": 0.0001805981234653833, "loss": 4.383, "step": 765950 }, { "epoch": 1.4294407031506955, "grad_norm": 2.4021854400634766, "learning_rate": 0.00018059563666107198, "loss": 4.5457, "step": 766000 }, { "epoch": 1.4295340087100739, "grad_norm": 1.1682896614074707, "learning_rate": 0.0001805931497145235, "loss": 4.7379, "step": 766050 }, { "epoch": 1.4296273142694524, "grad_norm": 1.1485880613327026, "learning_rate": 0.0001805906626257423, "loss": 4.3037, "step": 766100 }, { "epoch": 1.429720619828831, "grad_norm": 1.1957522630691528, "learning_rate": 0.00018058817539473275, "loss": 4.3706, "step": 766150 }, { "epoch": 1.4298139253882094, "grad_norm": 0.8990191221237183, "learning_rate": 0.0001805856880214993, "loss": 4.4765, "step": 766200 }, { "epoch": 1.429907230947588, "grad_norm": 1.3056612014770508, "learning_rate": 0.00018058320050604626, "loss": 4.551, "step": 766250 }, { "epoch": 1.4300005365069663, "grad_norm": 1.2101457118988037, "learning_rate": 0.00018058071284837802, "loss": 4.4733, "step": 766300 }, { "epoch": 1.430093842066345, "grad_norm": 1.1525806188583374, "learning_rate": 0.00018057822504849905, "loss": 4.43, "step": 766350 }, { "epoch": 1.4301871476257233, "grad_norm": 1.1641007661819458, "learning_rate": 0.00018057573710641368, "loss": 4.23, "step": 766400 }, { "epoch": 1.4302804531851019, "grad_norm": 1.185188889503479, "learning_rate": 0.00018057324902212628, "loss": 4.3892, "step": 766450 }, { "epoch": 1.4303737587444805, "grad_norm": 1.0136370658874512, "learning_rate": 0.0001805707607956413, "loss": 4.5823, "step": 766500 }, { "epoch": 1.4304670643038588, "grad_norm": 1.1873865127563477, "learning_rate": 0.0001805682724269631, "loss": 4.6697, "step": 766550 }, { "epoch": 1.4305603698632374, "grad_norm": 1.1872572898864746, "learning_rate": 0.00018056578391609606, "loss": 4.5222, "step": 766600 }, { "epoch": 1.4306536754226158, "grad_norm": 1.307830572128296, "learning_rate": 0.0001805632952630446, "loss": 4.6331, "step": 766650 }, { "epoch": 1.4307469809819944, "grad_norm": 1.146446704864502, "learning_rate": 0.00018056080646781312, "loss": 4.4745, "step": 766700 }, { "epoch": 1.4308402865413727, "grad_norm": 1.2394918203353882, "learning_rate": 0.00018055831753040598, "loss": 4.4994, "step": 766750 }, { "epoch": 1.4309335921007513, "grad_norm": 1.3265308141708374, "learning_rate": 0.00018055582845082757, "loss": 4.5611, "step": 766800 }, { "epoch": 1.43102689766013, "grad_norm": 0.923938512802124, "learning_rate": 0.0001805533392290823, "loss": 4.455, "step": 766850 }, { "epoch": 1.4311202032195083, "grad_norm": 1.1751686334609985, "learning_rate": 0.00018055084986517454, "loss": 4.4899, "step": 766900 }, { "epoch": 1.4312135087788869, "grad_norm": 0.7055365443229675, "learning_rate": 0.00018054836035910873, "loss": 4.3759, "step": 766950 }, { "epoch": 1.4313068143382652, "grad_norm": 1.2041410207748413, "learning_rate": 0.00018054587071088924, "loss": 4.4556, "step": 767000 }, { "epoch": 1.4314001198976438, "grad_norm": 1.4476786851882935, "learning_rate": 0.00018054338092052046, "loss": 4.656, "step": 767050 }, { "epoch": 1.4314934254570222, "grad_norm": 1.1249897480010986, "learning_rate": 0.00018054089098800677, "loss": 4.4067, "step": 767100 }, { "epoch": 1.4315867310164008, "grad_norm": 0.9483447670936584, "learning_rate": 0.00018053840091335262, "loss": 4.5332, "step": 767150 }, { "epoch": 1.4316800365757794, "grad_norm": 1.190340280532837, "learning_rate": 0.0001805359106965623, "loss": 4.5865, "step": 767200 }, { "epoch": 1.4317733421351577, "grad_norm": 1.1868788003921509, "learning_rate": 0.00018053342033764032, "loss": 4.688, "step": 767250 }, { "epoch": 1.4318666476945363, "grad_norm": 0.8160537481307983, "learning_rate": 0.00018053092983659098, "loss": 4.5932, "step": 767300 }, { "epoch": 1.4319599532539147, "grad_norm": 1.2115920782089233, "learning_rate": 0.00018052843919341875, "loss": 4.5434, "step": 767350 }, { "epoch": 1.4320532588132933, "grad_norm": 1.0527045726776123, "learning_rate": 0.000180525948408128, "loss": 4.4859, "step": 767400 }, { "epoch": 1.4321465643726716, "grad_norm": 1.1215989589691162, "learning_rate": 0.00018052345748072307, "loss": 4.5047, "step": 767450 }, { "epoch": 1.4322398699320502, "grad_norm": 0.9593571424484253, "learning_rate": 0.00018052096641120845, "loss": 4.4318, "step": 767500 }, { "epoch": 1.4323331754914288, "grad_norm": 1.6852697134017944, "learning_rate": 0.00018051847519958848, "loss": 4.4738, "step": 767550 }, { "epoch": 1.4324264810508072, "grad_norm": 1.297945499420166, "learning_rate": 0.00018051598384586758, "loss": 4.625, "step": 767600 }, { "epoch": 1.4325197866101858, "grad_norm": 1.1906694173812866, "learning_rate": 0.00018051349235005012, "loss": 4.4801, "step": 767650 }, { "epoch": 1.4326130921695641, "grad_norm": 1.102652668952942, "learning_rate": 0.00018051100071214053, "loss": 4.6859, "step": 767700 }, { "epoch": 1.4327063977289427, "grad_norm": 1.251621961593628, "learning_rate": 0.0001805085089321432, "loss": 4.5141, "step": 767750 }, { "epoch": 1.432799703288321, "grad_norm": 1.1050633192062378, "learning_rate": 0.00018050601701006248, "loss": 4.5224, "step": 767800 }, { "epoch": 1.4328930088476997, "grad_norm": 1.4277287721633911, "learning_rate": 0.00018050352494590284, "loss": 4.4855, "step": 767850 }, { "epoch": 1.4329863144070782, "grad_norm": 1.1989407539367676, "learning_rate": 0.00018050103273966863, "loss": 4.3445, "step": 767900 }, { "epoch": 1.4330796199664566, "grad_norm": 1.0351450443267822, "learning_rate": 0.00018049854039136424, "loss": 4.5356, "step": 767950 }, { "epoch": 1.4331729255258352, "grad_norm": 1.2078306674957275, "learning_rate": 0.0001804960479009941, "loss": 4.4853, "step": 768000 }, { "epoch": 1.4331729255258352, "eval_loss": 4.666540145874023, "eval_runtime": 232.7697, "eval_samples_per_second": 11.204, "eval_steps_per_second": 11.204, "eval_tts_loss": 7.623302169241286, "step": 768000 }, { "epoch": 1.4332662310852136, "grad_norm": 1.375489592552185, "learning_rate": 0.00018049355526856265, "loss": 4.602, "step": 768050 }, { "epoch": 1.4333595366445921, "grad_norm": 1.3766145706176758, "learning_rate": 0.0001804910624940742, "loss": 4.3655, "step": 768100 }, { "epoch": 1.4334528422039705, "grad_norm": 1.207764983177185, "learning_rate": 0.00018048856957753317, "loss": 4.6213, "step": 768150 }, { "epoch": 1.433546147763349, "grad_norm": 0.8710137605667114, "learning_rate": 0.000180486076518944, "loss": 4.5959, "step": 768200 }, { "epoch": 1.4336394533227277, "grad_norm": 1.1314088106155396, "learning_rate": 0.00018048358331831105, "loss": 4.4322, "step": 768250 }, { "epoch": 1.433732758882106, "grad_norm": 1.1112110614776611, "learning_rate": 0.00018048108997563875, "loss": 4.6632, "step": 768300 }, { "epoch": 1.4338260644414846, "grad_norm": 1.1782004833221436, "learning_rate": 0.0001804785964909315, "loss": 4.594, "step": 768350 }, { "epoch": 1.433919370000863, "grad_norm": 1.3366174697875977, "learning_rate": 0.00018047610286419364, "loss": 4.3303, "step": 768400 }, { "epoch": 1.4340126755602416, "grad_norm": 1.3267028331756592, "learning_rate": 0.00018047360909542967, "loss": 4.4794, "step": 768450 }, { "epoch": 1.43410598111962, "grad_norm": 1.0863854885101318, "learning_rate": 0.00018047111518464391, "loss": 4.6049, "step": 768500 }, { "epoch": 1.4341992866789985, "grad_norm": 1.4112571477890015, "learning_rate": 0.0001804686211318408, "loss": 4.485, "step": 768550 }, { "epoch": 1.4342925922383771, "grad_norm": 1.2201117277145386, "learning_rate": 0.00018046612693702475, "loss": 4.5829, "step": 768600 }, { "epoch": 1.4343858977977555, "grad_norm": 1.227774977684021, "learning_rate": 0.00018046363260020013, "loss": 4.5836, "step": 768650 }, { "epoch": 1.434479203357134, "grad_norm": 1.064635157585144, "learning_rate": 0.00018046113812137136, "loss": 4.4682, "step": 768700 }, { "epoch": 1.4345725089165124, "grad_norm": 0.8340311646461487, "learning_rate": 0.00018045864350054282, "loss": 4.6426, "step": 768750 }, { "epoch": 1.434665814475891, "grad_norm": 1.3065072298049927, "learning_rate": 0.00018045614873771895, "loss": 4.426, "step": 768800 }, { "epoch": 1.4347591200352694, "grad_norm": 1.2886563539505005, "learning_rate": 0.00018045365383290412, "loss": 4.5232, "step": 768850 }, { "epoch": 1.434852425594648, "grad_norm": 1.286910891532898, "learning_rate": 0.00018045115878610277, "loss": 4.6322, "step": 768900 }, { "epoch": 1.4349457311540266, "grad_norm": 1.2977832555770874, "learning_rate": 0.00018044866359731926, "loss": 4.7656, "step": 768950 }, { "epoch": 1.435039036713405, "grad_norm": 1.2744207382202148, "learning_rate": 0.000180446168266558, "loss": 4.519, "step": 769000 }, { "epoch": 1.4351323422727835, "grad_norm": 1.0086348056793213, "learning_rate": 0.00018044367279382347, "loss": 4.5083, "step": 769050 }, { "epoch": 1.4352256478321619, "grad_norm": 1.147972822189331, "learning_rate": 0.00018044117717911994, "loss": 4.4285, "step": 769100 }, { "epoch": 1.4353189533915405, "grad_norm": 0.9591060280799866, "learning_rate": 0.0001804386814224519, "loss": 4.4266, "step": 769150 }, { "epoch": 1.4354122589509188, "grad_norm": 1.1661403179168701, "learning_rate": 0.00018043618552382378, "loss": 4.5108, "step": 769200 }, { "epoch": 1.4355055645102974, "grad_norm": 1.1529183387756348, "learning_rate": 0.0001804336894832399, "loss": 4.4004, "step": 769250 }, { "epoch": 1.435598870069676, "grad_norm": 1.2013798952102661, "learning_rate": 0.00018043119330070476, "loss": 4.3847, "step": 769300 }, { "epoch": 1.4356921756290544, "grad_norm": 1.0918126106262207, "learning_rate": 0.0001804286969762227, "loss": 4.4228, "step": 769350 }, { "epoch": 1.435785481188433, "grad_norm": 1.2808181047439575, "learning_rate": 0.00018042620050979813, "loss": 4.619, "step": 769400 }, { "epoch": 1.4358787867478113, "grad_norm": 1.2201461791992188, "learning_rate": 0.00018042370390143546, "loss": 4.4418, "step": 769450 }, { "epoch": 1.43597209230719, "grad_norm": 1.0931472778320312, "learning_rate": 0.00018042120715113912, "loss": 4.5599, "step": 769500 }, { "epoch": 1.4360653978665683, "grad_norm": 1.2993226051330566, "learning_rate": 0.00018041871025891348, "loss": 4.5877, "step": 769550 }, { "epoch": 1.4361587034259469, "grad_norm": 1.2949659824371338, "learning_rate": 0.000180416213224763, "loss": 4.4791, "step": 769600 }, { "epoch": 1.4362520089853255, "grad_norm": 1.1886488199234009, "learning_rate": 0.00018041371604869203, "loss": 4.4473, "step": 769650 }, { "epoch": 1.4363453145447038, "grad_norm": 1.3612943887710571, "learning_rate": 0.000180411218730705, "loss": 4.4582, "step": 769700 }, { "epoch": 1.4364386201040824, "grad_norm": 1.0206992626190186, "learning_rate": 0.00018040872127080633, "loss": 4.4818, "step": 769750 }, { "epoch": 1.4365319256634608, "grad_norm": 1.0311999320983887, "learning_rate": 0.00018040622366900037, "loss": 4.5884, "step": 769800 }, { "epoch": 1.4366252312228394, "grad_norm": 1.1408036947250366, "learning_rate": 0.00018040372592529163, "loss": 4.4572, "step": 769850 }, { "epoch": 1.4367185367822177, "grad_norm": 1.3307262659072876, "learning_rate": 0.00018040122803968445, "loss": 4.5753, "step": 769900 }, { "epoch": 1.4368118423415963, "grad_norm": 1.342333197593689, "learning_rate": 0.00018039873001218323, "loss": 4.7378, "step": 769950 }, { "epoch": 1.436905147900975, "grad_norm": 1.4590122699737549, "learning_rate": 0.00018039623184279243, "loss": 4.3888, "step": 770000 }, { "epoch": 1.4369984534603533, "grad_norm": 1.4871413707733154, "learning_rate": 0.0001803937335315164, "loss": 4.5632, "step": 770050 }, { "epoch": 1.4370917590197319, "grad_norm": 1.1752780675888062, "learning_rate": 0.00018039123507835957, "loss": 4.4527, "step": 770100 }, { "epoch": 1.4371850645791102, "grad_norm": 1.0668138265609741, "learning_rate": 0.00018038873648332636, "loss": 4.6954, "step": 770150 }, { "epoch": 1.4372783701384888, "grad_norm": 1.3020116090774536, "learning_rate": 0.0001803862377464212, "loss": 4.4859, "step": 770200 }, { "epoch": 1.4373716756978672, "grad_norm": 0.9535107612609863, "learning_rate": 0.00018038373886764846, "loss": 4.4806, "step": 770250 }, { "epoch": 1.4374649812572458, "grad_norm": 1.5332229137420654, "learning_rate": 0.00018038123984701255, "loss": 4.6066, "step": 770300 }, { "epoch": 1.4375582868166243, "grad_norm": 1.3931421041488647, "learning_rate": 0.00018037874068451792, "loss": 4.5535, "step": 770350 }, { "epoch": 1.4376515923760027, "grad_norm": 1.5173532962799072, "learning_rate": 0.00018037624138016894, "loss": 4.3555, "step": 770400 }, { "epoch": 1.4377448979353813, "grad_norm": 1.3752037286758423, "learning_rate": 0.00018037374193397002, "loss": 4.3681, "step": 770450 }, { "epoch": 1.4378382034947597, "grad_norm": 1.3721927404403687, "learning_rate": 0.0001803712423459256, "loss": 4.5098, "step": 770500 }, { "epoch": 1.4379315090541382, "grad_norm": 1.2211096286773682, "learning_rate": 0.0001803687426160401, "loss": 4.6152, "step": 770550 }, { "epoch": 1.4380248146135166, "grad_norm": 1.0415723323822021, "learning_rate": 0.0001803662427443179, "loss": 4.5386, "step": 770600 }, { "epoch": 1.4381181201728952, "grad_norm": 1.4926120042800903, "learning_rate": 0.00018036374273076343, "loss": 4.609, "step": 770650 }, { "epoch": 1.4382114257322738, "grad_norm": 0.9463675022125244, "learning_rate": 0.00018036124257538106, "loss": 4.5833, "step": 770700 }, { "epoch": 1.4383047312916521, "grad_norm": 0.9115874767303467, "learning_rate": 0.00018035874227817527, "loss": 4.3483, "step": 770750 }, { "epoch": 1.4383980368510307, "grad_norm": 0.9490000605583191, "learning_rate": 0.00018035624183915044, "loss": 4.3509, "step": 770800 }, { "epoch": 1.438491342410409, "grad_norm": 1.2022995948791504, "learning_rate": 0.00018035374125831096, "loss": 4.575, "step": 770850 }, { "epoch": 1.4385846479697877, "grad_norm": 1.1890491247177124, "learning_rate": 0.00018035124053566128, "loss": 4.5672, "step": 770900 }, { "epoch": 1.438677953529166, "grad_norm": 1.3777884244918823, "learning_rate": 0.0001803487396712058, "loss": 4.4858, "step": 770950 }, { "epoch": 1.4387712590885446, "grad_norm": 1.0728179216384888, "learning_rate": 0.0001803462386649489, "loss": 4.4586, "step": 771000 }, { "epoch": 1.4387712590885446, "eval_loss": 4.667508125305176, "eval_runtime": 232.9969, "eval_samples_per_second": 11.193, "eval_steps_per_second": 11.193, "eval_tts_loss": 7.672249912400016, "step": 771000 }, { "epoch": 1.4388645646479232, "grad_norm": 1.0140492916107178, "learning_rate": 0.0001803437375168951, "loss": 4.4016, "step": 771050 }, { "epoch": 1.4389578702073016, "grad_norm": 1.2091118097305298, "learning_rate": 0.00018034123622704868, "loss": 4.5345, "step": 771100 }, { "epoch": 1.4390511757666802, "grad_norm": 1.4424926042556763, "learning_rate": 0.00018033873479541412, "loss": 4.4916, "step": 771150 }, { "epoch": 1.4391444813260585, "grad_norm": 1.3101881742477417, "learning_rate": 0.00018033623322199582, "loss": 4.409, "step": 771200 }, { "epoch": 1.4392377868854371, "grad_norm": 1.274510145187378, "learning_rate": 0.00018033373150679825, "loss": 4.5061, "step": 771250 }, { "epoch": 1.4393310924448155, "grad_norm": 1.1696717739105225, "learning_rate": 0.00018033122964982573, "loss": 4.3961, "step": 771300 }, { "epoch": 1.439424398004194, "grad_norm": 1.2649139165878296, "learning_rate": 0.00018032872765108278, "loss": 4.6118, "step": 771350 }, { "epoch": 1.4395177035635727, "grad_norm": 1.2144819498062134, "learning_rate": 0.00018032622551057372, "loss": 4.6652, "step": 771400 }, { "epoch": 1.439611009122951, "grad_norm": 1.3111515045166016, "learning_rate": 0.000180323723228303, "loss": 4.5438, "step": 771450 }, { "epoch": 1.4397043146823296, "grad_norm": 1.3187100887298584, "learning_rate": 0.00018032122080427506, "loss": 4.5365, "step": 771500 }, { "epoch": 1.439797620241708, "grad_norm": 1.166625738143921, "learning_rate": 0.00018031871823849426, "loss": 4.4482, "step": 771550 }, { "epoch": 1.4398909258010866, "grad_norm": 0.8904200792312622, "learning_rate": 0.00018031621553096512, "loss": 4.4477, "step": 771600 }, { "epoch": 1.439984231360465, "grad_norm": 1.0547096729278564, "learning_rate": 0.00018031371268169196, "loss": 4.5242, "step": 771650 }, { "epoch": 1.4400775369198435, "grad_norm": 1.1480703353881836, "learning_rate": 0.0001803112096906792, "loss": 4.4393, "step": 771700 }, { "epoch": 1.4401708424792221, "grad_norm": 1.396099328994751, "learning_rate": 0.00018030870655793134, "loss": 4.5087, "step": 771750 }, { "epoch": 1.4402641480386005, "grad_norm": 1.4342947006225586, "learning_rate": 0.00018030620328345268, "loss": 4.5932, "step": 771800 }, { "epoch": 1.440357453597979, "grad_norm": 1.3291150331497192, "learning_rate": 0.00018030369986724777, "loss": 4.4319, "step": 771850 }, { "epoch": 1.4404507591573574, "grad_norm": 1.1561295986175537, "learning_rate": 0.0001803011963093209, "loss": 4.4819, "step": 771900 }, { "epoch": 1.440544064716736, "grad_norm": 1.455711007118225, "learning_rate": 0.00018029869260967658, "loss": 4.4959, "step": 771950 }, { "epoch": 1.4406373702761144, "grad_norm": 1.4160335063934326, "learning_rate": 0.0001802961887683192, "loss": 4.5489, "step": 772000 }, { "epoch": 1.440730675835493, "grad_norm": 1.472607970237732, "learning_rate": 0.00018029368478525318, "loss": 4.3786, "step": 772050 }, { "epoch": 1.4408239813948716, "grad_norm": 1.3212214708328247, "learning_rate": 0.00018029118066048292, "loss": 4.7, "step": 772100 }, { "epoch": 1.44091728695425, "grad_norm": 0.7592604756355286, "learning_rate": 0.00018028867639401287, "loss": 4.37, "step": 772150 }, { "epoch": 1.4410105925136285, "grad_norm": 0.9890813827514648, "learning_rate": 0.00018028617198584737, "loss": 4.6758, "step": 772200 }, { "epoch": 1.4411038980730069, "grad_norm": 1.1591099500656128, "learning_rate": 0.000180283667435991, "loss": 4.5009, "step": 772250 }, { "epoch": 1.4411972036323855, "grad_norm": 1.0878084897994995, "learning_rate": 0.000180281162744448, "loss": 4.5268, "step": 772300 }, { "epoch": 1.4412905091917638, "grad_norm": 0.9479683637619019, "learning_rate": 0.00018027865791122292, "loss": 4.631, "step": 772350 }, { "epoch": 1.4413838147511424, "grad_norm": 1.2931561470031738, "learning_rate": 0.00018027615293632012, "loss": 4.4399, "step": 772400 }, { "epoch": 1.441477120310521, "grad_norm": 1.1678410768508911, "learning_rate": 0.00018027364781974406, "loss": 4.4775, "step": 772450 }, { "epoch": 1.4415704258698994, "grad_norm": 1.3034080266952515, "learning_rate": 0.00018027114256149911, "loss": 4.6236, "step": 772500 }, { "epoch": 1.441663731429278, "grad_norm": 1.0815435647964478, "learning_rate": 0.00018026863716158971, "loss": 4.5758, "step": 772550 }, { "epoch": 1.4417570369886563, "grad_norm": 1.4056761264801025, "learning_rate": 0.0001802661316200203, "loss": 4.4122, "step": 772600 }, { "epoch": 1.441850342548035, "grad_norm": 1.5499346256256104, "learning_rate": 0.00018026362593679532, "loss": 4.4597, "step": 772650 }, { "epoch": 1.4419436481074133, "grad_norm": 1.3256648778915405, "learning_rate": 0.00018026112011191914, "loss": 4.5118, "step": 772700 }, { "epoch": 1.4420369536667919, "grad_norm": 1.3465520143508911, "learning_rate": 0.00018025861414539617, "loss": 4.6021, "step": 772750 }, { "epoch": 1.4421302592261704, "grad_norm": 1.2570968866348267, "learning_rate": 0.0001802561080372309, "loss": 4.403, "step": 772800 }, { "epoch": 1.4422235647855488, "grad_norm": 1.5623722076416016, "learning_rate": 0.00018025360178742773, "loss": 4.3795, "step": 772850 }, { "epoch": 1.4423168703449274, "grad_norm": 1.0094107389450073, "learning_rate": 0.00018025109539599105, "loss": 4.3445, "step": 772900 }, { "epoch": 1.4424101759043058, "grad_norm": 1.232707142829895, "learning_rate": 0.00018024858886292534, "loss": 4.4519, "step": 772950 }, { "epoch": 1.4425034814636843, "grad_norm": 1.092909812927246, "learning_rate": 0.00018024608218823499, "loss": 4.549, "step": 773000 }, { "epoch": 1.4425967870230627, "grad_norm": 1.072887897491455, "learning_rate": 0.0001802435753719244, "loss": 4.4872, "step": 773050 }, { "epoch": 1.4426900925824413, "grad_norm": 1.34170663356781, "learning_rate": 0.000180241068413998, "loss": 4.6032, "step": 773100 }, { "epoch": 1.4427833981418199, "grad_norm": 1.1184030771255493, "learning_rate": 0.0001802385613144603, "loss": 4.5621, "step": 773150 }, { "epoch": 1.4428767037011982, "grad_norm": 1.3179848194122314, "learning_rate": 0.00018023605407331558, "loss": 4.5076, "step": 773200 }, { "epoch": 1.4429700092605768, "grad_norm": 1.343911051750183, "learning_rate": 0.00018023354669056838, "loss": 4.441, "step": 773250 }, { "epoch": 1.4430633148199552, "grad_norm": 1.1098930835723877, "learning_rate": 0.0001802310391662231, "loss": 4.5466, "step": 773300 }, { "epoch": 1.4431566203793338, "grad_norm": 1.3704051971435547, "learning_rate": 0.0001802285315002841, "loss": 4.3494, "step": 773350 }, { "epoch": 1.4432499259387122, "grad_norm": 1.4381294250488281, "learning_rate": 0.00018022602369275595, "loss": 4.791, "step": 773400 }, { "epoch": 1.4433432314980907, "grad_norm": 1.022616982460022, "learning_rate": 0.00018022351574364291, "loss": 4.4347, "step": 773450 }, { "epoch": 1.4434365370574693, "grad_norm": 0.8164454698562622, "learning_rate": 0.0001802210076529495, "loss": 4.6019, "step": 773500 }, { "epoch": 1.4435298426168477, "grad_norm": 1.213606834411621, "learning_rate": 0.00018021849942068013, "loss": 4.4205, "step": 773550 }, { "epoch": 1.4436231481762263, "grad_norm": 1.1292697191238403, "learning_rate": 0.00018021599104683922, "loss": 4.4517, "step": 773600 }, { "epoch": 1.4437164537356046, "grad_norm": 1.1262978315353394, "learning_rate": 0.0001802134825314312, "loss": 4.636, "step": 773650 }, { "epoch": 1.4438097592949832, "grad_norm": 0.8524104952812195, "learning_rate": 0.0001802109738744605, "loss": 4.4168, "step": 773700 }, { "epoch": 1.4439030648543616, "grad_norm": 1.4122551679611206, "learning_rate": 0.00018020846507593158, "loss": 4.4569, "step": 773750 }, { "epoch": 1.4439963704137402, "grad_norm": 1.1943992376327515, "learning_rate": 0.00018020595613584876, "loss": 4.6663, "step": 773800 }, { "epoch": 1.4440896759731188, "grad_norm": 1.3334665298461914, "learning_rate": 0.00018020344705421657, "loss": 4.3489, "step": 773850 }, { "epoch": 1.4441829815324971, "grad_norm": 1.3007961511611938, "learning_rate": 0.00018020093783103943, "loss": 4.379, "step": 773900 }, { "epoch": 1.4442762870918757, "grad_norm": 1.4516056776046753, "learning_rate": 0.00018019842846632173, "loss": 4.4332, "step": 773950 }, { "epoch": 1.444369592651254, "grad_norm": 1.1682356595993042, "learning_rate": 0.00018019591896006795, "loss": 4.5721, "step": 774000 }, { "epoch": 1.444369592651254, "eval_loss": 4.660179615020752, "eval_runtime": 229.8588, "eval_samples_per_second": 11.346, "eval_steps_per_second": 11.346, "eval_tts_loss": 7.65920675819614, "step": 774000 }, { "epoch": 1.4444628982106327, "grad_norm": 1.4106265306472778, "learning_rate": 0.00018019340931228243, "loss": 4.6384, "step": 774050 }, { "epoch": 1.444556203770011, "grad_norm": 1.27629816532135, "learning_rate": 0.00018019089952296967, "loss": 4.4958, "step": 774100 }, { "epoch": 1.4446495093293896, "grad_norm": 1.0064910650253296, "learning_rate": 0.0001801883895921341, "loss": 4.431, "step": 774150 }, { "epoch": 1.4447428148887682, "grad_norm": 1.3378223180770874, "learning_rate": 0.0001801858795197801, "loss": 4.4885, "step": 774200 }, { "epoch": 1.4448361204481466, "grad_norm": 0.9865140914916992, "learning_rate": 0.00018018336930591215, "loss": 4.4901, "step": 774250 }, { "epoch": 1.4449294260075252, "grad_norm": 1.2893882989883423, "learning_rate": 0.00018018085895053467, "loss": 4.3752, "step": 774300 }, { "epoch": 1.4450227315669035, "grad_norm": 0.925656795501709, "learning_rate": 0.00018017834845365208, "loss": 4.5121, "step": 774350 }, { "epoch": 1.4451160371262821, "grad_norm": 1.3008805513381958, "learning_rate": 0.0001801758378152688, "loss": 4.6022, "step": 774400 }, { "epoch": 1.4452093426856605, "grad_norm": 1.5284887552261353, "learning_rate": 0.00018017332703538927, "loss": 4.5198, "step": 774450 }, { "epoch": 1.445302648245039, "grad_norm": 1.242952585220337, "learning_rate": 0.00018017081611401795, "loss": 4.5759, "step": 774500 }, { "epoch": 1.4453959538044177, "grad_norm": 1.185299038887024, "learning_rate": 0.00018016830505115922, "loss": 4.5108, "step": 774550 }, { "epoch": 1.445489259363796, "grad_norm": 0.8266953229904175, "learning_rate": 0.00018016579384681755, "loss": 4.3293, "step": 774600 }, { "epoch": 1.4455825649231746, "grad_norm": 1.3259170055389404, "learning_rate": 0.00018016328250099734, "loss": 4.7487, "step": 774650 }, { "epoch": 1.445675870482553, "grad_norm": 1.6738992929458618, "learning_rate": 0.0001801607710137031, "loss": 4.5182, "step": 774700 }, { "epoch": 1.4457691760419316, "grad_norm": 1.1912956237792969, "learning_rate": 0.00018015825938493912, "loss": 4.407, "step": 774750 }, { "epoch": 1.44586248160131, "grad_norm": 1.1691521406173706, "learning_rate": 0.00018015574761470998, "loss": 4.2703, "step": 774800 }, { "epoch": 1.4459557871606885, "grad_norm": 1.2927955389022827, "learning_rate": 0.00018015323570302004, "loss": 4.6371, "step": 774850 }, { "epoch": 1.446049092720067, "grad_norm": 1.2651420831680298, "learning_rate": 0.00018015072364987374, "loss": 4.5727, "step": 774900 }, { "epoch": 1.4461423982794455, "grad_norm": 1.3653576374053955, "learning_rate": 0.0001801482114552755, "loss": 4.4497, "step": 774950 }, { "epoch": 1.446235703838824, "grad_norm": 1.3242924213409424, "learning_rate": 0.00018014569911922976, "loss": 4.6692, "step": 775000 }, { "epoch": 1.4463290093982024, "grad_norm": 1.3763872385025024, "learning_rate": 0.000180143186641741, "loss": 4.5189, "step": 775050 }, { "epoch": 1.446422314957581, "grad_norm": 1.2092957496643066, "learning_rate": 0.00018014067402281358, "loss": 4.7324, "step": 775100 }, { "epoch": 1.4465156205169594, "grad_norm": 1.3293168544769287, "learning_rate": 0.000180138161262452, "loss": 4.4721, "step": 775150 }, { "epoch": 1.446608926076338, "grad_norm": 1.1585038900375366, "learning_rate": 0.00018013564836066065, "loss": 4.5503, "step": 775200 }, { "epoch": 1.4467022316357165, "grad_norm": 1.2194722890853882, "learning_rate": 0.00018013313531744398, "loss": 4.5447, "step": 775250 }, { "epoch": 1.446795537195095, "grad_norm": 0.9149037003517151, "learning_rate": 0.00018013062213280642, "loss": 4.5923, "step": 775300 }, { "epoch": 1.4468888427544735, "grad_norm": 1.1425459384918213, "learning_rate": 0.00018012810880675245, "loss": 4.6864, "step": 775350 }, { "epoch": 1.4469821483138519, "grad_norm": 1.0171364545822144, "learning_rate": 0.00018012559533928644, "loss": 4.4864, "step": 775400 }, { "epoch": 1.4470754538732304, "grad_norm": 0.8627289533615112, "learning_rate": 0.00018012308173041285, "loss": 4.6055, "step": 775450 }, { "epoch": 1.4471687594326088, "grad_norm": 1.251126766204834, "learning_rate": 0.00018012056798013613, "loss": 4.5918, "step": 775500 }, { "epoch": 1.4472620649919874, "grad_norm": 0.9426770806312561, "learning_rate": 0.0001801180540884607, "loss": 4.3636, "step": 775550 }, { "epoch": 1.447355370551366, "grad_norm": 1.334795355796814, "learning_rate": 0.00018011554005539098, "loss": 4.6064, "step": 775600 }, { "epoch": 1.4474486761107443, "grad_norm": 1.294052004814148, "learning_rate": 0.00018011302588093146, "loss": 4.5361, "step": 775650 }, { "epoch": 1.447541981670123, "grad_norm": 1.2017312049865723, "learning_rate": 0.00018011051156508652, "loss": 4.4366, "step": 775700 }, { "epoch": 1.4476352872295013, "grad_norm": 1.2202517986297607, "learning_rate": 0.00018010799710786065, "loss": 4.4863, "step": 775750 }, { "epoch": 1.4477285927888799, "grad_norm": 1.129022240638733, "learning_rate": 0.00018010548250925825, "loss": 4.5169, "step": 775800 }, { "epoch": 1.4478218983482583, "grad_norm": 1.3236628770828247, "learning_rate": 0.00018010296776928378, "loss": 4.6282, "step": 775850 }, { "epoch": 1.4479152039076368, "grad_norm": 1.3330432176589966, "learning_rate": 0.00018010045288794163, "loss": 4.3919, "step": 775900 }, { "epoch": 1.4480085094670154, "grad_norm": 1.2738417387008667, "learning_rate": 0.0001800979378652363, "loss": 4.5189, "step": 775950 }, { "epoch": 1.4481018150263938, "grad_norm": 1.1751788854599, "learning_rate": 0.00018009542270117218, "loss": 4.5086, "step": 776000 }, { "epoch": 1.4481951205857724, "grad_norm": 1.1980212926864624, "learning_rate": 0.00018009290739575375, "loss": 4.6227, "step": 776050 }, { "epoch": 1.4482884261451507, "grad_norm": 1.1280452013015747, "learning_rate": 0.00018009039194898544, "loss": 4.471, "step": 776100 }, { "epoch": 1.4483817317045293, "grad_norm": 0.8283519744873047, "learning_rate": 0.00018008787636087167, "loss": 4.4137, "step": 776150 }, { "epoch": 1.4484750372639077, "grad_norm": 1.245788335800171, "learning_rate": 0.0001800853606314169, "loss": 4.5648, "step": 776200 }, { "epoch": 1.4485683428232863, "grad_norm": 1.241831660270691, "learning_rate": 0.00018008284476062557, "loss": 4.5316, "step": 776250 }, { "epoch": 1.4486616483826649, "grad_norm": 1.2635796070098877, "learning_rate": 0.0001800803287485021, "loss": 4.5293, "step": 776300 }, { "epoch": 1.4487549539420432, "grad_norm": 1.1093114614486694, "learning_rate": 0.00018007781259505093, "loss": 4.5829, "step": 776350 }, { "epoch": 1.4488482595014218, "grad_norm": 1.0754183530807495, "learning_rate": 0.0001800752963002765, "loss": 4.3999, "step": 776400 }, { "epoch": 1.4489415650608002, "grad_norm": 1.2647534608840942, "learning_rate": 0.0001800727798641833, "loss": 4.4648, "step": 776450 }, { "epoch": 1.4490348706201788, "grad_norm": 1.039973258972168, "learning_rate": 0.00018007026328677568, "loss": 4.4302, "step": 776500 }, { "epoch": 1.4491281761795571, "grad_norm": 1.205474853515625, "learning_rate": 0.00018006774656805817, "loss": 4.6899, "step": 776550 }, { "epoch": 1.4492214817389357, "grad_norm": 0.994087278842926, "learning_rate": 0.00018006522970803518, "loss": 4.4857, "step": 776600 }, { "epoch": 1.4493147872983143, "grad_norm": 0.917029857635498, "learning_rate": 0.00018006271270671112, "loss": 4.4131, "step": 776650 }, { "epoch": 1.4494080928576927, "grad_norm": 1.2148895263671875, "learning_rate": 0.00018006019556409048, "loss": 4.4065, "step": 776700 }, { "epoch": 1.4495013984170713, "grad_norm": 1.0337047576904297, "learning_rate": 0.00018005767828017767, "loss": 4.2981, "step": 776750 }, { "epoch": 1.4495947039764496, "grad_norm": 1.1114561557769775, "learning_rate": 0.00018005516085497716, "loss": 4.559, "step": 776800 }, { "epoch": 1.4496880095358282, "grad_norm": 0.9262935519218445, "learning_rate": 0.00018005264328849337, "loss": 4.3914, "step": 776850 }, { "epoch": 1.4497813150952066, "grad_norm": 0.9530255794525146, "learning_rate": 0.00018005012558073075, "loss": 4.4573, "step": 776900 }, { "epoch": 1.4498746206545852, "grad_norm": 1.5111445188522339, "learning_rate": 0.00018004760773169376, "loss": 4.525, "step": 776950 }, { "epoch": 1.4499679262139638, "grad_norm": 1.187482476234436, "learning_rate": 0.0001800450897413868, "loss": 4.4479, "step": 777000 }, { "epoch": 1.4499679262139638, "eval_loss": 4.65779972076416, "eval_runtime": 229.1541, "eval_samples_per_second": 11.381, "eval_steps_per_second": 11.381, "eval_tts_loss": 7.6011238636826075, "step": 777000 }, { "epoch": 1.4500612317733421, "grad_norm": 1.5328707695007324, "learning_rate": 0.00018004257160981438, "loss": 4.5038, "step": 777050 }, { "epoch": 1.4501545373327207, "grad_norm": 1.037157416343689, "learning_rate": 0.00018004005333698087, "loss": 4.5015, "step": 777100 }, { "epoch": 1.450247842892099, "grad_norm": 1.6141831874847412, "learning_rate": 0.00018003753492289076, "loss": 4.4084, "step": 777150 }, { "epoch": 1.4503411484514777, "grad_norm": 1.7018786668777466, "learning_rate": 0.00018003501636754848, "loss": 4.5983, "step": 777200 }, { "epoch": 1.450434454010856, "grad_norm": 1.6245293617248535, "learning_rate": 0.00018003249767095845, "loss": 4.6082, "step": 777250 }, { "epoch": 1.4505277595702346, "grad_norm": 1.0842900276184082, "learning_rate": 0.00018002997883312518, "loss": 4.7259, "step": 777300 }, { "epoch": 1.4506210651296132, "grad_norm": 1.2968006134033203, "learning_rate": 0.00018002745985405306, "loss": 4.4638, "step": 777350 }, { "epoch": 1.4507143706889916, "grad_norm": 0.9917076230049133, "learning_rate": 0.00018002494073374655, "loss": 4.6488, "step": 777400 }, { "epoch": 1.4508076762483701, "grad_norm": 1.1953271627426147, "learning_rate": 0.00018002242147221017, "loss": 4.2566, "step": 777450 }, { "epoch": 1.4509009818077485, "grad_norm": 1.1460000276565552, "learning_rate": 0.0001800199020694482, "loss": 4.5319, "step": 777500 }, { "epoch": 1.450994287367127, "grad_norm": 1.1301995515823364, "learning_rate": 0.00018001738252546522, "loss": 4.3195, "step": 777550 }, { "epoch": 1.4510875929265055, "grad_norm": 1.0220811367034912, "learning_rate": 0.00018001486284026567, "loss": 4.5177, "step": 777600 }, { "epoch": 1.451180898485884, "grad_norm": 1.4093170166015625, "learning_rate": 0.00018001234301385393, "loss": 4.4414, "step": 777650 }, { "epoch": 1.4512742040452626, "grad_norm": 1.1304699182510376, "learning_rate": 0.0001800098230462345, "loss": 4.4836, "step": 777700 }, { "epoch": 1.451367509604641, "grad_norm": 1.5177037715911865, "learning_rate": 0.00018000730293741182, "loss": 4.5023, "step": 777750 }, { "epoch": 1.4514608151640196, "grad_norm": 1.1220999956130981, "learning_rate": 0.0001800047826873903, "loss": 4.4653, "step": 777800 }, { "epoch": 1.451554120723398, "grad_norm": 1.1306718587875366, "learning_rate": 0.00018000226229617443, "loss": 4.3466, "step": 777850 }, { "epoch": 1.4516474262827765, "grad_norm": 1.0430713891983032, "learning_rate": 0.00017999974176376865, "loss": 4.3569, "step": 777900 }, { "epoch": 1.451740731842155, "grad_norm": 1.0632010698318481, "learning_rate": 0.00017999722109017738, "loss": 4.5061, "step": 777950 }, { "epoch": 1.4518340374015335, "grad_norm": 1.1130421161651611, "learning_rate": 0.00017999470027540513, "loss": 4.5507, "step": 778000 }, { "epoch": 1.451927342960912, "grad_norm": 1.2284832000732422, "learning_rate": 0.00017999217931945628, "loss": 4.4958, "step": 778050 }, { "epoch": 1.4520206485202904, "grad_norm": 1.815981149673462, "learning_rate": 0.0001799896582223353, "loss": 4.6424, "step": 778100 }, { "epoch": 1.452113954079669, "grad_norm": 0.9870684146881104, "learning_rate": 0.00017998713698404666, "loss": 4.4758, "step": 778150 }, { "epoch": 1.4522072596390474, "grad_norm": 0.9938642978668213, "learning_rate": 0.00017998461560459478, "loss": 4.4944, "step": 778200 }, { "epoch": 1.452300565198426, "grad_norm": 0.8919299244880676, "learning_rate": 0.00017998209408398417, "loss": 4.5574, "step": 778250 }, { "epoch": 1.4523938707578044, "grad_norm": 1.367987036705017, "learning_rate": 0.00017997957242221922, "loss": 4.3835, "step": 778300 }, { "epoch": 1.452487176317183, "grad_norm": 1.3675861358642578, "learning_rate": 0.00017997705061930438, "loss": 4.537, "step": 778350 }, { "epoch": 1.4525804818765615, "grad_norm": 1.2209423780441284, "learning_rate": 0.0001799745286752441, "loss": 4.5109, "step": 778400 }, { "epoch": 1.45267378743594, "grad_norm": 1.2261372804641724, "learning_rate": 0.0001799720065900429, "loss": 4.6316, "step": 778450 }, { "epoch": 1.4527670929953185, "grad_norm": 1.259387493133545, "learning_rate": 0.00017996948436370516, "loss": 4.5134, "step": 778500 }, { "epoch": 1.4528603985546968, "grad_norm": 0.9867664575576782, "learning_rate": 0.00017996696199623535, "loss": 4.5865, "step": 778550 }, { "epoch": 1.4529537041140754, "grad_norm": 1.0446977615356445, "learning_rate": 0.00017996443948763792, "loss": 4.6718, "step": 778600 }, { "epoch": 1.4530470096734538, "grad_norm": 0.9644551873207092, "learning_rate": 0.00017996191683791732, "loss": 4.5707, "step": 778650 }, { "epoch": 1.4531403152328324, "grad_norm": 1.3627334833145142, "learning_rate": 0.00017995939404707802, "loss": 4.5694, "step": 778700 }, { "epoch": 1.453233620792211, "grad_norm": 1.1479521989822388, "learning_rate": 0.00017995687111512445, "loss": 4.6246, "step": 778750 }, { "epoch": 1.4533269263515893, "grad_norm": 1.3104310035705566, "learning_rate": 0.0001799543480420611, "loss": 4.6224, "step": 778800 }, { "epoch": 1.4534202319109677, "grad_norm": 1.2776497602462769, "learning_rate": 0.00017995182482789235, "loss": 4.8036, "step": 778850 }, { "epoch": 1.4535135374703463, "grad_norm": 1.1575006246566772, "learning_rate": 0.0001799493014726227, "loss": 4.6373, "step": 778900 }, { "epoch": 1.4536068430297249, "grad_norm": 1.566729187965393, "learning_rate": 0.0001799467779762566, "loss": 4.6231, "step": 778950 }, { "epoch": 1.4537001485891032, "grad_norm": 1.1157954931259155, "learning_rate": 0.00017994425433879853, "loss": 4.6433, "step": 779000 }, { "epoch": 1.4537934541484818, "grad_norm": 1.2924871444702148, "learning_rate": 0.00017994173056025292, "loss": 4.3561, "step": 779050 }, { "epoch": 1.4538867597078604, "grad_norm": 0.717244565486908, "learning_rate": 0.0001799392066406242, "loss": 4.3715, "step": 779100 }, { "epoch": 1.4539800652672388, "grad_norm": 1.0869789123535156, "learning_rate": 0.00017993668257991688, "loss": 4.4943, "step": 779150 }, { "epoch": 1.4540733708266171, "grad_norm": 0.9677169322967529, "learning_rate": 0.00017993415837813532, "loss": 4.4834, "step": 779200 }, { "epoch": 1.4541666763859957, "grad_norm": 1.0717506408691406, "learning_rate": 0.00017993163403528408, "loss": 4.6366, "step": 779250 }, { "epoch": 1.4542599819453743, "grad_norm": 1.3315452337265015, "learning_rate": 0.00017992910955136758, "loss": 4.5631, "step": 779300 }, { "epoch": 1.4543532875047527, "grad_norm": 1.0410174131393433, "learning_rate": 0.00017992658492639024, "loss": 4.4536, "step": 779350 }, { "epoch": 1.4544465930641313, "grad_norm": 1.230658769607544, "learning_rate": 0.00017992406016035655, "loss": 4.5992, "step": 779400 }, { "epoch": 1.4545398986235099, "grad_norm": 0.9602358341217041, "learning_rate": 0.00017992153525327095, "loss": 4.5642, "step": 779450 }, { "epoch": 1.4546332041828882, "grad_norm": 1.2320843935012817, "learning_rate": 0.00017991901020513788, "loss": 4.3894, "step": 779500 }, { "epoch": 1.4547265097422666, "grad_norm": 1.4385770559310913, "learning_rate": 0.00017991648501596184, "loss": 4.6261, "step": 779550 }, { "epoch": 1.4548198153016452, "grad_norm": 1.3215069770812988, "learning_rate": 0.0001799139596857473, "loss": 4.4814, "step": 779600 }, { "epoch": 1.4549131208610238, "grad_norm": 1.1566412448883057, "learning_rate": 0.00017991143421449866, "loss": 4.5152, "step": 779650 }, { "epoch": 1.4550064264204021, "grad_norm": 1.4311187267303467, "learning_rate": 0.00017990890860222037, "loss": 4.4881, "step": 779700 }, { "epoch": 1.4550997319797807, "grad_norm": 1.353785753250122, "learning_rate": 0.00017990638284891695, "loss": 4.2909, "step": 779750 }, { "epoch": 1.4551930375391593, "grad_norm": 1.251240849494934, "learning_rate": 0.00017990385695459282, "loss": 4.3999, "step": 779800 }, { "epoch": 1.4552863430985377, "grad_norm": 1.0110318660736084, "learning_rate": 0.00017990133091925242, "loss": 4.5232, "step": 779850 }, { "epoch": 1.455379648657916, "grad_norm": 1.1490775346755981, "learning_rate": 0.00017989880474290025, "loss": 4.4757, "step": 779900 }, { "epoch": 1.4554729542172946, "grad_norm": 1.1822558641433716, "learning_rate": 0.00017989627842554076, "loss": 4.623, "step": 779950 }, { "epoch": 1.4555662597766732, "grad_norm": 1.4138990640640259, "learning_rate": 0.00017989375196717837, "loss": 4.4426, "step": 780000 }, { "epoch": 1.4555662597766732, "eval_loss": 4.660432815551758, "eval_runtime": 228.4677, "eval_samples_per_second": 11.415, "eval_steps_per_second": 11.415, "eval_tts_loss": 7.575179839453894, "step": 780000 }, { "epoch": 1.4556595653360516, "grad_norm": 1.2193901538848877, "learning_rate": 0.00017989122536781756, "loss": 4.5033, "step": 780050 }, { "epoch": 1.4557528708954302, "grad_norm": 1.24460768699646, "learning_rate": 0.00017988869862746283, "loss": 4.513, "step": 780100 }, { "epoch": 1.4558461764548087, "grad_norm": 1.1478904485702515, "learning_rate": 0.00017988617174611858, "loss": 4.444, "step": 780150 }, { "epoch": 1.455939482014187, "grad_norm": 0.9588996767997742, "learning_rate": 0.00017988364472378932, "loss": 4.4341, "step": 780200 }, { "epoch": 1.4560327875735655, "grad_norm": 0.9328056573867798, "learning_rate": 0.00017988111756047945, "loss": 4.5043, "step": 780250 }, { "epoch": 1.456126093132944, "grad_norm": 1.676249384880066, "learning_rate": 0.00017987859025619347, "loss": 4.5919, "step": 780300 }, { "epoch": 1.4562193986923226, "grad_norm": 1.2845678329467773, "learning_rate": 0.00017987606281093585, "loss": 4.5309, "step": 780350 }, { "epoch": 1.456312704251701, "grad_norm": 1.11307954788208, "learning_rate": 0.000179873535224711, "loss": 4.4624, "step": 780400 }, { "epoch": 1.4564060098110796, "grad_norm": 1.405887484550476, "learning_rate": 0.00017987100749752345, "loss": 4.7208, "step": 780450 }, { "epoch": 1.4564993153704582, "grad_norm": 1.025039792060852, "learning_rate": 0.0001798684796293776, "loss": 4.6323, "step": 780500 }, { "epoch": 1.4565926209298365, "grad_norm": 1.1266624927520752, "learning_rate": 0.00017986595162027794, "loss": 4.609, "step": 780550 }, { "epoch": 1.456685926489215, "grad_norm": 1.0085512399673462, "learning_rate": 0.00017986342347022896, "loss": 4.4753, "step": 780600 }, { "epoch": 1.4567792320485935, "grad_norm": 1.3571631908416748, "learning_rate": 0.00017986089517923505, "loss": 4.5302, "step": 780650 }, { "epoch": 1.456872537607972, "grad_norm": 0.9664684534072876, "learning_rate": 0.00017985836674730074, "loss": 4.4452, "step": 780700 }, { "epoch": 1.4569658431673504, "grad_norm": 1.2722972631454468, "learning_rate": 0.00017985583817443045, "loss": 4.6632, "step": 780750 }, { "epoch": 1.457059148726729, "grad_norm": 1.0899708271026611, "learning_rate": 0.00017985330946062865, "loss": 4.5601, "step": 780800 }, { "epoch": 1.4571524542861076, "grad_norm": 1.264857292175293, "learning_rate": 0.0001798507806058998, "loss": 4.5637, "step": 780850 }, { "epoch": 1.457245759845486, "grad_norm": 1.1603198051452637, "learning_rate": 0.0001798482516102484, "loss": 4.4476, "step": 780900 }, { "epoch": 1.4573390654048644, "grad_norm": 1.2498596906661987, "learning_rate": 0.00017984572247367887, "loss": 4.5798, "step": 780950 }, { "epoch": 1.457432370964243, "grad_norm": 1.1672319173812866, "learning_rate": 0.0001798431931961957, "loss": 4.5208, "step": 781000 }, { "epoch": 1.4575256765236215, "grad_norm": 1.1481189727783203, "learning_rate": 0.00017984066377780335, "loss": 4.4817, "step": 781050 }, { "epoch": 1.457618982083, "grad_norm": 1.4651484489440918, "learning_rate": 0.00017983813421850627, "loss": 4.3811, "step": 781100 }, { "epoch": 1.4577122876423785, "grad_norm": 1.3164863586425781, "learning_rate": 0.0001798356045183089, "loss": 4.444, "step": 781150 }, { "epoch": 1.457805593201757, "grad_norm": 1.1097347736358643, "learning_rate": 0.00017983307467721578, "loss": 4.5302, "step": 781200 }, { "epoch": 1.4578988987611354, "grad_norm": 1.145127296447754, "learning_rate": 0.00017983054469523132, "loss": 4.2713, "step": 781250 }, { "epoch": 1.4579922043205138, "grad_norm": 0.838314414024353, "learning_rate": 0.00017982801457236002, "loss": 4.6337, "step": 781300 }, { "epoch": 1.4580855098798924, "grad_norm": 1.0443519353866577, "learning_rate": 0.00017982548430860626, "loss": 4.5762, "step": 781350 }, { "epoch": 1.458178815439271, "grad_norm": 1.164795994758606, "learning_rate": 0.00017982295390397458, "loss": 4.4538, "step": 781400 }, { "epoch": 1.4582721209986493, "grad_norm": 1.3701648712158203, "learning_rate": 0.00017982042335846945, "loss": 4.3609, "step": 781450 }, { "epoch": 1.458365426558028, "grad_norm": 0.8058536648750305, "learning_rate": 0.0001798178926720953, "loss": 4.569, "step": 781500 }, { "epoch": 1.4584587321174065, "grad_norm": 1.5502588748931885, "learning_rate": 0.00017981536184485662, "loss": 4.5971, "step": 781550 }, { "epoch": 1.4585520376767849, "grad_norm": 1.0300568342208862, "learning_rate": 0.0001798128308767579, "loss": 4.5243, "step": 781600 }, { "epoch": 1.4586453432361632, "grad_norm": 1.3286131620407104, "learning_rate": 0.00017981029976780354, "loss": 4.4499, "step": 781650 }, { "epoch": 1.4587386487955418, "grad_norm": 1.405519962310791, "learning_rate": 0.00017980776851799808, "loss": 4.5507, "step": 781700 }, { "epoch": 1.4588319543549204, "grad_norm": 1.1173516511917114, "learning_rate": 0.00017980523712734595, "loss": 4.4496, "step": 781750 }, { "epoch": 1.4589252599142988, "grad_norm": 1.2720056772232056, "learning_rate": 0.00017980270559585156, "loss": 4.5271, "step": 781800 }, { "epoch": 1.4590185654736774, "grad_norm": 1.388696312904358, "learning_rate": 0.0001798001739235195, "loss": 4.4553, "step": 781850 }, { "epoch": 1.459111871033056, "grad_norm": 1.3939956426620483, "learning_rate": 0.00017979764211035413, "loss": 4.5954, "step": 781900 }, { "epoch": 1.4592051765924343, "grad_norm": 1.3362599611282349, "learning_rate": 0.00017979511015635998, "loss": 4.5476, "step": 781950 }, { "epoch": 1.4592984821518127, "grad_norm": 1.0439696311950684, "learning_rate": 0.0001797925780615415, "loss": 4.5117, "step": 782000 }, { "epoch": 1.4593917877111913, "grad_norm": 1.1576350927352905, "learning_rate": 0.00017979004582590317, "loss": 4.4788, "step": 782050 }, { "epoch": 1.4594850932705699, "grad_norm": 1.3793590068817139, "learning_rate": 0.00017978751344944945, "loss": 4.6109, "step": 782100 }, { "epoch": 1.4595783988299482, "grad_norm": 1.4899944067001343, "learning_rate": 0.0001797849809321848, "loss": 4.5398, "step": 782150 }, { "epoch": 1.4596717043893268, "grad_norm": 1.2425410747528076, "learning_rate": 0.00017978244827411368, "loss": 4.82, "step": 782200 }, { "epoch": 1.4597650099487054, "grad_norm": 1.2981122732162476, "learning_rate": 0.0001797799154752406, "loss": 4.5504, "step": 782250 }, { "epoch": 1.4598583155080838, "grad_norm": 1.2762749195098877, "learning_rate": 0.00017977738253557, "loss": 4.7921, "step": 782300 }, { "epoch": 1.4599516210674621, "grad_norm": 1.3224351406097412, "learning_rate": 0.00017977484945510636, "loss": 4.3478, "step": 782350 }, { "epoch": 1.4600449266268407, "grad_norm": 1.1675111055374146, "learning_rate": 0.00017977231623385413, "loss": 4.3632, "step": 782400 }, { "epoch": 1.4601382321862193, "grad_norm": 0.8151775598526001, "learning_rate": 0.00017976978287181783, "loss": 4.4394, "step": 782450 }, { "epoch": 1.4602315377455977, "grad_norm": 1.2272781133651733, "learning_rate": 0.00017976724936900188, "loss": 4.6061, "step": 782500 }, { "epoch": 1.4603248433049763, "grad_norm": 1.112001657485962, "learning_rate": 0.00017976471572541075, "loss": 4.7759, "step": 782550 }, { "epoch": 1.4604181488643548, "grad_norm": 1.382346272468567, "learning_rate": 0.00017976218194104896, "loss": 4.3982, "step": 782600 }, { "epoch": 1.4605114544237332, "grad_norm": 1.0208404064178467, "learning_rate": 0.00017975964801592096, "loss": 4.6629, "step": 782650 }, { "epoch": 1.4606047599831116, "grad_norm": 0.9783459305763245, "learning_rate": 0.0001797571139500312, "loss": 4.5397, "step": 782700 }, { "epoch": 1.4606980655424902, "grad_norm": 1.2478793859481812, "learning_rate": 0.00017975457974338417, "loss": 4.3274, "step": 782750 }, { "epoch": 1.4607913711018687, "grad_norm": 1.063894271850586, "learning_rate": 0.00017975204539598434, "loss": 4.5488, "step": 782800 }, { "epoch": 1.460884676661247, "grad_norm": 1.1485687494277954, "learning_rate": 0.00017974951090783616, "loss": 4.4946, "step": 782850 }, { "epoch": 1.4609779822206257, "grad_norm": 1.284026026725769, "learning_rate": 0.00017974697627894416, "loss": 4.5377, "step": 782900 }, { "epoch": 1.4610712877800043, "grad_norm": 1.5278762578964233, "learning_rate": 0.00017974444150931277, "loss": 4.4444, "step": 782950 }, { "epoch": 1.4611645933393826, "grad_norm": 0.8864188194274902, "learning_rate": 0.00017974190659894647, "loss": 4.6194, "step": 783000 }, { "epoch": 1.4611645933393826, "eval_loss": 4.664463043212891, "eval_runtime": 230.4508, "eval_samples_per_second": 11.317, "eval_steps_per_second": 11.317, "eval_tts_loss": 7.590670923171051, "step": 783000 }, { "epoch": 1.461257898898761, "grad_norm": 0.9270356893539429, "learning_rate": 0.00017973937154784975, "loss": 4.5106, "step": 783050 }, { "epoch": 1.4613512044581396, "grad_norm": 0.7343159914016724, "learning_rate": 0.00017973683635602705, "loss": 4.5012, "step": 783100 }, { "epoch": 1.4614445100175182, "grad_norm": 1.3723069429397583, "learning_rate": 0.00017973430102348288, "loss": 4.5665, "step": 783150 }, { "epoch": 1.4615378155768965, "grad_norm": 1.317427158355713, "learning_rate": 0.00017973176555022166, "loss": 4.5044, "step": 783200 }, { "epoch": 1.4616311211362751, "grad_norm": 1.1708024740219116, "learning_rate": 0.00017972922993624794, "loss": 4.3721, "step": 783250 }, { "epoch": 1.4617244266956537, "grad_norm": 1.0961261987686157, "learning_rate": 0.00017972669418156616, "loss": 4.5412, "step": 783300 }, { "epoch": 1.461817732255032, "grad_norm": 1.3925886154174805, "learning_rate": 0.00017972415828618077, "loss": 4.5058, "step": 783350 }, { "epoch": 1.4619110378144105, "grad_norm": 1.340004563331604, "learning_rate": 0.00017972162225009627, "loss": 4.3926, "step": 783400 }, { "epoch": 1.462004343373789, "grad_norm": 1.2730029821395874, "learning_rate": 0.00017971908607331715, "loss": 4.5313, "step": 783450 }, { "epoch": 1.4620976489331676, "grad_norm": 1.127732753753662, "learning_rate": 0.00017971654975584787, "loss": 4.4017, "step": 783500 }, { "epoch": 1.462190954492546, "grad_norm": 1.2324310541152954, "learning_rate": 0.0001797140132976929, "loss": 4.6286, "step": 783550 }, { "epoch": 1.4622842600519246, "grad_norm": 0.9814022779464722, "learning_rate": 0.0001797114766988567, "loss": 4.5723, "step": 783600 }, { "epoch": 1.4623775656113032, "grad_norm": 1.465378999710083, "learning_rate": 0.0001797089399593438, "loss": 4.5183, "step": 783650 }, { "epoch": 1.4624708711706815, "grad_norm": 1.1916699409484863, "learning_rate": 0.00017970640307915864, "loss": 4.3929, "step": 783700 }, { "epoch": 1.46256417673006, "grad_norm": 1.456707239151001, "learning_rate": 0.0001797038660583057, "loss": 4.5955, "step": 783750 }, { "epoch": 1.4626574822894385, "grad_norm": 0.9249477386474609, "learning_rate": 0.0001797013288967895, "loss": 4.5069, "step": 783800 }, { "epoch": 1.462750787848817, "grad_norm": 1.192003607749939, "learning_rate": 0.0001796987915946144, "loss": 4.5622, "step": 783850 }, { "epoch": 1.4628440934081954, "grad_norm": 1.504356026649475, "learning_rate": 0.000179696254151785, "loss": 4.455, "step": 783900 }, { "epoch": 1.462937398967574, "grad_norm": 1.142075777053833, "learning_rate": 0.00017969371656830573, "loss": 4.6414, "step": 783950 }, { "epoch": 1.4630307045269526, "grad_norm": 1.0895936489105225, "learning_rate": 0.00017969117884418108, "loss": 4.3516, "step": 784000 }, { "epoch": 1.463124010086331, "grad_norm": 1.0935274362564087, "learning_rate": 0.00017968864097941555, "loss": 4.5026, "step": 784050 }, { "epoch": 1.4632173156457093, "grad_norm": 1.4914964437484741, "learning_rate": 0.00017968610297401354, "loss": 4.6215, "step": 784100 }, { "epoch": 1.463310621205088, "grad_norm": 1.3076860904693604, "learning_rate": 0.0001796835648279796, "loss": 4.2752, "step": 784150 }, { "epoch": 1.4634039267644665, "grad_norm": 1.3308380842208862, "learning_rate": 0.00017968102654131816, "loss": 4.7631, "step": 784200 }, { "epoch": 1.4634972323238449, "grad_norm": 1.5649120807647705, "learning_rate": 0.00017967848811403377, "loss": 4.2629, "step": 784250 }, { "epoch": 1.4635905378832235, "grad_norm": 1.1743228435516357, "learning_rate": 0.00017967594954613084, "loss": 4.6159, "step": 784300 }, { "epoch": 1.463683843442602, "grad_norm": 1.2146271467208862, "learning_rate": 0.00017967341083761391, "loss": 4.781, "step": 784350 }, { "epoch": 1.4637771490019804, "grad_norm": 1.268615961074829, "learning_rate": 0.0001796708719884874, "loss": 4.561, "step": 784400 }, { "epoch": 1.4638704545613588, "grad_norm": 1.2980406284332275, "learning_rate": 0.0001796683329987558, "loss": 4.4158, "step": 784450 }, { "epoch": 1.4639637601207374, "grad_norm": 1.1251622438430786, "learning_rate": 0.00017966579386842366, "loss": 4.5298, "step": 784500 }, { "epoch": 1.464057065680116, "grad_norm": 1.2585062980651855, "learning_rate": 0.0001796632545974954, "loss": 4.3904, "step": 784550 }, { "epoch": 1.4641503712394943, "grad_norm": 1.199475884437561, "learning_rate": 0.0001796607151859755, "loss": 4.4228, "step": 784600 }, { "epoch": 1.464243676798873, "grad_norm": 1.377000093460083, "learning_rate": 0.00017965817563386844, "loss": 4.5454, "step": 784650 }, { "epoch": 1.4643369823582515, "grad_norm": 1.4283974170684814, "learning_rate": 0.00017965563594117873, "loss": 4.5071, "step": 784700 }, { "epoch": 1.4644302879176299, "grad_norm": 1.2653522491455078, "learning_rate": 0.00017965309610791085, "loss": 4.4487, "step": 784750 }, { "epoch": 1.4645235934770082, "grad_norm": 1.2174487113952637, "learning_rate": 0.00017965055613406927, "loss": 4.5817, "step": 784800 }, { "epoch": 1.4646168990363868, "grad_norm": 1.5407742261886597, "learning_rate": 0.00017964801601965847, "loss": 4.3862, "step": 784850 }, { "epoch": 1.4647102045957654, "grad_norm": 1.3092142343521118, "learning_rate": 0.00017964547576468292, "loss": 4.7279, "step": 784900 }, { "epoch": 1.4648035101551438, "grad_norm": 1.0659105777740479, "learning_rate": 0.00017964293536914714, "loss": 4.46, "step": 784950 }, { "epoch": 1.4648968157145223, "grad_norm": 1.1849784851074219, "learning_rate": 0.00017964039483305558, "loss": 4.5669, "step": 785000 }, { "epoch": 1.464990121273901, "grad_norm": 1.3678511381149292, "learning_rate": 0.00017963785415641274, "loss": 4.5699, "step": 785050 }, { "epoch": 1.4650834268332793, "grad_norm": 1.4592320919036865, "learning_rate": 0.0001796353133392231, "loss": 4.6324, "step": 785100 }, { "epoch": 1.4651767323926577, "grad_norm": 1.1375617980957031, "learning_rate": 0.00017963277238149114, "loss": 4.5655, "step": 785150 }, { "epoch": 1.4652700379520363, "grad_norm": 1.1511839628219604, "learning_rate": 0.00017963023128322137, "loss": 4.4632, "step": 785200 }, { "epoch": 1.4653633435114148, "grad_norm": 1.5093200206756592, "learning_rate": 0.00017962769004441822, "loss": 4.4832, "step": 785250 }, { "epoch": 1.4654566490707932, "grad_norm": 1.2856197357177734, "learning_rate": 0.00017962514866508623, "loss": 4.5414, "step": 785300 }, { "epoch": 1.4655499546301718, "grad_norm": 0.9719734787940979, "learning_rate": 0.00017962260714522987, "loss": 4.635, "step": 785350 }, { "epoch": 1.4656432601895504, "grad_norm": 1.0386929512023926, "learning_rate": 0.00017962006548485356, "loss": 4.3843, "step": 785400 }, { "epoch": 1.4657365657489287, "grad_norm": 1.189141869544983, "learning_rate": 0.00017961752368396194, "loss": 4.6375, "step": 785450 }, { "epoch": 1.465829871308307, "grad_norm": 1.1122851371765137, "learning_rate": 0.00017961498174255932, "loss": 4.5449, "step": 785500 }, { "epoch": 1.4659231768676857, "grad_norm": 0.9573639035224915, "learning_rate": 0.0001796124396606503, "loss": 4.4772, "step": 785550 }, { "epoch": 1.4660164824270643, "grad_norm": 1.1845438480377197, "learning_rate": 0.00017960989743823932, "loss": 4.5613, "step": 785600 }, { "epoch": 1.4661097879864426, "grad_norm": 1.3252160549163818, "learning_rate": 0.0001796073550753309, "loss": 4.5118, "step": 785650 }, { "epoch": 1.4662030935458212, "grad_norm": 1.2630422115325928, "learning_rate": 0.00017960481257192949, "loss": 4.5091, "step": 785700 }, { "epoch": 1.4662963991051998, "grad_norm": 1.1684879064559937, "learning_rate": 0.00017960226992803962, "loss": 4.6421, "step": 785750 }, { "epoch": 1.4663897046645782, "grad_norm": 1.3785613775253296, "learning_rate": 0.00017959972714366571, "loss": 4.63, "step": 785800 }, { "epoch": 1.4664830102239566, "grad_norm": 1.1019433736801147, "learning_rate": 0.0001795971842188123, "loss": 4.5992, "step": 785850 }, { "epoch": 1.4665763157833351, "grad_norm": 1.1086219549179077, "learning_rate": 0.00017959464115348386, "loss": 4.5469, "step": 785900 }, { "epoch": 1.4666696213427137, "grad_norm": 1.684064269065857, "learning_rate": 0.0001795920979476849, "loss": 4.439, "step": 785950 }, { "epoch": 1.466762926902092, "grad_norm": 0.9037010669708252, "learning_rate": 0.00017958955460141987, "loss": 4.5367, "step": 786000 }, { "epoch": 1.466762926902092, "eval_loss": 4.656131744384766, "eval_runtime": 229.1141, "eval_samples_per_second": 11.383, "eval_steps_per_second": 11.383, "eval_tts_loss": 7.611801532577106, "step": 786000 }, { "epoch": 1.4668562324614707, "grad_norm": 1.3579643964767456, "learning_rate": 0.0001795870111146933, "loss": 4.6129, "step": 786050 }, { "epoch": 1.4669495380208493, "grad_norm": 1.293036699295044, "learning_rate": 0.00017958446748750966, "loss": 4.6782, "step": 786100 }, { "epoch": 1.4670428435802276, "grad_norm": 1.2710316181182861, "learning_rate": 0.00017958192371987343, "loss": 4.4116, "step": 786150 }, { "epoch": 1.467136149139606, "grad_norm": 1.1489266157150269, "learning_rate": 0.0001795793798117891, "loss": 4.5693, "step": 786200 }, { "epoch": 1.4672294546989846, "grad_norm": 1.3038530349731445, "learning_rate": 0.0001795768357632612, "loss": 4.5196, "step": 786250 }, { "epoch": 1.4673227602583632, "grad_norm": 1.458841323852539, "learning_rate": 0.00017957429157429414, "loss": 4.3779, "step": 786300 }, { "epoch": 1.4674160658177415, "grad_norm": 1.4784303903579712, "learning_rate": 0.0001795717472448925, "loss": 4.7779, "step": 786350 }, { "epoch": 1.4675093713771201, "grad_norm": 1.1370017528533936, "learning_rate": 0.00017956920277506071, "loss": 4.3304, "step": 786400 }, { "epoch": 1.4676026769364987, "grad_norm": 1.1780179738998413, "learning_rate": 0.00017956665816480327, "loss": 4.5074, "step": 786450 }, { "epoch": 1.467695982495877, "grad_norm": 1.1078062057495117, "learning_rate": 0.00017956411341412467, "loss": 4.5818, "step": 786500 }, { "epoch": 1.4677892880552554, "grad_norm": 1.2079086303710938, "learning_rate": 0.00017956156852302942, "loss": 4.4613, "step": 786550 }, { "epoch": 1.467882593614634, "grad_norm": 0.7940014004707336, "learning_rate": 0.000179559023491522, "loss": 4.3144, "step": 786600 }, { "epoch": 1.4679758991740126, "grad_norm": 1.1313010454177856, "learning_rate": 0.00017955647831960692, "loss": 4.367, "step": 786650 }, { "epoch": 1.468069204733391, "grad_norm": 1.0747709274291992, "learning_rate": 0.0001795539330072886, "loss": 4.6407, "step": 786700 }, { "epoch": 1.4681625102927696, "grad_norm": 1.1071314811706543, "learning_rate": 0.00017955138755457165, "loss": 4.635, "step": 786750 }, { "epoch": 1.468255815852148, "grad_norm": 1.0203903913497925, "learning_rate": 0.00017954884196146048, "loss": 4.4835, "step": 786800 }, { "epoch": 1.4683491214115265, "grad_norm": 1.1635818481445312, "learning_rate": 0.0001795462962279596, "loss": 4.5654, "step": 786850 }, { "epoch": 1.4684424269709049, "grad_norm": 0.8528257608413696, "learning_rate": 0.0001795437503540735, "loss": 4.5521, "step": 786900 }, { "epoch": 1.4685357325302835, "grad_norm": 1.162811040878296, "learning_rate": 0.00017954120433980668, "loss": 4.7068, "step": 786950 }, { "epoch": 1.468629038089662, "grad_norm": 1.000416874885559, "learning_rate": 0.00017953865818516363, "loss": 4.5406, "step": 787000 }, { "epoch": 1.4687223436490404, "grad_norm": 0.9296506643295288, "learning_rate": 0.00017953611189014884, "loss": 4.4976, "step": 787050 }, { "epoch": 1.468815649208419, "grad_norm": 1.2321593761444092, "learning_rate": 0.00017953356545476678, "loss": 4.4909, "step": 787100 }, { "epoch": 1.4689089547677974, "grad_norm": 1.2934787273406982, "learning_rate": 0.00017953101887902202, "loss": 4.4486, "step": 787150 }, { "epoch": 1.469002260327176, "grad_norm": 1.2045717239379883, "learning_rate": 0.00017952847216291899, "loss": 4.5901, "step": 787200 }, { "epoch": 1.4690955658865543, "grad_norm": 1.2582340240478516, "learning_rate": 0.0001795259253064622, "loss": 4.4211, "step": 787250 }, { "epoch": 1.469188871445933, "grad_norm": 1.083167552947998, "learning_rate": 0.0001795233783096561, "loss": 4.3336, "step": 787300 }, { "epoch": 1.4692821770053115, "grad_norm": 1.4086263179779053, "learning_rate": 0.0001795208311725053, "loss": 4.3608, "step": 787350 }, { "epoch": 1.4693754825646899, "grad_norm": 1.2298734188079834, "learning_rate": 0.0001795182838950142, "loss": 4.4603, "step": 787400 }, { "epoch": 1.4694687881240684, "grad_norm": 1.2880661487579346, "learning_rate": 0.00017951573647718728, "loss": 4.4449, "step": 787450 }, { "epoch": 1.4695620936834468, "grad_norm": 1.333662986755371, "learning_rate": 0.00017951318891902912, "loss": 4.5949, "step": 787500 }, { "epoch": 1.4696553992428254, "grad_norm": 1.3357019424438477, "learning_rate": 0.00017951064122054419, "loss": 4.3364, "step": 787550 }, { "epoch": 1.4697487048022038, "grad_norm": 1.4096627235412598, "learning_rate": 0.00017950809338173693, "loss": 4.487, "step": 787600 }, { "epoch": 1.4698420103615824, "grad_norm": 1.102981686592102, "learning_rate": 0.00017950554540261185, "loss": 4.3938, "step": 787650 }, { "epoch": 1.469935315920961, "grad_norm": 1.043299674987793, "learning_rate": 0.00017950299728317353, "loss": 4.3562, "step": 787700 }, { "epoch": 1.4700286214803393, "grad_norm": 1.1658543348312378, "learning_rate": 0.0001795004490234264, "loss": 4.5779, "step": 787750 }, { "epoch": 1.470121927039718, "grad_norm": 0.7308076024055481, "learning_rate": 0.00017949790062337496, "loss": 4.3415, "step": 787800 }, { "epoch": 1.4702152325990963, "grad_norm": 1.1724220514297485, "learning_rate": 0.0001794953520830237, "loss": 4.5231, "step": 787850 }, { "epoch": 1.4703085381584748, "grad_norm": 1.3695881366729736, "learning_rate": 0.00017949280340237716, "loss": 4.3329, "step": 787900 }, { "epoch": 1.4704018437178532, "grad_norm": 0.9566438794136047, "learning_rate": 0.00017949025458143978, "loss": 4.4737, "step": 787950 }, { "epoch": 1.4704951492772318, "grad_norm": 1.286284327507019, "learning_rate": 0.00017948770562021608, "loss": 4.6317, "step": 788000 }, { "epoch": 1.4705884548366104, "grad_norm": 1.3841352462768555, "learning_rate": 0.00017948515651871057, "loss": 4.4761, "step": 788050 }, { "epoch": 1.4706817603959887, "grad_norm": 1.2112541198730469, "learning_rate": 0.00017948260727692776, "loss": 4.5475, "step": 788100 }, { "epoch": 1.4707750659553673, "grad_norm": 1.2862204313278198, "learning_rate": 0.00017948005789487215, "loss": 4.4451, "step": 788150 }, { "epoch": 1.4708683715147457, "grad_norm": 1.3253238201141357, "learning_rate": 0.00017947750837254818, "loss": 4.5634, "step": 788200 }, { "epoch": 1.4709616770741243, "grad_norm": 1.5073384046554565, "learning_rate": 0.0001794749587099604, "loss": 4.7093, "step": 788250 }, { "epoch": 1.4710549826335027, "grad_norm": 1.2184346914291382, "learning_rate": 0.00017947240890711332, "loss": 4.5302, "step": 788300 }, { "epoch": 1.4711482881928812, "grad_norm": 1.0097026824951172, "learning_rate": 0.00017946985896401144, "loss": 4.4732, "step": 788350 }, { "epoch": 1.4712415937522598, "grad_norm": 1.2293901443481445, "learning_rate": 0.00017946730888065922, "loss": 4.4851, "step": 788400 }, { "epoch": 1.4713348993116382, "grad_norm": 1.0063352584838867, "learning_rate": 0.00017946475865706117, "loss": 4.6846, "step": 788450 }, { "epoch": 1.4714282048710168, "grad_norm": 1.1784255504608154, "learning_rate": 0.0001794622082932218, "loss": 4.5811, "step": 788500 }, { "epoch": 1.4715215104303951, "grad_norm": 1.0941123962402344, "learning_rate": 0.00017945965778914564, "loss": 4.472, "step": 788550 }, { "epoch": 1.4716148159897737, "grad_norm": 1.2084468603134155, "learning_rate": 0.00017945710714483716, "loss": 4.5465, "step": 788600 }, { "epoch": 1.471708121549152, "grad_norm": 1.2276986837387085, "learning_rate": 0.00017945455636030085, "loss": 4.4089, "step": 788650 }, { "epoch": 1.4718014271085307, "grad_norm": 1.2694859504699707, "learning_rate": 0.00017945200543554126, "loss": 4.4665, "step": 788700 }, { "epoch": 1.4718947326679093, "grad_norm": 1.231745958328247, "learning_rate": 0.00017944945437056283, "loss": 4.4157, "step": 788750 }, { "epoch": 1.4719880382272876, "grad_norm": 1.3292145729064941, "learning_rate": 0.0001794469031653701, "loss": 4.68, "step": 788800 }, { "epoch": 1.4720813437866662, "grad_norm": 1.4053733348846436, "learning_rate": 0.00017944435181996757, "loss": 4.6235, "step": 788850 }, { "epoch": 1.4721746493460446, "grad_norm": 0.9814192056655884, "learning_rate": 0.00017944180033435973, "loss": 4.4021, "step": 788900 }, { "epoch": 1.4722679549054232, "grad_norm": 1.3460367918014526, "learning_rate": 0.00017943924870855108, "loss": 4.6169, "step": 788950 }, { "epoch": 1.4723612604648015, "grad_norm": 1.2339051961898804, "learning_rate": 0.00017943669694254618, "loss": 4.5133, "step": 789000 }, { "epoch": 1.4723612604648015, "eval_loss": 4.659470558166504, "eval_runtime": 230.5572, "eval_samples_per_second": 11.312, "eval_steps_per_second": 11.312, "eval_tts_loss": 7.628651398721157, "step": 789000 }, { "epoch": 1.4724545660241801, "grad_norm": 1.6879583597183228, "learning_rate": 0.00017943414503634946, "loss": 4.4692, "step": 789050 }, { "epoch": 1.4725478715835587, "grad_norm": 1.1456642150878906, "learning_rate": 0.00017943159298996542, "loss": 4.3057, "step": 789100 }, { "epoch": 1.472641177142937, "grad_norm": 1.1720296144485474, "learning_rate": 0.00017942904080339863, "loss": 4.5454, "step": 789150 }, { "epoch": 1.4727344827023157, "grad_norm": 1.6705217361450195, "learning_rate": 0.00017942648847665358, "loss": 4.3656, "step": 789200 }, { "epoch": 1.472827788261694, "grad_norm": 1.2254401445388794, "learning_rate": 0.0001794239360097347, "loss": 4.5714, "step": 789250 }, { "epoch": 1.4729210938210726, "grad_norm": 1.015315294265747, "learning_rate": 0.00017942138340264657, "loss": 4.2884, "step": 789300 }, { "epoch": 1.473014399380451, "grad_norm": 1.1678085327148438, "learning_rate": 0.0001794188306553937, "loss": 4.4241, "step": 789350 }, { "epoch": 1.4731077049398296, "grad_norm": 1.1887527704238892, "learning_rate": 0.00017941627776798052, "loss": 4.5137, "step": 789400 }, { "epoch": 1.4732010104992082, "grad_norm": 1.1687119007110596, "learning_rate": 0.00017941372474041163, "loss": 4.542, "step": 789450 }, { "epoch": 1.4732943160585865, "grad_norm": 1.3805161714553833, "learning_rate": 0.00017941117157269147, "loss": 4.4058, "step": 789500 }, { "epoch": 1.473387621617965, "grad_norm": 0.94623202085495, "learning_rate": 0.00017940861826482455, "loss": 4.7311, "step": 789550 }, { "epoch": 1.4734809271773435, "grad_norm": 1.4713083505630493, "learning_rate": 0.00017940606481681542, "loss": 4.6261, "step": 789600 }, { "epoch": 1.473574232736722, "grad_norm": 1.1844854354858398, "learning_rate": 0.00017940351122866852, "loss": 4.7558, "step": 789650 }, { "epoch": 1.4736675382961004, "grad_norm": 1.3652324676513672, "learning_rate": 0.0001794009575003884, "loss": 4.3928, "step": 789700 }, { "epoch": 1.473760843855479, "grad_norm": 1.1683701276779175, "learning_rate": 0.00017939840363197958, "loss": 4.3961, "step": 789750 }, { "epoch": 1.4738541494148576, "grad_norm": 1.3418668508529663, "learning_rate": 0.00017939584962344655, "loss": 4.7185, "step": 789800 }, { "epoch": 1.473947454974236, "grad_norm": 0.8877221941947937, "learning_rate": 0.00017939329547479382, "loss": 4.4371, "step": 789850 }, { "epoch": 1.4740407605336145, "grad_norm": 0.9251667857170105, "learning_rate": 0.00017939074118602589, "loss": 4.3857, "step": 789900 }, { "epoch": 1.474134066092993, "grad_norm": 0.878736674785614, "learning_rate": 0.00017938818675714726, "loss": 4.4157, "step": 789950 }, { "epoch": 1.4742273716523715, "grad_norm": 0.8711289763450623, "learning_rate": 0.00017938563218816245, "loss": 4.3508, "step": 790000 }, { "epoch": 1.4743206772117499, "grad_norm": 1.113279104232788, "learning_rate": 0.00017938307747907597, "loss": 4.5067, "step": 790050 }, { "epoch": 1.4744139827711285, "grad_norm": 0.8679510951042175, "learning_rate": 0.00017938052262989233, "loss": 4.5698, "step": 790100 }, { "epoch": 1.474507288330507, "grad_norm": 0.9091683626174927, "learning_rate": 0.00017937796764061603, "loss": 4.5454, "step": 790150 }, { "epoch": 1.4746005938898854, "grad_norm": 1.1004194021224976, "learning_rate": 0.0001793754125112516, "loss": 4.4198, "step": 790200 }, { "epoch": 1.474693899449264, "grad_norm": 1.4882850646972656, "learning_rate": 0.0001793728572418035, "loss": 4.3744, "step": 790250 }, { "epoch": 1.4747872050086424, "grad_norm": 1.417324423789978, "learning_rate": 0.0001793703018322763, "loss": 4.4455, "step": 790300 }, { "epoch": 1.474880510568021, "grad_norm": 1.2036219835281372, "learning_rate": 0.00017936774628267448, "loss": 4.5401, "step": 790350 }, { "epoch": 1.4749738161273993, "grad_norm": 1.1137529611587524, "learning_rate": 0.00017936519059300255, "loss": 4.4007, "step": 790400 }, { "epoch": 1.475067121686778, "grad_norm": 1.1449038982391357, "learning_rate": 0.000179362634763265, "loss": 4.4674, "step": 790450 }, { "epoch": 1.4751604272461565, "grad_norm": 1.043935775756836, "learning_rate": 0.0001793600787934664, "loss": 4.4383, "step": 790500 }, { "epoch": 1.4752537328055348, "grad_norm": 1.226982593536377, "learning_rate": 0.0001793575226836112, "loss": 4.5764, "step": 790550 }, { "epoch": 1.4753470383649134, "grad_norm": 1.2220382690429688, "learning_rate": 0.00017935496643370396, "loss": 4.703, "step": 790600 }, { "epoch": 1.4754403439242918, "grad_norm": 1.697373390197754, "learning_rate": 0.00017935241004374914, "loss": 4.4826, "step": 790650 }, { "epoch": 1.4755336494836704, "grad_norm": 1.0024616718292236, "learning_rate": 0.0001793498535137513, "loss": 4.5476, "step": 790700 }, { "epoch": 1.4756269550430487, "grad_norm": 1.2340408563613892, "learning_rate": 0.00017934729684371492, "loss": 4.4461, "step": 790750 }, { "epoch": 1.4757202606024273, "grad_norm": 1.2432856559753418, "learning_rate": 0.00017934474003364454, "loss": 4.5073, "step": 790800 }, { "epoch": 1.475813566161806, "grad_norm": 1.28501558303833, "learning_rate": 0.00017934218308354461, "loss": 4.5236, "step": 790850 }, { "epoch": 1.4759068717211843, "grad_norm": 1.2757658958435059, "learning_rate": 0.00017933962599341974, "loss": 4.4045, "step": 790900 }, { "epoch": 1.4760001772805629, "grad_norm": 1.0906072854995728, "learning_rate": 0.00017933706876327435, "loss": 4.8676, "step": 790950 }, { "epoch": 1.4760934828399412, "grad_norm": 1.0470792055130005, "learning_rate": 0.00017933451139311304, "loss": 4.5616, "step": 791000 }, { "epoch": 1.4761867883993198, "grad_norm": 1.125554084777832, "learning_rate": 0.00017933195388294024, "loss": 4.517, "step": 791050 }, { "epoch": 1.4762800939586982, "grad_norm": 0.957223653793335, "learning_rate": 0.00017932939623276051, "loss": 4.5438, "step": 791100 }, { "epoch": 1.4763733995180768, "grad_norm": 1.2896065711975098, "learning_rate": 0.00017932683844257836, "loss": 4.5723, "step": 791150 }, { "epoch": 1.4764667050774554, "grad_norm": 1.337079405784607, "learning_rate": 0.0001793242805123983, "loss": 4.5347, "step": 791200 }, { "epoch": 1.4765600106368337, "grad_norm": 1.2245360612869263, "learning_rate": 0.0001793217224422248, "loss": 4.2567, "step": 791250 }, { "epoch": 1.4766533161962123, "grad_norm": 1.1774864196777344, "learning_rate": 0.00017931916423206246, "loss": 4.6986, "step": 791300 }, { "epoch": 1.4767466217555907, "grad_norm": 1.060836672782898, "learning_rate": 0.00017931660588191574, "loss": 4.2436, "step": 791350 }, { "epoch": 1.4768399273149693, "grad_norm": 1.317107915878296, "learning_rate": 0.00017931404739178917, "loss": 4.4919, "step": 791400 }, { "epoch": 1.4769332328743476, "grad_norm": 0.9867435097694397, "learning_rate": 0.00017931148876168727, "loss": 4.619, "step": 791450 }, { "epoch": 1.4770265384337262, "grad_norm": 1.225031852722168, "learning_rate": 0.00017930892999161452, "loss": 4.5163, "step": 791500 }, { "epoch": 1.4771198439931048, "grad_norm": 1.4039103984832764, "learning_rate": 0.0001793063710815755, "loss": 4.6163, "step": 791550 }, { "epoch": 1.4772131495524832, "grad_norm": 0.9647724628448486, "learning_rate": 0.00017930381203157465, "loss": 4.5455, "step": 791600 }, { "epoch": 1.4773064551118618, "grad_norm": 1.5071864128112793, "learning_rate": 0.00017930125284161655, "loss": 4.5549, "step": 791650 }, { "epoch": 1.4773997606712401, "grad_norm": 1.3807916641235352, "learning_rate": 0.0001792986935117057, "loss": 4.4807, "step": 791700 }, { "epoch": 1.4774930662306187, "grad_norm": 1.5114126205444336, "learning_rate": 0.0001792961340418466, "loss": 4.6655, "step": 791750 }, { "epoch": 1.477586371789997, "grad_norm": 1.319785237312317, "learning_rate": 0.00017929357443204378, "loss": 4.5055, "step": 791800 }, { "epoch": 1.4776796773493757, "grad_norm": 1.2318146228790283, "learning_rate": 0.00017929101468230175, "loss": 4.4168, "step": 791850 }, { "epoch": 1.4777729829087543, "grad_norm": 1.2079684734344482, "learning_rate": 0.00017928845479262502, "loss": 4.5965, "step": 791900 }, { "epoch": 1.4778662884681326, "grad_norm": 1.3477561473846436, "learning_rate": 0.00017928589476301813, "loss": 4.5649, "step": 791950 }, { "epoch": 1.4779595940275112, "grad_norm": 1.3282418251037598, "learning_rate": 0.0001792833345934856, "loss": 4.5012, "step": 792000 }, { "epoch": 1.4779595940275112, "eval_loss": 4.662326335906982, "eval_runtime": 231.7953, "eval_samples_per_second": 11.251, "eval_steps_per_second": 11.251, "eval_tts_loss": 7.65926284381857, "step": 792000 }, { "epoch": 1.4780528995868896, "grad_norm": 1.1975152492523193, "learning_rate": 0.0001792807742840319, "loss": 4.6364, "step": 792050 }, { "epoch": 1.4781462051462682, "grad_norm": 1.2236460447311401, "learning_rate": 0.0001792782138346616, "loss": 4.5451, "step": 792100 }, { "epoch": 1.4782395107056465, "grad_norm": 1.312242031097412, "learning_rate": 0.0001792756532453792, "loss": 4.6007, "step": 792150 }, { "epoch": 1.478332816265025, "grad_norm": 1.1071360111236572, "learning_rate": 0.0001792730925161892, "loss": 4.508, "step": 792200 }, { "epoch": 1.4784261218244037, "grad_norm": 1.3665574789047241, "learning_rate": 0.00017927053164709618, "loss": 4.5225, "step": 792250 }, { "epoch": 1.478519427383782, "grad_norm": 1.171431064605713, "learning_rate": 0.0001792679706381046, "loss": 4.4192, "step": 792300 }, { "epoch": 1.4786127329431606, "grad_norm": 1.178812861442566, "learning_rate": 0.00017926540948921902, "loss": 4.3614, "step": 792350 }, { "epoch": 1.478706038502539, "grad_norm": 1.71433687210083, "learning_rate": 0.00017926284820044393, "loss": 4.5727, "step": 792400 }, { "epoch": 1.4787993440619176, "grad_norm": 1.0566483736038208, "learning_rate": 0.00017926028677178382, "loss": 4.7095, "step": 792450 }, { "epoch": 1.478892649621296, "grad_norm": 0.9818129539489746, "learning_rate": 0.0001792577252032433, "loss": 4.4276, "step": 792500 }, { "epoch": 1.4789859551806745, "grad_norm": 1.4989511966705322, "learning_rate": 0.00017925516349482683, "loss": 4.4005, "step": 792550 }, { "epoch": 1.4790792607400531, "grad_norm": 1.5087659358978271, "learning_rate": 0.00017925260164653893, "loss": 4.3041, "step": 792600 }, { "epoch": 1.4791725662994315, "grad_norm": 1.697153091430664, "learning_rate": 0.00017925003965838413, "loss": 4.6344, "step": 792650 }, { "epoch": 1.47926587185881, "grad_norm": 1.193044900894165, "learning_rate": 0.00017924747753036697, "loss": 4.3661, "step": 792700 }, { "epoch": 1.4793591774181885, "grad_norm": 0.9891694784164429, "learning_rate": 0.00017924491526249194, "loss": 4.4906, "step": 792750 }, { "epoch": 1.479452482977567, "grad_norm": 0.8586355447769165, "learning_rate": 0.00017924235285476359, "loss": 4.4099, "step": 792800 }, { "epoch": 1.4795457885369454, "grad_norm": 1.159881830215454, "learning_rate": 0.00017923979030718643, "loss": 4.6357, "step": 792850 }, { "epoch": 1.479639094096324, "grad_norm": 1.2207287549972534, "learning_rate": 0.000179237227619765, "loss": 4.369, "step": 792900 }, { "epoch": 1.4797323996557026, "grad_norm": 0.933800458908081, "learning_rate": 0.00017923466479250377, "loss": 4.65, "step": 792950 }, { "epoch": 1.479825705215081, "grad_norm": 1.2934019565582275, "learning_rate": 0.00017923210182540733, "loss": 4.5965, "step": 793000 }, { "epoch": 1.4799190107744595, "grad_norm": 1.1815540790557861, "learning_rate": 0.00017922953871848013, "loss": 4.6097, "step": 793050 }, { "epoch": 1.480012316333838, "grad_norm": 1.057965636253357, "learning_rate": 0.00017922697547172677, "loss": 4.5457, "step": 793100 }, { "epoch": 1.4801056218932165, "grad_norm": 1.0960670709609985, "learning_rate": 0.00017922441208515175, "loss": 4.6732, "step": 793150 }, { "epoch": 1.4801989274525948, "grad_norm": 1.1258125305175781, "learning_rate": 0.00017922184855875955, "loss": 4.3684, "step": 793200 }, { "epoch": 1.4802922330119734, "grad_norm": 1.3454194068908691, "learning_rate": 0.00017921928489255472, "loss": 4.6199, "step": 793250 }, { "epoch": 1.480385538571352, "grad_norm": 0.9908848404884338, "learning_rate": 0.00017921672108654182, "loss": 4.2746, "step": 793300 }, { "epoch": 1.4804788441307304, "grad_norm": 1.01479172706604, "learning_rate": 0.00017921415714072537, "loss": 4.4438, "step": 793350 }, { "epoch": 1.480572149690109, "grad_norm": 1.239121437072754, "learning_rate": 0.0001792115930551098, "loss": 4.4705, "step": 793400 }, { "epoch": 1.4806654552494873, "grad_norm": 1.5316052436828613, "learning_rate": 0.00017920902882969975, "loss": 4.4369, "step": 793450 }, { "epoch": 1.480758760808866, "grad_norm": 1.085817575454712, "learning_rate": 0.0001792064644644997, "loss": 4.6549, "step": 793500 }, { "epoch": 1.4808520663682443, "grad_norm": 1.0349148511886597, "learning_rate": 0.00017920389995951418, "loss": 4.3684, "step": 793550 }, { "epoch": 1.4809453719276229, "grad_norm": 0.976683497428894, "learning_rate": 0.0001792013353147477, "loss": 4.3673, "step": 793600 }, { "epoch": 1.4810386774870015, "grad_norm": 1.4018877744674683, "learning_rate": 0.00017919877053020481, "loss": 4.4934, "step": 793650 }, { "epoch": 1.4811319830463798, "grad_norm": 1.4406815767288208, "learning_rate": 0.00017919620560589002, "loss": 4.5587, "step": 793700 }, { "epoch": 1.4812252886057584, "grad_norm": 1.4577031135559082, "learning_rate": 0.00017919364054180784, "loss": 4.5569, "step": 793750 }, { "epoch": 1.4813185941651368, "grad_norm": 1.054131031036377, "learning_rate": 0.00017919107533796288, "loss": 4.3531, "step": 793800 }, { "epoch": 1.4814118997245154, "grad_norm": 1.5847495794296265, "learning_rate": 0.00017918850999435953, "loss": 4.4539, "step": 793850 }, { "epoch": 1.4815052052838937, "grad_norm": 1.0252007246017456, "learning_rate": 0.00017918594451100242, "loss": 4.6028, "step": 793900 }, { "epoch": 1.4815985108432723, "grad_norm": 1.2365378141403198, "learning_rate": 0.0001791833788878961, "loss": 4.6283, "step": 793950 }, { "epoch": 1.481691816402651, "grad_norm": 1.19844388961792, "learning_rate": 0.00017918081312504497, "loss": 4.4593, "step": 794000 }, { "epoch": 1.4817851219620293, "grad_norm": 1.2554502487182617, "learning_rate": 0.0001791782472224537, "loss": 4.5264, "step": 794050 }, { "epoch": 1.4818784275214079, "grad_norm": 0.8783405423164368, "learning_rate": 0.00017917568118012674, "loss": 4.514, "step": 794100 }, { "epoch": 1.4819717330807862, "grad_norm": 1.4418538808822632, "learning_rate": 0.0001791731149980686, "loss": 4.5873, "step": 794150 }, { "epoch": 1.4820650386401648, "grad_norm": 1.1350988149642944, "learning_rate": 0.00017917054867628388, "loss": 4.4134, "step": 794200 }, { "epoch": 1.4821583441995432, "grad_norm": 1.1560182571411133, "learning_rate": 0.00017916798221477706, "loss": 4.6927, "step": 794250 }, { "epoch": 1.4822516497589218, "grad_norm": 1.0147302150726318, "learning_rate": 0.0001791654156135527, "loss": 4.4419, "step": 794300 }, { "epoch": 1.4823449553183004, "grad_norm": 1.2965822219848633, "learning_rate": 0.00017916284887261528, "loss": 4.5846, "step": 794350 }, { "epoch": 1.4824382608776787, "grad_norm": 1.041540265083313, "learning_rate": 0.00017916028199196937, "loss": 4.6579, "step": 794400 }, { "epoch": 1.4825315664370573, "grad_norm": 1.0287343263626099, "learning_rate": 0.0001791577149716195, "loss": 4.7136, "step": 794450 }, { "epoch": 1.4826248719964357, "grad_norm": 0.9339358806610107, "learning_rate": 0.00017915514781157018, "loss": 4.4314, "step": 794500 }, { "epoch": 1.4827181775558143, "grad_norm": 1.3065624237060547, "learning_rate": 0.00017915258051182597, "loss": 4.5107, "step": 794550 }, { "epoch": 1.4828114831151926, "grad_norm": 1.12920343875885, "learning_rate": 0.00017915001307239136, "loss": 4.6571, "step": 794600 }, { "epoch": 1.4829047886745712, "grad_norm": 1.2127999067306519, "learning_rate": 0.00017914744549327093, "loss": 4.4306, "step": 794650 }, { "epoch": 1.4829980942339498, "grad_norm": 1.3566595315933228, "learning_rate": 0.0001791448777744692, "loss": 4.4493, "step": 794700 }, { "epoch": 1.4830913997933282, "grad_norm": 1.1283612251281738, "learning_rate": 0.0001791423099159906, "loss": 4.737, "step": 794750 }, { "epoch": 1.4831847053527067, "grad_norm": 1.2107841968536377, "learning_rate": 0.00017913974191783983, "loss": 4.6178, "step": 794800 }, { "epoch": 1.483278010912085, "grad_norm": 1.6039950847625732, "learning_rate": 0.00017913717378002133, "loss": 4.2511, "step": 794850 }, { "epoch": 1.4833713164714637, "grad_norm": 1.1689960956573486, "learning_rate": 0.0001791346055025396, "loss": 4.3968, "step": 794900 }, { "epoch": 1.483464622030842, "grad_norm": 1.2557625770568848, "learning_rate": 0.00017913203708539926, "loss": 4.5679, "step": 794950 }, { "epoch": 1.4835579275902206, "grad_norm": 1.2830575704574585, "learning_rate": 0.00017912946852860476, "loss": 4.5107, "step": 795000 }, { "epoch": 1.4835579275902206, "eval_loss": 4.662529468536377, "eval_runtime": 230.2426, "eval_samples_per_second": 11.327, "eval_steps_per_second": 11.327, "eval_tts_loss": 7.593943142487833, "step": 795000 } ], "logging_steps": 50, "max_steps": 3751118, "num_input_tokens_seen": 0, "num_train_epochs": 7, "save_steps": 3000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 3.873019690510975e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }