{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.8453483679691345, "eval_steps": 3000, "global_step": 453000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 9.330555937849166e-05, "grad_norm": 93.67707824707031, "learning_rate": 8.166666666666666e-07, "loss": 16.1123, "step": 50 }, { "epoch": 0.00018661111875698333, "grad_norm": 7.377334117889404, "learning_rate": 1.65e-06, "loss": 12.8959, "step": 100 }, { "epoch": 0.00027991667813547503, "grad_norm": 8.839871406555176, "learning_rate": 2.4833333333333334e-06, "loss": 11.8846, "step": 150 }, { "epoch": 0.00037322223751396665, "grad_norm": 16.873918533325195, "learning_rate": 3.3166666666666665e-06, "loss": 11.4614, "step": 200 }, { "epoch": 0.0004665277968924583, "grad_norm": 60.141456604003906, "learning_rate": 4.15e-06, "loss": 10.2624, "step": 250 }, { "epoch": 0.0005598333562709501, "grad_norm": 26.98830223083496, "learning_rate": 4.983333333333334e-06, "loss": 9.1782, "step": 300 }, { "epoch": 0.0006531389156494417, "grad_norm": 5.853992462158203, "learning_rate": 5.816666666666667e-06, "loss": 8.381, "step": 350 }, { "epoch": 0.0007464444750279333, "grad_norm": 38.342830657958984, "learning_rate": 6.650000000000001e-06, "loss": 7.5826, "step": 400 }, { "epoch": 0.000839750034406425, "grad_norm": 19.33097267150879, "learning_rate": 7.483333333333334e-06, "loss": 7.4673, "step": 450 }, { "epoch": 0.0009330555937849167, "grad_norm": 9.538616180419922, "learning_rate": 8.316666666666668e-06, "loss": 7.0077, "step": 500 }, { "epoch": 0.0010263611531634083, "grad_norm": 10.989666938781738, "learning_rate": 9.15e-06, "loss": 7.2275, "step": 550 }, { "epoch": 0.0011196667125419001, "grad_norm": 9.881364822387695, "learning_rate": 9.983333333333333e-06, "loss": 6.8267, "step": 600 }, { "epoch": 0.0012129722719203917, "grad_norm": 12.614446640014648, "learning_rate": 1.0816666666666666e-05, "loss": 7.133, "step": 650 }, { "epoch": 0.0013062778312988835, "grad_norm": 6.586646556854248, "learning_rate": 1.1650000000000002e-05, "loss": 7.0883, "step": 700 }, { "epoch": 0.001399583390677375, "grad_norm": 10.388212203979492, "learning_rate": 1.2483333333333335e-05, "loss": 6.8631, "step": 750 }, { "epoch": 0.0014928889500558666, "grad_norm": 5.741134166717529, "learning_rate": 1.3316666666666666e-05, "loss": 6.6987, "step": 800 }, { "epoch": 0.0015861945094343584, "grad_norm": 6.758159160614014, "learning_rate": 1.415e-05, "loss": 6.8654, "step": 850 }, { "epoch": 0.00167950006881285, "grad_norm": 11.188956260681152, "learning_rate": 1.4983333333333336e-05, "loss": 6.6844, "step": 900 }, { "epoch": 0.0017728056281913417, "grad_norm": 9.388659477233887, "learning_rate": 1.5816666666666667e-05, "loss": 6.6985, "step": 950 }, { "epoch": 0.0018661111875698333, "grad_norm": 17.585216522216797, "learning_rate": 1.665e-05, "loss": 6.5966, "step": 1000 }, { "epoch": 0.001959416746948325, "grad_norm": 5.970058917999268, "learning_rate": 1.7483333333333336e-05, "loss": 6.7767, "step": 1050 }, { "epoch": 0.0020527223063268167, "grad_norm": 13.584476470947266, "learning_rate": 1.8316666666666667e-05, "loss": 6.7776, "step": 1100 }, { "epoch": 0.0021460278657053082, "grad_norm": 8.928284645080566, "learning_rate": 1.915e-05, "loss": 6.5926, "step": 1150 }, { "epoch": 0.0022393334250838002, "grad_norm": 9.114434242248535, "learning_rate": 1.9983333333333336e-05, "loss": 6.687, "step": 1200 }, { "epoch": 0.002332638984462292, "grad_norm": 11.371248245239258, "learning_rate": 2.0816666666666667e-05, "loss": 6.4247, "step": 1250 }, { "epoch": 0.0024259445438407834, "grad_norm": 9.375133514404297, "learning_rate": 2.165e-05, "loss": 6.3891, "step": 1300 }, { "epoch": 0.002519250103219275, "grad_norm": 7.963316440582275, "learning_rate": 2.2483333333333335e-05, "loss": 6.7139, "step": 1350 }, { "epoch": 0.002612555662597767, "grad_norm": 9.820133209228516, "learning_rate": 2.3316666666666666e-05, "loss": 6.5634, "step": 1400 }, { "epoch": 0.0027058612219762585, "grad_norm": 10.639363288879395, "learning_rate": 2.415e-05, "loss": 6.6883, "step": 1450 }, { "epoch": 0.00279916678135475, "grad_norm": 6.419040679931641, "learning_rate": 2.4983333333333335e-05, "loss": 6.6271, "step": 1500 }, { "epoch": 0.0028924723407332416, "grad_norm": 8.272869110107422, "learning_rate": 2.5816666666666666e-05, "loss": 6.4248, "step": 1550 }, { "epoch": 0.002985777900111733, "grad_norm": 7.704101085662842, "learning_rate": 2.6650000000000004e-05, "loss": 6.3276, "step": 1600 }, { "epoch": 0.003079083459490225, "grad_norm": 9.784399032592773, "learning_rate": 2.748333333333333e-05, "loss": 6.2883, "step": 1650 }, { "epoch": 0.0031723890188687168, "grad_norm": 11.014293670654297, "learning_rate": 2.831666666666667e-05, "loss": 6.4566, "step": 1700 }, { "epoch": 0.0032656945782472083, "grad_norm": 10.516816139221191, "learning_rate": 2.915e-05, "loss": 6.2076, "step": 1750 }, { "epoch": 0.0033590001376257, "grad_norm": 11.228531837463379, "learning_rate": 2.9983333333333335e-05, "loss": 6.3539, "step": 1800 }, { "epoch": 0.003452305697004192, "grad_norm": 9.591064453125, "learning_rate": 3.0816666666666666e-05, "loss": 6.418, "step": 1850 }, { "epoch": 0.0035456112563826835, "grad_norm": 10.729142189025879, "learning_rate": 3.1650000000000004e-05, "loss": 6.0638, "step": 1900 }, { "epoch": 0.003638916815761175, "grad_norm": 8.057198524475098, "learning_rate": 3.2483333333333335e-05, "loss": 6.2542, "step": 1950 }, { "epoch": 0.0037322223751396666, "grad_norm": 8.599757194519043, "learning_rate": 3.3316666666666666e-05, "loss": 6.2936, "step": 2000 }, { "epoch": 0.0038255279345181586, "grad_norm": 10.339815139770508, "learning_rate": 3.415e-05, "loss": 6.1887, "step": 2050 }, { "epoch": 0.00391883349389665, "grad_norm": 8.34434700012207, "learning_rate": 3.4983333333333334e-05, "loss": 6.3519, "step": 2100 }, { "epoch": 0.004012139053275142, "grad_norm": 8.931737899780273, "learning_rate": 3.581666666666667e-05, "loss": 5.9798, "step": 2150 }, { "epoch": 0.004105444612653633, "grad_norm": 7.339552402496338, "learning_rate": 3.665e-05, "loss": 5.8089, "step": 2200 }, { "epoch": 0.004198750172032125, "grad_norm": 10.72778606414795, "learning_rate": 3.7483333333333334e-05, "loss": 6.154, "step": 2250 }, { "epoch": 0.0042920557314106164, "grad_norm": 7.218290328979492, "learning_rate": 3.8316666666666665e-05, "loss": 6.2445, "step": 2300 }, { "epoch": 0.004385361290789108, "grad_norm": 7.798749923706055, "learning_rate": 3.915e-05, "loss": 5.9981, "step": 2350 }, { "epoch": 0.0044786668501676005, "grad_norm": 7.12747049331665, "learning_rate": 3.9983333333333334e-05, "loss": 6.0511, "step": 2400 }, { "epoch": 0.004571972409546092, "grad_norm": 6.343996524810791, "learning_rate": 4.081666666666667e-05, "loss": 6.0378, "step": 2450 }, { "epoch": 0.004665277968924584, "grad_norm": 6.697947025299072, "learning_rate": 4.165e-05, "loss": 5.8112, "step": 2500 }, { "epoch": 0.004758583528303075, "grad_norm": 8.327080726623535, "learning_rate": 4.2483333333333334e-05, "loss": 6.2906, "step": 2550 }, { "epoch": 0.004851889087681567, "grad_norm": 5.573576927185059, "learning_rate": 4.3316666666666665e-05, "loss": 5.8131, "step": 2600 }, { "epoch": 0.004945194647060058, "grad_norm": 6.693691253662109, "learning_rate": 4.415e-05, "loss": 5.7015, "step": 2650 }, { "epoch": 0.00503850020643855, "grad_norm": 6.914066314697266, "learning_rate": 4.4983333333333334e-05, "loss": 5.8004, "step": 2700 }, { "epoch": 0.005131805765817041, "grad_norm": 8.469379425048828, "learning_rate": 4.581666666666667e-05, "loss": 5.9158, "step": 2750 }, { "epoch": 0.005225111325195534, "grad_norm": 7.3673295974731445, "learning_rate": 4.665e-05, "loss": 5.969, "step": 2800 }, { "epoch": 0.005318416884574025, "grad_norm": 7.18582010269165, "learning_rate": 4.748333333333333e-05, "loss": 6.0261, "step": 2850 }, { "epoch": 0.005411722443952517, "grad_norm": 5.982076644897461, "learning_rate": 4.831666666666667e-05, "loss": 6.023, "step": 2900 }, { "epoch": 0.0055050280033310086, "grad_norm": 5.787336826324463, "learning_rate": 4.915e-05, "loss": 5.8281, "step": 2950 }, { "epoch": 0.0055983335627095, "grad_norm": 4.3454718589782715, "learning_rate": 4.998333333333334e-05, "loss": 5.9889, "step": 3000 }, { "epoch": 0.0055983335627095, "eval_loss": 6.374537944793701, "eval_runtime": 232.2431, "eval_samples_per_second": 11.23, "eval_steps_per_second": 11.23, "eval_tts_loss": 6.28711945315583, "step": 3000 }, { "epoch": 0.005691639122087992, "grad_norm": 8.272449493408203, "learning_rate": 5.081666666666667e-05, "loss": 6.141, "step": 3050 }, { "epoch": 0.005784944681466483, "grad_norm": 5.653373718261719, "learning_rate": 5.1649999999999995e-05, "loss": 6.1385, "step": 3100 }, { "epoch": 0.005878250240844975, "grad_norm": 4.102393627166748, "learning_rate": 5.248333333333334e-05, "loss": 5.8571, "step": 3150 }, { "epoch": 0.005971555800223466, "grad_norm": 5.258531093597412, "learning_rate": 5.331666666666667e-05, "loss": 5.7932, "step": 3200 }, { "epoch": 0.006064861359601959, "grad_norm": 4.183366775512695, "learning_rate": 5.415e-05, "loss": 5.8642, "step": 3250 }, { "epoch": 0.00615816691898045, "grad_norm": 5.612656593322754, "learning_rate": 5.498333333333333e-05, "loss": 5.8893, "step": 3300 }, { "epoch": 0.006251472478358942, "grad_norm": 5.23427677154541, "learning_rate": 5.581666666666667e-05, "loss": 5.7405, "step": 3350 }, { "epoch": 0.0063447780377374335, "grad_norm": 5.158674240112305, "learning_rate": 5.665e-05, "loss": 5.8673, "step": 3400 }, { "epoch": 0.006438083597115925, "grad_norm": 5.116922378540039, "learning_rate": 5.748333333333333e-05, "loss": 5.7444, "step": 3450 }, { "epoch": 0.006531389156494417, "grad_norm": 5.790384292602539, "learning_rate": 5.831666666666668e-05, "loss": 5.9559, "step": 3500 }, { "epoch": 0.006624694715872908, "grad_norm": 4.801383972167969, "learning_rate": 5.915000000000001e-05, "loss": 6.1995, "step": 3550 }, { "epoch": 0.0067180002752514, "grad_norm": 3.692594528198242, "learning_rate": 5.998333333333334e-05, "loss": 5.733, "step": 3600 }, { "epoch": 0.006811305834629892, "grad_norm": 4.112286567687988, "learning_rate": 6.081666666666666e-05, "loss": 6.0937, "step": 3650 }, { "epoch": 0.006904611394008384, "grad_norm": 4.726688385009766, "learning_rate": 6.165000000000001e-05, "loss": 5.8487, "step": 3700 }, { "epoch": 0.006997916953386875, "grad_norm": 3.719205856323242, "learning_rate": 6.248333333333334e-05, "loss": 5.7952, "step": 3750 }, { "epoch": 0.007091222512765367, "grad_norm": 4.780358791351318, "learning_rate": 6.331666666666667e-05, "loss": 5.9006, "step": 3800 }, { "epoch": 0.0071845280721438585, "grad_norm": 5.655118942260742, "learning_rate": 6.415e-05, "loss": 5.7174, "step": 3850 }, { "epoch": 0.00727783363152235, "grad_norm": 4.822659015655518, "learning_rate": 6.498333333333335e-05, "loss": 5.9248, "step": 3900 }, { "epoch": 0.007371139190900842, "grad_norm": 5.104306697845459, "learning_rate": 6.581666666666668e-05, "loss": 5.8789, "step": 3950 }, { "epoch": 0.007464444750279333, "grad_norm": 4.1900105476379395, "learning_rate": 6.665000000000001e-05, "loss": 5.8902, "step": 4000 }, { "epoch": 0.007557750309657825, "grad_norm": 4.686253547668457, "learning_rate": 6.748333333333334e-05, "loss": 5.6845, "step": 4050 }, { "epoch": 0.007651055869036317, "grad_norm": 4.2270188331604, "learning_rate": 6.831666666666667e-05, "loss": 5.9528, "step": 4100 }, { "epoch": 0.007744361428414809, "grad_norm": 4.093949794769287, "learning_rate": 6.915e-05, "loss": 5.8561, "step": 4150 }, { "epoch": 0.0078376669877933, "grad_norm": 4.820592403411865, "learning_rate": 6.998333333333333e-05, "loss": 5.6467, "step": 4200 }, { "epoch": 0.007930972547171792, "grad_norm": 4.777779579162598, "learning_rate": 7.081666666666668e-05, "loss": 5.8914, "step": 4250 }, { "epoch": 0.008024278106550283, "grad_norm": 2.8841679096221924, "learning_rate": 7.165000000000001e-05, "loss": 5.7689, "step": 4300 }, { "epoch": 0.008117583665928775, "grad_norm": 3.7897868156433105, "learning_rate": 7.248333333333334e-05, "loss": 5.6131, "step": 4350 }, { "epoch": 0.008210889225307267, "grad_norm": 3.63638973236084, "learning_rate": 7.331666666666667e-05, "loss": 5.6802, "step": 4400 }, { "epoch": 0.008304194784685758, "grad_norm": 4.507631778717041, "learning_rate": 7.415000000000001e-05, "loss": 5.9008, "step": 4450 }, { "epoch": 0.00839750034406425, "grad_norm": 4.462691783905029, "learning_rate": 7.498333333333334e-05, "loss": 5.7764, "step": 4500 }, { "epoch": 0.008490805903442741, "grad_norm": 5.587769508361816, "learning_rate": 7.581666666666668e-05, "loss": 5.6366, "step": 4550 }, { "epoch": 0.008584111462821233, "grad_norm": 3.457714319229126, "learning_rate": 7.664999999999999e-05, "loss": 5.7527, "step": 4600 }, { "epoch": 0.008677417022199724, "grad_norm": 4.030301094055176, "learning_rate": 7.748333333333334e-05, "loss": 5.8454, "step": 4650 }, { "epoch": 0.008770722581578216, "grad_norm": 3.3396108150482178, "learning_rate": 7.831666666666667e-05, "loss": 5.9134, "step": 4700 }, { "epoch": 0.00886402814095671, "grad_norm": 3.4558310508728027, "learning_rate": 7.915e-05, "loss": 5.7332, "step": 4750 }, { "epoch": 0.008957333700335201, "grad_norm": 4.230519771575928, "learning_rate": 7.998333333333333e-05, "loss": 5.86, "step": 4800 }, { "epoch": 0.009050639259713692, "grad_norm": 3.126082181930542, "learning_rate": 8.081666666666667e-05, "loss": 5.7412, "step": 4850 }, { "epoch": 0.009143944819092184, "grad_norm": 4.2145609855651855, "learning_rate": 8.165e-05, "loss": 5.9317, "step": 4900 }, { "epoch": 0.009237250378470676, "grad_norm": 3.0938162803649902, "learning_rate": 8.248333333333334e-05, "loss": 5.6891, "step": 4950 }, { "epoch": 0.009330555937849167, "grad_norm": 4.230195045471191, "learning_rate": 8.331666666666668e-05, "loss": 5.7216, "step": 5000 }, { "epoch": 0.009423861497227659, "grad_norm": 3.7208573818206787, "learning_rate": 8.415000000000001e-05, "loss": 5.7874, "step": 5050 }, { "epoch": 0.00951716705660615, "grad_norm": 3.4725289344787598, "learning_rate": 8.498333333333334e-05, "loss": 5.8627, "step": 5100 }, { "epoch": 0.009610472615984642, "grad_norm": 3.3882782459259033, "learning_rate": 8.581666666666666e-05, "loss": 5.767, "step": 5150 }, { "epoch": 0.009703778175363133, "grad_norm": 3.7585625648498535, "learning_rate": 8.665e-05, "loss": 5.6776, "step": 5200 }, { "epoch": 0.009797083734741625, "grad_norm": 4.944695949554443, "learning_rate": 8.748333333333334e-05, "loss": 5.667, "step": 5250 }, { "epoch": 0.009890389294120117, "grad_norm": 3.8249406814575195, "learning_rate": 8.831666666666667e-05, "loss": 5.7428, "step": 5300 }, { "epoch": 0.009983694853498608, "grad_norm": 4.215453147888184, "learning_rate": 8.915e-05, "loss": 5.89, "step": 5350 }, { "epoch": 0.0100770004128771, "grad_norm": 3.0438342094421387, "learning_rate": 8.998333333333334e-05, "loss": 5.639, "step": 5400 }, { "epoch": 0.010170305972255591, "grad_norm": 4.088777542114258, "learning_rate": 9.081666666666667e-05, "loss": 5.5986, "step": 5450 }, { "epoch": 0.010263611531634083, "grad_norm": 3.3326492309570312, "learning_rate": 9.165e-05, "loss": 5.5374, "step": 5500 }, { "epoch": 0.010356917091012574, "grad_norm": 2.9593100547790527, "learning_rate": 9.248333333333334e-05, "loss": 5.4346, "step": 5550 }, { "epoch": 0.010450222650391068, "grad_norm": 3.170647144317627, "learning_rate": 9.331666666666668e-05, "loss": 5.8167, "step": 5600 }, { "epoch": 0.01054352820976956, "grad_norm": 2.9037222862243652, "learning_rate": 9.415e-05, "loss": 5.5978, "step": 5650 }, { "epoch": 0.01063683376914805, "grad_norm": 3.341003894805908, "learning_rate": 9.498333333333333e-05, "loss": 5.6149, "step": 5700 }, { "epoch": 0.010730139328526542, "grad_norm": 2.9810729026794434, "learning_rate": 9.581666666666667e-05, "loss": 5.7028, "step": 5750 }, { "epoch": 0.010823444887905034, "grad_norm": 3.4222352504730225, "learning_rate": 9.665e-05, "loss": 5.5117, "step": 5800 }, { "epoch": 0.010916750447283526, "grad_norm": 3.1818320751190186, "learning_rate": 9.748333333333334e-05, "loss": 5.7761, "step": 5850 }, { "epoch": 0.011010056006662017, "grad_norm": 3.1308321952819824, "learning_rate": 9.831666666666667e-05, "loss": 5.6791, "step": 5900 }, { "epoch": 0.011103361566040509, "grad_norm": 3.701707363128662, "learning_rate": 9.915000000000001e-05, "loss": 5.5769, "step": 5950 }, { "epoch": 0.011196667125419, "grad_norm": 2.931779384613037, "learning_rate": 9.998333333333334e-05, "loss": 5.4739, "step": 6000 }, { "epoch": 0.011196667125419, "eval_loss": 6.0997633934021, "eval_runtime": 229.8318, "eval_samples_per_second": 11.347, "eval_steps_per_second": 11.347, "eval_tts_loss": 6.378738236305035, "step": 6000 }, { "epoch": 0.011289972684797492, "grad_norm": 3.05470871925354, "learning_rate": 0.00010081666666666667, "loss": 5.5917, "step": 6050 }, { "epoch": 0.011383278244175983, "grad_norm": 3.2815566062927246, "learning_rate": 0.00010165, "loss": 5.4096, "step": 6100 }, { "epoch": 0.011476583803554475, "grad_norm": 2.6623780727386475, "learning_rate": 0.00010248333333333334, "loss": 5.4688, "step": 6150 }, { "epoch": 0.011569889362932967, "grad_norm": 3.355912923812866, "learning_rate": 0.00010331666666666667, "loss": 5.6187, "step": 6200 }, { "epoch": 0.011663194922311458, "grad_norm": 2.9381911754608154, "learning_rate": 0.00010415000000000001, "loss": 5.8301, "step": 6250 }, { "epoch": 0.01175650048168995, "grad_norm": 2.482517719268799, "learning_rate": 0.00010498333333333334, "loss": 5.5721, "step": 6300 }, { "epoch": 0.011849806041068441, "grad_norm": 3.363173246383667, "learning_rate": 0.00010581666666666667, "loss": 5.8946, "step": 6350 }, { "epoch": 0.011943111600446933, "grad_norm": 3.2402546405792236, "learning_rate": 0.00010665, "loss": 5.6961, "step": 6400 }, { "epoch": 0.012036417159825426, "grad_norm": 3.1428112983703613, "learning_rate": 0.00010748333333333333, "loss": 5.8315, "step": 6450 }, { "epoch": 0.012129722719203918, "grad_norm": 2.3944103717803955, "learning_rate": 0.00010831666666666667, "loss": 5.6921, "step": 6500 }, { "epoch": 0.01222302827858241, "grad_norm": 2.7807068824768066, "learning_rate": 0.00010915, "loss": 5.563, "step": 6550 }, { "epoch": 0.0123163338379609, "grad_norm": 2.692389965057373, "learning_rate": 0.00010998333333333335, "loss": 5.4729, "step": 6600 }, { "epoch": 0.012409639397339392, "grad_norm": 1.9794119596481323, "learning_rate": 0.00011081666666666669, "loss": 5.7415, "step": 6650 }, { "epoch": 0.012502944956717884, "grad_norm": 1.812415599822998, "learning_rate": 0.00011165000000000002, "loss": 5.4328, "step": 6700 }, { "epoch": 0.012596250516096376, "grad_norm": 3.0957512855529785, "learning_rate": 0.00011248333333333333, "loss": 5.3841, "step": 6750 }, { "epoch": 0.012689556075474867, "grad_norm": 3.101076126098633, "learning_rate": 0.00011331666666666667, "loss": 5.8037, "step": 6800 }, { "epoch": 0.012782861634853359, "grad_norm": 3.04792857170105, "learning_rate": 0.00011415, "loss": 5.7408, "step": 6850 }, { "epoch": 0.01287616719423185, "grad_norm": 3.4043712615966797, "learning_rate": 0.00011498333333333333, "loss": 5.6041, "step": 6900 }, { "epoch": 0.012969472753610342, "grad_norm": 1.666723608970642, "learning_rate": 0.00011581666666666666, "loss": 5.6662, "step": 6950 }, { "epoch": 0.013062778312988833, "grad_norm": 2.3120384216308594, "learning_rate": 0.00011665000000000002, "loss": 5.8835, "step": 7000 }, { "epoch": 0.013156083872367325, "grad_norm": 4.411382675170898, "learning_rate": 0.00011748333333333335, "loss": 5.7121, "step": 7050 }, { "epoch": 0.013249389431745816, "grad_norm": 2.192537546157837, "learning_rate": 0.00011831666666666668, "loss": 5.4821, "step": 7100 }, { "epoch": 0.013342694991124308, "grad_norm": 2.196122646331787, "learning_rate": 0.00011915000000000001, "loss": 5.7267, "step": 7150 }, { "epoch": 0.0134360005505028, "grad_norm": 2.2460808753967285, "learning_rate": 0.00011998333333333334, "loss": 5.5123, "step": 7200 }, { "epoch": 0.013529306109881291, "grad_norm": 2.052151679992676, "learning_rate": 0.00012081666666666667, "loss": 5.8528, "step": 7250 }, { "epoch": 0.013622611669259784, "grad_norm": 2.369872808456421, "learning_rate": 0.00012165, "loss": 5.5378, "step": 7300 }, { "epoch": 0.013715917228638276, "grad_norm": 2.383234977722168, "learning_rate": 0.00012248333333333335, "loss": 5.7572, "step": 7350 }, { "epoch": 0.013809222788016768, "grad_norm": 2.3634586334228516, "learning_rate": 0.0001233166666666667, "loss": 5.3572, "step": 7400 }, { "epoch": 0.01390252834739526, "grad_norm": 2.631533622741699, "learning_rate": 0.00012415, "loss": 5.4715, "step": 7450 }, { "epoch": 0.01399583390677375, "grad_norm": 1.962390661239624, "learning_rate": 0.00012498333333333335, "loss": 5.7204, "step": 7500 }, { "epoch": 0.014089139466152242, "grad_norm": 2.6363272666931152, "learning_rate": 0.00012581666666666667, "loss": 5.6266, "step": 7550 }, { "epoch": 0.014182445025530734, "grad_norm": 2.3159897327423096, "learning_rate": 0.00012665, "loss": 5.6025, "step": 7600 }, { "epoch": 0.014275750584909225, "grad_norm": 2.558704137802124, "learning_rate": 0.00012748333333333333, "loss": 5.622, "step": 7650 }, { "epoch": 0.014369056144287717, "grad_norm": 2.3983957767486572, "learning_rate": 0.00012831666666666665, "loss": 5.677, "step": 7700 }, { "epoch": 0.014462361703666209, "grad_norm": 2.4474925994873047, "learning_rate": 0.00012915000000000002, "loss": 5.735, "step": 7750 }, { "epoch": 0.0145556672630447, "grad_norm": 1.7076762914657593, "learning_rate": 0.00012998333333333334, "loss": 5.6703, "step": 7800 }, { "epoch": 0.014648972822423192, "grad_norm": 2.520663261413574, "learning_rate": 0.00013081666666666668, "loss": 5.6445, "step": 7850 }, { "epoch": 0.014742278381801683, "grad_norm": 1.7655445337295532, "learning_rate": 0.00013165, "loss": 5.5739, "step": 7900 }, { "epoch": 0.014835583941180175, "grad_norm": 2.880577564239502, "learning_rate": 0.00013248333333333335, "loss": 5.5834, "step": 7950 }, { "epoch": 0.014928889500558666, "grad_norm": 2.0654830932617188, "learning_rate": 0.00013331666666666666, "loss": 5.5844, "step": 8000 }, { "epoch": 0.015022195059937158, "grad_norm": 1.5936386585235596, "learning_rate": 0.00013415, "loss": 5.4989, "step": 8050 }, { "epoch": 0.01511550061931565, "grad_norm": 2.4710347652435303, "learning_rate": 0.00013498333333333335, "loss": 5.5193, "step": 8100 }, { "epoch": 0.015208806178694143, "grad_norm": 1.680249571800232, "learning_rate": 0.00013581666666666667, "loss": 5.4661, "step": 8150 }, { "epoch": 0.015302111738072634, "grad_norm": 1.9727201461791992, "learning_rate": 0.00013665000000000001, "loss": 5.7409, "step": 8200 }, { "epoch": 0.015395417297451126, "grad_norm": 2.1253304481506348, "learning_rate": 0.00013748333333333333, "loss": 5.5681, "step": 8250 }, { "epoch": 0.015488722856829618, "grad_norm": 2.3671233654022217, "learning_rate": 0.00013831666666666668, "loss": 5.5692, "step": 8300 }, { "epoch": 0.01558202841620811, "grad_norm": 2.448765516281128, "learning_rate": 0.00013915, "loss": 5.8129, "step": 8350 }, { "epoch": 0.0156753339755866, "grad_norm": 2.6392576694488525, "learning_rate": 0.00013998333333333334, "loss": 5.6357, "step": 8400 }, { "epoch": 0.01576863953496509, "grad_norm": 2.3633005619049072, "learning_rate": 0.00014081666666666666, "loss": 5.6419, "step": 8450 }, { "epoch": 0.015861945094343584, "grad_norm": 2.5643057823181152, "learning_rate": 0.00014165000000000003, "loss": 5.6119, "step": 8500 }, { "epoch": 0.015955250653722074, "grad_norm": 2.252556324005127, "learning_rate": 0.00014248333333333335, "loss": 5.5832, "step": 8550 }, { "epoch": 0.016048556213100567, "grad_norm": 2.5928521156311035, "learning_rate": 0.0001433166666666667, "loss": 5.6567, "step": 8600 }, { "epoch": 0.01614186177247906, "grad_norm": 2.565758228302002, "learning_rate": 0.00014415, "loss": 5.5772, "step": 8650 }, { "epoch": 0.01623516733185755, "grad_norm": 1.9352021217346191, "learning_rate": 0.00014498333333333332, "loss": 5.6935, "step": 8700 }, { "epoch": 0.016328472891236043, "grad_norm": 1.7188228368759155, "learning_rate": 0.00014581666666666667, "loss": 5.5983, "step": 8750 }, { "epoch": 0.016421778450614533, "grad_norm": 1.5790917873382568, "learning_rate": 0.00014665, "loss": 5.5048, "step": 8800 }, { "epoch": 0.016515084009993027, "grad_norm": 2.046977996826172, "learning_rate": 0.00014748333333333336, "loss": 5.4909, "step": 8850 }, { "epoch": 0.016608389569371516, "grad_norm": 2.3554604053497314, "learning_rate": 0.00014831666666666668, "loss": 5.5094, "step": 8900 }, { "epoch": 0.01670169512875001, "grad_norm": 2.019984245300293, "learning_rate": 0.00014915000000000002, "loss": 5.5235, "step": 8950 }, { "epoch": 0.0167950006881285, "grad_norm": 2.2459378242492676, "learning_rate": 0.00014998333333333334, "loss": 5.7603, "step": 9000 }, { "epoch": 0.0167950006881285, "eval_loss": 5.949121475219727, "eval_runtime": 232.0628, "eval_samples_per_second": 11.238, "eval_steps_per_second": 11.238, "eval_tts_loss": 6.626402078707777, "step": 9000 }, { "epoch": 0.016888306247506993, "grad_norm": 1.576752781867981, "learning_rate": 0.00015081666666666668, "loss": 5.425, "step": 9050 }, { "epoch": 0.016981611806885483, "grad_norm": 1.950769066810608, "learning_rate": 0.00015165, "loss": 5.5398, "step": 9100 }, { "epoch": 0.017074917366263976, "grad_norm": 1.870368242263794, "learning_rate": 0.00015248333333333334, "loss": 5.5754, "step": 9150 }, { "epoch": 0.017168222925642466, "grad_norm": 1.75543212890625, "learning_rate": 0.00015331666666666666, "loss": 5.5328, "step": 9200 }, { "epoch": 0.01726152848502096, "grad_norm": 1.626207947731018, "learning_rate": 0.00015415, "loss": 5.6149, "step": 9250 }, { "epoch": 0.01735483404439945, "grad_norm": 2.190242052078247, "learning_rate": 0.00015498333333333335, "loss": 5.4885, "step": 9300 }, { "epoch": 0.017448139603777942, "grad_norm": 1.4450881481170654, "learning_rate": 0.00015581666666666667, "loss": 5.4476, "step": 9350 }, { "epoch": 0.017541445163156432, "grad_norm": 1.9197875261306763, "learning_rate": 0.00015665, "loss": 5.611, "step": 9400 }, { "epoch": 0.017634750722534925, "grad_norm": 2.0102715492248535, "learning_rate": 0.00015748333333333333, "loss": 5.3812, "step": 9450 }, { "epoch": 0.01772805628191342, "grad_norm": 1.8926249742507935, "learning_rate": 0.00015831666666666667, "loss": 5.4527, "step": 9500 }, { "epoch": 0.01782136184129191, "grad_norm": 2.243831157684326, "learning_rate": 0.00015915, "loss": 5.4914, "step": 9550 }, { "epoch": 0.017914667400670402, "grad_norm": 1.8967549800872803, "learning_rate": 0.00015998333333333336, "loss": 5.5251, "step": 9600 }, { "epoch": 0.01800797296004889, "grad_norm": 1.785871982574463, "learning_rate": 0.00016081666666666668, "loss": 5.3667, "step": 9650 }, { "epoch": 0.018101278519427385, "grad_norm": 1.634459376335144, "learning_rate": 0.00016165000000000003, "loss": 5.7341, "step": 9700 }, { "epoch": 0.018194584078805875, "grad_norm": 2.1786787509918213, "learning_rate": 0.00016248333333333334, "loss": 5.4932, "step": 9750 }, { "epoch": 0.018287889638184368, "grad_norm": 2.0094616413116455, "learning_rate": 0.00016331666666666666, "loss": 5.5438, "step": 9800 }, { "epoch": 0.018381195197562858, "grad_norm": 1.5394549369812012, "learning_rate": 0.00016415, "loss": 5.3004, "step": 9850 }, { "epoch": 0.01847450075694135, "grad_norm": 2.034067392349243, "learning_rate": 0.00016498333333333332, "loss": 5.6939, "step": 9900 }, { "epoch": 0.01856780631631984, "grad_norm": 2.557887554168701, "learning_rate": 0.00016581666666666667, "loss": 5.4535, "step": 9950 }, { "epoch": 0.018661111875698334, "grad_norm": 1.7101596593856812, "learning_rate": 0.00016665, "loss": 5.5991, "step": 10000 }, { "epoch": 0.018754417435076824, "grad_norm": 1.1628295183181763, "learning_rate": 0.00016748333333333336, "loss": 5.5027, "step": 10050 }, { "epoch": 0.018847722994455317, "grad_norm": 1.6788240671157837, "learning_rate": 0.00016831666666666667, "loss": 5.3345, "step": 10100 }, { "epoch": 0.018941028553833807, "grad_norm": 2.166861057281494, "learning_rate": 0.00016915000000000002, "loss": 5.6807, "step": 10150 }, { "epoch": 0.0190343341132123, "grad_norm": 1.7033262252807617, "learning_rate": 0.00016998333333333334, "loss": 5.6921, "step": 10200 }, { "epoch": 0.01912763967259079, "grad_norm": 1.5504415035247803, "learning_rate": 0.00017081666666666668, "loss": 5.5782, "step": 10250 }, { "epoch": 0.019220945231969284, "grad_norm": 2.4242970943450928, "learning_rate": 0.00017165, "loss": 5.4636, "step": 10300 }, { "epoch": 0.019314250791347777, "grad_norm": 1.863472819328308, "learning_rate": 0.00017248333333333334, "loss": 5.4955, "step": 10350 }, { "epoch": 0.019407556350726267, "grad_norm": 1.856552004814148, "learning_rate": 0.0001733166666666667, "loss": 5.526, "step": 10400 }, { "epoch": 0.01950086191010476, "grad_norm": 2.124068260192871, "learning_rate": 0.00017415, "loss": 5.5683, "step": 10450 }, { "epoch": 0.01959416746948325, "grad_norm": 1.7942883968353271, "learning_rate": 0.00017498333333333335, "loss": 5.3387, "step": 10500 }, { "epoch": 0.019687473028861743, "grad_norm": 1.7919622659683228, "learning_rate": 0.00017581666666666667, "loss": 5.4876, "step": 10550 }, { "epoch": 0.019780778588240233, "grad_norm": 2.1417160034179688, "learning_rate": 0.00017665, "loss": 5.3508, "step": 10600 }, { "epoch": 0.019874084147618726, "grad_norm": 1.6931318044662476, "learning_rate": 0.00017748333333333333, "loss": 5.6171, "step": 10650 }, { "epoch": 0.019967389706997216, "grad_norm": 1.4930100440979004, "learning_rate": 0.00017831666666666667, "loss": 5.5452, "step": 10700 }, { "epoch": 0.02006069526637571, "grad_norm": 1.6434742212295532, "learning_rate": 0.00017915000000000002, "loss": 5.4924, "step": 10750 }, { "epoch": 0.0201540008257542, "grad_norm": 1.96444571018219, "learning_rate": 0.00017998333333333334, "loss": 5.3689, "step": 10800 }, { "epoch": 0.020247306385132693, "grad_norm": 2.2349231243133545, "learning_rate": 0.00018081666666666668, "loss": 5.3768, "step": 10850 }, { "epoch": 0.020340611944511183, "grad_norm": 1.5882576704025269, "learning_rate": 0.00018165, "loss": 5.747, "step": 10900 }, { "epoch": 0.020433917503889676, "grad_norm": 2.4228999614715576, "learning_rate": 0.00018248333333333334, "loss": 5.7638, "step": 10950 }, { "epoch": 0.020527223063268166, "grad_norm": 2.138207197189331, "learning_rate": 0.00018331666666666666, "loss": 5.7262, "step": 11000 }, { "epoch": 0.02062052862264666, "grad_norm": 1.9708890914916992, "learning_rate": 0.00018415, "loss": 5.5162, "step": 11050 }, { "epoch": 0.02071383418202515, "grad_norm": 1.5397372245788574, "learning_rate": 0.00018498333333333335, "loss": 5.5007, "step": 11100 }, { "epoch": 0.020807139741403642, "grad_norm": 1.579482913017273, "learning_rate": 0.0001858166666666667, "loss": 5.4921, "step": 11150 }, { "epoch": 0.020900445300782135, "grad_norm": 1.2825592756271362, "learning_rate": 0.00018665, "loss": 5.5486, "step": 11200 }, { "epoch": 0.020993750860160625, "grad_norm": 2.018866777420044, "learning_rate": 0.00018748333333333335, "loss": 5.4298, "step": 11250 }, { "epoch": 0.02108705641953912, "grad_norm": 1.8341054916381836, "learning_rate": 0.00018831666666666667, "loss": 5.5407, "step": 11300 }, { "epoch": 0.02118036197891761, "grad_norm": 1.6199865341186523, "learning_rate": 0.00018915000000000002, "loss": 5.3315, "step": 11350 }, { "epoch": 0.0212736675382961, "grad_norm": 1.5176050662994385, "learning_rate": 0.00018998333333333333, "loss": 5.4755, "step": 11400 }, { "epoch": 0.02136697309767459, "grad_norm": 1.8712049722671509, "learning_rate": 0.00019081666666666668, "loss": 5.4285, "step": 11450 }, { "epoch": 0.021460278657053085, "grad_norm": 1.531267762184143, "learning_rate": 0.00019165000000000002, "loss": 5.6335, "step": 11500 }, { "epoch": 0.021553584216431575, "grad_norm": 1.6521434783935547, "learning_rate": 0.00019248333333333334, "loss": 5.3548, "step": 11550 }, { "epoch": 0.021646889775810068, "grad_norm": 1.4107913970947266, "learning_rate": 0.00019331666666666669, "loss": 5.446, "step": 11600 }, { "epoch": 0.021740195335188558, "grad_norm": 1.8727481365203857, "learning_rate": 0.00019415, "loss": 5.5346, "step": 11650 }, { "epoch": 0.02183350089456705, "grad_norm": 2.040088653564453, "learning_rate": 0.00019498333333333335, "loss": 5.7011, "step": 11700 }, { "epoch": 0.02192680645394554, "grad_norm": 1.4047014713287354, "learning_rate": 0.00019581666666666666, "loss": 5.3325, "step": 11750 }, { "epoch": 0.022020112013324034, "grad_norm": 1.2155386209487915, "learning_rate": 0.00019665, "loss": 5.5547, "step": 11800 }, { "epoch": 0.022113417572702524, "grad_norm": 1.2685003280639648, "learning_rate": 0.00019748333333333335, "loss": 5.6444, "step": 11850 }, { "epoch": 0.022206723132081017, "grad_norm": 1.1052175760269165, "learning_rate": 0.00019831666666666667, "loss": 5.4218, "step": 11900 }, { "epoch": 0.022300028691459507, "grad_norm": 1.6609026193618774, "learning_rate": 0.00019915000000000002, "loss": 5.5406, "step": 11950 }, { "epoch": 0.022393334250838, "grad_norm": 1.7019258737564087, "learning_rate": 0.00019998333333333333, "loss": 5.4288, "step": 12000 }, { "epoch": 0.022393334250838, "eval_loss": 5.846170425415039, "eval_runtime": 230.7006, "eval_samples_per_second": 11.305, "eval_steps_per_second": 11.305, "eval_tts_loss": 6.596176531521512, "step": 12000 }, { "epoch": 0.022486639810216494, "grad_norm": 1.391849160194397, "learning_rate": 0.00019999999991525316, "loss": 5.438, "step": 12050 }, { "epoch": 0.022579945369594984, "grad_norm": 1.9549068212509155, "learning_rate": 0.0001999999996540592, "loss": 5.6524, "step": 12100 }, { "epoch": 0.022673250928973477, "grad_norm": 1.765764832496643, "learning_rate": 0.00019999999921638283, "loss": 5.529, "step": 12150 }, { "epoch": 0.022766556488351967, "grad_norm": 1.92975652217865, "learning_rate": 0.00019999999860222407, "loss": 5.1634, "step": 12200 }, { "epoch": 0.02285986204773046, "grad_norm": 1.627753734588623, "learning_rate": 0.00019999999781158292, "loss": 5.6305, "step": 12250 }, { "epoch": 0.02295316760710895, "grad_norm": 1.6781692504882812, "learning_rate": 0.00019999999684445935, "loss": 5.5936, "step": 12300 }, { "epoch": 0.023046473166487443, "grad_norm": 1.200257658958435, "learning_rate": 0.00019999999570085336, "loss": 5.4897, "step": 12350 }, { "epoch": 0.023139778725865933, "grad_norm": 1.7044615745544434, "learning_rate": 0.00019999999438076503, "loss": 5.2909, "step": 12400 }, { "epoch": 0.023233084285244426, "grad_norm": 1.4790157079696655, "learning_rate": 0.00019999999288419424, "loss": 5.3771, "step": 12450 }, { "epoch": 0.023326389844622916, "grad_norm": 1.5862489938735962, "learning_rate": 0.0001999999912111411, "loss": 5.2316, "step": 12500 }, { "epoch": 0.02341969540400141, "grad_norm": 1.9885330200195312, "learning_rate": 0.00019999998936160555, "loss": 5.3982, "step": 12550 }, { "epoch": 0.0235130009633799, "grad_norm": 1.4161193370819092, "learning_rate": 0.00019999998733558766, "loss": 5.4459, "step": 12600 }, { "epoch": 0.023606306522758393, "grad_norm": 1.5754268169403076, "learning_rate": 0.00019999998513308732, "loss": 5.3697, "step": 12650 }, { "epoch": 0.023699612082136882, "grad_norm": 1.0978269577026367, "learning_rate": 0.00019999998275410462, "loss": 5.4439, "step": 12700 }, { "epoch": 0.023792917641515376, "grad_norm": 1.4332829713821411, "learning_rate": 0.0001999999801986396, "loss": 5.2534, "step": 12750 }, { "epoch": 0.023886223200893866, "grad_norm": 1.157253623008728, "learning_rate": 0.00019999997746669216, "loss": 5.4755, "step": 12800 }, { "epoch": 0.02397952876027236, "grad_norm": 1.577779769897461, "learning_rate": 0.00019999997455826237, "loss": 5.4148, "step": 12850 }, { "epoch": 0.024072834319650852, "grad_norm": 1.4218207597732544, "learning_rate": 0.00019999997147335022, "loss": 5.4531, "step": 12900 }, { "epoch": 0.024166139879029342, "grad_norm": 2.003883123397827, "learning_rate": 0.0001999999682119557, "loss": 5.6188, "step": 12950 }, { "epoch": 0.024259445438407835, "grad_norm": 2.01855731010437, "learning_rate": 0.00019999996477407884, "loss": 5.5149, "step": 13000 }, { "epoch": 0.024352750997786325, "grad_norm": 1.6317418813705444, "learning_rate": 0.00019999996115971965, "loss": 5.3213, "step": 13050 }, { "epoch": 0.02444605655716482, "grad_norm": 1.6599922180175781, "learning_rate": 0.0001999999573688781, "loss": 5.4164, "step": 13100 }, { "epoch": 0.02453936211654331, "grad_norm": 1.4030259847640991, "learning_rate": 0.00019999995340155423, "loss": 5.254, "step": 13150 }, { "epoch": 0.0246326676759218, "grad_norm": 1.507149338722229, "learning_rate": 0.00019999994925774806, "loss": 5.2633, "step": 13200 }, { "epoch": 0.02472597323530029, "grad_norm": 1.6153885126113892, "learning_rate": 0.00019999994493745957, "loss": 5.4768, "step": 13250 }, { "epoch": 0.024819278794678785, "grad_norm": 1.2192950248718262, "learning_rate": 0.00019999994044068875, "loss": 5.465, "step": 13300 }, { "epoch": 0.024912584354057275, "grad_norm": 1.3811638355255127, "learning_rate": 0.00019999993576743565, "loss": 5.4459, "step": 13350 }, { "epoch": 0.025005889913435768, "grad_norm": 1.7365858554840088, "learning_rate": 0.00019999993091770026, "loss": 5.4616, "step": 13400 }, { "epoch": 0.025099195472814258, "grad_norm": 1.4087482690811157, "learning_rate": 0.00019999992589148256, "loss": 5.4608, "step": 13450 }, { "epoch": 0.02519250103219275, "grad_norm": 1.7531602382659912, "learning_rate": 0.00019999992068878264, "loss": 5.3973, "step": 13500 }, { "epoch": 0.02528580659157124, "grad_norm": 1.653184175491333, "learning_rate": 0.0001999999153096004, "loss": 5.219, "step": 13550 }, { "epoch": 0.025379112150949734, "grad_norm": 1.3364681005477905, "learning_rate": 0.00019999990975393592, "loss": 5.4375, "step": 13600 }, { "epoch": 0.025472417710328224, "grad_norm": 1.3480881452560425, "learning_rate": 0.0001999999040217892, "loss": 5.4107, "step": 13650 }, { "epoch": 0.025565723269706717, "grad_norm": 1.8013604879379272, "learning_rate": 0.00019999989811316025, "loss": 5.4488, "step": 13700 }, { "epoch": 0.02565902882908521, "grad_norm": 1.5909210443496704, "learning_rate": 0.00019999989202804908, "loss": 5.6496, "step": 13750 }, { "epoch": 0.0257523343884637, "grad_norm": 1.897382140159607, "learning_rate": 0.0001999998857664557, "loss": 5.3464, "step": 13800 }, { "epoch": 0.025845639947842194, "grad_norm": 1.844641089439392, "learning_rate": 0.0001999998793283801, "loss": 5.536, "step": 13850 }, { "epoch": 0.025938945507220684, "grad_norm": 1.41611909866333, "learning_rate": 0.00019999987271382232, "loss": 5.5876, "step": 13900 }, { "epoch": 0.026032251066599177, "grad_norm": 1.5440382957458496, "learning_rate": 0.00019999986592278234, "loss": 5.2344, "step": 13950 }, { "epoch": 0.026125556625977667, "grad_norm": 1.3486498594284058, "learning_rate": 0.0001999998589552602, "loss": 5.4819, "step": 14000 }, { "epoch": 0.02621886218535616, "grad_norm": 1.6134167909622192, "learning_rate": 0.0001999998518112559, "loss": 5.6369, "step": 14050 }, { "epoch": 0.02631216774473465, "grad_norm": 1.950914740562439, "learning_rate": 0.0001999998444907695, "loss": 5.3324, "step": 14100 }, { "epoch": 0.026405473304113143, "grad_norm": 1.515940546989441, "learning_rate": 0.00019999983699380095, "loss": 5.5848, "step": 14150 }, { "epoch": 0.026498778863491633, "grad_norm": 1.8684641122817993, "learning_rate": 0.00019999982932035023, "loss": 5.4585, "step": 14200 }, { "epoch": 0.026592084422870126, "grad_norm": 1.9417451620101929, "learning_rate": 0.00019999982147041748, "loss": 5.3855, "step": 14250 }, { "epoch": 0.026685389982248616, "grad_norm": 1.5790482759475708, "learning_rate": 0.0001999998134440026, "loss": 5.4521, "step": 14300 }, { "epoch": 0.02677869554162711, "grad_norm": 1.9342408180236816, "learning_rate": 0.00019999980524110564, "loss": 5.3272, "step": 14350 }, { "epoch": 0.0268720011010056, "grad_norm": 1.617722988128662, "learning_rate": 0.00019999979686172664, "loss": 5.2174, "step": 14400 }, { "epoch": 0.026965306660384093, "grad_norm": 1.0585445165634155, "learning_rate": 0.00019999978830586557, "loss": 5.1622, "step": 14450 }, { "epoch": 0.027058612219762582, "grad_norm": 1.661190390586853, "learning_rate": 0.00019999977957352248, "loss": 5.1873, "step": 14500 }, { "epoch": 0.027151917779141076, "grad_norm": 1.5143160820007324, "learning_rate": 0.0001999997706646974, "loss": 5.507, "step": 14550 }, { "epoch": 0.02724522333851957, "grad_norm": 1.7400487661361694, "learning_rate": 0.0001999997615793903, "loss": 5.0888, "step": 14600 }, { "epoch": 0.02733852889789806, "grad_norm": 1.344720482826233, "learning_rate": 0.0001999997523176012, "loss": 5.2325, "step": 14650 }, { "epoch": 0.027431834457276552, "grad_norm": 1.390224575996399, "learning_rate": 0.00019999974287933015, "loss": 5.2955, "step": 14700 }, { "epoch": 0.027525140016655042, "grad_norm": 1.685595154762268, "learning_rate": 0.00019999973326457714, "loss": 5.3243, "step": 14750 }, { "epoch": 0.027618445576033535, "grad_norm": 1.574565052986145, "learning_rate": 0.0001999997234733422, "loss": 5.5154, "step": 14800 }, { "epoch": 0.027711751135412025, "grad_norm": 1.8520110845565796, "learning_rate": 0.0001999997135056254, "loss": 5.4337, "step": 14850 }, { "epoch": 0.02780505669479052, "grad_norm": 1.9176346063613892, "learning_rate": 0.0001999997033614266, "loss": 5.2195, "step": 14900 }, { "epoch": 0.027898362254169008, "grad_norm": 1.4428346157073975, "learning_rate": 0.000199999693040746, "loss": 5.1029, "step": 14950 }, { "epoch": 0.0279916678135475, "grad_norm": 1.408713459968567, "learning_rate": 0.0001999996825435835, "loss": 5.4174, "step": 15000 }, { "epoch": 0.0279916678135475, "eval_loss": 5.698703289031982, "eval_runtime": 230.7647, "eval_samples_per_second": 11.302, "eval_steps_per_second": 11.302, "eval_tts_loss": 6.7004290100964115, "step": 15000 }, { "epoch": 0.02808497337292599, "grad_norm": 1.4793498516082764, "learning_rate": 0.00019999967186993915, "loss": 5.5412, "step": 15050 }, { "epoch": 0.028178278932304485, "grad_norm": 1.2378696203231812, "learning_rate": 0.000199999661019813, "loss": 5.3518, "step": 15100 }, { "epoch": 0.028271584491682974, "grad_norm": 1.4232487678527832, "learning_rate": 0.000199999649993205, "loss": 5.4232, "step": 15150 }, { "epoch": 0.028364890051061468, "grad_norm": 1.6765247583389282, "learning_rate": 0.00019999963879011525, "loss": 5.5725, "step": 15200 }, { "epoch": 0.028458195610439958, "grad_norm": 1.599683403968811, "learning_rate": 0.00019999962741054373, "loss": 5.4653, "step": 15250 }, { "epoch": 0.02855150116981845, "grad_norm": 1.5458168983459473, "learning_rate": 0.00019999961585449047, "loss": 5.0618, "step": 15300 }, { "epoch": 0.02864480672919694, "grad_norm": 1.09707510471344, "learning_rate": 0.0001999996041219555, "loss": 5.5357, "step": 15350 }, { "epoch": 0.028738112288575434, "grad_norm": 2.0242269039154053, "learning_rate": 0.00019999959221293878, "loss": 5.3426, "step": 15400 }, { "epoch": 0.028831417847953927, "grad_norm": 1.6951383352279663, "learning_rate": 0.00019999958012744038, "loss": 5.4521, "step": 15450 }, { "epoch": 0.028924723407332417, "grad_norm": 1.8190480470657349, "learning_rate": 0.00019999956786546035, "loss": 5.2914, "step": 15500 }, { "epoch": 0.02901802896671091, "grad_norm": 1.7063592672348022, "learning_rate": 0.00019999955542699863, "loss": 5.2852, "step": 15550 }, { "epoch": 0.0291113345260894, "grad_norm": 1.7153364419937134, "learning_rate": 0.00019999954281205533, "loss": 5.0807, "step": 15600 }, { "epoch": 0.029204640085467894, "grad_norm": 1.677153468132019, "learning_rate": 0.0001999995300206304, "loss": 5.1755, "step": 15650 }, { "epoch": 0.029297945644846383, "grad_norm": 1.5905139446258545, "learning_rate": 0.00019999951705272393, "loss": 5.2782, "step": 15700 }, { "epoch": 0.029391251204224877, "grad_norm": 1.6873728036880493, "learning_rate": 0.0001999995039083359, "loss": 5.51, "step": 15750 }, { "epoch": 0.029484556763603367, "grad_norm": 1.4348547458648682, "learning_rate": 0.00019999949058746632, "loss": 5.2227, "step": 15800 }, { "epoch": 0.02957786232298186, "grad_norm": 1.599501132965088, "learning_rate": 0.00019999947709011524, "loss": 5.4851, "step": 15850 }, { "epoch": 0.02967116788236035, "grad_norm": 1.2430881261825562, "learning_rate": 0.0001999994634162827, "loss": 5.059, "step": 15900 }, { "epoch": 0.029764473441738843, "grad_norm": 1.2983192205429077, "learning_rate": 0.00019999944956596867, "loss": 5.3391, "step": 15950 }, { "epoch": 0.029857779001117333, "grad_norm": 1.8286632299423218, "learning_rate": 0.00019999943553917324, "loss": 5.068, "step": 16000 }, { "epoch": 0.029951084560495826, "grad_norm": 1.7763135433197021, "learning_rate": 0.0001999994213358964, "loss": 5.2494, "step": 16050 }, { "epoch": 0.030044390119874316, "grad_norm": 1.5806595087051392, "learning_rate": 0.00019999940695613816, "loss": 5.3536, "step": 16100 }, { "epoch": 0.03013769567925281, "grad_norm": 1.542901635169983, "learning_rate": 0.00019999939239989857, "loss": 5.2556, "step": 16150 }, { "epoch": 0.0302310012386313, "grad_norm": 1.390985369682312, "learning_rate": 0.00019999937766717764, "loss": 5.155, "step": 16200 }, { "epoch": 0.030324306798009792, "grad_norm": 1.0642412900924683, "learning_rate": 0.0001999993627579754, "loss": 5.3142, "step": 16250 }, { "epoch": 0.030417612357388286, "grad_norm": 1.5385115146636963, "learning_rate": 0.00019999934767229188, "loss": 5.2755, "step": 16300 }, { "epoch": 0.030510917916766776, "grad_norm": 1.719664216041565, "learning_rate": 0.00019999933241012711, "loss": 5.3522, "step": 16350 }, { "epoch": 0.03060422347614527, "grad_norm": 1.6435158252716064, "learning_rate": 0.00019999931697148112, "loss": 5.1949, "step": 16400 }, { "epoch": 0.03069752903552376, "grad_norm": 1.2330780029296875, "learning_rate": 0.00019999930135635395, "loss": 5.0197, "step": 16450 }, { "epoch": 0.030790834594902252, "grad_norm": 1.433413028717041, "learning_rate": 0.0001999992855647456, "loss": 5.4277, "step": 16500 }, { "epoch": 0.030884140154280742, "grad_norm": 1.8009040355682373, "learning_rate": 0.00019999926959665608, "loss": 5.0877, "step": 16550 }, { "epoch": 0.030977445713659235, "grad_norm": 1.1439430713653564, "learning_rate": 0.00019999925345208546, "loss": 5.3289, "step": 16600 }, { "epoch": 0.031070751273037725, "grad_norm": 1.3732225894927979, "learning_rate": 0.00019999923713103374, "loss": 5.1679, "step": 16650 }, { "epoch": 0.03116405683241622, "grad_norm": 1.722701072692871, "learning_rate": 0.000199999220633501, "loss": 5.3267, "step": 16700 }, { "epoch": 0.03125736239179471, "grad_norm": 1.6895197629928589, "learning_rate": 0.00019999920395948718, "loss": 5.1153, "step": 16750 }, { "epoch": 0.0313506679511732, "grad_norm": 1.7803030014038086, "learning_rate": 0.00019999918710899238, "loss": 5.2153, "step": 16800 }, { "epoch": 0.03144397351055169, "grad_norm": 1.187790036201477, "learning_rate": 0.00019999917008201663, "loss": 5.2711, "step": 16850 }, { "epoch": 0.03153727906993018, "grad_norm": 1.3633534908294678, "learning_rate": 0.00019999915287855993, "loss": 5.5233, "step": 16900 }, { "epoch": 0.03163058462930868, "grad_norm": 1.4957891702651978, "learning_rate": 0.00019999913549862233, "loss": 5.2524, "step": 16950 }, { "epoch": 0.03172389018868717, "grad_norm": 1.6286832094192505, "learning_rate": 0.00019999911794220382, "loss": 5.1533, "step": 17000 }, { "epoch": 0.03181719574806566, "grad_norm": 1.2257238626480103, "learning_rate": 0.00019999910020930447, "loss": 5.1066, "step": 17050 }, { "epoch": 0.03191050130744415, "grad_norm": 1.537248134613037, "learning_rate": 0.00019999908229992433, "loss": 5.1558, "step": 17100 }, { "epoch": 0.032003806866822644, "grad_norm": 1.7594759464263916, "learning_rate": 0.00019999906421406337, "loss": 5.2488, "step": 17150 }, { "epoch": 0.032097112426201134, "grad_norm": 2.0060057640075684, "learning_rate": 0.00019999904595172167, "loss": 5.1244, "step": 17200 }, { "epoch": 0.032190417985579624, "grad_norm": 1.306092381477356, "learning_rate": 0.00019999902751289926, "loss": 5.2285, "step": 17250 }, { "epoch": 0.03228372354495812, "grad_norm": 1.605584979057312, "learning_rate": 0.00019999900889759617, "loss": 5.148, "step": 17300 }, { "epoch": 0.03237702910433661, "grad_norm": 1.3685698509216309, "learning_rate": 0.0001999989901058124, "loss": 5.2305, "step": 17350 }, { "epoch": 0.0324703346637151, "grad_norm": 1.7700085639953613, "learning_rate": 0.00019999897113754801, "loss": 5.2642, "step": 17400 }, { "epoch": 0.03256364022309359, "grad_norm": 1.8209742307662964, "learning_rate": 0.00019999895199280306, "loss": 5.2634, "step": 17450 }, { "epoch": 0.03265694578247209, "grad_norm": 1.7181389331817627, "learning_rate": 0.00019999893267157753, "loss": 5.3703, "step": 17500 }, { "epoch": 0.03275025134185058, "grad_norm": 1.6699622869491577, "learning_rate": 0.00019999891317387149, "loss": 5.1961, "step": 17550 }, { "epoch": 0.032843556901229067, "grad_norm": 1.7012041807174683, "learning_rate": 0.00019999889349968494, "loss": 5.2565, "step": 17600 }, { "epoch": 0.032936862460607556, "grad_norm": 1.4097135066986084, "learning_rate": 0.00019999887364901796, "loss": 5.3606, "step": 17650 }, { "epoch": 0.03303016801998605, "grad_norm": 1.8546066284179688, "learning_rate": 0.00019999885362187057, "loss": 5.1426, "step": 17700 }, { "epoch": 0.03312347357936454, "grad_norm": 1.6862845420837402, "learning_rate": 0.00019999883341824282, "loss": 5.4663, "step": 17750 }, { "epoch": 0.03321677913874303, "grad_norm": 1.4715492725372314, "learning_rate": 0.00019999881303813468, "loss": 5.2308, "step": 17800 }, { "epoch": 0.03331008469812152, "grad_norm": 1.50525963306427, "learning_rate": 0.00019999879248154624, "loss": 5.2341, "step": 17850 }, { "epoch": 0.03340339025750002, "grad_norm": 1.6435831785202026, "learning_rate": 0.00019999877174847752, "loss": 5.2826, "step": 17900 }, { "epoch": 0.03349669581687851, "grad_norm": 1.4661362171173096, "learning_rate": 0.00019999875083892858, "loss": 5.2187, "step": 17950 }, { "epoch": 0.033590001376257, "grad_norm": 1.255449652671814, "learning_rate": 0.00019999872975289944, "loss": 5.1868, "step": 18000 }, { "epoch": 0.033590001376257, "eval_loss": 5.565659046173096, "eval_runtime": 233.225, "eval_samples_per_second": 11.182, "eval_steps_per_second": 11.182, "eval_tts_loss": 6.789093455502569, "step": 18000 }, { "epoch": 0.033683306935635496, "grad_norm": 1.6058536767959595, "learning_rate": 0.00019999870849039014, "loss": 5.0283, "step": 18050 }, { "epoch": 0.033776612495013986, "grad_norm": 1.931518793106079, "learning_rate": 0.0001999986870514007, "loss": 5.0369, "step": 18100 }, { "epoch": 0.033869918054392475, "grad_norm": 1.3181180953979492, "learning_rate": 0.00019999866543593116, "loss": 5.0891, "step": 18150 }, { "epoch": 0.033963223613770965, "grad_norm": 1.6189442873001099, "learning_rate": 0.0001999986436439816, "loss": 5.2425, "step": 18200 }, { "epoch": 0.03405652917314946, "grad_norm": 1.4186607599258423, "learning_rate": 0.000199998621675552, "loss": 4.9802, "step": 18250 }, { "epoch": 0.03414983473252795, "grad_norm": 1.727707028388977, "learning_rate": 0.00019999859953064245, "loss": 5.1903, "step": 18300 }, { "epoch": 0.03424314029190644, "grad_norm": 2.1892459392547607, "learning_rate": 0.00019999857720925297, "loss": 5.0998, "step": 18350 }, { "epoch": 0.03433644585128493, "grad_norm": 1.567331075668335, "learning_rate": 0.00019999855471138358, "loss": 5.241, "step": 18400 }, { "epoch": 0.03442975141066343, "grad_norm": 1.5945489406585693, "learning_rate": 0.00019999853203703436, "loss": 5.1888, "step": 18450 }, { "epoch": 0.03452305697004192, "grad_norm": 1.468598484992981, "learning_rate": 0.0001999985091862053, "loss": 5.2305, "step": 18500 }, { "epoch": 0.03461636252942041, "grad_norm": 1.801831603050232, "learning_rate": 0.00019999848615889648, "loss": 5.0736, "step": 18550 }, { "epoch": 0.0347096680887989, "grad_norm": 1.983370065689087, "learning_rate": 0.00019999846295510793, "loss": 5.1701, "step": 18600 }, { "epoch": 0.034802973648177395, "grad_norm": 1.4538954496383667, "learning_rate": 0.00019999843957483967, "loss": 5.1838, "step": 18650 }, { "epoch": 0.034896279207555884, "grad_norm": 2.007187843322754, "learning_rate": 0.0001999984160180918, "loss": 5.1847, "step": 18700 }, { "epoch": 0.034989584766934374, "grad_norm": 1.3527206182479858, "learning_rate": 0.00019999839228486427, "loss": 5.21, "step": 18750 }, { "epoch": 0.035082890326312864, "grad_norm": 0.93910813331604, "learning_rate": 0.0001999983683751572, "loss": 5.217, "step": 18800 }, { "epoch": 0.03517619588569136, "grad_norm": 1.9449275732040405, "learning_rate": 0.00019999834428897063, "loss": 5.2404, "step": 18850 }, { "epoch": 0.03526950144506985, "grad_norm": 1.5186761617660522, "learning_rate": 0.00019999832002630453, "loss": 5.1827, "step": 18900 }, { "epoch": 0.03536280700444834, "grad_norm": 1.8738884925842285, "learning_rate": 0.000199998295587159, "loss": 5.2015, "step": 18950 }, { "epoch": 0.03545611256382684, "grad_norm": 1.6167317628860474, "learning_rate": 0.0001999982709715341, "loss": 5.1333, "step": 19000 }, { "epoch": 0.03554941812320533, "grad_norm": 1.5436670780181885, "learning_rate": 0.00019999824617942985, "loss": 5.2181, "step": 19050 }, { "epoch": 0.03564272368258382, "grad_norm": 1.9840049743652344, "learning_rate": 0.00019999822121084625, "loss": 4.9849, "step": 19100 }, { "epoch": 0.03573602924196231, "grad_norm": 2.000702381134033, "learning_rate": 0.00019999819606578343, "loss": 5.0625, "step": 19150 }, { "epoch": 0.035829334801340804, "grad_norm": 1.5434222221374512, "learning_rate": 0.00019999817074424137, "loss": 5.0164, "step": 19200 }, { "epoch": 0.03592264036071929, "grad_norm": 1.492538332939148, "learning_rate": 0.0001999981452462201, "loss": 5.2329, "step": 19250 }, { "epoch": 0.03601594592009778, "grad_norm": 1.3936853408813477, "learning_rate": 0.00019999811957171975, "loss": 5.2539, "step": 19300 }, { "epoch": 0.03610925147947627, "grad_norm": 1.3118730783462524, "learning_rate": 0.00019999809372074027, "loss": 5.1528, "step": 19350 }, { "epoch": 0.03620255703885477, "grad_norm": 1.0975465774536133, "learning_rate": 0.00019999806769328178, "loss": 5.1746, "step": 19400 }, { "epoch": 0.03629586259823326, "grad_norm": 1.3111509084701538, "learning_rate": 0.0001999980414893443, "loss": 5.1834, "step": 19450 }, { "epoch": 0.03638916815761175, "grad_norm": 1.45402193069458, "learning_rate": 0.00019999801510892784, "loss": 5.2616, "step": 19500 }, { "epoch": 0.03648247371699024, "grad_norm": 1.5854758024215698, "learning_rate": 0.0001999979885520325, "loss": 5.1249, "step": 19550 }, { "epoch": 0.036575779276368736, "grad_norm": 1.6653951406478882, "learning_rate": 0.00019999796181865832, "loss": 5.1441, "step": 19600 }, { "epoch": 0.036669084835747226, "grad_norm": 1.6704020500183105, "learning_rate": 0.0001999979349088053, "loss": 5.1533, "step": 19650 }, { "epoch": 0.036762390395125716, "grad_norm": 1.5533548593521118, "learning_rate": 0.00019999790782247354, "loss": 5.3171, "step": 19700 }, { "epoch": 0.03685569595450421, "grad_norm": 1.6590992212295532, "learning_rate": 0.0001999978805596631, "loss": 5.3232, "step": 19750 }, { "epoch": 0.0369490015138827, "grad_norm": 1.5630754232406616, "learning_rate": 0.00019999785312037392, "loss": 5.138, "step": 19800 }, { "epoch": 0.03704230707326119, "grad_norm": 1.771061897277832, "learning_rate": 0.00019999782550460618, "loss": 5.1169, "step": 19850 }, { "epoch": 0.03713561263263968, "grad_norm": 1.6629877090454102, "learning_rate": 0.00019999779771235983, "loss": 5.207, "step": 19900 }, { "epoch": 0.03722891819201818, "grad_norm": 1.2950810194015503, "learning_rate": 0.00019999776974363503, "loss": 5.1812, "step": 19950 }, { "epoch": 0.03732222375139667, "grad_norm": 0.9732388257980347, "learning_rate": 0.00019999774159843171, "loss": 5.2428, "step": 20000 }, { "epoch": 0.03741552931077516, "grad_norm": 1.3945764303207397, "learning_rate": 0.00019999771327675, "loss": 5.2425, "step": 20050 }, { "epoch": 0.03750883487015365, "grad_norm": 1.880929946899414, "learning_rate": 0.0001999976847785899, "loss": 5.3158, "step": 20100 }, { "epoch": 0.037602140429532145, "grad_norm": 1.3822988271713257, "learning_rate": 0.00019999765610395146, "loss": 5.1829, "step": 20150 }, { "epoch": 0.037695445988910635, "grad_norm": 1.7985529899597168, "learning_rate": 0.00019999762725283479, "loss": 5.0514, "step": 20200 }, { "epoch": 0.037788751548289125, "grad_norm": 1.568518877029419, "learning_rate": 0.0001999975982252399, "loss": 5.2333, "step": 20250 }, { "epoch": 0.037882057107667615, "grad_norm": 1.4257228374481201, "learning_rate": 0.00019999756902116684, "loss": 5.1142, "step": 20300 }, { "epoch": 0.03797536266704611, "grad_norm": 1.0664187669754028, "learning_rate": 0.00019999753964061566, "loss": 5.204, "step": 20350 }, { "epoch": 0.0380686682264246, "grad_norm": 1.2831335067749023, "learning_rate": 0.00019999751008358642, "loss": 5.2272, "step": 20400 }, { "epoch": 0.03816197378580309, "grad_norm": 1.7534427642822266, "learning_rate": 0.00019999748035007917, "loss": 5.1009, "step": 20450 }, { "epoch": 0.03825527934518158, "grad_norm": 1.8466085195541382, "learning_rate": 0.00019999745044009396, "loss": 5.1638, "step": 20500 }, { "epoch": 0.03834858490456008, "grad_norm": 1.4553605318069458, "learning_rate": 0.00019999742035363083, "loss": 5.0138, "step": 20550 }, { "epoch": 0.03844189046393857, "grad_norm": 1.5238099098205566, "learning_rate": 0.00019999739009068988, "loss": 5.1723, "step": 20600 }, { "epoch": 0.03853519602331706, "grad_norm": 1.434646487236023, "learning_rate": 0.0001999973596512711, "loss": 5.1743, "step": 20650 }, { "epoch": 0.038628501582695554, "grad_norm": 1.4512965679168701, "learning_rate": 0.0001999973290353746, "loss": 5.2082, "step": 20700 }, { "epoch": 0.038721807142074044, "grad_norm": 1.4691851139068604, "learning_rate": 0.0001999972982430004, "loss": 5.3011, "step": 20750 }, { "epoch": 0.038815112701452534, "grad_norm": 1.7574752569198608, "learning_rate": 0.00019999726727414855, "loss": 4.8797, "step": 20800 }, { "epoch": 0.038908418260831024, "grad_norm": 1.9155274629592896, "learning_rate": 0.00019999723612881915, "loss": 5.2745, "step": 20850 }, { "epoch": 0.03900172382020952, "grad_norm": 1.9480228424072266, "learning_rate": 0.0001999972048070122, "loss": 5.1093, "step": 20900 }, { "epoch": 0.03909502937958801, "grad_norm": 1.314173698425293, "learning_rate": 0.0001999971733087278, "loss": 4.9488, "step": 20950 }, { "epoch": 0.0391883349389665, "grad_norm": 1.5758183002471924, "learning_rate": 0.00019999714163396597, "loss": 5.0716, "step": 21000 }, { "epoch": 0.0391883349389665, "eval_loss": 5.446030616760254, "eval_runtime": 233.6236, "eval_samples_per_second": 11.163, "eval_steps_per_second": 11.163, "eval_tts_loss": 6.94306072605591, "step": 21000 }, { "epoch": 0.03928164049834499, "grad_norm": 1.767700433731079, "learning_rate": 0.00019999710978272676, "loss": 5.152, "step": 21050 }, { "epoch": 0.03937494605772349, "grad_norm": 1.171859622001648, "learning_rate": 0.00019999707775501026, "loss": 4.9265, "step": 21100 }, { "epoch": 0.039468251617101976, "grad_norm": 1.6448893547058105, "learning_rate": 0.00019999704555081653, "loss": 5.2328, "step": 21150 }, { "epoch": 0.039561557176480466, "grad_norm": 1.2383092641830444, "learning_rate": 0.00019999701317014557, "loss": 5.0509, "step": 21200 }, { "epoch": 0.039654862735858956, "grad_norm": 1.47544264793396, "learning_rate": 0.00019999698061299752, "loss": 5.0847, "step": 21250 }, { "epoch": 0.03974816829523745, "grad_norm": 1.4962610006332397, "learning_rate": 0.00019999694787937238, "loss": 5.1514, "step": 21300 }, { "epoch": 0.03984147385461594, "grad_norm": 1.7069225311279297, "learning_rate": 0.00019999691496927022, "loss": 5.0082, "step": 21350 }, { "epoch": 0.03993477941399443, "grad_norm": 1.970062017440796, "learning_rate": 0.00019999688188269111, "loss": 5.162, "step": 21400 }, { "epoch": 0.04002808497337293, "grad_norm": 1.9130045175552368, "learning_rate": 0.00019999684861963507, "loss": 5.1627, "step": 21450 }, { "epoch": 0.04012139053275142, "grad_norm": 1.7424521446228027, "learning_rate": 0.00019999681518010221, "loss": 5.1727, "step": 21500 }, { "epoch": 0.04021469609212991, "grad_norm": 1.2897844314575195, "learning_rate": 0.00019999678156409258, "loss": 5.0252, "step": 21550 }, { "epoch": 0.0403080016515084, "grad_norm": 1.493475079536438, "learning_rate": 0.0001999967477716062, "loss": 5.029, "step": 21600 }, { "epoch": 0.040401307210886896, "grad_norm": 1.2159661054611206, "learning_rate": 0.00019999671380264316, "loss": 5.0264, "step": 21650 }, { "epoch": 0.040494612770265385, "grad_norm": 1.4600199460983276, "learning_rate": 0.0001999966796572035, "loss": 4.9779, "step": 21700 }, { "epoch": 0.040587918329643875, "grad_norm": 1.526752233505249, "learning_rate": 0.00019999664533528732, "loss": 5.1797, "step": 21750 }, { "epoch": 0.040681223889022365, "grad_norm": 1.6486625671386719, "learning_rate": 0.00019999661083689466, "loss": 5.1361, "step": 21800 }, { "epoch": 0.04077452944840086, "grad_norm": 1.9790832996368408, "learning_rate": 0.00019999657616202556, "loss": 4.9149, "step": 21850 }, { "epoch": 0.04086783500777935, "grad_norm": 1.4455928802490234, "learning_rate": 0.00019999654131068012, "loss": 5.1656, "step": 21900 }, { "epoch": 0.04096114056715784, "grad_norm": 1.556361436843872, "learning_rate": 0.00019999650628285836, "loss": 5.1226, "step": 21950 }, { "epoch": 0.04105444612653633, "grad_norm": 1.2878562211990356, "learning_rate": 0.00019999647107856036, "loss": 5.0742, "step": 22000 }, { "epoch": 0.04114775168591483, "grad_norm": 1.4406765699386597, "learning_rate": 0.0001999964356977862, "loss": 5.167, "step": 22050 }, { "epoch": 0.04124105724529332, "grad_norm": 1.3805707693099976, "learning_rate": 0.0001999964001405359, "loss": 4.9802, "step": 22100 }, { "epoch": 0.04133436280467181, "grad_norm": 1.7562808990478516, "learning_rate": 0.00019999636440680955, "loss": 5.2414, "step": 22150 }, { "epoch": 0.0414276683640503, "grad_norm": 1.5996421575546265, "learning_rate": 0.00019999632849660722, "loss": 5.2813, "step": 22200 }, { "epoch": 0.041520973923428794, "grad_norm": 1.2786633968353271, "learning_rate": 0.00019999629240992898, "loss": 5.3503, "step": 22250 }, { "epoch": 0.041614279482807284, "grad_norm": 1.7056688070297241, "learning_rate": 0.00019999625614677488, "loss": 5.205, "step": 22300 }, { "epoch": 0.041707585042185774, "grad_norm": 1.5653877258300781, "learning_rate": 0.00019999621970714495, "loss": 5.1535, "step": 22350 }, { "epoch": 0.04180089060156427, "grad_norm": 1.6800984144210815, "learning_rate": 0.00019999618309103935, "loss": 5.0171, "step": 22400 }, { "epoch": 0.04189419616094276, "grad_norm": 1.608587384223938, "learning_rate": 0.000199996146298458, "loss": 5.0021, "step": 22450 }, { "epoch": 0.04198750172032125, "grad_norm": 1.1938655376434326, "learning_rate": 0.00019999610932940108, "loss": 4.7871, "step": 22500 }, { "epoch": 0.04208080727969974, "grad_norm": 2.0445282459259033, "learning_rate": 0.00019999607218386864, "loss": 5.2227, "step": 22550 }, { "epoch": 0.04217411283907824, "grad_norm": 1.440500020980835, "learning_rate": 0.0001999960348618607, "loss": 4.9996, "step": 22600 }, { "epoch": 0.04226741839845673, "grad_norm": 1.4636365175247192, "learning_rate": 0.0001999959973633774, "loss": 5.2787, "step": 22650 }, { "epoch": 0.04236072395783522, "grad_norm": 1.6443440914154053, "learning_rate": 0.0001999959596884187, "loss": 5.1626, "step": 22700 }, { "epoch": 0.04245402951721371, "grad_norm": 1.3920623064041138, "learning_rate": 0.00019999592183698477, "loss": 5.1757, "step": 22750 }, { "epoch": 0.0425473350765922, "grad_norm": 1.5439649820327759, "learning_rate": 0.00019999588380907563, "loss": 5.1367, "step": 22800 }, { "epoch": 0.04264064063597069, "grad_norm": 1.575374960899353, "learning_rate": 0.00019999584560469134, "loss": 5.0247, "step": 22850 }, { "epoch": 0.04273394619534918, "grad_norm": 1.206429123878479, "learning_rate": 0.00019999580722383198, "loss": 5.1514, "step": 22900 }, { "epoch": 0.04282725175472767, "grad_norm": 1.5674757957458496, "learning_rate": 0.0001999957686664976, "loss": 5.208, "step": 22950 }, { "epoch": 0.04292055731410617, "grad_norm": 1.2532552480697632, "learning_rate": 0.00019999572993268828, "loss": 4.9993, "step": 23000 }, { "epoch": 0.04301386287348466, "grad_norm": 1.4708222150802612, "learning_rate": 0.00019999569102240413, "loss": 5.121, "step": 23050 }, { "epoch": 0.04310716843286315, "grad_norm": 1.6088930368423462, "learning_rate": 0.00019999565193564513, "loss": 5.0844, "step": 23100 }, { "epoch": 0.043200473992241646, "grad_norm": 1.4088431596755981, "learning_rate": 0.00019999561267241146, "loss": 5.0259, "step": 23150 }, { "epoch": 0.043293779551620136, "grad_norm": 1.430076003074646, "learning_rate": 0.00019999557323270307, "loss": 5.2003, "step": 23200 }, { "epoch": 0.043387085110998626, "grad_norm": 1.7545416355133057, "learning_rate": 0.0001999955336165201, "loss": 4.9939, "step": 23250 }, { "epoch": 0.043480390670377116, "grad_norm": 1.3078407049179077, "learning_rate": 0.00019999549382386262, "loss": 5.2717, "step": 23300 }, { "epoch": 0.04357369622975561, "grad_norm": 1.6095589399337769, "learning_rate": 0.0001999954538547307, "loss": 5.2214, "step": 23350 }, { "epoch": 0.0436670017891341, "grad_norm": 1.4139171838760376, "learning_rate": 0.00019999541370912437, "loss": 5.1498, "step": 23400 }, { "epoch": 0.04376030734851259, "grad_norm": 1.6955764293670654, "learning_rate": 0.00019999537338704376, "loss": 5.2248, "step": 23450 }, { "epoch": 0.04385361290789108, "grad_norm": 1.632716178894043, "learning_rate": 0.0001999953328884889, "loss": 4.9997, "step": 23500 }, { "epoch": 0.04394691846726958, "grad_norm": 1.7285034656524658, "learning_rate": 0.00019999529221345985, "loss": 5.029, "step": 23550 }, { "epoch": 0.04404022402664807, "grad_norm": 1.7824739217758179, "learning_rate": 0.0001999952513619567, "loss": 4.9897, "step": 23600 }, { "epoch": 0.04413352958602656, "grad_norm": 1.517888069152832, "learning_rate": 0.00019999521033397957, "loss": 5.1368, "step": 23650 }, { "epoch": 0.04422683514540505, "grad_norm": 1.3025915622711182, "learning_rate": 0.00019999516912952846, "loss": 5.1239, "step": 23700 }, { "epoch": 0.044320140704783545, "grad_norm": 1.4753954410552979, "learning_rate": 0.00019999512774860348, "loss": 5.1521, "step": 23750 }, { "epoch": 0.044413446264162035, "grad_norm": 1.6234376430511475, "learning_rate": 0.00019999508619120468, "loss": 5.1371, "step": 23800 }, { "epoch": 0.044506751823540525, "grad_norm": 1.5803349018096924, "learning_rate": 0.00019999504445733217, "loss": 5.2971, "step": 23850 }, { "epoch": 0.044600057382919014, "grad_norm": 1.5795341730117798, "learning_rate": 0.00019999500254698596, "loss": 4.8945, "step": 23900 }, { "epoch": 0.04469336294229751, "grad_norm": 1.5893889665603638, "learning_rate": 0.0001999949604601662, "loss": 5.1103, "step": 23950 }, { "epoch": 0.044786668501676, "grad_norm": 1.1029318571090698, "learning_rate": 0.00019999491819687294, "loss": 5.0341, "step": 24000 }, { "epoch": 0.044786668501676, "eval_loss": 5.369950771331787, "eval_runtime": 228.2176, "eval_samples_per_second": 11.428, "eval_steps_per_second": 11.428, "eval_tts_loss": 7.0878655090813805, "step": 24000 }, { "epoch": 0.04487997406105449, "grad_norm": 1.5672223567962646, "learning_rate": 0.00019999487575710623, "loss": 5.1548, "step": 24050 }, { "epoch": 0.04497327962043299, "grad_norm": 1.3560975790023804, "learning_rate": 0.0001999948331408662, "loss": 5.1028, "step": 24100 }, { "epoch": 0.04506658517981148, "grad_norm": 1.365408182144165, "learning_rate": 0.0001999947903481528, "loss": 5.0995, "step": 24150 }, { "epoch": 0.04515989073918997, "grad_norm": 1.3482069969177246, "learning_rate": 0.00019999474737896625, "loss": 4.9473, "step": 24200 }, { "epoch": 0.04525319629856846, "grad_norm": 1.2890368700027466, "learning_rate": 0.00019999470423330654, "loss": 4.9009, "step": 24250 }, { "epoch": 0.045346501857946954, "grad_norm": 1.2973231077194214, "learning_rate": 0.00019999466091117377, "loss": 5.0809, "step": 24300 }, { "epoch": 0.045439807417325444, "grad_norm": 1.5761609077453613, "learning_rate": 0.000199994617412568, "loss": 4.9994, "step": 24350 }, { "epoch": 0.045533112976703934, "grad_norm": 1.3380223512649536, "learning_rate": 0.00019999457373748937, "loss": 4.916, "step": 24400 }, { "epoch": 0.04562641853608242, "grad_norm": 1.7922528982162476, "learning_rate": 0.00019999452988593789, "loss": 5.0799, "step": 24450 }, { "epoch": 0.04571972409546092, "grad_norm": 1.3923864364624023, "learning_rate": 0.00019999448585791369, "loss": 5.0239, "step": 24500 }, { "epoch": 0.04581302965483941, "grad_norm": 1.2720636129379272, "learning_rate": 0.0001999944416534168, "loss": 4.9941, "step": 24550 }, { "epoch": 0.0459063352142179, "grad_norm": 1.602018117904663, "learning_rate": 0.00019999439727244728, "loss": 5.1485, "step": 24600 }, { "epoch": 0.04599964077359639, "grad_norm": 1.5410170555114746, "learning_rate": 0.00019999435271500526, "loss": 5.2976, "step": 24650 }, { "epoch": 0.046092946332974886, "grad_norm": 1.6076123714447021, "learning_rate": 0.00019999430798109082, "loss": 4.9226, "step": 24700 }, { "epoch": 0.046186251892353376, "grad_norm": 1.3308855295181274, "learning_rate": 0.00019999426307070402, "loss": 5.0585, "step": 24750 }, { "epoch": 0.046279557451731866, "grad_norm": 1.0817416906356812, "learning_rate": 0.00019999421798384492, "loss": 4.9678, "step": 24800 }, { "epoch": 0.04637286301111036, "grad_norm": 1.4719314575195312, "learning_rate": 0.00019999417272051364, "loss": 4.734, "step": 24850 }, { "epoch": 0.04646616857048885, "grad_norm": 1.911425232887268, "learning_rate": 0.00019999412728071022, "loss": 5.023, "step": 24900 }, { "epoch": 0.04655947412986734, "grad_norm": 1.2537506818771362, "learning_rate": 0.0001999940816644348, "loss": 5.0473, "step": 24950 }, { "epoch": 0.04665277968924583, "grad_norm": 1.2723119258880615, "learning_rate": 0.00019999403587168737, "loss": 5.0075, "step": 25000 }, { "epoch": 0.04674608524862433, "grad_norm": 1.221137523651123, "learning_rate": 0.00019999398990246808, "loss": 5.0694, "step": 25050 }, { "epoch": 0.04683939080800282, "grad_norm": 1.6972732543945312, "learning_rate": 0.000199993943756777, "loss": 5.0661, "step": 25100 }, { "epoch": 0.04693269636738131, "grad_norm": 1.5249754190444946, "learning_rate": 0.00019999389743461418, "loss": 5.0894, "step": 25150 }, { "epoch": 0.0470260019267598, "grad_norm": 1.5337498188018799, "learning_rate": 0.00019999385093597975, "loss": 5.2557, "step": 25200 }, { "epoch": 0.047119307486138295, "grad_norm": 1.6065733432769775, "learning_rate": 0.00019999380426087378, "loss": 5.2814, "step": 25250 }, { "epoch": 0.047212613045516785, "grad_norm": 1.6499677896499634, "learning_rate": 0.0001999937574092963, "loss": 5.0703, "step": 25300 }, { "epoch": 0.047305918604895275, "grad_norm": 0.9138503670692444, "learning_rate": 0.00019999371038124746, "loss": 5.1375, "step": 25350 }, { "epoch": 0.047399224164273765, "grad_norm": 1.533087968826294, "learning_rate": 0.00019999366317672733, "loss": 5.0459, "step": 25400 }, { "epoch": 0.04749252972365226, "grad_norm": 1.8451659679412842, "learning_rate": 0.00019999361579573592, "loss": 4.9868, "step": 25450 }, { "epoch": 0.04758583528303075, "grad_norm": 1.0893833637237549, "learning_rate": 0.00019999356823827342, "loss": 4.9291, "step": 25500 }, { "epoch": 0.04767914084240924, "grad_norm": 1.099062204360962, "learning_rate": 0.00019999352050433986, "loss": 5.0681, "step": 25550 }, { "epoch": 0.04777244640178773, "grad_norm": 1.3058807849884033, "learning_rate": 0.00019999347259393532, "loss": 4.8777, "step": 25600 }, { "epoch": 0.04786575196116623, "grad_norm": 1.6822357177734375, "learning_rate": 0.0001999934245070599, "loss": 5.2018, "step": 25650 }, { "epoch": 0.04795905752054472, "grad_norm": 1.1984933614730835, "learning_rate": 0.0001999933762437137, "loss": 5.0762, "step": 25700 }, { "epoch": 0.04805236307992321, "grad_norm": 1.4918663501739502, "learning_rate": 0.00019999332780389677, "loss": 5.0357, "step": 25750 }, { "epoch": 0.048145668639301704, "grad_norm": 1.7189278602600098, "learning_rate": 0.0001999932791876092, "loss": 5.0787, "step": 25800 }, { "epoch": 0.048238974198680194, "grad_norm": 1.2158808708190918, "learning_rate": 0.00019999323039485113, "loss": 4.9543, "step": 25850 }, { "epoch": 0.048332279758058684, "grad_norm": 1.298972725868225, "learning_rate": 0.00019999318142562253, "loss": 4.8591, "step": 25900 }, { "epoch": 0.048425585317437174, "grad_norm": 1.8158042430877686, "learning_rate": 0.0001999931322799236, "loss": 4.8411, "step": 25950 }, { "epoch": 0.04851889087681567, "grad_norm": 1.2872735261917114, "learning_rate": 0.00019999308295775442, "loss": 4.9664, "step": 26000 }, { "epoch": 0.04861219643619416, "grad_norm": 1.3019319772720337, "learning_rate": 0.00019999303345911496, "loss": 5.1429, "step": 26050 }, { "epoch": 0.04870550199557265, "grad_norm": 1.3941707611083984, "learning_rate": 0.00019999298378400545, "loss": 5.0969, "step": 26100 }, { "epoch": 0.04879880755495114, "grad_norm": 1.3208707571029663, "learning_rate": 0.0001999929339324259, "loss": 5.0603, "step": 26150 }, { "epoch": 0.04889211311432964, "grad_norm": 1.7119523286819458, "learning_rate": 0.00019999288390437644, "loss": 5.0922, "step": 26200 }, { "epoch": 0.04898541867370813, "grad_norm": 1.4633029699325562, "learning_rate": 0.0001999928336998571, "loss": 5.0348, "step": 26250 }, { "epoch": 0.04907872423308662, "grad_norm": 1.695513367652893, "learning_rate": 0.00019999278331886804, "loss": 4.984, "step": 26300 }, { "epoch": 0.049172029792465106, "grad_norm": 1.6055431365966797, "learning_rate": 0.0001999927327614093, "loss": 5.0255, "step": 26350 }, { "epoch": 0.0492653353518436, "grad_norm": 1.5823806524276733, "learning_rate": 0.00019999268202748092, "loss": 4.9689, "step": 26400 }, { "epoch": 0.04935864091122209, "grad_norm": 1.099482774734497, "learning_rate": 0.00019999263111708312, "loss": 5.1056, "step": 26450 }, { "epoch": 0.04945194647060058, "grad_norm": 1.2539039850234985, "learning_rate": 0.0001999925800302159, "loss": 5.1003, "step": 26500 }, { "epoch": 0.04954525202997908, "grad_norm": 1.8513319492340088, "learning_rate": 0.0001999925287668794, "loss": 5.0544, "step": 26550 }, { "epoch": 0.04963855758935757, "grad_norm": 1.6248927116394043, "learning_rate": 0.00019999247732707364, "loss": 5.0591, "step": 26600 }, { "epoch": 0.04973186314873606, "grad_norm": 1.4035404920578003, "learning_rate": 0.00019999242571079874, "loss": 4.7997, "step": 26650 }, { "epoch": 0.04982516870811455, "grad_norm": 1.5909003019332886, "learning_rate": 0.00019999237391805482, "loss": 5.0233, "step": 26700 }, { "epoch": 0.049918474267493046, "grad_norm": 1.187412977218628, "learning_rate": 0.00019999232194884198, "loss": 4.9776, "step": 26750 }, { "epoch": 0.050011779826871536, "grad_norm": 0.9005943536758423, "learning_rate": 0.00019999226980316025, "loss": 4.9586, "step": 26800 }, { "epoch": 0.050105085386250026, "grad_norm": 1.1468946933746338, "learning_rate": 0.0001999922174810098, "loss": 5.0651, "step": 26850 }, { "epoch": 0.050198390945628515, "grad_norm": 1.2562627792358398, "learning_rate": 0.00019999216498239064, "loss": 4.8625, "step": 26900 }, { "epoch": 0.05029169650500701, "grad_norm": 1.2052425146102905, "learning_rate": 0.00019999211230730294, "loss": 4.9507, "step": 26950 }, { "epoch": 0.0503850020643855, "grad_norm": 1.3357577323913574, "learning_rate": 0.00019999205945574671, "loss": 4.9069, "step": 27000 }, { "epoch": 0.0503850020643855, "eval_loss": 5.300256252288818, "eval_runtime": 230.7705, "eval_samples_per_second": 11.301, "eval_steps_per_second": 11.301, "eval_tts_loss": 6.999222072712689, "step": 27000 }, { "epoch": 0.05047830762376399, "grad_norm": 1.0808535814285278, "learning_rate": 0.00019999200642772214, "loss": 4.9514, "step": 27050 }, { "epoch": 0.05057161318314248, "grad_norm": 1.6835181713104248, "learning_rate": 0.00019999195322322922, "loss": 5.1339, "step": 27100 }, { "epoch": 0.05066491874252098, "grad_norm": 1.3736001253128052, "learning_rate": 0.00019999189984226814, "loss": 4.8656, "step": 27150 }, { "epoch": 0.05075822430189947, "grad_norm": 1.404284954071045, "learning_rate": 0.00019999184628483892, "loss": 4.8601, "step": 27200 }, { "epoch": 0.05085152986127796, "grad_norm": 1.566685676574707, "learning_rate": 0.0001999917925509417, "loss": 4.9994, "step": 27250 }, { "epoch": 0.05094483542065645, "grad_norm": 1.4436407089233398, "learning_rate": 0.00019999173864057658, "loss": 4.9391, "step": 27300 }, { "epoch": 0.051038140980034945, "grad_norm": 1.1985220909118652, "learning_rate": 0.0001999916845537436, "loss": 4.9497, "step": 27350 }, { "epoch": 0.051131446539413435, "grad_norm": 1.4295037984848022, "learning_rate": 0.00019999163029044292, "loss": 5.0714, "step": 27400 }, { "epoch": 0.051224752098791924, "grad_norm": 1.8160730600357056, "learning_rate": 0.00019999157585067458, "loss": 4.9091, "step": 27450 }, { "epoch": 0.05131805765817042, "grad_norm": 1.26568603515625, "learning_rate": 0.00019999152123443872, "loss": 5.0946, "step": 27500 }, { "epoch": 0.05141136321754891, "grad_norm": 1.7634291648864746, "learning_rate": 0.00019999146644173542, "loss": 4.6454, "step": 27550 }, { "epoch": 0.0515046687769274, "grad_norm": 1.3475509881973267, "learning_rate": 0.00019999141147256476, "loss": 4.9264, "step": 27600 }, { "epoch": 0.05159797433630589, "grad_norm": 1.6052281856536865, "learning_rate": 0.00019999135632692687, "loss": 5.0712, "step": 27650 }, { "epoch": 0.05169127989568439, "grad_norm": 1.3022325038909912, "learning_rate": 0.00019999130100482182, "loss": 5.0866, "step": 27700 }, { "epoch": 0.05178458545506288, "grad_norm": 1.5663292407989502, "learning_rate": 0.00019999124550624972, "loss": 5.2636, "step": 27750 }, { "epoch": 0.05187789101444137, "grad_norm": 1.6265084743499756, "learning_rate": 0.00019999118983121065, "loss": 5.0156, "step": 27800 }, { "epoch": 0.05197119657381986, "grad_norm": 1.2428834438323975, "learning_rate": 0.00019999113397970475, "loss": 5.0459, "step": 27850 }, { "epoch": 0.052064502133198354, "grad_norm": 1.7661765813827515, "learning_rate": 0.0001999910779517321, "loss": 4.9641, "step": 27900 }, { "epoch": 0.052157807692576844, "grad_norm": 1.2354525327682495, "learning_rate": 0.00019999102174729277, "loss": 4.8599, "step": 27950 }, { "epoch": 0.05225111325195533, "grad_norm": 1.1708762645721436, "learning_rate": 0.0001999909653663869, "loss": 4.8535, "step": 28000 }, { "epoch": 0.05234441881133382, "grad_norm": 1.4422293901443481, "learning_rate": 0.00019999090880901455, "loss": 4.8957, "step": 28050 }, { "epoch": 0.05243772437071232, "grad_norm": 1.058984637260437, "learning_rate": 0.00019999085207517584, "loss": 5.0249, "step": 28100 }, { "epoch": 0.05253102993009081, "grad_norm": 1.315977692604065, "learning_rate": 0.00019999079516487087, "loss": 5.0057, "step": 28150 }, { "epoch": 0.0526243354894693, "grad_norm": 1.4370442628860474, "learning_rate": 0.00019999073807809974, "loss": 5.0101, "step": 28200 }, { "epoch": 0.05271764104884779, "grad_norm": 1.54827880859375, "learning_rate": 0.00019999068081486256, "loss": 5.054, "step": 28250 }, { "epoch": 0.052810946608226286, "grad_norm": 1.4660943746566772, "learning_rate": 0.00019999062337515945, "loss": 5.1078, "step": 28300 }, { "epoch": 0.052904252167604776, "grad_norm": 1.483725666999817, "learning_rate": 0.00019999056575899043, "loss": 4.935, "step": 28350 }, { "epoch": 0.052997557726983266, "grad_norm": 1.4430571794509888, "learning_rate": 0.00019999050796635568, "loss": 4.9998, "step": 28400 }, { "epoch": 0.05309086328636176, "grad_norm": 1.6983498334884644, "learning_rate": 0.00019999044999725527, "loss": 5.2524, "step": 28450 }, { "epoch": 0.05318416884574025, "grad_norm": 1.6908702850341797, "learning_rate": 0.0001999903918516893, "loss": 5.0378, "step": 28500 }, { "epoch": 0.05327747440511874, "grad_norm": 1.5972379446029663, "learning_rate": 0.00019999033352965793, "loss": 4.9017, "step": 28550 }, { "epoch": 0.05337077996449723, "grad_norm": 1.5607515573501587, "learning_rate": 0.00019999027503116116, "loss": 4.9577, "step": 28600 }, { "epoch": 0.05346408552387573, "grad_norm": 1.344556450843811, "learning_rate": 0.00019999021635619917, "loss": 5.0421, "step": 28650 }, { "epoch": 0.05355739108325422, "grad_norm": 1.3197835683822632, "learning_rate": 0.00019999015750477207, "loss": 5.1273, "step": 28700 }, { "epoch": 0.05365069664263271, "grad_norm": 1.4137849807739258, "learning_rate": 0.00019999009847687988, "loss": 4.8728, "step": 28750 }, { "epoch": 0.0537440022020112, "grad_norm": 1.2608588933944702, "learning_rate": 0.00019999003927252278, "loss": 5.0248, "step": 28800 }, { "epoch": 0.053837307761389695, "grad_norm": 1.033444881439209, "learning_rate": 0.00019998997989170087, "loss": 5.1288, "step": 28850 }, { "epoch": 0.053930613320768185, "grad_norm": 1.6949776411056519, "learning_rate": 0.00019998992033441423, "loss": 5.0369, "step": 28900 }, { "epoch": 0.054023918880146675, "grad_norm": 1.2738184928894043, "learning_rate": 0.00019998986060066297, "loss": 5.0106, "step": 28950 }, { "epoch": 0.054117224439525165, "grad_norm": 1.1779210567474365, "learning_rate": 0.0001999898006904472, "loss": 4.9383, "step": 29000 }, { "epoch": 0.05421052999890366, "grad_norm": 1.4241044521331787, "learning_rate": 0.00019998974060376706, "loss": 5.1722, "step": 29050 }, { "epoch": 0.05430383555828215, "grad_norm": 1.421637773513794, "learning_rate": 0.00019998968034062257, "loss": 5.1138, "step": 29100 }, { "epoch": 0.05439714111766064, "grad_norm": 1.2639997005462646, "learning_rate": 0.00019998961990101393, "loss": 5.1452, "step": 29150 }, { "epoch": 0.05449044667703914, "grad_norm": 1.241308569908142, "learning_rate": 0.0001999895592849412, "loss": 5.0645, "step": 29200 }, { "epoch": 0.05458375223641763, "grad_norm": 1.6818753480911255, "learning_rate": 0.00019998949849240444, "loss": 5.2461, "step": 29250 }, { "epoch": 0.05467705779579612, "grad_norm": 0.8954434990882874, "learning_rate": 0.00019998943752340387, "loss": 4.9691, "step": 29300 }, { "epoch": 0.05477036335517461, "grad_norm": 1.670721411705017, "learning_rate": 0.0001999893763779395, "loss": 5.0303, "step": 29350 }, { "epoch": 0.054863668914553104, "grad_norm": 1.2590948343276978, "learning_rate": 0.00019998931505601147, "loss": 4.961, "step": 29400 }, { "epoch": 0.054956974473931594, "grad_norm": 1.306104302406311, "learning_rate": 0.00019998925355761994, "loss": 4.889, "step": 29450 }, { "epoch": 0.055050280033310084, "grad_norm": 1.4491585493087769, "learning_rate": 0.00019998919188276494, "loss": 4.8524, "step": 29500 }, { "epoch": 0.055143585592688574, "grad_norm": 1.5823389291763306, "learning_rate": 0.0001999891300314466, "loss": 4.994, "step": 29550 }, { "epoch": 0.05523689115206707, "grad_norm": 1.5305702686309814, "learning_rate": 0.00019998906800366506, "loss": 4.8367, "step": 29600 }, { "epoch": 0.05533019671144556, "grad_norm": 1.538707971572876, "learning_rate": 0.0001999890057994204, "loss": 4.9698, "step": 29650 }, { "epoch": 0.05542350227082405, "grad_norm": 1.4228105545043945, "learning_rate": 0.00019998894341871274, "loss": 5.1362, "step": 29700 }, { "epoch": 0.05551680783020254, "grad_norm": 1.1229313611984253, "learning_rate": 0.00019998888086154217, "loss": 4.9158, "step": 29750 }, { "epoch": 0.05561011338958104, "grad_norm": 1.2107417583465576, "learning_rate": 0.00019998881812790883, "loss": 5.0441, "step": 29800 }, { "epoch": 0.05570341894895953, "grad_norm": 1.2901204824447632, "learning_rate": 0.00019998875521781283, "loss": 5.0762, "step": 29850 }, { "epoch": 0.055796724508338016, "grad_norm": 1.3414353132247925, "learning_rate": 0.00019998869213125428, "loss": 4.9442, "step": 29900 }, { "epoch": 0.055890030067716506, "grad_norm": 1.3812546730041504, "learning_rate": 0.00019998862886823327, "loss": 4.8799, "step": 29950 }, { "epoch": 0.055983335627095, "grad_norm": 1.2533543109893799, "learning_rate": 0.00019998856542874992, "loss": 4.8482, "step": 30000 }, { "epoch": 0.055983335627095, "eval_loss": 5.270205020904541, "eval_runtime": 231.0564, "eval_samples_per_second": 11.287, "eval_steps_per_second": 11.287, "eval_tts_loss": 7.119214073906258, "step": 30000 }, { "epoch": 0.05607664118647349, "grad_norm": 1.4817043542861938, "learning_rate": 0.00019998850181280432, "loss": 4.9294, "step": 30050 }, { "epoch": 0.05616994674585198, "grad_norm": 1.3725059032440186, "learning_rate": 0.00019998843802039663, "loss": 4.7408, "step": 30100 }, { "epoch": 0.05626325230523048, "grad_norm": 1.1192152500152588, "learning_rate": 0.00019998837405152696, "loss": 4.975, "step": 30150 }, { "epoch": 0.05635655786460897, "grad_norm": 1.4426796436309814, "learning_rate": 0.0001999883099061954, "loss": 5.1216, "step": 30200 }, { "epoch": 0.05644986342398746, "grad_norm": 1.060693383216858, "learning_rate": 0.00019998824558440203, "loss": 4.8414, "step": 30250 }, { "epoch": 0.05654316898336595, "grad_norm": 1.2075954675674438, "learning_rate": 0.00019998818108614704, "loss": 4.8929, "step": 30300 }, { "epoch": 0.056636474542744446, "grad_norm": 1.2428367137908936, "learning_rate": 0.00019998811641143047, "loss": 5.1015, "step": 30350 }, { "epoch": 0.056729780102122936, "grad_norm": 1.2443081140518188, "learning_rate": 0.0001999880515602525, "loss": 5.0629, "step": 30400 }, { "epoch": 0.056823085661501425, "grad_norm": 1.4452126026153564, "learning_rate": 0.0001999879865326132, "loss": 4.942, "step": 30450 }, { "epoch": 0.056916391220879915, "grad_norm": 0.8212096691131592, "learning_rate": 0.0001999879213285127, "loss": 4.9346, "step": 30500 }, { "epoch": 0.05700969678025841, "grad_norm": 1.2501893043518066, "learning_rate": 0.0001999878559479511, "loss": 4.8447, "step": 30550 }, { "epoch": 0.0571030023396369, "grad_norm": 1.0458223819732666, "learning_rate": 0.00019998779039092854, "loss": 4.929, "step": 30600 }, { "epoch": 0.05719630789901539, "grad_norm": 1.5575039386749268, "learning_rate": 0.0001999877246574451, "loss": 5.0728, "step": 30650 }, { "epoch": 0.05728961345839388, "grad_norm": 0.9992212057113647, "learning_rate": 0.00019998765874750096, "loss": 5.0842, "step": 30700 }, { "epoch": 0.05738291901777238, "grad_norm": 1.637300729751587, "learning_rate": 0.00019998759266109617, "loss": 4.9798, "step": 30750 }, { "epoch": 0.05747622457715087, "grad_norm": 1.372078537940979, "learning_rate": 0.00019998752639823088, "loss": 5.1645, "step": 30800 }, { "epoch": 0.05756953013652936, "grad_norm": 1.0558440685272217, "learning_rate": 0.0001999874599589052, "loss": 4.833, "step": 30850 }, { "epoch": 0.057662835695907855, "grad_norm": 1.3635761737823486, "learning_rate": 0.00019998739334311922, "loss": 4.7573, "step": 30900 }, { "epoch": 0.057756141255286345, "grad_norm": 1.2139973640441895, "learning_rate": 0.00019998732655087312, "loss": 5.1266, "step": 30950 }, { "epoch": 0.057849446814664834, "grad_norm": 1.2504178285598755, "learning_rate": 0.00019998725958216695, "loss": 5.1571, "step": 31000 }, { "epoch": 0.057942752374043324, "grad_norm": 1.4248679876327515, "learning_rate": 0.00019998719243700087, "loss": 5.1027, "step": 31050 }, { "epoch": 0.05803605793342182, "grad_norm": 4.532711029052734, "learning_rate": 0.000199987125115375, "loss": 5.1239, "step": 31100 }, { "epoch": 0.05812936349280031, "grad_norm": 1.717026710510254, "learning_rate": 0.00019998705761728943, "loss": 5.1622, "step": 31150 }, { "epoch": 0.0582226690521788, "grad_norm": 1.3576337099075317, "learning_rate": 0.00019998698994274435, "loss": 4.9084, "step": 31200 }, { "epoch": 0.05831597461155729, "grad_norm": 1.2413636445999146, "learning_rate": 0.00019998692209173977, "loss": 4.791, "step": 31250 }, { "epoch": 0.05840928017093579, "grad_norm": 1.3183064460754395, "learning_rate": 0.00019998685406427588, "loss": 5.1716, "step": 31300 }, { "epoch": 0.05850258573031428, "grad_norm": 1.0643527507781982, "learning_rate": 0.0001999867858603528, "loss": 4.8196, "step": 31350 }, { "epoch": 0.05859589128969277, "grad_norm": 1.1903074979782104, "learning_rate": 0.00019998671747997062, "loss": 4.8136, "step": 31400 }, { "epoch": 0.05868919684907126, "grad_norm": 1.2945053577423096, "learning_rate": 0.00019998664892312948, "loss": 4.9986, "step": 31450 }, { "epoch": 0.058782502408449754, "grad_norm": 1.5691232681274414, "learning_rate": 0.0001999865801898295, "loss": 5.0726, "step": 31500 }, { "epoch": 0.05887580796782824, "grad_norm": 1.3560112714767456, "learning_rate": 0.0001999865112800708, "loss": 4.9443, "step": 31550 }, { "epoch": 0.05896911352720673, "grad_norm": 1.1290074586868286, "learning_rate": 0.0001999864421938535, "loss": 4.9773, "step": 31600 }, { "epoch": 0.05906241908658522, "grad_norm": 1.4924343824386597, "learning_rate": 0.00019998637293117773, "loss": 5.1788, "step": 31650 }, { "epoch": 0.05915572464596372, "grad_norm": 1.2043700218200684, "learning_rate": 0.0001999863034920436, "loss": 4.8067, "step": 31700 }, { "epoch": 0.05924903020534221, "grad_norm": 1.2821531295776367, "learning_rate": 0.00019998623387645122, "loss": 5.01, "step": 31750 }, { "epoch": 0.0593423357647207, "grad_norm": 0.9705774188041687, "learning_rate": 0.00019998616408440075, "loss": 5.0704, "step": 31800 }, { "epoch": 0.059435641324099196, "grad_norm": 1.632771372795105, "learning_rate": 0.00019998609411589228, "loss": 5.1367, "step": 31850 }, { "epoch": 0.059528946883477686, "grad_norm": 1.5198719501495361, "learning_rate": 0.00019998602397092597, "loss": 4.806, "step": 31900 }, { "epoch": 0.059622252442856176, "grad_norm": 1.2326031923294067, "learning_rate": 0.00019998595364950191, "loss": 4.9858, "step": 31950 }, { "epoch": 0.059715558002234666, "grad_norm": 1.4651892185211182, "learning_rate": 0.00019998588315162018, "loss": 4.9011, "step": 32000 }, { "epoch": 0.05980886356161316, "grad_norm": 1.704825520515442, "learning_rate": 0.00019998581247728103, "loss": 5.1689, "step": 32050 }, { "epoch": 0.05990216912099165, "grad_norm": 1.4185510873794556, "learning_rate": 0.0001999857416264845, "loss": 4.984, "step": 32100 }, { "epoch": 0.05999547468037014, "grad_norm": 1.7740285396575928, "learning_rate": 0.00019998567059923073, "loss": 5.0549, "step": 32150 }, { "epoch": 0.06008878023974863, "grad_norm": 1.117332100868225, "learning_rate": 0.00019998559939551983, "loss": 4.9824, "step": 32200 }, { "epoch": 0.06018208579912713, "grad_norm": 1.4967530965805054, "learning_rate": 0.00019998552801535192, "loss": 5.0119, "step": 32250 }, { "epoch": 0.06027539135850562, "grad_norm": 1.1387840509414673, "learning_rate": 0.0001999854564587272, "loss": 5.0042, "step": 32300 }, { "epoch": 0.06036869691788411, "grad_norm": 1.2850226163864136, "learning_rate": 0.0001999853847256457, "loss": 4.7995, "step": 32350 }, { "epoch": 0.0604620024772626, "grad_norm": 1.3279659748077393, "learning_rate": 0.00019998531281610762, "loss": 4.9891, "step": 32400 }, { "epoch": 0.060555308036641095, "grad_norm": 1.3600449562072754, "learning_rate": 0.00019998524073011305, "loss": 4.9334, "step": 32450 }, { "epoch": 0.060648613596019585, "grad_norm": 1.053346872329712, "learning_rate": 0.0001999851684676621, "loss": 4.8668, "step": 32500 }, { "epoch": 0.060741919155398075, "grad_norm": 1.311558485031128, "learning_rate": 0.00019998509602875495, "loss": 4.9149, "step": 32550 }, { "epoch": 0.06083522471477657, "grad_norm": 1.0660821199417114, "learning_rate": 0.00019998502341339166, "loss": 4.8516, "step": 32600 }, { "epoch": 0.06092853027415506, "grad_norm": 1.450120210647583, "learning_rate": 0.00019998495062157243, "loss": 5.0558, "step": 32650 }, { "epoch": 0.06102183583353355, "grad_norm": 1.1899629831314087, "learning_rate": 0.00019998487765329734, "loss": 4.9774, "step": 32700 }, { "epoch": 0.06111514139291204, "grad_norm": 1.3648253679275513, "learning_rate": 0.00019998480450856654, "loss": 4.9348, "step": 32750 }, { "epoch": 0.06120844695229054, "grad_norm": 1.1677742004394531, "learning_rate": 0.00019998473118738015, "loss": 4.9243, "step": 32800 }, { "epoch": 0.06130175251166903, "grad_norm": 1.0506433248519897, "learning_rate": 0.0001999846576897383, "loss": 4.9175, "step": 32850 }, { "epoch": 0.06139505807104752, "grad_norm": 1.6663371324539185, "learning_rate": 0.00019998458401564113, "loss": 5.0618, "step": 32900 }, { "epoch": 0.06148836363042601, "grad_norm": 1.1988695859909058, "learning_rate": 0.00019998451016508875, "loss": 4.9605, "step": 32950 }, { "epoch": 0.061581669189804504, "grad_norm": 1.0045593976974487, "learning_rate": 0.0001999844361380813, "loss": 4.9856, "step": 33000 }, { "epoch": 0.061581669189804504, "eval_loss": 5.222363471984863, "eval_runtime": 230.7018, "eval_samples_per_second": 11.305, "eval_steps_per_second": 11.305, "eval_tts_loss": 7.172317200288908, "step": 33000 }, { "epoch": 0.061674974749182994, "grad_norm": 1.099878191947937, "learning_rate": 0.00019998436193461894, "loss": 4.8131, "step": 33050 }, { "epoch": 0.061768280308561484, "grad_norm": 1.5296263694763184, "learning_rate": 0.00019998428755470176, "loss": 4.8085, "step": 33100 }, { "epoch": 0.061861585867939974, "grad_norm": 1.1066418886184692, "learning_rate": 0.0001999842129983299, "loss": 4.9998, "step": 33150 }, { "epoch": 0.06195489142731847, "grad_norm": 1.0647575855255127, "learning_rate": 0.00019998413826550349, "loss": 4.9304, "step": 33200 }, { "epoch": 0.06204819698669696, "grad_norm": 1.4591343402862549, "learning_rate": 0.0001999840633562227, "loss": 4.7926, "step": 33250 }, { "epoch": 0.06214150254607545, "grad_norm": 1.5085718631744385, "learning_rate": 0.0001999839882704876, "loss": 4.9535, "step": 33300 }, { "epoch": 0.06223480810545394, "grad_norm": 1.4653303623199463, "learning_rate": 0.00019998391300829834, "loss": 5.0559, "step": 33350 }, { "epoch": 0.06232811366483244, "grad_norm": 1.4487497806549072, "learning_rate": 0.00019998383756965508, "loss": 4.7687, "step": 33400 }, { "epoch": 0.062421419224210926, "grad_norm": 1.0603256225585938, "learning_rate": 0.00019998376195455797, "loss": 5.0553, "step": 33450 }, { "epoch": 0.06251472478358942, "grad_norm": 1.2172232866287231, "learning_rate": 0.0001999836861630071, "loss": 5.016, "step": 33500 }, { "epoch": 0.06260803034296791, "grad_norm": 1.0388457775115967, "learning_rate": 0.0001999836101950026, "loss": 4.9595, "step": 33550 }, { "epoch": 0.0627013359023464, "grad_norm": 1.064448595046997, "learning_rate": 0.00019998353405054463, "loss": 4.9284, "step": 33600 }, { "epoch": 0.06279464146172489, "grad_norm": 1.3554415702819824, "learning_rate": 0.0001999834577296333, "loss": 4.9094, "step": 33650 }, { "epoch": 0.06288794702110338, "grad_norm": 1.323470950126648, "learning_rate": 0.00019998338123226878, "loss": 5.2709, "step": 33700 }, { "epoch": 0.06298125258048187, "grad_norm": 1.378735899925232, "learning_rate": 0.00019998330455845114, "loss": 5.002, "step": 33750 }, { "epoch": 0.06307455813986036, "grad_norm": 1.3667746782302856, "learning_rate": 0.00019998322770818064, "loss": 5.029, "step": 33800 }, { "epoch": 0.06316786369923887, "grad_norm": 1.2016565799713135, "learning_rate": 0.0001999831506814573, "loss": 4.7827, "step": 33850 }, { "epoch": 0.06326116925861736, "grad_norm": 1.6369880437850952, "learning_rate": 0.00019998307347828128, "loss": 4.9345, "step": 33900 }, { "epoch": 0.06335447481799585, "grad_norm": 1.6404528617858887, "learning_rate": 0.00019998299609865272, "loss": 4.9889, "step": 33950 }, { "epoch": 0.06344778037737434, "grad_norm": 1.322618007659912, "learning_rate": 0.00019998291854257175, "loss": 4.9924, "step": 34000 }, { "epoch": 0.06354108593675283, "grad_norm": 1.2779415845870972, "learning_rate": 0.00019998284081003856, "loss": 4.7519, "step": 34050 }, { "epoch": 0.06363439149613132, "grad_norm": 1.0011979341506958, "learning_rate": 0.00019998276290105325, "loss": 5.0784, "step": 34100 }, { "epoch": 0.0637276970555098, "grad_norm": 1.2990458011627197, "learning_rate": 0.00019998268481561594, "loss": 4.85, "step": 34150 }, { "epoch": 0.0638210026148883, "grad_norm": 0.8773101568222046, "learning_rate": 0.00019998260655372678, "loss": 4.9803, "step": 34200 }, { "epoch": 0.0639143081742668, "grad_norm": 0.890830934047699, "learning_rate": 0.00019998252811538595, "loss": 5.0072, "step": 34250 }, { "epoch": 0.06400761373364529, "grad_norm": 1.118335485458374, "learning_rate": 0.00019998244950059348, "loss": 4.9289, "step": 34300 }, { "epoch": 0.06410091929302378, "grad_norm": 1.284605860710144, "learning_rate": 0.00019998237070934964, "loss": 5.1104, "step": 34350 }, { "epoch": 0.06419422485240227, "grad_norm": 1.3694264888763428, "learning_rate": 0.00019998229174165446, "loss": 5.0689, "step": 34400 }, { "epoch": 0.06428753041178076, "grad_norm": 1.252570390701294, "learning_rate": 0.00019998221259750814, "loss": 4.9514, "step": 34450 }, { "epoch": 0.06438083597115925, "grad_norm": 1.339966893196106, "learning_rate": 0.00019998213327691085, "loss": 5.0985, "step": 34500 }, { "epoch": 0.06447414153053774, "grad_norm": 1.0073368549346924, "learning_rate": 0.00019998205377986267, "loss": 4.7595, "step": 34550 }, { "epoch": 0.06456744708991624, "grad_norm": 1.4048619270324707, "learning_rate": 0.00019998197410636372, "loss": 4.9045, "step": 34600 }, { "epoch": 0.06466075264929473, "grad_norm": 0.9788991808891296, "learning_rate": 0.00019998189425641423, "loss": 4.4886, "step": 34650 }, { "epoch": 0.06475405820867322, "grad_norm": 1.4313668012619019, "learning_rate": 0.00019998181423001423, "loss": 4.7508, "step": 34700 }, { "epoch": 0.06484736376805171, "grad_norm": 1.3244096040725708, "learning_rate": 0.00019998173402716396, "loss": 5.0475, "step": 34750 }, { "epoch": 0.0649406693274302, "grad_norm": 1.0660874843597412, "learning_rate": 0.00019998165364786352, "loss": 4.9674, "step": 34800 }, { "epoch": 0.06503397488680869, "grad_norm": 1.2351161241531372, "learning_rate": 0.00019998157309211305, "loss": 5.0454, "step": 34850 }, { "epoch": 0.06512728044618718, "grad_norm": 1.4965345859527588, "learning_rate": 0.00019998149235991267, "loss": 4.8981, "step": 34900 }, { "epoch": 0.06522058600556567, "grad_norm": 1.113448143005371, "learning_rate": 0.00019998141145126258, "loss": 5.1095, "step": 34950 }, { "epoch": 0.06531389156494417, "grad_norm": 1.1176252365112305, "learning_rate": 0.0001999813303661629, "loss": 5.0817, "step": 35000 }, { "epoch": 0.06540719712432266, "grad_norm": 1.1764382123947144, "learning_rate": 0.00019998124910461376, "loss": 4.8815, "step": 35050 }, { "epoch": 0.06550050268370115, "grad_norm": 1.3247041702270508, "learning_rate": 0.00019998116766661527, "loss": 4.7398, "step": 35100 }, { "epoch": 0.06559380824307964, "grad_norm": 1.155849575996399, "learning_rate": 0.00019998108605216762, "loss": 4.7499, "step": 35150 }, { "epoch": 0.06568711380245813, "grad_norm": 1.5121906995773315, "learning_rate": 0.000199981004261271, "loss": 4.8656, "step": 35200 }, { "epoch": 0.06578041936183662, "grad_norm": 1.2283687591552734, "learning_rate": 0.00019998092229392544, "loss": 5.0256, "step": 35250 }, { "epoch": 0.06587372492121511, "grad_norm": 1.1128344535827637, "learning_rate": 0.00019998084015013115, "loss": 5.0056, "step": 35300 }, { "epoch": 0.06596703048059362, "grad_norm": 0.7951078414916992, "learning_rate": 0.0001999807578298883, "loss": 4.8792, "step": 35350 }, { "epoch": 0.0660603360399721, "grad_norm": 1.197862148284912, "learning_rate": 0.000199980675333197, "loss": 4.7815, "step": 35400 }, { "epoch": 0.0661536415993506, "grad_norm": 1.3143230676651, "learning_rate": 0.0001999805926600574, "loss": 4.8473, "step": 35450 }, { "epoch": 0.06624694715872909, "grad_norm": 1.100462555885315, "learning_rate": 0.00019998050981046963, "loss": 5.1148, "step": 35500 }, { "epoch": 0.06634025271810758, "grad_norm": 0.8429204225540161, "learning_rate": 0.00019998042678443385, "loss": 5.0559, "step": 35550 }, { "epoch": 0.06643355827748607, "grad_norm": 1.61745023727417, "learning_rate": 0.0001999803435819502, "loss": 4.8, "step": 35600 }, { "epoch": 0.06652686383686456, "grad_norm": 1.1721575260162354, "learning_rate": 0.00019998026020301887, "loss": 4.746, "step": 35650 }, { "epoch": 0.06662016939624305, "grad_norm": 1.6032483577728271, "learning_rate": 0.00019998017664763995, "loss": 5.0657, "step": 35700 }, { "epoch": 0.06671347495562155, "grad_norm": 1.3949735164642334, "learning_rate": 0.0001999800929158136, "loss": 4.8643, "step": 35750 }, { "epoch": 0.06680678051500004, "grad_norm": 1.1598511934280396, "learning_rate": 0.00019998000900754001, "loss": 4.8064, "step": 35800 }, { "epoch": 0.06690008607437853, "grad_norm": 1.3976051807403564, "learning_rate": 0.00019997992492281927, "loss": 4.9548, "step": 35850 }, { "epoch": 0.06699339163375702, "grad_norm": 1.2537288665771484, "learning_rate": 0.0001999798406616516, "loss": 5.0992, "step": 35900 }, { "epoch": 0.06708669719313551, "grad_norm": 0.9580097198486328, "learning_rate": 0.00019997975622403704, "loss": 4.7681, "step": 35950 }, { "epoch": 0.067180002752514, "grad_norm": 1.2984302043914795, "learning_rate": 0.0001999796716099758, "loss": 4.7869, "step": 36000 }, { "epoch": 0.067180002752514, "eval_loss": 5.1954779624938965, "eval_runtime": 233.7952, "eval_samples_per_second": 11.155, "eval_steps_per_second": 11.155, "eval_tts_loss": 7.21724026617337, "step": 36000 }, { "epoch": 0.06727330831189249, "grad_norm": 1.1834443807601929, "learning_rate": 0.00019997958681946808, "loss": 4.9615, "step": 36050 }, { "epoch": 0.06736661387127099, "grad_norm": 1.4487770795822144, "learning_rate": 0.00019997950185251398, "loss": 4.9015, "step": 36100 }, { "epoch": 0.06745991943064948, "grad_norm": 1.003225326538086, "learning_rate": 0.00019997941670911362, "loss": 5.1782, "step": 36150 }, { "epoch": 0.06755322499002797, "grad_norm": 1.2664480209350586, "learning_rate": 0.0001999793313892672, "loss": 4.8188, "step": 36200 }, { "epoch": 0.06764653054940646, "grad_norm": 0.9945356845855713, "learning_rate": 0.00019997924589297482, "loss": 5.017, "step": 36250 }, { "epoch": 0.06773983610878495, "grad_norm": 1.232276201248169, "learning_rate": 0.00019997916022023674, "loss": 4.8427, "step": 36300 }, { "epoch": 0.06783314166816344, "grad_norm": 1.5773524045944214, "learning_rate": 0.00019997907437105297, "loss": 4.8048, "step": 36350 }, { "epoch": 0.06792644722754193, "grad_norm": 1.3550626039505005, "learning_rate": 0.00019997898834542372, "loss": 4.9573, "step": 36400 }, { "epoch": 0.06801975278692042, "grad_norm": 1.5464149713516235, "learning_rate": 0.0001999789021433492, "loss": 4.9219, "step": 36450 }, { "epoch": 0.06811305834629892, "grad_norm": 1.3397859334945679, "learning_rate": 0.00019997881576482945, "loss": 5.1099, "step": 36500 }, { "epoch": 0.06820636390567741, "grad_norm": 1.3835015296936035, "learning_rate": 0.0001999787292098647, "loss": 4.9419, "step": 36550 }, { "epoch": 0.0682996694650559, "grad_norm": 0.9244956374168396, "learning_rate": 0.0001999786424784551, "loss": 5.0481, "step": 36600 }, { "epoch": 0.0683929750244344, "grad_norm": 1.0417383909225464, "learning_rate": 0.00019997855557060077, "loss": 4.8458, "step": 36650 }, { "epoch": 0.06848628058381288, "grad_norm": 1.2598869800567627, "learning_rate": 0.0001999784684863019, "loss": 4.8585, "step": 36700 }, { "epoch": 0.06857958614319137, "grad_norm": 1.0697765350341797, "learning_rate": 0.00019997838122555862, "loss": 4.7396, "step": 36750 }, { "epoch": 0.06867289170256986, "grad_norm": 1.1050491333007812, "learning_rate": 0.0001999782937883711, "loss": 4.8543, "step": 36800 }, { "epoch": 0.06876619726194837, "grad_norm": 1.0602041482925415, "learning_rate": 0.00019997820617473946, "loss": 4.8054, "step": 36850 }, { "epoch": 0.06885950282132686, "grad_norm": 1.0901793241500854, "learning_rate": 0.0001999781183846639, "loss": 4.943, "step": 36900 }, { "epoch": 0.06895280838070535, "grad_norm": 1.1608269214630127, "learning_rate": 0.00019997803041814453, "loss": 4.8669, "step": 36950 }, { "epoch": 0.06904611394008384, "grad_norm": 1.428856611251831, "learning_rate": 0.00019997794227518152, "loss": 4.9257, "step": 37000 }, { "epoch": 0.06913941949946233, "grad_norm": 1.2141627073287964, "learning_rate": 0.00019997785395577507, "loss": 4.8884, "step": 37050 }, { "epoch": 0.06923272505884082, "grad_norm": 0.917578935623169, "learning_rate": 0.00019997776545992526, "loss": 4.7185, "step": 37100 }, { "epoch": 0.0693260306182193, "grad_norm": 1.1432592868804932, "learning_rate": 0.00019997767678763234, "loss": 5.1003, "step": 37150 }, { "epoch": 0.0694193361775978, "grad_norm": 1.5270577669143677, "learning_rate": 0.00019997758793889635, "loss": 4.7651, "step": 37200 }, { "epoch": 0.0695126417369763, "grad_norm": 0.8794492483139038, "learning_rate": 0.00019997749891371752, "loss": 4.9522, "step": 37250 }, { "epoch": 0.06960594729635479, "grad_norm": 1.5430872440338135, "learning_rate": 0.000199977409712096, "loss": 4.8259, "step": 37300 }, { "epoch": 0.06969925285573328, "grad_norm": 1.2974966764450073, "learning_rate": 0.000199977320334032, "loss": 5.0415, "step": 37350 }, { "epoch": 0.06979255841511177, "grad_norm": 1.3754868507385254, "learning_rate": 0.00019997723077952556, "loss": 4.9111, "step": 37400 }, { "epoch": 0.06988586397449026, "grad_norm": 1.1165885925292969, "learning_rate": 0.00019997714104857693, "loss": 5.0136, "step": 37450 }, { "epoch": 0.06997916953386875, "grad_norm": 1.127763032913208, "learning_rate": 0.0001999770511411862, "loss": 4.6376, "step": 37500 }, { "epoch": 0.07007247509324724, "grad_norm": 1.1348410844802856, "learning_rate": 0.00019997696105735358, "loss": 4.7574, "step": 37550 }, { "epoch": 0.07016578065262573, "grad_norm": 1.012832522392273, "learning_rate": 0.00019997687079707926, "loss": 4.9673, "step": 37600 }, { "epoch": 0.07025908621200423, "grad_norm": 0.8529366254806519, "learning_rate": 0.0001999767803603633, "loss": 4.9508, "step": 37650 }, { "epoch": 0.07035239177138272, "grad_norm": 1.2480065822601318, "learning_rate": 0.00019997668974720595, "loss": 4.8772, "step": 37700 }, { "epoch": 0.07044569733076121, "grad_norm": 1.318292498588562, "learning_rate": 0.00019997659895760732, "loss": 4.8596, "step": 37750 }, { "epoch": 0.0705390028901397, "grad_norm": 1.3873937129974365, "learning_rate": 0.0001999765079915676, "loss": 4.8429, "step": 37800 }, { "epoch": 0.07063230844951819, "grad_norm": 1.1860456466674805, "learning_rate": 0.0001999764168490869, "loss": 4.8605, "step": 37850 }, { "epoch": 0.07072561400889668, "grad_norm": 1.0608043670654297, "learning_rate": 0.00019997632553016544, "loss": 5.0347, "step": 37900 }, { "epoch": 0.07081891956827517, "grad_norm": 1.168198585510254, "learning_rate": 0.00019997623403480335, "loss": 4.7621, "step": 37950 }, { "epoch": 0.07091222512765367, "grad_norm": 0.9395182132720947, "learning_rate": 0.0001999761423630008, "loss": 4.9286, "step": 38000 }, { "epoch": 0.07100553068703216, "grad_norm": 1.2564831972122192, "learning_rate": 0.00019997605051475794, "loss": 4.9218, "step": 38050 }, { "epoch": 0.07109883624641065, "grad_norm": 1.497332215309143, "learning_rate": 0.00019997595849007496, "loss": 5.0628, "step": 38100 }, { "epoch": 0.07119214180578914, "grad_norm": 1.2820568084716797, "learning_rate": 0.00019997586628895197, "loss": 4.8333, "step": 38150 }, { "epoch": 0.07128544736516763, "grad_norm": 1.5082170963287354, "learning_rate": 0.0001999757739113892, "loss": 5.0072, "step": 38200 }, { "epoch": 0.07137875292454612, "grad_norm": 1.0686591863632202, "learning_rate": 0.00019997568135738677, "loss": 4.8872, "step": 38250 }, { "epoch": 0.07147205848392461, "grad_norm": 1.14346444606781, "learning_rate": 0.00019997558862694484, "loss": 4.8167, "step": 38300 }, { "epoch": 0.0715653640433031, "grad_norm": 0.6724060773849487, "learning_rate": 0.00019997549572006365, "loss": 4.8189, "step": 38350 }, { "epoch": 0.07165866960268161, "grad_norm": 1.5035301446914673, "learning_rate": 0.00019997540263674322, "loss": 4.9308, "step": 38400 }, { "epoch": 0.0717519751620601, "grad_norm": 1.1256290674209595, "learning_rate": 0.00019997530937698382, "loss": 4.8645, "step": 38450 }, { "epoch": 0.07184528072143859, "grad_norm": 0.9030829668045044, "learning_rate": 0.00019997521594078562, "loss": 4.9792, "step": 38500 }, { "epoch": 0.07193858628081708, "grad_norm": 1.1105825901031494, "learning_rate": 0.00019997512232814875, "loss": 4.9466, "step": 38550 }, { "epoch": 0.07203189184019557, "grad_norm": 1.272040843963623, "learning_rate": 0.00019997502853907336, "loss": 4.9014, "step": 38600 }, { "epoch": 0.07212519739957406, "grad_norm": 1.1471481323242188, "learning_rate": 0.00019997493457355963, "loss": 4.9466, "step": 38650 }, { "epoch": 0.07221850295895255, "grad_norm": 1.134016990661621, "learning_rate": 0.00019997484043160775, "loss": 4.8908, "step": 38700 }, { "epoch": 0.07231180851833105, "grad_norm": 1.214316964149475, "learning_rate": 0.00019997474611321787, "loss": 4.9729, "step": 38750 }, { "epoch": 0.07240511407770954, "grad_norm": 0.9872856140136719, "learning_rate": 0.00019997465161839014, "loss": 4.9633, "step": 38800 }, { "epoch": 0.07249841963708803, "grad_norm": 1.3392361402511597, "learning_rate": 0.00019997455694712474, "loss": 5.0015, "step": 38850 }, { "epoch": 0.07259172519646652, "grad_norm": 1.0813117027282715, "learning_rate": 0.00019997446209942186, "loss": 5.0432, "step": 38900 }, { "epoch": 0.07268503075584501, "grad_norm": 1.2037447690963745, "learning_rate": 0.0001999743670752816, "loss": 4.8759, "step": 38950 }, { "epoch": 0.0727783363152235, "grad_norm": 1.0193177461624146, "learning_rate": 0.00019997427187470423, "loss": 4.9456, "step": 39000 }, { "epoch": 0.0727783363152235, "eval_loss": 5.16133975982666, "eval_runtime": 230.6318, "eval_samples_per_second": 11.308, "eval_steps_per_second": 11.308, "eval_tts_loss": 7.213367073268178, "step": 39000 }, { "epoch": 0.07287164187460199, "grad_norm": 1.2097446918487549, "learning_rate": 0.00019997417649768982, "loss": 5.0498, "step": 39050 }, { "epoch": 0.07296494743398048, "grad_norm": 1.0760691165924072, "learning_rate": 0.00019997408094423862, "loss": 4.6899, "step": 39100 }, { "epoch": 0.07305825299335898, "grad_norm": 1.1207256317138672, "learning_rate": 0.0001999739852143507, "loss": 4.9026, "step": 39150 }, { "epoch": 0.07315155855273747, "grad_norm": 0.801072359085083, "learning_rate": 0.00019997388930802635, "loss": 4.8494, "step": 39200 }, { "epoch": 0.07324486411211596, "grad_norm": 1.1047637462615967, "learning_rate": 0.00019997379322526563, "loss": 4.7812, "step": 39250 }, { "epoch": 0.07333816967149445, "grad_norm": 1.152491807937622, "learning_rate": 0.00019997369696606878, "loss": 4.9095, "step": 39300 }, { "epoch": 0.07343147523087294, "grad_norm": 1.1126306056976318, "learning_rate": 0.00019997360053043592, "loss": 4.9798, "step": 39350 }, { "epoch": 0.07352478079025143, "grad_norm": 1.156316876411438, "learning_rate": 0.00019997350391836726, "loss": 4.9773, "step": 39400 }, { "epoch": 0.07361808634962992, "grad_norm": 1.0668503046035767, "learning_rate": 0.00019997340712986296, "loss": 5.1012, "step": 39450 }, { "epoch": 0.07371139190900843, "grad_norm": 1.3585624694824219, "learning_rate": 0.0001999733101649232, "loss": 4.6893, "step": 39500 }, { "epoch": 0.07380469746838692, "grad_norm": 1.5277858972549438, "learning_rate": 0.00019997321302354812, "loss": 4.8931, "step": 39550 }, { "epoch": 0.0738980030277654, "grad_norm": 1.2020436525344849, "learning_rate": 0.00019997311570573792, "loss": 5.1182, "step": 39600 }, { "epoch": 0.0739913085871439, "grad_norm": 1.1372538805007935, "learning_rate": 0.00019997301821149274, "loss": 4.9238, "step": 39650 }, { "epoch": 0.07408461414652238, "grad_norm": 0.7125515341758728, "learning_rate": 0.0001999729205408128, "loss": 4.8202, "step": 39700 }, { "epoch": 0.07417791970590087, "grad_norm": 0.9166264533996582, "learning_rate": 0.00019997282269369824, "loss": 4.8832, "step": 39750 }, { "epoch": 0.07427122526527936, "grad_norm": 1.099700927734375, "learning_rate": 0.00019997272467014922, "loss": 4.9073, "step": 39800 }, { "epoch": 0.07436453082465785, "grad_norm": 1.3260917663574219, "learning_rate": 0.00019997262647016597, "loss": 4.8271, "step": 39850 }, { "epoch": 0.07445783638403636, "grad_norm": 1.2185454368591309, "learning_rate": 0.0001999725280937486, "loss": 4.8525, "step": 39900 }, { "epoch": 0.07455114194341485, "grad_norm": 1.0137861967086792, "learning_rate": 0.0001999724295408973, "loss": 4.749, "step": 39950 }, { "epoch": 0.07464444750279334, "grad_norm": 0.8680387139320374, "learning_rate": 0.00019997233081161226, "loss": 5.0129, "step": 40000 }, { "epoch": 0.07473775306217183, "grad_norm": 1.314736008644104, "learning_rate": 0.00019997223190589366, "loss": 5.0179, "step": 40050 }, { "epoch": 0.07483105862155032, "grad_norm": 1.209912657737732, "learning_rate": 0.00019997213282374165, "loss": 5.0474, "step": 40100 }, { "epoch": 0.0749243641809288, "grad_norm": 0.9931349158287048, "learning_rate": 0.0001999720335651564, "loss": 4.8296, "step": 40150 }, { "epoch": 0.0750176697403073, "grad_norm": 0.9877680540084839, "learning_rate": 0.00019997193413013814, "loss": 4.7287, "step": 40200 }, { "epoch": 0.0751109752996858, "grad_norm": 1.2788828611373901, "learning_rate": 0.00019997183451868697, "loss": 4.7418, "step": 40250 }, { "epoch": 0.07520428085906429, "grad_norm": 1.1528019905090332, "learning_rate": 0.00019997173473080312, "loss": 4.7367, "step": 40300 }, { "epoch": 0.07529758641844278, "grad_norm": 1.3090732097625732, "learning_rate": 0.00019997163476648676, "loss": 4.8197, "step": 40350 }, { "epoch": 0.07539089197782127, "grad_norm": 1.2664506435394287, "learning_rate": 0.00019997153462573804, "loss": 4.6739, "step": 40400 }, { "epoch": 0.07548419753719976, "grad_norm": 1.1950277090072632, "learning_rate": 0.00019997143430855715, "loss": 5.007, "step": 40450 }, { "epoch": 0.07557750309657825, "grad_norm": 0.9949437975883484, "learning_rate": 0.00019997133381494426, "loss": 4.9069, "step": 40500 }, { "epoch": 0.07567080865595674, "grad_norm": 1.1204107999801636, "learning_rate": 0.00019997123314489956, "loss": 4.8598, "step": 40550 }, { "epoch": 0.07576411421533523, "grad_norm": 1.2911077737808228, "learning_rate": 0.00019997113229842323, "loss": 4.9724, "step": 40600 }, { "epoch": 0.07585741977471373, "grad_norm": 0.9205583930015564, "learning_rate": 0.00019997103127551548, "loss": 4.8516, "step": 40650 }, { "epoch": 0.07595072533409222, "grad_norm": 1.0593533515930176, "learning_rate": 0.0001999709300761764, "loss": 4.8309, "step": 40700 }, { "epoch": 0.07604403089347071, "grad_norm": 1.1807661056518555, "learning_rate": 0.00019997082870040625, "loss": 4.9056, "step": 40750 }, { "epoch": 0.0761373364528492, "grad_norm": 1.105985164642334, "learning_rate": 0.00019997072714820514, "loss": 4.8024, "step": 40800 }, { "epoch": 0.07623064201222769, "grad_norm": 1.2069274187088013, "learning_rate": 0.00019997062541957333, "loss": 4.9473, "step": 40850 }, { "epoch": 0.07632394757160618, "grad_norm": 1.3253543376922607, "learning_rate": 0.00019997052351451093, "loss": 4.9783, "step": 40900 }, { "epoch": 0.07641725313098467, "grad_norm": 1.22372567653656, "learning_rate": 0.00019997042143301815, "loss": 4.9517, "step": 40950 }, { "epoch": 0.07651055869036316, "grad_norm": 1.091848611831665, "learning_rate": 0.00019997031917509515, "loss": 4.7118, "step": 41000 }, { "epoch": 0.07660386424974167, "grad_norm": 1.0672491788864136, "learning_rate": 0.00019997021674074213, "loss": 4.9462, "step": 41050 }, { "epoch": 0.07669716980912016, "grad_norm": 1.0972808599472046, "learning_rate": 0.0001999701141299593, "loss": 4.8651, "step": 41100 }, { "epoch": 0.07679047536849865, "grad_norm": 1.0306609869003296, "learning_rate": 0.0001999700113427468, "loss": 4.9256, "step": 41150 }, { "epoch": 0.07688378092787714, "grad_norm": 1.1696428060531616, "learning_rate": 0.0001999699083791048, "loss": 4.7898, "step": 41200 }, { "epoch": 0.07697708648725562, "grad_norm": 1.2328746318817139, "learning_rate": 0.00019996980523903353, "loss": 4.942, "step": 41250 }, { "epoch": 0.07707039204663411, "grad_norm": 1.0950030088424683, "learning_rate": 0.00019996970192253313, "loss": 4.8404, "step": 41300 }, { "epoch": 0.0771636976060126, "grad_norm": 1.0388070344924927, "learning_rate": 0.00019996959842960379, "loss": 5.0406, "step": 41350 }, { "epoch": 0.07725700316539111, "grad_norm": 1.3518260717391968, "learning_rate": 0.0001999694947602457, "loss": 4.7539, "step": 41400 }, { "epoch": 0.0773503087247696, "grad_norm": 0.8718711733818054, "learning_rate": 0.00019996939091445904, "loss": 4.8804, "step": 41450 }, { "epoch": 0.07744361428414809, "grad_norm": 1.1106172800064087, "learning_rate": 0.00019996928689224402, "loss": 4.9871, "step": 41500 }, { "epoch": 0.07753691984352658, "grad_norm": 1.3443810939788818, "learning_rate": 0.00019996918269360076, "loss": 4.7499, "step": 41550 }, { "epoch": 0.07763022540290507, "grad_norm": 0.968607485294342, "learning_rate": 0.00019996907831852951, "loss": 4.7705, "step": 41600 }, { "epoch": 0.07772353096228356, "grad_norm": 1.1009609699249268, "learning_rate": 0.0001999689737670304, "loss": 4.8043, "step": 41650 }, { "epoch": 0.07781683652166205, "grad_norm": 0.9938008785247803, "learning_rate": 0.00019996886903910367, "loss": 4.9927, "step": 41700 }, { "epoch": 0.07791014208104054, "grad_norm": 1.0750339031219482, "learning_rate": 0.00019996876413474947, "loss": 4.9449, "step": 41750 }, { "epoch": 0.07800344764041904, "grad_norm": 1.2172019481658936, "learning_rate": 0.000199968659053968, "loss": 5.0049, "step": 41800 }, { "epoch": 0.07809675319979753, "grad_norm": 1.288533329963684, "learning_rate": 0.0001999685537967594, "loss": 4.6487, "step": 41850 }, { "epoch": 0.07819005875917602, "grad_norm": 1.087883710861206, "learning_rate": 0.00019996844836312394, "loss": 4.8504, "step": 41900 }, { "epoch": 0.07828336431855451, "grad_norm": 1.2223601341247559, "learning_rate": 0.00019996834275306174, "loss": 4.8588, "step": 41950 }, { "epoch": 0.078376669877933, "grad_norm": 0.8469075560569763, "learning_rate": 0.00019996823696657299, "loss": 4.9408, "step": 42000 }, { "epoch": 0.078376669877933, "eval_loss": 5.132350444793701, "eval_runtime": 230.8355, "eval_samples_per_second": 11.298, "eval_steps_per_second": 11.298, "eval_tts_loss": 7.264853509248697, "step": 42000 }, { "epoch": 0.07846997543731149, "grad_norm": 1.1458607912063599, "learning_rate": 0.0001999681310036579, "loss": 4.9473, "step": 42050 }, { "epoch": 0.07856328099668998, "grad_norm": 0.9783209562301636, "learning_rate": 0.00019996802486431665, "loss": 4.9335, "step": 42100 }, { "epoch": 0.07865658655606848, "grad_norm": 1.3118724822998047, "learning_rate": 0.00019996791854854943, "loss": 4.9134, "step": 42150 }, { "epoch": 0.07874989211544697, "grad_norm": 1.4106804132461548, "learning_rate": 0.0001999678120563564, "loss": 4.9419, "step": 42200 }, { "epoch": 0.07884319767482546, "grad_norm": 1.3338770866394043, "learning_rate": 0.0001999677053877378, "loss": 4.9377, "step": 42250 }, { "epoch": 0.07893650323420395, "grad_norm": 1.1981934309005737, "learning_rate": 0.00019996759854269377, "loss": 4.9804, "step": 42300 }, { "epoch": 0.07902980879358244, "grad_norm": 1.1797490119934082, "learning_rate": 0.00019996749152122455, "loss": 4.86, "step": 42350 }, { "epoch": 0.07912311435296093, "grad_norm": 0.7988554239273071, "learning_rate": 0.0001999673843233303, "loss": 5.0499, "step": 42400 }, { "epoch": 0.07921641991233942, "grad_norm": 0.9866506457328796, "learning_rate": 0.00019996727694901117, "loss": 4.8936, "step": 42450 }, { "epoch": 0.07930972547171791, "grad_norm": 1.507678747177124, "learning_rate": 0.0001999671693982674, "loss": 4.7716, "step": 42500 }, { "epoch": 0.07940303103109642, "grad_norm": 1.2090741395950317, "learning_rate": 0.00019996706167109917, "loss": 4.7974, "step": 42550 }, { "epoch": 0.0794963365904749, "grad_norm": 0.9674336314201355, "learning_rate": 0.00019996695376750667, "loss": 4.8457, "step": 42600 }, { "epoch": 0.0795896421498534, "grad_norm": 0.8843348622322083, "learning_rate": 0.00019996684568749008, "loss": 4.8981, "step": 42650 }, { "epoch": 0.07968294770923189, "grad_norm": 1.0419225692749023, "learning_rate": 0.00019996673743104956, "loss": 4.6177, "step": 42700 }, { "epoch": 0.07977625326861038, "grad_norm": 0.8469592928886414, "learning_rate": 0.0001999666289981854, "loss": 4.8767, "step": 42750 }, { "epoch": 0.07986955882798887, "grad_norm": 0.8583770394325256, "learning_rate": 0.00019996652038889768, "loss": 4.9546, "step": 42800 }, { "epoch": 0.07996286438736736, "grad_norm": 0.9213618636131287, "learning_rate": 0.00019996641160318666, "loss": 4.5851, "step": 42850 }, { "epoch": 0.08005616994674586, "grad_norm": 1.247279405593872, "learning_rate": 0.00019996630264105253, "loss": 4.8712, "step": 42900 }, { "epoch": 0.08014947550612435, "grad_norm": 1.3103673458099365, "learning_rate": 0.00019996619350249546, "loss": 4.9501, "step": 42950 }, { "epoch": 0.08024278106550284, "grad_norm": 0.9502874612808228, "learning_rate": 0.00019996608418751564, "loss": 4.9463, "step": 43000 }, { "epoch": 0.08033608662488133, "grad_norm": 1.1887316703796387, "learning_rate": 0.0001999659746961133, "loss": 4.7633, "step": 43050 }, { "epoch": 0.08042939218425982, "grad_norm": 1.132875680923462, "learning_rate": 0.00019996586502828856, "loss": 4.7873, "step": 43100 }, { "epoch": 0.08052269774363831, "grad_norm": 1.1734470129013062, "learning_rate": 0.00019996575518404166, "loss": 4.9746, "step": 43150 }, { "epoch": 0.0806160033030168, "grad_norm": 1.2493457794189453, "learning_rate": 0.0001999656451633728, "loss": 4.7152, "step": 43200 }, { "epoch": 0.08070930886239529, "grad_norm": 1.2277987003326416, "learning_rate": 0.00019996553496628216, "loss": 4.8031, "step": 43250 }, { "epoch": 0.08080261442177379, "grad_norm": 0.9987390041351318, "learning_rate": 0.00019996542459276996, "loss": 4.6355, "step": 43300 }, { "epoch": 0.08089591998115228, "grad_norm": 0.8293343186378479, "learning_rate": 0.0001999653140428364, "loss": 4.845, "step": 43350 }, { "epoch": 0.08098922554053077, "grad_norm": 1.0015592575073242, "learning_rate": 0.0001999652033164816, "loss": 4.8523, "step": 43400 }, { "epoch": 0.08108253109990926, "grad_norm": 1.1842169761657715, "learning_rate": 0.0001999650924137058, "loss": 5.0227, "step": 43450 }, { "epoch": 0.08117583665928775, "grad_norm": 1.0604639053344727, "learning_rate": 0.00019996498133450924, "loss": 4.9266, "step": 43500 }, { "epoch": 0.08126914221866624, "grad_norm": 1.1228868961334229, "learning_rate": 0.00019996487007889206, "loss": 4.9329, "step": 43550 }, { "epoch": 0.08136244777804473, "grad_norm": 1.149283766746521, "learning_rate": 0.0001999647586468545, "loss": 4.6841, "step": 43600 }, { "epoch": 0.08145575333742322, "grad_norm": 1.1383576393127441, "learning_rate": 0.00019996464703839667, "loss": 4.9698, "step": 43650 }, { "epoch": 0.08154905889680172, "grad_norm": 1.0777636766433716, "learning_rate": 0.00019996453525351887, "loss": 4.7305, "step": 43700 }, { "epoch": 0.08164236445618021, "grad_norm": 1.4025020599365234, "learning_rate": 0.00019996442329222125, "loss": 4.9614, "step": 43750 }, { "epoch": 0.0817356700155587, "grad_norm": 1.1779139041900635, "learning_rate": 0.00019996431115450403, "loss": 4.8422, "step": 43800 }, { "epoch": 0.0818289755749372, "grad_norm": 0.9133360981941223, "learning_rate": 0.00019996419884036733, "loss": 5.0592, "step": 43850 }, { "epoch": 0.08192228113431568, "grad_norm": 1.0266224145889282, "learning_rate": 0.00019996408634981148, "loss": 4.9319, "step": 43900 }, { "epoch": 0.08201558669369417, "grad_norm": 0.882277250289917, "learning_rate": 0.00019996397368283656, "loss": 4.8332, "step": 43950 }, { "epoch": 0.08210889225307266, "grad_norm": 1.1271204948425293, "learning_rate": 0.00019996386083944283, "loss": 5.1547, "step": 44000 }, { "epoch": 0.08220219781245117, "grad_norm": 1.0688495635986328, "learning_rate": 0.0001999637478196305, "loss": 5.0013, "step": 44050 }, { "epoch": 0.08229550337182966, "grad_norm": 1.0313043594360352, "learning_rate": 0.0001999636346233997, "loss": 4.9081, "step": 44100 }, { "epoch": 0.08238880893120815, "grad_norm": 0.9132486581802368, "learning_rate": 0.00019996352125075067, "loss": 4.5762, "step": 44150 }, { "epoch": 0.08248211449058664, "grad_norm": 1.090657353401184, "learning_rate": 0.00019996340770168363, "loss": 4.9538, "step": 44200 }, { "epoch": 0.08257542004996513, "grad_norm": 1.004305124282837, "learning_rate": 0.0001999632939761988, "loss": 4.9216, "step": 44250 }, { "epoch": 0.08266872560934362, "grad_norm": 1.309733510017395, "learning_rate": 0.00019996318007429632, "loss": 4.8428, "step": 44300 }, { "epoch": 0.0827620311687221, "grad_norm": 1.159638524055481, "learning_rate": 0.0001999630659959764, "loss": 4.9621, "step": 44350 }, { "epoch": 0.0828553367281006, "grad_norm": 1.147161602973938, "learning_rate": 0.00019996295174123927, "loss": 4.7534, "step": 44400 }, { "epoch": 0.0829486422874791, "grad_norm": 0.9718197584152222, "learning_rate": 0.00019996283731008513, "loss": 4.8755, "step": 44450 }, { "epoch": 0.08304194784685759, "grad_norm": 0.9277188181877136, "learning_rate": 0.00019996272270251416, "loss": 4.8344, "step": 44500 }, { "epoch": 0.08313525340623608, "grad_norm": 1.071356177330017, "learning_rate": 0.00019996260791852656, "loss": 4.8703, "step": 44550 }, { "epoch": 0.08322855896561457, "grad_norm": 1.029938817024231, "learning_rate": 0.00019996249295812257, "loss": 4.8064, "step": 44600 }, { "epoch": 0.08332186452499306, "grad_norm": 1.397524356842041, "learning_rate": 0.00019996237782130236, "loss": 4.8744, "step": 44650 }, { "epoch": 0.08341517008437155, "grad_norm": 1.0888935327529907, "learning_rate": 0.00019996226250806615, "loss": 4.8386, "step": 44700 }, { "epoch": 0.08350847564375004, "grad_norm": 0.727277934551239, "learning_rate": 0.0001999621470184141, "loss": 4.8482, "step": 44750 }, { "epoch": 0.08360178120312854, "grad_norm": 1.148036241531372, "learning_rate": 0.00019996203135234648, "loss": 4.8495, "step": 44800 }, { "epoch": 0.08369508676250703, "grad_norm": 1.073508858680725, "learning_rate": 0.00019996191550986347, "loss": 4.8289, "step": 44850 }, { "epoch": 0.08378839232188552, "grad_norm": 0.7582327723503113, "learning_rate": 0.00019996179949096528, "loss": 4.7814, "step": 44900 }, { "epoch": 0.08388169788126401, "grad_norm": 0.9114712476730347, "learning_rate": 0.00019996168329565207, "loss": 4.8783, "step": 44950 }, { "epoch": 0.0839750034406425, "grad_norm": 1.14763343334198, "learning_rate": 0.00019996156692392408, "loss": 4.6919, "step": 45000 }, { "epoch": 0.0839750034406425, "eval_loss": 5.113004684448242, "eval_runtime": 231.8555, "eval_samples_per_second": 11.248, "eval_steps_per_second": 11.248, "eval_tts_loss": 7.2781233883782805, "step": 45000 }, { "epoch": 0.08406830900002099, "grad_norm": 1.0536847114562988, "learning_rate": 0.00019996145037578153, "loss": 4.741, "step": 45050 }, { "epoch": 0.08416161455939948, "grad_norm": 1.0102460384368896, "learning_rate": 0.0001999613336512246, "loss": 4.9963, "step": 45100 }, { "epoch": 0.08425492011877797, "grad_norm": 1.3230600357055664, "learning_rate": 0.0001999612167502535, "loss": 4.8845, "step": 45150 }, { "epoch": 0.08434822567815647, "grad_norm": 1.181218147277832, "learning_rate": 0.00019996109967286845, "loss": 4.7497, "step": 45200 }, { "epoch": 0.08444153123753496, "grad_norm": 1.2599127292633057, "learning_rate": 0.00019996098241906963, "loss": 4.9527, "step": 45250 }, { "epoch": 0.08453483679691345, "grad_norm": 1.0429056882858276, "learning_rate": 0.0001999608649888573, "loss": 4.8438, "step": 45300 }, { "epoch": 0.08462814235629194, "grad_norm": 0.9174293279647827, "learning_rate": 0.00019996074738223157, "loss": 4.885, "step": 45350 }, { "epoch": 0.08472144791567043, "grad_norm": 0.7659981846809387, "learning_rate": 0.00019996062959919277, "loss": 4.8716, "step": 45400 }, { "epoch": 0.08481475347504892, "grad_norm": 1.2928154468536377, "learning_rate": 0.00019996051163974102, "loss": 4.9047, "step": 45450 }, { "epoch": 0.08490805903442741, "grad_norm": 1.145384669303894, "learning_rate": 0.00019996039350387657, "loss": 4.9983, "step": 45500 }, { "epoch": 0.08500136459380592, "grad_norm": 1.0447450876235962, "learning_rate": 0.0001999602751915996, "loss": 4.7498, "step": 45550 }, { "epoch": 0.0850946701531844, "grad_norm": 1.108702301979065, "learning_rate": 0.00019996015670291036, "loss": 5.0015, "step": 45600 }, { "epoch": 0.0851879757125629, "grad_norm": 1.2329564094543457, "learning_rate": 0.00019996003803780902, "loss": 4.7922, "step": 45650 }, { "epoch": 0.08528128127194139, "grad_norm": 1.1614669561386108, "learning_rate": 0.00019995991919629578, "loss": 5.0659, "step": 45700 }, { "epoch": 0.08537458683131988, "grad_norm": 1.1069921255111694, "learning_rate": 0.00019995980017837087, "loss": 5.134, "step": 45750 }, { "epoch": 0.08546789239069837, "grad_norm": 1.1076229810714722, "learning_rate": 0.00019995968098403453, "loss": 4.7386, "step": 45800 }, { "epoch": 0.08556119795007686, "grad_norm": 0.8227840662002563, "learning_rate": 0.00019995956161328692, "loss": 4.854, "step": 45850 }, { "epoch": 0.08565450350945535, "grad_norm": 0.9372345805168152, "learning_rate": 0.0001999594420661283, "loss": 4.8924, "step": 45900 }, { "epoch": 0.08574780906883385, "grad_norm": 1.0700907707214355, "learning_rate": 0.00019995932234255883, "loss": 4.7795, "step": 45950 }, { "epoch": 0.08584111462821234, "grad_norm": 1.173011064529419, "learning_rate": 0.00019995920244257872, "loss": 4.8779, "step": 46000 }, { "epoch": 0.08593442018759083, "grad_norm": 1.0453685522079468, "learning_rate": 0.00019995908236618824, "loss": 4.7893, "step": 46050 }, { "epoch": 0.08602772574696932, "grad_norm": 1.3147802352905273, "learning_rate": 0.00019995896211338757, "loss": 4.8669, "step": 46100 }, { "epoch": 0.08612103130634781, "grad_norm": 1.0655571222305298, "learning_rate": 0.00019995884168417692, "loss": 4.8534, "step": 46150 }, { "epoch": 0.0862143368657263, "grad_norm": 1.1845111846923828, "learning_rate": 0.00019995872107855648, "loss": 4.8177, "step": 46200 }, { "epoch": 0.08630764242510479, "grad_norm": 1.0939445495605469, "learning_rate": 0.0001999586002965265, "loss": 4.7689, "step": 46250 }, { "epoch": 0.08640094798448329, "grad_norm": 1.0770061016082764, "learning_rate": 0.00019995847933808718, "loss": 4.7746, "step": 46300 }, { "epoch": 0.08649425354386178, "grad_norm": 1.2694488763809204, "learning_rate": 0.00019995835820323872, "loss": 4.7283, "step": 46350 }, { "epoch": 0.08658755910324027, "grad_norm": 1.181217908859253, "learning_rate": 0.00019995823689198137, "loss": 4.9545, "step": 46400 }, { "epoch": 0.08668086466261876, "grad_norm": 1.1785484552383423, "learning_rate": 0.0001999581154043153, "loss": 4.7946, "step": 46450 }, { "epoch": 0.08677417022199725, "grad_norm": 0.8478593826293945, "learning_rate": 0.00019995799374024077, "loss": 4.5888, "step": 46500 }, { "epoch": 0.08686747578137574, "grad_norm": 0.7958502173423767, "learning_rate": 0.00019995787189975794, "loss": 4.7813, "step": 46550 }, { "epoch": 0.08696078134075423, "grad_norm": 1.2592016458511353, "learning_rate": 0.00019995774988286707, "loss": 4.7929, "step": 46600 }, { "epoch": 0.08705408690013272, "grad_norm": 0.9072364568710327, "learning_rate": 0.00019995762768956833, "loss": 4.7389, "step": 46650 }, { "epoch": 0.08714739245951122, "grad_norm": 0.8660083413124084, "learning_rate": 0.000199957505319862, "loss": 4.9873, "step": 46700 }, { "epoch": 0.08724069801888971, "grad_norm": 1.458755373954773, "learning_rate": 0.00019995738277374825, "loss": 4.851, "step": 46750 }, { "epoch": 0.0873340035782682, "grad_norm": 1.0809298753738403, "learning_rate": 0.0001999572600512273, "loss": 4.8688, "step": 46800 }, { "epoch": 0.0874273091376467, "grad_norm": 1.176781415939331, "learning_rate": 0.00019995713715229937, "loss": 4.848, "step": 46850 }, { "epoch": 0.08752061469702518, "grad_norm": 1.1064727306365967, "learning_rate": 0.0001999570140769647, "loss": 4.9472, "step": 46900 }, { "epoch": 0.08761392025640367, "grad_norm": 1.0191171169281006, "learning_rate": 0.00019995689082522348, "loss": 5.0196, "step": 46950 }, { "epoch": 0.08770722581578216, "grad_norm": 1.0902825593948364, "learning_rate": 0.00019995676739707597, "loss": 4.8434, "step": 47000 }, { "epoch": 0.08780053137516065, "grad_norm": 1.0799312591552734, "learning_rate": 0.0001999566437925223, "loss": 4.7215, "step": 47050 }, { "epoch": 0.08789383693453916, "grad_norm": 0.7429112792015076, "learning_rate": 0.00019995652001156278, "loss": 4.7088, "step": 47100 }, { "epoch": 0.08798714249391765, "grad_norm": 1.215327262878418, "learning_rate": 0.00019995639605419757, "loss": 4.7287, "step": 47150 }, { "epoch": 0.08808044805329614, "grad_norm": 0.9942157864570618, "learning_rate": 0.0001999562719204269, "loss": 4.8316, "step": 47200 }, { "epoch": 0.08817375361267463, "grad_norm": 1.1454399824142456, "learning_rate": 0.00019995614761025098, "loss": 4.9424, "step": 47250 }, { "epoch": 0.08826705917205312, "grad_norm": 1.0674301385879517, "learning_rate": 0.0001999560231236701, "loss": 4.8136, "step": 47300 }, { "epoch": 0.0883603647314316, "grad_norm": 1.0504412651062012, "learning_rate": 0.0001999558984606844, "loss": 4.8764, "step": 47350 }, { "epoch": 0.0884536702908101, "grad_norm": 1.0923010110855103, "learning_rate": 0.00019995577362129412, "loss": 4.8347, "step": 47400 }, { "epoch": 0.0885469758501886, "grad_norm": 0.7745321989059448, "learning_rate": 0.0001999556486054995, "loss": 4.9508, "step": 47450 }, { "epoch": 0.08864028140956709, "grad_norm": 0.992335319519043, "learning_rate": 0.00019995552341330074, "loss": 4.8245, "step": 47500 }, { "epoch": 0.08873358696894558, "grad_norm": 1.1324732303619385, "learning_rate": 0.00019995539804469807, "loss": 4.7327, "step": 47550 }, { "epoch": 0.08882689252832407, "grad_norm": 1.0124810934066772, "learning_rate": 0.00019995527249969174, "loss": 4.7606, "step": 47600 }, { "epoch": 0.08892019808770256, "grad_norm": 1.1177074909210205, "learning_rate": 0.00019995514677828192, "loss": 4.9257, "step": 47650 }, { "epoch": 0.08901350364708105, "grad_norm": 1.0227179527282715, "learning_rate": 0.00019995502088046885, "loss": 4.7303, "step": 47700 }, { "epoch": 0.08910680920645954, "grad_norm": 0.9861817359924316, "learning_rate": 0.00019995489480625277, "loss": 4.932, "step": 47750 }, { "epoch": 0.08920011476583803, "grad_norm": 1.1230319738388062, "learning_rate": 0.00019995476855563384, "loss": 4.7706, "step": 47800 }, { "epoch": 0.08929342032521653, "grad_norm": 0.8386600613594055, "learning_rate": 0.00019995464212861237, "loss": 4.7139, "step": 47850 }, { "epoch": 0.08938672588459502, "grad_norm": 1.322847604751587, "learning_rate": 0.00019995451552518853, "loss": 4.8769, "step": 47900 }, { "epoch": 0.08948003144397351, "grad_norm": 1.3985576629638672, "learning_rate": 0.00019995438874536258, "loss": 4.9127, "step": 47950 }, { "epoch": 0.089573337003352, "grad_norm": 1.1547143459320068, "learning_rate": 0.0001999542617891347, "loss": 4.8883, "step": 48000 }, { "epoch": 0.089573337003352, "eval_loss": 5.081268310546875, "eval_runtime": 231.9586, "eval_samples_per_second": 11.243, "eval_steps_per_second": 11.243, "eval_tts_loss": 7.271998688881428, "step": 48000 }, { "epoch": 0.08966664256273049, "grad_norm": 0.7190650701522827, "learning_rate": 0.00019995413465650515, "loss": 4.971, "step": 48050 }, { "epoch": 0.08975994812210898, "grad_norm": 1.1452605724334717, "learning_rate": 0.00019995400734747413, "loss": 4.9515, "step": 48100 }, { "epoch": 0.08985325368148747, "grad_norm": 0.7332000136375427, "learning_rate": 0.00019995387986204187, "loss": 4.7901, "step": 48150 }, { "epoch": 0.08994655924086598, "grad_norm": 1.2664438486099243, "learning_rate": 0.00019995375220020862, "loss": 4.7153, "step": 48200 }, { "epoch": 0.09003986480024447, "grad_norm": 0.9586982727050781, "learning_rate": 0.00019995362436197459, "loss": 4.8614, "step": 48250 }, { "epoch": 0.09013317035962295, "grad_norm": 1.139235019683838, "learning_rate": 0.00019995349634733994, "loss": 4.8637, "step": 48300 }, { "epoch": 0.09022647591900144, "grad_norm": 1.2145622968673706, "learning_rate": 0.000199953368156305, "loss": 4.8257, "step": 48350 }, { "epoch": 0.09031978147837993, "grad_norm": 1.1296002864837646, "learning_rate": 0.00019995323978886997, "loss": 4.8457, "step": 48400 }, { "epoch": 0.09041308703775842, "grad_norm": 0.7904325723648071, "learning_rate": 0.00019995311124503505, "loss": 4.8158, "step": 48450 }, { "epoch": 0.09050639259713691, "grad_norm": 1.2136282920837402, "learning_rate": 0.00019995298252480046, "loss": 4.7482, "step": 48500 }, { "epoch": 0.0905996981565154, "grad_norm": 1.312458872795105, "learning_rate": 0.00019995285362816643, "loss": 4.7005, "step": 48550 }, { "epoch": 0.09069300371589391, "grad_norm": 1.1623774766921997, "learning_rate": 0.00019995272455513324, "loss": 4.8313, "step": 48600 }, { "epoch": 0.0907863092752724, "grad_norm": 1.2712230682373047, "learning_rate": 0.00019995259530570104, "loss": 5.055, "step": 48650 }, { "epoch": 0.09087961483465089, "grad_norm": 1.1107847690582275, "learning_rate": 0.00019995246587987012, "loss": 4.9028, "step": 48700 }, { "epoch": 0.09097292039402938, "grad_norm": 1.1616772413253784, "learning_rate": 0.00019995233627764067, "loss": 4.7912, "step": 48750 }, { "epoch": 0.09106622595340787, "grad_norm": 0.872082531452179, "learning_rate": 0.00019995220649901297, "loss": 4.8532, "step": 48800 }, { "epoch": 0.09115953151278636, "grad_norm": 0.9554139971733093, "learning_rate": 0.00019995207654398714, "loss": 4.7248, "step": 48850 }, { "epoch": 0.09125283707216485, "grad_norm": 1.1058579683303833, "learning_rate": 0.00019995194641256355, "loss": 4.6829, "step": 48900 }, { "epoch": 0.09134614263154335, "grad_norm": 1.0344289541244507, "learning_rate": 0.00019995181610474238, "loss": 4.9281, "step": 48950 }, { "epoch": 0.09143944819092184, "grad_norm": 1.0128055810928345, "learning_rate": 0.00019995168562052379, "loss": 4.8823, "step": 49000 }, { "epoch": 0.09153275375030033, "grad_norm": 1.189562201499939, "learning_rate": 0.00019995155495990805, "loss": 4.8325, "step": 49050 }, { "epoch": 0.09162605930967882, "grad_norm": 1.1200377941131592, "learning_rate": 0.00019995142412289544, "loss": 5.0455, "step": 49100 }, { "epoch": 0.09171936486905731, "grad_norm": 1.298416256904602, "learning_rate": 0.00019995129310948612, "loss": 4.8515, "step": 49150 }, { "epoch": 0.0918126704284358, "grad_norm": 1.1552833318710327, "learning_rate": 0.00019995116191968038, "loss": 4.8612, "step": 49200 }, { "epoch": 0.09190597598781429, "grad_norm": 1.0328021049499512, "learning_rate": 0.00019995103055347842, "loss": 5.0486, "step": 49250 }, { "epoch": 0.09199928154719278, "grad_norm": 0.7570762038230896, "learning_rate": 0.00019995089901088048, "loss": 4.7099, "step": 49300 }, { "epoch": 0.09209258710657128, "grad_norm": 0.9117104411125183, "learning_rate": 0.00019995076729188677, "loss": 4.8546, "step": 49350 }, { "epoch": 0.09218589266594977, "grad_norm": 1.0194332599639893, "learning_rate": 0.00019995063539649758, "loss": 4.8198, "step": 49400 }, { "epoch": 0.09227919822532826, "grad_norm": 0.7058835625648499, "learning_rate": 0.00019995050332471308, "loss": 4.8416, "step": 49450 }, { "epoch": 0.09237250378470675, "grad_norm": 1.046873688697815, "learning_rate": 0.00019995037107653355, "loss": 4.7131, "step": 49500 }, { "epoch": 0.09246580934408524, "grad_norm": 1.1325424909591675, "learning_rate": 0.00019995023865195918, "loss": 4.8868, "step": 49550 }, { "epoch": 0.09255911490346373, "grad_norm": 0.933991551399231, "learning_rate": 0.00019995010605099024, "loss": 4.8612, "step": 49600 }, { "epoch": 0.09265242046284222, "grad_norm": 1.2743916511535645, "learning_rate": 0.0001999499732736269, "loss": 4.7381, "step": 49650 }, { "epoch": 0.09274572602222073, "grad_norm": 0.9663635492324829, "learning_rate": 0.0001999498403198695, "loss": 4.8429, "step": 49700 }, { "epoch": 0.09283903158159922, "grad_norm": 1.0755882263183594, "learning_rate": 0.0001999497071897182, "loss": 4.8957, "step": 49750 }, { "epoch": 0.0929323371409777, "grad_norm": 1.0110505819320679, "learning_rate": 0.0001999495738831733, "loss": 5.0335, "step": 49800 }, { "epoch": 0.0930256427003562, "grad_norm": 1.1059298515319824, "learning_rate": 0.0001999494404002349, "loss": 4.9632, "step": 49850 }, { "epoch": 0.09311894825973469, "grad_norm": 1.1788870096206665, "learning_rate": 0.00019994930674090337, "loss": 5.0718, "step": 49900 }, { "epoch": 0.09321225381911317, "grad_norm": 1.0631682872772217, "learning_rate": 0.0001999491729051789, "loss": 4.6773, "step": 49950 }, { "epoch": 0.09330555937849166, "grad_norm": 0.9250020384788513, "learning_rate": 0.00019994903889306172, "loss": 4.7364, "step": 50000 }, { "epoch": 0.09339886493787015, "grad_norm": 1.3172533512115479, "learning_rate": 0.0001999489047045521, "loss": 4.8955, "step": 50050 }, { "epoch": 0.09349217049724866, "grad_norm": 1.0621302127838135, "learning_rate": 0.00019994877033965023, "loss": 4.7293, "step": 50100 }, { "epoch": 0.09358547605662715, "grad_norm": 1.2123887538909912, "learning_rate": 0.00019994863579835638, "loss": 4.7187, "step": 50150 }, { "epoch": 0.09367878161600564, "grad_norm": 1.2413897514343262, "learning_rate": 0.00019994850108067075, "loss": 4.9169, "step": 50200 }, { "epoch": 0.09377208717538413, "grad_norm": 1.3045127391815186, "learning_rate": 0.0001999483661865936, "loss": 4.8983, "step": 50250 }, { "epoch": 0.09386539273476262, "grad_norm": 1.0260050296783447, "learning_rate": 0.00019994823111612518, "loss": 4.8541, "step": 50300 }, { "epoch": 0.09395869829414111, "grad_norm": 0.8468745350837708, "learning_rate": 0.00019994809586926576, "loss": 4.7221, "step": 50350 }, { "epoch": 0.0940520038535196, "grad_norm": 0.558533251285553, "learning_rate": 0.00019994796044601549, "loss": 4.8051, "step": 50400 }, { "epoch": 0.09414530941289809, "grad_norm": 1.0962811708450317, "learning_rate": 0.00019994782484637465, "loss": 5.0269, "step": 50450 }, { "epoch": 0.09423861497227659, "grad_norm": 1.1004583835601807, "learning_rate": 0.00019994768907034352, "loss": 4.6689, "step": 50500 }, { "epoch": 0.09433192053165508, "grad_norm": 0.585641086101532, "learning_rate": 0.00019994755311792227, "loss": 4.913, "step": 50550 }, { "epoch": 0.09442522609103357, "grad_norm": 1.0821267366409302, "learning_rate": 0.0001999474169891112, "loss": 4.8242, "step": 50600 }, { "epoch": 0.09451853165041206, "grad_norm": 0.7899723649024963, "learning_rate": 0.0001999472806839105, "loss": 4.8794, "step": 50650 }, { "epoch": 0.09461183720979055, "grad_norm": 1.1543865203857422, "learning_rate": 0.00019994714420232046, "loss": 4.9543, "step": 50700 }, { "epoch": 0.09470514276916904, "grad_norm": 0.8708122372627258, "learning_rate": 0.00019994700754434128, "loss": 4.6829, "step": 50750 }, { "epoch": 0.09479844832854753, "grad_norm": 1.07050359249115, "learning_rate": 0.00019994687070997325, "loss": 4.8473, "step": 50800 }, { "epoch": 0.09489175388792603, "grad_norm": 0.9256060123443604, "learning_rate": 0.00019994673369921656, "loss": 4.6233, "step": 50850 }, { "epoch": 0.09498505944730452, "grad_norm": 1.1149330139160156, "learning_rate": 0.00019994659651207143, "loss": 4.88, "step": 50900 }, { "epoch": 0.09507836500668301, "grad_norm": 1.1947017908096313, "learning_rate": 0.00019994645914853817, "loss": 4.7873, "step": 50950 }, { "epoch": 0.0951716705660615, "grad_norm": 0.9794861674308777, "learning_rate": 0.00019994632160861704, "loss": 4.8523, "step": 51000 }, { "epoch": 0.0951716705660615, "eval_loss": 5.061774730682373, "eval_runtime": 231.5203, "eval_samples_per_second": 11.265, "eval_steps_per_second": 11.265, "eval_tts_loss": 7.302327345133947, "step": 51000 }, { "epoch": 0.09526497612543999, "grad_norm": 1.0766355991363525, "learning_rate": 0.0001999461838923082, "loss": 4.662, "step": 51050 }, { "epoch": 0.09535828168481848, "grad_norm": 1.242478609085083, "learning_rate": 0.00019994604599961194, "loss": 5.0057, "step": 51100 }, { "epoch": 0.09545158724419697, "grad_norm": 1.0459357500076294, "learning_rate": 0.0001999459079305285, "loss": 4.8375, "step": 51150 }, { "epoch": 0.09554489280357546, "grad_norm": 1.0175572633743286, "learning_rate": 0.0001999457696850581, "loss": 4.7238, "step": 51200 }, { "epoch": 0.09563819836295397, "grad_norm": 1.033231258392334, "learning_rate": 0.000199945631263201, "loss": 5.1248, "step": 51250 }, { "epoch": 0.09573150392233246, "grad_norm": 1.0977270603179932, "learning_rate": 0.00019994549266495745, "loss": 4.7472, "step": 51300 }, { "epoch": 0.09582480948171095, "grad_norm": 1.0846363306045532, "learning_rate": 0.00019994535389032772, "loss": 4.7239, "step": 51350 }, { "epoch": 0.09591811504108944, "grad_norm": 0.9102774858474731, "learning_rate": 0.000199945214939312, "loss": 4.9151, "step": 51400 }, { "epoch": 0.09601142060046793, "grad_norm": 1.0620869398117065, "learning_rate": 0.00019994507581191058, "loss": 4.5926, "step": 51450 }, { "epoch": 0.09610472615984642, "grad_norm": 1.1233978271484375, "learning_rate": 0.00019994493650812368, "loss": 5.0843, "step": 51500 }, { "epoch": 0.0961980317192249, "grad_norm": 1.2142945528030396, "learning_rate": 0.00019994479702795155, "loss": 4.874, "step": 51550 }, { "epoch": 0.09629133727860341, "grad_norm": 1.21846342086792, "learning_rate": 0.00019994465737139443, "loss": 4.888, "step": 51600 }, { "epoch": 0.0963846428379819, "grad_norm": 1.1796367168426514, "learning_rate": 0.00019994451753845257, "loss": 4.9034, "step": 51650 }, { "epoch": 0.09647794839736039, "grad_norm": 0.9840373992919922, "learning_rate": 0.00019994437752912623, "loss": 4.759, "step": 51700 }, { "epoch": 0.09657125395673888, "grad_norm": 1.2641807794570923, "learning_rate": 0.00019994423734341567, "loss": 4.7452, "step": 51750 }, { "epoch": 0.09666455951611737, "grad_norm": 1.1618438959121704, "learning_rate": 0.0001999440969813211, "loss": 4.6881, "step": 51800 }, { "epoch": 0.09675786507549586, "grad_norm": 0.8106015920639038, "learning_rate": 0.0001999439564428428, "loss": 5.0901, "step": 51850 }, { "epoch": 0.09685117063487435, "grad_norm": 1.3673611879348755, "learning_rate": 0.00019994381572798096, "loss": 4.9073, "step": 51900 }, { "epoch": 0.09694447619425284, "grad_norm": 0.9756952524185181, "learning_rate": 0.0001999436748367359, "loss": 4.9472, "step": 51950 }, { "epoch": 0.09703778175363134, "grad_norm": 1.3207862377166748, "learning_rate": 0.00019994353376910783, "loss": 4.7792, "step": 52000 }, { "epoch": 0.09713108731300983, "grad_norm": 1.0687203407287598, "learning_rate": 0.00019994339252509704, "loss": 4.7979, "step": 52050 }, { "epoch": 0.09722439287238832, "grad_norm": 1.2489949464797974, "learning_rate": 0.00019994325110470371, "loss": 4.8534, "step": 52100 }, { "epoch": 0.09731769843176681, "grad_norm": 1.4121779203414917, "learning_rate": 0.00019994310950792815, "loss": 4.8141, "step": 52150 }, { "epoch": 0.0974110039911453, "grad_norm": 1.0948699712753296, "learning_rate": 0.00019994296773477056, "loss": 4.9626, "step": 52200 }, { "epoch": 0.09750430955052379, "grad_norm": 0.8050052523612976, "learning_rate": 0.00019994282578523126, "loss": 4.838, "step": 52250 }, { "epoch": 0.09759761510990228, "grad_norm": 1.2219855785369873, "learning_rate": 0.00019994268365931044, "loss": 5.0231, "step": 52300 }, { "epoch": 0.09769092066928078, "grad_norm": 1.0798107385635376, "learning_rate": 0.00019994254135700837, "loss": 4.9951, "step": 52350 }, { "epoch": 0.09778422622865927, "grad_norm": 1.1264784336090088, "learning_rate": 0.0001999423988783253, "loss": 4.9835, "step": 52400 }, { "epoch": 0.09787753178803776, "grad_norm": 1.0738651752471924, "learning_rate": 0.00019994225622326143, "loss": 4.6703, "step": 52450 }, { "epoch": 0.09797083734741625, "grad_norm": 0.9834763407707214, "learning_rate": 0.0001999421133918171, "loss": 4.8748, "step": 52500 }, { "epoch": 0.09806414290679474, "grad_norm": 0.9889846444129944, "learning_rate": 0.00019994197038399256, "loss": 5.0041, "step": 52550 }, { "epoch": 0.09815744846617323, "grad_norm": 0.6598125696182251, "learning_rate": 0.000199941827199788, "loss": 4.7931, "step": 52600 }, { "epoch": 0.09825075402555172, "grad_norm": 0.7698989510536194, "learning_rate": 0.00019994168383920372, "loss": 4.6956, "step": 52650 }, { "epoch": 0.09834405958493021, "grad_norm": 0.9986603260040283, "learning_rate": 0.00019994154030223994, "loss": 4.7647, "step": 52700 }, { "epoch": 0.09843736514430872, "grad_norm": 0.9387907385826111, "learning_rate": 0.0001999413965888969, "loss": 4.7919, "step": 52750 }, { "epoch": 0.0985306707036872, "grad_norm": 0.997485339641571, "learning_rate": 0.0001999412526991749, "loss": 4.864, "step": 52800 }, { "epoch": 0.0986239762630657, "grad_norm": 1.0263787508010864, "learning_rate": 0.0001999411086330742, "loss": 5.0764, "step": 52850 }, { "epoch": 0.09871728182244419, "grad_norm": 1.1817213296890259, "learning_rate": 0.000199940964390595, "loss": 4.9443, "step": 52900 }, { "epoch": 0.09881058738182268, "grad_norm": 1.2216764688491821, "learning_rate": 0.00019994081997173758, "loss": 4.9013, "step": 52950 }, { "epoch": 0.09890389294120117, "grad_norm": 1.020540475845337, "learning_rate": 0.00019994067537650221, "loss": 4.9208, "step": 53000 }, { "epoch": 0.09899719850057966, "grad_norm": 1.07760488986969, "learning_rate": 0.00019994053060488911, "loss": 4.8805, "step": 53050 }, { "epoch": 0.09909050405995816, "grad_norm": 0.9263754487037659, "learning_rate": 0.00019994038565689857, "loss": 4.6882, "step": 53100 }, { "epoch": 0.09918380961933665, "grad_norm": 1.0581586360931396, "learning_rate": 0.00019994024053253084, "loss": 4.7098, "step": 53150 }, { "epoch": 0.09927711517871514, "grad_norm": 0.9992542266845703, "learning_rate": 0.00019994009523178617, "loss": 4.8672, "step": 53200 }, { "epoch": 0.09937042073809363, "grad_norm": 0.9917731285095215, "learning_rate": 0.00019993994975466483, "loss": 4.7594, "step": 53250 }, { "epoch": 0.09946372629747212, "grad_norm": 0.8953468799591064, "learning_rate": 0.00019993980410116705, "loss": 4.9375, "step": 53300 }, { "epoch": 0.09955703185685061, "grad_norm": 1.0633111000061035, "learning_rate": 0.0001999396582712931, "loss": 4.7604, "step": 53350 }, { "epoch": 0.0996503374162291, "grad_norm": 0.9700153470039368, "learning_rate": 0.00019993951226504323, "loss": 4.8346, "step": 53400 }, { "epoch": 0.09974364297560759, "grad_norm": 1.198554277420044, "learning_rate": 0.00019993936608241771, "loss": 4.7809, "step": 53450 }, { "epoch": 0.09983694853498609, "grad_norm": 0.8726997971534729, "learning_rate": 0.0001999392197234168, "loss": 4.9039, "step": 53500 }, { "epoch": 0.09993025409436458, "grad_norm": 1.296087384223938, "learning_rate": 0.0001999390731880408, "loss": 4.7456, "step": 53550 }, { "epoch": 0.10002355965374307, "grad_norm": 0.9055943489074707, "learning_rate": 0.00019993892647628987, "loss": 4.964, "step": 53600 }, { "epoch": 0.10011686521312156, "grad_norm": 0.900660514831543, "learning_rate": 0.0001999387795881643, "loss": 4.92, "step": 53650 }, { "epoch": 0.10021017077250005, "grad_norm": 1.2182655334472656, "learning_rate": 0.0001999386325236644, "loss": 4.8738, "step": 53700 }, { "epoch": 0.10030347633187854, "grad_norm": 1.0977277755737305, "learning_rate": 0.0001999384852827904, "loss": 4.4823, "step": 53750 }, { "epoch": 0.10039678189125703, "grad_norm": 0.7490143775939941, "learning_rate": 0.00019993833786554252, "loss": 4.8447, "step": 53800 }, { "epoch": 0.10049008745063552, "grad_norm": 1.216125726699829, "learning_rate": 0.0001999381902719211, "loss": 4.7476, "step": 53850 }, { "epoch": 0.10058339301001402, "grad_norm": 0.9461069703102112, "learning_rate": 0.00019993804250192633, "loss": 4.7481, "step": 53900 }, { "epoch": 0.10067669856939251, "grad_norm": 0.9821112751960754, "learning_rate": 0.00019993789455555853, "loss": 4.6914, "step": 53950 }, { "epoch": 0.100770004128771, "grad_norm": 1.069184422492981, "learning_rate": 0.00019993774643281792, "loss": 4.8143, "step": 54000 }, { "epoch": 0.100770004128771, "eval_loss": 5.050092697143555, "eval_runtime": 233.5139, "eval_samples_per_second": 11.169, "eval_steps_per_second": 11.169, "eval_tts_loss": 7.309272037400504, "step": 54000 }, { "epoch": 0.1008633096881495, "grad_norm": 0.8678284883499146, "learning_rate": 0.00019993759813370475, "loss": 4.7695, "step": 54050 }, { "epoch": 0.10095661524752798, "grad_norm": 1.0255569219589233, "learning_rate": 0.00019993744965821932, "loss": 4.7333, "step": 54100 }, { "epoch": 0.10104992080690647, "grad_norm": 0.9796172976493835, "learning_rate": 0.0001999373010063619, "loss": 4.9283, "step": 54150 }, { "epoch": 0.10114322636628496, "grad_norm": 0.9378018975257874, "learning_rate": 0.00019993715217813272, "loss": 4.7371, "step": 54200 }, { "epoch": 0.10123653192566347, "grad_norm": 1.0029010772705078, "learning_rate": 0.00019993700317353201, "loss": 4.6258, "step": 54250 }, { "epoch": 0.10132983748504196, "grad_norm": 0.922479510307312, "learning_rate": 0.0001999368539925601, "loss": 5.0026, "step": 54300 }, { "epoch": 0.10142314304442045, "grad_norm": 1.0813497304916382, "learning_rate": 0.00019993670463521723, "loss": 4.8323, "step": 54350 }, { "epoch": 0.10151644860379894, "grad_norm": 1.0883569717407227, "learning_rate": 0.00019993655510150367, "loss": 4.8957, "step": 54400 }, { "epoch": 0.10160975416317743, "grad_norm": 0.792852520942688, "learning_rate": 0.00019993640539141966, "loss": 5.0127, "step": 54450 }, { "epoch": 0.10170305972255592, "grad_norm": 0.8241887092590332, "learning_rate": 0.0001999362555049655, "loss": 4.7821, "step": 54500 }, { "epoch": 0.1017963652819344, "grad_norm": 1.088869571685791, "learning_rate": 0.0001999361054421414, "loss": 4.8543, "step": 54550 }, { "epoch": 0.1018896708413129, "grad_norm": 1.0283081531524658, "learning_rate": 0.00019993595520294765, "loss": 4.7311, "step": 54600 }, { "epoch": 0.1019829764006914, "grad_norm": 0.7650019526481628, "learning_rate": 0.00019993580478738456, "loss": 4.7729, "step": 54650 }, { "epoch": 0.10207628196006989, "grad_norm": 0.8107566237449646, "learning_rate": 0.00019993565419545236, "loss": 5.0273, "step": 54700 }, { "epoch": 0.10216958751944838, "grad_norm": 1.2203528881072998, "learning_rate": 0.00019993550342715127, "loss": 4.6369, "step": 54750 }, { "epoch": 0.10226289307882687, "grad_norm": 0.9938797354698181, "learning_rate": 0.00019993535248248164, "loss": 4.7657, "step": 54800 }, { "epoch": 0.10235619863820536, "grad_norm": 1.1002684831619263, "learning_rate": 0.0001999352013614437, "loss": 4.8818, "step": 54850 }, { "epoch": 0.10244950419758385, "grad_norm": 0.9734085202217102, "learning_rate": 0.0001999350500640377, "loss": 4.8906, "step": 54900 }, { "epoch": 0.10254280975696234, "grad_norm": 0.9943709373474121, "learning_rate": 0.0001999348985902639, "loss": 4.8368, "step": 54950 }, { "epoch": 0.10263611531634084, "grad_norm": 0.8544298410415649, "learning_rate": 0.00019993474694012263, "loss": 4.62, "step": 55000 }, { "epoch": 0.10272942087571933, "grad_norm": 0.7274461388587952, "learning_rate": 0.0001999345951136141, "loss": 4.7548, "step": 55050 }, { "epoch": 0.10282272643509782, "grad_norm": 1.0700294971466064, "learning_rate": 0.00019993444311073859, "loss": 4.7764, "step": 55100 }, { "epoch": 0.10291603199447631, "grad_norm": 1.211059808731079, "learning_rate": 0.00019993429093149637, "loss": 4.7714, "step": 55150 }, { "epoch": 0.1030093375538548, "grad_norm": 1.0298805236816406, "learning_rate": 0.00019993413857588772, "loss": 4.8341, "step": 55200 }, { "epoch": 0.10310264311323329, "grad_norm": 1.2335697412490845, "learning_rate": 0.0001999339860439129, "loss": 4.7979, "step": 55250 }, { "epoch": 0.10319594867261178, "grad_norm": 1.05863618850708, "learning_rate": 0.00019993383333557216, "loss": 5.0299, "step": 55300 }, { "epoch": 0.10328925423199027, "grad_norm": 0.8664848804473877, "learning_rate": 0.0001999336804508658, "loss": 4.7784, "step": 55350 }, { "epoch": 0.10338255979136877, "grad_norm": 1.1443610191345215, "learning_rate": 0.00019993352738979407, "loss": 4.93, "step": 55400 }, { "epoch": 0.10347586535074726, "grad_norm": 1.0898007154464722, "learning_rate": 0.00019993337415235725, "loss": 4.6334, "step": 55450 }, { "epoch": 0.10356917091012575, "grad_norm": 0.9437859654426575, "learning_rate": 0.00019993322073855562, "loss": 4.8104, "step": 55500 }, { "epoch": 0.10366247646950424, "grad_norm": 1.0893656015396118, "learning_rate": 0.00019993306714838943, "loss": 4.8592, "step": 55550 }, { "epoch": 0.10375578202888273, "grad_norm": 1.0553542375564575, "learning_rate": 0.00019993291338185897, "loss": 4.8595, "step": 55600 }, { "epoch": 0.10384908758826122, "grad_norm": 0.9406054019927979, "learning_rate": 0.00019993275943896448, "loss": 4.7845, "step": 55650 }, { "epoch": 0.10394239314763971, "grad_norm": 0.9452017545700073, "learning_rate": 0.00019993260531970627, "loss": 4.6355, "step": 55700 }, { "epoch": 0.10403569870701822, "grad_norm": 0.8077996969223022, "learning_rate": 0.00019993245102408459, "loss": 4.9368, "step": 55750 }, { "epoch": 0.10412900426639671, "grad_norm": 0.9253317713737488, "learning_rate": 0.00019993229655209974, "loss": 4.6939, "step": 55800 }, { "epoch": 0.1042223098257752, "grad_norm": 1.2367135286331177, "learning_rate": 0.0001999321419037519, "loss": 5.0054, "step": 55850 }, { "epoch": 0.10431561538515369, "grad_norm": 0.9509602189064026, "learning_rate": 0.00019993198707904148, "loss": 4.7925, "step": 55900 }, { "epoch": 0.10440892094453218, "grad_norm": 0.9606555700302124, "learning_rate": 0.00019993183207796868, "loss": 4.7034, "step": 55950 }, { "epoch": 0.10450222650391067, "grad_norm": 0.7478787899017334, "learning_rate": 0.00019993167690053376, "loss": 4.9531, "step": 56000 }, { "epoch": 0.10459553206328916, "grad_norm": 0.9634212851524353, "learning_rate": 0.00019993152154673704, "loss": 4.9075, "step": 56050 }, { "epoch": 0.10468883762266765, "grad_norm": 0.83271723985672, "learning_rate": 0.00019993136601657873, "loss": 4.6116, "step": 56100 }, { "epoch": 0.10478214318204615, "grad_norm": 1.1285144090652466, "learning_rate": 0.0001999312103100592, "loss": 4.6186, "step": 56150 }, { "epoch": 0.10487544874142464, "grad_norm": 0.8766820430755615, "learning_rate": 0.00019993105442717862, "loss": 4.7467, "step": 56200 }, { "epoch": 0.10496875430080313, "grad_norm": 1.0494366884231567, "learning_rate": 0.00019993089836793732, "loss": 4.7853, "step": 56250 }, { "epoch": 0.10506205986018162, "grad_norm": 0.9477689862251282, "learning_rate": 0.00019993074213233556, "loss": 4.8802, "step": 56300 }, { "epoch": 0.10515536541956011, "grad_norm": 1.2510788440704346, "learning_rate": 0.00019993058572037364, "loss": 4.8521, "step": 56350 }, { "epoch": 0.1052486709789386, "grad_norm": 1.0136951208114624, "learning_rate": 0.0001999304291320518, "loss": 4.6612, "step": 56400 }, { "epoch": 0.10534197653831709, "grad_norm": 1.1237655878067017, "learning_rate": 0.00019993027236737038, "loss": 5.043, "step": 56450 }, { "epoch": 0.10543528209769558, "grad_norm": 0.8546143770217896, "learning_rate": 0.00019993011542632956, "loss": 4.6913, "step": 56500 }, { "epoch": 0.10552858765707408, "grad_norm": 1.2368345260620117, "learning_rate": 0.00019992995830892972, "loss": 4.9539, "step": 56550 }, { "epoch": 0.10562189321645257, "grad_norm": 1.055706262588501, "learning_rate": 0.00019992980101517102, "loss": 4.8411, "step": 56600 }, { "epoch": 0.10571519877583106, "grad_norm": 0.8533573746681213, "learning_rate": 0.00019992964354505387, "loss": 4.6453, "step": 56650 }, { "epoch": 0.10580850433520955, "grad_norm": 1.2660565376281738, "learning_rate": 0.00019992948589857843, "loss": 4.9287, "step": 56700 }, { "epoch": 0.10590180989458804, "grad_norm": 1.2691367864608765, "learning_rate": 0.00019992932807574507, "loss": 4.7645, "step": 56750 }, { "epoch": 0.10599511545396653, "grad_norm": 1.0898069143295288, "learning_rate": 0.00019992917007655403, "loss": 4.924, "step": 56800 }, { "epoch": 0.10608842101334502, "grad_norm": 1.1141910552978516, "learning_rate": 0.00019992901190100555, "loss": 4.7186, "step": 56850 }, { "epoch": 0.10618172657272353, "grad_norm": 1.047428846359253, "learning_rate": 0.00019992885354909998, "loss": 4.5938, "step": 56900 }, { "epoch": 0.10627503213210202, "grad_norm": 1.3936903476715088, "learning_rate": 0.00019992869502083755, "loss": 4.661, "step": 56950 }, { "epoch": 0.1063683376914805, "grad_norm": 1.2814745903015137, "learning_rate": 0.00019992853631621858, "loss": 5.1075, "step": 57000 }, { "epoch": 0.1063683376914805, "eval_loss": 5.03396463394165, "eval_runtime": 231.8405, "eval_samples_per_second": 11.249, "eval_steps_per_second": 11.249, "eval_tts_loss": 7.338024720789556, "step": 57000 }, { "epoch": 0.106461643250859, "grad_norm": 0.9817240834236145, "learning_rate": 0.0001999283774352433, "loss": 4.6815, "step": 57050 }, { "epoch": 0.10655494881023748, "grad_norm": 0.8881374001502991, "learning_rate": 0.000199928218377912, "loss": 4.7038, "step": 57100 }, { "epoch": 0.10664825436961597, "grad_norm": 0.8835790157318115, "learning_rate": 0.00019992805914422501, "loss": 4.805, "step": 57150 }, { "epoch": 0.10674155992899446, "grad_norm": 1.0517323017120361, "learning_rate": 0.0001999278997341826, "loss": 4.8259, "step": 57200 }, { "epoch": 0.10683486548837295, "grad_norm": 0.9587489366531372, "learning_rate": 0.000199927740147785, "loss": 4.7815, "step": 57250 }, { "epoch": 0.10692817104775146, "grad_norm": 1.1249165534973145, "learning_rate": 0.00019992758038503249, "loss": 4.5363, "step": 57300 }, { "epoch": 0.10702147660712995, "grad_norm": 1.0124765634536743, "learning_rate": 0.00019992742044592542, "loss": 4.6893, "step": 57350 }, { "epoch": 0.10711478216650844, "grad_norm": 0.969052255153656, "learning_rate": 0.00019992726033046405, "loss": 4.7008, "step": 57400 }, { "epoch": 0.10720808772588693, "grad_norm": 1.0661418437957764, "learning_rate": 0.00019992710003864864, "loss": 4.9145, "step": 57450 }, { "epoch": 0.10730139328526542, "grad_norm": 0.9736279845237732, "learning_rate": 0.00019992693957047948, "loss": 4.5799, "step": 57500 }, { "epoch": 0.10739469884464391, "grad_norm": 0.8981762528419495, "learning_rate": 0.0001999267789259568, "loss": 4.9882, "step": 57550 }, { "epoch": 0.1074880044040224, "grad_norm": 1.3107281923294067, "learning_rate": 0.000199926618105081, "loss": 4.5749, "step": 57600 }, { "epoch": 0.1075813099634009, "grad_norm": 0.8268171548843384, "learning_rate": 0.00019992645710785228, "loss": 4.7451, "step": 57650 }, { "epoch": 0.10767461552277939, "grad_norm": 0.8941455483436584, "learning_rate": 0.00019992629593427097, "loss": 4.7114, "step": 57700 }, { "epoch": 0.10776792108215788, "grad_norm": 1.0125592947006226, "learning_rate": 0.0001999261345843373, "loss": 4.7143, "step": 57750 }, { "epoch": 0.10786122664153637, "grad_norm": 1.0623835325241089, "learning_rate": 0.0001999259730580516, "loss": 4.8096, "step": 57800 }, { "epoch": 0.10795453220091486, "grad_norm": 0.9189125299453735, "learning_rate": 0.00019992581135541414, "loss": 4.5356, "step": 57850 }, { "epoch": 0.10804783776029335, "grad_norm": 1.0694884061813354, "learning_rate": 0.0001999256494764252, "loss": 4.902, "step": 57900 }, { "epoch": 0.10814114331967184, "grad_norm": 1.200424313545227, "learning_rate": 0.00019992548742108507, "loss": 4.6708, "step": 57950 }, { "epoch": 0.10823444887905033, "grad_norm": 1.0127471685409546, "learning_rate": 0.00019992532518939404, "loss": 4.6113, "step": 58000 }, { "epoch": 0.10832775443842883, "grad_norm": 1.081398367881775, "learning_rate": 0.0001999251627813524, "loss": 4.7227, "step": 58050 }, { "epoch": 0.10842105999780732, "grad_norm": 1.0527719259262085, "learning_rate": 0.00019992500019696042, "loss": 4.7023, "step": 58100 }, { "epoch": 0.10851436555718581, "grad_norm": 0.9371632933616638, "learning_rate": 0.0001999248374362184, "loss": 4.7177, "step": 58150 }, { "epoch": 0.1086076711165643, "grad_norm": 1.2352396249771118, "learning_rate": 0.00019992467449912664, "loss": 4.7167, "step": 58200 }, { "epoch": 0.10870097667594279, "grad_norm": 1.295750379562378, "learning_rate": 0.0001999245113856854, "loss": 4.6888, "step": 58250 }, { "epoch": 0.10879428223532128, "grad_norm": 0.9914165735244751, "learning_rate": 0.00019992434809589494, "loss": 4.6626, "step": 58300 }, { "epoch": 0.10888758779469977, "grad_norm": 1.038947582244873, "learning_rate": 0.00019992418462975566, "loss": 4.8624, "step": 58350 }, { "epoch": 0.10898089335407828, "grad_norm": 0.9562987685203552, "learning_rate": 0.00019992402098726776, "loss": 4.8079, "step": 58400 }, { "epoch": 0.10907419891345677, "grad_norm": 1.1640965938568115, "learning_rate": 0.0001999238571684315, "loss": 4.749, "step": 58450 }, { "epoch": 0.10916750447283526, "grad_norm": 0.8574897646903992, "learning_rate": 0.00019992369317324725, "loss": 4.5478, "step": 58500 }, { "epoch": 0.10926081003221375, "grad_norm": 1.1693881750106812, "learning_rate": 0.00019992352900171523, "loss": 4.8086, "step": 58550 }, { "epoch": 0.10935411559159224, "grad_norm": 1.0062192678451538, "learning_rate": 0.0001999233646538358, "loss": 4.7885, "step": 58600 }, { "epoch": 0.10944742115097073, "grad_norm": 1.1878728866577148, "learning_rate": 0.0001999232001296092, "loss": 4.7585, "step": 58650 }, { "epoch": 0.10954072671034921, "grad_norm": 0.6779685020446777, "learning_rate": 0.00019992303542903575, "loss": 4.8446, "step": 58700 }, { "epoch": 0.1096340322697277, "grad_norm": 1.022136926651001, "learning_rate": 0.0001999228705521157, "loss": 4.7668, "step": 58750 }, { "epoch": 0.10972733782910621, "grad_norm": 0.8543698787689209, "learning_rate": 0.00019992270549884936, "loss": 4.9229, "step": 58800 }, { "epoch": 0.1098206433884847, "grad_norm": 1.0954478979110718, "learning_rate": 0.00019992254026923704, "loss": 4.7663, "step": 58850 }, { "epoch": 0.10991394894786319, "grad_norm": 1.051382303237915, "learning_rate": 0.00019992237486327904, "loss": 4.9512, "step": 58900 }, { "epoch": 0.11000725450724168, "grad_norm": 1.2128174304962158, "learning_rate": 0.0001999222092809756, "loss": 4.677, "step": 58950 }, { "epoch": 0.11010056006662017, "grad_norm": 0.8626731038093567, "learning_rate": 0.00019992204352232707, "loss": 4.6943, "step": 59000 }, { "epoch": 0.11019386562599866, "grad_norm": 1.2140212059020996, "learning_rate": 0.00019992187758733368, "loss": 4.8939, "step": 59050 }, { "epoch": 0.11028717118537715, "grad_norm": 1.127719759941101, "learning_rate": 0.00019992171147599578, "loss": 4.8684, "step": 59100 }, { "epoch": 0.11038047674475565, "grad_norm": 1.0840810537338257, "learning_rate": 0.00019992154518831364, "loss": 4.947, "step": 59150 }, { "epoch": 0.11047378230413414, "grad_norm": 1.0713996887207031, "learning_rate": 0.00019992137872428754, "loss": 4.7424, "step": 59200 }, { "epoch": 0.11056708786351263, "grad_norm": 0.8163297176361084, "learning_rate": 0.0001999212120839178, "loss": 4.8426, "step": 59250 }, { "epoch": 0.11066039342289112, "grad_norm": 0.9037017226219177, "learning_rate": 0.0001999210452672047, "loss": 5.0381, "step": 59300 }, { "epoch": 0.11075369898226961, "grad_norm": 1.091113567352295, "learning_rate": 0.00019992087827414853, "loss": 4.5411, "step": 59350 }, { "epoch": 0.1108470045416481, "grad_norm": 0.872543454170227, "learning_rate": 0.0001999207111047496, "loss": 4.6965, "step": 59400 }, { "epoch": 0.11094031010102659, "grad_norm": 0.9569900035858154, "learning_rate": 0.0001999205437590082, "loss": 5.0155, "step": 59450 }, { "epoch": 0.11103361566040508, "grad_norm": 0.8814337253570557, "learning_rate": 0.00019992037623692461, "loss": 4.9238, "step": 59500 }, { "epoch": 0.11112692121978358, "grad_norm": 1.0400547981262207, "learning_rate": 0.00019992020853849916, "loss": 4.7072, "step": 59550 }, { "epoch": 0.11122022677916207, "grad_norm": 1.0725575685501099, "learning_rate": 0.0001999200406637321, "loss": 4.7132, "step": 59600 }, { "epoch": 0.11131353233854056, "grad_norm": 1.1183878183364868, "learning_rate": 0.00019991987261262377, "loss": 4.8993, "step": 59650 }, { "epoch": 0.11140683789791905, "grad_norm": 1.005928874015808, "learning_rate": 0.0001999197043851744, "loss": 4.7833, "step": 59700 }, { "epoch": 0.11150014345729754, "grad_norm": 0.9358251690864563, "learning_rate": 0.0001999195359813844, "loss": 4.6399, "step": 59750 }, { "epoch": 0.11159344901667603, "grad_norm": 0.9199875593185425, "learning_rate": 0.00019991936740125396, "loss": 4.9291, "step": 59800 }, { "epoch": 0.11168675457605452, "grad_norm": 1.1673732995986938, "learning_rate": 0.00019991919864478344, "loss": 4.8272, "step": 59850 }, { "epoch": 0.11178006013543301, "grad_norm": 1.0224826335906982, "learning_rate": 0.0001999190297119731, "loss": 4.6601, "step": 59900 }, { "epoch": 0.11187336569481152, "grad_norm": 1.0551832914352417, "learning_rate": 0.00019991886060282324, "loss": 4.9511, "step": 59950 }, { "epoch": 0.11196667125419, "grad_norm": 0.7255629301071167, "learning_rate": 0.00019991869131733418, "loss": 4.8392, "step": 60000 }, { "epoch": 0.11196667125419, "eval_loss": 5.020105838775635, "eval_runtime": 233.6602, "eval_samples_per_second": 11.162, "eval_steps_per_second": 11.162, "eval_tts_loss": 7.413437219613741, "step": 60000 }, { "epoch": 0.1120599768135685, "grad_norm": 1.0283355712890625, "learning_rate": 0.00019991852185550623, "loss": 4.8832, "step": 60050 }, { "epoch": 0.11215328237294699, "grad_norm": 0.8838055729866028, "learning_rate": 0.00019991835221733965, "loss": 4.6478, "step": 60100 }, { "epoch": 0.11224658793232548, "grad_norm": 1.2853875160217285, "learning_rate": 0.00019991818240283477, "loss": 4.7012, "step": 60150 }, { "epoch": 0.11233989349170397, "grad_norm": 1.3384721279144287, "learning_rate": 0.0001999180124119919, "loss": 4.8132, "step": 60200 }, { "epoch": 0.11243319905108246, "grad_norm": 1.0799473524093628, "learning_rate": 0.00019991784224481128, "loss": 4.793, "step": 60250 }, { "epoch": 0.11252650461046096, "grad_norm": 0.8523072004318237, "learning_rate": 0.00019991767190129326, "loss": 4.9316, "step": 60300 }, { "epoch": 0.11261981016983945, "grad_norm": 0.9551059007644653, "learning_rate": 0.00019991750138143813, "loss": 4.8224, "step": 60350 }, { "epoch": 0.11271311572921794, "grad_norm": 1.0353354215621948, "learning_rate": 0.0001999173306852462, "loss": 4.6761, "step": 60400 }, { "epoch": 0.11280642128859643, "grad_norm": 0.6829168796539307, "learning_rate": 0.00019991715981271778, "loss": 4.8235, "step": 60450 }, { "epoch": 0.11289972684797492, "grad_norm": 0.7613895535469055, "learning_rate": 0.00019991698876385314, "loss": 4.759, "step": 60500 }, { "epoch": 0.11299303240735341, "grad_norm": 0.9478106498718262, "learning_rate": 0.0001999168175386526, "loss": 4.6671, "step": 60550 }, { "epoch": 0.1130863379667319, "grad_norm": 0.9536855220794678, "learning_rate": 0.00019991664613711643, "loss": 4.8579, "step": 60600 }, { "epoch": 0.11317964352611039, "grad_norm": 1.144857406616211, "learning_rate": 0.00019991647455924497, "loss": 4.9233, "step": 60650 }, { "epoch": 0.11327294908548889, "grad_norm": 1.1187453269958496, "learning_rate": 0.00019991630280503852, "loss": 4.7061, "step": 60700 }, { "epoch": 0.11336625464486738, "grad_norm": 1.1460309028625488, "learning_rate": 0.0001999161308744974, "loss": 4.8344, "step": 60750 }, { "epoch": 0.11345956020424587, "grad_norm": 0.9266796112060547, "learning_rate": 0.00019991595876762185, "loss": 4.6502, "step": 60800 }, { "epoch": 0.11355286576362436, "grad_norm": 0.9658082127571106, "learning_rate": 0.00019991578648441224, "loss": 4.9107, "step": 60850 }, { "epoch": 0.11364617132300285, "grad_norm": 0.9839101433753967, "learning_rate": 0.00019991561402486885, "loss": 4.6822, "step": 60900 }, { "epoch": 0.11373947688238134, "grad_norm": 1.0026122331619263, "learning_rate": 0.00019991544138899195, "loss": 4.6888, "step": 60950 }, { "epoch": 0.11383278244175983, "grad_norm": 1.139105200767517, "learning_rate": 0.0001999152685767819, "loss": 4.6663, "step": 61000 }, { "epoch": 0.11392608800113833, "grad_norm": 0.9084557294845581, "learning_rate": 0.00019991509558823898, "loss": 4.77, "step": 61050 }, { "epoch": 0.11401939356051682, "grad_norm": 1.2279844284057617, "learning_rate": 0.0001999149224233635, "loss": 4.8361, "step": 61100 }, { "epoch": 0.11411269911989531, "grad_norm": 0.8564290404319763, "learning_rate": 0.00019991474908215574, "loss": 4.8206, "step": 61150 }, { "epoch": 0.1142060046792738, "grad_norm": 0.92253577709198, "learning_rate": 0.00019991457556461605, "loss": 4.8973, "step": 61200 }, { "epoch": 0.1142993102386523, "grad_norm": 1.0969637632369995, "learning_rate": 0.0001999144018707447, "loss": 4.9673, "step": 61250 }, { "epoch": 0.11439261579803078, "grad_norm": 1.1410199403762817, "learning_rate": 0.00019991422800054204, "loss": 4.6975, "step": 61300 }, { "epoch": 0.11448592135740927, "grad_norm": 1.0572748184204102, "learning_rate": 0.00019991405395400835, "loss": 4.7044, "step": 61350 }, { "epoch": 0.11457922691678776, "grad_norm": 1.115541696548462, "learning_rate": 0.0001999138797311439, "loss": 4.7546, "step": 61400 }, { "epoch": 0.11467253247616627, "grad_norm": 0.9594268798828125, "learning_rate": 0.00019991370533194904, "loss": 4.7057, "step": 61450 }, { "epoch": 0.11476583803554476, "grad_norm": 1.0258264541625977, "learning_rate": 0.00019991353075642408, "loss": 4.7853, "step": 61500 }, { "epoch": 0.11485914359492325, "grad_norm": 1.0233960151672363, "learning_rate": 0.00019991335600456932, "loss": 4.7917, "step": 61550 }, { "epoch": 0.11495244915430174, "grad_norm": 0.901678204536438, "learning_rate": 0.00019991318107638506, "loss": 4.9225, "step": 61600 }, { "epoch": 0.11504575471368023, "grad_norm": 0.7477099299430847, "learning_rate": 0.00019991300597187165, "loss": 4.7892, "step": 61650 }, { "epoch": 0.11513906027305872, "grad_norm": 1.0212405920028687, "learning_rate": 0.00019991283069102932, "loss": 4.789, "step": 61700 }, { "epoch": 0.1152323658324372, "grad_norm": 0.9930858016014099, "learning_rate": 0.00019991265523385842, "loss": 4.7956, "step": 61750 }, { "epoch": 0.11532567139181571, "grad_norm": 0.9863542318344116, "learning_rate": 0.0001999124796003593, "loss": 4.7178, "step": 61800 }, { "epoch": 0.1154189769511942, "grad_norm": 0.8697104454040527, "learning_rate": 0.0001999123037905322, "loss": 4.7792, "step": 61850 }, { "epoch": 0.11551228251057269, "grad_norm": 1.0396831035614014, "learning_rate": 0.0001999121278043775, "loss": 4.8915, "step": 61900 }, { "epoch": 0.11560558806995118, "grad_norm": 1.0583255290985107, "learning_rate": 0.00019991195164189546, "loss": 4.7479, "step": 61950 }, { "epoch": 0.11569889362932967, "grad_norm": 0.8265174627304077, "learning_rate": 0.0001999117753030864, "loss": 4.7255, "step": 62000 }, { "epoch": 0.11579219918870816, "grad_norm": 1.1002154350280762, "learning_rate": 0.00019991159878795064, "loss": 4.8505, "step": 62050 }, { "epoch": 0.11588550474808665, "grad_norm": 1.1061955690383911, "learning_rate": 0.00019991142209648848, "loss": 4.6528, "step": 62100 }, { "epoch": 0.11597881030746514, "grad_norm": 1.06044340133667, "learning_rate": 0.00019991124522870024, "loss": 4.8613, "step": 62150 }, { "epoch": 0.11607211586684364, "grad_norm": 1.1321583986282349, "learning_rate": 0.00019991106818458622, "loss": 4.6643, "step": 62200 }, { "epoch": 0.11616542142622213, "grad_norm": 1.0335965156555176, "learning_rate": 0.00019991089096414678, "loss": 4.7951, "step": 62250 }, { "epoch": 0.11625872698560062, "grad_norm": 1.1257051229476929, "learning_rate": 0.00019991071356738216, "loss": 4.8458, "step": 62300 }, { "epoch": 0.11635203254497911, "grad_norm": 0.9923104047775269, "learning_rate": 0.00019991053599429273, "loss": 4.7476, "step": 62350 }, { "epoch": 0.1164453381043576, "grad_norm": 1.1084821224212646, "learning_rate": 0.00019991035824487878, "loss": 4.6966, "step": 62400 }, { "epoch": 0.11653864366373609, "grad_norm": 0.9355782866477966, "learning_rate": 0.00019991018031914064, "loss": 4.7396, "step": 62450 }, { "epoch": 0.11663194922311458, "grad_norm": 1.1987919807434082, "learning_rate": 0.0001999100022170786, "loss": 4.7583, "step": 62500 }, { "epoch": 0.11672525478249308, "grad_norm": 1.2184990644454956, "learning_rate": 0.000199909823938693, "loss": 5.0186, "step": 62550 }, { "epoch": 0.11681856034187157, "grad_norm": 1.0517642498016357, "learning_rate": 0.00019990964548398412, "loss": 4.9411, "step": 62600 }, { "epoch": 0.11691186590125006, "grad_norm": 1.2433594465255737, "learning_rate": 0.00019990946685295228, "loss": 4.7959, "step": 62650 }, { "epoch": 0.11700517146062855, "grad_norm": 1.1244089603424072, "learning_rate": 0.00019990928804559786, "loss": 4.6779, "step": 62700 }, { "epoch": 0.11709847702000704, "grad_norm": 1.0606389045715332, "learning_rate": 0.00019990910906192107, "loss": 4.8322, "step": 62750 }, { "epoch": 0.11719178257938553, "grad_norm": 1.1183406114578247, "learning_rate": 0.00019990892990192233, "loss": 4.7931, "step": 62800 }, { "epoch": 0.11728508813876402, "grad_norm": 0.8485277891159058, "learning_rate": 0.0001999087505656019, "loss": 4.8164, "step": 62850 }, { "epoch": 0.11737839369814251, "grad_norm": 0.9464021921157837, "learning_rate": 0.0001999085710529601, "loss": 4.7665, "step": 62900 }, { "epoch": 0.11747169925752102, "grad_norm": 0.7202494144439697, "learning_rate": 0.00019990839136399725, "loss": 4.7638, "step": 62950 }, { "epoch": 0.11756500481689951, "grad_norm": 1.0719358921051025, "learning_rate": 0.00019990821149871364, "loss": 4.8894, "step": 63000 }, { "epoch": 0.11756500481689951, "eval_loss": 5.0077104568481445, "eval_runtime": 233.0923, "eval_samples_per_second": 11.189, "eval_steps_per_second": 11.189, "eval_tts_loss": 7.356751820346786, "step": 63000 }, { "epoch": 0.117658310376278, "grad_norm": 1.0613762140274048, "learning_rate": 0.00019990803145710965, "loss": 4.8346, "step": 63050 }, { "epoch": 0.11775161593565649, "grad_norm": 0.9940460324287415, "learning_rate": 0.00019990785123918557, "loss": 4.8047, "step": 63100 }, { "epoch": 0.11784492149503498, "grad_norm": 1.0382481813430786, "learning_rate": 0.0001999076708449417, "loss": 4.6179, "step": 63150 }, { "epoch": 0.11793822705441347, "grad_norm": 1.2504392862319946, "learning_rate": 0.00019990749027437837, "loss": 4.7242, "step": 63200 }, { "epoch": 0.11803153261379196, "grad_norm": 1.3020495176315308, "learning_rate": 0.00019990730952749587, "loss": 4.7754, "step": 63250 }, { "epoch": 0.11812483817317045, "grad_norm": 0.9937262535095215, "learning_rate": 0.00019990712860429457, "loss": 4.6131, "step": 63300 }, { "epoch": 0.11821814373254895, "grad_norm": 0.7087301015853882, "learning_rate": 0.0001999069475047748, "loss": 4.8183, "step": 63350 }, { "epoch": 0.11831144929192744, "grad_norm": 1.2310640811920166, "learning_rate": 0.0001999067662289368, "loss": 4.7773, "step": 63400 }, { "epoch": 0.11840475485130593, "grad_norm": 1.2428086996078491, "learning_rate": 0.00019990658477678097, "loss": 4.7297, "step": 63450 }, { "epoch": 0.11849806041068442, "grad_norm": 0.7657245397567749, "learning_rate": 0.00019990640314830756, "loss": 4.9017, "step": 63500 }, { "epoch": 0.11859136597006291, "grad_norm": 1.0270503759384155, "learning_rate": 0.00019990622134351694, "loss": 4.7397, "step": 63550 }, { "epoch": 0.1186846715294414, "grad_norm": 0.7919912338256836, "learning_rate": 0.00019990603936240943, "loss": 4.8849, "step": 63600 }, { "epoch": 0.11877797708881989, "grad_norm": 1.0639008283615112, "learning_rate": 0.00019990585720498536, "loss": 4.8832, "step": 63650 }, { "epoch": 0.11887128264819839, "grad_norm": 1.0238081216812134, "learning_rate": 0.00019990567487124502, "loss": 4.6403, "step": 63700 }, { "epoch": 0.11896458820757688, "grad_norm": 1.0058157444000244, "learning_rate": 0.00019990549236118872, "loss": 4.5824, "step": 63750 }, { "epoch": 0.11905789376695537, "grad_norm": 0.7168673276901245, "learning_rate": 0.00019990530967481683, "loss": 4.8433, "step": 63800 }, { "epoch": 0.11915119932633386, "grad_norm": 0.9038125276565552, "learning_rate": 0.0001999051268121296, "loss": 4.8536, "step": 63850 }, { "epoch": 0.11924450488571235, "grad_norm": 0.9761154651641846, "learning_rate": 0.00019990494377312745, "loss": 4.7876, "step": 63900 }, { "epoch": 0.11933781044509084, "grad_norm": 1.0363209247589111, "learning_rate": 0.00019990476055781065, "loss": 4.8438, "step": 63950 }, { "epoch": 0.11943111600446933, "grad_norm": 0.9672982096672058, "learning_rate": 0.0001999045771661795, "loss": 4.7065, "step": 64000 }, { "epoch": 0.11952442156384782, "grad_norm": 1.009544014930725, "learning_rate": 0.00019990439359823437, "loss": 4.8137, "step": 64050 }, { "epoch": 0.11961772712322633, "grad_norm": 0.8747443556785583, "learning_rate": 0.00019990420985397552, "loss": 4.6859, "step": 64100 }, { "epoch": 0.11971103268260481, "grad_norm": 1.121934413909912, "learning_rate": 0.0001999040259334034, "loss": 4.8379, "step": 64150 }, { "epoch": 0.1198043382419833, "grad_norm": 0.8262189030647278, "learning_rate": 0.0001999038418365182, "loss": 4.7602, "step": 64200 }, { "epoch": 0.1198976438013618, "grad_norm": 0.9570197463035583, "learning_rate": 0.0001999036575633203, "loss": 4.8409, "step": 64250 }, { "epoch": 0.11999094936074028, "grad_norm": 0.7180740833282471, "learning_rate": 0.00019990347311381004, "loss": 4.9055, "step": 64300 }, { "epoch": 0.12008425492011877, "grad_norm": 0.7505408525466919, "learning_rate": 0.0001999032884879877, "loss": 4.8526, "step": 64350 }, { "epoch": 0.12017756047949726, "grad_norm": 0.9490662813186646, "learning_rate": 0.00019990310368585366, "loss": 4.6729, "step": 64400 }, { "epoch": 0.12027086603887577, "grad_norm": 1.0772948265075684, "learning_rate": 0.0001999029187074082, "loss": 4.6781, "step": 64450 }, { "epoch": 0.12036417159825426, "grad_norm": 0.6764106154441833, "learning_rate": 0.00019990273355265168, "loss": 4.8261, "step": 64500 }, { "epoch": 0.12045747715763275, "grad_norm": 0.8827162981033325, "learning_rate": 0.00019990254822158443, "loss": 4.8375, "step": 64550 }, { "epoch": 0.12055078271701124, "grad_norm": 0.9409701228141785, "learning_rate": 0.00019990236271420674, "loss": 4.8701, "step": 64600 }, { "epoch": 0.12064408827638973, "grad_norm": 0.8678632974624634, "learning_rate": 0.00019990217703051894, "loss": 4.9678, "step": 64650 }, { "epoch": 0.12073739383576822, "grad_norm": 1.1174262762069702, "learning_rate": 0.00019990199117052141, "loss": 4.8362, "step": 64700 }, { "epoch": 0.1208306993951467, "grad_norm": 0.8138838410377502, "learning_rate": 0.00019990180513421445, "loss": 4.9131, "step": 64750 }, { "epoch": 0.1209240049545252, "grad_norm": 1.0765495300292969, "learning_rate": 0.00019990161892159835, "loss": 4.8251, "step": 64800 }, { "epoch": 0.1210173105139037, "grad_norm": 1.0338910818099976, "learning_rate": 0.0001999014325326735, "loss": 4.4701, "step": 64850 }, { "epoch": 0.12111061607328219, "grad_norm": 0.7882349491119385, "learning_rate": 0.00019990124596744018, "loss": 4.6385, "step": 64900 }, { "epoch": 0.12120392163266068, "grad_norm": 1.1244629621505737, "learning_rate": 0.00019990105922589876, "loss": 4.6897, "step": 64950 }, { "epoch": 0.12129722719203917, "grad_norm": 1.0500319004058838, "learning_rate": 0.0001999008723080495, "loss": 4.6929, "step": 65000 }, { "epoch": 0.12139053275141766, "grad_norm": 0.9561808705329895, "learning_rate": 0.0001999006852138928, "loss": 4.866, "step": 65050 }, { "epoch": 0.12148383831079615, "grad_norm": 1.0104138851165771, "learning_rate": 0.000199900497943429, "loss": 4.8589, "step": 65100 }, { "epoch": 0.12157714387017464, "grad_norm": 0.975612998008728, "learning_rate": 0.00019990031049665837, "loss": 4.9448, "step": 65150 }, { "epoch": 0.12167044942955314, "grad_norm": 0.8425891399383545, "learning_rate": 0.0001999001228735813, "loss": 4.6693, "step": 65200 }, { "epoch": 0.12176375498893163, "grad_norm": 1.091673493385315, "learning_rate": 0.00019989993507419805, "loss": 4.7452, "step": 65250 }, { "epoch": 0.12185706054831012, "grad_norm": 1.0499212741851807, "learning_rate": 0.00019989974709850903, "loss": 4.9866, "step": 65300 }, { "epoch": 0.12195036610768861, "grad_norm": 0.9292690753936768, "learning_rate": 0.0001998995589465145, "loss": 4.9873, "step": 65350 }, { "epoch": 0.1220436716670671, "grad_norm": 0.9211663603782654, "learning_rate": 0.00019989937061821485, "loss": 4.8513, "step": 65400 }, { "epoch": 0.12213697722644559, "grad_norm": 1.0245435237884521, "learning_rate": 0.00019989918211361037, "loss": 4.813, "step": 65450 }, { "epoch": 0.12223028278582408, "grad_norm": 0.8180379867553711, "learning_rate": 0.00019989899343270145, "loss": 4.8255, "step": 65500 }, { "epoch": 0.12232358834520257, "grad_norm": 1.1613210439682007, "learning_rate": 0.00019989880457548832, "loss": 5.14, "step": 65550 }, { "epoch": 0.12241689390458108, "grad_norm": 1.1680033206939697, "learning_rate": 0.00019989861554197143, "loss": 4.7817, "step": 65600 }, { "epoch": 0.12251019946395957, "grad_norm": 1.1515231132507324, "learning_rate": 0.00019989842633215106, "loss": 4.8963, "step": 65650 }, { "epoch": 0.12260350502333806, "grad_norm": 0.9831023216247559, "learning_rate": 0.00019989823694602753, "loss": 4.613, "step": 65700 }, { "epoch": 0.12269681058271655, "grad_norm": 1.052646279335022, "learning_rate": 0.0001998980473836012, "loss": 4.7509, "step": 65750 }, { "epoch": 0.12279011614209503, "grad_norm": 0.8736461400985718, "learning_rate": 0.00019989785764487237, "loss": 4.8964, "step": 65800 }, { "epoch": 0.12288342170147352, "grad_norm": 0.9044458866119385, "learning_rate": 0.0001998976677298414, "loss": 4.8151, "step": 65850 }, { "epoch": 0.12297672726085201, "grad_norm": 0.8736430406570435, "learning_rate": 0.00019989747763850865, "loss": 4.6468, "step": 65900 }, { "epoch": 0.12307003282023052, "grad_norm": 1.0747185945510864, "learning_rate": 0.0001998972873708744, "loss": 4.8394, "step": 65950 }, { "epoch": 0.12316333837960901, "grad_norm": 1.0368112325668335, "learning_rate": 0.00019989709692693906, "loss": 4.7296, "step": 66000 }, { "epoch": 0.12316333837960901, "eval_loss": 4.993486404418945, "eval_runtime": 231.2413, "eval_samples_per_second": 11.278, "eval_steps_per_second": 11.278, "eval_tts_loss": 7.411762504489875, "step": 66000 }, { "epoch": 0.1232566439389875, "grad_norm": 1.0206661224365234, "learning_rate": 0.0001998969063067029, "loss": 4.8384, "step": 66050 }, { "epoch": 0.12334994949836599, "grad_norm": 1.2967489957809448, "learning_rate": 0.00019989671551016625, "loss": 4.6992, "step": 66100 }, { "epoch": 0.12344325505774448, "grad_norm": 0.8370565176010132, "learning_rate": 0.00019989652453732948, "loss": 4.7676, "step": 66150 }, { "epoch": 0.12353656061712297, "grad_norm": 1.1219327449798584, "learning_rate": 0.00019989633338819295, "loss": 4.913, "step": 66200 }, { "epoch": 0.12362986617650146, "grad_norm": 1.0357359647750854, "learning_rate": 0.00019989614206275696, "loss": 4.7011, "step": 66250 }, { "epoch": 0.12372317173587995, "grad_norm": 1.0418895483016968, "learning_rate": 0.00019989595056102183, "loss": 4.858, "step": 66300 }, { "epoch": 0.12381647729525845, "grad_norm": 0.9631625413894653, "learning_rate": 0.00019989575888298795, "loss": 4.9325, "step": 66350 }, { "epoch": 0.12390978285463694, "grad_norm": 1.2581260204315186, "learning_rate": 0.00019989556702865565, "loss": 4.7477, "step": 66400 }, { "epoch": 0.12400308841401543, "grad_norm": 0.7702711224555969, "learning_rate": 0.0001998953749980252, "loss": 4.9714, "step": 66450 }, { "epoch": 0.12409639397339392, "grad_norm": 1.0711308717727661, "learning_rate": 0.00019989518279109705, "loss": 4.836, "step": 66500 }, { "epoch": 0.12418969953277241, "grad_norm": 1.082642912864685, "learning_rate": 0.00019989499040787146, "loss": 4.7201, "step": 66550 }, { "epoch": 0.1242830050921509, "grad_norm": 1.1112911701202393, "learning_rate": 0.00019989479784834877, "loss": 4.6956, "step": 66600 }, { "epoch": 0.12437631065152939, "grad_norm": 0.9402649998664856, "learning_rate": 0.00019989460511252936, "loss": 4.6172, "step": 66650 }, { "epoch": 0.12446961621090788, "grad_norm": 1.1066251993179321, "learning_rate": 0.00019989441220041353, "loss": 4.7511, "step": 66700 }, { "epoch": 0.12456292177028638, "grad_norm": 1.2885663509368896, "learning_rate": 0.00019989421911200162, "loss": 4.8006, "step": 66750 }, { "epoch": 0.12465622732966487, "grad_norm": 0.717598557472229, "learning_rate": 0.00019989402584729405, "loss": 4.6387, "step": 66800 }, { "epoch": 0.12474953288904336, "grad_norm": 1.0635030269622803, "learning_rate": 0.00019989383240629106, "loss": 4.6984, "step": 66850 }, { "epoch": 0.12484283844842185, "grad_norm": 1.0408589839935303, "learning_rate": 0.00019989363878899306, "loss": 4.696, "step": 66900 }, { "epoch": 0.12493614400780034, "grad_norm": 0.8848728537559509, "learning_rate": 0.00019989344499540032, "loss": 4.6144, "step": 66950 }, { "epoch": 0.12502944956717885, "grad_norm": 0.9865880608558655, "learning_rate": 0.00019989325102551326, "loss": 4.8072, "step": 67000 }, { "epoch": 0.12512275512655732, "grad_norm": 2.2153608798980713, "learning_rate": 0.00019989305687933217, "loss": 4.7434, "step": 67050 }, { "epoch": 0.12521606068593583, "grad_norm": 0.9598948955535889, "learning_rate": 0.00019989286255685742, "loss": 4.7729, "step": 67100 }, { "epoch": 0.1253093662453143, "grad_norm": 1.0703397989273071, "learning_rate": 0.00019989266805808936, "loss": 4.9212, "step": 67150 }, { "epoch": 0.1254026718046928, "grad_norm": 0.954094648361206, "learning_rate": 0.0001998924733830283, "loss": 4.7929, "step": 67200 }, { "epoch": 0.12549597736407128, "grad_norm": 0.7662084102630615, "learning_rate": 0.0001998922785316746, "loss": 4.9457, "step": 67250 }, { "epoch": 0.12558928292344979, "grad_norm": 1.1596436500549316, "learning_rate": 0.0001998920835040286, "loss": 4.658, "step": 67300 }, { "epoch": 0.1256825884828283, "grad_norm": 1.0783514976501465, "learning_rate": 0.00019989188830009068, "loss": 4.9466, "step": 67350 }, { "epoch": 0.12577589404220677, "grad_norm": 0.8832924365997314, "learning_rate": 0.00019989169291986114, "loss": 4.7456, "step": 67400 }, { "epoch": 0.12586919960158527, "grad_norm": 0.9739583730697632, "learning_rate": 0.00019989149736334032, "loss": 4.9385, "step": 67450 }, { "epoch": 0.12596250516096374, "grad_norm": 1.0089017152786255, "learning_rate": 0.0001998913016305286, "loss": 4.6565, "step": 67500 }, { "epoch": 0.12605581072034225, "grad_norm": 0.9786615371704102, "learning_rate": 0.0001998911057214263, "loss": 4.7881, "step": 67550 }, { "epoch": 0.12614911627972072, "grad_norm": 0.9184353351593018, "learning_rate": 0.00019989090963603377, "loss": 4.779, "step": 67600 }, { "epoch": 0.12624242183909923, "grad_norm": 0.8026026487350464, "learning_rate": 0.00019989071337435136, "loss": 4.6157, "step": 67650 }, { "epoch": 0.12633572739847773, "grad_norm": 0.8393520712852478, "learning_rate": 0.00019989051693637943, "loss": 4.5899, "step": 67700 }, { "epoch": 0.1264290329578562, "grad_norm": 0.9405346512794495, "learning_rate": 0.00019989032032211834, "loss": 4.5016, "step": 67750 }, { "epoch": 0.1265223385172347, "grad_norm": 1.0097512006759644, "learning_rate": 0.00019989012353156834, "loss": 4.8901, "step": 67800 }, { "epoch": 0.1266156440766132, "grad_norm": 1.1094014644622803, "learning_rate": 0.0001998899265647299, "loss": 4.8235, "step": 67850 }, { "epoch": 0.1267089496359917, "grad_norm": 1.045980453491211, "learning_rate": 0.00019988972942160333, "loss": 4.8714, "step": 67900 }, { "epoch": 0.12680225519537017, "grad_norm": 0.897870659828186, "learning_rate": 0.0001998895321021889, "loss": 4.78, "step": 67950 }, { "epoch": 0.12689556075474867, "grad_norm": 0.8801417946815491, "learning_rate": 0.0001998893346064871, "loss": 4.7189, "step": 68000 }, { "epoch": 0.12698886631412717, "grad_norm": 0.8801950812339783, "learning_rate": 0.00019988913693449816, "loss": 4.817, "step": 68050 }, { "epoch": 0.12708217187350565, "grad_norm": 1.0675042867660522, "learning_rate": 0.00019988893908622245, "loss": 4.9578, "step": 68100 }, { "epoch": 0.12717547743288415, "grad_norm": 0.8822778463363647, "learning_rate": 0.00019988874106166037, "loss": 4.9018, "step": 68150 }, { "epoch": 0.12726878299226263, "grad_norm": 0.8060069680213928, "learning_rate": 0.00019988854286081225, "loss": 4.6307, "step": 68200 }, { "epoch": 0.12736208855164113, "grad_norm": 1.0484051704406738, "learning_rate": 0.00019988834448367838, "loss": 4.928, "step": 68250 }, { "epoch": 0.1274553941110196, "grad_norm": 1.0150614976882935, "learning_rate": 0.0001998881459302592, "loss": 4.7119, "step": 68300 }, { "epoch": 0.1275486996703981, "grad_norm": 0.8359283208847046, "learning_rate": 0.000199887947200555, "loss": 4.6128, "step": 68350 }, { "epoch": 0.1276420052297766, "grad_norm": 0.9599153399467468, "learning_rate": 0.00019988774829456617, "loss": 4.7391, "step": 68400 }, { "epoch": 0.1277353107891551, "grad_norm": 1.055501103401184, "learning_rate": 0.000199887549212293, "loss": 4.7485, "step": 68450 }, { "epoch": 0.1278286163485336, "grad_norm": 0.6903030872344971, "learning_rate": 0.0001998873499537359, "loss": 4.8311, "step": 68500 }, { "epoch": 0.12792192190791207, "grad_norm": 1.1893553733825684, "learning_rate": 0.0001998871505188952, "loss": 4.8558, "step": 68550 }, { "epoch": 0.12801522746729058, "grad_norm": 1.0602223873138428, "learning_rate": 0.00019988695090777126, "loss": 4.8572, "step": 68600 }, { "epoch": 0.12810853302666905, "grad_norm": 1.0478551387786865, "learning_rate": 0.00019988675112036442, "loss": 4.8788, "step": 68650 }, { "epoch": 0.12820183858604756, "grad_norm": 0.6287881731987, "learning_rate": 0.00019988655115667507, "loss": 4.8575, "step": 68700 }, { "epoch": 0.12829514414542603, "grad_norm": 1.2025055885314941, "learning_rate": 0.0001998863510167035, "loss": 4.5946, "step": 68750 }, { "epoch": 0.12838844970480454, "grad_norm": 1.10360586643219, "learning_rate": 0.0001998861507004501, "loss": 4.8477, "step": 68800 }, { "epoch": 0.12848175526418304, "grad_norm": 1.1787991523742676, "learning_rate": 0.00019988595020791524, "loss": 4.5538, "step": 68850 }, { "epoch": 0.12857506082356152, "grad_norm": 1.1096701622009277, "learning_rate": 0.00019988574953909923, "loss": 4.8392, "step": 68900 }, { "epoch": 0.12866836638294002, "grad_norm": 1.0045229196548462, "learning_rate": 0.00019988554869400247, "loss": 4.8437, "step": 68950 }, { "epoch": 0.1287616719423185, "grad_norm": 0.7254719734191895, "learning_rate": 0.00019988534767262527, "loss": 4.6926, "step": 69000 }, { "epoch": 0.1287616719423185, "eval_loss": 4.980347156524658, "eval_runtime": 233.3792, "eval_samples_per_second": 11.175, "eval_steps_per_second": 11.175, "eval_tts_loss": 7.437576909478114, "step": 69000 }, { "epoch": 0.128854977501697, "grad_norm": 1.043537974357605, "learning_rate": 0.00019988514647496802, "loss": 4.7274, "step": 69050 }, { "epoch": 0.12894828306107547, "grad_norm": 1.02115797996521, "learning_rate": 0.00019988494510103104, "loss": 4.7884, "step": 69100 }, { "epoch": 0.12904158862045398, "grad_norm": 1.1048721075057983, "learning_rate": 0.00019988474355081475, "loss": 4.4432, "step": 69150 }, { "epoch": 0.12913489417983248, "grad_norm": 0.9006786942481995, "learning_rate": 0.0001998845418243194, "loss": 4.703, "step": 69200 }, { "epoch": 0.12922819973921096, "grad_norm": 0.9208207130432129, "learning_rate": 0.00019988433992154545, "loss": 4.6751, "step": 69250 }, { "epoch": 0.12932150529858946, "grad_norm": 0.9180870652198792, "learning_rate": 0.00019988413784249322, "loss": 4.5272, "step": 69300 }, { "epoch": 0.12941481085796794, "grad_norm": 1.102773904800415, "learning_rate": 0.00019988393558716306, "loss": 4.7114, "step": 69350 }, { "epoch": 0.12950811641734644, "grad_norm": 0.9914569854736328, "learning_rate": 0.00019988373315555532, "loss": 4.8268, "step": 69400 }, { "epoch": 0.12960142197672492, "grad_norm": 0.9935908913612366, "learning_rate": 0.00019988353054767037, "loss": 4.8687, "step": 69450 }, { "epoch": 0.12969472753610342, "grad_norm": 0.8525168895721436, "learning_rate": 0.00019988332776350856, "loss": 4.7935, "step": 69500 }, { "epoch": 0.12978803309548193, "grad_norm": 0.8857572674751282, "learning_rate": 0.00019988312480307026, "loss": 4.7981, "step": 69550 }, { "epoch": 0.1298813386548604, "grad_norm": 1.1465469598770142, "learning_rate": 0.0001998829216663558, "loss": 4.6701, "step": 69600 }, { "epoch": 0.1299746442142389, "grad_norm": 0.91568523645401, "learning_rate": 0.00019988271835336558, "loss": 4.7749, "step": 69650 }, { "epoch": 0.13006794977361738, "grad_norm": 0.9274788498878479, "learning_rate": 0.00019988251486409992, "loss": 4.6415, "step": 69700 }, { "epoch": 0.13016125533299588, "grad_norm": 1.1520894765853882, "learning_rate": 0.0001998823111985592, "loss": 4.5849, "step": 69750 }, { "epoch": 0.13025456089237436, "grad_norm": 1.3699066638946533, "learning_rate": 0.0001998821073567438, "loss": 4.7611, "step": 69800 }, { "epoch": 0.13034786645175286, "grad_norm": 0.9259943962097168, "learning_rate": 0.00019988190333865402, "loss": 4.8313, "step": 69850 }, { "epoch": 0.13044117201113134, "grad_norm": 0.8618115782737732, "learning_rate": 0.00019988169914429026, "loss": 4.6103, "step": 69900 }, { "epoch": 0.13053447757050984, "grad_norm": 1.0242559909820557, "learning_rate": 0.0001998814947736529, "loss": 4.9478, "step": 69950 }, { "epoch": 0.13062778312988835, "grad_norm": 1.2048753499984741, "learning_rate": 0.00019988129022674228, "loss": 4.9015, "step": 70000 }, { "epoch": 0.13072108868926682, "grad_norm": 0.8471056818962097, "learning_rate": 0.0001998810855035587, "loss": 4.5826, "step": 70050 }, { "epoch": 0.13081439424864533, "grad_norm": 0.9561209082603455, "learning_rate": 0.00019988088060410265, "loss": 4.7941, "step": 70100 }, { "epoch": 0.1309076998080238, "grad_norm": 0.6860484480857849, "learning_rate": 0.00019988067552837439, "loss": 4.6454, "step": 70150 }, { "epoch": 0.1310010053674023, "grad_norm": 1.028444528579712, "learning_rate": 0.0001998804702763743, "loss": 4.8362, "step": 70200 }, { "epoch": 0.13109431092678078, "grad_norm": 0.9688596129417419, "learning_rate": 0.00019988026484810275, "loss": 4.8212, "step": 70250 }, { "epoch": 0.1311876164861593, "grad_norm": 1.0312488079071045, "learning_rate": 0.00019988005924356013, "loss": 4.8123, "step": 70300 }, { "epoch": 0.1312809220455378, "grad_norm": 0.8787838220596313, "learning_rate": 0.0001998798534627468, "loss": 4.6147, "step": 70350 }, { "epoch": 0.13137422760491627, "grad_norm": 1.1073654890060425, "learning_rate": 0.00019987964750566305, "loss": 4.7933, "step": 70400 }, { "epoch": 0.13146753316429477, "grad_norm": 1.085288643836975, "learning_rate": 0.00019987944137230934, "loss": 4.8248, "step": 70450 }, { "epoch": 0.13156083872367325, "grad_norm": 1.2170988321304321, "learning_rate": 0.00019987923506268595, "loss": 4.783, "step": 70500 }, { "epoch": 0.13165414428305175, "grad_norm": 1.0648763179779053, "learning_rate": 0.00019987902857679333, "loss": 4.9425, "step": 70550 }, { "epoch": 0.13174744984243023, "grad_norm": 1.1931159496307373, "learning_rate": 0.00019987882191463177, "loss": 4.6804, "step": 70600 }, { "epoch": 0.13184075540180873, "grad_norm": 0.9888035655021667, "learning_rate": 0.00019987861507620168, "loss": 4.7612, "step": 70650 }, { "epoch": 0.13193406096118723, "grad_norm": 0.9140545725822449, "learning_rate": 0.0001998784080615034, "loss": 4.6891, "step": 70700 }, { "epoch": 0.1320273665205657, "grad_norm": 1.096142053604126, "learning_rate": 0.00019987820087053728, "loss": 4.9494, "step": 70750 }, { "epoch": 0.1321206720799442, "grad_norm": 1.1193305253982544, "learning_rate": 0.00019987799350330374, "loss": 4.8832, "step": 70800 }, { "epoch": 0.1322139776393227, "grad_norm": 1.2808094024658203, "learning_rate": 0.0001998777859598031, "loss": 4.9684, "step": 70850 }, { "epoch": 0.1323072831987012, "grad_norm": 0.852861225605011, "learning_rate": 0.00019987757824003577, "loss": 4.911, "step": 70900 }, { "epoch": 0.13240058875807967, "grad_norm": 0.8314254283905029, "learning_rate": 0.00019987737034400204, "loss": 4.6268, "step": 70950 }, { "epoch": 0.13249389431745817, "grad_norm": 1.0888328552246094, "learning_rate": 0.00019987716227170234, "loss": 4.837, "step": 71000 }, { "epoch": 0.13258719987683665, "grad_norm": 0.9562662243843079, "learning_rate": 0.00019987695402313706, "loss": 4.8852, "step": 71050 }, { "epoch": 0.13268050543621515, "grad_norm": 1.0483977794647217, "learning_rate": 0.00019987674559830652, "loss": 4.6314, "step": 71100 }, { "epoch": 0.13277381099559366, "grad_norm": 1.0416483879089355, "learning_rate": 0.00019987653699721106, "loss": 4.6485, "step": 71150 }, { "epoch": 0.13286711655497213, "grad_norm": 0.9139002561569214, "learning_rate": 0.00019987632821985113, "loss": 4.4885, "step": 71200 }, { "epoch": 0.13296042211435063, "grad_norm": 1.0126796960830688, "learning_rate": 0.00019987611926622703, "loss": 4.7368, "step": 71250 }, { "epoch": 0.1330537276737291, "grad_norm": 0.8071941137313843, "learning_rate": 0.00019987591013633912, "loss": 4.7529, "step": 71300 }, { "epoch": 0.13314703323310761, "grad_norm": 0.8318772912025452, "learning_rate": 0.00019987570083018788, "loss": 4.686, "step": 71350 }, { "epoch": 0.1332403387924861, "grad_norm": 1.0917198657989502, "learning_rate": 0.00019987549134777355, "loss": 4.6129, "step": 71400 }, { "epoch": 0.1333336443518646, "grad_norm": 1.0430625677108765, "learning_rate": 0.00019987528168909655, "loss": 4.5787, "step": 71450 }, { "epoch": 0.1334269499112431, "grad_norm": 1.307949423789978, "learning_rate": 0.00019987507185415727, "loss": 4.7455, "step": 71500 }, { "epoch": 0.13352025547062157, "grad_norm": 0.9685766696929932, "learning_rate": 0.00019987486184295606, "loss": 4.7564, "step": 71550 }, { "epoch": 0.13361356103000008, "grad_norm": 0.9537215232849121, "learning_rate": 0.00019987465165549327, "loss": 4.7439, "step": 71600 }, { "epoch": 0.13370686658937855, "grad_norm": 0.8062765598297119, "learning_rate": 0.00019987444129176933, "loss": 4.6103, "step": 71650 }, { "epoch": 0.13380017214875706, "grad_norm": 1.0107309818267822, "learning_rate": 0.00019987423075178455, "loss": 4.8355, "step": 71700 }, { "epoch": 0.13389347770813553, "grad_norm": 0.8485686779022217, "learning_rate": 0.00019987402003553934, "loss": 4.7971, "step": 71750 }, { "epoch": 0.13398678326751404, "grad_norm": 1.2249001264572144, "learning_rate": 0.00019987380914303403, "loss": 4.823, "step": 71800 }, { "epoch": 0.13408008882689254, "grad_norm": 1.0090044736862183, "learning_rate": 0.00019987359807426905, "loss": 4.9343, "step": 71850 }, { "epoch": 0.13417339438627102, "grad_norm": 1.018228530883789, "learning_rate": 0.00019987338682924476, "loss": 4.8365, "step": 71900 }, { "epoch": 0.13426669994564952, "grad_norm": 0.947043240070343, "learning_rate": 0.00019987317540796147, "loss": 4.7334, "step": 71950 }, { "epoch": 0.134360005505028, "grad_norm": 1.1589524745941162, "learning_rate": 0.00019987296381041963, "loss": 4.6444, "step": 72000 }, { "epoch": 0.134360005505028, "eval_loss": 4.982792377471924, "eval_runtime": 233.4346, "eval_samples_per_second": 11.172, "eval_steps_per_second": 11.172, "eval_tts_loss": 7.417768215367777, "step": 72000 }, { "epoch": 0.1344533110644065, "grad_norm": 1.0631495714187622, "learning_rate": 0.00019987275203661956, "loss": 4.8445, "step": 72050 }, { "epoch": 0.13454661662378498, "grad_norm": 1.056353211402893, "learning_rate": 0.00019987254008656165, "loss": 4.8737, "step": 72100 }, { "epoch": 0.13463992218316348, "grad_norm": 1.192908525466919, "learning_rate": 0.0001998723279602463, "loss": 4.7873, "step": 72150 }, { "epoch": 0.13473322774254198, "grad_norm": 1.1714168787002563, "learning_rate": 0.00019987211565767384, "loss": 4.7542, "step": 72200 }, { "epoch": 0.13482653330192046, "grad_norm": 0.7499253153800964, "learning_rate": 0.00019987190317884469, "loss": 4.8558, "step": 72250 }, { "epoch": 0.13491983886129896, "grad_norm": 1.1434695720672607, "learning_rate": 0.0001998716905237592, "loss": 4.684, "step": 72300 }, { "epoch": 0.13501314442067744, "grad_norm": 0.9762905836105347, "learning_rate": 0.00019987147769241775, "loss": 4.6689, "step": 72350 }, { "epoch": 0.13510644998005594, "grad_norm": 1.05728018283844, "learning_rate": 0.00019987126468482073, "loss": 4.7129, "step": 72400 }, { "epoch": 0.13519975553943442, "grad_norm": 1.0053168535232544, "learning_rate": 0.00019987105150096849, "loss": 4.6524, "step": 72450 }, { "epoch": 0.13529306109881292, "grad_norm": 0.9576205611228943, "learning_rate": 0.0001998708381408614, "loss": 4.6339, "step": 72500 }, { "epoch": 0.1353863666581914, "grad_norm": 1.0552775859832764, "learning_rate": 0.00019987062460449984, "loss": 4.6316, "step": 72550 }, { "epoch": 0.1354796722175699, "grad_norm": 1.1562297344207764, "learning_rate": 0.00019987041089188426, "loss": 4.7064, "step": 72600 }, { "epoch": 0.1355729777769484, "grad_norm": 1.0458017587661743, "learning_rate": 0.00019987019700301494, "loss": 4.8765, "step": 72650 }, { "epoch": 0.13566628333632688, "grad_norm": 0.8299524784088135, "learning_rate": 0.00019986998293789233, "loss": 4.8115, "step": 72700 }, { "epoch": 0.13575958889570539, "grad_norm": 1.1313036680221558, "learning_rate": 0.0001998697686965167, "loss": 4.7096, "step": 72750 }, { "epoch": 0.13585289445508386, "grad_norm": 1.0760210752487183, "learning_rate": 0.00019986955427888855, "loss": 4.886, "step": 72800 }, { "epoch": 0.13594620001446237, "grad_norm": 1.1115272045135498, "learning_rate": 0.0001998693396850082, "loss": 4.8623, "step": 72850 }, { "epoch": 0.13603950557384084, "grad_norm": 0.6817601919174194, "learning_rate": 0.00019986912491487603, "loss": 4.7888, "step": 72900 }, { "epoch": 0.13613281113321934, "grad_norm": 1.2269861698150635, "learning_rate": 0.00019986890996849245, "loss": 4.8229, "step": 72950 }, { "epoch": 0.13622611669259785, "grad_norm": 1.13284170627594, "learning_rate": 0.00019986869484585782, "loss": 4.7043, "step": 73000 }, { "epoch": 0.13631942225197632, "grad_norm": 1.085242748260498, "learning_rate": 0.0001998684795469725, "loss": 4.8635, "step": 73050 }, { "epoch": 0.13641272781135483, "grad_norm": 0.9383891224861145, "learning_rate": 0.00019986826407183687, "loss": 4.7206, "step": 73100 }, { "epoch": 0.1365060333707333, "grad_norm": 0.9130173921585083, "learning_rate": 0.00019986804842045134, "loss": 4.783, "step": 73150 }, { "epoch": 0.1365993389301118, "grad_norm": 0.9875626564025879, "learning_rate": 0.0001998678325928163, "loss": 4.6706, "step": 73200 }, { "epoch": 0.13669264448949028, "grad_norm": 0.775663435459137, "learning_rate": 0.00019986761658893207, "loss": 4.5753, "step": 73250 }, { "epoch": 0.1367859500488688, "grad_norm": 0.9947422742843628, "learning_rate": 0.00019986740040879908, "loss": 4.59, "step": 73300 }, { "epoch": 0.1368792556082473, "grad_norm": 0.7354251742362976, "learning_rate": 0.00019986718405241773, "loss": 4.5268, "step": 73350 }, { "epoch": 0.13697256116762577, "grad_norm": 0.9858702421188354, "learning_rate": 0.00019986696751978834, "loss": 4.7284, "step": 73400 }, { "epoch": 0.13706586672700427, "grad_norm": 1.1084253787994385, "learning_rate": 0.00019986675081091134, "loss": 4.864, "step": 73450 }, { "epoch": 0.13715917228638275, "grad_norm": 1.0216785669326782, "learning_rate": 0.00019986653392578706, "loss": 4.6343, "step": 73500 }, { "epoch": 0.13725247784576125, "grad_norm": 1.0096267461776733, "learning_rate": 0.00019986631686441596, "loss": 4.7856, "step": 73550 }, { "epoch": 0.13734578340513973, "grad_norm": 0.9455108642578125, "learning_rate": 0.0001998660996267984, "loss": 4.6775, "step": 73600 }, { "epoch": 0.13743908896451823, "grad_norm": 1.4153708219528198, "learning_rate": 0.0001998658822129347, "loss": 4.6638, "step": 73650 }, { "epoch": 0.13753239452389673, "grad_norm": 0.8518409729003906, "learning_rate": 0.00019986566462282533, "loss": 4.9213, "step": 73700 }, { "epoch": 0.1376257000832752, "grad_norm": 1.1150931119918823, "learning_rate": 0.0001998654468564706, "loss": 4.7676, "step": 73750 }, { "epoch": 0.1377190056426537, "grad_norm": 0.9590360522270203, "learning_rate": 0.00019986522891387093, "loss": 4.9231, "step": 73800 }, { "epoch": 0.1378123112020322, "grad_norm": 0.7029619216918945, "learning_rate": 0.0001998650107950267, "loss": 4.692, "step": 73850 }, { "epoch": 0.1379056167614107, "grad_norm": 0.8161832690238953, "learning_rate": 0.00019986479249993832, "loss": 4.7669, "step": 73900 }, { "epoch": 0.13799892232078917, "grad_norm": 1.2208623886108398, "learning_rate": 0.00019986457402860618, "loss": 4.774, "step": 73950 }, { "epoch": 0.13809222788016767, "grad_norm": 0.6693832278251648, "learning_rate": 0.00019986435538103058, "loss": 4.6447, "step": 74000 }, { "epoch": 0.13818553343954615, "grad_norm": 1.1013367176055908, "learning_rate": 0.000199864136557212, "loss": 4.9094, "step": 74050 }, { "epoch": 0.13827883899892465, "grad_norm": 0.6470247507095337, "learning_rate": 0.00019986391755715075, "loss": 4.7323, "step": 74100 }, { "epoch": 0.13837214455830316, "grad_norm": 1.0333960056304932, "learning_rate": 0.0001998636983808473, "loss": 4.6899, "step": 74150 }, { "epoch": 0.13846545011768163, "grad_norm": 1.0309245586395264, "learning_rate": 0.00019986347902830196, "loss": 4.8887, "step": 74200 }, { "epoch": 0.13855875567706014, "grad_norm": 1.0553001165390015, "learning_rate": 0.00019986325949951516, "loss": 4.654, "step": 74250 }, { "epoch": 0.1386520612364386, "grad_norm": 1.3011856079101562, "learning_rate": 0.0001998630397944873, "loss": 4.8259, "step": 74300 }, { "epoch": 0.13874536679581712, "grad_norm": 1.04082190990448, "learning_rate": 0.00019986281991321873, "loss": 4.9014, "step": 74350 }, { "epoch": 0.1388386723551956, "grad_norm": 0.9634092450141907, "learning_rate": 0.00019986259985570987, "loss": 4.7242, "step": 74400 }, { "epoch": 0.1389319779145741, "grad_norm": 1.2973626852035522, "learning_rate": 0.00019986237962196107, "loss": 4.7606, "step": 74450 }, { "epoch": 0.1390252834739526, "grad_norm": 0.8108509182929993, "learning_rate": 0.00019986215921197277, "loss": 4.7773, "step": 74500 }, { "epoch": 0.13911858903333107, "grad_norm": 1.048374891281128, "learning_rate": 0.0001998619386257453, "loss": 4.8176, "step": 74550 }, { "epoch": 0.13921189459270958, "grad_norm": 0.9757275581359863, "learning_rate": 0.0001998617178632791, "loss": 4.5193, "step": 74600 }, { "epoch": 0.13930520015208805, "grad_norm": 1.088984727859497, "learning_rate": 0.00019986149692457453, "loss": 4.738, "step": 74650 }, { "epoch": 0.13939850571146656, "grad_norm": 1.0206042528152466, "learning_rate": 0.00019986127580963199, "loss": 4.7663, "step": 74700 }, { "epoch": 0.13949181127084503, "grad_norm": 0.859443724155426, "learning_rate": 0.00019986105451845186, "loss": 4.7207, "step": 74750 }, { "epoch": 0.13958511683022354, "grad_norm": 0.9786943793296814, "learning_rate": 0.00019986083305103457, "loss": 4.8639, "step": 74800 }, { "epoch": 0.13967842238960204, "grad_norm": 1.2023348808288574, "learning_rate": 0.00019986061140738043, "loss": 4.7091, "step": 74850 }, { "epoch": 0.13977172794898052, "grad_norm": 1.1484274864196777, "learning_rate": 0.00019986038958748993, "loss": 4.6658, "step": 74900 }, { "epoch": 0.13986503350835902, "grad_norm": 1.0033177137374878, "learning_rate": 0.0001998601675913634, "loss": 4.8025, "step": 74950 }, { "epoch": 0.1399583390677375, "grad_norm": 1.1093839406967163, "learning_rate": 0.00019985994541900124, "loss": 4.7705, "step": 75000 }, { "epoch": 0.1399583390677375, "eval_loss": 4.963533878326416, "eval_runtime": 229.2936, "eval_samples_per_second": 11.374, "eval_steps_per_second": 11.374, "eval_tts_loss": 7.357835572455764, "step": 75000 }, { "epoch": 0.140051644627116, "grad_norm": 0.8389537930488586, "learning_rate": 0.00019985972307040383, "loss": 4.8843, "step": 75050 }, { "epoch": 0.14014495018649448, "grad_norm": 1.009690761566162, "learning_rate": 0.00019985950054557164, "loss": 4.6147, "step": 75100 }, { "epoch": 0.14023825574587298, "grad_norm": 1.0867621898651123, "learning_rate": 0.00019985927784450493, "loss": 4.7178, "step": 75150 }, { "epoch": 0.14033156130525146, "grad_norm": 0.7208017706871033, "learning_rate": 0.0001998590549672042, "loss": 4.918, "step": 75200 }, { "epoch": 0.14042486686462996, "grad_norm": 0.9712778925895691, "learning_rate": 0.0001998588319136698, "loss": 4.8301, "step": 75250 }, { "epoch": 0.14051817242400846, "grad_norm": 0.7589443922042847, "learning_rate": 0.00019985860868390213, "loss": 4.7432, "step": 75300 }, { "epoch": 0.14061147798338694, "grad_norm": 0.7034851312637329, "learning_rate": 0.0001998583852779016, "loss": 4.6202, "step": 75350 }, { "epoch": 0.14070478354276544, "grad_norm": 1.2004735469818115, "learning_rate": 0.0001998581616956686, "loss": 4.4097, "step": 75400 }, { "epoch": 0.14079808910214392, "grad_norm": 1.1603310108184814, "learning_rate": 0.0001998579379372035, "loss": 4.8106, "step": 75450 }, { "epoch": 0.14089139466152242, "grad_norm": 0.9474809169769287, "learning_rate": 0.00019985771400250668, "loss": 4.6189, "step": 75500 }, { "epoch": 0.1409847002209009, "grad_norm": 0.9576531052589417, "learning_rate": 0.0001998574898915786, "loss": 4.6265, "step": 75550 }, { "epoch": 0.1410780057802794, "grad_norm": 0.7832580804824829, "learning_rate": 0.00019985726560441962, "loss": 4.656, "step": 75600 }, { "epoch": 0.1411713113396579, "grad_norm": 1.3878700733184814, "learning_rate": 0.00019985704114103015, "loss": 4.5926, "step": 75650 }, { "epoch": 0.14126461689903638, "grad_norm": 0.8612073659896851, "learning_rate": 0.00019985681650141053, "loss": 4.8437, "step": 75700 }, { "epoch": 0.1413579224584149, "grad_norm": 1.2866979837417603, "learning_rate": 0.00019985659168556124, "loss": 4.9809, "step": 75750 }, { "epoch": 0.14145122801779336, "grad_norm": 1.1674981117248535, "learning_rate": 0.0001998563666934826, "loss": 4.7633, "step": 75800 }, { "epoch": 0.14154453357717187, "grad_norm": 1.0047423839569092, "learning_rate": 0.0001998561415251751, "loss": 4.6555, "step": 75850 }, { "epoch": 0.14163783913655034, "grad_norm": 0.8399094343185425, "learning_rate": 0.00019985591618063903, "loss": 5.0474, "step": 75900 }, { "epoch": 0.14173114469592885, "grad_norm": 1.3186641931533813, "learning_rate": 0.00019985569065987484, "loss": 4.7141, "step": 75950 }, { "epoch": 0.14182445025530735, "grad_norm": 1.2126065492630005, "learning_rate": 0.00019985546496288295, "loss": 4.5208, "step": 76000 }, { "epoch": 0.14191775581468583, "grad_norm": 1.0247917175292969, "learning_rate": 0.0001998552390896637, "loss": 4.6857, "step": 76050 }, { "epoch": 0.14201106137406433, "grad_norm": 0.8075127005577087, "learning_rate": 0.00019985501304021756, "loss": 4.6593, "step": 76100 }, { "epoch": 0.1421043669334428, "grad_norm": 0.8528395295143127, "learning_rate": 0.00019985478681454487, "loss": 4.7271, "step": 76150 }, { "epoch": 0.1421976724928213, "grad_norm": 0.9807457327842712, "learning_rate": 0.00019985456041264605, "loss": 4.5915, "step": 76200 }, { "epoch": 0.14229097805219978, "grad_norm": 1.1197192668914795, "learning_rate": 0.0001998543338345215, "loss": 4.4433, "step": 76250 }, { "epoch": 0.1423842836115783, "grad_norm": 1.044595718383789, "learning_rate": 0.00019985410708017165, "loss": 4.7313, "step": 76300 }, { "epoch": 0.1424775891709568, "grad_norm": 0.8226972818374634, "learning_rate": 0.00019985388014959684, "loss": 4.5888, "step": 76350 }, { "epoch": 0.14257089473033527, "grad_norm": 0.9106892943382263, "learning_rate": 0.00019985365304279752, "loss": 4.7119, "step": 76400 }, { "epoch": 0.14266420028971377, "grad_norm": 1.2753225564956665, "learning_rate": 0.00019985342575977406, "loss": 4.7376, "step": 76450 }, { "epoch": 0.14275750584909225, "grad_norm": 0.9028472304344177, "learning_rate": 0.00019985319830052688, "loss": 4.6413, "step": 76500 }, { "epoch": 0.14285081140847075, "grad_norm": 1.0299334526062012, "learning_rate": 0.00019985297066505637, "loss": 4.6844, "step": 76550 }, { "epoch": 0.14294411696784923, "grad_norm": 1.1871273517608643, "learning_rate": 0.00019985274285336293, "loss": 4.8803, "step": 76600 }, { "epoch": 0.14303742252722773, "grad_norm": 0.5859182476997375, "learning_rate": 0.000199852514865447, "loss": 4.627, "step": 76650 }, { "epoch": 0.1431307280866062, "grad_norm": 0.8444706201553345, "learning_rate": 0.00019985228670130894, "loss": 4.7149, "step": 76700 }, { "epoch": 0.1432240336459847, "grad_norm": 0.889392614364624, "learning_rate": 0.0001998520583609491, "loss": 4.6281, "step": 76750 }, { "epoch": 0.14331733920536321, "grad_norm": 0.9411704540252686, "learning_rate": 0.00019985182984436802, "loss": 4.6086, "step": 76800 }, { "epoch": 0.1434106447647417, "grad_norm": 0.8570734262466431, "learning_rate": 0.000199851601151566, "loss": 4.7565, "step": 76850 }, { "epoch": 0.1435039503241202, "grad_norm": 0.8784626722335815, "learning_rate": 0.0001998513722825435, "loss": 4.891, "step": 76900 }, { "epoch": 0.14359725588349867, "grad_norm": 1.0024194717407227, "learning_rate": 0.00019985114323730087, "loss": 4.7093, "step": 76950 }, { "epoch": 0.14369056144287717, "grad_norm": 1.1379250288009644, "learning_rate": 0.00019985091401583855, "loss": 4.8616, "step": 77000 }, { "epoch": 0.14378386700225565, "grad_norm": 0.9281609058380127, "learning_rate": 0.00019985068461815693, "loss": 4.7413, "step": 77050 }, { "epoch": 0.14387717256163415, "grad_norm": 1.4726687669754028, "learning_rate": 0.00019985045504425643, "loss": 4.7395, "step": 77100 }, { "epoch": 0.14397047812101266, "grad_norm": 1.1294130086898804, "learning_rate": 0.00019985022529413744, "loss": 4.8413, "step": 77150 }, { "epoch": 0.14406378368039113, "grad_norm": 0.7833269834518433, "learning_rate": 0.00019984999536780038, "loss": 4.7995, "step": 77200 }, { "epoch": 0.14415708923976964, "grad_norm": 0.861925482749939, "learning_rate": 0.00019984976526524567, "loss": 4.7661, "step": 77250 }, { "epoch": 0.1442503947991481, "grad_norm": 1.15879225730896, "learning_rate": 0.00019984953498647365, "loss": 4.8043, "step": 77300 }, { "epoch": 0.14434370035852662, "grad_norm": 0.854045569896698, "learning_rate": 0.0001998493045314848, "loss": 4.6388, "step": 77350 }, { "epoch": 0.1444370059179051, "grad_norm": 1.0074321031570435, "learning_rate": 0.00019984907390027944, "loss": 4.7215, "step": 77400 }, { "epoch": 0.1445303114772836, "grad_norm": 0.9455142617225647, "learning_rate": 0.0001998488430928581, "loss": 4.7325, "step": 77450 }, { "epoch": 0.1446236170366621, "grad_norm": 1.0225214958190918, "learning_rate": 0.00019984861210922109, "loss": 4.7662, "step": 77500 }, { "epoch": 0.14471692259604058, "grad_norm": 0.8790587186813354, "learning_rate": 0.00019984838094936884, "loss": 4.7582, "step": 77550 }, { "epoch": 0.14481022815541908, "grad_norm": 0.7881041169166565, "learning_rate": 0.00019984814961330178, "loss": 4.7529, "step": 77600 }, { "epoch": 0.14490353371479756, "grad_norm": 0.9009997844696045, "learning_rate": 0.00019984791810102032, "loss": 4.8476, "step": 77650 }, { "epoch": 0.14499683927417606, "grad_norm": 1.1333011388778687, "learning_rate": 0.00019984768641252484, "loss": 4.7743, "step": 77700 }, { "epoch": 0.14509014483355454, "grad_norm": 1.0094295740127563, "learning_rate": 0.00019984745454781577, "loss": 4.7856, "step": 77750 }, { "epoch": 0.14518345039293304, "grad_norm": 1.1273713111877441, "learning_rate": 0.00019984722250689348, "loss": 4.6921, "step": 77800 }, { "epoch": 0.14527675595231151, "grad_norm": 1.0432093143463135, "learning_rate": 0.00019984699028975845, "loss": 4.6715, "step": 77850 }, { "epoch": 0.14537006151169002, "grad_norm": 1.0306214094161987, "learning_rate": 0.00019984675789641102, "loss": 4.9655, "step": 77900 }, { "epoch": 0.14546336707106852, "grad_norm": 0.5953212380409241, "learning_rate": 0.00019984652532685165, "loss": 4.6442, "step": 77950 }, { "epoch": 0.145556672630447, "grad_norm": 0.900357723236084, "learning_rate": 0.00019984629258108073, "loss": 4.6404, "step": 78000 }, { "epoch": 0.145556672630447, "eval_loss": 4.966691017150879, "eval_runtime": 229.295, "eval_samples_per_second": 11.374, "eval_steps_per_second": 11.374, "eval_tts_loss": 7.41557245793541, "step": 78000 }, { "epoch": 0.1456499781898255, "grad_norm": 0.9374669194221497, "learning_rate": 0.00019984605965909865, "loss": 4.6393, "step": 78050 }, { "epoch": 0.14574328374920398, "grad_norm": 0.7869539856910706, "learning_rate": 0.00019984582656090586, "loss": 4.5879, "step": 78100 }, { "epoch": 0.14583658930858248, "grad_norm": 0.6721176505088806, "learning_rate": 0.00019984559328650278, "loss": 4.8905, "step": 78150 }, { "epoch": 0.14592989486796096, "grad_norm": 0.9904831647872925, "learning_rate": 0.00019984535983588977, "loss": 4.8523, "step": 78200 }, { "epoch": 0.14602320042733946, "grad_norm": 0.9806076884269714, "learning_rate": 0.00019984512620906727, "loss": 4.8636, "step": 78250 }, { "epoch": 0.14611650598671796, "grad_norm": 1.2187144756317139, "learning_rate": 0.0001998448924060357, "loss": 4.7684, "step": 78300 }, { "epoch": 0.14620981154609644, "grad_norm": 0.9885228276252747, "learning_rate": 0.00019984465842679545, "loss": 4.743, "step": 78350 }, { "epoch": 0.14630311710547494, "grad_norm": 0.8706178665161133, "learning_rate": 0.00019984442427134694, "loss": 4.7663, "step": 78400 }, { "epoch": 0.14639642266485342, "grad_norm": 0.9276357889175415, "learning_rate": 0.0001998441899396906, "loss": 4.9867, "step": 78450 }, { "epoch": 0.14648972822423192, "grad_norm": 1.1178739070892334, "learning_rate": 0.00019984395543182687, "loss": 4.7049, "step": 78500 }, { "epoch": 0.1465830337836104, "grad_norm": 0.9989400506019592, "learning_rate": 0.00019984372074775605, "loss": 4.6692, "step": 78550 }, { "epoch": 0.1466763393429889, "grad_norm": 1.0636882781982422, "learning_rate": 0.0001998434858874787, "loss": 4.8389, "step": 78600 }, { "epoch": 0.1467696449023674, "grad_norm": 1.0234490633010864, "learning_rate": 0.00019984325085099515, "loss": 4.7642, "step": 78650 }, { "epoch": 0.14686295046174588, "grad_norm": 0.9406095743179321, "learning_rate": 0.00019984301563830579, "loss": 4.8147, "step": 78700 }, { "epoch": 0.1469562560211244, "grad_norm": 1.0061767101287842, "learning_rate": 0.0001998427802494111, "loss": 4.8082, "step": 78750 }, { "epoch": 0.14704956158050286, "grad_norm": 0.9918909668922424, "learning_rate": 0.00019984254468431148, "loss": 4.7704, "step": 78800 }, { "epoch": 0.14714286713988137, "grad_norm": 1.0421675443649292, "learning_rate": 0.00019984230894300731, "loss": 4.916, "step": 78850 }, { "epoch": 0.14723617269925984, "grad_norm": 1.1399683952331543, "learning_rate": 0.00019984207302549907, "loss": 4.7578, "step": 78900 }, { "epoch": 0.14732947825863835, "grad_norm": 1.0678455829620361, "learning_rate": 0.0001998418369317871, "loss": 4.838, "step": 78950 }, { "epoch": 0.14742278381801685, "grad_norm": 0.9881036877632141, "learning_rate": 0.00019984160066187188, "loss": 4.8102, "step": 79000 }, { "epoch": 0.14751608937739533, "grad_norm": 1.053001880645752, "learning_rate": 0.0001998413642157538, "loss": 4.7473, "step": 79050 }, { "epoch": 0.14760939493677383, "grad_norm": 0.9560543298721313, "learning_rate": 0.00019984112759343327, "loss": 4.7536, "step": 79100 }, { "epoch": 0.1477027004961523, "grad_norm": 0.8099980354309082, "learning_rate": 0.00019984089079491075, "loss": 4.5886, "step": 79150 }, { "epoch": 0.1477960060555308, "grad_norm": 0.8475550413131714, "learning_rate": 0.0001998406538201866, "loss": 4.628, "step": 79200 }, { "epoch": 0.14788931161490929, "grad_norm": 1.233866572380066, "learning_rate": 0.00019984041666926125, "loss": 4.6666, "step": 79250 }, { "epoch": 0.1479826171742878, "grad_norm": 1.088966727256775, "learning_rate": 0.00019984017934213515, "loss": 4.7315, "step": 79300 }, { "epoch": 0.14807592273366627, "grad_norm": 1.3375529050827026, "learning_rate": 0.0001998399418388087, "loss": 4.6797, "step": 79350 }, { "epoch": 0.14816922829304477, "grad_norm": 0.8874202370643616, "learning_rate": 0.00019983970415928234, "loss": 4.8035, "step": 79400 }, { "epoch": 0.14826253385242327, "grad_norm": 0.9835208058357239, "learning_rate": 0.00019983946630355642, "loss": 4.8454, "step": 79450 }, { "epoch": 0.14835583941180175, "grad_norm": 1.30135977268219, "learning_rate": 0.00019983922827163144, "loss": 4.9588, "step": 79500 }, { "epoch": 0.14844914497118025, "grad_norm": 0.8822344541549683, "learning_rate": 0.00019983899006350777, "loss": 4.6737, "step": 79550 }, { "epoch": 0.14854245053055873, "grad_norm": 0.812828779220581, "learning_rate": 0.0001998387516791859, "loss": 4.7829, "step": 79600 }, { "epoch": 0.14863575608993723, "grad_norm": 0.9404298663139343, "learning_rate": 0.00019983851311866616, "loss": 4.4065, "step": 79650 }, { "epoch": 0.1487290616493157, "grad_norm": 1.5870780944824219, "learning_rate": 0.000199838274381949, "loss": 4.8494, "step": 79700 }, { "epoch": 0.1488223672086942, "grad_norm": 0.9650738835334778, "learning_rate": 0.0001998380354690349, "loss": 4.7546, "step": 79750 }, { "epoch": 0.14891567276807272, "grad_norm": 1.1705528497695923, "learning_rate": 0.00019983779637992418, "loss": 4.9383, "step": 79800 }, { "epoch": 0.1490089783274512, "grad_norm": 1.1494523286819458, "learning_rate": 0.00019983755711461734, "loss": 4.7803, "step": 79850 }, { "epoch": 0.1491022838868297, "grad_norm": 0.7461511492729187, "learning_rate": 0.00019983731767311477, "loss": 4.7261, "step": 79900 }, { "epoch": 0.14919558944620817, "grad_norm": 1.043479323387146, "learning_rate": 0.0001998370780554169, "loss": 4.8492, "step": 79950 }, { "epoch": 0.14928889500558667, "grad_norm": 1.0631600618362427, "learning_rate": 0.0001998368382615242, "loss": 4.7619, "step": 80000 }, { "epoch": 0.14938220056496515, "grad_norm": 1.0687527656555176, "learning_rate": 0.00019983659829143698, "loss": 4.9802, "step": 80050 }, { "epoch": 0.14947550612434365, "grad_norm": 0.7543812394142151, "learning_rate": 0.0001998363581451558, "loss": 5.0034, "step": 80100 }, { "epoch": 0.14956881168372216, "grad_norm": 0.6798197627067566, "learning_rate": 0.00019983611782268095, "loss": 4.7076, "step": 80150 }, { "epoch": 0.14966211724310063, "grad_norm": 0.8588387370109558, "learning_rate": 0.00019983587732401293, "loss": 4.8063, "step": 80200 }, { "epoch": 0.14975542280247914, "grad_norm": 1.1029499769210815, "learning_rate": 0.0001998356366491522, "loss": 4.6521, "step": 80250 }, { "epoch": 0.1498487283618576, "grad_norm": 0.8040598034858704, "learning_rate": 0.0001998353957980991, "loss": 4.6823, "step": 80300 }, { "epoch": 0.14994203392123612, "grad_norm": 1.1279397010803223, "learning_rate": 0.00019983515477085408, "loss": 4.7486, "step": 80350 }, { "epoch": 0.1500353394806146, "grad_norm": 1.1090079545974731, "learning_rate": 0.00019983491356741762, "loss": 4.7731, "step": 80400 }, { "epoch": 0.1501286450399931, "grad_norm": 1.3547964096069336, "learning_rate": 0.00019983467218779007, "loss": 4.8618, "step": 80450 }, { "epoch": 0.1502219505993716, "grad_norm": 1.263022780418396, "learning_rate": 0.00019983443063197192, "loss": 4.8745, "step": 80500 }, { "epoch": 0.15031525615875008, "grad_norm": 1.056709885597229, "learning_rate": 0.0001998341888999635, "loss": 4.7292, "step": 80550 }, { "epoch": 0.15040856171812858, "grad_norm": 1.0344963073730469, "learning_rate": 0.0001998339469917654, "loss": 4.5219, "step": 80600 }, { "epoch": 0.15050186727750706, "grad_norm": 0.7098541259765625, "learning_rate": 0.0001998337049073779, "loss": 4.6419, "step": 80650 }, { "epoch": 0.15059517283688556, "grad_norm": 1.0774918794631958, "learning_rate": 0.00019983346264680147, "loss": 4.7146, "step": 80700 }, { "epoch": 0.15068847839626404, "grad_norm": 0.8869603872299194, "learning_rate": 0.00019983322021003655, "loss": 4.8143, "step": 80750 }, { "epoch": 0.15078178395564254, "grad_norm": 1.0924726724624634, "learning_rate": 0.00019983297759708357, "loss": 4.7092, "step": 80800 }, { "epoch": 0.15087508951502102, "grad_norm": 1.0143225193023682, "learning_rate": 0.00019983273480794295, "loss": 4.6599, "step": 80850 }, { "epoch": 0.15096839507439952, "grad_norm": 1.0644515752792358, "learning_rate": 0.0001998324918426151, "loss": 4.6655, "step": 80900 }, { "epoch": 0.15106170063377802, "grad_norm": 1.2146971225738525, "learning_rate": 0.00019983224870110048, "loss": 4.7266, "step": 80950 }, { "epoch": 0.1511550061931565, "grad_norm": 0.9861900210380554, "learning_rate": 0.00019983200538339953, "loss": 4.8545, "step": 81000 }, { "epoch": 0.1511550061931565, "eval_loss": 4.951722145080566, "eval_runtime": 230.9557, "eval_samples_per_second": 11.292, "eval_steps_per_second": 11.292, "eval_tts_loss": 7.390448370325648, "step": 81000 }, { "epoch": 0.151248311752535, "grad_norm": 1.09833824634552, "learning_rate": 0.00019983176188951264, "loss": 4.7289, "step": 81050 }, { "epoch": 0.15134161731191348, "grad_norm": 1.1554728746414185, "learning_rate": 0.00019983151821944027, "loss": 4.7042, "step": 81100 }, { "epoch": 0.15143492287129198, "grad_norm": 0.9898795485496521, "learning_rate": 0.00019983127437318282, "loss": 4.7149, "step": 81150 }, { "epoch": 0.15152822843067046, "grad_norm": 0.9649356603622437, "learning_rate": 0.00019983103035074072, "loss": 4.9477, "step": 81200 }, { "epoch": 0.15162153399004896, "grad_norm": 0.9268784523010254, "learning_rate": 0.00019983078615211442, "loss": 4.6484, "step": 81250 }, { "epoch": 0.15171483954942747, "grad_norm": 1.1820745468139648, "learning_rate": 0.00019983054177730435, "loss": 4.6602, "step": 81300 }, { "epoch": 0.15180814510880594, "grad_norm": 0.8073734045028687, "learning_rate": 0.00019983029722631096, "loss": 4.6794, "step": 81350 }, { "epoch": 0.15190145066818445, "grad_norm": 0.8954252600669861, "learning_rate": 0.00019983005249913467, "loss": 4.6629, "step": 81400 }, { "epoch": 0.15199475622756292, "grad_norm": 1.1588051319122314, "learning_rate": 0.0001998298075957759, "loss": 4.9659, "step": 81450 }, { "epoch": 0.15208806178694143, "grad_norm": 1.1768730878829956, "learning_rate": 0.00019982956251623506, "loss": 4.7299, "step": 81500 }, { "epoch": 0.1521813673463199, "grad_norm": 0.9447085857391357, "learning_rate": 0.00019982931726051262, "loss": 4.7048, "step": 81550 }, { "epoch": 0.1522746729056984, "grad_norm": 0.8608984351158142, "learning_rate": 0.000199829071828609, "loss": 4.7718, "step": 81600 }, { "epoch": 0.1523679784650769, "grad_norm": 0.7421265840530396, "learning_rate": 0.0001998288262205246, "loss": 4.8859, "step": 81650 }, { "epoch": 0.15246128402445538, "grad_norm": 0.9788615107536316, "learning_rate": 0.00019982858043625994, "loss": 4.8217, "step": 81700 }, { "epoch": 0.1525545895838339, "grad_norm": 0.8593958616256714, "learning_rate": 0.00019982833447581538, "loss": 4.918, "step": 81750 }, { "epoch": 0.15264789514321236, "grad_norm": 1.2187819480895996, "learning_rate": 0.0001998280883391914, "loss": 4.6866, "step": 81800 }, { "epoch": 0.15274120070259087, "grad_norm": 1.0192663669586182, "learning_rate": 0.0001998278420263884, "loss": 4.7098, "step": 81850 }, { "epoch": 0.15283450626196934, "grad_norm": 1.1101956367492676, "learning_rate": 0.00019982759553740677, "loss": 4.6928, "step": 81900 }, { "epoch": 0.15292781182134785, "grad_norm": 0.8426311016082764, "learning_rate": 0.00019982734887224703, "loss": 4.7077, "step": 81950 }, { "epoch": 0.15302111738072632, "grad_norm": 1.0216084718704224, "learning_rate": 0.00019982710203090962, "loss": 4.822, "step": 82000 }, { "epoch": 0.15311442294010483, "grad_norm": 1.1248880624771118, "learning_rate": 0.00019982685501339492, "loss": 4.7781, "step": 82050 }, { "epoch": 0.15320772849948333, "grad_norm": 0.8825070261955261, "learning_rate": 0.00019982660781970335, "loss": 4.7426, "step": 82100 }, { "epoch": 0.1533010340588618, "grad_norm": 1.2149336338043213, "learning_rate": 0.00019982636044983543, "loss": 4.7091, "step": 82150 }, { "epoch": 0.1533943396182403, "grad_norm": 1.1825565099716187, "learning_rate": 0.00019982611290379152, "loss": 4.7934, "step": 82200 }, { "epoch": 0.1534876451776188, "grad_norm": 0.9500203728675842, "learning_rate": 0.0001998258651815721, "loss": 4.5296, "step": 82250 }, { "epoch": 0.1535809507369973, "grad_norm": 1.058222770690918, "learning_rate": 0.00019982561728317758, "loss": 4.7336, "step": 82300 }, { "epoch": 0.15367425629637577, "grad_norm": 1.1481443643569946, "learning_rate": 0.00019982536920860842, "loss": 4.6843, "step": 82350 }, { "epoch": 0.15376756185575427, "grad_norm": 1.009384036064148, "learning_rate": 0.00019982512095786503, "loss": 4.6789, "step": 82400 }, { "epoch": 0.15386086741513277, "grad_norm": 1.0169785022735596, "learning_rate": 0.00019982487253094791, "loss": 4.8183, "step": 82450 }, { "epoch": 0.15395417297451125, "grad_norm": 1.4280728101730347, "learning_rate": 0.0001998246239278574, "loss": 4.7322, "step": 82500 }, { "epoch": 0.15404747853388975, "grad_norm": 0.9956715106964111, "learning_rate": 0.00019982437514859402, "loss": 4.467, "step": 82550 }, { "epoch": 0.15414078409326823, "grad_norm": 0.9345201253890991, "learning_rate": 0.0001998241261931582, "loss": 4.8818, "step": 82600 }, { "epoch": 0.15423408965264673, "grad_norm": 0.8383892774581909, "learning_rate": 0.00019982387706155035, "loss": 5.0506, "step": 82650 }, { "epoch": 0.1543273952120252, "grad_norm": 1.0288281440734863, "learning_rate": 0.00019982362775377094, "loss": 4.7276, "step": 82700 }, { "epoch": 0.1544207007714037, "grad_norm": 1.5287654399871826, "learning_rate": 0.00019982337826982035, "loss": 4.7515, "step": 82750 }, { "epoch": 0.15451400633078222, "grad_norm": 1.1659990549087524, "learning_rate": 0.00019982312860969908, "loss": 4.7312, "step": 82800 }, { "epoch": 0.1546073118901607, "grad_norm": 1.0841422080993652, "learning_rate": 0.00019982287877340756, "loss": 4.6208, "step": 82850 }, { "epoch": 0.1547006174495392, "grad_norm": 0.9620239734649658, "learning_rate": 0.0001998226287609462, "loss": 4.7568, "step": 82900 }, { "epoch": 0.15479392300891767, "grad_norm": 1.0376754999160767, "learning_rate": 0.0001998223785723155, "loss": 4.9583, "step": 82950 }, { "epoch": 0.15488722856829618, "grad_norm": 0.8780238032341003, "learning_rate": 0.00019982212820751584, "loss": 4.7517, "step": 83000 }, { "epoch": 0.15498053412767465, "grad_norm": 0.9946292042732239, "learning_rate": 0.00019982187766654768, "loss": 4.7237, "step": 83050 }, { "epoch": 0.15507383968705316, "grad_norm": 0.7709692716598511, "learning_rate": 0.00019982162694941148, "loss": 4.9062, "step": 83100 }, { "epoch": 0.15516714524643166, "grad_norm": 1.010310411453247, "learning_rate": 0.00019982137605610768, "loss": 4.7754, "step": 83150 }, { "epoch": 0.15526045080581014, "grad_norm": 0.752420961856842, "learning_rate": 0.00019982112498663672, "loss": 4.762, "step": 83200 }, { "epoch": 0.15535375636518864, "grad_norm": 0.9181608557701111, "learning_rate": 0.00019982087374099906, "loss": 4.6671, "step": 83250 }, { "epoch": 0.15544706192456711, "grad_norm": 1.0234555006027222, "learning_rate": 0.00019982062231919509, "loss": 4.963, "step": 83300 }, { "epoch": 0.15554036748394562, "grad_norm": 1.1121389865875244, "learning_rate": 0.00019982037072122525, "loss": 4.5522, "step": 83350 }, { "epoch": 0.1556336730433241, "grad_norm": 1.1406534910202026, "learning_rate": 0.00019982011894709007, "loss": 4.8627, "step": 83400 }, { "epoch": 0.1557269786027026, "grad_norm": 0.9664844870567322, "learning_rate": 0.00019981986699678993, "loss": 4.6783, "step": 83450 }, { "epoch": 0.15582028416208107, "grad_norm": 0.931291937828064, "learning_rate": 0.0001998196148703253, "loss": 4.6384, "step": 83500 }, { "epoch": 0.15591358972145958, "grad_norm": 1.102245569229126, "learning_rate": 0.0001998193625676966, "loss": 4.5632, "step": 83550 }, { "epoch": 0.15600689528083808, "grad_norm": 0.9392174482345581, "learning_rate": 0.00019981911008890428, "loss": 4.71, "step": 83600 }, { "epoch": 0.15610020084021656, "grad_norm": 0.7961680293083191, "learning_rate": 0.00019981885743394884, "loss": 4.7738, "step": 83650 }, { "epoch": 0.15619350639959506, "grad_norm": 0.9825389385223389, "learning_rate": 0.00019981860460283063, "loss": 4.4743, "step": 83700 }, { "epoch": 0.15628681195897354, "grad_norm": 0.8377106189727783, "learning_rate": 0.00019981835159555016, "loss": 4.7894, "step": 83750 }, { "epoch": 0.15638011751835204, "grad_norm": 0.9133685231208801, "learning_rate": 0.00019981809841210788, "loss": 4.7747, "step": 83800 }, { "epoch": 0.15647342307773052, "grad_norm": 0.8195970058441162, "learning_rate": 0.0001998178450525042, "loss": 4.592, "step": 83850 }, { "epoch": 0.15656672863710902, "grad_norm": 1.0992616415023804, "learning_rate": 0.0001998175915167396, "loss": 4.8078, "step": 83900 }, { "epoch": 0.15666003419648752, "grad_norm": 1.0138832330703735, "learning_rate": 0.0001998173378048145, "loss": 4.6467, "step": 83950 }, { "epoch": 0.156753339755866, "grad_norm": 0.8786870837211609, "learning_rate": 0.0001998170839167294, "loss": 4.7446, "step": 84000 }, { "epoch": 0.156753339755866, "eval_loss": 4.946982383728027, "eval_runtime": 229.0908, "eval_samples_per_second": 11.384, "eval_steps_per_second": 11.384, "eval_tts_loss": 7.394625851078862, "step": 84000 }, { "epoch": 0.1568466453152445, "grad_norm": 1.0230584144592285, "learning_rate": 0.00019981682985248465, "loss": 4.6526, "step": 84050 }, { "epoch": 0.15693995087462298, "grad_norm": 0.7539162039756775, "learning_rate": 0.0001998165756120808, "loss": 4.6949, "step": 84100 }, { "epoch": 0.15703325643400148, "grad_norm": 1.1452572345733643, "learning_rate": 0.00019981632119551822, "loss": 4.6745, "step": 84150 }, { "epoch": 0.15712656199337996, "grad_norm": 0.9784168601036072, "learning_rate": 0.00019981606660279745, "loss": 4.9249, "step": 84200 }, { "epoch": 0.15721986755275846, "grad_norm": 1.1359683275222778, "learning_rate": 0.00019981581183391885, "loss": 4.761, "step": 84250 }, { "epoch": 0.15731317311213697, "grad_norm": 1.1261489391326904, "learning_rate": 0.0001998155568888829, "loss": 4.6123, "step": 84300 }, { "epoch": 0.15740647867151544, "grad_norm": 1.306156039237976, "learning_rate": 0.00019981530176769007, "loss": 4.6882, "step": 84350 }, { "epoch": 0.15749978423089395, "grad_norm": 0.861440896987915, "learning_rate": 0.0001998150464703408, "loss": 4.6409, "step": 84400 }, { "epoch": 0.15759308979027242, "grad_norm": 0.6180737018585205, "learning_rate": 0.00019981479099683553, "loss": 4.6318, "step": 84450 }, { "epoch": 0.15768639534965093, "grad_norm": 1.0346062183380127, "learning_rate": 0.00019981453534717475, "loss": 4.8614, "step": 84500 }, { "epoch": 0.1577797009090294, "grad_norm": 1.3572213649749756, "learning_rate": 0.00019981427952135884, "loss": 4.7108, "step": 84550 }, { "epoch": 0.1578730064684079, "grad_norm": 0.7147801518440247, "learning_rate": 0.0001998140235193883, "loss": 4.6901, "step": 84600 }, { "epoch": 0.15796631202778638, "grad_norm": 1.1168407201766968, "learning_rate": 0.00019981376734126355, "loss": 4.7444, "step": 84650 }, { "epoch": 0.15805961758716489, "grad_norm": 1.0936293601989746, "learning_rate": 0.00019981351098698509, "loss": 4.8869, "step": 84700 }, { "epoch": 0.1581529231465434, "grad_norm": 0.9468141198158264, "learning_rate": 0.00019981325445655331, "loss": 4.7834, "step": 84750 }, { "epoch": 0.15824622870592187, "grad_norm": 1.188549518585205, "learning_rate": 0.00019981299774996875, "loss": 4.7705, "step": 84800 }, { "epoch": 0.15833953426530037, "grad_norm": 1.0123662948608398, "learning_rate": 0.0001998127408672318, "loss": 4.7251, "step": 84850 }, { "epoch": 0.15843283982467884, "grad_norm": 1.1224631071090698, "learning_rate": 0.0001998124838083429, "loss": 4.7438, "step": 84900 }, { "epoch": 0.15852614538405735, "grad_norm": 0.9828144311904907, "learning_rate": 0.00019981222657330255, "loss": 4.7185, "step": 84950 }, { "epoch": 0.15861945094343582, "grad_norm": 0.932032585144043, "learning_rate": 0.00019981196916211115, "loss": 4.7946, "step": 85000 }, { "epoch": 0.15871275650281433, "grad_norm": 1.082952618598938, "learning_rate": 0.0001998117115747692, "loss": 4.7448, "step": 85050 }, { "epoch": 0.15880606206219283, "grad_norm": 1.2148911952972412, "learning_rate": 0.00019981145381127715, "loss": 4.6534, "step": 85100 }, { "epoch": 0.1588993676215713, "grad_norm": 0.8082234859466553, "learning_rate": 0.00019981119587163547, "loss": 4.6324, "step": 85150 }, { "epoch": 0.1589926731809498, "grad_norm": 1.083858847618103, "learning_rate": 0.00019981093775584457, "loss": 4.6805, "step": 85200 }, { "epoch": 0.1590859787403283, "grad_norm": 1.0355510711669922, "learning_rate": 0.00019981067946390493, "loss": 5.026, "step": 85250 }, { "epoch": 0.1591792842997068, "grad_norm": 1.0597902536392212, "learning_rate": 0.00019981042099581698, "loss": 4.8371, "step": 85300 }, { "epoch": 0.15927258985908527, "grad_norm": 0.8655616641044617, "learning_rate": 0.0001998101623515812, "loss": 4.8269, "step": 85350 }, { "epoch": 0.15936589541846377, "grad_norm": 1.1475633382797241, "learning_rate": 0.00019980990353119809, "loss": 4.6525, "step": 85400 }, { "epoch": 0.15945920097784227, "grad_norm": 0.9865615963935852, "learning_rate": 0.00019980964453466803, "loss": 4.6893, "step": 85450 }, { "epoch": 0.15955250653722075, "grad_norm": 0.8663655519485474, "learning_rate": 0.00019980938536199151, "loss": 4.736, "step": 85500 }, { "epoch": 0.15964581209659925, "grad_norm": 0.9567622542381287, "learning_rate": 0.000199809126013169, "loss": 4.5879, "step": 85550 }, { "epoch": 0.15973911765597773, "grad_norm": 0.9500153660774231, "learning_rate": 0.00019980886648820093, "loss": 4.7534, "step": 85600 }, { "epoch": 0.15983242321535623, "grad_norm": 0.9333567023277283, "learning_rate": 0.0001998086067870878, "loss": 4.671, "step": 85650 }, { "epoch": 0.1599257287747347, "grad_norm": 0.9707118272781372, "learning_rate": 0.00019980834690983, "loss": 4.7252, "step": 85700 }, { "epoch": 0.1600190343341132, "grad_norm": 0.801749587059021, "learning_rate": 0.00019980808685642805, "loss": 4.7634, "step": 85750 }, { "epoch": 0.16011233989349172, "grad_norm": 1.0327675342559814, "learning_rate": 0.0001998078266268824, "loss": 4.7954, "step": 85800 }, { "epoch": 0.1602056454528702, "grad_norm": 0.9381875991821289, "learning_rate": 0.00019980756622119347, "loss": 4.6373, "step": 85850 }, { "epoch": 0.1602989510122487, "grad_norm": 0.9350003600120544, "learning_rate": 0.00019980730563936175, "loss": 4.7086, "step": 85900 }, { "epoch": 0.16039225657162717, "grad_norm": 1.1545840501785278, "learning_rate": 0.00019980704488138776, "loss": 4.7647, "step": 85950 }, { "epoch": 0.16048556213100568, "grad_norm": 1.1162843704223633, "learning_rate": 0.00019980678394727184, "loss": 4.701, "step": 86000 }, { "epoch": 0.16057886769038415, "grad_norm": 0.9487619400024414, "learning_rate": 0.0001998065228370145, "loss": 4.847, "step": 86050 }, { "epoch": 0.16067217324976266, "grad_norm": 1.1424440145492554, "learning_rate": 0.00019980626155061623, "loss": 4.7775, "step": 86100 }, { "epoch": 0.16076547880914113, "grad_norm": 1.0836292505264282, "learning_rate": 0.00019980600008807746, "loss": 4.7517, "step": 86150 }, { "epoch": 0.16085878436851964, "grad_norm": 1.026230812072754, "learning_rate": 0.00019980573844939866, "loss": 4.6033, "step": 86200 }, { "epoch": 0.16095208992789814, "grad_norm": 0.7633348107337952, "learning_rate": 0.0001998054766345803, "loss": 4.7064, "step": 86250 }, { "epoch": 0.16104539548727662, "grad_norm": 0.9830362796783447, "learning_rate": 0.00019980521464362284, "loss": 4.6753, "step": 86300 }, { "epoch": 0.16113870104665512, "grad_norm": 0.7929043173789978, "learning_rate": 0.00019980495247652673, "loss": 4.4158, "step": 86350 }, { "epoch": 0.1612320066060336, "grad_norm": 0.906753420829773, "learning_rate": 0.00019980469013329243, "loss": 4.5392, "step": 86400 }, { "epoch": 0.1613253121654121, "grad_norm": 0.841719388961792, "learning_rate": 0.00019980442761392041, "loss": 4.7356, "step": 86450 }, { "epoch": 0.16141861772479058, "grad_norm": 0.9761390686035156, "learning_rate": 0.00019980416491841116, "loss": 4.8662, "step": 86500 }, { "epoch": 0.16151192328416908, "grad_norm": 1.2850239276885986, "learning_rate": 0.0001998039020467651, "loss": 4.6927, "step": 86550 }, { "epoch": 0.16160522884354758, "grad_norm": 0.9129481911659241, "learning_rate": 0.00019980363899898277, "loss": 4.855, "step": 86600 }, { "epoch": 0.16169853440292606, "grad_norm": 1.1330456733703613, "learning_rate": 0.0001998033757750645, "loss": 5.0053, "step": 86650 }, { "epoch": 0.16179183996230456, "grad_norm": 1.0761449337005615, "learning_rate": 0.00019980311237501085, "loss": 4.6144, "step": 86700 }, { "epoch": 0.16188514552168304, "grad_norm": 0.9955719709396362, "learning_rate": 0.0001998028487988223, "loss": 4.7949, "step": 86750 }, { "epoch": 0.16197845108106154, "grad_norm": 1.190447211265564, "learning_rate": 0.00019980258504649927, "loss": 4.7975, "step": 86800 }, { "epoch": 0.16207175664044002, "grad_norm": 1.0047932863235474, "learning_rate": 0.00019980232111804224, "loss": 4.7366, "step": 86850 }, { "epoch": 0.16216506219981852, "grad_norm": 0.9887164831161499, "learning_rate": 0.00019980205701345164, "loss": 4.7679, "step": 86900 }, { "epoch": 0.16225836775919703, "grad_norm": 1.305762529373169, "learning_rate": 0.00019980179273272803, "loss": 4.8636, "step": 86950 }, { "epoch": 0.1623516733185755, "grad_norm": 1.0135672092437744, "learning_rate": 0.00019980152827587177, "loss": 4.7288, "step": 87000 }, { "epoch": 0.1623516733185755, "eval_loss": 4.931732654571533, "eval_runtime": 230.4634, "eval_samples_per_second": 11.316, "eval_steps_per_second": 11.316, "eval_tts_loss": 7.40806279529566, "step": 87000 }, { "epoch": 0.162444978877954, "grad_norm": 1.1869860887527466, "learning_rate": 0.0001998012636428834, "loss": 4.6017, "step": 87050 }, { "epoch": 0.16253828443733248, "grad_norm": 0.9774503707885742, "learning_rate": 0.00019980099883376334, "loss": 4.7084, "step": 87100 }, { "epoch": 0.16263158999671098, "grad_norm": 0.8578800559043884, "learning_rate": 0.00019980073384851207, "loss": 4.9125, "step": 87150 }, { "epoch": 0.16272489555608946, "grad_norm": 1.0579946041107178, "learning_rate": 0.00019980046868713013, "loss": 4.7317, "step": 87200 }, { "epoch": 0.16281820111546796, "grad_norm": 0.6333214044570923, "learning_rate": 0.00019980020334961783, "loss": 4.5953, "step": 87250 }, { "epoch": 0.16291150667484644, "grad_norm": 1.0423671007156372, "learning_rate": 0.0001997999378359758, "loss": 4.7238, "step": 87300 }, { "epoch": 0.16300481223422494, "grad_norm": 0.7020459175109863, "learning_rate": 0.0001997996721462044, "loss": 4.6665, "step": 87350 }, { "epoch": 0.16309811779360345, "grad_norm": 0.9732704162597656, "learning_rate": 0.00019979940628030415, "loss": 4.6746, "step": 87400 }, { "epoch": 0.16319142335298192, "grad_norm": 0.8247692584991455, "learning_rate": 0.00019979914023827552, "loss": 4.8856, "step": 87450 }, { "epoch": 0.16328472891236043, "grad_norm": 1.1054822206497192, "learning_rate": 0.00019979887402011898, "loss": 4.639, "step": 87500 }, { "epoch": 0.1633780344717389, "grad_norm": 1.0070406198501587, "learning_rate": 0.00019979860762583496, "loss": 4.7724, "step": 87550 }, { "epoch": 0.1634713400311174, "grad_norm": 0.9917934536933899, "learning_rate": 0.00019979834105542397, "loss": 4.8537, "step": 87600 }, { "epoch": 0.16356464559049588, "grad_norm": 1.4356805086135864, "learning_rate": 0.00019979807430888644, "loss": 4.816, "step": 87650 }, { "epoch": 0.1636579511498744, "grad_norm": 1.1016113758087158, "learning_rate": 0.0001997978073862229, "loss": 4.8199, "step": 87700 }, { "epoch": 0.1637512567092529, "grad_norm": 1.1834055185317993, "learning_rate": 0.00019979754028743378, "loss": 4.5439, "step": 87750 }, { "epoch": 0.16384456226863137, "grad_norm": 1.2742072343826294, "learning_rate": 0.00019979727301251959, "loss": 4.7865, "step": 87800 }, { "epoch": 0.16393786782800987, "grad_norm": 1.1467939615249634, "learning_rate": 0.0001997970055614807, "loss": 4.848, "step": 87850 }, { "epoch": 0.16403117338738835, "grad_norm": 0.9691332578659058, "learning_rate": 0.00019979673793431773, "loss": 4.866, "step": 87900 }, { "epoch": 0.16412447894676685, "grad_norm": 1.061712622642517, "learning_rate": 0.00019979647013103105, "loss": 4.5297, "step": 87950 }, { "epoch": 0.16421778450614533, "grad_norm": 0.9618381857872009, "learning_rate": 0.00019979620215162115, "loss": 4.6956, "step": 88000 }, { "epoch": 0.16431109006552383, "grad_norm": 1.0956761837005615, "learning_rate": 0.0001997959339960885, "loss": 4.7234, "step": 88050 }, { "epoch": 0.16440439562490233, "grad_norm": 1.304779291152954, "learning_rate": 0.00019979566566443363, "loss": 4.7132, "step": 88100 }, { "epoch": 0.1644977011842808, "grad_norm": 0.8119908571243286, "learning_rate": 0.00019979539715665693, "loss": 4.7707, "step": 88150 }, { "epoch": 0.1645910067436593, "grad_norm": 0.9809343218803406, "learning_rate": 0.00019979512847275897, "loss": 4.6912, "step": 88200 }, { "epoch": 0.1646843123030378, "grad_norm": 1.1159480810165405, "learning_rate": 0.00019979485961274008, "loss": 4.7362, "step": 88250 }, { "epoch": 0.1647776178624163, "grad_norm": 1.0415050983428955, "learning_rate": 0.0001997945905766009, "loss": 4.8182, "step": 88300 }, { "epoch": 0.16487092342179477, "grad_norm": 0.856876015663147, "learning_rate": 0.00019979432136434179, "loss": 4.475, "step": 88350 }, { "epoch": 0.16496422898117327, "grad_norm": 0.9265991449356079, "learning_rate": 0.0001997940519759633, "loss": 4.5788, "step": 88400 }, { "epoch": 0.16505753454055178, "grad_norm": 1.082276701927185, "learning_rate": 0.00019979378241146583, "loss": 4.658, "step": 88450 }, { "epoch": 0.16515084009993025, "grad_norm": 0.916746199131012, "learning_rate": 0.0001997935126708499, "loss": 4.7118, "step": 88500 }, { "epoch": 0.16524414565930876, "grad_norm": 0.7613538503646851, "learning_rate": 0.00019979324275411597, "loss": 4.6433, "step": 88550 }, { "epoch": 0.16533745121868723, "grad_norm": 0.9194437861442566, "learning_rate": 0.00019979297266126453, "loss": 4.8783, "step": 88600 }, { "epoch": 0.16543075677806574, "grad_norm": 0.907049298286438, "learning_rate": 0.00019979270239229606, "loss": 5.006, "step": 88650 }, { "epoch": 0.1655240623374442, "grad_norm": 0.8706242442131042, "learning_rate": 0.00019979243194721103, "loss": 4.6975, "step": 88700 }, { "epoch": 0.16561736789682271, "grad_norm": 1.136927843093872, "learning_rate": 0.0001997921613260099, "loss": 4.7098, "step": 88750 }, { "epoch": 0.1657106734562012, "grad_norm": 0.9624984264373779, "learning_rate": 0.0001997918905286932, "loss": 4.7689, "step": 88800 }, { "epoch": 0.1658039790155797, "grad_norm": 1.003145694732666, "learning_rate": 0.00019979161955526134, "loss": 4.6874, "step": 88850 }, { "epoch": 0.1658972845749582, "grad_norm": 0.9634168148040771, "learning_rate": 0.00019979134840571484, "loss": 4.7637, "step": 88900 }, { "epoch": 0.16599059013433667, "grad_norm": 1.0608584880828857, "learning_rate": 0.00019979107708005417, "loss": 4.6663, "step": 88950 }, { "epoch": 0.16608389569371518, "grad_norm": 0.7554059028625488, "learning_rate": 0.00019979080557827982, "loss": 4.7785, "step": 89000 }, { "epoch": 0.16617720125309365, "grad_norm": 1.0240588188171387, "learning_rate": 0.00019979053390039225, "loss": 4.7544, "step": 89050 }, { "epoch": 0.16627050681247216, "grad_norm": 1.253793478012085, "learning_rate": 0.00019979026204639192, "loss": 4.5589, "step": 89100 }, { "epoch": 0.16636381237185063, "grad_norm": 0.729597270488739, "learning_rate": 0.00019978999001627937, "loss": 4.5692, "step": 89150 }, { "epoch": 0.16645711793122914, "grad_norm": 0.9666997194290161, "learning_rate": 0.00019978971781005504, "loss": 4.7101, "step": 89200 }, { "epoch": 0.16655042349060764, "grad_norm": 1.3623840808868408, "learning_rate": 0.0001997894454277194, "loss": 4.6743, "step": 89250 }, { "epoch": 0.16664372904998612, "grad_norm": 0.9426568746566772, "learning_rate": 0.00019978917286927295, "loss": 4.6291, "step": 89300 }, { "epoch": 0.16673703460936462, "grad_norm": 0.9860442876815796, "learning_rate": 0.00019978890013471617, "loss": 4.7023, "step": 89350 }, { "epoch": 0.1668303401687431, "grad_norm": 0.9578588008880615, "learning_rate": 0.00019978862722404956, "loss": 4.8009, "step": 89400 }, { "epoch": 0.1669236457281216, "grad_norm": 1.0634305477142334, "learning_rate": 0.00019978835413727358, "loss": 4.8068, "step": 89450 }, { "epoch": 0.16701695128750008, "grad_norm": 0.5965625643730164, "learning_rate": 0.0001997880808743887, "loss": 4.5365, "step": 89500 }, { "epoch": 0.16711025684687858, "grad_norm": 0.9046986699104309, "learning_rate": 0.0001997878074353954, "loss": 4.7543, "step": 89550 }, { "epoch": 0.16720356240625708, "grad_norm": 1.0531772375106812, "learning_rate": 0.0001997875338202942, "loss": 4.7865, "step": 89600 }, { "epoch": 0.16729686796563556, "grad_norm": 1.0829286575317383, "learning_rate": 0.00019978726002908555, "loss": 4.5878, "step": 89650 }, { "epoch": 0.16739017352501406, "grad_norm": 0.9807512164115906, "learning_rate": 0.00019978698606176994, "loss": 4.7111, "step": 89700 }, { "epoch": 0.16748347908439254, "grad_norm": 0.7498921155929565, "learning_rate": 0.00019978671191834787, "loss": 4.5111, "step": 89750 }, { "epoch": 0.16757678464377104, "grad_norm": 1.3092355728149414, "learning_rate": 0.0001997864375988198, "loss": 4.6699, "step": 89800 }, { "epoch": 0.16767009020314952, "grad_norm": 1.1617666482925415, "learning_rate": 0.00019978616310318623, "loss": 4.7742, "step": 89850 }, { "epoch": 0.16776339576252802, "grad_norm": 1.0421655178070068, "learning_rate": 0.00019978588843144767, "loss": 4.8136, "step": 89900 }, { "epoch": 0.16785670132190653, "grad_norm": 1.0589338541030884, "learning_rate": 0.00019978561358360454, "loss": 4.961, "step": 89950 }, { "epoch": 0.167950006881285, "grad_norm": 0.8682652115821838, "learning_rate": 0.00019978533855965735, "loss": 4.6367, "step": 90000 }, { "epoch": 0.167950006881285, "eval_loss": 4.925217628479004, "eval_runtime": 229.1885, "eval_samples_per_second": 11.379, "eval_steps_per_second": 11.379, "eval_tts_loss": 7.469426870473198, "step": 90000 }, { "epoch": 0.1680433124406635, "grad_norm": 1.0179359912872314, "learning_rate": 0.00019978506335960664, "loss": 4.7742, "step": 90050 }, { "epoch": 0.16813661800004198, "grad_norm": 1.0394161939620972, "learning_rate": 0.00019978478798345282, "loss": 4.7354, "step": 90100 }, { "epoch": 0.16822992355942049, "grad_norm": 1.118033766746521, "learning_rate": 0.0001997845124311964, "loss": 4.6653, "step": 90150 }, { "epoch": 0.16832322911879896, "grad_norm": 0.7835417985916138, "learning_rate": 0.00019978423670283788, "loss": 4.7592, "step": 90200 }, { "epoch": 0.16841653467817747, "grad_norm": 1.0521881580352783, "learning_rate": 0.00019978396079837777, "loss": 4.7853, "step": 90250 }, { "epoch": 0.16850984023755594, "grad_norm": 0.9498222470283508, "learning_rate": 0.0001997836847178165, "loss": 4.7828, "step": 90300 }, { "epoch": 0.16860314579693444, "grad_norm": 0.9034340381622314, "learning_rate": 0.00019978340846115456, "loss": 4.6206, "step": 90350 }, { "epoch": 0.16869645135631295, "grad_norm": 1.0530623197555542, "learning_rate": 0.00019978313202839249, "loss": 4.6888, "step": 90400 }, { "epoch": 0.16878975691569142, "grad_norm": 1.2802088260650635, "learning_rate": 0.00019978285541953077, "loss": 4.665, "step": 90450 }, { "epoch": 0.16888306247506993, "grad_norm": 0.7782902717590332, "learning_rate": 0.00019978257863456983, "loss": 4.7043, "step": 90500 }, { "epoch": 0.1689763680344484, "grad_norm": 0.930324912071228, "learning_rate": 0.00019978230167351022, "loss": 4.7948, "step": 90550 }, { "epoch": 0.1690696735938269, "grad_norm": 1.2882273197174072, "learning_rate": 0.00019978202453635238, "loss": 4.8561, "step": 90600 }, { "epoch": 0.16916297915320538, "grad_norm": 0.7437873482704163, "learning_rate": 0.00019978174722309685, "loss": 4.8459, "step": 90650 }, { "epoch": 0.1692562847125839, "grad_norm": 0.9222044944763184, "learning_rate": 0.00019978146973374407, "loss": 4.581, "step": 90700 }, { "epoch": 0.1693495902719624, "grad_norm": 1.0944164991378784, "learning_rate": 0.00019978119206829454, "loss": 4.86, "step": 90750 }, { "epoch": 0.16944289583134087, "grad_norm": 0.8687542080879211, "learning_rate": 0.00019978091422674882, "loss": 4.7375, "step": 90800 }, { "epoch": 0.16953620139071937, "grad_norm": 1.013555884361267, "learning_rate": 0.0001997806362091073, "loss": 4.8501, "step": 90850 }, { "epoch": 0.16962950695009785, "grad_norm": 0.9529797434806824, "learning_rate": 0.00019978035801537054, "loss": 4.9599, "step": 90900 }, { "epoch": 0.16972281250947635, "grad_norm": 1.0254067182540894, "learning_rate": 0.00019978007964553899, "loss": 4.6487, "step": 90950 }, { "epoch": 0.16981611806885483, "grad_norm": 1.2346069812774658, "learning_rate": 0.00019977980109961313, "loss": 4.9191, "step": 91000 }, { "epoch": 0.16990942362823333, "grad_norm": 0.7727384567260742, "learning_rate": 0.00019977952237759352, "loss": 4.676, "step": 91050 }, { "epoch": 0.17000272918761183, "grad_norm": 0.8196634650230408, "learning_rate": 0.00019977924347948058, "loss": 4.7945, "step": 91100 }, { "epoch": 0.1700960347469903, "grad_norm": 0.937310516834259, "learning_rate": 0.00019977896440527486, "loss": 4.8049, "step": 91150 }, { "epoch": 0.1701893403063688, "grad_norm": 0.6780749559402466, "learning_rate": 0.0001997786851549768, "loss": 4.9489, "step": 91200 }, { "epoch": 0.1702826458657473, "grad_norm": 1.1923000812530518, "learning_rate": 0.00019977840572858691, "loss": 4.6109, "step": 91250 }, { "epoch": 0.1703759514251258, "grad_norm": 1.0224426984786987, "learning_rate": 0.00019977812612610572, "loss": 4.7127, "step": 91300 }, { "epoch": 0.17046925698450427, "grad_norm": 0.8383764624595642, "learning_rate": 0.0001997778463475337, "loss": 4.6056, "step": 91350 }, { "epoch": 0.17056256254388277, "grad_norm": 1.2207326889038086, "learning_rate": 0.00019977756639287132, "loss": 4.7927, "step": 91400 }, { "epoch": 0.17065586810326125, "grad_norm": 1.1539119482040405, "learning_rate": 0.0001997772862621191, "loss": 4.7997, "step": 91450 }, { "epoch": 0.17074917366263975, "grad_norm": 0.9894577860832214, "learning_rate": 0.0001997770059552775, "loss": 4.7092, "step": 91500 }, { "epoch": 0.17084247922201826, "grad_norm": 1.1220641136169434, "learning_rate": 0.00019977672547234705, "loss": 4.7147, "step": 91550 }, { "epoch": 0.17093578478139673, "grad_norm": 0.8501759171485901, "learning_rate": 0.00019977644481332824, "loss": 4.7409, "step": 91600 }, { "epoch": 0.17102909034077524, "grad_norm": 0.859946608543396, "learning_rate": 0.00019977616397822158, "loss": 4.5836, "step": 91650 }, { "epoch": 0.1711223959001537, "grad_norm": 0.7580205798149109, "learning_rate": 0.00019977588296702753, "loss": 4.6373, "step": 91700 }, { "epoch": 0.17121570145953222, "grad_norm": 1.3309026956558228, "learning_rate": 0.0001997756017797466, "loss": 4.6238, "step": 91750 }, { "epoch": 0.1713090070189107, "grad_norm": 0.965422511100769, "learning_rate": 0.0001997753204163793, "loss": 4.8401, "step": 91800 }, { "epoch": 0.1714023125782892, "grad_norm": 0.9164948463439941, "learning_rate": 0.0001997750388769261, "loss": 4.8671, "step": 91850 }, { "epoch": 0.1714956181376677, "grad_norm": 0.8471038937568665, "learning_rate": 0.00019977475716138752, "loss": 4.7976, "step": 91900 }, { "epoch": 0.17158892369704618, "grad_norm": 1.1867235898971558, "learning_rate": 0.00019977447526976406, "loss": 4.8716, "step": 91950 }, { "epoch": 0.17168222925642468, "grad_norm": 1.0447583198547363, "learning_rate": 0.00019977419320205616, "loss": 4.5796, "step": 92000 }, { "epoch": 0.17177553481580315, "grad_norm": 0.894728422164917, "learning_rate": 0.00019977391095826443, "loss": 4.7049, "step": 92050 }, { "epoch": 0.17186884037518166, "grad_norm": 1.03791344165802, "learning_rate": 0.0001997736285383893, "loss": 4.6183, "step": 92100 }, { "epoch": 0.17196214593456013, "grad_norm": 0.7112425565719604, "learning_rate": 0.00019977334594243124, "loss": 4.5619, "step": 92150 }, { "epoch": 0.17205545149393864, "grad_norm": 0.7274053692817688, "learning_rate": 0.00019977306317039078, "loss": 4.7356, "step": 92200 }, { "epoch": 0.17214875705331714, "grad_norm": 0.9669164419174194, "learning_rate": 0.0001997727802222684, "loss": 4.5581, "step": 92250 }, { "epoch": 0.17224206261269562, "grad_norm": 0.9389071464538574, "learning_rate": 0.00019977249709806464, "loss": 4.7506, "step": 92300 }, { "epoch": 0.17233536817207412, "grad_norm": 0.9951257109642029, "learning_rate": 0.00019977221379778, "loss": 4.9339, "step": 92350 }, { "epoch": 0.1724286737314526, "grad_norm": 1.1136345863342285, "learning_rate": 0.00019977193032141494, "loss": 4.8545, "step": 92400 }, { "epoch": 0.1725219792908311, "grad_norm": 0.7983369827270508, "learning_rate": 0.00019977164666896998, "loss": 4.6367, "step": 92450 }, { "epoch": 0.17261528485020958, "grad_norm": 0.905382513999939, "learning_rate": 0.00019977136284044563, "loss": 4.4375, "step": 92500 }, { "epoch": 0.17270859040958808, "grad_norm": 1.1273249387741089, "learning_rate": 0.00019977107883584235, "loss": 4.777, "step": 92550 }, { "epoch": 0.17280189596896658, "grad_norm": 0.7400512099266052, "learning_rate": 0.00019977079465516068, "loss": 4.6677, "step": 92600 }, { "epoch": 0.17289520152834506, "grad_norm": 1.0424542427062988, "learning_rate": 0.00019977051029840116, "loss": 4.6719, "step": 92650 }, { "epoch": 0.17298850708772356, "grad_norm": 1.0340840816497803, "learning_rate": 0.00019977022576556417, "loss": 4.7587, "step": 92700 }, { "epoch": 0.17308181264710204, "grad_norm": 1.0534294843673706, "learning_rate": 0.00019976994105665033, "loss": 4.7536, "step": 92750 }, { "epoch": 0.17317511820648054, "grad_norm": 0.754612922668457, "learning_rate": 0.0001997696561716601, "loss": 4.5673, "step": 92800 }, { "epoch": 0.17326842376585902, "grad_norm": 1.1167819499969482, "learning_rate": 0.00019976937111059395, "loss": 4.735, "step": 92850 }, { "epoch": 0.17336172932523752, "grad_norm": 1.0311816930770874, "learning_rate": 0.00019976908587345245, "loss": 4.8155, "step": 92900 }, { "epoch": 0.173455034884616, "grad_norm": 1.6865220069885254, "learning_rate": 0.00019976880046023606, "loss": 4.5625, "step": 92950 }, { "epoch": 0.1735483404439945, "grad_norm": 1.0632766485214233, "learning_rate": 0.0001997685148709453, "loss": 4.6878, "step": 93000 }, { "epoch": 0.1735483404439945, "eval_loss": 4.916426658630371, "eval_runtime": 229.9701, "eval_samples_per_second": 11.341, "eval_steps_per_second": 11.341, "eval_tts_loss": 7.440112192040892, "step": 93000 }, { "epoch": 0.173641646003373, "grad_norm": 0.9786635637283325, "learning_rate": 0.00019976822910558063, "loss": 4.7599, "step": 93050 }, { "epoch": 0.17373495156275148, "grad_norm": 0.9855509400367737, "learning_rate": 0.00019976794316414265, "loss": 4.4347, "step": 93100 }, { "epoch": 0.17382825712213, "grad_norm": 0.8809786438941956, "learning_rate": 0.00019976765704663176, "loss": 4.6688, "step": 93150 }, { "epoch": 0.17392156268150846, "grad_norm": 1.0891468524932861, "learning_rate": 0.00019976737075304851, "loss": 4.7232, "step": 93200 }, { "epoch": 0.17401486824088697, "grad_norm": 1.038535237312317, "learning_rate": 0.00019976708428339342, "loss": 4.7486, "step": 93250 }, { "epoch": 0.17410817380026544, "grad_norm": 1.0178803205490112, "learning_rate": 0.00019976679763766697, "loss": 4.6426, "step": 93300 }, { "epoch": 0.17420147935964395, "grad_norm": 0.7876224517822266, "learning_rate": 0.00019976651081586968, "loss": 4.7747, "step": 93350 }, { "epoch": 0.17429478491902245, "grad_norm": 0.6796728372573853, "learning_rate": 0.00019976622381800207, "loss": 4.8263, "step": 93400 }, { "epoch": 0.17438809047840093, "grad_norm": 0.796763002872467, "learning_rate": 0.0001997659366440646, "loss": 4.5471, "step": 93450 }, { "epoch": 0.17448139603777943, "grad_norm": 0.7914952039718628, "learning_rate": 0.00019976564929405783, "loss": 4.8249, "step": 93500 }, { "epoch": 0.1745747015971579, "grad_norm": 0.7674325704574585, "learning_rate": 0.00019976536176798224, "loss": 4.6139, "step": 93550 }, { "epoch": 0.1746680071565364, "grad_norm": 0.8790803551673889, "learning_rate": 0.00019976507406583834, "loss": 4.6569, "step": 93600 }, { "epoch": 0.17476131271591488, "grad_norm": 1.362616777420044, "learning_rate": 0.00019976478618762662, "loss": 4.7218, "step": 93650 }, { "epoch": 0.1748546182752934, "grad_norm": 1.1918220520019531, "learning_rate": 0.00019976449813334762, "loss": 4.9003, "step": 93700 }, { "epoch": 0.1749479238346719, "grad_norm": 0.9692675471305847, "learning_rate": 0.00019976420990300183, "loss": 4.6867, "step": 93750 }, { "epoch": 0.17504122939405037, "grad_norm": 1.2216622829437256, "learning_rate": 0.00019976392149658977, "loss": 4.6528, "step": 93800 }, { "epoch": 0.17513453495342887, "grad_norm": 0.9711513519287109, "learning_rate": 0.00019976363291411195, "loss": 4.7519, "step": 93850 }, { "epoch": 0.17522784051280735, "grad_norm": 1.0663491487503052, "learning_rate": 0.00019976334415556886, "loss": 4.5517, "step": 93900 }, { "epoch": 0.17532114607218585, "grad_norm": 0.9361185431480408, "learning_rate": 0.00019976305522096101, "loss": 4.7161, "step": 93950 }, { "epoch": 0.17541445163156433, "grad_norm": 0.7986595630645752, "learning_rate": 0.00019976276611028895, "loss": 4.7228, "step": 94000 }, { "epoch": 0.17550775719094283, "grad_norm": 0.9233583211898804, "learning_rate": 0.00019976247682355316, "loss": 4.8464, "step": 94050 }, { "epoch": 0.1756010627503213, "grad_norm": 1.2323817014694214, "learning_rate": 0.00019976218736075414, "loss": 4.5892, "step": 94100 }, { "epoch": 0.1756943683096998, "grad_norm": 1.0528844594955444, "learning_rate": 0.00019976189772189242, "loss": 4.8237, "step": 94150 }, { "epoch": 0.17578767386907831, "grad_norm": 1.1424999237060547, "learning_rate": 0.00019976160790696848, "loss": 4.6996, "step": 94200 }, { "epoch": 0.1758809794284568, "grad_norm": 0.8401368856430054, "learning_rate": 0.00019976131791598286, "loss": 4.8017, "step": 94250 }, { "epoch": 0.1759742849878353, "grad_norm": 0.7714797854423523, "learning_rate": 0.00019976102774893608, "loss": 4.5253, "step": 94300 }, { "epoch": 0.17606759054721377, "grad_norm": 1.129595398902893, "learning_rate": 0.00019976073740582864, "loss": 4.8271, "step": 94350 }, { "epoch": 0.17616089610659227, "grad_norm": 0.9854161739349365, "learning_rate": 0.00019976044688666106, "loss": 4.6552, "step": 94400 }, { "epoch": 0.17625420166597075, "grad_norm": 1.1386884450912476, "learning_rate": 0.00019976015619143384, "loss": 4.6417, "step": 94450 }, { "epoch": 0.17634750722534925, "grad_norm": 1.0306202173233032, "learning_rate": 0.0001997598653201475, "loss": 4.7774, "step": 94500 }, { "epoch": 0.17644081278472776, "grad_norm": 1.3340245485305786, "learning_rate": 0.00019975957427280255, "loss": 4.9127, "step": 94550 }, { "epoch": 0.17653411834410623, "grad_norm": 1.3886502981185913, "learning_rate": 0.0001997592830493995, "loss": 4.5873, "step": 94600 }, { "epoch": 0.17662742390348474, "grad_norm": 1.117301344871521, "learning_rate": 0.00019975899164993886, "loss": 4.7842, "step": 94650 }, { "epoch": 0.1767207294628632, "grad_norm": 1.071517825126648, "learning_rate": 0.00019975870007442117, "loss": 4.6684, "step": 94700 }, { "epoch": 0.17681403502224172, "grad_norm": 1.0764856338500977, "learning_rate": 0.0001997584083228469, "loss": 4.6823, "step": 94750 }, { "epoch": 0.1769073405816202, "grad_norm": 0.9086782932281494, "learning_rate": 0.00019975811639521662, "loss": 4.863, "step": 94800 }, { "epoch": 0.1770006461409987, "grad_norm": 0.9636552929878235, "learning_rate": 0.00019975782429153082, "loss": 4.685, "step": 94850 }, { "epoch": 0.1770939517003772, "grad_norm": 0.620227038860321, "learning_rate": 0.00019975753201179, "loss": 4.6816, "step": 94900 }, { "epoch": 0.17718725725975568, "grad_norm": 0.8294452428817749, "learning_rate": 0.00019975723955599468, "loss": 4.6858, "step": 94950 }, { "epoch": 0.17728056281913418, "grad_norm": 1.1841310262680054, "learning_rate": 0.0001997569469241454, "loss": 4.6845, "step": 95000 }, { "epoch": 0.17737386837851266, "grad_norm": 1.1353343725204468, "learning_rate": 0.00019975665411624266, "loss": 4.9758, "step": 95050 }, { "epoch": 0.17746717393789116, "grad_norm": 1.098388910293579, "learning_rate": 0.00019975636113228696, "loss": 4.807, "step": 95100 }, { "epoch": 0.17756047949726964, "grad_norm": 1.1415269374847412, "learning_rate": 0.00019975606797227886, "loss": 4.6393, "step": 95150 }, { "epoch": 0.17765378505664814, "grad_norm": 1.01333487033844, "learning_rate": 0.00019975577463621883, "loss": 4.7383, "step": 95200 }, { "epoch": 0.17774709061602664, "grad_norm": 0.9192553758621216, "learning_rate": 0.00019975548112410743, "loss": 4.7246, "step": 95250 }, { "epoch": 0.17784039617540512, "grad_norm": 1.233699917793274, "learning_rate": 0.00019975518743594516, "loss": 4.6757, "step": 95300 }, { "epoch": 0.17793370173478362, "grad_norm": 1.0001840591430664, "learning_rate": 0.00019975489357173252, "loss": 4.6488, "step": 95350 }, { "epoch": 0.1780270072941621, "grad_norm": 0.9523444771766663, "learning_rate": 0.00019975459953147005, "loss": 4.6393, "step": 95400 }, { "epoch": 0.1781203128535406, "grad_norm": 1.081119179725647, "learning_rate": 0.00019975430531515826, "loss": 4.583, "step": 95450 }, { "epoch": 0.17821361841291908, "grad_norm": 0.9086116552352905, "learning_rate": 0.00019975401092279767, "loss": 4.659, "step": 95500 }, { "epoch": 0.17830692397229758, "grad_norm": 0.9077746868133545, "learning_rate": 0.00019975371635438884, "loss": 4.5218, "step": 95550 }, { "epoch": 0.17840022953167606, "grad_norm": 1.0118244886398315, "learning_rate": 0.00019975342160993224, "loss": 4.9855, "step": 95600 }, { "epoch": 0.17849353509105456, "grad_norm": 0.7483125329017639, "learning_rate": 0.0001997531266894284, "loss": 4.5592, "step": 95650 }, { "epoch": 0.17858684065043307, "grad_norm": 0.9252362847328186, "learning_rate": 0.00019975283159287782, "loss": 4.6356, "step": 95700 }, { "epoch": 0.17868014620981154, "grad_norm": 0.7764643430709839, "learning_rate": 0.00019975253632028105, "loss": 4.6681, "step": 95750 }, { "epoch": 0.17877345176919004, "grad_norm": 0.9048830270767212, "learning_rate": 0.00019975224087163861, "loss": 4.5292, "step": 95800 }, { "epoch": 0.17886675732856852, "grad_norm": 0.9486568570137024, "learning_rate": 0.00019975194524695102, "loss": 4.7868, "step": 95850 }, { "epoch": 0.17896006288794702, "grad_norm": 0.8098161816596985, "learning_rate": 0.0001997516494462188, "loss": 4.6501, "step": 95900 }, { "epoch": 0.1790533684473255, "grad_norm": 1.2166732549667358, "learning_rate": 0.00019975135346944248, "loss": 4.4709, "step": 95950 }, { "epoch": 0.179146674006704, "grad_norm": 0.9364936351776123, "learning_rate": 0.00019975105731662255, "loss": 4.592, "step": 96000 }, { "epoch": 0.179146674006704, "eval_loss": 4.917341232299805, "eval_runtime": 231.2993, "eval_samples_per_second": 11.275, "eval_steps_per_second": 11.275, "eval_tts_loss": 7.391886383147933, "step": 96000 }, { "epoch": 0.1792399795660825, "grad_norm": 1.0819836854934692, "learning_rate": 0.00019975076098775957, "loss": 4.6746, "step": 96050 }, { "epoch": 0.17933328512546098, "grad_norm": 1.1346968412399292, "learning_rate": 0.00019975046448285407, "loss": 4.7887, "step": 96100 }, { "epoch": 0.1794265906848395, "grad_norm": 1.0858914852142334, "learning_rate": 0.00019975016780190652, "loss": 4.5072, "step": 96150 }, { "epoch": 0.17951989624421796, "grad_norm": 1.002199649810791, "learning_rate": 0.0001997498709449175, "loss": 4.7632, "step": 96200 }, { "epoch": 0.17961320180359647, "grad_norm": 0.8308196067810059, "learning_rate": 0.00019974957391188746, "loss": 4.6145, "step": 96250 }, { "epoch": 0.17970650736297494, "grad_norm": 0.9789503216743469, "learning_rate": 0.00019974927670281703, "loss": 4.6052, "step": 96300 }, { "epoch": 0.17979981292235345, "grad_norm": 1.1850559711456299, "learning_rate": 0.00019974897931770663, "loss": 4.6967, "step": 96350 }, { "epoch": 0.17989311848173195, "grad_norm": 1.0554805994033813, "learning_rate": 0.00019974868175655686, "loss": 4.8473, "step": 96400 }, { "epoch": 0.17998642404111043, "grad_norm": 1.0979632139205933, "learning_rate": 0.00019974838401936822, "loss": 4.8164, "step": 96450 }, { "epoch": 0.18007972960048893, "grad_norm": 1.1470582485198975, "learning_rate": 0.00019974808610614123, "loss": 4.674, "step": 96500 }, { "epoch": 0.1801730351598674, "grad_norm": 0.7371369004249573, "learning_rate": 0.00019974778801687643, "loss": 4.7965, "step": 96550 }, { "epoch": 0.1802663407192459, "grad_norm": 0.9566873908042908, "learning_rate": 0.00019974748975157432, "loss": 4.8604, "step": 96600 }, { "epoch": 0.18035964627862439, "grad_norm": 1.2158596515655518, "learning_rate": 0.00019974719131023543, "loss": 4.7986, "step": 96650 }, { "epoch": 0.1804529518380029, "grad_norm": 1.0148197412490845, "learning_rate": 0.0001997468926928603, "loss": 4.9964, "step": 96700 }, { "epoch": 0.1805462573973814, "grad_norm": 1.0894036293029785, "learning_rate": 0.00019974659389944947, "loss": 4.7985, "step": 96750 }, { "epoch": 0.18063956295675987, "grad_norm": 1.0293515920639038, "learning_rate": 0.00019974629493000348, "loss": 4.7646, "step": 96800 }, { "epoch": 0.18073286851613837, "grad_norm": 1.054319977760315, "learning_rate": 0.00019974599578452277, "loss": 4.6976, "step": 96850 }, { "epoch": 0.18082617407551685, "grad_norm": 1.062825083732605, "learning_rate": 0.00019974569646300799, "loss": 4.8446, "step": 96900 }, { "epoch": 0.18091947963489535, "grad_norm": 0.7005245089530945, "learning_rate": 0.00019974539696545956, "loss": 4.5875, "step": 96950 }, { "epoch": 0.18101278519427383, "grad_norm": 1.2943309545516968, "learning_rate": 0.0001997450972918781, "loss": 4.6447, "step": 97000 }, { "epoch": 0.18110609075365233, "grad_norm": 0.8457721471786499, "learning_rate": 0.00019974479744226404, "loss": 4.8779, "step": 97050 }, { "epoch": 0.1811993963130308, "grad_norm": 1.1781150102615356, "learning_rate": 0.000199744497416618, "loss": 4.7332, "step": 97100 }, { "epoch": 0.1812927018724093, "grad_norm": 0.9106166958808899, "learning_rate": 0.00019974419721494044, "loss": 4.5838, "step": 97150 }, { "epoch": 0.18138600743178782, "grad_norm": 1.1041089296340942, "learning_rate": 0.00019974389683723195, "loss": 4.5042, "step": 97200 }, { "epoch": 0.1814793129911663, "grad_norm": 0.9294777512550354, "learning_rate": 0.00019974359628349304, "loss": 4.8125, "step": 97250 }, { "epoch": 0.1815726185505448, "grad_norm": 1.158308506011963, "learning_rate": 0.0001997432955537242, "loss": 4.9712, "step": 97300 }, { "epoch": 0.18166592410992327, "grad_norm": 0.9689257740974426, "learning_rate": 0.00019974299464792602, "loss": 4.7698, "step": 97350 }, { "epoch": 0.18175922966930178, "grad_norm": 1.2290277481079102, "learning_rate": 0.00019974269356609902, "loss": 4.585, "step": 97400 }, { "epoch": 0.18185253522868025, "grad_norm": 1.2505154609680176, "learning_rate": 0.00019974239230824368, "loss": 4.7192, "step": 97450 }, { "epoch": 0.18194584078805875, "grad_norm": 1.1231937408447266, "learning_rate": 0.00019974209087436057, "loss": 4.6296, "step": 97500 }, { "epoch": 0.18203914634743726, "grad_norm": 1.0043781995773315, "learning_rate": 0.00019974178926445023, "loss": 4.7468, "step": 97550 }, { "epoch": 0.18213245190681573, "grad_norm": 0.986024796962738, "learning_rate": 0.0001997414874785132, "loss": 4.6579, "step": 97600 }, { "epoch": 0.18222575746619424, "grad_norm": 0.9532257914543152, "learning_rate": 0.00019974118551654998, "loss": 4.8355, "step": 97650 }, { "epoch": 0.18231906302557271, "grad_norm": 1.6420865058898926, "learning_rate": 0.0001997408833785611, "loss": 4.793, "step": 97700 }, { "epoch": 0.18241236858495122, "grad_norm": 1.155208706855774, "learning_rate": 0.00019974058106454713, "loss": 4.7287, "step": 97750 }, { "epoch": 0.1825056741443297, "grad_norm": 0.7323354482650757, "learning_rate": 0.00019974027857450858, "loss": 4.7684, "step": 97800 }, { "epoch": 0.1825989797037082, "grad_norm": 0.9008813500404358, "learning_rate": 0.000199739975908446, "loss": 4.858, "step": 97850 }, { "epoch": 0.1826922852630867, "grad_norm": 1.181888461112976, "learning_rate": 0.0001997396730663599, "loss": 4.7723, "step": 97900 }, { "epoch": 0.18278559082246518, "grad_norm": 0.7667856812477112, "learning_rate": 0.00019973937004825082, "loss": 4.5993, "step": 97950 }, { "epoch": 0.18287889638184368, "grad_norm": 1.1447443962097168, "learning_rate": 0.00019973906685411932, "loss": 4.5683, "step": 98000 }, { "epoch": 0.18297220194122216, "grad_norm": 0.7537021040916443, "learning_rate": 0.00019973876348396589, "loss": 4.7089, "step": 98050 }, { "epoch": 0.18306550750060066, "grad_norm": 0.937552809715271, "learning_rate": 0.00019973845993779112, "loss": 4.8778, "step": 98100 }, { "epoch": 0.18315881305997914, "grad_norm": 1.2293739318847656, "learning_rate": 0.0001997381562155955, "loss": 4.762, "step": 98150 }, { "epoch": 0.18325211861935764, "grad_norm": 1.0932244062423706, "learning_rate": 0.0001997378523173796, "loss": 4.8538, "step": 98200 }, { "epoch": 0.18334542417873612, "grad_norm": 1.1234608888626099, "learning_rate": 0.00019973754824314393, "loss": 4.6881, "step": 98250 }, { "epoch": 0.18343872973811462, "grad_norm": 0.9274517893791199, "learning_rate": 0.00019973724399288903, "loss": 4.8478, "step": 98300 }, { "epoch": 0.18353203529749312, "grad_norm": 0.9792783260345459, "learning_rate": 0.00019973693956661545, "loss": 4.5753, "step": 98350 }, { "epoch": 0.1836253408568716, "grad_norm": 1.3715656995773315, "learning_rate": 0.00019973663496432374, "loss": 4.6483, "step": 98400 }, { "epoch": 0.1837186464162501, "grad_norm": 0.7978438138961792, "learning_rate": 0.00019973633018601438, "loss": 4.6874, "step": 98450 }, { "epoch": 0.18381195197562858, "grad_norm": 0.9657984375953674, "learning_rate": 0.00019973602523168797, "loss": 4.7344, "step": 98500 }, { "epoch": 0.18390525753500708, "grad_norm": 1.0131194591522217, "learning_rate": 0.00019973572010134502, "loss": 4.6663, "step": 98550 }, { "epoch": 0.18399856309438556, "grad_norm": 1.2889825105667114, "learning_rate": 0.00019973541479498608, "loss": 4.6407, "step": 98600 }, { "epoch": 0.18409186865376406, "grad_norm": 1.0131268501281738, "learning_rate": 0.00019973510931261168, "loss": 4.8937, "step": 98650 }, { "epoch": 0.18418517421314257, "grad_norm": 0.9884664416313171, "learning_rate": 0.00019973480365422235, "loss": 4.7954, "step": 98700 }, { "epoch": 0.18427847977252104, "grad_norm": 1.0592337846755981, "learning_rate": 0.00019973449781981865, "loss": 4.7603, "step": 98750 }, { "epoch": 0.18437178533189955, "grad_norm": 0.9447229504585266, "learning_rate": 0.00019973419180940112, "loss": 4.5579, "step": 98800 }, { "epoch": 0.18446509089127802, "grad_norm": 1.186189889907837, "learning_rate": 0.00019973388562297026, "loss": 4.6064, "step": 98850 }, { "epoch": 0.18455839645065653, "grad_norm": 1.156333088874817, "learning_rate": 0.00019973357926052665, "loss": 4.6359, "step": 98900 }, { "epoch": 0.184651702010035, "grad_norm": 1.2002015113830566, "learning_rate": 0.00019973327272207082, "loss": 4.5993, "step": 98950 }, { "epoch": 0.1847450075694135, "grad_norm": 1.1841974258422852, "learning_rate": 0.00019973296600760333, "loss": 4.6125, "step": 99000 }, { "epoch": 0.1847450075694135, "eval_loss": 4.913722991943359, "eval_runtime": 229.7391, "eval_samples_per_second": 11.352, "eval_steps_per_second": 11.352, "eval_tts_loss": 7.435562890975742, "step": 99000 }, { "epoch": 0.184838313128792, "grad_norm": 1.0555998086929321, "learning_rate": 0.00019973265911712472, "loss": 4.7896, "step": 99050 }, { "epoch": 0.18493161868817048, "grad_norm": 1.1442583799362183, "learning_rate": 0.00019973235205063546, "loss": 4.695, "step": 99100 }, { "epoch": 0.185024924247549, "grad_norm": 1.149874210357666, "learning_rate": 0.0001997320448081362, "loss": 4.7661, "step": 99150 }, { "epoch": 0.18511822980692746, "grad_norm": 0.9762941002845764, "learning_rate": 0.00019973173738962742, "loss": 4.6591, "step": 99200 }, { "epoch": 0.18521153536630597, "grad_norm": 1.0889984369277954, "learning_rate": 0.00019973142979510963, "loss": 4.6857, "step": 99250 }, { "epoch": 0.18530484092568444, "grad_norm": 1.1313124895095825, "learning_rate": 0.00019973112202458345, "loss": 4.5218, "step": 99300 }, { "epoch": 0.18539814648506295, "grad_norm": 0.9728614687919617, "learning_rate": 0.0001997308140780494, "loss": 4.6974, "step": 99350 }, { "epoch": 0.18549145204444145, "grad_norm": 1.0745929479599, "learning_rate": 0.000199730505955508, "loss": 4.6735, "step": 99400 }, { "epoch": 0.18558475760381993, "grad_norm": 0.6565387845039368, "learning_rate": 0.00019973019765695977, "loss": 4.6161, "step": 99450 }, { "epoch": 0.18567806316319843, "grad_norm": 1.0615601539611816, "learning_rate": 0.00019972988918240533, "loss": 4.8634, "step": 99500 }, { "epoch": 0.1857713687225769, "grad_norm": 0.948047399520874, "learning_rate": 0.00019972958053184517, "loss": 4.8276, "step": 99550 }, { "epoch": 0.1858646742819554, "grad_norm": 1.0971624851226807, "learning_rate": 0.00019972927170527983, "loss": 4.9053, "step": 99600 }, { "epoch": 0.1859579798413339, "grad_norm": 0.9675891399383545, "learning_rate": 0.0001997289627027099, "loss": 4.7391, "step": 99650 }, { "epoch": 0.1860512854007124, "grad_norm": 0.8968614339828491, "learning_rate": 0.0001997286535241359, "loss": 4.6259, "step": 99700 }, { "epoch": 0.18614459096009087, "grad_norm": 0.772883951663971, "learning_rate": 0.00019972834416955837, "loss": 4.6393, "step": 99750 }, { "epoch": 0.18623789651946937, "grad_norm": 0.9643422365188599, "learning_rate": 0.00019972803463897784, "loss": 4.793, "step": 99800 }, { "epoch": 0.18633120207884787, "grad_norm": 1.0831654071807861, "learning_rate": 0.00019972772493239492, "loss": 4.7557, "step": 99850 }, { "epoch": 0.18642450763822635, "grad_norm": 1.0079478025436401, "learning_rate": 0.00019972741504981008, "loss": 4.4274, "step": 99900 }, { "epoch": 0.18651781319760485, "grad_norm": 0.9231737852096558, "learning_rate": 0.00019972710499122388, "loss": 4.8235, "step": 99950 }, { "epoch": 0.18661111875698333, "grad_norm": 1.2413074970245361, "learning_rate": 0.00019972679475663694, "loss": 4.676, "step": 100000 }, { "epoch": 0.18670442431636183, "grad_norm": 1.0384629964828491, "learning_rate": 0.0001997264843460497, "loss": 4.7109, "step": 100050 }, { "epoch": 0.1867977298757403, "grad_norm": 1.113856554031372, "learning_rate": 0.0001997261737594628, "loss": 4.5818, "step": 100100 }, { "epoch": 0.1868910354351188, "grad_norm": 1.0480083227157593, "learning_rate": 0.00019972586299687675, "loss": 4.6197, "step": 100150 }, { "epoch": 0.18698434099449732, "grad_norm": 1.1027523279190063, "learning_rate": 0.00019972555205829208, "loss": 4.714, "step": 100200 }, { "epoch": 0.1870776465538758, "grad_norm": 1.1065394878387451, "learning_rate": 0.00019972524094370936, "loss": 4.6669, "step": 100250 }, { "epoch": 0.1871709521132543, "grad_norm": 1.1278355121612549, "learning_rate": 0.00019972492965312916, "loss": 4.7749, "step": 100300 }, { "epoch": 0.18726425767263277, "grad_norm": 0.9157338738441467, "learning_rate": 0.00019972461818655197, "loss": 4.886, "step": 100350 }, { "epoch": 0.18735756323201128, "grad_norm": 0.8030517101287842, "learning_rate": 0.0001997243065439784, "loss": 4.7571, "step": 100400 }, { "epoch": 0.18745086879138975, "grad_norm": 0.9438421726226807, "learning_rate": 0.000199723994725409, "loss": 4.8845, "step": 100450 }, { "epoch": 0.18754417435076826, "grad_norm": 1.0755733251571655, "learning_rate": 0.00019972368273084426, "loss": 4.8225, "step": 100500 }, { "epoch": 0.18763747991014676, "grad_norm": 1.0262365341186523, "learning_rate": 0.00019972337056028475, "loss": 4.7597, "step": 100550 }, { "epoch": 0.18773078546952524, "grad_norm": 0.8555694222450256, "learning_rate": 0.00019972305821373108, "loss": 4.6938, "step": 100600 }, { "epoch": 0.18782409102890374, "grad_norm": 0.9016522169113159, "learning_rate": 0.00019972274569118374, "loss": 4.5223, "step": 100650 }, { "epoch": 0.18791739658828222, "grad_norm": 1.187323808670044, "learning_rate": 0.0001997224329926433, "loss": 4.7224, "step": 100700 }, { "epoch": 0.18801070214766072, "grad_norm": 1.1199045181274414, "learning_rate": 0.0001997221201181103, "loss": 4.5454, "step": 100750 }, { "epoch": 0.1881040077070392, "grad_norm": 0.8698449730873108, "learning_rate": 0.00019972180706758535, "loss": 4.7941, "step": 100800 }, { "epoch": 0.1881973132664177, "grad_norm": 1.17020583152771, "learning_rate": 0.00019972149384106893, "loss": 4.5653, "step": 100850 }, { "epoch": 0.18829061882579617, "grad_norm": 1.278030514717102, "learning_rate": 0.00019972118043856162, "loss": 4.4897, "step": 100900 }, { "epoch": 0.18838392438517468, "grad_norm": 0.8267766237258911, "learning_rate": 0.00019972086686006395, "loss": 4.5364, "step": 100950 }, { "epoch": 0.18847722994455318, "grad_norm": 1.0478912591934204, "learning_rate": 0.00019972055310557652, "loss": 4.8908, "step": 101000 }, { "epoch": 0.18857053550393166, "grad_norm": 1.0098049640655518, "learning_rate": 0.00019972023917509986, "loss": 4.6964, "step": 101050 }, { "epoch": 0.18866384106331016, "grad_norm": 0.9657384753227234, "learning_rate": 0.00019971992506863452, "loss": 4.5589, "step": 101100 }, { "epoch": 0.18875714662268864, "grad_norm": 1.05173921585083, "learning_rate": 0.00019971961078618108, "loss": 4.5888, "step": 101150 }, { "epoch": 0.18885045218206714, "grad_norm": 0.9653334617614746, "learning_rate": 0.00019971929632774002, "loss": 4.4529, "step": 101200 }, { "epoch": 0.18894375774144562, "grad_norm": 0.8235491514205933, "learning_rate": 0.000199718981693312, "loss": 4.6389, "step": 101250 }, { "epoch": 0.18903706330082412, "grad_norm": 0.921193540096283, "learning_rate": 0.0001997186668828975, "loss": 4.7759, "step": 101300 }, { "epoch": 0.18913036886020262, "grad_norm": 0.6879051327705383, "learning_rate": 0.0001997183518964971, "loss": 4.9006, "step": 101350 }, { "epoch": 0.1892236744195811, "grad_norm": 1.0606837272644043, "learning_rate": 0.00019971803673411135, "loss": 4.7399, "step": 101400 }, { "epoch": 0.1893169799789596, "grad_norm": 0.8906022906303406, "learning_rate": 0.00019971772139574083, "loss": 4.5369, "step": 101450 }, { "epoch": 0.18941028553833808, "grad_norm": 0.6738720536231995, "learning_rate": 0.0001997174058813861, "loss": 4.7231, "step": 101500 }, { "epoch": 0.18950359109771658, "grad_norm": 1.0936765670776367, "learning_rate": 0.00019971709019104764, "loss": 4.6172, "step": 101550 }, { "epoch": 0.18959689665709506, "grad_norm": 0.9944111108779907, "learning_rate": 0.0001997167743247261, "loss": 4.7858, "step": 101600 }, { "epoch": 0.18969020221647356, "grad_norm": 0.7380058765411377, "learning_rate": 0.00019971645828242195, "loss": 4.8101, "step": 101650 }, { "epoch": 0.18978350777585207, "grad_norm": 1.133588194847107, "learning_rate": 0.00019971614206413586, "loss": 4.5909, "step": 101700 }, { "epoch": 0.18987681333523054, "grad_norm": 0.9205106496810913, "learning_rate": 0.00019971582566986827, "loss": 4.4921, "step": 101750 }, { "epoch": 0.18997011889460905, "grad_norm": 1.0584709644317627, "learning_rate": 0.0001997155090996198, "loss": 4.7779, "step": 101800 }, { "epoch": 0.19006342445398752, "grad_norm": 1.113847017288208, "learning_rate": 0.00019971519235339105, "loss": 4.8153, "step": 101850 }, { "epoch": 0.19015673001336603, "grad_norm": 1.1247295141220093, "learning_rate": 0.00019971487543118247, "loss": 4.8936, "step": 101900 }, { "epoch": 0.1902500355727445, "grad_norm": 1.1859774589538574, "learning_rate": 0.00019971455833299473, "loss": 4.5792, "step": 101950 }, { "epoch": 0.190343341132123, "grad_norm": 0.8291439414024353, "learning_rate": 0.00019971424105882831, "loss": 4.5756, "step": 102000 }, { "epoch": 0.190343341132123, "eval_loss": 4.920861721038818, "eval_runtime": 229.1235, "eval_samples_per_second": 11.383, "eval_steps_per_second": 11.383, "eval_tts_loss": 7.3611460567746265, "step": 102000 }, { "epoch": 0.1904366466915015, "grad_norm": 0.6810999512672424, "learning_rate": 0.0001997139236086838, "loss": 4.8005, "step": 102050 }, { "epoch": 0.19052995225087999, "grad_norm": 1.0597984790802002, "learning_rate": 0.00019971360598256178, "loss": 4.7247, "step": 102100 }, { "epoch": 0.1906232578102585, "grad_norm": 0.9405797719955444, "learning_rate": 0.00019971328818046277, "loss": 4.628, "step": 102150 }, { "epoch": 0.19071656336963697, "grad_norm": 1.037429928779602, "learning_rate": 0.00019971297020238737, "loss": 4.8803, "step": 102200 }, { "epoch": 0.19080986892901547, "grad_norm": 1.3440206050872803, "learning_rate": 0.00019971265204833612, "loss": 4.7855, "step": 102250 }, { "epoch": 0.19090317448839395, "grad_norm": 1.0140005350112915, "learning_rate": 0.00019971233371830958, "loss": 4.6445, "step": 102300 }, { "epoch": 0.19099648004777245, "grad_norm": 1.2359912395477295, "learning_rate": 0.0001997120152123083, "loss": 4.8863, "step": 102350 }, { "epoch": 0.19108978560715092, "grad_norm": 0.9746628403663635, "learning_rate": 0.00019971169653033288, "loss": 4.7716, "step": 102400 }, { "epoch": 0.19118309116652943, "grad_norm": 1.2258669137954712, "learning_rate": 0.00019971137767238386, "loss": 4.9284, "step": 102450 }, { "epoch": 0.19127639672590793, "grad_norm": 0.9276184439659119, "learning_rate": 0.0001997110586384618, "loss": 4.5106, "step": 102500 }, { "epoch": 0.1913697022852864, "grad_norm": 1.0533397197723389, "learning_rate": 0.00019971073942856723, "loss": 4.8912, "step": 102550 }, { "epoch": 0.1914630078446649, "grad_norm": 1.0632177591323853, "learning_rate": 0.00019971042004270081, "loss": 4.764, "step": 102600 }, { "epoch": 0.1915563134040434, "grad_norm": 1.2086039781570435, "learning_rate": 0.000199710100480863, "loss": 4.8496, "step": 102650 }, { "epoch": 0.1916496189634219, "grad_norm": 1.030631184577942, "learning_rate": 0.00019970978074305444, "loss": 4.6516, "step": 102700 }, { "epoch": 0.19174292452280037, "grad_norm": 1.213368535041809, "learning_rate": 0.00019970946082927565, "loss": 4.5534, "step": 102750 }, { "epoch": 0.19183623008217887, "grad_norm": 0.9166072607040405, "learning_rate": 0.0001997091407395272, "loss": 4.6505, "step": 102800 }, { "epoch": 0.19192953564155738, "grad_norm": 1.179938793182373, "learning_rate": 0.00019970882047380965, "loss": 4.8351, "step": 102850 }, { "epoch": 0.19202284120093585, "grad_norm": 0.961173415184021, "learning_rate": 0.00019970850003212358, "loss": 4.7488, "step": 102900 }, { "epoch": 0.19211614676031435, "grad_norm": 0.8187737464904785, "learning_rate": 0.00019970817941446955, "loss": 4.8423, "step": 102950 }, { "epoch": 0.19220945231969283, "grad_norm": 1.1100882291793823, "learning_rate": 0.00019970785862084814, "loss": 4.7505, "step": 103000 }, { "epoch": 0.19230275787907133, "grad_norm": 0.842174232006073, "learning_rate": 0.00019970753765125992, "loss": 4.4804, "step": 103050 }, { "epoch": 0.1923960634384498, "grad_norm": 1.1668494939804077, "learning_rate": 0.0001997072165057054, "loss": 4.6164, "step": 103100 }, { "epoch": 0.19248936899782831, "grad_norm": 1.0573556423187256, "learning_rate": 0.0001997068951841852, "loss": 4.7643, "step": 103150 }, { "epoch": 0.19258267455720682, "grad_norm": 1.0298389196395874, "learning_rate": 0.00019970657368669989, "loss": 4.5873, "step": 103200 }, { "epoch": 0.1926759801165853, "grad_norm": 1.0277085304260254, "learning_rate": 0.00019970625201325002, "loss": 4.6273, "step": 103250 }, { "epoch": 0.1927692856759638, "grad_norm": 0.9712157249450684, "learning_rate": 0.00019970593016383617, "loss": 4.4374, "step": 103300 }, { "epoch": 0.19286259123534227, "grad_norm": 0.8751323223114014, "learning_rate": 0.00019970560813845886, "loss": 4.6056, "step": 103350 }, { "epoch": 0.19295589679472078, "grad_norm": 1.0707224607467651, "learning_rate": 0.00019970528593711874, "loss": 4.9276, "step": 103400 }, { "epoch": 0.19304920235409925, "grad_norm": 1.120894193649292, "learning_rate": 0.00019970496355981628, "loss": 4.5104, "step": 103450 }, { "epoch": 0.19314250791347776, "grad_norm": 0.9309062361717224, "learning_rate": 0.00019970464100655213, "loss": 4.775, "step": 103500 }, { "epoch": 0.19323581347285623, "grad_norm": 1.098646879196167, "learning_rate": 0.00019970431827732684, "loss": 4.6697, "step": 103550 }, { "epoch": 0.19332911903223474, "grad_norm": 1.0255775451660156, "learning_rate": 0.00019970399537214096, "loss": 4.7094, "step": 103600 }, { "epoch": 0.19342242459161324, "grad_norm": 0.9653803110122681, "learning_rate": 0.00019970367229099506, "loss": 4.6016, "step": 103650 }, { "epoch": 0.19351573015099172, "grad_norm": 1.1465661525726318, "learning_rate": 0.00019970334903388976, "loss": 4.892, "step": 103700 }, { "epoch": 0.19360903571037022, "grad_norm": 1.0665031671524048, "learning_rate": 0.00019970302560082557, "loss": 4.8088, "step": 103750 }, { "epoch": 0.1937023412697487, "grad_norm": 1.180091381072998, "learning_rate": 0.00019970270199180307, "loss": 4.7068, "step": 103800 }, { "epoch": 0.1937956468291272, "grad_norm": 0.9207305312156677, "learning_rate": 0.00019970237820682285, "loss": 4.6804, "step": 103850 }, { "epoch": 0.19388895238850568, "grad_norm": 0.6405048370361328, "learning_rate": 0.0001997020542458855, "loss": 4.6396, "step": 103900 }, { "epoch": 0.19398225794788418, "grad_norm": 0.9103243947029114, "learning_rate": 0.00019970173010899154, "loss": 4.7101, "step": 103950 }, { "epoch": 0.19407556350726268, "grad_norm": 1.086352825164795, "learning_rate": 0.00019970140579614158, "loss": 4.8022, "step": 104000 }, { "epoch": 0.19416886906664116, "grad_norm": 0.9352219104766846, "learning_rate": 0.0001997010813073362, "loss": 4.759, "step": 104050 }, { "epoch": 0.19426217462601966, "grad_norm": 0.8962497115135193, "learning_rate": 0.00019970075664257593, "loss": 4.7071, "step": 104100 }, { "epoch": 0.19435548018539814, "grad_norm": 1.1224924325942993, "learning_rate": 0.00019970043180186139, "loss": 4.8363, "step": 104150 }, { "epoch": 0.19444878574477664, "grad_norm": 0.8941819071769714, "learning_rate": 0.00019970010678519308, "loss": 4.7968, "step": 104200 }, { "epoch": 0.19454209130415512, "grad_norm": 1.0338774919509888, "learning_rate": 0.00019969978159257168, "loss": 4.7946, "step": 104250 }, { "epoch": 0.19463539686353362, "grad_norm": 1.0611014366149902, "learning_rate": 0.0001996994562239977, "loss": 4.7344, "step": 104300 }, { "epoch": 0.19472870242291213, "grad_norm": 0.9688970446586609, "learning_rate": 0.00019969913067947168, "loss": 4.6926, "step": 104350 }, { "epoch": 0.1948220079822906, "grad_norm": 0.716977059841156, "learning_rate": 0.00019969880495899426, "loss": 4.775, "step": 104400 }, { "epoch": 0.1949153135416691, "grad_norm": 1.210655927658081, "learning_rate": 0.000199698479062566, "loss": 4.7247, "step": 104450 }, { "epoch": 0.19500861910104758, "grad_norm": 1.2916840314865112, "learning_rate": 0.0001996981529901875, "loss": 4.6479, "step": 104500 }, { "epoch": 0.19510192466042608, "grad_norm": 1.0678060054779053, "learning_rate": 0.00019969782674185923, "loss": 4.7438, "step": 104550 }, { "epoch": 0.19519523021980456, "grad_norm": 0.8926165699958801, "learning_rate": 0.0001996975003175819, "loss": 4.7003, "step": 104600 }, { "epoch": 0.19528853577918306, "grad_norm": 0.8851156234741211, "learning_rate": 0.000199697173717356, "loss": 4.7207, "step": 104650 }, { "epoch": 0.19538184133856157, "grad_norm": 0.8675005435943604, "learning_rate": 0.0001996968469411821, "loss": 4.4109, "step": 104700 }, { "epoch": 0.19547514689794004, "grad_norm": 0.9853196740150452, "learning_rate": 0.00019969651998906084, "loss": 4.6939, "step": 104750 }, { "epoch": 0.19556845245731855, "grad_norm": 0.8662623167037964, "learning_rate": 0.0001996961928609928, "loss": 4.763, "step": 104800 }, { "epoch": 0.19566175801669702, "grad_norm": 1.1778194904327393, "learning_rate": 0.00019969586555697845, "loss": 4.6761, "step": 104850 }, { "epoch": 0.19575506357607553, "grad_norm": 1.2361232042312622, "learning_rate": 0.0001996955380770185, "loss": 4.7453, "step": 104900 }, { "epoch": 0.195848369135454, "grad_norm": 1.2205801010131836, "learning_rate": 0.00019969521042111343, "loss": 4.8602, "step": 104950 }, { "epoch": 0.1959416746948325, "grad_norm": 1.0522173643112183, "learning_rate": 0.00019969488258926384, "loss": 4.8686, "step": 105000 }, { "epoch": 0.1959416746948325, "eval_loss": 4.90353536605835, "eval_runtime": 230.4752, "eval_samples_per_second": 11.316, "eval_steps_per_second": 11.316, "eval_tts_loss": 7.433045758504386, "step": 105000 }, { "epoch": 0.19603498025421098, "grad_norm": 0.8923590183258057, "learning_rate": 0.00019969455458147036, "loss": 4.6476, "step": 105050 }, { "epoch": 0.1961282858135895, "grad_norm": 1.0533981323242188, "learning_rate": 0.00019969422639773353, "loss": 4.7786, "step": 105100 }, { "epoch": 0.196221591372968, "grad_norm": 1.0477826595306396, "learning_rate": 0.00019969389803805391, "loss": 4.6004, "step": 105150 }, { "epoch": 0.19631489693234647, "grad_norm": 1.089428186416626, "learning_rate": 0.00019969356950243214, "loss": 4.6546, "step": 105200 }, { "epoch": 0.19640820249172497, "grad_norm": 0.7010145783424377, "learning_rate": 0.00019969324079086872, "loss": 4.6761, "step": 105250 }, { "epoch": 0.19650150805110345, "grad_norm": 0.9401247501373291, "learning_rate": 0.0001996929119033643, "loss": 4.5199, "step": 105300 }, { "epoch": 0.19659481361048195, "grad_norm": 1.1464163064956665, "learning_rate": 0.00019969258283991943, "loss": 4.7447, "step": 105350 }, { "epoch": 0.19668811916986043, "grad_norm": 0.9506011605262756, "learning_rate": 0.0001996922536005347, "loss": 4.9001, "step": 105400 }, { "epoch": 0.19678142472923893, "grad_norm": 1.0676487684249878, "learning_rate": 0.00019969192418521064, "loss": 4.5418, "step": 105450 }, { "epoch": 0.19687473028861743, "grad_norm": 1.0197359323501587, "learning_rate": 0.0001996915945939479, "loss": 4.706, "step": 105500 }, { "epoch": 0.1969680358479959, "grad_norm": 1.120069980621338, "learning_rate": 0.00019969126482674706, "loss": 4.6804, "step": 105550 }, { "epoch": 0.1970613414073744, "grad_norm": 0.9666004776954651, "learning_rate": 0.00019969093488360863, "loss": 4.792, "step": 105600 }, { "epoch": 0.1971546469667529, "grad_norm": 1.1865931749343872, "learning_rate": 0.0001996906047645333, "loss": 4.6457, "step": 105650 }, { "epoch": 0.1972479525261314, "grad_norm": 0.9605655074119568, "learning_rate": 0.00019969027446952155, "loss": 4.8505, "step": 105700 }, { "epoch": 0.19734125808550987, "grad_norm": 0.6585858464241028, "learning_rate": 0.00019968994399857404, "loss": 4.5811, "step": 105750 }, { "epoch": 0.19743456364488837, "grad_norm": 1.0499614477157593, "learning_rate": 0.0001996896133516913, "loss": 4.661, "step": 105800 }, { "epoch": 0.19752786920426688, "grad_norm": 1.0251386165618896, "learning_rate": 0.0001996892825288739, "loss": 4.8558, "step": 105850 }, { "epoch": 0.19762117476364535, "grad_norm": 1.2540901899337769, "learning_rate": 0.00019968895153012252, "loss": 4.5407, "step": 105900 }, { "epoch": 0.19771448032302386, "grad_norm": 0.6922102570533752, "learning_rate": 0.00019968862035543765, "loss": 4.5388, "step": 105950 }, { "epoch": 0.19780778588240233, "grad_norm": 1.0659449100494385, "learning_rate": 0.00019968828900481992, "loss": 4.6475, "step": 106000 }, { "epoch": 0.19790109144178084, "grad_norm": 0.8389386534690857, "learning_rate": 0.00019968795747826986, "loss": 4.555, "step": 106050 }, { "epoch": 0.1979943970011593, "grad_norm": 0.8430420160293579, "learning_rate": 0.00019968762577578814, "loss": 4.7839, "step": 106100 }, { "epoch": 0.19808770256053781, "grad_norm": 1.0360347032546997, "learning_rate": 0.00019968729389737528, "loss": 4.7264, "step": 106150 }, { "epoch": 0.19818100811991632, "grad_norm": 0.9396027326583862, "learning_rate": 0.0001996869618430319, "loss": 4.5931, "step": 106200 }, { "epoch": 0.1982743136792948, "grad_norm": 0.8887374997138977, "learning_rate": 0.00019968662961275855, "loss": 4.6601, "step": 106250 }, { "epoch": 0.1983676192386733, "grad_norm": 1.1754348278045654, "learning_rate": 0.00019968629720655586, "loss": 4.7664, "step": 106300 }, { "epoch": 0.19846092479805177, "grad_norm": 0.6698839068412781, "learning_rate": 0.0001996859646244244, "loss": 4.5077, "step": 106350 }, { "epoch": 0.19855423035743028, "grad_norm": 0.944202184677124, "learning_rate": 0.00019968563186636474, "loss": 4.6335, "step": 106400 }, { "epoch": 0.19864753591680875, "grad_norm": 1.1213611364364624, "learning_rate": 0.00019968529893237745, "loss": 4.6235, "step": 106450 }, { "epoch": 0.19874084147618726, "grad_norm": 1.068108081817627, "learning_rate": 0.0001996849658224632, "loss": 4.6891, "step": 106500 }, { "epoch": 0.19883414703556573, "grad_norm": 1.1877700090408325, "learning_rate": 0.0001996846325366225, "loss": 4.5616, "step": 106550 }, { "epoch": 0.19892745259494424, "grad_norm": 0.9596836566925049, "learning_rate": 0.00019968429907485597, "loss": 4.612, "step": 106600 }, { "epoch": 0.19902075815432274, "grad_norm": 0.8319620490074158, "learning_rate": 0.0001996839654371642, "loss": 4.6712, "step": 106650 }, { "epoch": 0.19911406371370122, "grad_norm": 0.921432375907898, "learning_rate": 0.00019968363162354774, "loss": 4.7998, "step": 106700 }, { "epoch": 0.19920736927307972, "grad_norm": 1.089003324508667, "learning_rate": 0.00019968329763400726, "loss": 4.5917, "step": 106750 }, { "epoch": 0.1993006748324582, "grad_norm": 1.233829140663147, "learning_rate": 0.00019968296346854327, "loss": 4.8402, "step": 106800 }, { "epoch": 0.1993939803918367, "grad_norm": 1.0184001922607422, "learning_rate": 0.0001996826291271564, "loss": 4.7715, "step": 106850 }, { "epoch": 0.19948728595121518, "grad_norm": 1.058369517326355, "learning_rate": 0.0001996822946098472, "loss": 4.6193, "step": 106900 }, { "epoch": 0.19958059151059368, "grad_norm": 1.2557109594345093, "learning_rate": 0.00019968195991661633, "loss": 4.9265, "step": 106950 }, { "epoch": 0.19967389706997218, "grad_norm": 0.7956662178039551, "learning_rate": 0.00019968162504746434, "loss": 4.8568, "step": 107000 }, { "epoch": 0.19976720262935066, "grad_norm": 0.9638285040855408, "learning_rate": 0.00019968129000239179, "loss": 4.6949, "step": 107050 }, { "epoch": 0.19986050818872916, "grad_norm": 1.0805922746658325, "learning_rate": 0.0001996809547813993, "loss": 4.7621, "step": 107100 }, { "epoch": 0.19995381374810764, "grad_norm": 0.9019108414649963, "learning_rate": 0.00019968061938448748, "loss": 4.4901, "step": 107150 }, { "epoch": 0.20004711930748614, "grad_norm": 0.7746037840843201, "learning_rate": 0.0001996802838116569, "loss": 4.4827, "step": 107200 }, { "epoch": 0.20014042486686462, "grad_norm": 1.065016508102417, "learning_rate": 0.00019967994806290818, "loss": 4.8291, "step": 107250 }, { "epoch": 0.20023373042624312, "grad_norm": 1.0848886966705322, "learning_rate": 0.00019967961213824187, "loss": 4.7838, "step": 107300 }, { "epoch": 0.20032703598562163, "grad_norm": 0.9382323026657104, "learning_rate": 0.0001996792760376586, "loss": 4.6825, "step": 107350 }, { "epoch": 0.2004203415450001, "grad_norm": 1.0932012796401978, "learning_rate": 0.00019967893976115893, "loss": 4.8105, "step": 107400 }, { "epoch": 0.2005136471043786, "grad_norm": 0.8511140942573547, "learning_rate": 0.00019967860330874344, "loss": 4.6117, "step": 107450 }, { "epoch": 0.20060695266375708, "grad_norm": 1.1722475290298462, "learning_rate": 0.00019967826668041281, "loss": 4.6226, "step": 107500 }, { "epoch": 0.20070025822313559, "grad_norm": 1.120883822441101, "learning_rate": 0.00019967792987616757, "loss": 4.8838, "step": 107550 }, { "epoch": 0.20079356378251406, "grad_norm": 0.9528012871742249, "learning_rate": 0.0001996775928960083, "loss": 4.7664, "step": 107600 }, { "epoch": 0.20088686934189257, "grad_norm": 0.9668023586273193, "learning_rate": 0.0001996772557399356, "loss": 4.6156, "step": 107650 }, { "epoch": 0.20098017490127104, "grad_norm": 0.8092545866966248, "learning_rate": 0.00019967691840795014, "loss": 4.8754, "step": 107700 }, { "epoch": 0.20107348046064955, "grad_norm": 1.1513761281967163, "learning_rate": 0.0001996765809000524, "loss": 4.7403, "step": 107750 }, { "epoch": 0.20116678602002805, "grad_norm": 1.2313563823699951, "learning_rate": 0.00019967624321624307, "loss": 4.7554, "step": 107800 }, { "epoch": 0.20126009157940652, "grad_norm": 1.2222445011138916, "learning_rate": 0.0001996759053565227, "loss": 4.6705, "step": 107850 }, { "epoch": 0.20135339713878503, "grad_norm": 1.0731276273727417, "learning_rate": 0.00019967556732089189, "loss": 4.7103, "step": 107900 }, { "epoch": 0.2014467026981635, "grad_norm": 0.8914058804512024, "learning_rate": 0.00019967522910935127, "loss": 4.7109, "step": 107950 }, { "epoch": 0.201540008257542, "grad_norm": 0.9752347469329834, "learning_rate": 0.00019967489072190137, "loss": 4.8808, "step": 108000 }, { "epoch": 0.201540008257542, "eval_loss": 4.898900508880615, "eval_runtime": 231.0932, "eval_samples_per_second": 11.285, "eval_steps_per_second": 11.285, "eval_tts_loss": 7.3902602159986674, "step": 108000 }, { "epoch": 0.20163331381692048, "grad_norm": 1.1584863662719727, "learning_rate": 0.00019967455215854285, "loss": 4.6683, "step": 108050 }, { "epoch": 0.201726619376299, "grad_norm": 1.2001816034317017, "learning_rate": 0.00019967421341927628, "loss": 4.6347, "step": 108100 }, { "epoch": 0.2018199249356775, "grad_norm": 1.0059622526168823, "learning_rate": 0.00019967387450410225, "loss": 4.9153, "step": 108150 }, { "epoch": 0.20191323049505597, "grad_norm": 0.6907665729522705, "learning_rate": 0.00019967353541302138, "loss": 4.6052, "step": 108200 }, { "epoch": 0.20200653605443447, "grad_norm": 0.9193375706672668, "learning_rate": 0.00019967319614603427, "loss": 4.7746, "step": 108250 }, { "epoch": 0.20209984161381295, "grad_norm": 0.8793426752090454, "learning_rate": 0.00019967285670314147, "loss": 4.6992, "step": 108300 }, { "epoch": 0.20219314717319145, "grad_norm": 1.1513392925262451, "learning_rate": 0.00019967251708434365, "loss": 4.9056, "step": 108350 }, { "epoch": 0.20228645273256993, "grad_norm": 1.0788843631744385, "learning_rate": 0.00019967217728964135, "loss": 4.6675, "step": 108400 }, { "epoch": 0.20237975829194843, "grad_norm": 0.9058928489685059, "learning_rate": 0.0001996718373190352, "loss": 4.7488, "step": 108450 }, { "epoch": 0.20247306385132693, "grad_norm": 0.7392362952232361, "learning_rate": 0.0001996714971725258, "loss": 4.6251, "step": 108500 }, { "epoch": 0.2025663694107054, "grad_norm": 0.9372434616088867, "learning_rate": 0.00019967115685011377, "loss": 4.9667, "step": 108550 }, { "epoch": 0.20265967497008391, "grad_norm": 1.1257874965667725, "learning_rate": 0.00019967081635179968, "loss": 4.7603, "step": 108600 }, { "epoch": 0.2027529805294624, "grad_norm": 0.8157814741134644, "learning_rate": 0.0001996704756775841, "loss": 4.5966, "step": 108650 }, { "epoch": 0.2028462860888409, "grad_norm": 1.0224437713623047, "learning_rate": 0.0001996701348274677, "loss": 4.6797, "step": 108700 }, { "epoch": 0.20293959164821937, "grad_norm": 1.2594748735427856, "learning_rate": 0.00019966979380145103, "loss": 4.6246, "step": 108750 }, { "epoch": 0.20303289720759787, "grad_norm": 0.936982274055481, "learning_rate": 0.00019966945259953473, "loss": 4.7162, "step": 108800 }, { "epoch": 0.20312620276697638, "grad_norm": 1.1596720218658447, "learning_rate": 0.00019966911122171936, "loss": 4.6277, "step": 108850 }, { "epoch": 0.20321950832635485, "grad_norm": 0.9620667099952698, "learning_rate": 0.00019966876966800556, "loss": 4.7468, "step": 108900 }, { "epoch": 0.20331281388573336, "grad_norm": 1.2302641868591309, "learning_rate": 0.00019966842793839394, "loss": 4.6094, "step": 108950 }, { "epoch": 0.20340611944511183, "grad_norm": 1.131081461906433, "learning_rate": 0.00019966808603288504, "loss": 4.628, "step": 109000 }, { "epoch": 0.20349942500449034, "grad_norm": 0.9959622621536255, "learning_rate": 0.00019966774395147955, "loss": 4.8578, "step": 109050 }, { "epoch": 0.2035927305638688, "grad_norm": 1.1797744035720825, "learning_rate": 0.00019966740169417803, "loss": 4.8179, "step": 109100 }, { "epoch": 0.20368603612324732, "grad_norm": 1.0603262186050415, "learning_rate": 0.00019966705926098102, "loss": 4.6757, "step": 109150 }, { "epoch": 0.2037793416826258, "grad_norm": 1.2537606954574585, "learning_rate": 0.00019966671665188924, "loss": 4.5657, "step": 109200 }, { "epoch": 0.2038726472420043, "grad_norm": 0.8469638228416443, "learning_rate": 0.0001996663738669032, "loss": 4.5299, "step": 109250 }, { "epoch": 0.2039659528013828, "grad_norm": 1.1526724100112915, "learning_rate": 0.0001996660309060236, "loss": 4.6851, "step": 109300 }, { "epoch": 0.20405925836076128, "grad_norm": 0.9065315127372742, "learning_rate": 0.00019966568776925097, "loss": 4.6405, "step": 109350 }, { "epoch": 0.20415256392013978, "grad_norm": 1.0660070180892944, "learning_rate": 0.00019966534445658591, "loss": 4.8246, "step": 109400 }, { "epoch": 0.20424586947951825, "grad_norm": 0.9208210110664368, "learning_rate": 0.0001996650009680291, "loss": 4.8364, "step": 109450 }, { "epoch": 0.20433917503889676, "grad_norm": 1.1117137670516968, "learning_rate": 0.00019966465730358106, "loss": 4.6232, "step": 109500 }, { "epoch": 0.20443248059827523, "grad_norm": 0.7682850360870361, "learning_rate": 0.00019966431346324246, "loss": 4.6112, "step": 109550 }, { "epoch": 0.20452578615765374, "grad_norm": 0.9791049957275391, "learning_rate": 0.0001996639694470139, "loss": 4.739, "step": 109600 }, { "epoch": 0.20461909171703224, "grad_norm": 0.7025151252746582, "learning_rate": 0.00019966362525489592, "loss": 4.625, "step": 109650 }, { "epoch": 0.20471239727641072, "grad_norm": 0.9386808276176453, "learning_rate": 0.00019966328088688923, "loss": 4.7254, "step": 109700 }, { "epoch": 0.20480570283578922, "grad_norm": 0.9682460427284241, "learning_rate": 0.00019966293634299435, "loss": 4.6955, "step": 109750 }, { "epoch": 0.2048990083951677, "grad_norm": 0.9403902292251587, "learning_rate": 0.00019966259162321196, "loss": 4.648, "step": 109800 }, { "epoch": 0.2049923139545462, "grad_norm": 0.8865294456481934, "learning_rate": 0.00019966224672754259, "loss": 4.6593, "step": 109850 }, { "epoch": 0.20508561951392468, "grad_norm": 0.9685643315315247, "learning_rate": 0.0001996619016559869, "loss": 4.8078, "step": 109900 }, { "epoch": 0.20517892507330318, "grad_norm": 1.1610651016235352, "learning_rate": 0.00019966155640854552, "loss": 4.7153, "step": 109950 }, { "epoch": 0.20527223063268168, "grad_norm": 1.3731842041015625, "learning_rate": 0.000199661210985219, "loss": 4.6649, "step": 110000 }, { "epoch": 0.20536553619206016, "grad_norm": 0.9859686493873596, "learning_rate": 0.00019966086538600797, "loss": 4.649, "step": 110050 }, { "epoch": 0.20545884175143866, "grad_norm": 0.9828516244888306, "learning_rate": 0.00019966051961091305, "loss": 4.6669, "step": 110100 }, { "epoch": 0.20555214731081714, "grad_norm": 1.1860650777816772, "learning_rate": 0.00019966017365993487, "loss": 4.6617, "step": 110150 }, { "epoch": 0.20564545287019564, "grad_norm": 0.7568049430847168, "learning_rate": 0.00019965982753307403, "loss": 4.5716, "step": 110200 }, { "epoch": 0.20573875842957412, "grad_norm": 1.4290099143981934, "learning_rate": 0.0001996594812303311, "loss": 4.7415, "step": 110250 }, { "epoch": 0.20583206398895262, "grad_norm": 1.090683937072754, "learning_rate": 0.00019965913475170673, "loss": 4.7548, "step": 110300 }, { "epoch": 0.2059253695483311, "grad_norm": 1.1351553201675415, "learning_rate": 0.00019965878809720153, "loss": 4.6024, "step": 110350 }, { "epoch": 0.2060186751077096, "grad_norm": 0.8278642296791077, "learning_rate": 0.0001996584412668161, "loss": 4.7692, "step": 110400 }, { "epoch": 0.2061119806670881, "grad_norm": 0.9527729749679565, "learning_rate": 0.00019965809426055104, "loss": 4.6836, "step": 110450 }, { "epoch": 0.20620528622646658, "grad_norm": 0.9294393062591553, "learning_rate": 0.000199657747078407, "loss": 4.7799, "step": 110500 }, { "epoch": 0.2062985917858451, "grad_norm": 0.8470353484153748, "learning_rate": 0.00019965739972038456, "loss": 4.6028, "step": 110550 }, { "epoch": 0.20639189734522356, "grad_norm": 1.0058228969573975, "learning_rate": 0.00019965705218648436, "loss": 4.7502, "step": 110600 }, { "epoch": 0.20648520290460207, "grad_norm": 1.0412211418151855, "learning_rate": 0.000199656704476707, "loss": 4.7988, "step": 110650 }, { "epoch": 0.20657850846398054, "grad_norm": 1.1575844287872314, "learning_rate": 0.00019965635659105306, "loss": 4.905, "step": 110700 }, { "epoch": 0.20667181402335905, "grad_norm": 1.1507608890533447, "learning_rate": 0.0001996560085295232, "loss": 4.708, "step": 110750 }, { "epoch": 0.20676511958273755, "grad_norm": 1.1214720010757446, "learning_rate": 0.00019965566029211807, "loss": 4.7185, "step": 110800 }, { "epoch": 0.20685842514211603, "grad_norm": 0.8996137380599976, "learning_rate": 0.00019965531187883816, "loss": 4.6913, "step": 110850 }, { "epoch": 0.20695173070149453, "grad_norm": 1.0486303567886353, "learning_rate": 0.0001996549632896842, "loss": 4.9886, "step": 110900 }, { "epoch": 0.207045036260873, "grad_norm": 0.9184983968734741, "learning_rate": 0.00019965461452465676, "loss": 4.4322, "step": 110950 }, { "epoch": 0.2071383418202515, "grad_norm": 1.004927396774292, "learning_rate": 0.00019965426558375647, "loss": 4.7627, "step": 111000 }, { "epoch": 0.2071383418202515, "eval_loss": 4.893270015716553, "eval_runtime": 228.2153, "eval_samples_per_second": 11.428, "eval_steps_per_second": 11.428, "eval_tts_loss": 7.3785387509319635, "step": 111000 }, { "epoch": 0.20723164737962999, "grad_norm": 1.0662178993225098, "learning_rate": 0.0001996539164669839, "loss": 4.5831, "step": 111050 }, { "epoch": 0.2073249529390085, "grad_norm": 1.0260896682739258, "learning_rate": 0.00019965356717433971, "loss": 4.6671, "step": 111100 }, { "epoch": 0.207418258498387, "grad_norm": 0.8488427400588989, "learning_rate": 0.00019965321770582455, "loss": 4.8138, "step": 111150 }, { "epoch": 0.20751156405776547, "grad_norm": 0.9575690627098083, "learning_rate": 0.00019965286806143898, "loss": 4.5885, "step": 111200 }, { "epoch": 0.20760486961714397, "grad_norm": 0.9709866046905518, "learning_rate": 0.00019965251824118361, "loss": 4.721, "step": 111250 }, { "epoch": 0.20769817517652245, "grad_norm": 1.058888554573059, "learning_rate": 0.00019965216824505908, "loss": 4.681, "step": 111300 }, { "epoch": 0.20779148073590095, "grad_norm": 0.9510217905044556, "learning_rate": 0.00019965181807306604, "loss": 4.84, "step": 111350 }, { "epoch": 0.20788478629527943, "grad_norm": 1.249549388885498, "learning_rate": 0.00019965146772520505, "loss": 4.5905, "step": 111400 }, { "epoch": 0.20797809185465793, "grad_norm": 0.9404703378677368, "learning_rate": 0.00019965111720147673, "loss": 4.7374, "step": 111450 }, { "epoch": 0.20807139741403644, "grad_norm": 0.9950361251831055, "learning_rate": 0.00019965076650188177, "loss": 4.7764, "step": 111500 }, { "epoch": 0.2081647029734149, "grad_norm": 1.015062689781189, "learning_rate": 0.0001996504156264207, "loss": 4.7205, "step": 111550 }, { "epoch": 0.20825800853279341, "grad_norm": 1.0537135601043701, "learning_rate": 0.00019965006457509422, "loss": 4.7283, "step": 111600 }, { "epoch": 0.2083513140921719, "grad_norm": 0.9111753106117249, "learning_rate": 0.00019964971334790287, "loss": 4.5574, "step": 111650 }, { "epoch": 0.2084446196515504, "grad_norm": 1.2378709316253662, "learning_rate": 0.00019964936194484737, "loss": 4.5554, "step": 111700 }, { "epoch": 0.20853792521092887, "grad_norm": 0.8984060883522034, "learning_rate": 0.00019964901036592823, "loss": 4.6334, "step": 111750 }, { "epoch": 0.20863123077030737, "grad_norm": 0.6208471655845642, "learning_rate": 0.0001996486586111461, "loss": 4.5307, "step": 111800 }, { "epoch": 0.20872453632968585, "grad_norm": 1.0021463632583618, "learning_rate": 0.00019964830668050168, "loss": 4.8079, "step": 111850 }, { "epoch": 0.20881784188906435, "grad_norm": 0.6217372417449951, "learning_rate": 0.00019964795457399549, "loss": 4.5041, "step": 111900 }, { "epoch": 0.20891114744844286, "grad_norm": 1.1446421146392822, "learning_rate": 0.00019964760229162823, "loss": 4.656, "step": 111950 }, { "epoch": 0.20900445300782133, "grad_norm": 0.9120539426803589, "learning_rate": 0.00019964724983340045, "loss": 4.6973, "step": 112000 }, { "epoch": 0.20909775856719984, "grad_norm": 0.6660919189453125, "learning_rate": 0.0001996468971993128, "loss": 4.584, "step": 112050 }, { "epoch": 0.2091910641265783, "grad_norm": 1.0424081087112427, "learning_rate": 0.00019964654438936594, "loss": 4.7912, "step": 112100 }, { "epoch": 0.20928436968595682, "grad_norm": 0.8505585789680481, "learning_rate": 0.00019964619140356047, "loss": 4.6274, "step": 112150 }, { "epoch": 0.2093776752453353, "grad_norm": 1.0397758483886719, "learning_rate": 0.00019964583824189699, "loss": 4.4355, "step": 112200 }, { "epoch": 0.2094709808047138, "grad_norm": 1.0277235507965088, "learning_rate": 0.00019964548490437612, "loss": 4.7227, "step": 112250 }, { "epoch": 0.2095642863640923, "grad_norm": 1.0164071321487427, "learning_rate": 0.0001996451313909985, "loss": 4.6742, "step": 112300 }, { "epoch": 0.20965759192347078, "grad_norm": 1.1814101934432983, "learning_rate": 0.0001996447777017648, "loss": 4.5611, "step": 112350 }, { "epoch": 0.20975089748284928, "grad_norm": 0.9306598901748657, "learning_rate": 0.00019964442383667557, "loss": 4.749, "step": 112400 }, { "epoch": 0.20984420304222776, "grad_norm": 0.7609649896621704, "learning_rate": 0.0001996440697957315, "loss": 4.8032, "step": 112450 }, { "epoch": 0.20993750860160626, "grad_norm": 0.7726313471794128, "learning_rate": 0.00019964371557893313, "loss": 4.6712, "step": 112500 }, { "epoch": 0.21003081416098474, "grad_norm": 0.718001127243042, "learning_rate": 0.00019964336118628116, "loss": 4.3786, "step": 112550 }, { "epoch": 0.21012411972036324, "grad_norm": 1.221449375152588, "learning_rate": 0.00019964300661777618, "loss": 4.6134, "step": 112600 }, { "epoch": 0.21021742527974174, "grad_norm": 0.9702039361000061, "learning_rate": 0.0001996426518734188, "loss": 4.4521, "step": 112650 }, { "epoch": 0.21031073083912022, "grad_norm": 1.0901730060577393, "learning_rate": 0.00019964229695320973, "loss": 4.7553, "step": 112700 }, { "epoch": 0.21040403639849872, "grad_norm": 1.4701701402664185, "learning_rate": 0.00019964194185714953, "loss": 4.4546, "step": 112750 }, { "epoch": 0.2104973419578772, "grad_norm": 1.0534309148788452, "learning_rate": 0.0001996415865852388, "loss": 4.618, "step": 112800 }, { "epoch": 0.2105906475172557, "grad_norm": 0.875619649887085, "learning_rate": 0.00019964123113747824, "loss": 4.7289, "step": 112850 }, { "epoch": 0.21068395307663418, "grad_norm": 1.1053240299224854, "learning_rate": 0.00019964087551386841, "loss": 4.6377, "step": 112900 }, { "epoch": 0.21077725863601268, "grad_norm": 1.1777266263961792, "learning_rate": 0.00019964051971440997, "loss": 4.9208, "step": 112950 }, { "epoch": 0.21087056419539116, "grad_norm": 0.9478060007095337, "learning_rate": 0.00019964016373910358, "loss": 4.5935, "step": 113000 }, { "epoch": 0.21096386975476966, "grad_norm": 1.104618787765503, "learning_rate": 0.0001996398075879498, "loss": 4.7408, "step": 113050 }, { "epoch": 0.21105717531414817, "grad_norm": 0.9139821529388428, "learning_rate": 0.00019963945126094928, "loss": 4.8311, "step": 113100 }, { "epoch": 0.21115048087352664, "grad_norm": 1.1723382472991943, "learning_rate": 0.00019963909475810267, "loss": 4.7919, "step": 113150 }, { "epoch": 0.21124378643290515, "grad_norm": 1.1692183017730713, "learning_rate": 0.0001996387380794106, "loss": 4.8049, "step": 113200 }, { "epoch": 0.21133709199228362, "grad_norm": 0.8972973823547363, "learning_rate": 0.0001996383812248737, "loss": 4.6703, "step": 113250 }, { "epoch": 0.21143039755166212, "grad_norm": 0.8258083462715149, "learning_rate": 0.00019963802419449258, "loss": 4.6485, "step": 113300 }, { "epoch": 0.2115237031110406, "grad_norm": 1.101173758506775, "learning_rate": 0.00019963766698826785, "loss": 4.5944, "step": 113350 }, { "epoch": 0.2116170086704191, "grad_norm": 0.9222469925880432, "learning_rate": 0.00019963730960620023, "loss": 4.6801, "step": 113400 }, { "epoch": 0.2117103142297976, "grad_norm": 0.8024877309799194, "learning_rate": 0.00019963695204829022, "loss": 4.6408, "step": 113450 }, { "epoch": 0.21180361978917608, "grad_norm": 1.1959446668624878, "learning_rate": 0.00019963659431453858, "loss": 4.7532, "step": 113500 }, { "epoch": 0.2118969253485546, "grad_norm": 1.03960382938385, "learning_rate": 0.00019963623640494586, "loss": 4.531, "step": 113550 }, { "epoch": 0.21199023090793306, "grad_norm": 0.9562681317329407, "learning_rate": 0.0001996358783195127, "loss": 4.525, "step": 113600 }, { "epoch": 0.21208353646731157, "grad_norm": 0.9814481735229492, "learning_rate": 0.00019963552005823975, "loss": 4.5483, "step": 113650 }, { "epoch": 0.21217684202669004, "grad_norm": 0.948824942111969, "learning_rate": 0.00019963516162112765, "loss": 4.8581, "step": 113700 }, { "epoch": 0.21227014758606855, "grad_norm": 1.0382654666900635, "learning_rate": 0.00019963480300817704, "loss": 4.7724, "step": 113750 }, { "epoch": 0.21236345314544705, "grad_norm": 1.0841362476348877, "learning_rate": 0.00019963444421938848, "loss": 4.8569, "step": 113800 }, { "epoch": 0.21245675870482553, "grad_norm": 1.138662338256836, "learning_rate": 0.00019963408525476272, "loss": 4.828, "step": 113850 }, { "epoch": 0.21255006426420403, "grad_norm": 0.9034907817840576, "learning_rate": 0.00019963372611430028, "loss": 4.7168, "step": 113900 }, { "epoch": 0.2126433698235825, "grad_norm": 0.9675794243812561, "learning_rate": 0.0001996333667980019, "loss": 4.9587, "step": 113950 }, { "epoch": 0.212736675382961, "grad_norm": 0.9928723573684692, "learning_rate": 0.0001996330073058681, "loss": 4.6761, "step": 114000 }, { "epoch": 0.212736675382961, "eval_loss": 4.887726783752441, "eval_runtime": 228.3892, "eval_samples_per_second": 11.419, "eval_steps_per_second": 11.419, "eval_tts_loss": 7.461525186266396, "step": 114000 }, { "epoch": 0.2128299809423395, "grad_norm": 0.9984357953071594, "learning_rate": 0.0001996326476378996, "loss": 4.7151, "step": 114050 }, { "epoch": 0.212923286501718, "grad_norm": 0.9879376888275146, "learning_rate": 0.000199632287794097, "loss": 4.5944, "step": 114100 }, { "epoch": 0.2130165920610965, "grad_norm": 0.9942005276679993, "learning_rate": 0.00019963192777446094, "loss": 4.676, "step": 114150 }, { "epoch": 0.21310989762047497, "grad_norm": 1.000113606452942, "learning_rate": 0.00019963156757899205, "loss": 4.5756, "step": 114200 }, { "epoch": 0.21320320317985347, "grad_norm": 1.1187829971313477, "learning_rate": 0.000199631207207691, "loss": 4.5309, "step": 114250 }, { "epoch": 0.21329650873923195, "grad_norm": 1.0895246267318726, "learning_rate": 0.0001996308466605584, "loss": 4.4908, "step": 114300 }, { "epoch": 0.21338981429861045, "grad_norm": 0.7784030437469482, "learning_rate": 0.00019963048593759486, "loss": 4.4745, "step": 114350 }, { "epoch": 0.21348311985798893, "grad_norm": 1.0729563236236572, "learning_rate": 0.00019963012503880107, "loss": 4.5451, "step": 114400 }, { "epoch": 0.21357642541736743, "grad_norm": 1.1839163303375244, "learning_rate": 0.00019962976396417764, "loss": 4.7851, "step": 114450 }, { "epoch": 0.2136697309767459, "grad_norm": 1.1860408782958984, "learning_rate": 0.00019962940271372518, "loss": 4.6238, "step": 114500 }, { "epoch": 0.2137630365361244, "grad_norm": 0.9122569561004639, "learning_rate": 0.00019962904128744438, "loss": 4.6063, "step": 114550 }, { "epoch": 0.21385634209550292, "grad_norm": 0.8689755797386169, "learning_rate": 0.00019962867968533585, "loss": 4.5032, "step": 114600 }, { "epoch": 0.2139496476548814, "grad_norm": 0.8875397443771362, "learning_rate": 0.0001996283179074002, "loss": 4.7033, "step": 114650 }, { "epoch": 0.2140429532142599, "grad_norm": 0.7156913876533508, "learning_rate": 0.00019962795595363813, "loss": 4.8056, "step": 114700 }, { "epoch": 0.21413625877363837, "grad_norm": 1.1040927171707153, "learning_rate": 0.00019962759382405025, "loss": 4.6738, "step": 114750 }, { "epoch": 0.21422956433301688, "grad_norm": 0.9362927675247192, "learning_rate": 0.00019962723151863718, "loss": 4.8287, "step": 114800 }, { "epoch": 0.21432286989239535, "grad_norm": 0.8515648245811462, "learning_rate": 0.00019962686903739958, "loss": 4.4636, "step": 114850 }, { "epoch": 0.21441617545177385, "grad_norm": 1.0025912523269653, "learning_rate": 0.0001996265063803381, "loss": 4.7396, "step": 114900 }, { "epoch": 0.21450948101115236, "grad_norm": 0.7344082593917847, "learning_rate": 0.00019962614354745336, "loss": 4.5112, "step": 114950 }, { "epoch": 0.21460278657053083, "grad_norm": 1.0541155338287354, "learning_rate": 0.00019962578053874598, "loss": 4.6548, "step": 115000 }, { "epoch": 0.21469609212990934, "grad_norm": 0.9491979479789734, "learning_rate": 0.00019962541735421666, "loss": 4.7796, "step": 115050 }, { "epoch": 0.21478939768928781, "grad_norm": 0.8249531388282776, "learning_rate": 0.000199625053993866, "loss": 4.7001, "step": 115100 }, { "epoch": 0.21488270324866632, "grad_norm": 0.862076461315155, "learning_rate": 0.00019962469045769465, "loss": 4.4974, "step": 115150 }, { "epoch": 0.2149760088080448, "grad_norm": 1.1294341087341309, "learning_rate": 0.00019962432674570325, "loss": 4.6363, "step": 115200 }, { "epoch": 0.2150693143674233, "grad_norm": 0.7202677130699158, "learning_rate": 0.0001996239628578924, "loss": 4.6216, "step": 115250 }, { "epoch": 0.2151626199268018, "grad_norm": 0.8325316905975342, "learning_rate": 0.00019962359879426285, "loss": 4.6454, "step": 115300 }, { "epoch": 0.21525592548618028, "grad_norm": 1.1335790157318115, "learning_rate": 0.00019962323455481514, "loss": 4.6516, "step": 115350 }, { "epoch": 0.21534923104555878, "grad_norm": 0.8559627532958984, "learning_rate": 0.00019962287013954998, "loss": 4.7461, "step": 115400 }, { "epoch": 0.21544253660493726, "grad_norm": 1.0715285539627075, "learning_rate": 0.00019962250554846796, "loss": 4.7282, "step": 115450 }, { "epoch": 0.21553584216431576, "grad_norm": 1.0772089958190918, "learning_rate": 0.00019962214078156975, "loss": 4.808, "step": 115500 }, { "epoch": 0.21562914772369424, "grad_norm": 0.8694863319396973, "learning_rate": 0.000199621775838856, "loss": 4.6563, "step": 115550 }, { "epoch": 0.21572245328307274, "grad_norm": 0.7618568539619446, "learning_rate": 0.00019962141072032733, "loss": 4.5163, "step": 115600 }, { "epoch": 0.21581575884245124, "grad_norm": 0.7407212257385254, "learning_rate": 0.0001996210454259844, "loss": 4.7637, "step": 115650 }, { "epoch": 0.21590906440182972, "grad_norm": 0.86847984790802, "learning_rate": 0.00019962067995582788, "loss": 4.6896, "step": 115700 }, { "epoch": 0.21600236996120822, "grad_norm": 1.2252628803253174, "learning_rate": 0.00019962031430985838, "loss": 4.5515, "step": 115750 }, { "epoch": 0.2160956755205867, "grad_norm": 0.8714708089828491, "learning_rate": 0.00019961994848807652, "loss": 4.699, "step": 115800 }, { "epoch": 0.2161889810799652, "grad_norm": 0.9541438221931458, "learning_rate": 0.000199619582490483, "loss": 4.7368, "step": 115850 }, { "epoch": 0.21628228663934368, "grad_norm": 1.139961838722229, "learning_rate": 0.00019961921631707845, "loss": 4.7307, "step": 115900 }, { "epoch": 0.21637559219872218, "grad_norm": 1.154788613319397, "learning_rate": 0.00019961884996786352, "loss": 4.7677, "step": 115950 }, { "epoch": 0.21646889775810066, "grad_norm": 0.9261074662208557, "learning_rate": 0.00019961848344283883, "loss": 4.5874, "step": 116000 }, { "epoch": 0.21656220331747916, "grad_norm": 1.3883506059646606, "learning_rate": 0.00019961811674200505, "loss": 4.6135, "step": 116050 }, { "epoch": 0.21665550887685767, "grad_norm": 1.012757658958435, "learning_rate": 0.0001996177498653628, "loss": 4.7373, "step": 116100 }, { "epoch": 0.21674881443623614, "grad_norm": 1.0129855871200562, "learning_rate": 0.0001996173828129128, "loss": 4.7325, "step": 116150 }, { "epoch": 0.21684211999561465, "grad_norm": 1.0152562856674194, "learning_rate": 0.00019961701558465563, "loss": 4.5226, "step": 116200 }, { "epoch": 0.21693542555499312, "grad_norm": 1.094165563583374, "learning_rate": 0.00019961664818059194, "loss": 4.7992, "step": 116250 }, { "epoch": 0.21702873111437163, "grad_norm": 1.0749988555908203, "learning_rate": 0.0001996162806007224, "loss": 4.7296, "step": 116300 }, { "epoch": 0.2171220366737501, "grad_norm": 0.8733508586883545, "learning_rate": 0.00019961591284504767, "loss": 4.8671, "step": 116350 }, { "epoch": 0.2172153422331286, "grad_norm": 0.7462573647499084, "learning_rate": 0.00019961554491356836, "loss": 4.8757, "step": 116400 }, { "epoch": 0.2173086477925071, "grad_norm": 0.8342740535736084, "learning_rate": 0.00019961517680628515, "loss": 4.5841, "step": 116450 }, { "epoch": 0.21740195335188559, "grad_norm": 1.1010075807571411, "learning_rate": 0.00019961480852319867, "loss": 4.5726, "step": 116500 }, { "epoch": 0.2174952589112641, "grad_norm": 0.8100841045379639, "learning_rate": 0.0001996144400643096, "loss": 4.6927, "step": 116550 }, { "epoch": 0.21758856447064256, "grad_norm": 1.0698412656784058, "learning_rate": 0.00019961407142961854, "loss": 4.8108, "step": 116600 }, { "epoch": 0.21768187003002107, "grad_norm": 0.9264890551567078, "learning_rate": 0.00019961370261912623, "loss": 4.555, "step": 116650 }, { "epoch": 0.21777517558939954, "grad_norm": 0.8985602855682373, "learning_rate": 0.00019961333363283318, "loss": 4.7173, "step": 116700 }, { "epoch": 0.21786848114877805, "grad_norm": 0.9962376356124878, "learning_rate": 0.0001996129644707402, "loss": 4.5897, "step": 116750 }, { "epoch": 0.21796178670815655, "grad_norm": 1.108918309211731, "learning_rate": 0.00019961259513284781, "loss": 4.7698, "step": 116800 }, { "epoch": 0.21805509226753503, "grad_norm": 1.3311587572097778, "learning_rate": 0.00019961222561915677, "loss": 4.6247, "step": 116850 }, { "epoch": 0.21814839782691353, "grad_norm": 0.602816104888916, "learning_rate": 0.00019961185592966762, "loss": 4.9767, "step": 116900 }, { "epoch": 0.218241703386292, "grad_norm": 1.00581693649292, "learning_rate": 0.0001996114860643811, "loss": 4.5662, "step": 116950 }, { "epoch": 0.2183350089456705, "grad_norm": 0.9467529654502869, "learning_rate": 0.00019961111602329786, "loss": 4.6823, "step": 117000 }, { "epoch": 0.2183350089456705, "eval_loss": 4.885964393615723, "eval_runtime": 228.2045, "eval_samples_per_second": 11.428, "eval_steps_per_second": 11.428, "eval_tts_loss": 7.423409835071205, "step": 117000 }, { "epoch": 0.218428314505049, "grad_norm": 0.9018577933311462, "learning_rate": 0.00019961074580641846, "loss": 4.7069, "step": 117050 }, { "epoch": 0.2185216200644275, "grad_norm": 1.015038251876831, "learning_rate": 0.0001996103754137437, "loss": 4.772, "step": 117100 }, { "epoch": 0.21861492562380597, "grad_norm": 1.082078456878662, "learning_rate": 0.00019961000484527413, "loss": 4.5817, "step": 117150 }, { "epoch": 0.21870823118318447, "grad_norm": 1.0951112508773804, "learning_rate": 0.0001996096341010104, "loss": 4.76, "step": 117200 }, { "epoch": 0.21880153674256297, "grad_norm": 1.023581862449646, "learning_rate": 0.00019960926318095322, "loss": 4.547, "step": 117250 }, { "epoch": 0.21889484230194145, "grad_norm": 0.8084960579872131, "learning_rate": 0.00019960889208510318, "loss": 4.8218, "step": 117300 }, { "epoch": 0.21898814786131995, "grad_norm": 1.2182049751281738, "learning_rate": 0.000199608520813461, "loss": 4.6656, "step": 117350 }, { "epoch": 0.21908145342069843, "grad_norm": 0.9085681438446045, "learning_rate": 0.00019960814936602732, "loss": 4.8121, "step": 117400 }, { "epoch": 0.21917475898007693, "grad_norm": 0.9307880997657776, "learning_rate": 0.00019960777774280278, "loss": 4.6564, "step": 117450 }, { "epoch": 0.2192680645394554, "grad_norm": 0.8649754524230957, "learning_rate": 0.00019960740594378803, "loss": 4.7026, "step": 117500 }, { "epoch": 0.2193613700988339, "grad_norm": 0.7863887548446655, "learning_rate": 0.00019960703396898371, "loss": 4.5972, "step": 117550 }, { "epoch": 0.21945467565821242, "grad_norm": 0.9783292412757874, "learning_rate": 0.00019960666181839055, "loss": 4.6181, "step": 117600 }, { "epoch": 0.2195479812175909, "grad_norm": 1.023344874382019, "learning_rate": 0.00019960628949200913, "loss": 4.5983, "step": 117650 }, { "epoch": 0.2196412867769694, "grad_norm": 0.9733097553253174, "learning_rate": 0.0001996059169898401, "loss": 4.8129, "step": 117700 }, { "epoch": 0.21973459233634787, "grad_norm": 0.9498526453971863, "learning_rate": 0.00019960554431188422, "loss": 4.5947, "step": 117750 }, { "epoch": 0.21982789789572638, "grad_norm": 1.1250981092453003, "learning_rate": 0.00019960517145814203, "loss": 4.4665, "step": 117800 }, { "epoch": 0.21992120345510485, "grad_norm": 0.7923647165298462, "learning_rate": 0.00019960479842861428, "loss": 4.5954, "step": 117850 }, { "epoch": 0.22001450901448336, "grad_norm": 1.0662282705307007, "learning_rate": 0.00019960442522330154, "loss": 4.755, "step": 117900 }, { "epoch": 0.22010781457386186, "grad_norm": 1.0039044618606567, "learning_rate": 0.00019960405184220456, "loss": 4.7219, "step": 117950 }, { "epoch": 0.22020112013324034, "grad_norm": 1.210990309715271, "learning_rate": 0.00019960367828532392, "loss": 4.6545, "step": 118000 }, { "epoch": 0.22029442569261884, "grad_norm": 0.9507876634597778, "learning_rate": 0.00019960330455266035, "loss": 4.7404, "step": 118050 }, { "epoch": 0.22038773125199732, "grad_norm": 1.1549371480941772, "learning_rate": 0.00019960293064421445, "loss": 4.8518, "step": 118100 }, { "epoch": 0.22048103681137582, "grad_norm": 1.0183074474334717, "learning_rate": 0.0001996025565599869, "loss": 4.7184, "step": 118150 }, { "epoch": 0.2205743423707543, "grad_norm": 1.1634994745254517, "learning_rate": 0.00019960218229997837, "loss": 4.6093, "step": 118200 }, { "epoch": 0.2206676479301328, "grad_norm": 0.969548225402832, "learning_rate": 0.00019960180786418951, "loss": 4.8916, "step": 118250 }, { "epoch": 0.2207609534895113, "grad_norm": 1.2633224725723267, "learning_rate": 0.000199601433252621, "loss": 4.6768, "step": 118300 }, { "epoch": 0.22085425904888978, "grad_norm": 0.9347729682922363, "learning_rate": 0.00019960105846527346, "loss": 4.6532, "step": 118350 }, { "epoch": 0.22094756460826828, "grad_norm": 1.0598022937774658, "learning_rate": 0.0001996006835021476, "loss": 4.6961, "step": 118400 }, { "epoch": 0.22104087016764676, "grad_norm": 1.2343388795852661, "learning_rate": 0.00019960030836324404, "loss": 4.6699, "step": 118450 }, { "epoch": 0.22113417572702526, "grad_norm": 0.871600866317749, "learning_rate": 0.00019959993304856347, "loss": 4.6907, "step": 118500 }, { "epoch": 0.22122748128640374, "grad_norm": 0.9813231825828552, "learning_rate": 0.00019959955755810652, "loss": 4.6616, "step": 118550 }, { "epoch": 0.22132078684578224, "grad_norm": 1.049269199371338, "learning_rate": 0.00019959918189187392, "loss": 4.6005, "step": 118600 }, { "epoch": 0.22141409240516072, "grad_norm": 0.7890821695327759, "learning_rate": 0.00019959880604986628, "loss": 4.685, "step": 118650 }, { "epoch": 0.22150739796453922, "grad_norm": 0.8772531151771545, "learning_rate": 0.00019959843003208424, "loss": 4.7188, "step": 118700 }, { "epoch": 0.22160070352391772, "grad_norm": 1.0186238288879395, "learning_rate": 0.00019959805383852852, "loss": 4.8515, "step": 118750 }, { "epoch": 0.2216940090832962, "grad_norm": 1.081235647201538, "learning_rate": 0.00019959767746919975, "loss": 4.7645, "step": 118800 }, { "epoch": 0.2217873146426747, "grad_norm": 0.9641054272651672, "learning_rate": 0.0001995973009240986, "loss": 4.6056, "step": 118850 }, { "epoch": 0.22188062020205318, "grad_norm": 0.8466959595680237, "learning_rate": 0.00019959692420322577, "loss": 4.7974, "step": 118900 }, { "epoch": 0.22197392576143168, "grad_norm": 1.1167409420013428, "learning_rate": 0.00019959654730658185, "loss": 4.8381, "step": 118950 }, { "epoch": 0.22206723132081016, "grad_norm": 0.8874496221542358, "learning_rate": 0.0001995961702341676, "loss": 4.5208, "step": 119000 }, { "epoch": 0.22216053688018866, "grad_norm": 1.0184682607650757, "learning_rate": 0.0001995957929859836, "loss": 4.766, "step": 119050 }, { "epoch": 0.22225384243956717, "grad_norm": 0.8084706664085388, "learning_rate": 0.00019959541556203054, "loss": 4.6748, "step": 119100 }, { "epoch": 0.22234714799894564, "grad_norm": 0.9491915106773376, "learning_rate": 0.0001995950379623091, "loss": 4.7225, "step": 119150 }, { "epoch": 0.22244045355832415, "grad_norm": 0.9508479833602905, "learning_rate": 0.00019959466018682, "loss": 4.785, "step": 119200 }, { "epoch": 0.22253375911770262, "grad_norm": 0.8194633722305298, "learning_rate": 0.0001995942822355638, "loss": 4.7103, "step": 119250 }, { "epoch": 0.22262706467708113, "grad_norm": 0.9935794472694397, "learning_rate": 0.0001995939041085412, "loss": 4.7287, "step": 119300 }, { "epoch": 0.2227203702364596, "grad_norm": 0.7997922897338867, "learning_rate": 0.0001995935258057529, "loss": 4.621, "step": 119350 }, { "epoch": 0.2228136757958381, "grad_norm": 0.9721159338951111, "learning_rate": 0.00019959314732719956, "loss": 4.7279, "step": 119400 }, { "epoch": 0.2229069813552166, "grad_norm": 1.006213665008545, "learning_rate": 0.0001995927686728818, "loss": 4.8168, "step": 119450 }, { "epoch": 0.22300028691459509, "grad_norm": 0.976470947265625, "learning_rate": 0.00019959238984280038, "loss": 4.6645, "step": 119500 }, { "epoch": 0.2230935924739736, "grad_norm": 0.9182271957397461, "learning_rate": 0.00019959201083695588, "loss": 4.6798, "step": 119550 }, { "epoch": 0.22318689803335207, "grad_norm": 0.873684287071228, "learning_rate": 0.00019959163165534903, "loss": 4.4998, "step": 119600 }, { "epoch": 0.22328020359273057, "grad_norm": 1.1776008605957031, "learning_rate": 0.00019959125229798046, "loss": 4.7635, "step": 119650 }, { "epoch": 0.22337350915210905, "grad_norm": 1.159077763557434, "learning_rate": 0.00019959087276485087, "loss": 4.6824, "step": 119700 }, { "epoch": 0.22346681471148755, "grad_norm": 1.2135952711105347, "learning_rate": 0.0001995904930559609, "loss": 4.7218, "step": 119750 }, { "epoch": 0.22356012027086603, "grad_norm": 0.8527218699455261, "learning_rate": 0.0001995901131713112, "loss": 4.7581, "step": 119800 }, { "epoch": 0.22365342583024453, "grad_norm": 1.1286325454711914, "learning_rate": 0.0001995897331109025, "loss": 4.6019, "step": 119850 }, { "epoch": 0.22374673138962303, "grad_norm": 0.7209294438362122, "learning_rate": 0.00019958935287473547, "loss": 4.4614, "step": 119900 }, { "epoch": 0.2238400369490015, "grad_norm": 1.012035608291626, "learning_rate": 0.00019958897246281074, "loss": 4.743, "step": 119950 }, { "epoch": 0.22393334250838, "grad_norm": 0.9427865147590637, "learning_rate": 0.000199588591875129, "loss": 4.5508, "step": 120000 }, { "epoch": 0.22393334250838, "eval_loss": 4.883612632751465, "eval_runtime": 229.5565, "eval_samples_per_second": 11.361, "eval_steps_per_second": 11.361, "eval_tts_loss": 7.450549049201519, "step": 120000 }, { "epoch": 0.2240266480677585, "grad_norm": 1.0907127857208252, "learning_rate": 0.0001995882111116909, "loss": 4.6219, "step": 120050 }, { "epoch": 0.224119953627137, "grad_norm": 1.2542510032653809, "learning_rate": 0.00019958783017249714, "loss": 4.5626, "step": 120100 }, { "epoch": 0.22421325918651547, "grad_norm": 0.9580556750297546, "learning_rate": 0.00019958744905754837, "loss": 4.7883, "step": 120150 }, { "epoch": 0.22430656474589397, "grad_norm": 1.1819099187850952, "learning_rate": 0.00019958706776684532, "loss": 4.7903, "step": 120200 }, { "epoch": 0.22439987030527248, "grad_norm": 0.9004278182983398, "learning_rate": 0.00019958668630038856, "loss": 4.6342, "step": 120250 }, { "epoch": 0.22449317586465095, "grad_norm": 0.898798942565918, "learning_rate": 0.00019958630465817886, "loss": 4.9032, "step": 120300 }, { "epoch": 0.22458648142402945, "grad_norm": 1.324995994567871, "learning_rate": 0.00019958592284021685, "loss": 4.7004, "step": 120350 }, { "epoch": 0.22467978698340793, "grad_norm": 0.9683493971824646, "learning_rate": 0.00019958554084650317, "loss": 4.7933, "step": 120400 }, { "epoch": 0.22477309254278643, "grad_norm": 0.9748978614807129, "learning_rate": 0.00019958515867703858, "loss": 4.8761, "step": 120450 }, { "epoch": 0.2248663981021649, "grad_norm": 1.192099690437317, "learning_rate": 0.00019958477633182372, "loss": 4.6939, "step": 120500 }, { "epoch": 0.22495970366154341, "grad_norm": 1.2291393280029297, "learning_rate": 0.0001995843938108592, "loss": 4.743, "step": 120550 }, { "epoch": 0.22505300922092192, "grad_norm": 1.2293897867202759, "learning_rate": 0.00019958401111414578, "loss": 4.8312, "step": 120600 }, { "epoch": 0.2251463147803004, "grad_norm": 1.0013765096664429, "learning_rate": 0.0001995836282416841, "loss": 4.71, "step": 120650 }, { "epoch": 0.2252396203396789, "grad_norm": 1.1996088027954102, "learning_rate": 0.00019958324519347485, "loss": 4.8127, "step": 120700 }, { "epoch": 0.22533292589905737, "grad_norm": 0.9375223517417908, "learning_rate": 0.00019958286196951867, "loss": 4.6523, "step": 120750 }, { "epoch": 0.22542623145843588, "grad_norm": 0.7946412563323975, "learning_rate": 0.00019958247856981628, "loss": 4.5345, "step": 120800 }, { "epoch": 0.22551953701781435, "grad_norm": 0.9169266223907471, "learning_rate": 0.00019958209499436832, "loss": 4.5954, "step": 120850 }, { "epoch": 0.22561284257719286, "grad_norm": 1.0124138593673706, "learning_rate": 0.0001995817112431755, "loss": 4.7285, "step": 120900 }, { "epoch": 0.22570614813657136, "grad_norm": 0.8469855189323425, "learning_rate": 0.00019958132731623843, "loss": 4.5862, "step": 120950 }, { "epoch": 0.22579945369594984, "grad_norm": 0.9922230839729309, "learning_rate": 0.00019958094321355791, "loss": 4.7852, "step": 121000 }, { "epoch": 0.22589275925532834, "grad_norm": 1.1002434492111206, "learning_rate": 0.00019958055893513451, "loss": 4.7454, "step": 121050 }, { "epoch": 0.22598606481470682, "grad_norm": 1.2257192134857178, "learning_rate": 0.00019958017448096896, "loss": 4.7457, "step": 121100 }, { "epoch": 0.22607937037408532, "grad_norm": 1.063035249710083, "learning_rate": 0.0001995797898510619, "loss": 4.5511, "step": 121150 }, { "epoch": 0.2261726759334638, "grad_norm": 0.769403874874115, "learning_rate": 0.00019957940504541406, "loss": 4.6497, "step": 121200 }, { "epoch": 0.2262659814928423, "grad_norm": 0.9175217151641846, "learning_rate": 0.0001995790200640261, "loss": 4.6908, "step": 121250 }, { "epoch": 0.22635928705222078, "grad_norm": 1.0887384414672852, "learning_rate": 0.00019957863490689865, "loss": 4.9673, "step": 121300 }, { "epoch": 0.22645259261159928, "grad_norm": 1.0084506273269653, "learning_rate": 0.00019957824957403245, "loss": 4.6811, "step": 121350 }, { "epoch": 0.22654589817097778, "grad_norm": 1.035234808921814, "learning_rate": 0.00019957786406542817, "loss": 4.5367, "step": 121400 }, { "epoch": 0.22663920373035626, "grad_norm": 1.082025408744812, "learning_rate": 0.0001995774783810865, "loss": 4.8276, "step": 121450 }, { "epoch": 0.22673250928973476, "grad_norm": 0.8627831339836121, "learning_rate": 0.00019957709252100807, "loss": 4.5622, "step": 121500 }, { "epoch": 0.22682581484911324, "grad_norm": 0.6274713277816772, "learning_rate": 0.0001995767064851936, "loss": 4.6064, "step": 121550 }, { "epoch": 0.22691912040849174, "grad_norm": 0.6322067975997925, "learning_rate": 0.00019957632027364376, "loss": 4.6496, "step": 121600 }, { "epoch": 0.22701242596787022, "grad_norm": 0.7151324152946472, "learning_rate": 0.00019957593388635923, "loss": 4.4901, "step": 121650 }, { "epoch": 0.22710573152724872, "grad_norm": 0.8204741477966309, "learning_rate": 0.00019957554732334071, "loss": 4.8227, "step": 121700 }, { "epoch": 0.22719903708662723, "grad_norm": 1.0234726667404175, "learning_rate": 0.0001995751605845889, "loss": 4.6125, "step": 121750 }, { "epoch": 0.2272923426460057, "grad_norm": 1.1819202899932861, "learning_rate": 0.0001995747736701044, "loss": 4.7185, "step": 121800 }, { "epoch": 0.2273856482053842, "grad_norm": 0.9932510256767273, "learning_rate": 0.00019957438657988797, "loss": 4.9552, "step": 121850 }, { "epoch": 0.22747895376476268, "grad_norm": 1.0149807929992676, "learning_rate": 0.00019957399931394028, "loss": 4.7491, "step": 121900 }, { "epoch": 0.22757225932414119, "grad_norm": 0.8650890588760376, "learning_rate": 0.00019957361187226199, "loss": 4.7261, "step": 121950 }, { "epoch": 0.22766556488351966, "grad_norm": 0.8572404980659485, "learning_rate": 0.00019957322425485376, "loss": 4.6074, "step": 122000 }, { "epoch": 0.22775887044289816, "grad_norm": 0.8673201203346252, "learning_rate": 0.00019957283646171634, "loss": 4.5425, "step": 122050 }, { "epoch": 0.22785217600227667, "grad_norm": 1.1532055139541626, "learning_rate": 0.0001995724484928504, "loss": 4.6619, "step": 122100 }, { "epoch": 0.22794548156165514, "grad_norm": 1.0270742177963257, "learning_rate": 0.00019957206034825658, "loss": 4.7036, "step": 122150 }, { "epoch": 0.22803878712103365, "grad_norm": 1.0229623317718506, "learning_rate": 0.0001995716720279356, "loss": 4.5638, "step": 122200 }, { "epoch": 0.22813209268041212, "grad_norm": 1.0269454717636108, "learning_rate": 0.00019957128353188816, "loss": 4.4661, "step": 122250 }, { "epoch": 0.22822539823979063, "grad_norm": 1.051167368888855, "learning_rate": 0.0001995708948601149, "loss": 4.5614, "step": 122300 }, { "epoch": 0.2283187037991691, "grad_norm": 0.9413858652114868, "learning_rate": 0.00019957050601261656, "loss": 4.5145, "step": 122350 }, { "epoch": 0.2284120093585476, "grad_norm": 0.7322555184364319, "learning_rate": 0.00019957011698939377, "loss": 4.7114, "step": 122400 }, { "epoch": 0.2285053149179261, "grad_norm": 0.9706824421882629, "learning_rate": 0.00019956972779044727, "loss": 4.5615, "step": 122450 }, { "epoch": 0.2285986204773046, "grad_norm": 1.0096715688705444, "learning_rate": 0.0001995693384157777, "loss": 4.7756, "step": 122500 }, { "epoch": 0.2286919260366831, "grad_norm": 0.901025652885437, "learning_rate": 0.00019956894886538574, "loss": 4.7176, "step": 122550 }, { "epoch": 0.22878523159606157, "grad_norm": 0.9558797478675842, "learning_rate": 0.00019956855913927214, "loss": 4.8155, "step": 122600 }, { "epoch": 0.22887853715544007, "grad_norm": 0.6272359490394592, "learning_rate": 0.00019956816923743755, "loss": 4.7247, "step": 122650 }, { "epoch": 0.22897184271481855, "grad_norm": 1.197523832321167, "learning_rate": 0.00019956777915988269, "loss": 4.6914, "step": 122700 }, { "epoch": 0.22906514827419705, "grad_norm": 0.9652473330497742, "learning_rate": 0.00019956738890660817, "loss": 4.6038, "step": 122750 }, { "epoch": 0.22915845383357553, "grad_norm": 1.0007468461990356, "learning_rate": 0.00019956699847761472, "loss": 4.6983, "step": 122800 }, { "epoch": 0.22925175939295403, "grad_norm": 0.9335245490074158, "learning_rate": 0.0001995666078729031, "loss": 4.631, "step": 122850 }, { "epoch": 0.22934506495233253, "grad_norm": 0.9081475734710693, "learning_rate": 0.00019956621709247386, "loss": 4.8258, "step": 122900 }, { "epoch": 0.229438370511711, "grad_norm": 0.7586884498596191, "learning_rate": 0.00019956582613632785, "loss": 4.5945, "step": 122950 }, { "epoch": 0.2295316760710895, "grad_norm": 0.909521222114563, "learning_rate": 0.0001995654350044656, "loss": 4.6523, "step": 123000 }, { "epoch": 0.2295316760710895, "eval_loss": 4.8781867027282715, "eval_runtime": 230.72, "eval_samples_per_second": 11.304, "eval_steps_per_second": 11.304, "eval_tts_loss": 7.44082480463045, "step": 123000 }, { "epoch": 0.229624981630468, "grad_norm": 1.208822250366211, "learning_rate": 0.00019956504369688792, "loss": 4.9531, "step": 123050 }, { "epoch": 0.2297182871898465, "grad_norm": 1.0203379392623901, "learning_rate": 0.00019956465221359544, "loss": 4.6531, "step": 123100 }, { "epoch": 0.22981159274922497, "grad_norm": 0.9764498472213745, "learning_rate": 0.00019956426055458887, "loss": 4.5436, "step": 123150 }, { "epoch": 0.22990489830860347, "grad_norm": 1.1145259141921997, "learning_rate": 0.00019956386871986892, "loss": 4.5113, "step": 123200 }, { "epoch": 0.22999820386798198, "grad_norm": 0.9552769064903259, "learning_rate": 0.00019956347670943624, "loss": 4.7847, "step": 123250 }, { "epoch": 0.23009150942736045, "grad_norm": 1.0565723180770874, "learning_rate": 0.00019956308452329154, "loss": 4.8575, "step": 123300 }, { "epoch": 0.23018481498673896, "grad_norm": 1.123518705368042, "learning_rate": 0.0001995626921614355, "loss": 4.6688, "step": 123350 }, { "epoch": 0.23027812054611743, "grad_norm": 0.911044716835022, "learning_rate": 0.00019956229962386885, "loss": 4.7472, "step": 123400 }, { "epoch": 0.23037142610549594, "grad_norm": 1.0663899183273315, "learning_rate": 0.00019956190691059223, "loss": 4.7779, "step": 123450 }, { "epoch": 0.2304647316648744, "grad_norm": 0.9321340918540955, "learning_rate": 0.0001995615140216064, "loss": 4.6964, "step": 123500 }, { "epoch": 0.23055803722425292, "grad_norm": 0.8135243058204651, "learning_rate": 0.00019956112095691203, "loss": 4.6287, "step": 123550 }, { "epoch": 0.23065134278363142, "grad_norm": 1.0988374948501587, "learning_rate": 0.00019956072771650977, "loss": 4.8658, "step": 123600 }, { "epoch": 0.2307446483430099, "grad_norm": 0.8417083024978638, "learning_rate": 0.00019956033430040034, "loss": 4.7001, "step": 123650 }, { "epoch": 0.2308379539023884, "grad_norm": 1.0932660102844238, "learning_rate": 0.00019955994070858447, "loss": 4.5109, "step": 123700 }, { "epoch": 0.23093125946176687, "grad_norm": 1.016459584236145, "learning_rate": 0.00019955954694106278, "loss": 4.7203, "step": 123750 }, { "epoch": 0.23102456502114538, "grad_norm": 1.1212247610092163, "learning_rate": 0.000199559152997836, "loss": 4.5223, "step": 123800 }, { "epoch": 0.23111787058052385, "grad_norm": 0.849577784538269, "learning_rate": 0.00019955875887890488, "loss": 4.6992, "step": 123850 }, { "epoch": 0.23121117613990236, "grad_norm": 0.7211328744888306, "learning_rate": 0.00019955836458427007, "loss": 4.554, "step": 123900 }, { "epoch": 0.23130448169928083, "grad_norm": 0.7936710715293884, "learning_rate": 0.00019955797011393222, "loss": 4.7927, "step": 123950 }, { "epoch": 0.23139778725865934, "grad_norm": 0.9986937642097473, "learning_rate": 0.0001995575754678921, "loss": 4.4808, "step": 124000 }, { "epoch": 0.23149109281803784, "grad_norm": 0.8931700587272644, "learning_rate": 0.00019955718064615036, "loss": 4.7492, "step": 124050 }, { "epoch": 0.23158439837741632, "grad_norm": 1.1811097860336304, "learning_rate": 0.00019955678564870772, "loss": 4.5177, "step": 124100 }, { "epoch": 0.23167770393679482, "grad_norm": 0.7191671133041382, "learning_rate": 0.0001995563904755649, "loss": 4.8179, "step": 124150 }, { "epoch": 0.2317710094961733, "grad_norm": 1.1839138269424438, "learning_rate": 0.00019955599512672253, "loss": 4.6699, "step": 124200 }, { "epoch": 0.2318643150555518, "grad_norm": 0.9712040424346924, "learning_rate": 0.00019955559960218135, "loss": 4.6827, "step": 124250 }, { "epoch": 0.23195762061493028, "grad_norm": 1.1487324237823486, "learning_rate": 0.00019955520390194207, "loss": 4.5614, "step": 124300 }, { "epoch": 0.23205092617430878, "grad_norm": 1.104325532913208, "learning_rate": 0.00019955480802600538, "loss": 4.6895, "step": 124350 }, { "epoch": 0.23214423173368728, "grad_norm": 1.0275545120239258, "learning_rate": 0.00019955441197437191, "loss": 4.6102, "step": 124400 }, { "epoch": 0.23223753729306576, "grad_norm": 1.1445039510726929, "learning_rate": 0.00019955401574704248, "loss": 4.6854, "step": 124450 }, { "epoch": 0.23233084285244426, "grad_norm": 0.9036419987678528, "learning_rate": 0.00019955361934401774, "loss": 4.6985, "step": 124500 }, { "epoch": 0.23242414841182274, "grad_norm": 0.825995683670044, "learning_rate": 0.00019955322276529836, "loss": 4.8375, "step": 124550 }, { "epoch": 0.23251745397120124, "grad_norm": 1.3338888883590698, "learning_rate": 0.00019955282601088502, "loss": 4.851, "step": 124600 }, { "epoch": 0.23261075953057972, "grad_norm": 1.4017682075500488, "learning_rate": 0.0001995524290807785, "loss": 4.7548, "step": 124650 }, { "epoch": 0.23270406508995822, "grad_norm": 0.6777539849281311, "learning_rate": 0.00019955203197497947, "loss": 4.6612, "step": 124700 }, { "epoch": 0.23279737064933673, "grad_norm": 0.94838947057724, "learning_rate": 0.0001995516346934886, "loss": 4.8645, "step": 124750 }, { "epoch": 0.2328906762087152, "grad_norm": 1.0013059377670288, "learning_rate": 0.0001995512372363066, "loss": 4.6178, "step": 124800 }, { "epoch": 0.2329839817680937, "grad_norm": 0.8147088885307312, "learning_rate": 0.0001995508396034342, "loss": 4.4805, "step": 124850 }, { "epoch": 0.23307728732747218, "grad_norm": 0.6364634037017822, "learning_rate": 0.0001995504417948721, "loss": 4.614, "step": 124900 }, { "epoch": 0.23317059288685069, "grad_norm": 1.0096015930175781, "learning_rate": 0.00019955004381062096, "loss": 4.8126, "step": 124950 }, { "epoch": 0.23326389844622916, "grad_norm": 1.0549802780151367, "learning_rate": 0.00019954964565068147, "loss": 4.5155, "step": 125000 }, { "epoch": 0.23335720400560767, "grad_norm": 1.0715384483337402, "learning_rate": 0.00019954924731505443, "loss": 4.8361, "step": 125050 }, { "epoch": 0.23345050956498617, "grad_norm": 1.089919090270996, "learning_rate": 0.00019954884880374048, "loss": 4.4423, "step": 125100 }, { "epoch": 0.23354381512436465, "grad_norm": 1.1438995599746704, "learning_rate": 0.00019954845011674032, "loss": 4.5797, "step": 125150 }, { "epoch": 0.23363712068374315, "grad_norm": 0.8875765800476074, "learning_rate": 0.00019954805125405468, "loss": 4.7633, "step": 125200 }, { "epoch": 0.23373042624312163, "grad_norm": 1.0435429811477661, "learning_rate": 0.00019954765221568424, "loss": 4.5587, "step": 125250 }, { "epoch": 0.23382373180250013, "grad_norm": 0.8705243468284607, "learning_rate": 0.00019954725300162966, "loss": 4.7671, "step": 125300 }, { "epoch": 0.2339170373618786, "grad_norm": 1.100628137588501, "learning_rate": 0.00019954685361189177, "loss": 4.5598, "step": 125350 }, { "epoch": 0.2340103429212571, "grad_norm": 0.7905378341674805, "learning_rate": 0.00019954645404647115, "loss": 4.6633, "step": 125400 }, { "epoch": 0.23410364848063558, "grad_norm": 1.0382624864578247, "learning_rate": 0.00019954605430536857, "loss": 4.8594, "step": 125450 }, { "epoch": 0.2341969540400141, "grad_norm": 1.0554344654083252, "learning_rate": 0.0001995456543885847, "loss": 4.8501, "step": 125500 }, { "epoch": 0.2342902595993926, "grad_norm": 0.6847901344299316, "learning_rate": 0.0001995452542961203, "loss": 4.686, "step": 125550 }, { "epoch": 0.23438356515877107, "grad_norm": 0.9976589679718018, "learning_rate": 0.000199544854027976, "loss": 4.7711, "step": 125600 }, { "epoch": 0.23447687071814957, "grad_norm": 1.0457299947738647, "learning_rate": 0.0001995444535841526, "loss": 4.5277, "step": 125650 }, { "epoch": 0.23457017627752805, "grad_norm": 0.7479730844497681, "learning_rate": 0.00019954405296465072, "loss": 4.7013, "step": 125700 }, { "epoch": 0.23466348183690655, "grad_norm": 1.141799807548523, "learning_rate": 0.0001995436521694711, "loss": 4.8921, "step": 125750 }, { "epoch": 0.23475678739628503, "grad_norm": 0.865399181842804, "learning_rate": 0.00019954325119861442, "loss": 4.6746, "step": 125800 }, { "epoch": 0.23485009295566353, "grad_norm": 1.0194836854934692, "learning_rate": 0.00019954285005208147, "loss": 4.8391, "step": 125850 }, { "epoch": 0.23494339851504203, "grad_norm": 0.7085402011871338, "learning_rate": 0.00019954244872987287, "loss": 4.7155, "step": 125900 }, { "epoch": 0.2350367040744205, "grad_norm": 0.9328494071960449, "learning_rate": 0.0001995420472319894, "loss": 4.7267, "step": 125950 }, { "epoch": 0.23513000963379901, "grad_norm": 0.9679514169692993, "learning_rate": 0.0001995416455584317, "loss": 4.6913, "step": 126000 }, { "epoch": 0.23513000963379901, "eval_loss": 4.869121551513672, "eval_runtime": 230.4744, "eval_samples_per_second": 11.316, "eval_steps_per_second": 11.316, "eval_tts_loss": 7.37138597899048, "step": 126000 }, { "epoch": 0.2352233151931775, "grad_norm": 1.0236223936080933, "learning_rate": 0.00019954124370920054, "loss": 4.6602, "step": 126050 }, { "epoch": 0.235316620752556, "grad_norm": 0.9021424651145935, "learning_rate": 0.0001995408416842966, "loss": 4.3545, "step": 126100 }, { "epoch": 0.23540992631193447, "grad_norm": 1.059244155883789, "learning_rate": 0.00019954043948372055, "loss": 4.7123, "step": 126150 }, { "epoch": 0.23550323187131297, "grad_norm": 0.8630461096763611, "learning_rate": 0.00019954003710747316, "loss": 4.7759, "step": 126200 }, { "epoch": 0.23559653743069148, "grad_norm": 1.192002534866333, "learning_rate": 0.00019953963455555514, "loss": 4.7526, "step": 126250 }, { "epoch": 0.23568984299006995, "grad_norm": 1.1485189199447632, "learning_rate": 0.00019953923182796713, "loss": 4.7111, "step": 126300 }, { "epoch": 0.23578314854944846, "grad_norm": 0.9921421408653259, "learning_rate": 0.00019953882892470993, "loss": 4.6672, "step": 126350 }, { "epoch": 0.23587645410882693, "grad_norm": 1.0147913694381714, "learning_rate": 0.0001995384258457842, "loss": 4.8141, "step": 126400 }, { "epoch": 0.23596975966820544, "grad_norm": 1.1433029174804688, "learning_rate": 0.00019953802259119066, "loss": 4.6498, "step": 126450 }, { "epoch": 0.2360630652275839, "grad_norm": 0.8153554201126099, "learning_rate": 0.00019953761916093003, "loss": 4.501, "step": 126500 }, { "epoch": 0.23615637078696242, "grad_norm": 0.8511623740196228, "learning_rate": 0.00019953721555500303, "loss": 4.6241, "step": 126550 }, { "epoch": 0.2362496763463409, "grad_norm": 0.9093624353408813, "learning_rate": 0.00019953681177341034, "loss": 4.8207, "step": 126600 }, { "epoch": 0.2363429819057194, "grad_norm": 1.1618552207946777, "learning_rate": 0.0001995364078161527, "loss": 4.7207, "step": 126650 }, { "epoch": 0.2364362874650979, "grad_norm": 1.0167028903961182, "learning_rate": 0.0001995360036832308, "loss": 4.7409, "step": 126700 }, { "epoch": 0.23652959302447638, "grad_norm": 1.2850828170776367, "learning_rate": 0.00019953559937464535, "loss": 4.8472, "step": 126750 }, { "epoch": 0.23662289858385488, "grad_norm": 0.9665905237197876, "learning_rate": 0.00019953519489039715, "loss": 4.658, "step": 126800 }, { "epoch": 0.23671620414323336, "grad_norm": 0.9830486178398132, "learning_rate": 0.0001995347902304868, "loss": 4.6645, "step": 126850 }, { "epoch": 0.23680950970261186, "grad_norm": 1.1448428630828857, "learning_rate": 0.00019953438539491505, "loss": 4.7673, "step": 126900 }, { "epoch": 0.23690281526199033, "grad_norm": 1.1206557750701904, "learning_rate": 0.00019953398038368266, "loss": 4.8454, "step": 126950 }, { "epoch": 0.23699612082136884, "grad_norm": 0.9602530002593994, "learning_rate": 0.00019953357519679026, "loss": 4.588, "step": 127000 }, { "epoch": 0.23708942638074734, "grad_norm": 0.816632091999054, "learning_rate": 0.00019953316983423863, "loss": 4.6631, "step": 127050 }, { "epoch": 0.23718273194012582, "grad_norm": 0.996062159538269, "learning_rate": 0.00019953276429602847, "loss": 4.6397, "step": 127100 }, { "epoch": 0.23727603749950432, "grad_norm": 1.116538405418396, "learning_rate": 0.0001995323585821605, "loss": 4.7699, "step": 127150 }, { "epoch": 0.2373693430588828, "grad_norm": 1.0206782817840576, "learning_rate": 0.00019953195269263544, "loss": 4.568, "step": 127200 }, { "epoch": 0.2374626486182613, "grad_norm": 0.8355110287666321, "learning_rate": 0.00019953154662745397, "loss": 4.5723, "step": 127250 }, { "epoch": 0.23755595417763978, "grad_norm": 0.7962259650230408, "learning_rate": 0.00019953114038661685, "loss": 4.6924, "step": 127300 }, { "epoch": 0.23764925973701828, "grad_norm": 1.0387051105499268, "learning_rate": 0.0001995307339701248, "loss": 4.6205, "step": 127350 }, { "epoch": 0.23774256529639679, "grad_norm": 0.9013943672180176, "learning_rate": 0.0001995303273779785, "loss": 4.705, "step": 127400 }, { "epoch": 0.23783587085577526, "grad_norm": 0.9052708148956299, "learning_rate": 0.00019952992061017866, "loss": 4.6817, "step": 127450 }, { "epoch": 0.23792917641515376, "grad_norm": 1.073463797569275, "learning_rate": 0.00019952951366672606, "loss": 4.8822, "step": 127500 }, { "epoch": 0.23802248197453224, "grad_norm": 0.8236823081970215, "learning_rate": 0.00019952910654762136, "loss": 4.7687, "step": 127550 }, { "epoch": 0.23811578753391074, "grad_norm": 1.148167371749878, "learning_rate": 0.0001995286992528653, "loss": 4.6313, "step": 127600 }, { "epoch": 0.23820909309328922, "grad_norm": 0.5891329050064087, "learning_rate": 0.00019952829178245863, "loss": 4.6894, "step": 127650 }, { "epoch": 0.23830239865266772, "grad_norm": 0.9606289267539978, "learning_rate": 0.00019952788413640196, "loss": 4.7802, "step": 127700 }, { "epoch": 0.23839570421204623, "grad_norm": 0.8213685750961304, "learning_rate": 0.00019952747631469617, "loss": 4.6354, "step": 127750 }, { "epoch": 0.2384890097714247, "grad_norm": 0.9802691340446472, "learning_rate": 0.00019952706831734187, "loss": 4.5251, "step": 127800 }, { "epoch": 0.2385823153308032, "grad_norm": 0.8711889386177063, "learning_rate": 0.00019952666014433977, "loss": 4.6781, "step": 127850 }, { "epoch": 0.23867562089018168, "grad_norm": 1.1614457368850708, "learning_rate": 0.0001995262517956907, "loss": 4.5533, "step": 127900 }, { "epoch": 0.2387689264495602, "grad_norm": 0.9102240204811096, "learning_rate": 0.00019952584327139524, "loss": 4.6598, "step": 127950 }, { "epoch": 0.23886223200893866, "grad_norm": 0.8396139144897461, "learning_rate": 0.00019952543457145422, "loss": 4.4948, "step": 128000 }, { "epoch": 0.23895553756831717, "grad_norm": 0.927024781703949, "learning_rate": 0.0001995250256958683, "loss": 4.5239, "step": 128050 }, { "epoch": 0.23904884312769564, "grad_norm": 1.0152112245559692, "learning_rate": 0.00019952461664463823, "loss": 4.6425, "step": 128100 }, { "epoch": 0.23914214868707415, "grad_norm": 1.4888454675674438, "learning_rate": 0.0001995242074177647, "loss": 4.7832, "step": 128150 }, { "epoch": 0.23923545424645265, "grad_norm": 1.1071109771728516, "learning_rate": 0.0001995237980152485, "loss": 4.6113, "step": 128200 }, { "epoch": 0.23932875980583113, "grad_norm": 0.9352720379829407, "learning_rate": 0.00019952338843709028, "loss": 4.6275, "step": 128250 }, { "epoch": 0.23942206536520963, "grad_norm": 0.9158457517623901, "learning_rate": 0.00019952297868329079, "loss": 4.6196, "step": 128300 }, { "epoch": 0.2395153709245881, "grad_norm": 0.46742984652519226, "learning_rate": 0.00019952256875385076, "loss": 4.6612, "step": 128350 }, { "epoch": 0.2396086764839666, "grad_norm": 1.0727510452270508, "learning_rate": 0.0001995221586487709, "loss": 4.7063, "step": 128400 }, { "epoch": 0.23970198204334509, "grad_norm": 0.9847069382667542, "learning_rate": 0.00019952174836805197, "loss": 4.9514, "step": 128450 }, { "epoch": 0.2397952876027236, "grad_norm": 1.0134080648422241, "learning_rate": 0.00019952133791169464, "loss": 4.4549, "step": 128500 }, { "epoch": 0.2398885931621021, "grad_norm": 0.7838501334190369, "learning_rate": 0.00019952092727969965, "loss": 4.348, "step": 128550 }, { "epoch": 0.23998189872148057, "grad_norm": 0.9687092900276184, "learning_rate": 0.00019952051647206776, "loss": 4.5199, "step": 128600 }, { "epoch": 0.24007520428085907, "grad_norm": 0.833861231803894, "learning_rate": 0.00019952010548879968, "loss": 4.812, "step": 128650 }, { "epoch": 0.24016850984023755, "grad_norm": 1.0054713487625122, "learning_rate": 0.00019951969432989607, "loss": 4.6566, "step": 128700 }, { "epoch": 0.24026181539961605, "grad_norm": 0.8803058862686157, "learning_rate": 0.00019951928299535775, "loss": 4.4073, "step": 128750 }, { "epoch": 0.24035512095899453, "grad_norm": 1.0025216341018677, "learning_rate": 0.0001995188714851854, "loss": 4.6065, "step": 128800 }, { "epoch": 0.24044842651837303, "grad_norm": 0.9606624245643616, "learning_rate": 0.0001995184597993798, "loss": 4.6934, "step": 128850 }, { "epoch": 0.24054173207775154, "grad_norm": 0.8765124678611755, "learning_rate": 0.00019951804793794155, "loss": 4.6806, "step": 128900 }, { "epoch": 0.24063503763713, "grad_norm": 0.9700292944908142, "learning_rate": 0.0001995176359008715, "loss": 4.7322, "step": 128950 }, { "epoch": 0.24072834319650852, "grad_norm": 0.7968749403953552, "learning_rate": 0.0001995172236881703, "loss": 4.648, "step": 129000 }, { "epoch": 0.24072834319650852, "eval_loss": 4.864802837371826, "eval_runtime": 229.1704, "eval_samples_per_second": 11.38, "eval_steps_per_second": 11.38, "eval_tts_loss": 7.457818870576944, "step": 129000 }, { "epoch": 0.240821648755887, "grad_norm": 0.8393458724021912, "learning_rate": 0.00019951681129983876, "loss": 4.6002, "step": 129050 }, { "epoch": 0.2409149543152655, "grad_norm": 1.1570461988449097, "learning_rate": 0.00019951639873587752, "loss": 4.5618, "step": 129100 }, { "epoch": 0.24100825987464397, "grad_norm": 1.0129081010818481, "learning_rate": 0.00019951598599628735, "loss": 4.4886, "step": 129150 }, { "epoch": 0.24110156543402247, "grad_norm": 0.7932485342025757, "learning_rate": 0.00019951557308106897, "loss": 4.6398, "step": 129200 }, { "epoch": 0.24119487099340095, "grad_norm": 0.9829528331756592, "learning_rate": 0.00019951515999022315, "loss": 4.6797, "step": 129250 }, { "epoch": 0.24128817655277945, "grad_norm": 0.748111367225647, "learning_rate": 0.00019951474672375054, "loss": 4.5282, "step": 129300 }, { "epoch": 0.24138148211215796, "grad_norm": 0.9890366792678833, "learning_rate": 0.00019951433328165195, "loss": 4.7415, "step": 129350 }, { "epoch": 0.24147478767153643, "grad_norm": 0.8397785425186157, "learning_rate": 0.00019951391966392806, "loss": 4.704, "step": 129400 }, { "epoch": 0.24156809323091494, "grad_norm": 0.9623748064041138, "learning_rate": 0.0001995135058705796, "loss": 4.6239, "step": 129450 }, { "epoch": 0.2416613987902934, "grad_norm": 0.6656234860420227, "learning_rate": 0.00019951309190160732, "loss": 4.5185, "step": 129500 }, { "epoch": 0.24175470434967192, "grad_norm": 1.2473630905151367, "learning_rate": 0.00019951267775701194, "loss": 4.6633, "step": 129550 }, { "epoch": 0.2418480099090504, "grad_norm": 1.0328094959259033, "learning_rate": 0.00019951226343679419, "loss": 4.6325, "step": 129600 }, { "epoch": 0.2419413154684289, "grad_norm": 0.8722690343856812, "learning_rate": 0.00019951184894095481, "loss": 4.7876, "step": 129650 }, { "epoch": 0.2420346210278074, "grad_norm": 0.8723517060279846, "learning_rate": 0.00019951143426949453, "loss": 4.6523, "step": 129700 }, { "epoch": 0.24212792658718588, "grad_norm": 0.9116722941398621, "learning_rate": 0.00019951101942241407, "loss": 4.6544, "step": 129750 }, { "epoch": 0.24222123214656438, "grad_norm": 0.9696924686431885, "learning_rate": 0.0001995106043997142, "loss": 4.5401, "step": 129800 }, { "epoch": 0.24231453770594286, "grad_norm": 0.9570209383964539, "learning_rate": 0.00019951018920139556, "loss": 4.6208, "step": 129850 }, { "epoch": 0.24240784326532136, "grad_norm": 1.2670087814331055, "learning_rate": 0.00019950977382745898, "loss": 4.6398, "step": 129900 }, { "epoch": 0.24250114882469984, "grad_norm": 0.9587291479110718, "learning_rate": 0.00019950935827790515, "loss": 4.6175, "step": 129950 }, { "epoch": 0.24259445438407834, "grad_norm": 0.772126317024231, "learning_rate": 0.00019950894255273485, "loss": 4.4946, "step": 130000 }, { "epoch": 0.24268775994345684, "grad_norm": 1.0134305953979492, "learning_rate": 0.00019950852665194874, "loss": 4.7511, "step": 130050 }, { "epoch": 0.24278106550283532, "grad_norm": 0.9733072519302368, "learning_rate": 0.0001995081105755476, "loss": 4.7236, "step": 130100 }, { "epoch": 0.24287437106221382, "grad_norm": 0.7325246334075928, "learning_rate": 0.00019950769432353215, "loss": 4.783, "step": 130150 }, { "epoch": 0.2429676766215923, "grad_norm": 1.1216601133346558, "learning_rate": 0.00019950727789590312, "loss": 4.8321, "step": 130200 }, { "epoch": 0.2430609821809708, "grad_norm": 1.0440672636032104, "learning_rate": 0.00019950686129266128, "loss": 4.7706, "step": 130250 }, { "epoch": 0.24315428774034928, "grad_norm": 0.9024654030799866, "learning_rate": 0.0001995064445138073, "loss": 4.7164, "step": 130300 }, { "epoch": 0.24324759329972778, "grad_norm": 0.8911654949188232, "learning_rate": 0.00019950602755934197, "loss": 4.5, "step": 130350 }, { "epoch": 0.24334089885910629, "grad_norm": 0.9491771459579468, "learning_rate": 0.00019950561042926602, "loss": 4.6981, "step": 130400 }, { "epoch": 0.24343420441848476, "grad_norm": 0.9986402988433838, "learning_rate": 0.0001995051931235802, "loss": 4.8321, "step": 130450 }, { "epoch": 0.24352750997786327, "grad_norm": 0.9581927061080933, "learning_rate": 0.0001995047756422852, "loss": 4.7905, "step": 130500 }, { "epoch": 0.24362081553724174, "grad_norm": 0.8729482293128967, "learning_rate": 0.00019950435798538174, "loss": 4.5298, "step": 130550 }, { "epoch": 0.24371412109662025, "grad_norm": 1.103448510169983, "learning_rate": 0.00019950394015287064, "loss": 4.5703, "step": 130600 }, { "epoch": 0.24380742665599872, "grad_norm": 1.0000076293945312, "learning_rate": 0.00019950352214475258, "loss": 4.4457, "step": 130650 }, { "epoch": 0.24390073221537723, "grad_norm": 0.767560601234436, "learning_rate": 0.0001995031039610283, "loss": 5.0027, "step": 130700 }, { "epoch": 0.2439940377747557, "grad_norm": 0.7368218898773193, "learning_rate": 0.00019950268560169857, "loss": 4.6973, "step": 130750 }, { "epoch": 0.2440873433341342, "grad_norm": 0.9467607140541077, "learning_rate": 0.0001995022670667641, "loss": 4.4799, "step": 130800 }, { "epoch": 0.2441806488935127, "grad_norm": 0.8919101357460022, "learning_rate": 0.00019950184835622564, "loss": 4.563, "step": 130850 }, { "epoch": 0.24427395445289118, "grad_norm": 0.9824000000953674, "learning_rate": 0.00019950142947008394, "loss": 4.3836, "step": 130900 }, { "epoch": 0.2443672600122697, "grad_norm": 0.9101470112800598, "learning_rate": 0.00019950101040833973, "loss": 4.5129, "step": 130950 }, { "epoch": 0.24446056557164816, "grad_norm": 0.9340775609016418, "learning_rate": 0.00019950059117099372, "loss": 4.5774, "step": 131000 }, { "epoch": 0.24455387113102667, "grad_norm": 0.7431853413581848, "learning_rate": 0.00019950017175804664, "loss": 4.6977, "step": 131050 }, { "epoch": 0.24464717669040514, "grad_norm": 1.0146548748016357, "learning_rate": 0.00019949975216949933, "loss": 4.6435, "step": 131100 }, { "epoch": 0.24474048224978365, "grad_norm": 1.0537569522857666, "learning_rate": 0.00019949933240535244, "loss": 4.6641, "step": 131150 }, { "epoch": 0.24483378780916215, "grad_norm": 1.1558393239974976, "learning_rate": 0.00019949891246560673, "loss": 4.7927, "step": 131200 }, { "epoch": 0.24492709336854063, "grad_norm": 0.8795566558837891, "learning_rate": 0.00019949849235026296, "loss": 4.7991, "step": 131250 }, { "epoch": 0.24502039892791913, "grad_norm": 1.0229332447052002, "learning_rate": 0.00019949807205932185, "loss": 4.6722, "step": 131300 }, { "epoch": 0.2451137044872976, "grad_norm": 0.985419750213623, "learning_rate": 0.00019949765159278413, "loss": 4.6676, "step": 131350 }, { "epoch": 0.2452070100466761, "grad_norm": 0.8589661121368408, "learning_rate": 0.00019949723095065062, "loss": 4.5871, "step": 131400 }, { "epoch": 0.2453003156060546, "grad_norm": 1.0366382598876953, "learning_rate": 0.00019949681013292194, "loss": 4.7374, "step": 131450 }, { "epoch": 0.2453936211654331, "grad_norm": 1.379646897315979, "learning_rate": 0.00019949638913959892, "loss": 4.7077, "step": 131500 }, { "epoch": 0.2454869267248116, "grad_norm": 1.0151523351669312, "learning_rate": 0.0001994959679706823, "loss": 4.6403, "step": 131550 }, { "epoch": 0.24558023228419007, "grad_norm": 0.8096948266029358, "learning_rate": 0.00019949554662617277, "loss": 4.6464, "step": 131600 }, { "epoch": 0.24567353784356857, "grad_norm": 1.1682924032211304, "learning_rate": 0.00019949512510607115, "loss": 4.6433, "step": 131650 }, { "epoch": 0.24576684340294705, "grad_norm": 1.0436276197433472, "learning_rate": 0.00019949470341037808, "loss": 4.6682, "step": 131700 }, { "epoch": 0.24586014896232555, "grad_norm": 1.0496689081192017, "learning_rate": 0.00019949428153909443, "loss": 4.7054, "step": 131750 }, { "epoch": 0.24595345452170403, "grad_norm": 1.0121614933013916, "learning_rate": 0.00019949385949222083, "loss": 4.656, "step": 131800 }, { "epoch": 0.24604676008108253, "grad_norm": 0.8388519287109375, "learning_rate": 0.00019949343726975808, "loss": 4.5378, "step": 131850 }, { "epoch": 0.24614006564046104, "grad_norm": 0.9778211116790771, "learning_rate": 0.00019949301487170694, "loss": 4.6026, "step": 131900 }, { "epoch": 0.2462333711998395, "grad_norm": 1.1076475381851196, "learning_rate": 0.0001994925922980681, "loss": 4.7711, "step": 131950 }, { "epoch": 0.24632667675921802, "grad_norm": 1.092414379119873, "learning_rate": 0.0001994921695488424, "loss": 4.6508, "step": 132000 }, { "epoch": 0.24632667675921802, "eval_loss": 4.861928939819336, "eval_runtime": 233.537, "eval_samples_per_second": 11.167, "eval_steps_per_second": 11.167, "eval_tts_loss": 7.461830320030999, "step": 132000 }, { "epoch": 0.2464199823185965, "grad_norm": 1.5649731159210205, "learning_rate": 0.0001994917466240305, "loss": 4.7985, "step": 132050 }, { "epoch": 0.246513287877975, "grad_norm": 0.7559811472892761, "learning_rate": 0.00019949132352363317, "loss": 4.545, "step": 132100 }, { "epoch": 0.24660659343735347, "grad_norm": 1.1505687236785889, "learning_rate": 0.00019949090024765114, "loss": 4.7491, "step": 132150 }, { "epoch": 0.24669989899673198, "grad_norm": 0.856060802936554, "learning_rate": 0.0001994904767960852, "loss": 4.6556, "step": 132200 }, { "epoch": 0.24679320455611045, "grad_norm": 1.4347225427627563, "learning_rate": 0.00019949005316893608, "loss": 4.6111, "step": 132250 }, { "epoch": 0.24688651011548896, "grad_norm": 1.2799773216247559, "learning_rate": 0.0001994896293662045, "loss": 4.8259, "step": 132300 }, { "epoch": 0.24697981567486746, "grad_norm": 0.7314300537109375, "learning_rate": 0.00019948920538789125, "loss": 4.5923, "step": 132350 }, { "epoch": 0.24707312123424593, "grad_norm": 1.4587212800979614, "learning_rate": 0.00019948878123399703, "loss": 4.6689, "step": 132400 }, { "epoch": 0.24716642679362444, "grad_norm": 0.7531588077545166, "learning_rate": 0.00019948835690452264, "loss": 4.6059, "step": 132450 }, { "epoch": 0.24725973235300291, "grad_norm": 1.1337380409240723, "learning_rate": 0.00019948793239946881, "loss": 4.6628, "step": 132500 }, { "epoch": 0.24735303791238142, "grad_norm": 0.6850484609603882, "learning_rate": 0.00019948750771883628, "loss": 4.7061, "step": 132550 }, { "epoch": 0.2474463434717599, "grad_norm": 1.0581914186477661, "learning_rate": 0.00019948708286262581, "loss": 4.7761, "step": 132600 }, { "epoch": 0.2475396490311384, "grad_norm": 0.8422567844390869, "learning_rate": 0.00019948665783083814, "loss": 4.8732, "step": 132650 }, { "epoch": 0.2476329545905169, "grad_norm": 1.0814651250839233, "learning_rate": 0.000199486232623474, "loss": 4.8206, "step": 132700 }, { "epoch": 0.24772626014989538, "grad_norm": 1.1069611310958862, "learning_rate": 0.00019948580724053422, "loss": 4.8655, "step": 132750 }, { "epoch": 0.24781956570927388, "grad_norm": 1.0033888816833496, "learning_rate": 0.00019948538168201944, "loss": 4.6903, "step": 132800 }, { "epoch": 0.24791287126865236, "grad_norm": 0.9167322516441345, "learning_rate": 0.00019948495594793048, "loss": 4.4695, "step": 132850 }, { "epoch": 0.24800617682803086, "grad_norm": 0.970410943031311, "learning_rate": 0.00019948453003826807, "loss": 4.647, "step": 132900 }, { "epoch": 0.24809948238740934, "grad_norm": 1.0341490507125854, "learning_rate": 0.00019948410395303302, "loss": 4.6685, "step": 132950 }, { "epoch": 0.24819278794678784, "grad_norm": 0.9720880389213562, "learning_rate": 0.000199483677692226, "loss": 4.6414, "step": 133000 }, { "epoch": 0.24828609350616634, "grad_norm": 0.8272822499275208, "learning_rate": 0.00019948325125584778, "loss": 4.8563, "step": 133050 }, { "epoch": 0.24837939906554482, "grad_norm": 1.2242487668991089, "learning_rate": 0.0001994828246438991, "loss": 4.6999, "step": 133100 }, { "epoch": 0.24847270462492332, "grad_norm": 0.9121090769767761, "learning_rate": 0.0001994823978563808, "loss": 4.752, "step": 133150 }, { "epoch": 0.2485660101843018, "grad_norm": 0.7485866546630859, "learning_rate": 0.00019948197089329352, "loss": 4.6905, "step": 133200 }, { "epoch": 0.2486593157436803, "grad_norm": 0.6626511812210083, "learning_rate": 0.0001994815437546381, "loss": 4.5935, "step": 133250 }, { "epoch": 0.24875262130305878, "grad_norm": 1.1593636274337769, "learning_rate": 0.00019948111644041525, "loss": 4.9399, "step": 133300 }, { "epoch": 0.24884592686243728, "grad_norm": 1.0495115518569946, "learning_rate": 0.00019948068895062575, "loss": 4.6613, "step": 133350 }, { "epoch": 0.24893923242181576, "grad_norm": 1.1008858680725098, "learning_rate": 0.00019948026128527034, "loss": 4.9013, "step": 133400 }, { "epoch": 0.24903253798119426, "grad_norm": 0.784603476524353, "learning_rate": 0.00019947983344434974, "loss": 4.6166, "step": 133450 }, { "epoch": 0.24912584354057277, "grad_norm": 0.9832925796508789, "learning_rate": 0.00019947940542786477, "loss": 4.7428, "step": 133500 }, { "epoch": 0.24921914909995124, "grad_norm": 0.9270418286323547, "learning_rate": 0.0001994789772358161, "loss": 4.4206, "step": 133550 }, { "epoch": 0.24931245465932975, "grad_norm": 1.2622601985931396, "learning_rate": 0.00019947854886820458, "loss": 4.6319, "step": 133600 }, { "epoch": 0.24940576021870822, "grad_norm": 0.7396588325500488, "learning_rate": 0.00019947812032503092, "loss": 4.786, "step": 133650 }, { "epoch": 0.24949906577808673, "grad_norm": 0.9550930261611938, "learning_rate": 0.00019947769160629587, "loss": 4.7091, "step": 133700 }, { "epoch": 0.2495923713374652, "grad_norm": 1.0118082761764526, "learning_rate": 0.00019947726271200025, "loss": 4.7579, "step": 133750 }, { "epoch": 0.2496856768968437, "grad_norm": 0.838410496711731, "learning_rate": 0.0001994768336421447, "loss": 4.7016, "step": 133800 }, { "epoch": 0.2497789824562222, "grad_norm": 0.8363553881645203, "learning_rate": 0.00019947640439673008, "loss": 4.6482, "step": 133850 }, { "epoch": 0.24987228801560069, "grad_norm": 1.0128004550933838, "learning_rate": 0.0001994759749757571, "loss": 4.8391, "step": 133900 }, { "epoch": 0.2499655935749792, "grad_norm": 0.8684887290000916, "learning_rate": 0.00019947554537922652, "loss": 4.5435, "step": 133950 }, { "epoch": 0.2500588991343577, "grad_norm": 0.8165358901023865, "learning_rate": 0.0001994751156071391, "loss": 4.6167, "step": 134000 }, { "epoch": 0.25015220469373617, "grad_norm": 0.9833562970161438, "learning_rate": 0.00019947468565949562, "loss": 4.7929, "step": 134050 }, { "epoch": 0.25024551025311464, "grad_norm": 0.8654858469963074, "learning_rate": 0.00019947425553629682, "loss": 4.8782, "step": 134100 }, { "epoch": 0.2503388158124931, "grad_norm": 0.8595501184463501, "learning_rate": 0.00019947382523754348, "loss": 4.6882, "step": 134150 }, { "epoch": 0.25043212137187165, "grad_norm": 0.9764409065246582, "learning_rate": 0.0001994733947632363, "loss": 4.723, "step": 134200 }, { "epoch": 0.25052542693125013, "grad_norm": 1.361682415008545, "learning_rate": 0.0001994729641133761, "loss": 4.9449, "step": 134250 }, { "epoch": 0.2506187324906286, "grad_norm": 1.0243170261383057, "learning_rate": 0.00019947253328796364, "loss": 4.8382, "step": 134300 }, { "epoch": 0.25071203805000714, "grad_norm": 0.9603681564331055, "learning_rate": 0.00019947210228699963, "loss": 4.7538, "step": 134350 }, { "epoch": 0.2508053436093856, "grad_norm": 0.8127009272575378, "learning_rate": 0.0001994716711104849, "loss": 4.74, "step": 134400 }, { "epoch": 0.2508986491687641, "grad_norm": 1.0043110847473145, "learning_rate": 0.00019947123975842012, "loss": 4.6537, "step": 134450 }, { "epoch": 0.25099195472814256, "grad_norm": 0.88910973072052, "learning_rate": 0.00019947080823080616, "loss": 4.6681, "step": 134500 }, { "epoch": 0.2510852602875211, "grad_norm": 1.072106957435608, "learning_rate": 0.00019947037652764369, "loss": 4.6476, "step": 134550 }, { "epoch": 0.25117856584689957, "grad_norm": 1.160860538482666, "learning_rate": 0.0001994699446489335, "loss": 4.8706, "step": 134600 }, { "epoch": 0.25127187140627805, "grad_norm": 1.0427870750427246, "learning_rate": 0.0001994695125946764, "loss": 4.7866, "step": 134650 }, { "epoch": 0.2513651769656566, "grad_norm": 0.8088847398757935, "learning_rate": 0.00019946908036487307, "loss": 4.5036, "step": 134700 }, { "epoch": 0.25145848252503505, "grad_norm": 0.8402940034866333, "learning_rate": 0.00019946864795952434, "loss": 4.4938, "step": 134750 }, { "epoch": 0.25155178808441353, "grad_norm": 0.9996683597564697, "learning_rate": 0.00019946821537863095, "loss": 4.6048, "step": 134800 }, { "epoch": 0.251645093643792, "grad_norm": 0.9219832420349121, "learning_rate": 0.00019946778262219366, "loss": 4.6674, "step": 134850 }, { "epoch": 0.25173839920317054, "grad_norm": 0.9935362935066223, "learning_rate": 0.00019946734969021322, "loss": 4.8539, "step": 134900 }, { "epoch": 0.251831704762549, "grad_norm": 1.0310291051864624, "learning_rate": 0.00019946691658269042, "loss": 4.5868, "step": 134950 }, { "epoch": 0.2519250103219275, "grad_norm": 0.9530249238014221, "learning_rate": 0.00019946648329962602, "loss": 4.6397, "step": 135000 }, { "epoch": 0.2519250103219275, "eval_loss": 4.854243755340576, "eval_runtime": 229.3514, "eval_samples_per_second": 11.371, "eval_steps_per_second": 11.371, "eval_tts_loss": 7.440514219263454, "step": 135000 }, { "epoch": 0.252018315881306, "grad_norm": 0.7310658097267151, "learning_rate": 0.00019946604984102077, "loss": 4.6717, "step": 135050 }, { "epoch": 0.2521116214406845, "grad_norm": 0.7537361979484558, "learning_rate": 0.00019946561620687544, "loss": 4.6167, "step": 135100 }, { "epoch": 0.252204927000063, "grad_norm": 0.9830604195594788, "learning_rate": 0.00019946518239719082, "loss": 4.4883, "step": 135150 }, { "epoch": 0.25229823255944145, "grad_norm": 0.8440564870834351, "learning_rate": 0.00019946474841196765, "loss": 4.7197, "step": 135200 }, { "epoch": 0.25239153811882, "grad_norm": 1.2416318655014038, "learning_rate": 0.00019946431425120665, "loss": 4.5972, "step": 135250 }, { "epoch": 0.25248484367819846, "grad_norm": 1.1739615201950073, "learning_rate": 0.00019946387991490871, "loss": 4.4314, "step": 135300 }, { "epoch": 0.25257814923757693, "grad_norm": 1.079363226890564, "learning_rate": 0.0001994634454030745, "loss": 4.7442, "step": 135350 }, { "epoch": 0.25267145479695546, "grad_norm": 1.0851373672485352, "learning_rate": 0.00019946301071570478, "loss": 4.7188, "step": 135400 }, { "epoch": 0.25276476035633394, "grad_norm": 0.8519002199172974, "learning_rate": 0.00019946257585280036, "loss": 4.5734, "step": 135450 }, { "epoch": 0.2528580659157124, "grad_norm": 1.175642728805542, "learning_rate": 0.00019946214081436204, "loss": 4.8021, "step": 135500 }, { "epoch": 0.2529513714750909, "grad_norm": 0.984249472618103, "learning_rate": 0.0001994617056003905, "loss": 4.7219, "step": 135550 }, { "epoch": 0.2530446770344694, "grad_norm": 1.4463765621185303, "learning_rate": 0.00019946127021088655, "loss": 4.5299, "step": 135600 }, { "epoch": 0.2531379825938479, "grad_norm": 1.2375439405441284, "learning_rate": 0.00019946083464585096, "loss": 4.6073, "step": 135650 }, { "epoch": 0.2532312881532264, "grad_norm": 1.073440432548523, "learning_rate": 0.0001994603989052845, "loss": 4.5566, "step": 135700 }, { "epoch": 0.2533245937126049, "grad_norm": 1.0174309015274048, "learning_rate": 0.00019945996298918793, "loss": 4.6937, "step": 135750 }, { "epoch": 0.2534178992719834, "grad_norm": 0.9919219613075256, "learning_rate": 0.00019945952689756204, "loss": 4.665, "step": 135800 }, { "epoch": 0.25351120483136186, "grad_norm": 0.8121524453163147, "learning_rate": 0.0001994590906304076, "loss": 4.882, "step": 135850 }, { "epoch": 0.25360451039074033, "grad_norm": 0.5900788903236389, "learning_rate": 0.00019945865418772533, "loss": 4.7844, "step": 135900 }, { "epoch": 0.25369781595011887, "grad_norm": 1.1510095596313477, "learning_rate": 0.00019945821756951605, "loss": 4.729, "step": 135950 }, { "epoch": 0.25379112150949734, "grad_norm": 0.9832910895347595, "learning_rate": 0.0001994577807757805, "loss": 4.7004, "step": 136000 }, { "epoch": 0.2538844270688758, "grad_norm": 0.8559871315956116, "learning_rate": 0.0001994573438065195, "loss": 4.7026, "step": 136050 }, { "epoch": 0.25397773262825435, "grad_norm": 0.9402391910552979, "learning_rate": 0.00019945690666173376, "loss": 4.5792, "step": 136100 }, { "epoch": 0.2540710381876328, "grad_norm": 0.9144654870033264, "learning_rate": 0.00019945646934142408, "loss": 4.6403, "step": 136150 }, { "epoch": 0.2541643437470113, "grad_norm": 1.0262720584869385, "learning_rate": 0.00019945603184559125, "loss": 4.766, "step": 136200 }, { "epoch": 0.2542576493063898, "grad_norm": 0.9860953092575073, "learning_rate": 0.00019945559417423604, "loss": 4.7118, "step": 136250 }, { "epoch": 0.2543509548657683, "grad_norm": 0.9816266298294067, "learning_rate": 0.00019945515632735917, "loss": 4.6694, "step": 136300 }, { "epoch": 0.2544442604251468, "grad_norm": 1.0225788354873657, "learning_rate": 0.00019945471830496144, "loss": 4.6733, "step": 136350 }, { "epoch": 0.25453756598452526, "grad_norm": 0.8615061640739441, "learning_rate": 0.00019945428010704368, "loss": 4.4516, "step": 136400 }, { "epoch": 0.2546308715439038, "grad_norm": 0.9485636353492737, "learning_rate": 0.00019945384173360658, "loss": 4.7265, "step": 136450 }, { "epoch": 0.25472417710328227, "grad_norm": 0.9879016876220703, "learning_rate": 0.00019945340318465097, "loss": 4.6516, "step": 136500 }, { "epoch": 0.25481748266266074, "grad_norm": 0.886114776134491, "learning_rate": 0.00019945296446017755, "loss": 4.7834, "step": 136550 }, { "epoch": 0.2549107882220392, "grad_norm": 0.901164174079895, "learning_rate": 0.00019945252556018722, "loss": 4.7377, "step": 136600 }, { "epoch": 0.25500409378141775, "grad_norm": 1.0700454711914062, "learning_rate": 0.00019945208648468064, "loss": 4.629, "step": 136650 }, { "epoch": 0.2550973993407962, "grad_norm": 0.761396586894989, "learning_rate": 0.00019945164723365863, "loss": 4.5535, "step": 136700 }, { "epoch": 0.2551907049001747, "grad_norm": 0.9955791234970093, "learning_rate": 0.00019945120780712197, "loss": 4.8143, "step": 136750 }, { "epoch": 0.2552840104595532, "grad_norm": 0.7975375652313232, "learning_rate": 0.00019945076820507144, "loss": 4.6698, "step": 136800 }, { "epoch": 0.2553773160189317, "grad_norm": 0.8464979529380798, "learning_rate": 0.00019945032842750776, "loss": 4.6989, "step": 136850 }, { "epoch": 0.2554706215783102, "grad_norm": 0.9093438386917114, "learning_rate": 0.0001994498884744318, "loss": 4.7043, "step": 136900 }, { "epoch": 0.25556392713768866, "grad_norm": 0.7675898671150208, "learning_rate": 0.00019944944834584426, "loss": 4.8434, "step": 136950 }, { "epoch": 0.2556572326970672, "grad_norm": 1.0677566528320312, "learning_rate": 0.00019944900804174593, "loss": 4.7249, "step": 137000 }, { "epoch": 0.25575053825644567, "grad_norm": 0.7189279794692993, "learning_rate": 0.00019944856756213763, "loss": 4.5614, "step": 137050 }, { "epoch": 0.25584384381582415, "grad_norm": 1.0350892543792725, "learning_rate": 0.0001994481269070201, "loss": 4.963, "step": 137100 }, { "epoch": 0.2559371493752026, "grad_norm": 0.7921010255813599, "learning_rate": 0.0001994476860763941, "loss": 4.5003, "step": 137150 }, { "epoch": 0.25603045493458115, "grad_norm": 1.237960934638977, "learning_rate": 0.00019944724507026045, "loss": 4.6863, "step": 137200 }, { "epoch": 0.25612376049395963, "grad_norm": 0.9462816119194031, "learning_rate": 0.0001994468038886199, "loss": 4.6557, "step": 137250 }, { "epoch": 0.2562170660533381, "grad_norm": 1.1308284997940063, "learning_rate": 0.00019944636253147325, "loss": 4.5256, "step": 137300 }, { "epoch": 0.25631037161271664, "grad_norm": 0.8221685886383057, "learning_rate": 0.00019944592099882126, "loss": 4.4556, "step": 137350 }, { "epoch": 0.2564036771720951, "grad_norm": 0.8622978925704956, "learning_rate": 0.00019944547929066472, "loss": 4.5635, "step": 137400 }, { "epoch": 0.2564969827314736, "grad_norm": 1.0377954244613647, "learning_rate": 0.0001994450374070044, "loss": 4.7431, "step": 137450 }, { "epoch": 0.25659028829085206, "grad_norm": 0.9822484254837036, "learning_rate": 0.00019944459534784113, "loss": 4.5507, "step": 137500 }, { "epoch": 0.2566835938502306, "grad_norm": 1.063193917274475, "learning_rate": 0.0001994441531131756, "loss": 4.6944, "step": 137550 }, { "epoch": 0.25677689940960907, "grad_norm": 0.9671196937561035, "learning_rate": 0.00019944371070300865, "loss": 4.7503, "step": 137600 }, { "epoch": 0.25687020496898755, "grad_norm": 1.0232065916061401, "learning_rate": 0.00019944326811734104, "loss": 4.4999, "step": 137650 }, { "epoch": 0.2569635105283661, "grad_norm": 0.7935104966163635, "learning_rate": 0.00019944282535617358, "loss": 4.7047, "step": 137700 }, { "epoch": 0.25705681608774456, "grad_norm": 1.06540846824646, "learning_rate": 0.00019944238241950704, "loss": 4.8058, "step": 137750 }, { "epoch": 0.25715012164712303, "grad_norm": 1.0669400691986084, "learning_rate": 0.00019944193930734216, "loss": 4.6499, "step": 137800 }, { "epoch": 0.2572434272065015, "grad_norm": 0.9878630042076111, "learning_rate": 0.00019944149601967977, "loss": 4.5647, "step": 137850 }, { "epoch": 0.25733673276588004, "grad_norm": 1.0921598672866821, "learning_rate": 0.00019944105255652065, "loss": 4.6141, "step": 137900 }, { "epoch": 0.2574300383252585, "grad_norm": 0.7695943117141724, "learning_rate": 0.00019944060891786554, "loss": 4.5943, "step": 137950 }, { "epoch": 0.257523343884637, "grad_norm": 1.0480443239212036, "learning_rate": 0.00019944016510371526, "loss": 4.7454, "step": 138000 }, { "epoch": 0.257523343884637, "eval_loss": 4.861461162567139, "eval_runtime": 230.715, "eval_samples_per_second": 11.304, "eval_steps_per_second": 11.304, "eval_tts_loss": 7.436345217680913, "step": 138000 }, { "epoch": 0.2576166494440155, "grad_norm": 1.0323460102081299, "learning_rate": 0.0001994397211140706, "loss": 4.7979, "step": 138050 }, { "epoch": 0.257709955003394, "grad_norm": 1.2108174562454224, "learning_rate": 0.0001994392769489323, "loss": 4.5913, "step": 138100 }, { "epoch": 0.2578032605627725, "grad_norm": 0.9902898073196411, "learning_rate": 0.0001994388326083012, "loss": 4.4259, "step": 138150 }, { "epoch": 0.25789656612215095, "grad_norm": 1.0889242887496948, "learning_rate": 0.00019943838809217805, "loss": 4.6503, "step": 138200 }, { "epoch": 0.2579898716815295, "grad_norm": 0.991397500038147, "learning_rate": 0.00019943794340056362, "loss": 4.6364, "step": 138250 }, { "epoch": 0.25808317724090796, "grad_norm": 1.141694188117981, "learning_rate": 0.00019943749853345875, "loss": 4.7025, "step": 138300 }, { "epoch": 0.25817648280028643, "grad_norm": 0.8368604183197021, "learning_rate": 0.0001994370534908642, "loss": 4.6701, "step": 138350 }, { "epoch": 0.25826978835966496, "grad_norm": 1.0592689514160156, "learning_rate": 0.00019943660827278072, "loss": 4.569, "step": 138400 }, { "epoch": 0.25836309391904344, "grad_norm": 0.7820909023284912, "learning_rate": 0.00019943616287920914, "loss": 4.7223, "step": 138450 }, { "epoch": 0.2584563994784219, "grad_norm": 1.0248945951461792, "learning_rate": 0.0001994357173101502, "loss": 4.6945, "step": 138500 }, { "epoch": 0.2585497050378004, "grad_norm": 0.9100630283355713, "learning_rate": 0.00019943527156560475, "loss": 4.5899, "step": 138550 }, { "epoch": 0.2586430105971789, "grad_norm": 1.1249892711639404, "learning_rate": 0.00019943482564557352, "loss": 4.6453, "step": 138600 }, { "epoch": 0.2587363161565574, "grad_norm": 1.033903956413269, "learning_rate": 0.00019943437955005733, "loss": 4.7613, "step": 138650 }, { "epoch": 0.2588296217159359, "grad_norm": 1.0414884090423584, "learning_rate": 0.00019943393327905695, "loss": 4.6118, "step": 138700 }, { "epoch": 0.2589229272753144, "grad_norm": 0.7779209017753601, "learning_rate": 0.00019943348683257317, "loss": 4.7519, "step": 138750 }, { "epoch": 0.2590162328346929, "grad_norm": 1.0751841068267822, "learning_rate": 0.0001994330402106068, "loss": 4.5788, "step": 138800 }, { "epoch": 0.25910953839407136, "grad_norm": 0.7722923159599304, "learning_rate": 0.00019943259341315857, "loss": 4.4937, "step": 138850 }, { "epoch": 0.25920284395344984, "grad_norm": 0.8846758008003235, "learning_rate": 0.00019943214644022934, "loss": 4.7343, "step": 138900 }, { "epoch": 0.25929614951282837, "grad_norm": 1.157002329826355, "learning_rate": 0.00019943169929181987, "loss": 4.8432, "step": 138950 }, { "epoch": 0.25938945507220684, "grad_norm": 0.8861170411109924, "learning_rate": 0.00019943125196793094, "loss": 4.7967, "step": 139000 }, { "epoch": 0.2594827606315853, "grad_norm": 0.9354442954063416, "learning_rate": 0.00019943080446856333, "loss": 4.6719, "step": 139050 }, { "epoch": 0.25957606619096385, "grad_norm": 1.0083680152893066, "learning_rate": 0.00019943035679371787, "loss": 4.6639, "step": 139100 }, { "epoch": 0.2596693717503423, "grad_norm": 1.0523858070373535, "learning_rate": 0.0001994299089433953, "loss": 4.6234, "step": 139150 }, { "epoch": 0.2597626773097208, "grad_norm": 0.9569562077522278, "learning_rate": 0.00019942946091759648, "loss": 4.4507, "step": 139200 }, { "epoch": 0.2598559828690993, "grad_norm": 1.1821283102035522, "learning_rate": 0.00019942901271632213, "loss": 4.7459, "step": 139250 }, { "epoch": 0.2599492884284778, "grad_norm": 0.9408050179481506, "learning_rate": 0.00019942856433957304, "loss": 4.6761, "step": 139300 }, { "epoch": 0.2600425939878563, "grad_norm": 0.939137876033783, "learning_rate": 0.00019942811578735006, "loss": 4.6466, "step": 139350 }, { "epoch": 0.26013589954723476, "grad_norm": 0.8856296539306641, "learning_rate": 0.00019942766705965395, "loss": 4.9821, "step": 139400 }, { "epoch": 0.26022920510661324, "grad_norm": 0.6106601357460022, "learning_rate": 0.00019942721815648548, "loss": 4.5543, "step": 139450 }, { "epoch": 0.26032251066599177, "grad_norm": 0.8569189310073853, "learning_rate": 0.0001994267690778455, "loss": 4.6647, "step": 139500 }, { "epoch": 0.26041581622537024, "grad_norm": 1.102795958518982, "learning_rate": 0.00019942631982373474, "loss": 4.4512, "step": 139550 }, { "epoch": 0.2605091217847487, "grad_norm": 0.9730409979820251, "learning_rate": 0.00019942587039415403, "loss": 4.9565, "step": 139600 }, { "epoch": 0.26060242734412725, "grad_norm": 0.7721739411354065, "learning_rate": 0.00019942542078910415, "loss": 4.7068, "step": 139650 }, { "epoch": 0.26069573290350573, "grad_norm": 0.70539790391922, "learning_rate": 0.0001994249710085859, "loss": 4.5592, "step": 139700 }, { "epoch": 0.2607890384628842, "grad_norm": 1.0219758749008179, "learning_rate": 0.00019942452105260007, "loss": 4.6835, "step": 139750 }, { "epoch": 0.2608823440222627, "grad_norm": 1.2075634002685547, "learning_rate": 0.00019942407092114745, "loss": 4.5386, "step": 139800 }, { "epoch": 0.2609756495816412, "grad_norm": 0.8083966970443726, "learning_rate": 0.00019942362061422885, "loss": 4.6374, "step": 139850 }, { "epoch": 0.2610689551410197, "grad_norm": 0.7901577353477478, "learning_rate": 0.00019942317013184506, "loss": 4.7071, "step": 139900 }, { "epoch": 0.26116226070039816, "grad_norm": 1.2326618432998657, "learning_rate": 0.00019942271947399683, "loss": 4.6927, "step": 139950 }, { "epoch": 0.2612555662597767, "grad_norm": 0.7666860818862915, "learning_rate": 0.00019942226864068502, "loss": 4.8306, "step": 140000 }, { "epoch": 0.26134887181915517, "grad_norm": 0.9988517761230469, "learning_rate": 0.00019942181763191042, "loss": 4.6557, "step": 140050 }, { "epoch": 0.26144217737853365, "grad_norm": 0.7240771651268005, "learning_rate": 0.0001994213664476738, "loss": 4.4553, "step": 140100 }, { "epoch": 0.2615354829379121, "grad_norm": 1.13454270362854, "learning_rate": 0.00019942091508797595, "loss": 4.5737, "step": 140150 }, { "epoch": 0.26162878849729065, "grad_norm": 0.8832876682281494, "learning_rate": 0.00019942046355281767, "loss": 4.7567, "step": 140200 }, { "epoch": 0.26172209405666913, "grad_norm": 1.089231014251709, "learning_rate": 0.00019942001184219976, "loss": 4.5035, "step": 140250 }, { "epoch": 0.2618153996160476, "grad_norm": 1.0818380117416382, "learning_rate": 0.00019941955995612305, "loss": 4.8077, "step": 140300 }, { "epoch": 0.26190870517542614, "grad_norm": 1.0916963815689087, "learning_rate": 0.00019941910789458825, "loss": 4.7335, "step": 140350 }, { "epoch": 0.2620020107348046, "grad_norm": 1.029809594154358, "learning_rate": 0.00019941865565759628, "loss": 4.7261, "step": 140400 }, { "epoch": 0.2620953162941831, "grad_norm": 0.786016047000885, "learning_rate": 0.00019941820324514785, "loss": 4.6044, "step": 140450 }, { "epoch": 0.26218862185356157, "grad_norm": 0.9927809834480286, "learning_rate": 0.0001994177506572438, "loss": 4.5903, "step": 140500 }, { "epoch": 0.2622819274129401, "grad_norm": 1.1766144037246704, "learning_rate": 0.0001994172978938849, "loss": 4.4297, "step": 140550 }, { "epoch": 0.2623752329723186, "grad_norm": 0.8965726494789124, "learning_rate": 0.00019941684495507198, "loss": 4.7621, "step": 140600 }, { "epoch": 0.26246853853169705, "grad_norm": 0.9904941916465759, "learning_rate": 0.00019941639184080582, "loss": 4.5519, "step": 140650 }, { "epoch": 0.2625618440910756, "grad_norm": 0.9716530442237854, "learning_rate": 0.0001994159385510872, "loss": 4.6065, "step": 140700 }, { "epoch": 0.26265514965045406, "grad_norm": 0.7707858085632324, "learning_rate": 0.00019941548508591695, "loss": 4.6938, "step": 140750 }, { "epoch": 0.26274845520983253, "grad_norm": 1.0510951280593872, "learning_rate": 0.00019941503144529587, "loss": 4.8028, "step": 140800 }, { "epoch": 0.262841760769211, "grad_norm": 0.8451001048088074, "learning_rate": 0.00019941457762922475, "loss": 4.4955, "step": 140850 }, { "epoch": 0.26293506632858954, "grad_norm": 0.8992056250572205, "learning_rate": 0.00019941412363770438, "loss": 4.6839, "step": 140900 }, { "epoch": 0.263028371887968, "grad_norm": 0.8834390640258789, "learning_rate": 0.00019941366947073559, "loss": 4.6869, "step": 140950 }, { "epoch": 0.2631216774473465, "grad_norm": 1.1744734048843384, "learning_rate": 0.00019941321512831914, "loss": 4.6612, "step": 141000 }, { "epoch": 0.2631216774473465, "eval_loss": 4.854221820831299, "eval_runtime": 230.4865, "eval_samples_per_second": 11.315, "eval_steps_per_second": 11.315, "eval_tts_loss": 7.429972851733202, "step": 141000 }, { "epoch": 0.263214983006725, "grad_norm": 1.01349675655365, "learning_rate": 0.00019941276061045588, "loss": 4.7109, "step": 141050 }, { "epoch": 0.2633082885661035, "grad_norm": 0.7250033617019653, "learning_rate": 0.0001994123059171466, "loss": 4.6136, "step": 141100 }, { "epoch": 0.263401594125482, "grad_norm": 0.8900704979896545, "learning_rate": 0.00019941185104839208, "loss": 4.4201, "step": 141150 }, { "epoch": 0.26349489968486045, "grad_norm": 0.9894741773605347, "learning_rate": 0.00019941139600419315, "loss": 4.5358, "step": 141200 }, { "epoch": 0.263588205244239, "grad_norm": 0.728403627872467, "learning_rate": 0.00019941094078455057, "loss": 4.4723, "step": 141250 }, { "epoch": 0.26368151080361746, "grad_norm": 0.9740440249443054, "learning_rate": 0.00019941048538946516, "loss": 4.9644, "step": 141300 }, { "epoch": 0.26377481636299593, "grad_norm": 0.9161767959594727, "learning_rate": 0.00019941002981893778, "loss": 4.7106, "step": 141350 }, { "epoch": 0.26386812192237447, "grad_norm": 1.2737964391708374, "learning_rate": 0.00019940957407296918, "loss": 4.569, "step": 141400 }, { "epoch": 0.26396142748175294, "grad_norm": 0.902451753616333, "learning_rate": 0.00019940911815156012, "loss": 4.6479, "step": 141450 }, { "epoch": 0.2640547330411314, "grad_norm": 0.8375920653343201, "learning_rate": 0.00019940866205471152, "loss": 4.7673, "step": 141500 }, { "epoch": 0.2641480386005099, "grad_norm": 0.6975705027580261, "learning_rate": 0.0001994082057824241, "loss": 4.638, "step": 141550 }, { "epoch": 0.2642413441598884, "grad_norm": 1.0885804891586304, "learning_rate": 0.00019940774933469867, "loss": 4.7292, "step": 141600 }, { "epoch": 0.2643346497192669, "grad_norm": 1.1714972257614136, "learning_rate": 0.00019940729271153608, "loss": 4.7405, "step": 141650 }, { "epoch": 0.2644279552786454, "grad_norm": 1.126559853553772, "learning_rate": 0.00019940683591293713, "loss": 4.58, "step": 141700 }, { "epoch": 0.2645212608380239, "grad_norm": 1.1203880310058594, "learning_rate": 0.00019940637893890256, "loss": 4.6255, "step": 141750 }, { "epoch": 0.2646145663974024, "grad_norm": 0.9070851802825928, "learning_rate": 0.00019940592178943324, "loss": 4.8198, "step": 141800 }, { "epoch": 0.26470787195678086, "grad_norm": 0.8276423215866089, "learning_rate": 0.00019940546446452998, "loss": 4.6199, "step": 141850 }, { "epoch": 0.26480117751615934, "grad_norm": 1.2383900880813599, "learning_rate": 0.0001994050069641935, "loss": 4.6207, "step": 141900 }, { "epoch": 0.26489448307553787, "grad_norm": 0.9433544278144836, "learning_rate": 0.00019940454928842475, "loss": 4.7679, "step": 141950 }, { "epoch": 0.26498778863491634, "grad_norm": 1.12177574634552, "learning_rate": 0.00019940409143722444, "loss": 4.8803, "step": 142000 }, { "epoch": 0.2650810941942948, "grad_norm": 0.8541791439056396, "learning_rate": 0.00019940363341059336, "loss": 4.7773, "step": 142050 }, { "epoch": 0.2651743997536733, "grad_norm": 1.2318682670593262, "learning_rate": 0.0001994031752085324, "loss": 4.7115, "step": 142100 }, { "epoch": 0.2652677053130518, "grad_norm": 1.1669272184371948, "learning_rate": 0.0001994027168310423, "loss": 4.6872, "step": 142150 }, { "epoch": 0.2653610108724303, "grad_norm": 0.7266700267791748, "learning_rate": 0.0001994022582781239, "loss": 4.4811, "step": 142200 }, { "epoch": 0.2654543164318088, "grad_norm": 0.6903916001319885, "learning_rate": 0.000199401799549778, "loss": 4.7914, "step": 142250 }, { "epoch": 0.2655476219911873, "grad_norm": 1.0394556522369385, "learning_rate": 0.00019940134064600547, "loss": 4.686, "step": 142300 }, { "epoch": 0.2656409275505658, "grad_norm": 1.02165687084198, "learning_rate": 0.00019940088156680702, "loss": 4.5818, "step": 142350 }, { "epoch": 0.26573423310994426, "grad_norm": 0.8454614281654358, "learning_rate": 0.0001994004223121835, "loss": 4.6726, "step": 142400 }, { "epoch": 0.26582753866932274, "grad_norm": 1.206411600112915, "learning_rate": 0.00019939996288213574, "loss": 4.6144, "step": 142450 }, { "epoch": 0.26592084422870127, "grad_norm": 0.9744541049003601, "learning_rate": 0.00019939950327666452, "loss": 4.902, "step": 142500 }, { "epoch": 0.26601414978807975, "grad_norm": 0.7370103001594543, "learning_rate": 0.00019939904349577069, "loss": 4.5981, "step": 142550 }, { "epoch": 0.2661074553474582, "grad_norm": 0.9200401902198792, "learning_rate": 0.000199398583539455, "loss": 4.5697, "step": 142600 }, { "epoch": 0.26620076090683675, "grad_norm": 1.0769352912902832, "learning_rate": 0.00019939812340771833, "loss": 4.7748, "step": 142650 }, { "epoch": 0.26629406646621523, "grad_norm": 1.1400402784347534, "learning_rate": 0.00019939766310056144, "loss": 4.7504, "step": 142700 }, { "epoch": 0.2663873720255937, "grad_norm": 0.8837327361106873, "learning_rate": 0.0001993972026179852, "loss": 4.7623, "step": 142750 }, { "epoch": 0.2664806775849722, "grad_norm": 0.9852148294448853, "learning_rate": 0.00019939674195999034, "loss": 4.7179, "step": 142800 }, { "epoch": 0.2665739831443507, "grad_norm": 1.1082642078399658, "learning_rate": 0.00019939628112657776, "loss": 4.6306, "step": 142850 }, { "epoch": 0.2666672887037292, "grad_norm": 1.2532259225845337, "learning_rate": 0.0001993958201177482, "loss": 4.7182, "step": 142900 }, { "epoch": 0.26676059426310766, "grad_norm": 1.113945484161377, "learning_rate": 0.00019939535893350252, "loss": 4.7235, "step": 142950 }, { "epoch": 0.2668538998224862, "grad_norm": 0.853884220123291, "learning_rate": 0.00019939489757384154, "loss": 4.7894, "step": 143000 }, { "epoch": 0.26694720538186467, "grad_norm": 1.0670225620269775, "learning_rate": 0.00019939443603876603, "loss": 4.6292, "step": 143050 }, { "epoch": 0.26704051094124315, "grad_norm": 1.090808629989624, "learning_rate": 0.00019939397432827685, "loss": 4.7517, "step": 143100 }, { "epoch": 0.2671338165006216, "grad_norm": 0.8925207853317261, "learning_rate": 0.0001993935124423748, "loss": 4.5034, "step": 143150 }, { "epoch": 0.26722712206000016, "grad_norm": 0.7665544748306274, "learning_rate": 0.00019939305038106064, "loss": 4.8442, "step": 143200 }, { "epoch": 0.26732042761937863, "grad_norm": 1.0686653852462769, "learning_rate": 0.00019939258814433527, "loss": 4.5459, "step": 143250 }, { "epoch": 0.2674137331787571, "grad_norm": 1.0229724645614624, "learning_rate": 0.00019939212573219948, "loss": 4.7637, "step": 143300 }, { "epoch": 0.26750703873813564, "grad_norm": 1.1056723594665527, "learning_rate": 0.00019939166314465406, "loss": 4.4069, "step": 143350 }, { "epoch": 0.2676003442975141, "grad_norm": 0.7907326221466064, "learning_rate": 0.00019939120038169983, "loss": 4.6325, "step": 143400 }, { "epoch": 0.2676936498568926, "grad_norm": 1.0753734111785889, "learning_rate": 0.00019939073744333765, "loss": 4.6872, "step": 143450 }, { "epoch": 0.26778695541627107, "grad_norm": 1.119105339050293, "learning_rate": 0.00019939027432956828, "loss": 4.64, "step": 143500 }, { "epoch": 0.2678802609756496, "grad_norm": 0.839352011680603, "learning_rate": 0.00019938981104039257, "loss": 4.6258, "step": 143550 }, { "epoch": 0.2679735665350281, "grad_norm": 1.0468686819076538, "learning_rate": 0.00019938934757581132, "loss": 4.6909, "step": 143600 }, { "epoch": 0.26806687209440655, "grad_norm": 1.086549162864685, "learning_rate": 0.0001993888839358254, "loss": 4.6586, "step": 143650 }, { "epoch": 0.2681601776537851, "grad_norm": 0.9070858359336853, "learning_rate": 0.0001993884201204356, "loss": 4.8447, "step": 143700 }, { "epoch": 0.26825348321316356, "grad_norm": 1.0874264240264893, "learning_rate": 0.00019938795612964267, "loss": 4.6228, "step": 143750 }, { "epoch": 0.26834678877254203, "grad_norm": 0.9141344428062439, "learning_rate": 0.0001993874919634475, "loss": 4.6042, "step": 143800 }, { "epoch": 0.2684400943319205, "grad_norm": 1.0117084980010986, "learning_rate": 0.0001993870276218509, "loss": 4.5892, "step": 143850 }, { "epoch": 0.26853339989129904, "grad_norm": 1.1900421380996704, "learning_rate": 0.0001993865631048537, "loss": 4.6517, "step": 143900 }, { "epoch": 0.2686267054506775, "grad_norm": 1.108844518661499, "learning_rate": 0.0001993860984124567, "loss": 4.6209, "step": 143950 }, { "epoch": 0.268720011010056, "grad_norm": 0.9162120223045349, "learning_rate": 0.00019938563354466072, "loss": 4.7472, "step": 144000 }, { "epoch": 0.268720011010056, "eval_loss": 4.846428394317627, "eval_runtime": 231.517, "eval_samples_per_second": 11.265, "eval_steps_per_second": 11.265, "eval_tts_loss": 7.418669227682099, "step": 144000 }, { "epoch": 0.2688133165694345, "grad_norm": 0.8355111479759216, "learning_rate": 0.00019938516850146655, "loss": 4.5845, "step": 144050 }, { "epoch": 0.268906622128813, "grad_norm": 0.6543982625007629, "learning_rate": 0.0001993847032828751, "loss": 4.5369, "step": 144100 }, { "epoch": 0.2689999276881915, "grad_norm": 0.947825014591217, "learning_rate": 0.0001993842378888871, "loss": 4.6122, "step": 144150 }, { "epoch": 0.26909323324756995, "grad_norm": 1.1164684295654297, "learning_rate": 0.0001993837723195034, "loss": 4.6844, "step": 144200 }, { "epoch": 0.2691865388069485, "grad_norm": 0.8201596736907959, "learning_rate": 0.00019938330657472486, "loss": 4.6992, "step": 144250 }, { "epoch": 0.26927984436632696, "grad_norm": 1.0226540565490723, "learning_rate": 0.00019938284065455225, "loss": 4.8901, "step": 144300 }, { "epoch": 0.26937314992570544, "grad_norm": 0.8372309803962708, "learning_rate": 0.0001993823745589864, "loss": 4.776, "step": 144350 }, { "epoch": 0.26946645548508397, "grad_norm": 0.6700859069824219, "learning_rate": 0.0001993819082880282, "loss": 4.5808, "step": 144400 }, { "epoch": 0.26955976104446244, "grad_norm": 0.9853873252868652, "learning_rate": 0.00019938144184167836, "loss": 4.7041, "step": 144450 }, { "epoch": 0.2696530666038409, "grad_norm": 1.0174555778503418, "learning_rate": 0.0001993809752199378, "loss": 4.6368, "step": 144500 }, { "epoch": 0.2697463721632194, "grad_norm": 1.0446439981460571, "learning_rate": 0.00019938050842280728, "loss": 4.8199, "step": 144550 }, { "epoch": 0.2698396777225979, "grad_norm": 1.1562541723251343, "learning_rate": 0.00019938004145028768, "loss": 4.8258, "step": 144600 }, { "epoch": 0.2699329832819764, "grad_norm": 1.0433145761489868, "learning_rate": 0.00019937957430237976, "loss": 4.5957, "step": 144650 }, { "epoch": 0.2700262888413549, "grad_norm": 0.8568206429481506, "learning_rate": 0.0001993791069790844, "loss": 4.6801, "step": 144700 }, { "epoch": 0.2701195944007334, "grad_norm": 1.0583845376968384, "learning_rate": 0.0001993786394804024, "loss": 4.7475, "step": 144750 }, { "epoch": 0.2702128999601119, "grad_norm": 1.1356602907180786, "learning_rate": 0.00019937817180633457, "loss": 4.5852, "step": 144800 }, { "epoch": 0.27030620551949036, "grad_norm": 1.091326117515564, "learning_rate": 0.00019937770395688179, "loss": 4.6252, "step": 144850 }, { "epoch": 0.27039951107886884, "grad_norm": 1.135559320449829, "learning_rate": 0.0001993772359320448, "loss": 4.7506, "step": 144900 }, { "epoch": 0.27049281663824737, "grad_norm": 0.910708487033844, "learning_rate": 0.0001993767677318245, "loss": 4.6904, "step": 144950 }, { "epoch": 0.27058612219762584, "grad_norm": 1.0501621961593628, "learning_rate": 0.0001993762993562217, "loss": 4.8702, "step": 145000 }, { "epoch": 0.2706794277570043, "grad_norm": 0.691247820854187, "learning_rate": 0.0001993758308052372, "loss": 4.666, "step": 145050 }, { "epoch": 0.2707727333163828, "grad_norm": 0.9486430287361145, "learning_rate": 0.00019937536207887183, "loss": 4.7462, "step": 145100 }, { "epoch": 0.27086603887576133, "grad_norm": 0.9459395408630371, "learning_rate": 0.00019937489317712646, "loss": 4.8276, "step": 145150 }, { "epoch": 0.2709593444351398, "grad_norm": 1.2889015674591064, "learning_rate": 0.00019937442410000185, "loss": 4.6273, "step": 145200 }, { "epoch": 0.2710526499945183, "grad_norm": 1.114570140838623, "learning_rate": 0.00019937395484749892, "loss": 4.5248, "step": 145250 }, { "epoch": 0.2711459555538968, "grad_norm": 1.1005711555480957, "learning_rate": 0.00019937348541961843, "loss": 4.6502, "step": 145300 }, { "epoch": 0.2712392611132753, "grad_norm": 0.806015133857727, "learning_rate": 0.00019937301581636122, "loss": 4.6222, "step": 145350 }, { "epoch": 0.27133256667265376, "grad_norm": 0.9318636059761047, "learning_rate": 0.00019937254603772812, "loss": 4.5528, "step": 145400 }, { "epoch": 0.27142587223203224, "grad_norm": 0.697272539138794, "learning_rate": 0.00019937207608371996, "loss": 4.7631, "step": 145450 }, { "epoch": 0.27151917779141077, "grad_norm": 0.9297815561294556, "learning_rate": 0.00019937160595433757, "loss": 4.5448, "step": 145500 }, { "epoch": 0.27161248335078925, "grad_norm": 1.144652247428894, "learning_rate": 0.00019937113564958176, "loss": 4.6709, "step": 145550 }, { "epoch": 0.2717057889101677, "grad_norm": 0.9392854571342468, "learning_rate": 0.0001993706651694534, "loss": 4.7134, "step": 145600 }, { "epoch": 0.27179909446954625, "grad_norm": 0.9400250315666199, "learning_rate": 0.00019937019451395332, "loss": 4.5493, "step": 145650 }, { "epoch": 0.27189240002892473, "grad_norm": 1.1566085815429688, "learning_rate": 0.0001993697236830823, "loss": 4.939, "step": 145700 }, { "epoch": 0.2719857055883032, "grad_norm": 0.7615442872047424, "learning_rate": 0.00019936925267684122, "loss": 4.7826, "step": 145750 }, { "epoch": 0.2720790111476817, "grad_norm": 0.8968701362609863, "learning_rate": 0.0001993687814952309, "loss": 4.5258, "step": 145800 }, { "epoch": 0.2721723167070602, "grad_norm": 1.0293611288070679, "learning_rate": 0.00019936831013825214, "loss": 4.6597, "step": 145850 }, { "epoch": 0.2722656222664387, "grad_norm": 0.9446884989738464, "learning_rate": 0.00019936783860590582, "loss": 4.6033, "step": 145900 }, { "epoch": 0.27235892782581717, "grad_norm": 1.0046056509017944, "learning_rate": 0.00019936736689819275, "loss": 4.5402, "step": 145950 }, { "epoch": 0.2724522333851957, "grad_norm": 0.7611010074615479, "learning_rate": 0.00019936689501511375, "loss": 4.6937, "step": 146000 }, { "epoch": 0.2725455389445742, "grad_norm": 0.9381779432296753, "learning_rate": 0.00019936642295666966, "loss": 4.5509, "step": 146050 }, { "epoch": 0.27263884450395265, "grad_norm": 0.862014889717102, "learning_rate": 0.00019936595072286134, "loss": 4.6046, "step": 146100 }, { "epoch": 0.2727321500633311, "grad_norm": 1.1331309080123901, "learning_rate": 0.00019936547831368956, "loss": 4.4339, "step": 146150 }, { "epoch": 0.27282545562270966, "grad_norm": 0.9985886812210083, "learning_rate": 0.00019936500572915523, "loss": 4.7718, "step": 146200 }, { "epoch": 0.27291876118208813, "grad_norm": 1.0175446271896362, "learning_rate": 0.00019936453296925914, "loss": 4.5348, "step": 146250 }, { "epoch": 0.2730120667414666, "grad_norm": 1.1573448181152344, "learning_rate": 0.00019936406003400215, "loss": 4.5544, "step": 146300 }, { "epoch": 0.27310537230084514, "grad_norm": 0.7975154519081116, "learning_rate": 0.00019936358692338506, "loss": 4.6486, "step": 146350 }, { "epoch": 0.2731986778602236, "grad_norm": 0.9577890038490295, "learning_rate": 0.0001993631136374087, "loss": 4.661, "step": 146400 }, { "epoch": 0.2732919834196021, "grad_norm": 1.0218697786331177, "learning_rate": 0.00019936264017607398, "loss": 4.6003, "step": 146450 }, { "epoch": 0.27338528897898057, "grad_norm": 1.3306013345718384, "learning_rate": 0.00019936216653938162, "loss": 4.7019, "step": 146500 }, { "epoch": 0.2734785945383591, "grad_norm": 1.0977168083190918, "learning_rate": 0.00019936169272733256, "loss": 4.7338, "step": 146550 }, { "epoch": 0.2735719000977376, "grad_norm": 1.1367696523666382, "learning_rate": 0.0001993612187399276, "loss": 4.9661, "step": 146600 }, { "epoch": 0.27366520565711605, "grad_norm": 1.0126125812530518, "learning_rate": 0.00019936074457716757, "loss": 4.7346, "step": 146650 }, { "epoch": 0.2737585112164946, "grad_norm": 0.9203944802284241, "learning_rate": 0.0001993602702390533, "loss": 4.7422, "step": 146700 }, { "epoch": 0.27385181677587306, "grad_norm": 0.8814552426338196, "learning_rate": 0.00019935979572558563, "loss": 4.5732, "step": 146750 }, { "epoch": 0.27394512233525153, "grad_norm": 0.9684024453163147, "learning_rate": 0.00019935932103676542, "loss": 4.6788, "step": 146800 }, { "epoch": 0.27403842789463, "grad_norm": 0.8495442271232605, "learning_rate": 0.00019935884617259349, "loss": 4.4585, "step": 146850 }, { "epoch": 0.27413173345400854, "grad_norm": 0.5965443849563599, "learning_rate": 0.00019935837113307066, "loss": 4.4931, "step": 146900 }, { "epoch": 0.274225039013387, "grad_norm": 0.9253427386283875, "learning_rate": 0.00019935789591819778, "loss": 4.7332, "step": 146950 }, { "epoch": 0.2743183445727655, "grad_norm": 0.959456205368042, "learning_rate": 0.0001993574205279757, "loss": 4.6347, "step": 147000 }, { "epoch": 0.2743183445727655, "eval_loss": 4.84540319442749, "eval_runtime": 228.6502, "eval_samples_per_second": 11.406, "eval_steps_per_second": 11.406, "eval_tts_loss": 7.421910372788395, "step": 147000 }, { "epoch": 0.274411650132144, "grad_norm": 0.8804351687431335, "learning_rate": 0.0001993569449624053, "loss": 4.5913, "step": 147050 }, { "epoch": 0.2745049556915225, "grad_norm": 1.257071852684021, "learning_rate": 0.00019935646922148733, "loss": 4.6649, "step": 147100 }, { "epoch": 0.274598261250901, "grad_norm": 1.0313574075698853, "learning_rate": 0.0001993559933052227, "loss": 4.6562, "step": 147150 }, { "epoch": 0.27469156681027945, "grad_norm": 0.9633307456970215, "learning_rate": 0.0001993555172136122, "loss": 4.6321, "step": 147200 }, { "epoch": 0.274784872369658, "grad_norm": 0.7918700575828552, "learning_rate": 0.0001993550409466567, "loss": 4.7728, "step": 147250 }, { "epoch": 0.27487817792903646, "grad_norm": 0.6969190239906311, "learning_rate": 0.00019935456450435703, "loss": 4.7156, "step": 147300 }, { "epoch": 0.27497148348841494, "grad_norm": 1.4013220071792603, "learning_rate": 0.00019935408788671405, "loss": 4.5602, "step": 147350 }, { "epoch": 0.27506478904779347, "grad_norm": 1.0779790878295898, "learning_rate": 0.00019935361109372859, "loss": 4.7721, "step": 147400 }, { "epoch": 0.27515809460717194, "grad_norm": 1.0877084732055664, "learning_rate": 0.00019935313412540147, "loss": 4.7129, "step": 147450 }, { "epoch": 0.2752514001665504, "grad_norm": 1.1705236434936523, "learning_rate": 0.00019935265698173358, "loss": 4.7317, "step": 147500 }, { "epoch": 0.2753447057259289, "grad_norm": 1.0509397983551025, "learning_rate": 0.0001993521796627257, "loss": 4.7795, "step": 147550 }, { "epoch": 0.2754380112853074, "grad_norm": 1.1017162799835205, "learning_rate": 0.00019935170216837873, "loss": 4.7745, "step": 147600 }, { "epoch": 0.2755313168446859, "grad_norm": 0.9443593621253967, "learning_rate": 0.0001993512244986935, "loss": 4.6332, "step": 147650 }, { "epoch": 0.2756246224040644, "grad_norm": 1.0674967765808105, "learning_rate": 0.00019935074665367077, "loss": 4.721, "step": 147700 }, { "epoch": 0.27571792796344285, "grad_norm": 0.976577877998352, "learning_rate": 0.00019935026863331153, "loss": 4.5112, "step": 147750 }, { "epoch": 0.2758112335228214, "grad_norm": 0.9510547518730164, "learning_rate": 0.0001993497904376165, "loss": 4.8486, "step": 147800 }, { "epoch": 0.27590453908219986, "grad_norm": 1.0854049921035767, "learning_rate": 0.00019934931206658658, "loss": 4.4141, "step": 147850 }, { "epoch": 0.27599784464157834, "grad_norm": 1.5401407480239868, "learning_rate": 0.0001993488335202226, "loss": 4.741, "step": 147900 }, { "epoch": 0.27609115020095687, "grad_norm": 1.0566352605819702, "learning_rate": 0.00019934835479852546, "loss": 4.9034, "step": 147950 }, { "epoch": 0.27618445576033535, "grad_norm": 1.0547597408294678, "learning_rate": 0.0001993478759014959, "loss": 4.6906, "step": 148000 }, { "epoch": 0.2762777613197138, "grad_norm": 0.8163520097732544, "learning_rate": 0.00019934739682913485, "loss": 4.8077, "step": 148050 }, { "epoch": 0.2763710668790923, "grad_norm": 1.1130653619766235, "learning_rate": 0.00019934691758144308, "loss": 4.7433, "step": 148100 }, { "epoch": 0.27646437243847083, "grad_norm": 0.769314706325531, "learning_rate": 0.00019934643815842153, "loss": 4.6217, "step": 148150 }, { "epoch": 0.2765576779978493, "grad_norm": 1.1348406076431274, "learning_rate": 0.000199345958560071, "loss": 4.5468, "step": 148200 }, { "epoch": 0.2766509835572278, "grad_norm": 0.9563210010528564, "learning_rate": 0.00019934547878639228, "loss": 4.6116, "step": 148250 }, { "epoch": 0.2767442891166063, "grad_norm": 0.9107294082641602, "learning_rate": 0.00019934499883738632, "loss": 4.8486, "step": 148300 }, { "epoch": 0.2768375946759848, "grad_norm": 1.0701485872268677, "learning_rate": 0.0001993445187130539, "loss": 4.7653, "step": 148350 }, { "epoch": 0.27693090023536326, "grad_norm": 1.1094465255737305, "learning_rate": 0.00019934403841339587, "loss": 4.451, "step": 148400 }, { "epoch": 0.27702420579474174, "grad_norm": 1.283995270729065, "learning_rate": 0.00019934355793841313, "loss": 4.8218, "step": 148450 }, { "epoch": 0.27711751135412027, "grad_norm": 0.9863646030426025, "learning_rate": 0.00019934307728810646, "loss": 4.831, "step": 148500 }, { "epoch": 0.27721081691349875, "grad_norm": 1.0133435726165771, "learning_rate": 0.00019934259646247674, "loss": 4.4905, "step": 148550 }, { "epoch": 0.2773041224728772, "grad_norm": 0.9534834623336792, "learning_rate": 0.00019934211546152484, "loss": 4.7916, "step": 148600 }, { "epoch": 0.27739742803225576, "grad_norm": 1.2967839241027832, "learning_rate": 0.00019934163428525156, "loss": 4.7967, "step": 148650 }, { "epoch": 0.27749073359163423, "grad_norm": 0.869767427444458, "learning_rate": 0.00019934115293365778, "loss": 4.7234, "step": 148700 }, { "epoch": 0.2775840391510127, "grad_norm": 0.9812904000282288, "learning_rate": 0.00019934067140674436, "loss": 4.4395, "step": 148750 }, { "epoch": 0.2776773447103912, "grad_norm": 0.9434531331062317, "learning_rate": 0.0001993401897045121, "loss": 4.6072, "step": 148800 }, { "epoch": 0.2777706502697697, "grad_norm": 0.6807860732078552, "learning_rate": 0.00019933970782696193, "loss": 4.6445, "step": 148850 }, { "epoch": 0.2778639558291482, "grad_norm": 0.8833329677581787, "learning_rate": 0.00019933922577409463, "loss": 4.4624, "step": 148900 }, { "epoch": 0.27795726138852667, "grad_norm": 1.2914021015167236, "learning_rate": 0.0001993387435459111, "loss": 4.8027, "step": 148950 }, { "epoch": 0.2780505669479052, "grad_norm": 1.0718110799789429, "learning_rate": 0.00019933826114241213, "loss": 4.6699, "step": 149000 }, { "epoch": 0.2781438725072837, "grad_norm": 0.8196322321891785, "learning_rate": 0.00019933777856359863, "loss": 4.5286, "step": 149050 }, { "epoch": 0.27823717806666215, "grad_norm": 1.0307577848434448, "learning_rate": 0.00019933729580947143, "loss": 4.6294, "step": 149100 }, { "epoch": 0.2783304836260406, "grad_norm": 0.874259889125824, "learning_rate": 0.00019933681288003137, "loss": 4.6817, "step": 149150 }, { "epoch": 0.27842378918541916, "grad_norm": 0.8164685964584351, "learning_rate": 0.0001993363297752793, "loss": 4.6156, "step": 149200 }, { "epoch": 0.27851709474479763, "grad_norm": 1.4516046047210693, "learning_rate": 0.0001993358464952161, "loss": 4.8161, "step": 149250 }, { "epoch": 0.2786104003041761, "grad_norm": 1.4525372982025146, "learning_rate": 0.00019933536303984263, "loss": 4.6891, "step": 149300 }, { "epoch": 0.27870370586355464, "grad_norm": 1.0817906856536865, "learning_rate": 0.0001993348794091597, "loss": 4.9178, "step": 149350 }, { "epoch": 0.2787970114229331, "grad_norm": 0.8734576106071472, "learning_rate": 0.0001993343956031682, "loss": 4.7184, "step": 149400 }, { "epoch": 0.2788903169823116, "grad_norm": 0.8996809720993042, "learning_rate": 0.00019933391162186897, "loss": 4.5792, "step": 149450 }, { "epoch": 0.27898362254169007, "grad_norm": 1.0781534910202026, "learning_rate": 0.00019933342746526286, "loss": 4.6787, "step": 149500 }, { "epoch": 0.2790769281010686, "grad_norm": 0.9181458353996277, "learning_rate": 0.00019933294313335072, "loss": 4.6463, "step": 149550 }, { "epoch": 0.2791702336604471, "grad_norm": 1.0423184633255005, "learning_rate": 0.00019933245862613344, "loss": 4.5832, "step": 149600 }, { "epoch": 0.27926353921982555, "grad_norm": 0.9878393411636353, "learning_rate": 0.00019933197394361181, "loss": 4.5294, "step": 149650 }, { "epoch": 0.2793568447792041, "grad_norm": 0.9291845560073853, "learning_rate": 0.00019933148908578678, "loss": 4.7486, "step": 149700 }, { "epoch": 0.27945015033858256, "grad_norm": 1.0207691192626953, "learning_rate": 0.0001993310040526591, "loss": 4.7574, "step": 149750 }, { "epoch": 0.27954345589796104, "grad_norm": 1.0741710662841797, "learning_rate": 0.00019933051884422968, "loss": 4.6548, "step": 149800 }, { "epoch": 0.2796367614573395, "grad_norm": 1.0750080347061157, "learning_rate": 0.0001993300334604994, "loss": 4.6506, "step": 149850 }, { "epoch": 0.27973006701671804, "grad_norm": 1.0452051162719727, "learning_rate": 0.00019932954790146908, "loss": 4.7486, "step": 149900 }, { "epoch": 0.2798233725760965, "grad_norm": 0.8522201776504517, "learning_rate": 0.0001993290621671396, "loss": 4.6639, "step": 149950 }, { "epoch": 0.279916678135475, "grad_norm": 0.8773214817047119, "learning_rate": 0.00019932857625751176, "loss": 4.7012, "step": 150000 }, { "epoch": 0.279916678135475, "eval_loss": 4.83031702041626, "eval_runtime": 229.2208, "eval_samples_per_second": 11.378, "eval_steps_per_second": 11.378, "eval_tts_loss": 7.386615539569568, "step": 150000 }, { "epoch": 0.2800099836948535, "grad_norm": 0.9581308364868164, "learning_rate": 0.00019932809017258647, "loss": 4.8594, "step": 150050 }, { "epoch": 0.280103289254232, "grad_norm": 0.9009203314781189, "learning_rate": 0.00019932760391236457, "loss": 4.6758, "step": 150100 }, { "epoch": 0.2801965948136105, "grad_norm": 0.9050132632255554, "learning_rate": 0.00019932711747684695, "loss": 4.5833, "step": 150150 }, { "epoch": 0.28028990037298895, "grad_norm": 1.0961129665374756, "learning_rate": 0.00019932663086603446, "loss": 4.7339, "step": 150200 }, { "epoch": 0.2803832059323675, "grad_norm": 1.4378759860992432, "learning_rate": 0.00019932614407992792, "loss": 4.8097, "step": 150250 }, { "epoch": 0.28047651149174596, "grad_norm": 0.941024899482727, "learning_rate": 0.0001993256571185282, "loss": 4.5834, "step": 150300 }, { "epoch": 0.28056981705112444, "grad_norm": 0.8255406618118286, "learning_rate": 0.0001993251699818362, "loss": 4.5948, "step": 150350 }, { "epoch": 0.2806631226105029, "grad_norm": 0.9121527671813965, "learning_rate": 0.00019932468266985274, "loss": 4.8421, "step": 150400 }, { "epoch": 0.28075642816988144, "grad_norm": 1.2012547254562378, "learning_rate": 0.0001993241951825787, "loss": 4.7063, "step": 150450 }, { "epoch": 0.2808497337292599, "grad_norm": 1.0221003293991089, "learning_rate": 0.0001993237075200149, "loss": 4.5613, "step": 150500 }, { "epoch": 0.2809430392886384, "grad_norm": 0.6838630437850952, "learning_rate": 0.00019932321968216225, "loss": 4.53, "step": 150550 }, { "epoch": 0.28103634484801693, "grad_norm": 0.9109449982643127, "learning_rate": 0.0001993227316690216, "loss": 4.7778, "step": 150600 }, { "epoch": 0.2811296504073954, "grad_norm": 1.0373647212982178, "learning_rate": 0.00019932224348059383, "loss": 4.7198, "step": 150650 }, { "epoch": 0.2812229559667739, "grad_norm": 1.202540397644043, "learning_rate": 0.00019932175511687975, "loss": 4.5706, "step": 150700 }, { "epoch": 0.28131626152615236, "grad_norm": 1.1028730869293213, "learning_rate": 0.00019932126657788024, "loss": 4.6268, "step": 150750 }, { "epoch": 0.2814095670855309, "grad_norm": 0.88832688331604, "learning_rate": 0.00019932077786359616, "loss": 4.7166, "step": 150800 }, { "epoch": 0.28150287264490936, "grad_norm": 1.0212421417236328, "learning_rate": 0.00019932028897402844, "loss": 4.7143, "step": 150850 }, { "epoch": 0.28159617820428784, "grad_norm": 0.8306617736816406, "learning_rate": 0.00019931979990917782, "loss": 4.7177, "step": 150900 }, { "epoch": 0.28168948376366637, "grad_norm": 1.1900157928466797, "learning_rate": 0.00019931931066904528, "loss": 4.6889, "step": 150950 }, { "epoch": 0.28178278932304485, "grad_norm": 0.8949201703071594, "learning_rate": 0.00019931882125363162, "loss": 4.72, "step": 151000 }, { "epoch": 0.2818760948824233, "grad_norm": 1.0268856287002563, "learning_rate": 0.00019931833166293768, "loss": 4.7837, "step": 151050 }, { "epoch": 0.2819694004418018, "grad_norm": 1.0028334856033325, "learning_rate": 0.00019931784189696438, "loss": 4.4346, "step": 151100 }, { "epoch": 0.28206270600118033, "grad_norm": 0.6958338618278503, "learning_rate": 0.00019931735195571258, "loss": 4.599, "step": 151150 }, { "epoch": 0.2821560115605588, "grad_norm": 1.0011975765228271, "learning_rate": 0.00019931686183918314, "loss": 4.6019, "step": 151200 }, { "epoch": 0.2822493171199373, "grad_norm": 1.0134965181350708, "learning_rate": 0.00019931637154737688, "loss": 4.5689, "step": 151250 }, { "epoch": 0.2823426226793158, "grad_norm": 1.1758054494857788, "learning_rate": 0.00019931588108029473, "loss": 4.863, "step": 151300 }, { "epoch": 0.2824359282386943, "grad_norm": 0.6513984203338623, "learning_rate": 0.0001993153904379375, "loss": 4.8059, "step": 151350 }, { "epoch": 0.28252923379807277, "grad_norm": 0.9191519618034363, "learning_rate": 0.0001993148996203061, "loss": 4.5621, "step": 151400 }, { "epoch": 0.28262253935745124, "grad_norm": 1.069206953048706, "learning_rate": 0.00019931440862740136, "loss": 4.54, "step": 151450 }, { "epoch": 0.2827158449168298, "grad_norm": 0.8649135828018188, "learning_rate": 0.00019931391745922416, "loss": 4.4516, "step": 151500 }, { "epoch": 0.28280915047620825, "grad_norm": 0.8605431914329529, "learning_rate": 0.0001993134261157754, "loss": 4.6004, "step": 151550 }, { "epoch": 0.2829024560355867, "grad_norm": 1.1754326820373535, "learning_rate": 0.00019931293459705588, "loss": 4.7162, "step": 151600 }, { "epoch": 0.28299576159496526, "grad_norm": 0.9741180539131165, "learning_rate": 0.0001993124429030665, "loss": 4.8181, "step": 151650 }, { "epoch": 0.28308906715434373, "grad_norm": 1.0868324041366577, "learning_rate": 0.00019931195103380819, "loss": 4.602, "step": 151700 }, { "epoch": 0.2831823727137222, "grad_norm": 0.9874230027198792, "learning_rate": 0.00019931145898928173, "loss": 4.6277, "step": 151750 }, { "epoch": 0.2832756782731007, "grad_norm": 1.13186776638031, "learning_rate": 0.00019931096676948802, "loss": 4.5465, "step": 151800 }, { "epoch": 0.2833689838324792, "grad_norm": 0.8239442110061646, "learning_rate": 0.00019931047437442794, "loss": 4.7412, "step": 151850 }, { "epoch": 0.2834622893918577, "grad_norm": 1.0195013284683228, "learning_rate": 0.00019930998180410232, "loss": 4.4785, "step": 151900 }, { "epoch": 0.28355559495123617, "grad_norm": 1.3029369115829468, "learning_rate": 0.00019930948905851206, "loss": 4.8227, "step": 151950 }, { "epoch": 0.2836489005106147, "grad_norm": 0.8565980792045593, "learning_rate": 0.00019930899613765805, "loss": 4.5811, "step": 152000 }, { "epoch": 0.2837422060699932, "grad_norm": 0.9807513356208801, "learning_rate": 0.0001993085030415411, "loss": 4.733, "step": 152050 }, { "epoch": 0.28383551162937165, "grad_norm": 1.0454181432724, "learning_rate": 0.00019930800977016215, "loss": 4.703, "step": 152100 }, { "epoch": 0.2839288171887501, "grad_norm": 1.0807571411132812, "learning_rate": 0.00019930751632352202, "loss": 4.4999, "step": 152150 }, { "epoch": 0.28402212274812866, "grad_norm": 1.1262049674987793, "learning_rate": 0.0001993070227016216, "loss": 4.6597, "step": 152200 }, { "epoch": 0.28411542830750713, "grad_norm": 0.8939054012298584, "learning_rate": 0.00019930652890446176, "loss": 4.6779, "step": 152250 }, { "epoch": 0.2842087338668856, "grad_norm": 0.9293799996376038, "learning_rate": 0.00019930603493204337, "loss": 4.551, "step": 152300 }, { "epoch": 0.28430203942626414, "grad_norm": 1.1074103116989136, "learning_rate": 0.0001993055407843673, "loss": 4.805, "step": 152350 }, { "epoch": 0.2843953449856426, "grad_norm": 1.147536039352417, "learning_rate": 0.0001993050464614344, "loss": 4.7777, "step": 152400 }, { "epoch": 0.2844886505450211, "grad_norm": 0.8999912142753601, "learning_rate": 0.00019930455196324562, "loss": 4.7226, "step": 152450 }, { "epoch": 0.28458195610439957, "grad_norm": 0.8882901072502136, "learning_rate": 0.00019930405728980176, "loss": 4.5275, "step": 152500 }, { "epoch": 0.2846752616637781, "grad_norm": 0.958745539188385, "learning_rate": 0.00019930356244110369, "loss": 4.5713, "step": 152550 }, { "epoch": 0.2847685672231566, "grad_norm": 0.8023204803466797, "learning_rate": 0.00019930306741715233, "loss": 4.6426, "step": 152600 }, { "epoch": 0.28486187278253505, "grad_norm": 0.7979447245597839, "learning_rate": 0.0001993025722179485, "loss": 4.81, "step": 152650 }, { "epoch": 0.2849551783419136, "grad_norm": 0.8460389375686646, "learning_rate": 0.00019930207684349313, "loss": 4.7702, "step": 152700 }, { "epoch": 0.28504848390129206, "grad_norm": 0.905491054058075, "learning_rate": 0.00019930158129378707, "loss": 4.6398, "step": 152750 }, { "epoch": 0.28514178946067054, "grad_norm": 0.5452645421028137, "learning_rate": 0.00019930108556883116, "loss": 4.4575, "step": 152800 }, { "epoch": 0.285235095020049, "grad_norm": 1.016485333442688, "learning_rate": 0.00019930058966862634, "loss": 4.6945, "step": 152850 }, { "epoch": 0.28532840057942754, "grad_norm": 0.6815197467803955, "learning_rate": 0.00019930009359317343, "loss": 4.6492, "step": 152900 }, { "epoch": 0.285421706138806, "grad_norm": 0.9308417439460754, "learning_rate": 0.00019929959734247331, "loss": 4.6306, "step": 152950 }, { "epoch": 0.2855150116981845, "grad_norm": 0.9671070575714111, "learning_rate": 0.0001992991009165269, "loss": 4.6504, "step": 153000 }, { "epoch": 0.2855150116981845, "eval_loss": 4.835600852966309, "eval_runtime": 230.5993, "eval_samples_per_second": 11.31, "eval_steps_per_second": 11.31, "eval_tts_loss": 7.43258083978685, "step": 153000 }, { "epoch": 0.28560831725756297, "grad_norm": 0.8776199221611023, "learning_rate": 0.00019929860431533504, "loss": 4.5989, "step": 153050 }, { "epoch": 0.2857016228169415, "grad_norm": 0.8226851224899292, "learning_rate": 0.00019929810753889865, "loss": 4.3491, "step": 153100 }, { "epoch": 0.28579492837632, "grad_norm": 1.3178420066833496, "learning_rate": 0.00019929761058721853, "loss": 4.5217, "step": 153150 }, { "epoch": 0.28588823393569845, "grad_norm": 1.0944714546203613, "learning_rate": 0.0001992971134602956, "loss": 4.7149, "step": 153200 }, { "epoch": 0.285981539495077, "grad_norm": 0.8839026689529419, "learning_rate": 0.00019929661615813077, "loss": 4.5549, "step": 153250 }, { "epoch": 0.28607484505445546, "grad_norm": 0.9436131119728088, "learning_rate": 0.00019929611868072487, "loss": 4.7396, "step": 153300 }, { "epoch": 0.28616815061383394, "grad_norm": 0.8555207252502441, "learning_rate": 0.00019929562102807878, "loss": 4.6732, "step": 153350 }, { "epoch": 0.2862614561732124, "grad_norm": 0.8614400625228882, "learning_rate": 0.0001992951232001934, "loss": 4.5358, "step": 153400 }, { "epoch": 0.28635476173259095, "grad_norm": 0.7893537282943726, "learning_rate": 0.00019929462519706958, "loss": 4.7613, "step": 153450 }, { "epoch": 0.2864480672919694, "grad_norm": 1.0133888721466064, "learning_rate": 0.00019929412701870822, "loss": 4.7695, "step": 153500 }, { "epoch": 0.2865413728513479, "grad_norm": 1.0339868068695068, "learning_rate": 0.00019929362866511022, "loss": 4.6753, "step": 153550 }, { "epoch": 0.28663467841072643, "grad_norm": 0.8841448426246643, "learning_rate": 0.00019929313013627642, "loss": 4.5938, "step": 153600 }, { "epoch": 0.2867279839701049, "grad_norm": 0.8612989783287048, "learning_rate": 0.00019929263143220774, "loss": 4.4805, "step": 153650 }, { "epoch": 0.2868212895294834, "grad_norm": 0.9694012403488159, "learning_rate": 0.000199292132552905, "loss": 4.8494, "step": 153700 }, { "epoch": 0.28691459508886186, "grad_norm": 1.2090984582901, "learning_rate": 0.00019929163349836915, "loss": 4.7134, "step": 153750 }, { "epoch": 0.2870079006482404, "grad_norm": 1.0495941638946533, "learning_rate": 0.00019929113426860103, "loss": 4.7221, "step": 153800 }, { "epoch": 0.28710120620761886, "grad_norm": 0.8493571281433105, "learning_rate": 0.00019929063486360152, "loss": 4.4686, "step": 153850 }, { "epoch": 0.28719451176699734, "grad_norm": 1.2186065912246704, "learning_rate": 0.0001992901352833715, "loss": 4.7176, "step": 153900 }, { "epoch": 0.28728781732637587, "grad_norm": 1.0181117057800293, "learning_rate": 0.00019928963552791188, "loss": 4.484, "step": 153950 }, { "epoch": 0.28738112288575435, "grad_norm": 1.2028471231460571, "learning_rate": 0.0001992891355972235, "loss": 4.8875, "step": 154000 }, { "epoch": 0.2874744284451328, "grad_norm": 1.1655298471450806, "learning_rate": 0.00019928863549130733, "loss": 4.5263, "step": 154050 }, { "epoch": 0.2875677340045113, "grad_norm": 1.0849800109863281, "learning_rate": 0.00019928813521016413, "loss": 4.4706, "step": 154100 }, { "epoch": 0.28766103956388983, "grad_norm": 0.9207818508148193, "learning_rate": 0.00019928763475379486, "loss": 4.605, "step": 154150 }, { "epoch": 0.2877543451232683, "grad_norm": 0.8862490653991699, "learning_rate": 0.00019928713412220036, "loss": 4.5169, "step": 154200 }, { "epoch": 0.2878476506826468, "grad_norm": 1.0077502727508545, "learning_rate": 0.00019928663331538154, "loss": 4.7856, "step": 154250 }, { "epoch": 0.2879409562420253, "grad_norm": 1.1006149053573608, "learning_rate": 0.00019928613233333934, "loss": 4.7136, "step": 154300 }, { "epoch": 0.2880342618014038, "grad_norm": 1.0672188997268677, "learning_rate": 0.00019928563117607454, "loss": 4.6588, "step": 154350 }, { "epoch": 0.28812756736078227, "grad_norm": 1.0344702005386353, "learning_rate": 0.00019928512984358808, "loss": 4.5211, "step": 154400 }, { "epoch": 0.28822087292016074, "grad_norm": 0.9514538645744324, "learning_rate": 0.00019928462833588081, "loss": 4.586, "step": 154450 }, { "epoch": 0.2883141784795393, "grad_norm": 0.9031883478164673, "learning_rate": 0.00019928412665295368, "loss": 4.6923, "step": 154500 }, { "epoch": 0.28840748403891775, "grad_norm": 0.9801226258277893, "learning_rate": 0.00019928362479480752, "loss": 4.6889, "step": 154550 }, { "epoch": 0.2885007895982962, "grad_norm": 0.9183834791183472, "learning_rate": 0.00019928312276144326, "loss": 4.5817, "step": 154600 }, { "epoch": 0.28859409515767476, "grad_norm": 0.8800172209739685, "learning_rate": 0.0001992826205528617, "loss": 4.6275, "step": 154650 }, { "epoch": 0.28868740071705323, "grad_norm": 1.0279831886291504, "learning_rate": 0.00019928211816906384, "loss": 4.6246, "step": 154700 }, { "epoch": 0.2887807062764317, "grad_norm": 0.9106327295303345, "learning_rate": 0.00019928161561005045, "loss": 4.5679, "step": 154750 }, { "epoch": 0.2888740118358102, "grad_norm": 0.9070572257041931, "learning_rate": 0.0001992811128758225, "loss": 4.6825, "step": 154800 }, { "epoch": 0.2889673173951887, "grad_norm": 0.9388306140899658, "learning_rate": 0.00019928060996638093, "loss": 4.4394, "step": 154850 }, { "epoch": 0.2890606229545672, "grad_norm": 0.9213298559188843, "learning_rate": 0.00019928010688172645, "loss": 4.6287, "step": 154900 }, { "epoch": 0.28915392851394567, "grad_norm": 0.9049386382102966, "learning_rate": 0.00019927960362186008, "loss": 4.7666, "step": 154950 }, { "epoch": 0.2892472340733242, "grad_norm": 0.9011629223823547, "learning_rate": 0.0001992791001867827, "loss": 4.5634, "step": 155000 }, { "epoch": 0.2893405396327027, "grad_norm": 1.0195951461791992, "learning_rate": 0.00019927859657649518, "loss": 4.6655, "step": 155050 }, { "epoch": 0.28943384519208115, "grad_norm": 0.7028415203094482, "learning_rate": 0.00019927809279099838, "loss": 4.4955, "step": 155100 }, { "epoch": 0.2895271507514596, "grad_norm": 0.9944257140159607, "learning_rate": 0.0001992775888302932, "loss": 4.9458, "step": 155150 }, { "epoch": 0.28962045631083816, "grad_norm": 1.0102462768554688, "learning_rate": 0.0001992770846943806, "loss": 4.748, "step": 155200 }, { "epoch": 0.28971376187021664, "grad_norm": 0.8883141875267029, "learning_rate": 0.00019927658038326138, "loss": 4.4985, "step": 155250 }, { "epoch": 0.2898070674295951, "grad_norm": 0.8198468089103699, "learning_rate": 0.0001992760758969365, "loss": 4.6353, "step": 155300 }, { "epoch": 0.28990037298897364, "grad_norm": 0.7499061226844788, "learning_rate": 0.00019927557123540677, "loss": 4.5579, "step": 155350 }, { "epoch": 0.2899936785483521, "grad_norm": 0.7764875888824463, "learning_rate": 0.0001992750663986731, "loss": 4.4778, "step": 155400 }, { "epoch": 0.2900869841077306, "grad_norm": 0.8831614255905151, "learning_rate": 0.00019927456138673646, "loss": 4.403, "step": 155450 }, { "epoch": 0.29018028966710907, "grad_norm": 1.0157040357589722, "learning_rate": 0.00019927405619959767, "loss": 4.6915, "step": 155500 }, { "epoch": 0.2902735952264876, "grad_norm": 1.1390964984893799, "learning_rate": 0.00019927355083725762, "loss": 4.5186, "step": 155550 }, { "epoch": 0.2903669007858661, "grad_norm": 1.154977560043335, "learning_rate": 0.00019927304529971727, "loss": 4.7172, "step": 155600 }, { "epoch": 0.29046020634524455, "grad_norm": 1.095638632774353, "learning_rate": 0.00019927253958697743, "loss": 4.8124, "step": 155650 }, { "epoch": 0.29055351190462303, "grad_norm": 0.9511516094207764, "learning_rate": 0.00019927203369903901, "loss": 4.4195, "step": 155700 }, { "epoch": 0.29064681746400156, "grad_norm": 1.0165811777114868, "learning_rate": 0.00019927152763590296, "loss": 4.6677, "step": 155750 }, { "epoch": 0.29074012302338004, "grad_norm": 0.8672209978103638, "learning_rate": 0.0001992710213975701, "loss": 4.7442, "step": 155800 }, { "epoch": 0.2908334285827585, "grad_norm": 1.1338589191436768, "learning_rate": 0.00019927051498404134, "loss": 4.6876, "step": 155850 }, { "epoch": 0.29092673414213704, "grad_norm": 1.0696067810058594, "learning_rate": 0.00019927000839531763, "loss": 4.5303, "step": 155900 }, { "epoch": 0.2910200397015155, "grad_norm": 0.9519825577735901, "learning_rate": 0.00019926950163139977, "loss": 4.581, "step": 155950 }, { "epoch": 0.291113345260894, "grad_norm": 1.0160168409347534, "learning_rate": 0.00019926899469228877, "loss": 4.589, "step": 156000 }, { "epoch": 0.291113345260894, "eval_loss": 4.831135272979736, "eval_runtime": 231.6301, "eval_samples_per_second": 11.259, "eval_steps_per_second": 11.259, "eval_tts_loss": 7.382750091658608, "step": 156000 }, { "epoch": 0.2912066508202725, "grad_norm": 1.5486959218978882, "learning_rate": 0.00019926848757798542, "loss": 4.5012, "step": 156050 }, { "epoch": 0.291299956379651, "grad_norm": 1.007653832435608, "learning_rate": 0.00019926798028849064, "loss": 4.5957, "step": 156100 }, { "epoch": 0.2913932619390295, "grad_norm": 0.7587629556655884, "learning_rate": 0.00019926747282380538, "loss": 4.5975, "step": 156150 }, { "epoch": 0.29148656749840796, "grad_norm": 0.8650146126747131, "learning_rate": 0.0001992669651839305, "loss": 4.5823, "step": 156200 }, { "epoch": 0.2915798730577865, "grad_norm": 0.8940017819404602, "learning_rate": 0.00019926645736886687, "loss": 4.5223, "step": 156250 }, { "epoch": 0.29167317861716496, "grad_norm": 1.1759480237960815, "learning_rate": 0.0001992659493786154, "loss": 4.5113, "step": 156300 }, { "epoch": 0.29176648417654344, "grad_norm": 0.9425550699234009, "learning_rate": 0.000199265441213177, "loss": 4.62, "step": 156350 }, { "epoch": 0.2918597897359219, "grad_norm": 1.2574522495269775, "learning_rate": 0.00019926493287255258, "loss": 4.7219, "step": 156400 }, { "epoch": 0.29195309529530045, "grad_norm": 0.6988387703895569, "learning_rate": 0.00019926442435674302, "loss": 4.5019, "step": 156450 }, { "epoch": 0.2920464008546789, "grad_norm": 1.215088963508606, "learning_rate": 0.0001992639156657492, "loss": 4.7895, "step": 156500 }, { "epoch": 0.2921397064140574, "grad_norm": 1.1339821815490723, "learning_rate": 0.00019926340679957206, "loss": 4.7092, "step": 156550 }, { "epoch": 0.29223301197343593, "grad_norm": 1.184877872467041, "learning_rate": 0.00019926289775821246, "loss": 4.848, "step": 156600 }, { "epoch": 0.2923263175328144, "grad_norm": 1.0538442134857178, "learning_rate": 0.0001992623885416713, "loss": 4.7515, "step": 156650 }, { "epoch": 0.2924196230921929, "grad_norm": 0.875169038772583, "learning_rate": 0.0001992618791499495, "loss": 4.5665, "step": 156700 }, { "epoch": 0.29251292865157136, "grad_norm": 0.8852382898330688, "learning_rate": 0.00019926136958304795, "loss": 4.4925, "step": 156750 }, { "epoch": 0.2926062342109499, "grad_norm": 0.683463454246521, "learning_rate": 0.00019926085984096754, "loss": 4.5527, "step": 156800 }, { "epoch": 0.29269953977032837, "grad_norm": 0.848481297492981, "learning_rate": 0.00019926034992370922, "loss": 4.5793, "step": 156850 }, { "epoch": 0.29279284532970684, "grad_norm": 0.9897000193595886, "learning_rate": 0.0001992598398312738, "loss": 4.6143, "step": 156900 }, { "epoch": 0.2928861508890854, "grad_norm": 0.9709441065788269, "learning_rate": 0.00019925932956366222, "loss": 4.5603, "step": 156950 }, { "epoch": 0.29297945644846385, "grad_norm": 0.9135065674781799, "learning_rate": 0.00019925881912087544, "loss": 4.7722, "step": 157000 }, { "epoch": 0.2930727620078423, "grad_norm": 0.9769490361213684, "learning_rate": 0.00019925830850291428, "loss": 4.534, "step": 157050 }, { "epoch": 0.2931660675672208, "grad_norm": 1.041882872581482, "learning_rate": 0.00019925779770977963, "loss": 4.6919, "step": 157100 }, { "epoch": 0.29325937312659933, "grad_norm": 0.6851228475570679, "learning_rate": 0.0001992572867414725, "loss": 4.5882, "step": 157150 }, { "epoch": 0.2933526786859778, "grad_norm": 1.007688283920288, "learning_rate": 0.00019925677559799368, "loss": 4.5231, "step": 157200 }, { "epoch": 0.2934459842453563, "grad_norm": 1.0649223327636719, "learning_rate": 0.00019925626427934413, "loss": 4.7773, "step": 157250 }, { "epoch": 0.2935392898047348, "grad_norm": 0.9082051515579224, "learning_rate": 0.00019925575278552474, "loss": 4.5285, "step": 157300 }, { "epoch": 0.2936325953641133, "grad_norm": 1.0188795328140259, "learning_rate": 0.00019925524111653642, "loss": 4.7059, "step": 157350 }, { "epoch": 0.29372590092349177, "grad_norm": 0.9674282670021057, "learning_rate": 0.00019925472927238006, "loss": 4.5048, "step": 157400 }, { "epoch": 0.29381920648287024, "grad_norm": 0.9318031072616577, "learning_rate": 0.00019925421725305656, "loss": 4.5211, "step": 157450 }, { "epoch": 0.2939125120422488, "grad_norm": 1.3290448188781738, "learning_rate": 0.00019925370505856683, "loss": 4.6717, "step": 157500 }, { "epoch": 0.29400581760162725, "grad_norm": 0.9480525851249695, "learning_rate": 0.00019925319268891177, "loss": 4.4553, "step": 157550 }, { "epoch": 0.2940991231610057, "grad_norm": 0.9672073721885681, "learning_rate": 0.00019925268014409227, "loss": 4.8206, "step": 157600 }, { "epoch": 0.29419242872038426, "grad_norm": 0.8860331177711487, "learning_rate": 0.0001992521674241093, "loss": 4.6907, "step": 157650 }, { "epoch": 0.29428573427976273, "grad_norm": 0.9548688530921936, "learning_rate": 0.0001992516545289637, "loss": 4.6107, "step": 157700 }, { "epoch": 0.2943790398391412, "grad_norm": 0.9516497850418091, "learning_rate": 0.0001992511414586564, "loss": 4.5693, "step": 157750 }, { "epoch": 0.2944723453985197, "grad_norm": 1.0647341012954712, "learning_rate": 0.0001992506282131883, "loss": 4.5256, "step": 157800 }, { "epoch": 0.2945656509578982, "grad_norm": 0.9625911712646484, "learning_rate": 0.0001992501147925603, "loss": 4.592, "step": 157850 }, { "epoch": 0.2946589565172767, "grad_norm": 1.109753131866455, "learning_rate": 0.00019924960119677328, "loss": 4.3946, "step": 157900 }, { "epoch": 0.29475226207665517, "grad_norm": 0.7586527466773987, "learning_rate": 0.0001992490874258282, "loss": 4.6777, "step": 157950 }, { "epoch": 0.2948455676360337, "grad_norm": 1.0405488014221191, "learning_rate": 0.00019924857347972593, "loss": 4.6913, "step": 158000 }, { "epoch": 0.2949388731954122, "grad_norm": 0.7991945743560791, "learning_rate": 0.00019924805935846746, "loss": 4.6677, "step": 158050 }, { "epoch": 0.29503217875479065, "grad_norm": 1.0138425827026367, "learning_rate": 0.00019924754506205354, "loss": 4.7778, "step": 158100 }, { "epoch": 0.29512548431416913, "grad_norm": 0.9910553693771362, "learning_rate": 0.00019924703059048522, "loss": 4.5255, "step": 158150 }, { "epoch": 0.29521878987354766, "grad_norm": 1.0533186197280884, "learning_rate": 0.00019924651594376334, "loss": 4.779, "step": 158200 }, { "epoch": 0.29531209543292614, "grad_norm": 0.8881279230117798, "learning_rate": 0.0001992460011218888, "loss": 4.8608, "step": 158250 }, { "epoch": 0.2954054009923046, "grad_norm": 0.9535214304924011, "learning_rate": 0.00019924548612486258, "loss": 4.6377, "step": 158300 }, { "epoch": 0.2954987065516831, "grad_norm": 1.084489107131958, "learning_rate": 0.00019924497095268547, "loss": 4.6459, "step": 158350 }, { "epoch": 0.2955920121110616, "grad_norm": 1.0378689765930176, "learning_rate": 0.0001992444556053585, "loss": 4.7632, "step": 158400 }, { "epoch": 0.2956853176704401, "grad_norm": 0.878700315952301, "learning_rate": 0.00019924394008288254, "loss": 4.4921, "step": 158450 }, { "epoch": 0.29577862322981857, "grad_norm": 1.0042437314987183, "learning_rate": 0.00019924342438525848, "loss": 4.7155, "step": 158500 }, { "epoch": 0.2958719287891971, "grad_norm": 1.3455873727798462, "learning_rate": 0.00019924290851248722, "loss": 4.6473, "step": 158550 }, { "epoch": 0.2959652343485756, "grad_norm": 0.9573052525520325, "learning_rate": 0.0001992423924645697, "loss": 4.8455, "step": 158600 }, { "epoch": 0.29605853990795405, "grad_norm": 1.201541781425476, "learning_rate": 0.00019924187624150677, "loss": 4.6125, "step": 158650 }, { "epoch": 0.29615184546733253, "grad_norm": 1.1214284896850586, "learning_rate": 0.00019924135984329946, "loss": 4.5251, "step": 158700 }, { "epoch": 0.29624515102671106, "grad_norm": 0.8579356670379639, "learning_rate": 0.00019924084326994858, "loss": 4.6285, "step": 158750 }, { "epoch": 0.29633845658608954, "grad_norm": 1.1231144666671753, "learning_rate": 0.00019924032652145508, "loss": 4.5755, "step": 158800 }, { "epoch": 0.296431762145468, "grad_norm": 0.8243206739425659, "learning_rate": 0.00019923980959781984, "loss": 4.5558, "step": 158850 }, { "epoch": 0.29652506770484655, "grad_norm": 1.0935394763946533, "learning_rate": 0.00019923929249904385, "loss": 4.913, "step": 158900 }, { "epoch": 0.296618373264225, "grad_norm": 0.9777284264564514, "learning_rate": 0.00019923877522512795, "loss": 4.6809, "step": 158950 }, { "epoch": 0.2967116788236035, "grad_norm": 0.9716470241546631, "learning_rate": 0.00019923825777607305, "loss": 4.4853, "step": 159000 }, { "epoch": 0.2967116788236035, "eval_loss": 4.834835529327393, "eval_runtime": 229.1261, "eval_samples_per_second": 11.382, "eval_steps_per_second": 11.382, "eval_tts_loss": 7.493346262103278, "step": 159000 }, { "epoch": 0.296804984382982, "grad_norm": 1.0158995389938354, "learning_rate": 0.00019923774015188014, "loss": 4.6832, "step": 159050 }, { "epoch": 0.2968982899423605, "grad_norm": 1.0746111869812012, "learning_rate": 0.00019923722235255, "loss": 4.9126, "step": 159100 }, { "epoch": 0.296991595501739, "grad_norm": 0.8816997408866882, "learning_rate": 0.00019923670437808368, "loss": 4.6653, "step": 159150 }, { "epoch": 0.29708490106111746, "grad_norm": 0.8043944835662842, "learning_rate": 0.00019923618622848203, "loss": 4.6432, "step": 159200 }, { "epoch": 0.297178206620496, "grad_norm": 1.2232578992843628, "learning_rate": 0.00019923566790374598, "loss": 4.7212, "step": 159250 }, { "epoch": 0.29727151217987446, "grad_norm": 0.991773784160614, "learning_rate": 0.0001992351494038764, "loss": 4.7941, "step": 159300 }, { "epoch": 0.29736481773925294, "grad_norm": 0.9849087595939636, "learning_rate": 0.00019923463072887428, "loss": 4.6429, "step": 159350 }, { "epoch": 0.2974581232986314, "grad_norm": 1.1423635482788086, "learning_rate": 0.0001992341118787405, "loss": 4.875, "step": 159400 }, { "epoch": 0.29755142885800995, "grad_norm": 0.6247135996818542, "learning_rate": 0.00019923359285347594, "loss": 4.4517, "step": 159450 }, { "epoch": 0.2976447344173884, "grad_norm": 0.9761686325073242, "learning_rate": 0.00019923307365308154, "loss": 4.5767, "step": 159500 }, { "epoch": 0.2977380399767669, "grad_norm": 0.8199406266212463, "learning_rate": 0.00019923255427755826, "loss": 4.7449, "step": 159550 }, { "epoch": 0.29783134553614543, "grad_norm": 1.3415900468826294, "learning_rate": 0.00019923203472690698, "loss": 4.5032, "step": 159600 }, { "epoch": 0.2979246510955239, "grad_norm": 0.806204617023468, "learning_rate": 0.00019923151500112862, "loss": 4.6618, "step": 159650 }, { "epoch": 0.2980179566549024, "grad_norm": 1.2007147073745728, "learning_rate": 0.0001992309951002241, "loss": 4.8032, "step": 159700 }, { "epoch": 0.29811126221428086, "grad_norm": 0.580830991268158, "learning_rate": 0.00019923047502419432, "loss": 4.5456, "step": 159750 }, { "epoch": 0.2982045677736594, "grad_norm": 1.0669281482696533, "learning_rate": 0.0001992299547730402, "loss": 4.4379, "step": 159800 }, { "epoch": 0.29829787333303787, "grad_norm": 0.7621221542358398, "learning_rate": 0.0001992294343467627, "loss": 4.698, "step": 159850 }, { "epoch": 0.29839117889241634, "grad_norm": 1.0081877708435059, "learning_rate": 0.00019922891374536268, "loss": 4.4803, "step": 159900 }, { "epoch": 0.2984844844517949, "grad_norm": 1.1483149528503418, "learning_rate": 0.00019922839296884114, "loss": 4.7129, "step": 159950 }, { "epoch": 0.29857779001117335, "grad_norm": 1.0426714420318604, "learning_rate": 0.00019922787201719888, "loss": 4.5651, "step": 160000 }, { "epoch": 0.2986710955705518, "grad_norm": 0.8986678123474121, "learning_rate": 0.00019922735089043691, "loss": 4.779, "step": 160050 }, { "epoch": 0.2987644011299303, "grad_norm": 0.9667002558708191, "learning_rate": 0.00019922682958855613, "loss": 4.6677, "step": 160100 }, { "epoch": 0.29885770668930883, "grad_norm": 0.9710828065872192, "learning_rate": 0.00019922630811155747, "loss": 4.7131, "step": 160150 }, { "epoch": 0.2989510122486873, "grad_norm": 1.078286051750183, "learning_rate": 0.00019922578645944182, "loss": 4.6679, "step": 160200 }, { "epoch": 0.2990443178080658, "grad_norm": 0.694128692150116, "learning_rate": 0.00019922526463221008, "loss": 4.5504, "step": 160250 }, { "epoch": 0.2991376233674443, "grad_norm": 1.0354713201522827, "learning_rate": 0.00019922474262986325, "loss": 4.4172, "step": 160300 }, { "epoch": 0.2992309289268228, "grad_norm": 0.9620189666748047, "learning_rate": 0.0001992242204524022, "loss": 4.6703, "step": 160350 }, { "epoch": 0.29932423448620127, "grad_norm": 0.9035593271255493, "learning_rate": 0.0001992236980998279, "loss": 4.6122, "step": 160400 }, { "epoch": 0.29941754004557974, "grad_norm": 0.9520270824432373, "learning_rate": 0.00019922317557214116, "loss": 4.7732, "step": 160450 }, { "epoch": 0.2995108456049583, "grad_norm": 0.8644099831581116, "learning_rate": 0.00019922265286934302, "loss": 4.5148, "step": 160500 }, { "epoch": 0.29960415116433675, "grad_norm": 0.7828960418701172, "learning_rate": 0.00019922212999143435, "loss": 4.7247, "step": 160550 }, { "epoch": 0.2996974567237152, "grad_norm": 0.9005388617515564, "learning_rate": 0.00019922160693841608, "loss": 4.6693, "step": 160600 }, { "epoch": 0.29979076228309376, "grad_norm": 0.6794205904006958, "learning_rate": 0.0001992210837102891, "loss": 4.5782, "step": 160650 }, { "epoch": 0.29988406784247224, "grad_norm": 1.0133854150772095, "learning_rate": 0.0001992205603070544, "loss": 4.562, "step": 160700 }, { "epoch": 0.2999773734018507, "grad_norm": 0.9719075560569763, "learning_rate": 0.00019922003672871285, "loss": 4.7801, "step": 160750 }, { "epoch": 0.3000706789612292, "grad_norm": 0.9305253028869629, "learning_rate": 0.00019921951297526542, "loss": 4.4637, "step": 160800 }, { "epoch": 0.3001639845206077, "grad_norm": 0.8835601210594177, "learning_rate": 0.00019921898904671297, "loss": 4.6572, "step": 160850 }, { "epoch": 0.3002572900799862, "grad_norm": 1.0033226013183594, "learning_rate": 0.0001992184649430565, "loss": 4.4474, "step": 160900 }, { "epoch": 0.30035059563936467, "grad_norm": 1.0143111944198608, "learning_rate": 0.00019921794066429688, "loss": 4.6021, "step": 160950 }, { "epoch": 0.3004439011987432, "grad_norm": 1.0978728532791138, "learning_rate": 0.00019921741621043503, "loss": 4.7851, "step": 161000 }, { "epoch": 0.3005372067581217, "grad_norm": 0.6397398710250854, "learning_rate": 0.00019921689158147195, "loss": 4.7239, "step": 161050 }, { "epoch": 0.30063051231750015, "grad_norm": 1.0616174936294556, "learning_rate": 0.00019921636677740846, "loss": 4.7843, "step": 161100 }, { "epoch": 0.30072381787687863, "grad_norm": 1.0118709802627563, "learning_rate": 0.00019921584179824557, "loss": 4.2376, "step": 161150 }, { "epoch": 0.30081712343625716, "grad_norm": 1.0035194158554077, "learning_rate": 0.00019921531664398418, "loss": 4.3126, "step": 161200 }, { "epoch": 0.30091042899563564, "grad_norm": 1.0038213729858398, "learning_rate": 0.0001992147913146252, "loss": 4.673, "step": 161250 }, { "epoch": 0.3010037345550141, "grad_norm": 1.1764312982559204, "learning_rate": 0.00019921426581016958, "loss": 4.6918, "step": 161300 }, { "epoch": 0.3010970401143926, "grad_norm": 0.8991124629974365, "learning_rate": 0.00019921374013061823, "loss": 4.4333, "step": 161350 }, { "epoch": 0.3011903456737711, "grad_norm": 0.8967214226722717, "learning_rate": 0.0001992132142759721, "loss": 4.5326, "step": 161400 }, { "epoch": 0.3012836512331496, "grad_norm": 1.1561260223388672, "learning_rate": 0.0001992126882462321, "loss": 4.8165, "step": 161450 }, { "epoch": 0.3013769567925281, "grad_norm": 1.163358211517334, "learning_rate": 0.00019921216204139914, "loss": 4.48, "step": 161500 }, { "epoch": 0.3014702623519066, "grad_norm": 0.9330498576164246, "learning_rate": 0.0001992116356614742, "loss": 4.5813, "step": 161550 }, { "epoch": 0.3015635679112851, "grad_norm": 0.8502509593963623, "learning_rate": 0.00019921110910645816, "loss": 4.8287, "step": 161600 }, { "epoch": 0.30165687347066356, "grad_norm": 0.6995953917503357, "learning_rate": 0.000199210582376352, "loss": 4.683, "step": 161650 }, { "epoch": 0.30175017903004203, "grad_norm": 1.1503678560256958, "learning_rate": 0.0001992100554711566, "loss": 4.7569, "step": 161700 }, { "epoch": 0.30184348458942056, "grad_norm": 0.8800851702690125, "learning_rate": 0.0001992095283908729, "loss": 4.6177, "step": 161750 }, { "epoch": 0.30193679014879904, "grad_norm": 0.9621112942695618, "learning_rate": 0.00019920900113550185, "loss": 4.6983, "step": 161800 }, { "epoch": 0.3020300957081775, "grad_norm": 1.1952325105667114, "learning_rate": 0.00019920847370504435, "loss": 4.7141, "step": 161850 }, { "epoch": 0.30212340126755605, "grad_norm": 0.8587202429771423, "learning_rate": 0.0001992079460995014, "loss": 4.7418, "step": 161900 }, { "epoch": 0.3022167068269345, "grad_norm": 0.8307145237922668, "learning_rate": 0.0001992074183188738, "loss": 4.7172, "step": 161950 }, { "epoch": 0.302310012386313, "grad_norm": 0.9355490207672119, "learning_rate": 0.00019920689036316262, "loss": 4.7346, "step": 162000 }, { "epoch": 0.302310012386313, "eval_loss": 4.827625274658203, "eval_runtime": 232.9861, "eval_samples_per_second": 11.194, "eval_steps_per_second": 11.194, "eval_tts_loss": 7.458007696548911, "step": 162000 }, { "epoch": 0.3024033179456915, "grad_norm": 1.3112372159957886, "learning_rate": 0.00019920636223236872, "loss": 4.6245, "step": 162050 }, { "epoch": 0.30249662350507, "grad_norm": 0.8583440780639648, "learning_rate": 0.00019920583392649306, "loss": 4.6192, "step": 162100 }, { "epoch": 0.3025899290644485, "grad_norm": 0.9936934113502502, "learning_rate": 0.00019920530544553653, "loss": 4.7136, "step": 162150 }, { "epoch": 0.30268323462382696, "grad_norm": 0.9335197806358337, "learning_rate": 0.00019920477678950013, "loss": 4.6699, "step": 162200 }, { "epoch": 0.3027765401832055, "grad_norm": 0.9630725383758545, "learning_rate": 0.0001992042479583847, "loss": 4.6637, "step": 162250 }, { "epoch": 0.30286984574258397, "grad_norm": 0.7474913597106934, "learning_rate": 0.00019920371895219128, "loss": 4.7166, "step": 162300 }, { "epoch": 0.30296315130196244, "grad_norm": 0.7774944305419922, "learning_rate": 0.00019920318977092075, "loss": 4.8446, "step": 162350 }, { "epoch": 0.3030564568613409, "grad_norm": 1.0125433206558228, "learning_rate": 0.00019920266041457401, "loss": 4.6259, "step": 162400 }, { "epoch": 0.30314976242071945, "grad_norm": 1.1934956312179565, "learning_rate": 0.00019920213088315202, "loss": 4.6795, "step": 162450 }, { "epoch": 0.3032430679800979, "grad_norm": 0.9211152195930481, "learning_rate": 0.00019920160117665577, "loss": 4.6288, "step": 162500 }, { "epoch": 0.3033363735394764, "grad_norm": 0.9939976930618286, "learning_rate": 0.00019920107129508613, "loss": 4.7214, "step": 162550 }, { "epoch": 0.30342967909885493, "grad_norm": 0.8724084496498108, "learning_rate": 0.00019920054123844408, "loss": 4.4638, "step": 162600 }, { "epoch": 0.3035229846582334, "grad_norm": 0.994696855545044, "learning_rate": 0.00019920001100673046, "loss": 4.8186, "step": 162650 }, { "epoch": 0.3036162902176119, "grad_norm": 0.9215883612632751, "learning_rate": 0.00019919948059994634, "loss": 4.5902, "step": 162700 }, { "epoch": 0.30370959577699036, "grad_norm": 0.9584205746650696, "learning_rate": 0.00019919895001809258, "loss": 4.6455, "step": 162750 }, { "epoch": 0.3038029013363689, "grad_norm": 0.8970978856086731, "learning_rate": 0.00019919841926117012, "loss": 4.5378, "step": 162800 }, { "epoch": 0.30389620689574737, "grad_norm": 0.7979338765144348, "learning_rate": 0.0001991978883291799, "loss": 4.6594, "step": 162850 }, { "epoch": 0.30398951245512584, "grad_norm": 0.7578598856925964, "learning_rate": 0.00019919735722212284, "loss": 4.832, "step": 162900 }, { "epoch": 0.3040828180145044, "grad_norm": 1.0004535913467407, "learning_rate": 0.00019919682593999994, "loss": 4.695, "step": 162950 }, { "epoch": 0.30417612357388285, "grad_norm": 0.9248453974723816, "learning_rate": 0.00019919629448281209, "loss": 4.3721, "step": 163000 }, { "epoch": 0.3042694291332613, "grad_norm": 1.3004566431045532, "learning_rate": 0.00019919576285056022, "loss": 4.7349, "step": 163050 }, { "epoch": 0.3043627346926398, "grad_norm": 0.8813239336013794, "learning_rate": 0.0001991952310432453, "loss": 4.6711, "step": 163100 }, { "epoch": 0.30445604025201833, "grad_norm": 1.0755641460418701, "learning_rate": 0.00019919469906086822, "loss": 4.8154, "step": 163150 }, { "epoch": 0.3045493458113968, "grad_norm": 0.9873396158218384, "learning_rate": 0.00019919416690343, "loss": 4.3667, "step": 163200 }, { "epoch": 0.3046426513707753, "grad_norm": 0.7341350317001343, "learning_rate": 0.0001991936345709315, "loss": 4.5203, "step": 163250 }, { "epoch": 0.3047359569301538, "grad_norm": 1.094083547592163, "learning_rate": 0.00019919310206337366, "loss": 4.6661, "step": 163300 }, { "epoch": 0.3048292624895323, "grad_norm": 1.1205472946166992, "learning_rate": 0.00019919256938075747, "loss": 4.6381, "step": 163350 }, { "epoch": 0.30492256804891077, "grad_norm": 0.9988998174667358, "learning_rate": 0.00019919203652308385, "loss": 4.7326, "step": 163400 }, { "epoch": 0.30501587360828925, "grad_norm": 0.9494795203208923, "learning_rate": 0.00019919150349035375, "loss": 4.4937, "step": 163450 }, { "epoch": 0.3051091791676678, "grad_norm": 0.9424142241477966, "learning_rate": 0.0001991909702825681, "loss": 4.3626, "step": 163500 }, { "epoch": 0.30520248472704625, "grad_norm": 1.1510084867477417, "learning_rate": 0.00019919043689972782, "loss": 4.6443, "step": 163550 }, { "epoch": 0.30529579028642473, "grad_norm": 1.1867023706436157, "learning_rate": 0.0001991899033418339, "loss": 4.6603, "step": 163600 }, { "epoch": 0.30538909584580326, "grad_norm": 1.285799503326416, "learning_rate": 0.00019918936960888723, "loss": 4.7631, "step": 163650 }, { "epoch": 0.30548240140518174, "grad_norm": 1.011475920677185, "learning_rate": 0.00019918883570088877, "loss": 4.4944, "step": 163700 }, { "epoch": 0.3055757069645602, "grad_norm": 0.7682061791419983, "learning_rate": 0.0001991883016178395, "loss": 4.5854, "step": 163750 }, { "epoch": 0.3056690125239387, "grad_norm": 1.2138035297393799, "learning_rate": 0.0001991877673597403, "loss": 4.5669, "step": 163800 }, { "epoch": 0.3057623180833172, "grad_norm": 1.1199895143508911, "learning_rate": 0.00019918723292659216, "loss": 4.8251, "step": 163850 }, { "epoch": 0.3058556236426957, "grad_norm": 0.9371145367622375, "learning_rate": 0.00019918669831839601, "loss": 4.6153, "step": 163900 }, { "epoch": 0.30594892920207417, "grad_norm": 0.8127995729446411, "learning_rate": 0.00019918616353515277, "loss": 4.4168, "step": 163950 }, { "epoch": 0.30604223476145265, "grad_norm": 0.9031199216842651, "learning_rate": 0.00019918562857686344, "loss": 4.707, "step": 164000 }, { "epoch": 0.3061355403208312, "grad_norm": 1.1019059419631958, "learning_rate": 0.00019918509344352887, "loss": 4.577, "step": 164050 }, { "epoch": 0.30622884588020965, "grad_norm": 0.9607084393501282, "learning_rate": 0.0001991845581351501, "loss": 4.6266, "step": 164100 }, { "epoch": 0.30632215143958813, "grad_norm": 0.6045005917549133, "learning_rate": 0.00019918402265172805, "loss": 4.743, "step": 164150 }, { "epoch": 0.30641545699896666, "grad_norm": 0.9607996940612793, "learning_rate": 0.0001991834869932636, "loss": 4.637, "step": 164200 }, { "epoch": 0.30650876255834514, "grad_norm": 1.1145765781402588, "learning_rate": 0.0001991829511597578, "loss": 4.6641, "step": 164250 }, { "epoch": 0.3066020681177236, "grad_norm": 1.0428991317749023, "learning_rate": 0.0001991824151512115, "loss": 4.6888, "step": 164300 }, { "epoch": 0.3066953736771021, "grad_norm": 0.8830058574676514, "learning_rate": 0.00019918187896762575, "loss": 4.6689, "step": 164350 }, { "epoch": 0.3067886792364806, "grad_norm": 0.9978854656219482, "learning_rate": 0.0001991813426090014, "loss": 4.7586, "step": 164400 }, { "epoch": 0.3068819847958591, "grad_norm": 0.9433168768882751, "learning_rate": 0.00019918080607533943, "loss": 4.813, "step": 164450 }, { "epoch": 0.3069752903552376, "grad_norm": 0.64937824010849, "learning_rate": 0.00019918026936664078, "loss": 4.3087, "step": 164500 }, { "epoch": 0.3070685959146161, "grad_norm": 0.9438527822494507, "learning_rate": 0.00019917973248290643, "loss": 4.4295, "step": 164550 }, { "epoch": 0.3071619014739946, "grad_norm": 0.8068161606788635, "learning_rate": 0.00019917919542413728, "loss": 4.4977, "step": 164600 }, { "epoch": 0.30725520703337306, "grad_norm": 0.896666407585144, "learning_rate": 0.00019917865819033432, "loss": 4.6681, "step": 164650 }, { "epoch": 0.30734851259275153, "grad_norm": 1.1718946695327759, "learning_rate": 0.00019917812078149843, "loss": 4.565, "step": 164700 }, { "epoch": 0.30744181815213006, "grad_norm": 1.3140202760696411, "learning_rate": 0.00019917758319763066, "loss": 4.4831, "step": 164750 }, { "epoch": 0.30753512371150854, "grad_norm": 1.0430924892425537, "learning_rate": 0.0001991770454387319, "loss": 4.4573, "step": 164800 }, { "epoch": 0.307628429270887, "grad_norm": 0.9410809278488159, "learning_rate": 0.0001991765075048031, "loss": 4.5014, "step": 164850 }, { "epoch": 0.30772173483026555, "grad_norm": 0.7173283696174622, "learning_rate": 0.00019917596939584523, "loss": 4.7214, "step": 164900 }, { "epoch": 0.307815040389644, "grad_norm": 1.0757492780685425, "learning_rate": 0.00019917543111185921, "loss": 4.5144, "step": 164950 }, { "epoch": 0.3079083459490225, "grad_norm": 0.9228636026382446, "learning_rate": 0.000199174892652846, "loss": 4.6586, "step": 165000 }, { "epoch": 0.3079083459490225, "eval_loss": 4.838285446166992, "eval_runtime": 229.3082, "eval_samples_per_second": 11.373, "eval_steps_per_second": 11.373, "eval_tts_loss": 7.4116918211060066, "step": 165000 }, { "epoch": 0.308001651508401, "grad_norm": 1.0502113103866577, "learning_rate": 0.00019917435401880655, "loss": 4.6569, "step": 165050 }, { "epoch": 0.3080949570677795, "grad_norm": 0.9943327903747559, "learning_rate": 0.0001991738152097418, "loss": 4.5486, "step": 165100 }, { "epoch": 0.308188262627158, "grad_norm": 1.2838720083236694, "learning_rate": 0.00019917327622565276, "loss": 4.6833, "step": 165150 }, { "epoch": 0.30828156818653646, "grad_norm": 0.9620805978775024, "learning_rate": 0.00019917273706654033, "loss": 4.6248, "step": 165200 }, { "epoch": 0.308374873745915, "grad_norm": 1.1069529056549072, "learning_rate": 0.00019917219773240544, "loss": 4.4957, "step": 165250 }, { "epoch": 0.30846817930529347, "grad_norm": 0.6974848508834839, "learning_rate": 0.0001991716582232491, "loss": 4.7632, "step": 165300 }, { "epoch": 0.30856148486467194, "grad_norm": 1.0856902599334717, "learning_rate": 0.0001991711185390722, "loss": 4.6712, "step": 165350 }, { "epoch": 0.3086547904240504, "grad_norm": 1.2169914245605469, "learning_rate": 0.00019917057867987575, "loss": 4.5188, "step": 165400 }, { "epoch": 0.30874809598342895, "grad_norm": 0.9107069969177246, "learning_rate": 0.00019917003864566065, "loss": 4.7021, "step": 165450 }, { "epoch": 0.3088414015428074, "grad_norm": 0.910029947757721, "learning_rate": 0.00019916949843642794, "loss": 4.788, "step": 165500 }, { "epoch": 0.3089347071021859, "grad_norm": 1.238258957862854, "learning_rate": 0.00019916895805217844, "loss": 4.7447, "step": 165550 }, { "epoch": 0.30902801266156443, "grad_norm": 1.087679386138916, "learning_rate": 0.00019916841749291323, "loss": 4.6314, "step": 165600 }, { "epoch": 0.3091213182209429, "grad_norm": 0.9426969885826111, "learning_rate": 0.0001991678767586332, "loss": 4.4512, "step": 165650 }, { "epoch": 0.3092146237803214, "grad_norm": 0.944223940372467, "learning_rate": 0.00019916733584933932, "loss": 4.7139, "step": 165700 }, { "epoch": 0.30930792933969986, "grad_norm": 1.0137444734573364, "learning_rate": 0.00019916679476503254, "loss": 4.6459, "step": 165750 }, { "epoch": 0.3094012348990784, "grad_norm": 1.2591444253921509, "learning_rate": 0.00019916625350571384, "loss": 4.6593, "step": 165800 }, { "epoch": 0.30949454045845687, "grad_norm": 0.8761984705924988, "learning_rate": 0.00019916571207138412, "loss": 4.4277, "step": 165850 }, { "epoch": 0.30958784601783534, "grad_norm": 0.8550477027893066, "learning_rate": 0.00019916517046204434, "loss": 4.5806, "step": 165900 }, { "epoch": 0.3096811515772139, "grad_norm": 0.9725227355957031, "learning_rate": 0.00019916462867769553, "loss": 4.6497, "step": 165950 }, { "epoch": 0.30977445713659235, "grad_norm": 1.0279942750930786, "learning_rate": 0.00019916408671833858, "loss": 4.6852, "step": 166000 }, { "epoch": 0.3098677626959708, "grad_norm": 0.8512559533119202, "learning_rate": 0.0001991635445839745, "loss": 4.6959, "step": 166050 }, { "epoch": 0.3099610682553493, "grad_norm": 0.7771254181861877, "learning_rate": 0.00019916300227460417, "loss": 4.7392, "step": 166100 }, { "epoch": 0.31005437381472783, "grad_norm": 1.103044867515564, "learning_rate": 0.00019916245979022858, "loss": 4.3448, "step": 166150 }, { "epoch": 0.3101476793741063, "grad_norm": 1.1170908212661743, "learning_rate": 0.00019916191713084872, "loss": 4.4304, "step": 166200 }, { "epoch": 0.3102409849334848, "grad_norm": 1.0523208379745483, "learning_rate": 0.00019916137429646554, "loss": 4.7645, "step": 166250 }, { "epoch": 0.3103342904928633, "grad_norm": 1.361086368560791, "learning_rate": 0.00019916083128707995, "loss": 4.4946, "step": 166300 }, { "epoch": 0.3104275960522418, "grad_norm": 0.9509695172309875, "learning_rate": 0.00019916028810269297, "loss": 4.7445, "step": 166350 }, { "epoch": 0.31052090161162027, "grad_norm": 1.121620774269104, "learning_rate": 0.0001991597447433055, "loss": 4.8506, "step": 166400 }, { "epoch": 0.31061420717099875, "grad_norm": 1.0928542613983154, "learning_rate": 0.00019915920120891855, "loss": 4.6162, "step": 166450 }, { "epoch": 0.3107075127303773, "grad_norm": 1.339594841003418, "learning_rate": 0.00019915865749953304, "loss": 4.6252, "step": 166500 }, { "epoch": 0.31080081828975575, "grad_norm": 0.9323852062225342, "learning_rate": 0.00019915811361514996, "loss": 4.7945, "step": 166550 }, { "epoch": 0.31089412384913423, "grad_norm": 1.001569390296936, "learning_rate": 0.00019915756955577027, "loss": 4.5336, "step": 166600 }, { "epoch": 0.3109874294085127, "grad_norm": 0.9864224195480347, "learning_rate": 0.00019915702532139487, "loss": 4.5236, "step": 166650 }, { "epoch": 0.31108073496789124, "grad_norm": 0.9965605735778809, "learning_rate": 0.0001991564809120248, "loss": 4.6922, "step": 166700 }, { "epoch": 0.3111740405272697, "grad_norm": 1.0161752700805664, "learning_rate": 0.000199155936327661, "loss": 4.7956, "step": 166750 }, { "epoch": 0.3112673460866482, "grad_norm": 1.1669565439224243, "learning_rate": 0.00019915539156830443, "loss": 4.8049, "step": 166800 }, { "epoch": 0.3113606516460267, "grad_norm": 1.1081312894821167, "learning_rate": 0.000199154846633956, "loss": 4.6178, "step": 166850 }, { "epoch": 0.3114539572054052, "grad_norm": 0.9586905837059021, "learning_rate": 0.00019915430152461674, "loss": 4.8232, "step": 166900 }, { "epoch": 0.3115472627647837, "grad_norm": 1.1186403036117554, "learning_rate": 0.00019915375624028755, "loss": 4.6214, "step": 166950 }, { "epoch": 0.31164056832416215, "grad_norm": 1.1981947422027588, "learning_rate": 0.00019915321078096946, "loss": 4.7213, "step": 167000 }, { "epoch": 0.3117338738835407, "grad_norm": 0.7325296401977539, "learning_rate": 0.0001991526651466634, "loss": 4.6811, "step": 167050 }, { "epoch": 0.31182717944291916, "grad_norm": 1.0409541130065918, "learning_rate": 0.00019915211933737033, "loss": 4.5533, "step": 167100 }, { "epoch": 0.31192048500229763, "grad_norm": 0.8175338506698608, "learning_rate": 0.0001991515733530912, "loss": 4.5909, "step": 167150 }, { "epoch": 0.31201379056167616, "grad_norm": 1.0503971576690674, "learning_rate": 0.000199151027193827, "loss": 4.4858, "step": 167200 }, { "epoch": 0.31210709612105464, "grad_norm": 1.094265103340149, "learning_rate": 0.00019915048085957868, "loss": 4.5226, "step": 167250 }, { "epoch": 0.3122004016804331, "grad_norm": 0.8987294435501099, "learning_rate": 0.0001991499343503472, "loss": 4.7354, "step": 167300 }, { "epoch": 0.3122937072398116, "grad_norm": 1.0248451232910156, "learning_rate": 0.00019914938766613354, "loss": 4.5532, "step": 167350 }, { "epoch": 0.3123870127991901, "grad_norm": 0.8451243042945862, "learning_rate": 0.00019914884080693868, "loss": 4.5967, "step": 167400 }, { "epoch": 0.3124803183585686, "grad_norm": 0.8834465742111206, "learning_rate": 0.0001991482937727635, "loss": 4.2495, "step": 167450 }, { "epoch": 0.3125736239179471, "grad_norm": 0.8053299784660339, "learning_rate": 0.00019914774656360908, "loss": 4.9166, "step": 167500 }, { "epoch": 0.3126669294773256, "grad_norm": 1.0274240970611572, "learning_rate": 0.00019914719917947632, "loss": 4.7979, "step": 167550 }, { "epoch": 0.3127602350367041, "grad_norm": 0.8318864107131958, "learning_rate": 0.0001991466516203662, "loss": 4.5883, "step": 167600 }, { "epoch": 0.31285354059608256, "grad_norm": 1.0550457239151, "learning_rate": 0.00019914610388627969, "loss": 4.4031, "step": 167650 }, { "epoch": 0.31294684615546103, "grad_norm": 1.0894427299499512, "learning_rate": 0.00019914555597721773, "loss": 4.87, "step": 167700 }, { "epoch": 0.31304015171483957, "grad_norm": 0.8881887197494507, "learning_rate": 0.00019914500789318133, "loss": 4.8212, "step": 167750 }, { "epoch": 0.31313345727421804, "grad_norm": 0.9769049286842346, "learning_rate": 0.00019914445963417144, "loss": 4.5287, "step": 167800 }, { "epoch": 0.3132267628335965, "grad_norm": 1.0178450345993042, "learning_rate": 0.00019914391120018902, "loss": 4.5699, "step": 167850 }, { "epoch": 0.31332006839297505, "grad_norm": 0.9004485011100769, "learning_rate": 0.00019914336259123502, "loss": 4.6477, "step": 167900 }, { "epoch": 0.3134133739523535, "grad_norm": 1.098066806793213, "learning_rate": 0.00019914281380731047, "loss": 4.4141, "step": 167950 }, { "epoch": 0.313506679511732, "grad_norm": 0.9474552273750305, "learning_rate": 0.00019914226484841625, "loss": 4.4733, "step": 168000 }, { "epoch": 0.313506679511732, "eval_loss": 4.8274312019348145, "eval_runtime": 228.3749, "eval_samples_per_second": 11.42, "eval_steps_per_second": 11.42, "eval_tts_loss": 7.468050231518612, "step": 168000 }, { "epoch": 0.3135999850711105, "grad_norm": 1.2087820768356323, "learning_rate": 0.00019914171571455342, "loss": 4.6384, "step": 168050 }, { "epoch": 0.313693290630489, "grad_norm": 1.0665926933288574, "learning_rate": 0.00019914116640572287, "loss": 4.6831, "step": 168100 }, { "epoch": 0.3137865961898675, "grad_norm": 0.769088625907898, "learning_rate": 0.00019914061692192562, "loss": 4.7609, "step": 168150 }, { "epoch": 0.31387990174924596, "grad_norm": 1.0530633926391602, "learning_rate": 0.00019914006726316264, "loss": 4.9207, "step": 168200 }, { "epoch": 0.3139732073086245, "grad_norm": 0.705093502998352, "learning_rate": 0.00019913951742943486, "loss": 4.642, "step": 168250 }, { "epoch": 0.31406651286800297, "grad_norm": 0.9531348347663879, "learning_rate": 0.0001991389674207433, "loss": 4.9607, "step": 168300 }, { "epoch": 0.31415981842738144, "grad_norm": 0.8925808668136597, "learning_rate": 0.0001991384172370889, "loss": 4.4425, "step": 168350 }, { "epoch": 0.3142531239867599, "grad_norm": 0.8512563705444336, "learning_rate": 0.00019913786687847265, "loss": 4.5021, "step": 168400 }, { "epoch": 0.31434642954613845, "grad_norm": 1.233641266822815, "learning_rate": 0.00019913731634489548, "loss": 4.7805, "step": 168450 }, { "epoch": 0.3144397351055169, "grad_norm": 0.7689288854598999, "learning_rate": 0.0001991367656363584, "loss": 4.5333, "step": 168500 }, { "epoch": 0.3145330406648954, "grad_norm": 1.1268550157546997, "learning_rate": 0.00019913621475286237, "loss": 4.7591, "step": 168550 }, { "epoch": 0.31462634622427393, "grad_norm": 1.1939990520477295, "learning_rate": 0.00019913566369440837, "loss": 4.6831, "step": 168600 }, { "epoch": 0.3147196517836524, "grad_norm": 0.9455341100692749, "learning_rate": 0.00019913511246099735, "loss": 4.7392, "step": 168650 }, { "epoch": 0.3148129573430309, "grad_norm": 0.9590946435928345, "learning_rate": 0.00019913456105263033, "loss": 4.4734, "step": 168700 }, { "epoch": 0.31490626290240936, "grad_norm": 1.0091745853424072, "learning_rate": 0.00019913400946930824, "loss": 4.6294, "step": 168750 }, { "epoch": 0.3149995684617879, "grad_norm": 0.9442651867866516, "learning_rate": 0.00019913345771103207, "loss": 4.5777, "step": 168800 }, { "epoch": 0.31509287402116637, "grad_norm": 0.7429641485214233, "learning_rate": 0.00019913290577780278, "loss": 4.722, "step": 168850 }, { "epoch": 0.31518617958054485, "grad_norm": 1.2271195650100708, "learning_rate": 0.00019913235366962137, "loss": 4.8243, "step": 168900 }, { "epoch": 0.3152794851399234, "grad_norm": 1.0357447862625122, "learning_rate": 0.00019913180138648876, "loss": 4.5255, "step": 168950 }, { "epoch": 0.31537279069930185, "grad_norm": 1.144982933998108, "learning_rate": 0.00019913124892840603, "loss": 4.5941, "step": 169000 }, { "epoch": 0.31546609625868033, "grad_norm": 0.9120845794677734, "learning_rate": 0.00019913069629537403, "loss": 4.6719, "step": 169050 }, { "epoch": 0.3155594018180588, "grad_norm": 1.1090995073318481, "learning_rate": 0.00019913014348739384, "loss": 4.7637, "step": 169100 }, { "epoch": 0.31565270737743734, "grad_norm": 1.0764634609222412, "learning_rate": 0.00019912959050446633, "loss": 4.5816, "step": 169150 }, { "epoch": 0.3157460129368158, "grad_norm": 0.9311011433601379, "learning_rate": 0.00019912903734659257, "loss": 4.5876, "step": 169200 }, { "epoch": 0.3158393184961943, "grad_norm": 0.9119871854782104, "learning_rate": 0.00019912848401377353, "loss": 4.428, "step": 169250 }, { "epoch": 0.31593262405557276, "grad_norm": 0.9250431060791016, "learning_rate": 0.00019912793050601014, "loss": 4.528, "step": 169300 }, { "epoch": 0.3160259296149513, "grad_norm": 1.0978806018829346, "learning_rate": 0.00019912737682330336, "loss": 4.8015, "step": 169350 }, { "epoch": 0.31611923517432977, "grad_norm": 0.9182941913604736, "learning_rate": 0.00019912682296565422, "loss": 4.6908, "step": 169400 }, { "epoch": 0.31621254073370825, "grad_norm": 0.8654223084449768, "learning_rate": 0.0001991262689330637, "loss": 4.5184, "step": 169450 }, { "epoch": 0.3163058462930868, "grad_norm": 0.9357742071151733, "learning_rate": 0.00019912571472553272, "loss": 4.7359, "step": 169500 }, { "epoch": 0.31639915185246525, "grad_norm": 0.9018232226371765, "learning_rate": 0.00019912516034306232, "loss": 4.497, "step": 169550 }, { "epoch": 0.31649245741184373, "grad_norm": 1.170499324798584, "learning_rate": 0.00019912460578565345, "loss": 4.6351, "step": 169600 }, { "epoch": 0.3165857629712222, "grad_norm": 1.0474013090133667, "learning_rate": 0.0001991240510533071, "loss": 4.6865, "step": 169650 }, { "epoch": 0.31667906853060074, "grad_norm": 0.7851083874702454, "learning_rate": 0.00019912349614602425, "loss": 4.6992, "step": 169700 }, { "epoch": 0.3167723740899792, "grad_norm": 1.0943204164505005, "learning_rate": 0.00019912294106380586, "loss": 4.683, "step": 169750 }, { "epoch": 0.3168656796493577, "grad_norm": 1.037070870399475, "learning_rate": 0.00019912238580665288, "loss": 4.6626, "step": 169800 }, { "epoch": 0.3169589852087362, "grad_norm": 0.8870090842247009, "learning_rate": 0.00019912183037456638, "loss": 4.6193, "step": 169850 }, { "epoch": 0.3170522907681147, "grad_norm": 0.9513359069824219, "learning_rate": 0.00019912127476754728, "loss": 4.694, "step": 169900 }, { "epoch": 0.3171455963274932, "grad_norm": 1.1024489402770996, "learning_rate": 0.00019912071898559654, "loss": 4.6326, "step": 169950 }, { "epoch": 0.31723890188687165, "grad_norm": 1.0496639013290405, "learning_rate": 0.0001991201630287152, "loss": 4.4681, "step": 170000 }, { "epoch": 0.3173322074462502, "grad_norm": 0.8167015910148621, "learning_rate": 0.00019911960689690422, "loss": 4.6452, "step": 170050 }, { "epoch": 0.31742551300562866, "grad_norm": 0.811915397644043, "learning_rate": 0.00019911905059016456, "loss": 4.8109, "step": 170100 }, { "epoch": 0.31751881856500713, "grad_norm": 1.01187264919281, "learning_rate": 0.00019911849410849723, "loss": 4.6726, "step": 170150 }, { "epoch": 0.31761212412438566, "grad_norm": 1.0084149837493896, "learning_rate": 0.0001991179374519032, "loss": 4.5055, "step": 170200 }, { "epoch": 0.31770542968376414, "grad_norm": 0.9547125101089478, "learning_rate": 0.00019911738062038342, "loss": 4.5173, "step": 170250 }, { "epoch": 0.3177987352431426, "grad_norm": 0.9173873066902161, "learning_rate": 0.00019911682361393895, "loss": 4.8112, "step": 170300 }, { "epoch": 0.3178920408025211, "grad_norm": 1.2184867858886719, "learning_rate": 0.0001991162664325707, "loss": 4.4778, "step": 170350 }, { "epoch": 0.3179853463618996, "grad_norm": 1.0674309730529785, "learning_rate": 0.0001991157090762797, "loss": 4.5748, "step": 170400 }, { "epoch": 0.3180786519212781, "grad_norm": 0.7090270519256592, "learning_rate": 0.00019911515154506687, "loss": 4.4085, "step": 170450 }, { "epoch": 0.3181719574806566, "grad_norm": 0.9503317475318909, "learning_rate": 0.00019911459383893326, "loss": 4.5279, "step": 170500 }, { "epoch": 0.3182652630400351, "grad_norm": 0.6449097990989685, "learning_rate": 0.00019911403595787984, "loss": 4.5365, "step": 170550 }, { "epoch": 0.3183585685994136, "grad_norm": 1.0800449848175049, "learning_rate": 0.00019911347790190758, "loss": 4.5558, "step": 170600 }, { "epoch": 0.31845187415879206, "grad_norm": 0.8874872326850891, "learning_rate": 0.00019911291967101747, "loss": 4.6167, "step": 170650 }, { "epoch": 0.31854517971817053, "grad_norm": 0.8549138903617859, "learning_rate": 0.00019911236126521047, "loss": 4.6317, "step": 170700 }, { "epoch": 0.31863848527754907, "grad_norm": 1.0165430307388306, "learning_rate": 0.00019911180268448764, "loss": 4.6746, "step": 170750 }, { "epoch": 0.31873179083692754, "grad_norm": 1.049067497253418, "learning_rate": 0.00019911124392884988, "loss": 4.7359, "step": 170800 }, { "epoch": 0.318825096396306, "grad_norm": 1.071564793586731, "learning_rate": 0.00019911068499829825, "loss": 4.5707, "step": 170850 }, { "epoch": 0.31891840195568455, "grad_norm": 0.9487912654876709, "learning_rate": 0.00019911012589283366, "loss": 4.8717, "step": 170900 }, { "epoch": 0.319011707515063, "grad_norm": 1.0813241004943848, "learning_rate": 0.00019910956661245717, "loss": 4.6888, "step": 170950 }, { "epoch": 0.3191050130744415, "grad_norm": 1.02992844581604, "learning_rate": 0.0001991090071571697, "loss": 4.6039, "step": 171000 }, { "epoch": 0.3191050130744415, "eval_loss": 4.821337699890137, "eval_runtime": 229.1039, "eval_samples_per_second": 11.383, "eval_steps_per_second": 11.383, "eval_tts_loss": 7.466205600627692, "step": 171000 }, { "epoch": 0.31919831863382, "grad_norm": 1.2654902935028076, "learning_rate": 0.0001991084475269723, "loss": 4.5334, "step": 171050 }, { "epoch": 0.3192916241931985, "grad_norm": 0.9937511086463928, "learning_rate": 0.00019910788772186592, "loss": 4.6948, "step": 171100 }, { "epoch": 0.319384929752577, "grad_norm": 1.1610994338989258, "learning_rate": 0.00019910732774185155, "loss": 4.3105, "step": 171150 }, { "epoch": 0.31947823531195546, "grad_norm": 0.9847679734230042, "learning_rate": 0.00019910676758693022, "loss": 4.7063, "step": 171200 }, { "epoch": 0.319571540871334, "grad_norm": 1.15132474899292, "learning_rate": 0.00019910620725710284, "loss": 4.4479, "step": 171250 }, { "epoch": 0.31966484643071247, "grad_norm": 1.0646486282348633, "learning_rate": 0.00019910564675237048, "loss": 4.6407, "step": 171300 }, { "epoch": 0.31975815199009094, "grad_norm": 0.982887864112854, "learning_rate": 0.00019910508607273405, "loss": 4.4841, "step": 171350 }, { "epoch": 0.3198514575494694, "grad_norm": 1.1266651153564453, "learning_rate": 0.0001991045252181946, "loss": 4.8636, "step": 171400 }, { "epoch": 0.31994476310884795, "grad_norm": 0.9004781246185303, "learning_rate": 0.0001991039641887531, "loss": 4.7329, "step": 171450 }, { "epoch": 0.3200380686682264, "grad_norm": 0.904808759689331, "learning_rate": 0.00019910340298441055, "loss": 4.7403, "step": 171500 }, { "epoch": 0.3201313742276049, "grad_norm": 1.0027724504470825, "learning_rate": 0.0001991028416051679, "loss": 4.7172, "step": 171550 }, { "epoch": 0.32022467978698343, "grad_norm": 0.9115274548530579, "learning_rate": 0.00019910228005102623, "loss": 4.7143, "step": 171600 }, { "epoch": 0.3203179853463619, "grad_norm": 1.1834417581558228, "learning_rate": 0.0001991017183219864, "loss": 4.7278, "step": 171650 }, { "epoch": 0.3204112909057404, "grad_norm": 1.0323907136917114, "learning_rate": 0.00019910115641804952, "loss": 4.7366, "step": 171700 }, { "epoch": 0.32050459646511886, "grad_norm": 1.0845869779586792, "learning_rate": 0.00019910059433921653, "loss": 4.7036, "step": 171750 }, { "epoch": 0.3205979020244974, "grad_norm": 1.2340019941329956, "learning_rate": 0.00019910003208548844, "loss": 4.3835, "step": 171800 }, { "epoch": 0.32069120758387587, "grad_norm": 1.1320146322250366, "learning_rate": 0.0001990994696568662, "loss": 4.5122, "step": 171850 }, { "epoch": 0.32078451314325435, "grad_norm": 1.0517185926437378, "learning_rate": 0.00019909890705335087, "loss": 4.4093, "step": 171900 }, { "epoch": 0.3208778187026328, "grad_norm": 0.9432281255722046, "learning_rate": 0.00019909834427494342, "loss": 4.7337, "step": 171950 }, { "epoch": 0.32097112426201135, "grad_norm": 1.0861693620681763, "learning_rate": 0.00019909778132164475, "loss": 4.5552, "step": 172000 }, { "epoch": 0.32106442982138983, "grad_norm": 1.0468862056732178, "learning_rate": 0.00019909721819345603, "loss": 4.8132, "step": 172050 }, { "epoch": 0.3211577353807683, "grad_norm": 0.941402018070221, "learning_rate": 0.00019909665489037807, "loss": 4.7838, "step": 172100 }, { "epoch": 0.32125104094014684, "grad_norm": 0.92326420545578, "learning_rate": 0.000199096091412412, "loss": 4.476, "step": 172150 }, { "epoch": 0.3213443464995253, "grad_norm": 0.8617150187492371, "learning_rate": 0.00019909552775955876, "loss": 4.5015, "step": 172200 }, { "epoch": 0.3214376520589038, "grad_norm": 0.7417050004005432, "learning_rate": 0.00019909496393181932, "loss": 4.8644, "step": 172250 }, { "epoch": 0.32153095761828226, "grad_norm": 1.1346899271011353, "learning_rate": 0.00019909439992919474, "loss": 4.8367, "step": 172300 }, { "epoch": 0.3216242631776608, "grad_norm": 0.9893891215324402, "learning_rate": 0.00019909383575168597, "loss": 4.7965, "step": 172350 }, { "epoch": 0.3217175687370393, "grad_norm": 0.8581725358963013, "learning_rate": 0.000199093271399294, "loss": 4.6347, "step": 172400 }, { "epoch": 0.32181087429641775, "grad_norm": 0.9298039078712463, "learning_rate": 0.00019909270687201987, "loss": 4.7319, "step": 172450 }, { "epoch": 0.3219041798557963, "grad_norm": 1.0204685926437378, "learning_rate": 0.00019909214216986453, "loss": 4.6937, "step": 172500 }, { "epoch": 0.32199748541517476, "grad_norm": 1.0665239095687866, "learning_rate": 0.000199091577292829, "loss": 4.7268, "step": 172550 }, { "epoch": 0.32209079097455323, "grad_norm": 1.0132877826690674, "learning_rate": 0.0001990910122409143, "loss": 4.509, "step": 172600 }, { "epoch": 0.3221840965339317, "grad_norm": 1.0414512157440186, "learning_rate": 0.00019909044701412135, "loss": 4.4304, "step": 172650 }, { "epoch": 0.32227740209331024, "grad_norm": 0.763314962387085, "learning_rate": 0.0001990898816124512, "loss": 4.5164, "step": 172700 }, { "epoch": 0.3223707076526887, "grad_norm": 0.8184088468551636, "learning_rate": 0.00019908931603590486, "loss": 4.7437, "step": 172750 }, { "epoch": 0.3224640132120672, "grad_norm": 0.8922542333602905, "learning_rate": 0.00019908875028448332, "loss": 4.4946, "step": 172800 }, { "epoch": 0.3225573187714457, "grad_norm": 0.7964805364608765, "learning_rate": 0.00019908818435818756, "loss": 4.6092, "step": 172850 }, { "epoch": 0.3226506243308242, "grad_norm": 0.9622323513031006, "learning_rate": 0.0001990876182570186, "loss": 4.7243, "step": 172900 }, { "epoch": 0.3227439298902027, "grad_norm": 1.1066397428512573, "learning_rate": 0.0001990870519809774, "loss": 4.6053, "step": 172950 }, { "epoch": 0.32283723544958115, "grad_norm": 1.3279519081115723, "learning_rate": 0.00019908648553006503, "loss": 4.5265, "step": 173000 }, { "epoch": 0.3229305410089597, "grad_norm": 0.8593128323554993, "learning_rate": 0.00019908591890428244, "loss": 4.8224, "step": 173050 }, { "epoch": 0.32302384656833816, "grad_norm": 0.9954770803451538, "learning_rate": 0.00019908535210363062, "loss": 4.7258, "step": 173100 }, { "epoch": 0.32311715212771663, "grad_norm": 0.6582159996032715, "learning_rate": 0.0001990847851281106, "loss": 4.6627, "step": 173150 }, { "epoch": 0.32321045768709517, "grad_norm": 1.021619439125061, "learning_rate": 0.00019908421797772336, "loss": 4.7628, "step": 173200 }, { "epoch": 0.32330376324647364, "grad_norm": 1.0340412855148315, "learning_rate": 0.0001990836506524699, "loss": 4.4405, "step": 173250 }, { "epoch": 0.3233970688058521, "grad_norm": 1.0030783414840698, "learning_rate": 0.00019908308315235126, "loss": 4.7434, "step": 173300 }, { "epoch": 0.3234903743652306, "grad_norm": 1.1510215997695923, "learning_rate": 0.0001990825154773684, "loss": 4.6975, "step": 173350 }, { "epoch": 0.3235836799246091, "grad_norm": 0.7062892317771912, "learning_rate": 0.00019908194762752234, "loss": 4.6726, "step": 173400 }, { "epoch": 0.3236769854839876, "grad_norm": 1.1423907279968262, "learning_rate": 0.00019908137960281407, "loss": 4.524, "step": 173450 }, { "epoch": 0.3237702910433661, "grad_norm": 0.9802839159965515, "learning_rate": 0.00019908081140324456, "loss": 4.5331, "step": 173500 }, { "epoch": 0.3238635966027446, "grad_norm": 0.7058094143867493, "learning_rate": 0.00019908024302881493, "loss": 4.4531, "step": 173550 }, { "epoch": 0.3239569021621231, "grad_norm": 0.9642864465713501, "learning_rate": 0.00019907967447952603, "loss": 4.5526, "step": 173600 }, { "epoch": 0.32405020772150156, "grad_norm": 0.8382852077484131, "learning_rate": 0.000199079105755379, "loss": 4.6682, "step": 173650 }, { "epoch": 0.32414351328088004, "grad_norm": 1.2792258262634277, "learning_rate": 0.00019907853685637475, "loss": 4.6946, "step": 173700 }, { "epoch": 0.32423681884025857, "grad_norm": 1.1190528869628906, "learning_rate": 0.00019907796778251432, "loss": 4.5769, "step": 173750 }, { "epoch": 0.32433012439963704, "grad_norm": 1.1447540521621704, "learning_rate": 0.0001990773985337987, "loss": 4.8288, "step": 173800 }, { "epoch": 0.3244234299590155, "grad_norm": 1.0323656797409058, "learning_rate": 0.00019907682911022892, "loss": 4.6316, "step": 173850 }, { "epoch": 0.32451673551839405, "grad_norm": 0.9703738689422607, "learning_rate": 0.00019907625951180594, "loss": 4.6256, "step": 173900 }, { "epoch": 0.3246100410777725, "grad_norm": 1.1295841932296753, "learning_rate": 0.0001990756897385308, "loss": 4.542, "step": 173950 }, { "epoch": 0.324703346637151, "grad_norm": 1.204323172569275, "learning_rate": 0.00019907511979040454, "loss": 4.655, "step": 174000 }, { "epoch": 0.324703346637151, "eval_loss": 4.816260814666748, "eval_runtime": 230.4264, "eval_samples_per_second": 11.318, "eval_steps_per_second": 11.318, "eval_tts_loss": 7.435452688951045, "step": 174000 }, { "epoch": 0.3247966521965295, "grad_norm": 1.0280636548995972, "learning_rate": 0.00019907454966742812, "loss": 4.6904, "step": 174050 }, { "epoch": 0.324889957755908, "grad_norm": 1.1565958261489868, "learning_rate": 0.00019907397936960253, "loss": 4.3498, "step": 174100 }, { "epoch": 0.3249832633152865, "grad_norm": 1.0304701328277588, "learning_rate": 0.0001990734088969288, "loss": 4.5535, "step": 174150 }, { "epoch": 0.32507656887466496, "grad_norm": 0.8157714605331421, "learning_rate": 0.00019907283824940794, "loss": 4.5031, "step": 174200 }, { "epoch": 0.3251698744340435, "grad_norm": 0.6631931662559509, "learning_rate": 0.00019907226742704097, "loss": 4.8029, "step": 174250 }, { "epoch": 0.32526317999342197, "grad_norm": 1.1516517400741577, "learning_rate": 0.00019907169642982886, "loss": 4.7112, "step": 174300 }, { "epoch": 0.32535648555280045, "grad_norm": 0.7090115547180176, "learning_rate": 0.00019907112525777265, "loss": 4.8284, "step": 174350 }, { "epoch": 0.3254497911121789, "grad_norm": 1.1411478519439697, "learning_rate": 0.00019907055391087334, "loss": 4.7277, "step": 174400 }, { "epoch": 0.32554309667155745, "grad_norm": 0.763681948184967, "learning_rate": 0.00019906998238913192, "loss": 4.8038, "step": 174450 }, { "epoch": 0.32563640223093593, "grad_norm": 1.0577315092086792, "learning_rate": 0.0001990694106925494, "loss": 4.6134, "step": 174500 }, { "epoch": 0.3257297077903144, "grad_norm": 0.9963567852973938, "learning_rate": 0.00019906883882112682, "loss": 4.472, "step": 174550 }, { "epoch": 0.3258230133496929, "grad_norm": 0.8183549046516418, "learning_rate": 0.00019906826677486516, "loss": 4.4465, "step": 174600 }, { "epoch": 0.3259163189090714, "grad_norm": 0.908412516117096, "learning_rate": 0.00019906769455376545, "loss": 4.5436, "step": 174650 }, { "epoch": 0.3260096244684499, "grad_norm": 1.11734139919281, "learning_rate": 0.0001990671221578287, "loss": 4.8297, "step": 174700 }, { "epoch": 0.32610293002782836, "grad_norm": 1.123195767402649, "learning_rate": 0.0001990665495870559, "loss": 4.5935, "step": 174750 }, { "epoch": 0.3261962355872069, "grad_norm": 0.9942184090614319, "learning_rate": 0.00019906597684144807, "loss": 4.6195, "step": 174800 }, { "epoch": 0.32628954114658537, "grad_norm": 1.0256868600845337, "learning_rate": 0.00019906540392100624, "loss": 4.5741, "step": 174850 }, { "epoch": 0.32638284670596385, "grad_norm": 1.351872205734253, "learning_rate": 0.00019906483082573136, "loss": 4.5077, "step": 174900 }, { "epoch": 0.3264761522653423, "grad_norm": 0.7622151374816895, "learning_rate": 0.00019906425755562452, "loss": 4.5419, "step": 174950 }, { "epoch": 0.32656945782472085, "grad_norm": 0.8031182885169983, "learning_rate": 0.00019906368411068672, "loss": 4.5477, "step": 175000 }, { "epoch": 0.32666276338409933, "grad_norm": 1.1167811155319214, "learning_rate": 0.0001990631104909189, "loss": 4.7045, "step": 175050 }, { "epoch": 0.3267560689434778, "grad_norm": 1.2244101762771606, "learning_rate": 0.00019906253669632214, "loss": 4.7501, "step": 175100 }, { "epoch": 0.32684937450285634, "grad_norm": 0.8345790505409241, "learning_rate": 0.00019906196272689744, "loss": 4.5942, "step": 175150 }, { "epoch": 0.3269426800622348, "grad_norm": 0.927778422832489, "learning_rate": 0.00019906138858264577, "loss": 4.7552, "step": 175200 }, { "epoch": 0.3270359856216133, "grad_norm": 1.1129297018051147, "learning_rate": 0.00019906081426356823, "loss": 4.5184, "step": 175250 }, { "epoch": 0.32712929118099177, "grad_norm": 0.8951523900032043, "learning_rate": 0.00019906023976966577, "loss": 4.7909, "step": 175300 }, { "epoch": 0.3272225967403703, "grad_norm": 1.09775710105896, "learning_rate": 0.0001990596651009394, "loss": 4.61, "step": 175350 }, { "epoch": 0.3273159022997488, "grad_norm": 1.0176907777786255, "learning_rate": 0.00019905909025739015, "loss": 4.7255, "step": 175400 }, { "epoch": 0.32740920785912725, "grad_norm": 1.1389837265014648, "learning_rate": 0.00019905851523901907, "loss": 4.4602, "step": 175450 }, { "epoch": 0.3275025134185058, "grad_norm": 0.8852382302284241, "learning_rate": 0.00019905794004582708, "loss": 4.629, "step": 175500 }, { "epoch": 0.32759581897788426, "grad_norm": 1.0725276470184326, "learning_rate": 0.00019905736467781533, "loss": 4.5922, "step": 175550 }, { "epoch": 0.32768912453726273, "grad_norm": 0.8179895281791687, "learning_rate": 0.00019905678913498468, "loss": 4.6167, "step": 175600 }, { "epoch": 0.3277824300966412, "grad_norm": 0.7662099599838257, "learning_rate": 0.0001990562134173363, "loss": 4.46, "step": 175650 }, { "epoch": 0.32787573565601974, "grad_norm": 1.104725956916809, "learning_rate": 0.00019905563752487112, "loss": 4.6232, "step": 175700 }, { "epoch": 0.3279690412153982, "grad_norm": 1.0175747871398926, "learning_rate": 0.0001990550614575901, "loss": 4.5933, "step": 175750 }, { "epoch": 0.3280623467747767, "grad_norm": 1.1158710718154907, "learning_rate": 0.00019905448521549438, "loss": 4.6522, "step": 175800 }, { "epoch": 0.3281556523341552, "grad_norm": 1.6744157075881958, "learning_rate": 0.00019905390879858495, "loss": 4.7029, "step": 175850 }, { "epoch": 0.3282489578935337, "grad_norm": 0.9802636504173279, "learning_rate": 0.00019905333220686275, "loss": 4.5862, "step": 175900 }, { "epoch": 0.3283422634529122, "grad_norm": 0.7022011280059814, "learning_rate": 0.00019905275544032888, "loss": 4.5594, "step": 175950 }, { "epoch": 0.32843556901229065, "grad_norm": 0.8897641897201538, "learning_rate": 0.00019905217849898433, "loss": 4.6417, "step": 176000 }, { "epoch": 0.3285288745716692, "grad_norm": 1.1504569053649902, "learning_rate": 0.00019905160138283007, "loss": 4.6595, "step": 176050 }, { "epoch": 0.32862218013104766, "grad_norm": 1.0295069217681885, "learning_rate": 0.0001990510240918672, "loss": 4.5644, "step": 176100 }, { "epoch": 0.32871548569042613, "grad_norm": 0.9090021848678589, "learning_rate": 0.0001990504466260967, "loss": 4.4895, "step": 176150 }, { "epoch": 0.32880879124980467, "grad_norm": 0.9149088263511658, "learning_rate": 0.00019904986898551955, "loss": 4.6901, "step": 176200 }, { "epoch": 0.32890209680918314, "grad_norm": 0.8863251209259033, "learning_rate": 0.00019904929117013685, "loss": 4.6697, "step": 176250 }, { "epoch": 0.3289954023685616, "grad_norm": 1.07883882522583, "learning_rate": 0.00019904871317994957, "loss": 4.598, "step": 176300 }, { "epoch": 0.3290887079279401, "grad_norm": 0.9367828369140625, "learning_rate": 0.00019904813501495875, "loss": 4.6257, "step": 176350 }, { "epoch": 0.3291820134873186, "grad_norm": 1.0381193161010742, "learning_rate": 0.00019904755667516538, "loss": 4.6984, "step": 176400 }, { "epoch": 0.3292753190466971, "grad_norm": 0.847889244556427, "learning_rate": 0.00019904697816057053, "loss": 4.4849, "step": 176450 }, { "epoch": 0.3293686246060756, "grad_norm": 1.0124871730804443, "learning_rate": 0.00019904639947117514, "loss": 4.7115, "step": 176500 }, { "epoch": 0.3294619301654541, "grad_norm": 0.8684128522872925, "learning_rate": 0.0001990458206069803, "loss": 4.5261, "step": 176550 }, { "epoch": 0.3295552357248326, "grad_norm": 0.7277292609214783, "learning_rate": 0.00019904524156798704, "loss": 4.4937, "step": 176600 }, { "epoch": 0.32964854128421106, "grad_norm": 1.0759245157241821, "learning_rate": 0.00019904466235419634, "loss": 4.6008, "step": 176650 }, { "epoch": 0.32974184684358954, "grad_norm": 0.9541880488395691, "learning_rate": 0.00019904408296560924, "loss": 4.542, "step": 176700 }, { "epoch": 0.32983515240296807, "grad_norm": 0.8675011992454529, "learning_rate": 0.00019904350340222676, "loss": 4.5248, "step": 176750 }, { "epoch": 0.32992845796234654, "grad_norm": 0.9355060458183289, "learning_rate": 0.00019904292366404995, "loss": 4.5585, "step": 176800 }, { "epoch": 0.330021763521725, "grad_norm": 1.1286976337432861, "learning_rate": 0.00019904234375107976, "loss": 4.6191, "step": 176850 }, { "epoch": 0.33011506908110355, "grad_norm": 0.9177277684211731, "learning_rate": 0.00019904176366331727, "loss": 4.698, "step": 176900 }, { "epoch": 0.330208374640482, "grad_norm": 0.9858579039573669, "learning_rate": 0.00019904118340076355, "loss": 4.6723, "step": 176950 }, { "epoch": 0.3303016801998605, "grad_norm": 1.172903299331665, "learning_rate": 0.00019904060296341947, "loss": 4.5899, "step": 177000 }, { "epoch": 0.3303016801998605, "eval_loss": 4.8154120445251465, "eval_runtime": 228.9954, "eval_samples_per_second": 11.389, "eval_steps_per_second": 11.389, "eval_tts_loss": 7.542438022716634, "step": 177000 }, { "epoch": 0.330394985759239, "grad_norm": 1.2343531847000122, "learning_rate": 0.00019904002235128623, "loss": 4.5554, "step": 177050 }, { "epoch": 0.3304882913186175, "grad_norm": 1.2554324865341187, "learning_rate": 0.00019903944156436477, "loss": 4.6312, "step": 177100 }, { "epoch": 0.330581596877996, "grad_norm": 1.062771201133728, "learning_rate": 0.0001990388606026561, "loss": 4.7273, "step": 177150 }, { "epoch": 0.33067490243737446, "grad_norm": 0.921607494354248, "learning_rate": 0.0001990382794661613, "loss": 4.5602, "step": 177200 }, { "epoch": 0.330768207996753, "grad_norm": 1.0047394037246704, "learning_rate": 0.00019903769815488132, "loss": 4.4348, "step": 177250 }, { "epoch": 0.33086151355613147, "grad_norm": 0.6674817204475403, "learning_rate": 0.00019903711666881724, "loss": 4.6014, "step": 177300 }, { "epoch": 0.33095481911550995, "grad_norm": 1.1274831295013428, "learning_rate": 0.0001990365350079701, "loss": 4.6538, "step": 177350 }, { "epoch": 0.3310481246748884, "grad_norm": 0.9511878490447998, "learning_rate": 0.00019903595317234088, "loss": 4.8159, "step": 177400 }, { "epoch": 0.33114143023426695, "grad_norm": 1.1247509717941284, "learning_rate": 0.0001990353711619306, "loss": 4.6459, "step": 177450 }, { "epoch": 0.33123473579364543, "grad_norm": 1.111100673675537, "learning_rate": 0.00019903478897674036, "loss": 4.6574, "step": 177500 }, { "epoch": 0.3313280413530239, "grad_norm": 0.9724700450897217, "learning_rate": 0.00019903420661677114, "loss": 4.6917, "step": 177550 }, { "epoch": 0.3314213469124024, "grad_norm": 1.2161903381347656, "learning_rate": 0.00019903362408202395, "loss": 4.6725, "step": 177600 }, { "epoch": 0.3315146524717809, "grad_norm": 0.881798267364502, "learning_rate": 0.00019903304137249984, "loss": 4.4201, "step": 177650 }, { "epoch": 0.3316079580311594, "grad_norm": 1.228102684020996, "learning_rate": 0.00019903245848819988, "loss": 4.6403, "step": 177700 }, { "epoch": 0.33170126359053786, "grad_norm": 0.9526916146278381, "learning_rate": 0.00019903187542912504, "loss": 4.5728, "step": 177750 }, { "epoch": 0.3317945691499164, "grad_norm": 0.8187074661254883, "learning_rate": 0.00019903129219527632, "loss": 4.6696, "step": 177800 }, { "epoch": 0.33188787470929487, "grad_norm": 0.9834941029548645, "learning_rate": 0.00019903070878665485, "loss": 4.4247, "step": 177850 }, { "epoch": 0.33198118026867335, "grad_norm": 1.224478006362915, "learning_rate": 0.00019903012520326155, "loss": 4.6612, "step": 177900 }, { "epoch": 0.3320744858280518, "grad_norm": 0.7598810791969299, "learning_rate": 0.00019902954144509753, "loss": 4.563, "step": 177950 }, { "epoch": 0.33216779138743036, "grad_norm": 1.0683156251907349, "learning_rate": 0.0001990289575121638, "loss": 4.6277, "step": 178000 }, { "epoch": 0.33226109694680883, "grad_norm": 1.1003293991088867, "learning_rate": 0.0001990283734044614, "loss": 4.7756, "step": 178050 }, { "epoch": 0.3323544025061873, "grad_norm": 0.9530490040779114, "learning_rate": 0.00019902778912199128, "loss": 4.5517, "step": 178100 }, { "epoch": 0.33244770806556584, "grad_norm": 0.7573721408843994, "learning_rate": 0.0001990272046647546, "loss": 4.4958, "step": 178150 }, { "epoch": 0.3325410136249443, "grad_norm": 1.020322561264038, "learning_rate": 0.00019902662003275232, "loss": 4.5234, "step": 178200 }, { "epoch": 0.3326343191843228, "grad_norm": 0.9242296814918518, "learning_rate": 0.00019902603522598547, "loss": 4.6549, "step": 178250 }, { "epoch": 0.33272762474370127, "grad_norm": 1.179311752319336, "learning_rate": 0.00019902545024445508, "loss": 4.5256, "step": 178300 }, { "epoch": 0.3328209303030798, "grad_norm": 1.0598257780075073, "learning_rate": 0.00019902486508816222, "loss": 4.7532, "step": 178350 }, { "epoch": 0.3329142358624583, "grad_norm": 0.7825391292572021, "learning_rate": 0.0001990242797571079, "loss": 4.6937, "step": 178400 }, { "epoch": 0.33300754142183675, "grad_norm": 1.1020742654800415, "learning_rate": 0.00019902369425129311, "loss": 4.6548, "step": 178450 }, { "epoch": 0.3331008469812153, "grad_norm": 0.9558370113372803, "learning_rate": 0.00019902310857071895, "loss": 4.8421, "step": 178500 }, { "epoch": 0.33319415254059376, "grad_norm": 0.8280293941497803, "learning_rate": 0.00019902252271538644, "loss": 4.6858, "step": 178550 }, { "epoch": 0.33328745809997223, "grad_norm": 0.8872970342636108, "learning_rate": 0.00019902193668529656, "loss": 4.4993, "step": 178600 }, { "epoch": 0.3333807636593507, "grad_norm": 1.1399357318878174, "learning_rate": 0.00019902135048045042, "loss": 4.7029, "step": 178650 }, { "epoch": 0.33347406921872924, "grad_norm": 1.1143829822540283, "learning_rate": 0.00019902076410084903, "loss": 4.6988, "step": 178700 }, { "epoch": 0.3335673747781077, "grad_norm": 1.012130856513977, "learning_rate": 0.00019902017754649338, "loss": 4.6413, "step": 178750 }, { "epoch": 0.3336606803374862, "grad_norm": 1.0381497144699097, "learning_rate": 0.00019901959081738456, "loss": 4.5631, "step": 178800 }, { "epoch": 0.3337539858968647, "grad_norm": 0.788299024105072, "learning_rate": 0.00019901900391352357, "loss": 4.5039, "step": 178850 }, { "epoch": 0.3338472914562432, "grad_norm": 0.8483543992042542, "learning_rate": 0.0001990184168349115, "loss": 4.7003, "step": 178900 }, { "epoch": 0.3339405970156217, "grad_norm": 0.7698935866355896, "learning_rate": 0.0001990178295815493, "loss": 4.5215, "step": 178950 }, { "epoch": 0.33403390257500015, "grad_norm": 1.0911551713943481, "learning_rate": 0.00019901724215343807, "loss": 4.5445, "step": 179000 }, { "epoch": 0.3341272081343787, "grad_norm": 0.824788510799408, "learning_rate": 0.00019901665455057882, "loss": 4.5568, "step": 179050 }, { "epoch": 0.33422051369375716, "grad_norm": 0.88285893201828, "learning_rate": 0.00019901606677297262, "loss": 4.5505, "step": 179100 }, { "epoch": 0.33431381925313564, "grad_norm": 0.9610635638237, "learning_rate": 0.00019901547882062047, "loss": 4.6942, "step": 179150 }, { "epoch": 0.33440712481251417, "grad_norm": 0.9467892050743103, "learning_rate": 0.0001990148906935234, "loss": 4.5793, "step": 179200 }, { "epoch": 0.33450043037189264, "grad_norm": 1.0193818807601929, "learning_rate": 0.00019901430239168254, "loss": 4.5554, "step": 179250 }, { "epoch": 0.3345937359312711, "grad_norm": 1.2325133085250854, "learning_rate": 0.0001990137139150988, "loss": 4.6663, "step": 179300 }, { "epoch": 0.3346870414906496, "grad_norm": 0.6962212324142456, "learning_rate": 0.00019901312526377329, "loss": 4.5595, "step": 179350 }, { "epoch": 0.3347803470500281, "grad_norm": 1.0818673372268677, "learning_rate": 0.00019901253643770704, "loss": 4.5388, "step": 179400 }, { "epoch": 0.3348736526094066, "grad_norm": 0.9971811175346375, "learning_rate": 0.0001990119474369011, "loss": 4.4762, "step": 179450 }, { "epoch": 0.3349669581687851, "grad_norm": 0.9982391595840454, "learning_rate": 0.00019901135826135645, "loss": 4.5818, "step": 179500 }, { "epoch": 0.3350602637281636, "grad_norm": 2.8972480297088623, "learning_rate": 0.0001990107689110742, "loss": 4.5753, "step": 179550 }, { "epoch": 0.3351535692875421, "grad_norm": 0.804068922996521, "learning_rate": 0.00019901017938605536, "loss": 4.8127, "step": 179600 }, { "epoch": 0.33524687484692056, "grad_norm": 0.8897950053215027, "learning_rate": 0.000199009589686301, "loss": 4.6659, "step": 179650 }, { "epoch": 0.33534018040629904, "grad_norm": 1.096867561340332, "learning_rate": 0.00019900899981181212, "loss": 4.6793, "step": 179700 }, { "epoch": 0.33543348596567757, "grad_norm": 0.9849563241004944, "learning_rate": 0.00019900840976258977, "loss": 4.6498, "step": 179750 }, { "epoch": 0.33552679152505605, "grad_norm": 1.0917383432388306, "learning_rate": 0.000199007819538635, "loss": 4.4898, "step": 179800 }, { "epoch": 0.3356200970844345, "grad_norm": 1.0085644721984863, "learning_rate": 0.00019900722913994884, "loss": 4.7175, "step": 179850 }, { "epoch": 0.33571340264381305, "grad_norm": 0.8432663679122925, "learning_rate": 0.00019900663856653233, "loss": 4.5698, "step": 179900 }, { "epoch": 0.33580670820319153, "grad_norm": 1.037923812866211, "learning_rate": 0.00019900604781838656, "loss": 4.6207, "step": 179950 }, { "epoch": 0.33590001376257, "grad_norm": 0.9545549154281616, "learning_rate": 0.0001990054568955125, "loss": 4.6257, "step": 180000 }, { "epoch": 0.33590001376257, "eval_loss": 4.818175315856934, "eval_runtime": 228.2553, "eval_samples_per_second": 11.426, "eval_steps_per_second": 11.426, "eval_tts_loss": 7.496765146657618, "step": 180000 }, { "epoch": 0.3359933193219485, "grad_norm": 0.8834198713302612, "learning_rate": 0.00019900486579791128, "loss": 4.5444, "step": 180050 }, { "epoch": 0.336086624881327, "grad_norm": 1.05471932888031, "learning_rate": 0.00019900427452558385, "loss": 4.5951, "step": 180100 }, { "epoch": 0.3361799304407055, "grad_norm": 1.0553511381149292, "learning_rate": 0.0001990036830785313, "loss": 4.6423, "step": 180150 }, { "epoch": 0.33627323600008396, "grad_norm": 0.8488613963127136, "learning_rate": 0.00019900309145675472, "loss": 4.8316, "step": 180200 }, { "epoch": 0.33636654155946244, "grad_norm": 0.8496327996253967, "learning_rate": 0.00019900249966025503, "loss": 4.5013, "step": 180250 }, { "epoch": 0.33645984711884097, "grad_norm": 1.085153579711914, "learning_rate": 0.0001990019076890334, "loss": 4.5181, "step": 180300 }, { "epoch": 0.33655315267821945, "grad_norm": 1.0941439867019653, "learning_rate": 0.0001990013155430908, "loss": 4.4706, "step": 180350 }, { "epoch": 0.3366464582375979, "grad_norm": 0.9088996052742004, "learning_rate": 0.00019900072322242827, "loss": 4.8162, "step": 180400 }, { "epoch": 0.33673976379697645, "grad_norm": 0.8877679705619812, "learning_rate": 0.00019900013072704693, "loss": 4.4462, "step": 180450 }, { "epoch": 0.33683306935635493, "grad_norm": 1.0370676517486572, "learning_rate": 0.0001989995380569478, "loss": 4.6237, "step": 180500 }, { "epoch": 0.3369263749157334, "grad_norm": 0.7951025366783142, "learning_rate": 0.00019899894521213183, "loss": 4.5673, "step": 180550 }, { "epoch": 0.3370196804751119, "grad_norm": 1.1994744539260864, "learning_rate": 0.00019899835219260018, "loss": 4.4731, "step": 180600 }, { "epoch": 0.3371129860344904, "grad_norm": 1.1491518020629883, "learning_rate": 0.00019899775899835388, "loss": 4.5261, "step": 180650 }, { "epoch": 0.3372062915938689, "grad_norm": 0.9701890349388123, "learning_rate": 0.0001989971656293939, "loss": 4.8223, "step": 180700 }, { "epoch": 0.33729959715324737, "grad_norm": 0.7716856002807617, "learning_rate": 0.00019899657208572138, "loss": 4.7053, "step": 180750 }, { "epoch": 0.3373929027126259, "grad_norm": 1.2134188413619995, "learning_rate": 0.00019899597836733733, "loss": 4.8142, "step": 180800 }, { "epoch": 0.3374862082720044, "grad_norm": 1.0805442333221436, "learning_rate": 0.00019899538447424277, "loss": 4.5745, "step": 180850 }, { "epoch": 0.33757951383138285, "grad_norm": 0.9581602215766907, "learning_rate": 0.0001989947904064388, "loss": 4.7698, "step": 180900 }, { "epoch": 0.3376728193907613, "grad_norm": 1.0755600929260254, "learning_rate": 0.00019899419616392644, "loss": 4.6043, "step": 180950 }, { "epoch": 0.33776612495013986, "grad_norm": 0.9674501419067383, "learning_rate": 0.00019899360174670673, "loss": 4.683, "step": 181000 }, { "epoch": 0.33785943050951833, "grad_norm": 0.8484556078910828, "learning_rate": 0.00019899300715478077, "loss": 4.6563, "step": 181050 }, { "epoch": 0.3379527360688968, "grad_norm": 0.9506059885025024, "learning_rate": 0.00019899241238814953, "loss": 4.5925, "step": 181100 }, { "epoch": 0.33804604162827534, "grad_norm": 1.1092076301574707, "learning_rate": 0.0001989918174468141, "loss": 4.689, "step": 181150 }, { "epoch": 0.3381393471876538, "grad_norm": 0.8176827430725098, "learning_rate": 0.00019899122233077554, "loss": 4.5785, "step": 181200 }, { "epoch": 0.3382326527470323, "grad_norm": 0.9832103252410889, "learning_rate": 0.00019899062704003488, "loss": 4.4976, "step": 181250 }, { "epoch": 0.33832595830641077, "grad_norm": 0.6890990734100342, "learning_rate": 0.00019899003157459322, "loss": 4.5854, "step": 181300 }, { "epoch": 0.3384192638657893, "grad_norm": 1.0396530628204346, "learning_rate": 0.00019898943593445154, "loss": 4.6799, "step": 181350 }, { "epoch": 0.3385125694251678, "grad_norm": 1.0980576276779175, "learning_rate": 0.00019898884011961094, "loss": 4.5739, "step": 181400 }, { "epoch": 0.33860587498454625, "grad_norm": 0.8679540753364563, "learning_rate": 0.00019898824413007243, "loss": 4.5388, "step": 181450 }, { "epoch": 0.3386991805439248, "grad_norm": 0.857035219669342, "learning_rate": 0.0001989876479658371, "loss": 4.4769, "step": 181500 }, { "epoch": 0.33879248610330326, "grad_norm": 0.9443413019180298, "learning_rate": 0.00019898705162690598, "loss": 4.4387, "step": 181550 }, { "epoch": 0.33888579166268173, "grad_norm": 1.0914325714111328, "learning_rate": 0.00019898645511328016, "loss": 4.8009, "step": 181600 }, { "epoch": 0.3389790972220602, "grad_norm": 0.9410198926925659, "learning_rate": 0.00019898585842496064, "loss": 4.7744, "step": 181650 }, { "epoch": 0.33907240278143874, "grad_norm": 0.9771658778190613, "learning_rate": 0.0001989852615619485, "loss": 4.505, "step": 181700 }, { "epoch": 0.3391657083408172, "grad_norm": 0.8809220194816589, "learning_rate": 0.0001989846645242448, "loss": 4.6773, "step": 181750 }, { "epoch": 0.3392590139001957, "grad_norm": 0.9671509265899658, "learning_rate": 0.00019898406731185057, "loss": 4.573, "step": 181800 }, { "epoch": 0.3393523194595742, "grad_norm": 1.0768080949783325, "learning_rate": 0.0001989834699247669, "loss": 4.8166, "step": 181850 }, { "epoch": 0.3394456250189527, "grad_norm": 0.941792905330658, "learning_rate": 0.00019898287236299478, "loss": 4.7192, "step": 181900 }, { "epoch": 0.3395389305783312, "grad_norm": 0.8942563533782959, "learning_rate": 0.00019898227462653533, "loss": 4.5277, "step": 181950 }, { "epoch": 0.33963223613770965, "grad_norm": 0.9910199642181396, "learning_rate": 0.00019898167671538957, "loss": 4.6392, "step": 182000 }, { "epoch": 0.3397255416970882, "grad_norm": 1.0308103561401367, "learning_rate": 0.0001989810786295586, "loss": 4.6123, "step": 182050 }, { "epoch": 0.33981884725646666, "grad_norm": 0.8738309144973755, "learning_rate": 0.00019898048036904342, "loss": 4.5485, "step": 182100 }, { "epoch": 0.33991215281584514, "grad_norm": 0.8555976152420044, "learning_rate": 0.00019897988193384513, "loss": 4.6508, "step": 182150 }, { "epoch": 0.34000545837522367, "grad_norm": 0.9000989198684692, "learning_rate": 0.00019897928332396477, "loss": 4.7006, "step": 182200 }, { "epoch": 0.34009876393460214, "grad_norm": 0.8779635429382324, "learning_rate": 0.00019897868453940337, "loss": 4.7278, "step": 182250 }, { "epoch": 0.3401920694939806, "grad_norm": 1.2050081491470337, "learning_rate": 0.00019897808558016198, "loss": 4.7024, "step": 182300 }, { "epoch": 0.3402853750533591, "grad_norm": 0.8261715173721313, "learning_rate": 0.0001989774864462417, "loss": 4.6134, "step": 182350 }, { "epoch": 0.3403786806127376, "grad_norm": 1.1884026527404785, "learning_rate": 0.0001989768871376436, "loss": 4.5208, "step": 182400 }, { "epoch": 0.3404719861721161, "grad_norm": 1.0583335161209106, "learning_rate": 0.00019897628765436868, "loss": 4.628, "step": 182450 }, { "epoch": 0.3405652917314946, "grad_norm": 0.8147032856941223, "learning_rate": 0.00019897568799641805, "loss": 4.5084, "step": 182500 }, { "epoch": 0.3406585972908731, "grad_norm": 0.9693218469619751, "learning_rate": 0.00019897508816379273, "loss": 4.7667, "step": 182550 }, { "epoch": 0.3407519028502516, "grad_norm": 0.9098610877990723, "learning_rate": 0.0001989744881564938, "loss": 4.8388, "step": 182600 }, { "epoch": 0.34084520840963006, "grad_norm": 1.03932785987854, "learning_rate": 0.00019897388797452233, "loss": 4.6463, "step": 182650 }, { "epoch": 0.34093851396900854, "grad_norm": 0.9611286520957947, "learning_rate": 0.00019897328761787932, "loss": 4.7354, "step": 182700 }, { "epoch": 0.34103181952838707, "grad_norm": 1.1163007020950317, "learning_rate": 0.00019897268708656592, "loss": 4.6021, "step": 182750 }, { "epoch": 0.34112512508776555, "grad_norm": 0.9765385985374451, "learning_rate": 0.0001989720863805831, "loss": 4.5286, "step": 182800 }, { "epoch": 0.341218430647144, "grad_norm": 0.8786355257034302, "learning_rate": 0.00019897148549993202, "loss": 4.6033, "step": 182850 }, { "epoch": 0.3413117362065225, "grad_norm": 1.1458057165145874, "learning_rate": 0.00019897088444461363, "loss": 4.7663, "step": 182900 }, { "epoch": 0.34140504176590103, "grad_norm": 0.8316869139671326, "learning_rate": 0.00019897028321462905, "loss": 4.5552, "step": 182950 }, { "epoch": 0.3414983473252795, "grad_norm": 0.9072234630584717, "learning_rate": 0.00019896968180997934, "loss": 4.6009, "step": 183000 }, { "epoch": 0.3414983473252795, "eval_loss": 4.803169250488281, "eval_runtime": 229.034, "eval_samples_per_second": 11.387, "eval_steps_per_second": 11.387, "eval_tts_loss": 7.459224440081708, "step": 183000 }, { "epoch": 0.341591652884658, "grad_norm": 1.083984613418579, "learning_rate": 0.00019896908023066557, "loss": 4.469, "step": 183050 }, { "epoch": 0.3416849584440365, "grad_norm": 0.8765878677368164, "learning_rate": 0.00019896847847668877, "loss": 4.4339, "step": 183100 }, { "epoch": 0.341778264003415, "grad_norm": 0.9148985743522644, "learning_rate": 0.00019896787654805005, "loss": 4.6406, "step": 183150 }, { "epoch": 0.34187156956279346, "grad_norm": 0.8526818156242371, "learning_rate": 0.00019896727444475038, "loss": 4.7993, "step": 183200 }, { "epoch": 0.34196487512217194, "grad_norm": 0.9056457877159119, "learning_rate": 0.00019896667216679094, "loss": 4.4859, "step": 183250 }, { "epoch": 0.34205818068155047, "grad_norm": 1.0208319425582886, "learning_rate": 0.00019896606971417274, "loss": 4.7113, "step": 183300 }, { "epoch": 0.34215148624092895, "grad_norm": 0.9874845743179321, "learning_rate": 0.0001989654670868968, "loss": 4.5711, "step": 183350 }, { "epoch": 0.3422447918003074, "grad_norm": 0.988519549369812, "learning_rate": 0.00019896486428496425, "loss": 4.6615, "step": 183400 }, { "epoch": 0.34233809735968596, "grad_norm": 1.0255403518676758, "learning_rate": 0.0001989642613083761, "loss": 4.7423, "step": 183450 }, { "epoch": 0.34243140291906443, "grad_norm": 0.8926201462745667, "learning_rate": 0.0001989636581571335, "loss": 4.5053, "step": 183500 }, { "epoch": 0.3425247084784429, "grad_norm": 1.1894657611846924, "learning_rate": 0.0001989630548312374, "loss": 4.6994, "step": 183550 }, { "epoch": 0.3426180140378214, "grad_norm": 1.066410779953003, "learning_rate": 0.00019896245133068895, "loss": 4.6081, "step": 183600 }, { "epoch": 0.3427113195971999, "grad_norm": 1.3439453840255737, "learning_rate": 0.00019896184765548917, "loss": 4.6694, "step": 183650 }, { "epoch": 0.3428046251565784, "grad_norm": 1.0843331813812256, "learning_rate": 0.00019896124380563916, "loss": 4.5804, "step": 183700 }, { "epoch": 0.34289793071595687, "grad_norm": 1.108058214187622, "learning_rate": 0.00019896063978113996, "loss": 4.5375, "step": 183750 }, { "epoch": 0.3429912362753354, "grad_norm": 1.012802004814148, "learning_rate": 0.00019896003558199263, "loss": 4.5888, "step": 183800 }, { "epoch": 0.3430845418347139, "grad_norm": 0.9825184941291809, "learning_rate": 0.0001989594312081983, "loss": 4.6243, "step": 183850 }, { "epoch": 0.34317784739409235, "grad_norm": 0.7795212268829346, "learning_rate": 0.0001989588266597579, "loss": 4.805, "step": 183900 }, { "epoch": 0.3432711529534708, "grad_norm": 1.6016143560409546, "learning_rate": 0.00019895822193667265, "loss": 4.6542, "step": 183950 }, { "epoch": 0.34336445851284936, "grad_norm": 1.0046799182891846, "learning_rate": 0.00019895761703894355, "loss": 4.5883, "step": 184000 }, { "epoch": 0.34345776407222783, "grad_norm": 0.7315368056297302, "learning_rate": 0.00019895701196657163, "loss": 4.6325, "step": 184050 }, { "epoch": 0.3435510696316063, "grad_norm": 0.8666321635246277, "learning_rate": 0.00019895640671955804, "loss": 4.5639, "step": 184100 }, { "epoch": 0.34364437519098484, "grad_norm": 1.0640748739242554, "learning_rate": 0.00019895580129790377, "loss": 4.5574, "step": 184150 }, { "epoch": 0.3437376807503633, "grad_norm": 0.8577877879142761, "learning_rate": 0.00019895519570160996, "loss": 4.6898, "step": 184200 }, { "epoch": 0.3438309863097418, "grad_norm": 0.7682492733001709, "learning_rate": 0.0001989545899306776, "loss": 4.7193, "step": 184250 }, { "epoch": 0.34392429186912027, "grad_norm": 1.053924322128296, "learning_rate": 0.00019895398398510778, "loss": 4.8604, "step": 184300 }, { "epoch": 0.3440175974284988, "grad_norm": 0.9362823367118835, "learning_rate": 0.00019895337786490162, "loss": 4.3165, "step": 184350 }, { "epoch": 0.3441109029878773, "grad_norm": 0.9843419194221497, "learning_rate": 0.0001989527715700602, "loss": 4.3989, "step": 184400 }, { "epoch": 0.34420420854725575, "grad_norm": 1.27715265750885, "learning_rate": 0.0001989521651005845, "loss": 4.6078, "step": 184450 }, { "epoch": 0.3442975141066343, "grad_norm": 0.9157127141952515, "learning_rate": 0.00019895155845647567, "loss": 4.4988, "step": 184500 }, { "epoch": 0.34439081966601276, "grad_norm": 0.9360002875328064, "learning_rate": 0.00019895095163773472, "loss": 4.6132, "step": 184550 }, { "epoch": 0.34448412522539124, "grad_norm": 1.044089674949646, "learning_rate": 0.00019895034464436276, "loss": 4.351, "step": 184600 }, { "epoch": 0.3445774307847697, "grad_norm": 1.0638103485107422, "learning_rate": 0.00019894973747636084, "loss": 4.7351, "step": 184650 }, { "epoch": 0.34467073634414824, "grad_norm": 1.2173103094100952, "learning_rate": 0.00019894913013373007, "loss": 4.4605, "step": 184700 }, { "epoch": 0.3447640419035267, "grad_norm": 0.8456254601478577, "learning_rate": 0.00019894852261647148, "loss": 4.8245, "step": 184750 }, { "epoch": 0.3448573474629052, "grad_norm": 1.3228527307510376, "learning_rate": 0.00019894791492458617, "loss": 4.6516, "step": 184800 }, { "epoch": 0.3449506530222837, "grad_norm": 0.8793355226516724, "learning_rate": 0.0001989473070580752, "loss": 4.7263, "step": 184850 }, { "epoch": 0.3450439585816622, "grad_norm": 0.733553409576416, "learning_rate": 0.00019894669901693966, "loss": 4.5974, "step": 184900 }, { "epoch": 0.3451372641410407, "grad_norm": 0.9788616895675659, "learning_rate": 0.00019894609080118057, "loss": 4.4838, "step": 184950 }, { "epoch": 0.34523056970041915, "grad_norm": 1.099247932434082, "learning_rate": 0.00019894548241079907, "loss": 4.5225, "step": 185000 }, { "epoch": 0.3453238752597977, "grad_norm": 0.9463627934455872, "learning_rate": 0.00019894487384579618, "loss": 4.8215, "step": 185050 }, { "epoch": 0.34541718081917616, "grad_norm": 0.9453304409980774, "learning_rate": 0.000198944265106173, "loss": 4.4883, "step": 185100 }, { "epoch": 0.34551048637855464, "grad_norm": 1.1768165826797485, "learning_rate": 0.0001989436561919306, "loss": 4.8437, "step": 185150 }, { "epoch": 0.34560379193793317, "grad_norm": 0.7917547821998596, "learning_rate": 0.00019894304710307007, "loss": 4.3636, "step": 185200 }, { "epoch": 0.34569709749731165, "grad_norm": 0.802765429019928, "learning_rate": 0.00019894243783959246, "loss": 4.6204, "step": 185250 }, { "epoch": 0.3457904030566901, "grad_norm": 0.9512039422988892, "learning_rate": 0.00019894182840149885, "loss": 4.4626, "step": 185300 }, { "epoch": 0.3458837086160686, "grad_norm": 0.9868209362030029, "learning_rate": 0.00019894121878879034, "loss": 4.8379, "step": 185350 }, { "epoch": 0.34597701417544713, "grad_norm": 1.1373767852783203, "learning_rate": 0.000198940609001468, "loss": 4.4935, "step": 185400 }, { "epoch": 0.3460703197348256, "grad_norm": 0.8839678168296814, "learning_rate": 0.00019893999903953288, "loss": 4.7252, "step": 185450 }, { "epoch": 0.3461636252942041, "grad_norm": 0.8720890283584595, "learning_rate": 0.00019893938890298607, "loss": 4.7323, "step": 185500 }, { "epoch": 0.34625693085358256, "grad_norm": 0.9795693159103394, "learning_rate": 0.00019893877859182864, "loss": 4.8515, "step": 185550 }, { "epoch": 0.3463502364129611, "grad_norm": 1.0126317739486694, "learning_rate": 0.00019893816810606166, "loss": 4.7654, "step": 185600 }, { "epoch": 0.34644354197233956, "grad_norm": 1.0032151937484741, "learning_rate": 0.00019893755744568626, "loss": 4.638, "step": 185650 }, { "epoch": 0.34653684753171804, "grad_norm": 0.7030245661735535, "learning_rate": 0.00019893694661070349, "loss": 4.7943, "step": 185700 }, { "epoch": 0.34663015309109657, "grad_norm": 0.9134154915809631, "learning_rate": 0.00019893633560111437, "loss": 4.7419, "step": 185750 }, { "epoch": 0.34672345865047505, "grad_norm": 0.6795088052749634, "learning_rate": 0.00019893572441692005, "loss": 4.593, "step": 185800 }, { "epoch": 0.3468167642098535, "grad_norm": 0.9457302093505859, "learning_rate": 0.00019893511305812159, "loss": 4.6045, "step": 185850 }, { "epoch": 0.346910069769232, "grad_norm": 1.1078051328659058, "learning_rate": 0.00019893450152472003, "loss": 4.7646, "step": 185900 }, { "epoch": 0.34700337532861053, "grad_norm": 0.8041667342185974, "learning_rate": 0.00019893388981671653, "loss": 4.6, "step": 185950 }, { "epoch": 0.347096680887989, "grad_norm": 1.0429846048355103, "learning_rate": 0.00019893327793411207, "loss": 4.6922, "step": 186000 }, { "epoch": 0.347096680887989, "eval_loss": 4.805553913116455, "eval_runtime": 227.9003, "eval_samples_per_second": 11.444, "eval_steps_per_second": 11.444, "eval_tts_loss": 7.4744320140360205, "step": 186000 }, { "epoch": 0.3471899864473675, "grad_norm": 1.188132643699646, "learning_rate": 0.0001989326658769078, "loss": 4.6133, "step": 186050 }, { "epoch": 0.347283292006746, "grad_norm": 1.105686902999878, "learning_rate": 0.00019893205364510482, "loss": 4.5588, "step": 186100 }, { "epoch": 0.3473765975661245, "grad_norm": 1.0260311365127563, "learning_rate": 0.00019893144123870414, "loss": 4.5501, "step": 186150 }, { "epoch": 0.34746990312550297, "grad_norm": 1.099698781967163, "learning_rate": 0.00019893082865770687, "loss": 4.6763, "step": 186200 }, { "epoch": 0.34756320868488144, "grad_norm": 1.0089246034622192, "learning_rate": 0.0001989302159021141, "loss": 4.6909, "step": 186250 }, { "epoch": 0.34765651424426, "grad_norm": 1.0701543092727661, "learning_rate": 0.00019892960297192692, "loss": 4.6315, "step": 186300 }, { "epoch": 0.34774981980363845, "grad_norm": 1.1907925605773926, "learning_rate": 0.00019892898986714638, "loss": 4.4744, "step": 186350 }, { "epoch": 0.3478431253630169, "grad_norm": 1.0035996437072754, "learning_rate": 0.00019892837658777362, "loss": 4.6013, "step": 186400 }, { "epoch": 0.34793643092239546, "grad_norm": 1.0162055492401123, "learning_rate": 0.00019892776313380965, "loss": 4.4274, "step": 186450 }, { "epoch": 0.34802973648177393, "grad_norm": 1.1313321590423584, "learning_rate": 0.0001989271495052556, "loss": 4.7547, "step": 186500 }, { "epoch": 0.3481230420411524, "grad_norm": 0.8484470248222351, "learning_rate": 0.0001989265357021125, "loss": 4.8249, "step": 186550 }, { "epoch": 0.3482163476005309, "grad_norm": 0.9785260558128357, "learning_rate": 0.00019892592172438152, "loss": 4.7419, "step": 186600 }, { "epoch": 0.3483096531599094, "grad_norm": 0.9270712733268738, "learning_rate": 0.00019892530757206368, "loss": 4.7485, "step": 186650 }, { "epoch": 0.3484029587192879, "grad_norm": 1.1253376007080078, "learning_rate": 0.00019892469324516004, "loss": 4.6492, "step": 186700 }, { "epoch": 0.34849626427866637, "grad_norm": 0.8123875856399536, "learning_rate": 0.00019892407874367175, "loss": 4.5601, "step": 186750 }, { "epoch": 0.3485895698380449, "grad_norm": 0.9842862486839294, "learning_rate": 0.00019892346406759988, "loss": 4.6747, "step": 186800 }, { "epoch": 0.3486828753974234, "grad_norm": 1.0450785160064697, "learning_rate": 0.0001989228492169455, "loss": 4.6732, "step": 186850 }, { "epoch": 0.34877618095680185, "grad_norm": 1.2406466007232666, "learning_rate": 0.00019892223419170968, "loss": 4.5914, "step": 186900 }, { "epoch": 0.3488694865161803, "grad_norm": 1.202197790145874, "learning_rate": 0.0001989216189918935, "loss": 4.4839, "step": 186950 }, { "epoch": 0.34896279207555886, "grad_norm": 0.9749503135681152, "learning_rate": 0.00019892100361749813, "loss": 4.5923, "step": 187000 }, { "epoch": 0.34905609763493733, "grad_norm": 0.7598263621330261, "learning_rate": 0.00019892038806852455, "loss": 4.582, "step": 187050 }, { "epoch": 0.3491494031943158, "grad_norm": 1.0366371870040894, "learning_rate": 0.00019891977234497393, "loss": 4.6446, "step": 187100 }, { "epoch": 0.34924270875369434, "grad_norm": 0.82780522108078, "learning_rate": 0.0001989191564468473, "loss": 4.5483, "step": 187150 }, { "epoch": 0.3493360143130728, "grad_norm": 1.048930287361145, "learning_rate": 0.00019891854037414573, "loss": 4.6858, "step": 187200 }, { "epoch": 0.3494293198724513, "grad_norm": 1.0502203702926636, "learning_rate": 0.00019891792412687038, "loss": 4.821, "step": 187250 }, { "epoch": 0.34952262543182977, "grad_norm": 1.079811930656433, "learning_rate": 0.00019891730770502228, "loss": 4.6184, "step": 187300 }, { "epoch": 0.3496159309912083, "grad_norm": 0.9850397109985352, "learning_rate": 0.00019891669110860255, "loss": 4.7636, "step": 187350 }, { "epoch": 0.3497092365505868, "grad_norm": 1.421713948249817, "learning_rate": 0.00019891607433761226, "loss": 4.8562, "step": 187400 }, { "epoch": 0.34980254210996525, "grad_norm": 0.9206252098083496, "learning_rate": 0.00019891545739205249, "loss": 4.47, "step": 187450 }, { "epoch": 0.3498958476693438, "grad_norm": 0.9205539226531982, "learning_rate": 0.00019891484027192435, "loss": 4.5468, "step": 187500 }, { "epoch": 0.34998915322872226, "grad_norm": 1.1044520139694214, "learning_rate": 0.00019891422297722893, "loss": 4.7065, "step": 187550 }, { "epoch": 0.35008245878810074, "grad_norm": 1.2163364887237549, "learning_rate": 0.0001989136055079673, "loss": 4.5895, "step": 187600 }, { "epoch": 0.3501757643474792, "grad_norm": 1.0337022542953491, "learning_rate": 0.00019891298786414057, "loss": 4.752, "step": 187650 }, { "epoch": 0.35026906990685774, "grad_norm": 0.9887984395027161, "learning_rate": 0.0001989123700457498, "loss": 4.498, "step": 187700 }, { "epoch": 0.3503623754662362, "grad_norm": 0.8707548975944519, "learning_rate": 0.00019891175205279614, "loss": 4.4742, "step": 187750 }, { "epoch": 0.3504556810256147, "grad_norm": 0.9432954788208008, "learning_rate": 0.00019891113388528062, "loss": 4.7022, "step": 187800 }, { "epoch": 0.3505489865849932, "grad_norm": 1.063997745513916, "learning_rate": 0.00019891051554320438, "loss": 4.4942, "step": 187850 }, { "epoch": 0.3506422921443717, "grad_norm": 1.1391524076461792, "learning_rate": 0.00019890989702656844, "loss": 4.6385, "step": 187900 }, { "epoch": 0.3507355977037502, "grad_norm": 1.218582034111023, "learning_rate": 0.00019890927833537397, "loss": 4.665, "step": 187950 }, { "epoch": 0.35082890326312866, "grad_norm": 0.8306301832199097, "learning_rate": 0.00019890865946962198, "loss": 4.7438, "step": 188000 }, { "epoch": 0.3509222088225072, "grad_norm": 0.9843810796737671, "learning_rate": 0.00019890804042931362, "loss": 4.5901, "step": 188050 }, { "epoch": 0.35101551438188566, "grad_norm": 0.9525366425514221, "learning_rate": 0.00019890742121445002, "loss": 4.6783, "step": 188100 }, { "epoch": 0.35110881994126414, "grad_norm": 1.0917189121246338, "learning_rate": 0.00019890680182503218, "loss": 4.6924, "step": 188150 }, { "epoch": 0.3512021255006426, "grad_norm": 1.1915470361709595, "learning_rate": 0.00019890618226106125, "loss": 4.4894, "step": 188200 }, { "epoch": 0.35129543106002115, "grad_norm": 1.0157119035720825, "learning_rate": 0.00019890556252253832, "loss": 4.5125, "step": 188250 }, { "epoch": 0.3513887366193996, "grad_norm": 1.2147808074951172, "learning_rate": 0.0001989049426094645, "loss": 4.474, "step": 188300 }, { "epoch": 0.3514820421787781, "grad_norm": 0.9446842074394226, "learning_rate": 0.00019890432252184082, "loss": 4.6241, "step": 188350 }, { "epoch": 0.35157534773815663, "grad_norm": 0.6960589289665222, "learning_rate": 0.00019890370225966844, "loss": 4.5706, "step": 188400 }, { "epoch": 0.3516686532975351, "grad_norm": 0.8260095119476318, "learning_rate": 0.00019890308182294838, "loss": 4.7308, "step": 188450 }, { "epoch": 0.3517619588569136, "grad_norm": 0.9620236158370972, "learning_rate": 0.00019890246121168184, "loss": 4.5599, "step": 188500 }, { "epoch": 0.35185526441629206, "grad_norm": 0.8720660209655762, "learning_rate": 0.00019890184042586985, "loss": 4.5758, "step": 188550 }, { "epoch": 0.3519485699756706, "grad_norm": 1.1599175930023193, "learning_rate": 0.00019890121946551347, "loss": 4.7073, "step": 188600 }, { "epoch": 0.35204187553504906, "grad_norm": 1.200167179107666, "learning_rate": 0.00019890059833061388, "loss": 4.6389, "step": 188650 }, { "epoch": 0.35213518109442754, "grad_norm": 0.7152448296546936, "learning_rate": 0.0001988999770211721, "loss": 4.6249, "step": 188700 }, { "epoch": 0.35222848665380607, "grad_norm": 0.6849294900894165, "learning_rate": 0.0001988993555371893, "loss": 4.6701, "step": 188750 }, { "epoch": 0.35232179221318455, "grad_norm": 1.2049897909164429, "learning_rate": 0.00019889873387866653, "loss": 4.6354, "step": 188800 }, { "epoch": 0.352415097772563, "grad_norm": 0.9554798603057861, "learning_rate": 0.0001988981120456049, "loss": 4.4661, "step": 188850 }, { "epoch": 0.3525084033319415, "grad_norm": 1.052039623260498, "learning_rate": 0.0001988974900380055, "loss": 4.6176, "step": 188900 }, { "epoch": 0.35260170889132003, "grad_norm": 1.091506838798523, "learning_rate": 0.00019889686785586942, "loss": 4.801, "step": 188950 }, { "epoch": 0.3526950144506985, "grad_norm": 1.3394882678985596, "learning_rate": 0.00019889624549919777, "loss": 4.722, "step": 189000 }, { "epoch": 0.3526950144506985, "eval_loss": 4.806208610534668, "eval_runtime": 230.3394, "eval_samples_per_second": 11.322, "eval_steps_per_second": 11.322, "eval_tts_loss": 7.53537485083748, "step": 189000 }, { "epoch": 0.352788320010077, "grad_norm": 0.8463157415390015, "learning_rate": 0.00019889562296799166, "loss": 4.6327, "step": 189050 }, { "epoch": 0.3528816255694555, "grad_norm": 1.0070785284042358, "learning_rate": 0.00019889500026225217, "loss": 4.6743, "step": 189100 }, { "epoch": 0.352974931128834, "grad_norm": 0.8200750946998596, "learning_rate": 0.0001988943773819804, "loss": 4.3624, "step": 189150 }, { "epoch": 0.35306823668821247, "grad_norm": 0.9505131840705872, "learning_rate": 0.00019889375432717747, "loss": 4.8476, "step": 189200 }, { "epoch": 0.35316154224759094, "grad_norm": 0.9525319337844849, "learning_rate": 0.00019889313109784445, "loss": 4.511, "step": 189250 }, { "epoch": 0.3532548478069695, "grad_norm": 0.9823152422904968, "learning_rate": 0.00019889250769398247, "loss": 4.6123, "step": 189300 }, { "epoch": 0.35334815336634795, "grad_norm": 1.0408767461776733, "learning_rate": 0.0001988918841155926, "loss": 4.6926, "step": 189350 }, { "epoch": 0.3534414589257264, "grad_norm": 1.0379136800765991, "learning_rate": 0.00019889126036267595, "loss": 4.6514, "step": 189400 }, { "epoch": 0.35353476448510496, "grad_norm": 0.7835113406181335, "learning_rate": 0.00019889063643523366, "loss": 4.5634, "step": 189450 }, { "epoch": 0.35362807004448343, "grad_norm": 1.5030099153518677, "learning_rate": 0.00019889001233326677, "loss": 4.5985, "step": 189500 }, { "epoch": 0.3537213756038619, "grad_norm": 0.9151412844657898, "learning_rate": 0.0001988893880567764, "loss": 4.7988, "step": 189550 }, { "epoch": 0.3538146811632404, "grad_norm": 0.9782933592796326, "learning_rate": 0.00019888876360576367, "loss": 4.6997, "step": 189600 }, { "epoch": 0.3539079867226189, "grad_norm": 0.9720525145530701, "learning_rate": 0.00019888813898022968, "loss": 4.8163, "step": 189650 }, { "epoch": 0.3540012922819974, "grad_norm": 1.0412356853485107, "learning_rate": 0.00019888751418017554, "loss": 4.7658, "step": 189700 }, { "epoch": 0.35409459784137587, "grad_norm": 0.5875607132911682, "learning_rate": 0.0001988868892056023, "loss": 4.8729, "step": 189750 }, { "epoch": 0.3541879034007544, "grad_norm": 1.0475389957427979, "learning_rate": 0.00019888626405651112, "loss": 4.517, "step": 189800 }, { "epoch": 0.3542812089601329, "grad_norm": 1.3192886114120483, "learning_rate": 0.0001988856387329031, "loss": 4.5192, "step": 189850 }, { "epoch": 0.35437451451951135, "grad_norm": 1.179880142211914, "learning_rate": 0.00019888501323477928, "loss": 4.4085, "step": 189900 }, { "epoch": 0.35446782007888983, "grad_norm": 1.334714651107788, "learning_rate": 0.00019888438756214086, "loss": 4.7391, "step": 189950 }, { "epoch": 0.35456112563826836, "grad_norm": 1.2655901908874512, "learning_rate": 0.00019888376171498887, "loss": 4.756, "step": 190000 }, { "epoch": 0.35465443119764684, "grad_norm": 1.0103163719177246, "learning_rate": 0.00019888313569332443, "loss": 4.6622, "step": 190050 }, { "epoch": 0.3547477367570253, "grad_norm": 0.8582555055618286, "learning_rate": 0.00019888250949714865, "loss": 4.4321, "step": 190100 }, { "epoch": 0.35484104231640384, "grad_norm": 1.1723741292953491, "learning_rate": 0.00019888188312646267, "loss": 4.5144, "step": 190150 }, { "epoch": 0.3549343478757823, "grad_norm": 1.05409574508667, "learning_rate": 0.00019888125658126758, "loss": 4.618, "step": 190200 }, { "epoch": 0.3550276534351608, "grad_norm": 1.0219589471817017, "learning_rate": 0.00019888062986156443, "loss": 4.632, "step": 190250 }, { "epoch": 0.35512095899453927, "grad_norm": 1.1344385147094727, "learning_rate": 0.00019888000296735436, "loss": 4.4802, "step": 190300 }, { "epoch": 0.3552142645539178, "grad_norm": 1.0866692066192627, "learning_rate": 0.00019887937589863853, "loss": 4.6449, "step": 190350 }, { "epoch": 0.3553075701132963, "grad_norm": 1.0079729557037354, "learning_rate": 0.00019887874865541795, "loss": 4.6405, "step": 190400 }, { "epoch": 0.35540087567267475, "grad_norm": 1.0848729610443115, "learning_rate": 0.00019887812123769384, "loss": 4.6839, "step": 190450 }, { "epoch": 0.3554941812320533, "grad_norm": 0.8781519532203674, "learning_rate": 0.00019887749364546717, "loss": 4.5597, "step": 190500 }, { "epoch": 0.35558748679143176, "grad_norm": 0.9141809940338135, "learning_rate": 0.00019887686587873916, "loss": 4.577, "step": 190550 }, { "epoch": 0.35568079235081024, "grad_norm": 0.9193921685218811, "learning_rate": 0.00019887623793751088, "loss": 4.6156, "step": 190600 }, { "epoch": 0.3557740979101887, "grad_norm": 0.9867975115776062, "learning_rate": 0.00019887560982178343, "loss": 4.4942, "step": 190650 }, { "epoch": 0.35586740346956725, "grad_norm": 0.9040959477424622, "learning_rate": 0.00019887498153155794, "loss": 4.8029, "step": 190700 }, { "epoch": 0.3559607090289457, "grad_norm": 0.9523190855979919, "learning_rate": 0.00019887435306683549, "loss": 4.5111, "step": 190750 }, { "epoch": 0.3560540145883242, "grad_norm": 1.135631799697876, "learning_rate": 0.00019887372442761722, "loss": 4.7086, "step": 190800 }, { "epoch": 0.3561473201477027, "grad_norm": 1.1821600198745728, "learning_rate": 0.0001988730956139042, "loss": 4.524, "step": 190850 }, { "epoch": 0.3562406257070812, "grad_norm": 1.0236471891403198, "learning_rate": 0.00019887246662569758, "loss": 4.7486, "step": 190900 }, { "epoch": 0.3563339312664597, "grad_norm": 1.091172695159912, "learning_rate": 0.00019887183746299846, "loss": 4.572, "step": 190950 }, { "epoch": 0.35642723682583816, "grad_norm": 1.0206681489944458, "learning_rate": 0.00019887120812580792, "loss": 4.6126, "step": 191000 }, { "epoch": 0.3565205423852167, "grad_norm": 1.0186724662780762, "learning_rate": 0.00019887057861412712, "loss": 4.5049, "step": 191050 }, { "epoch": 0.35661384794459516, "grad_norm": 0.817267656326294, "learning_rate": 0.0001988699489279571, "loss": 4.6261, "step": 191100 }, { "epoch": 0.35670715350397364, "grad_norm": 1.2809371948242188, "learning_rate": 0.0001988693190672991, "loss": 4.7523, "step": 191150 }, { "epoch": 0.3568004590633521, "grad_norm": 0.6853193044662476, "learning_rate": 0.00019886868903215408, "loss": 4.4515, "step": 191200 }, { "epoch": 0.35689376462273065, "grad_norm": 1.1701523065567017, "learning_rate": 0.00019886805882252324, "loss": 4.6187, "step": 191250 }, { "epoch": 0.3569870701821091, "grad_norm": 0.8031530976295471, "learning_rate": 0.0001988674284384077, "loss": 4.5504, "step": 191300 }, { "epoch": 0.3570803757414876, "grad_norm": 0.855839192867279, "learning_rate": 0.0001988667978798085, "loss": 4.3099, "step": 191350 }, { "epoch": 0.35717368130086613, "grad_norm": 1.233931541442871, "learning_rate": 0.00019886616714672681, "loss": 4.5372, "step": 191400 }, { "epoch": 0.3572669868602446, "grad_norm": 1.1093002557754517, "learning_rate": 0.00019886553623916372, "loss": 4.8146, "step": 191450 }, { "epoch": 0.3573602924196231, "grad_norm": 1.2051666975021362, "learning_rate": 0.00019886490515712038, "loss": 4.7033, "step": 191500 }, { "epoch": 0.35745359797900156, "grad_norm": 1.0855170488357544, "learning_rate": 0.00019886427390059788, "loss": 4.5491, "step": 191550 }, { "epoch": 0.3575469035383801, "grad_norm": 1.0766313076019287, "learning_rate": 0.00019886364246959732, "loss": 4.5545, "step": 191600 }, { "epoch": 0.35764020909775857, "grad_norm": 0.8042473793029785, "learning_rate": 0.00019886301086411982, "loss": 4.8249, "step": 191650 }, { "epoch": 0.35773351465713704, "grad_norm": 0.9476457834243774, "learning_rate": 0.00019886237908416652, "loss": 4.8088, "step": 191700 }, { "epoch": 0.3578268202165156, "grad_norm": 1.0122379064559937, "learning_rate": 0.0001988617471297385, "loss": 4.4633, "step": 191750 }, { "epoch": 0.35792012577589405, "grad_norm": 1.0943959951400757, "learning_rate": 0.00019886111500083692, "loss": 4.4537, "step": 191800 }, { "epoch": 0.3580134313352725, "grad_norm": 0.9375251531600952, "learning_rate": 0.00019886048269746281, "loss": 4.7861, "step": 191850 }, { "epoch": 0.358106736894651, "grad_norm": 1.0100510120391846, "learning_rate": 0.00019885985021961738, "loss": 4.6012, "step": 191900 }, { "epoch": 0.35820004245402953, "grad_norm": 1.1179404258728027, "learning_rate": 0.00019885921756730173, "loss": 4.771, "step": 191950 }, { "epoch": 0.358293348013408, "grad_norm": 1.1079944372177124, "learning_rate": 0.00019885858474051694, "loss": 4.5788, "step": 192000 }, { "epoch": 0.358293348013408, "eval_loss": 4.794641971588135, "eval_runtime": 230.5927, "eval_samples_per_second": 11.31, "eval_steps_per_second": 11.31, "eval_tts_loss": 7.481033778209468, "step": 192000 }, { "epoch": 0.3583866535727865, "grad_norm": 1.0501129627227783, "learning_rate": 0.00019885795173926412, "loss": 4.589, "step": 192050 }, { "epoch": 0.358479959132165, "grad_norm": 0.9563352465629578, "learning_rate": 0.00019885731856354442, "loss": 4.7751, "step": 192100 }, { "epoch": 0.3585732646915435, "grad_norm": 1.539056658744812, "learning_rate": 0.00019885668521335898, "loss": 4.6909, "step": 192150 }, { "epoch": 0.35866657025092197, "grad_norm": 1.1265029907226562, "learning_rate": 0.00019885605168870884, "loss": 4.4262, "step": 192200 }, { "epoch": 0.35875987581030044, "grad_norm": 0.973284900188446, "learning_rate": 0.00019885541798959517, "loss": 4.8141, "step": 192250 }, { "epoch": 0.358853181369679, "grad_norm": 0.6788082718849182, "learning_rate": 0.0001988547841160191, "loss": 4.6917, "step": 192300 }, { "epoch": 0.35894648692905745, "grad_norm": 0.7461669445037842, "learning_rate": 0.0001988541500679817, "loss": 4.7468, "step": 192350 }, { "epoch": 0.3590397924884359, "grad_norm": 1.1224901676177979, "learning_rate": 0.00019885351584548414, "loss": 4.5873, "step": 192400 }, { "epoch": 0.35913309804781446, "grad_norm": 1.1169931888580322, "learning_rate": 0.00019885288144852754, "loss": 4.7951, "step": 192450 }, { "epoch": 0.35922640360719293, "grad_norm": 0.9802398085594177, "learning_rate": 0.00019885224687711298, "loss": 4.5911, "step": 192500 }, { "epoch": 0.3593197091665714, "grad_norm": 0.9591051340103149, "learning_rate": 0.00019885161213124158, "loss": 4.4823, "step": 192550 }, { "epoch": 0.3594130147259499, "grad_norm": 1.2416292428970337, "learning_rate": 0.0001988509772109145, "loss": 4.7929, "step": 192600 }, { "epoch": 0.3595063202853284, "grad_norm": 1.196144461631775, "learning_rate": 0.00019885034211613282, "loss": 4.6477, "step": 192650 }, { "epoch": 0.3595996258447069, "grad_norm": 1.0120468139648438, "learning_rate": 0.0001988497068468977, "loss": 4.7484, "step": 192700 }, { "epoch": 0.35969293140408537, "grad_norm": 0.7862088680267334, "learning_rate": 0.00019884907140321024, "loss": 4.5306, "step": 192750 }, { "epoch": 0.3597862369634639, "grad_norm": 1.0786339044570923, "learning_rate": 0.00019884843578507155, "loss": 4.861, "step": 192800 }, { "epoch": 0.3598795425228424, "grad_norm": 1.0364389419555664, "learning_rate": 0.00019884779999248278, "loss": 4.6999, "step": 192850 }, { "epoch": 0.35997284808222085, "grad_norm": 0.887571394443512, "learning_rate": 0.00019884716402544502, "loss": 4.6757, "step": 192900 }, { "epoch": 0.36006615364159933, "grad_norm": 0.6657192707061768, "learning_rate": 0.00019884652788395942, "loss": 4.5924, "step": 192950 }, { "epoch": 0.36015945920097786, "grad_norm": 1.246086597442627, "learning_rate": 0.00019884589156802707, "loss": 4.5792, "step": 193000 }, { "epoch": 0.36025276476035634, "grad_norm": 0.8808768391609192, "learning_rate": 0.00019884525507764912, "loss": 4.6044, "step": 193050 }, { "epoch": 0.3603460703197348, "grad_norm": 0.8309186697006226, "learning_rate": 0.00019884461841282672, "loss": 4.6616, "step": 193100 }, { "epoch": 0.36043937587911334, "grad_norm": 0.9923182129859924, "learning_rate": 0.0001988439815735609, "loss": 4.6831, "step": 193150 }, { "epoch": 0.3605326814384918, "grad_norm": 0.8415591716766357, "learning_rate": 0.0001988433445598529, "loss": 4.648, "step": 193200 }, { "epoch": 0.3606259869978703, "grad_norm": 1.2085046768188477, "learning_rate": 0.00019884270737170377, "loss": 4.429, "step": 193250 }, { "epoch": 0.36071929255724877, "grad_norm": 0.8806056976318359, "learning_rate": 0.00019884207000911465, "loss": 4.4001, "step": 193300 }, { "epoch": 0.3608125981166273, "grad_norm": 1.0138684511184692, "learning_rate": 0.00019884143247208665, "loss": 4.6307, "step": 193350 }, { "epoch": 0.3609059036760058, "grad_norm": 0.8585569858551025, "learning_rate": 0.00019884079476062094, "loss": 4.5959, "step": 193400 }, { "epoch": 0.36099920923538426, "grad_norm": 0.9443112015724182, "learning_rate": 0.00019884015687471863, "loss": 4.7276, "step": 193450 }, { "epoch": 0.3610925147947628, "grad_norm": 1.2051678895950317, "learning_rate": 0.00019883951881438078, "loss": 4.795, "step": 193500 }, { "epoch": 0.36118582035414126, "grad_norm": 0.7949090600013733, "learning_rate": 0.00019883888057960864, "loss": 4.7397, "step": 193550 }, { "epoch": 0.36127912591351974, "grad_norm": 0.9525768756866455, "learning_rate": 0.00019883824217040323, "loss": 4.6333, "step": 193600 }, { "epoch": 0.3613724314728982, "grad_norm": 1.2094985246658325, "learning_rate": 0.0001988376035867657, "loss": 4.5322, "step": 193650 }, { "epoch": 0.36146573703227675, "grad_norm": 1.0645558834075928, "learning_rate": 0.00019883696482869722, "loss": 4.7803, "step": 193700 }, { "epoch": 0.3615590425916552, "grad_norm": 1.0554648637771606, "learning_rate": 0.00019883632589619887, "loss": 4.5194, "step": 193750 }, { "epoch": 0.3616523481510337, "grad_norm": 0.8595163822174072, "learning_rate": 0.0001988356867892718, "loss": 4.6625, "step": 193800 }, { "epoch": 0.3617456537104122, "grad_norm": 0.9319025874137878, "learning_rate": 0.00019883504750791713, "loss": 4.7127, "step": 193850 }, { "epoch": 0.3618389592697907, "grad_norm": 1.2545781135559082, "learning_rate": 0.000198834408052136, "loss": 4.409, "step": 193900 }, { "epoch": 0.3619322648291692, "grad_norm": 1.0958348512649536, "learning_rate": 0.00019883376842192953, "loss": 4.629, "step": 193950 }, { "epoch": 0.36202557038854766, "grad_norm": 1.1290863752365112, "learning_rate": 0.00019883312861729887, "loss": 4.7757, "step": 194000 }, { "epoch": 0.3621188759479262, "grad_norm": 1.168502926826477, "learning_rate": 0.0001988324886382451, "loss": 4.5306, "step": 194050 }, { "epoch": 0.36221218150730466, "grad_norm": 1.0988664627075195, "learning_rate": 0.00019883184848476938, "loss": 4.5708, "step": 194100 }, { "epoch": 0.36230548706668314, "grad_norm": 1.10366952419281, "learning_rate": 0.00019883120815687285, "loss": 4.7049, "step": 194150 }, { "epoch": 0.3623987926260616, "grad_norm": 1.0056383609771729, "learning_rate": 0.0001988305676545566, "loss": 4.8057, "step": 194200 }, { "epoch": 0.36249209818544015, "grad_norm": 0.930894672870636, "learning_rate": 0.00019882992697782182, "loss": 4.5286, "step": 194250 }, { "epoch": 0.3625854037448186, "grad_norm": 1.0776501893997192, "learning_rate": 0.0001988292861266696, "loss": 4.56, "step": 194300 }, { "epoch": 0.3626787093041971, "grad_norm": 1.0516318082809448, "learning_rate": 0.00019882864510110105, "loss": 4.5943, "step": 194350 }, { "epoch": 0.36277201486357563, "grad_norm": 1.0238537788391113, "learning_rate": 0.00019882800390111735, "loss": 4.6839, "step": 194400 }, { "epoch": 0.3628653204229541, "grad_norm": 1.1146060228347778, "learning_rate": 0.00019882736252671963, "loss": 4.6616, "step": 194450 }, { "epoch": 0.3629586259823326, "grad_norm": 1.2036668062210083, "learning_rate": 0.00019882672097790898, "loss": 4.5828, "step": 194500 }, { "epoch": 0.36305193154171106, "grad_norm": 0.9518182873725891, "learning_rate": 0.00019882607925468656, "loss": 4.5861, "step": 194550 }, { "epoch": 0.3631452371010896, "grad_norm": 0.8470454812049866, "learning_rate": 0.0001988254373570535, "loss": 4.7907, "step": 194600 }, { "epoch": 0.36323854266046807, "grad_norm": 0.9194369316101074, "learning_rate": 0.00019882479528501095, "loss": 4.4232, "step": 194650 }, { "epoch": 0.36333184821984654, "grad_norm": 1.164222002029419, "learning_rate": 0.00019882415303856002, "loss": 4.4877, "step": 194700 }, { "epoch": 0.3634251537792251, "grad_norm": 0.9167250394821167, "learning_rate": 0.00019882351061770184, "loss": 4.5728, "step": 194750 }, { "epoch": 0.36351845933860355, "grad_norm": 0.8055945634841919, "learning_rate": 0.00019882286802243754, "loss": 4.6115, "step": 194800 }, { "epoch": 0.363611764897982, "grad_norm": 0.8844305872917175, "learning_rate": 0.0001988222252527683, "loss": 4.6615, "step": 194850 }, { "epoch": 0.3637050704573605, "grad_norm": 1.018107533454895, "learning_rate": 0.0001988215823086952, "loss": 4.6986, "step": 194900 }, { "epoch": 0.36379837601673903, "grad_norm": 0.80892413854599, "learning_rate": 0.00019882093919021938, "loss": 4.5929, "step": 194950 }, { "epoch": 0.3638916815761175, "grad_norm": 0.9718637466430664, "learning_rate": 0.00019882029589734203, "loss": 4.5956, "step": 195000 }, { "epoch": 0.3638916815761175, "eval_loss": 4.797578811645508, "eval_runtime": 230.7126, "eval_samples_per_second": 11.304, "eval_steps_per_second": 11.304, "eval_tts_loss": 7.486753235031148, "step": 195000 }, { "epoch": 0.363984987135496, "grad_norm": 0.9084148406982422, "learning_rate": 0.0001988196524300642, "loss": 4.5751, "step": 195050 }, { "epoch": 0.3640782926948745, "grad_norm": 1.0336567163467407, "learning_rate": 0.0001988190087883871, "loss": 4.7999, "step": 195100 }, { "epoch": 0.364171598254253, "grad_norm": 0.9693698287010193, "learning_rate": 0.00019881836497231183, "loss": 4.3249, "step": 195150 }, { "epoch": 0.36426490381363147, "grad_norm": 1.1222256422042847, "learning_rate": 0.00019881772098183956, "loss": 4.587, "step": 195200 }, { "epoch": 0.36435820937300994, "grad_norm": 1.0114097595214844, "learning_rate": 0.00019881707681697135, "loss": 4.6811, "step": 195250 }, { "epoch": 0.3644515149323885, "grad_norm": 0.9249664545059204, "learning_rate": 0.00019881643247770842, "loss": 4.8543, "step": 195300 }, { "epoch": 0.36454482049176695, "grad_norm": 1.0353797674179077, "learning_rate": 0.00019881578796405186, "loss": 4.6228, "step": 195350 }, { "epoch": 0.36463812605114543, "grad_norm": 1.0884183645248413, "learning_rate": 0.00019881514327600282, "loss": 4.7143, "step": 195400 }, { "epoch": 0.36473143161052396, "grad_norm": 1.049566388130188, "learning_rate": 0.00019881449841356247, "loss": 4.6686, "step": 195450 }, { "epoch": 0.36482473716990244, "grad_norm": 0.990344226360321, "learning_rate": 0.00019881385337673188, "loss": 4.585, "step": 195500 }, { "epoch": 0.3649180427292809, "grad_norm": 1.0016299486160278, "learning_rate": 0.00019881320816551225, "loss": 4.5803, "step": 195550 }, { "epoch": 0.3650113482886594, "grad_norm": 0.9439250826835632, "learning_rate": 0.0001988125627799047, "loss": 4.5019, "step": 195600 }, { "epoch": 0.3651046538480379, "grad_norm": 1.0324641466140747, "learning_rate": 0.00019881191721991034, "loss": 4.708, "step": 195650 }, { "epoch": 0.3651979594074164, "grad_norm": 1.1070998907089233, "learning_rate": 0.0001988112714855303, "loss": 4.6045, "step": 195700 }, { "epoch": 0.36529126496679487, "grad_norm": 0.7762126922607422, "learning_rate": 0.0001988106255767658, "loss": 4.9177, "step": 195750 }, { "epoch": 0.3653845705261734, "grad_norm": 1.03263258934021, "learning_rate": 0.00019880997949361793, "loss": 4.7215, "step": 195800 }, { "epoch": 0.3654778760855519, "grad_norm": 0.9962678551673889, "learning_rate": 0.00019880933323608783, "loss": 4.49, "step": 195850 }, { "epoch": 0.36557118164493035, "grad_norm": 0.8162546157836914, "learning_rate": 0.00019880868680417663, "loss": 4.7263, "step": 195900 }, { "epoch": 0.36566448720430883, "grad_norm": 0.7211110591888428, "learning_rate": 0.0001988080401978855, "loss": 4.7463, "step": 195950 }, { "epoch": 0.36575779276368736, "grad_norm": 0.9785488843917847, "learning_rate": 0.00019880739341721556, "loss": 4.7713, "step": 196000 }, { "epoch": 0.36585109832306584, "grad_norm": 0.8973463773727417, "learning_rate": 0.00019880674646216795, "loss": 4.7429, "step": 196050 }, { "epoch": 0.3659444038824443, "grad_norm": 1.1833984851837158, "learning_rate": 0.00019880609933274382, "loss": 4.5631, "step": 196100 }, { "epoch": 0.36603770944182284, "grad_norm": 1.098944067955017, "learning_rate": 0.0001988054520289443, "loss": 4.4881, "step": 196150 }, { "epoch": 0.3661310150012013, "grad_norm": 0.6843777894973755, "learning_rate": 0.00019880480455077059, "loss": 4.525, "step": 196200 }, { "epoch": 0.3662243205605798, "grad_norm": 0.9798048138618469, "learning_rate": 0.00019880415689822373, "loss": 4.5468, "step": 196250 }, { "epoch": 0.3663176261199583, "grad_norm": 1.0058033466339111, "learning_rate": 0.00019880350907130494, "loss": 4.5075, "step": 196300 }, { "epoch": 0.3664109316793368, "grad_norm": 0.8590142130851746, "learning_rate": 0.00019880286107001535, "loss": 4.5171, "step": 196350 }, { "epoch": 0.3665042372387153, "grad_norm": 0.8589456081390381, "learning_rate": 0.00019880221289435607, "loss": 4.672, "step": 196400 }, { "epoch": 0.36659754279809376, "grad_norm": 1.3543787002563477, "learning_rate": 0.00019880156454432827, "loss": 4.523, "step": 196450 }, { "epoch": 0.36669084835747223, "grad_norm": 0.9485663771629333, "learning_rate": 0.00019880091601993312, "loss": 4.8311, "step": 196500 }, { "epoch": 0.36678415391685076, "grad_norm": 0.8935502767562866, "learning_rate": 0.00019880026732117174, "loss": 4.604, "step": 196550 }, { "epoch": 0.36687745947622924, "grad_norm": 0.9361104369163513, "learning_rate": 0.00019879961844804524, "loss": 4.6418, "step": 196600 }, { "epoch": 0.3669707650356077, "grad_norm": 0.6516358852386475, "learning_rate": 0.00019879896940055481, "loss": 4.7734, "step": 196650 }, { "epoch": 0.36706407059498625, "grad_norm": 1.2583208084106445, "learning_rate": 0.0001987983201787016, "loss": 4.6789, "step": 196700 }, { "epoch": 0.3671573761543647, "grad_norm": 1.1848986148834229, "learning_rate": 0.00019879767078248673, "loss": 4.6755, "step": 196750 }, { "epoch": 0.3672506817137432, "grad_norm": 0.871610164642334, "learning_rate": 0.00019879702121191134, "loss": 4.5271, "step": 196800 }, { "epoch": 0.3673439872731217, "grad_norm": 1.1172187328338623, "learning_rate": 0.00019879637146697662, "loss": 4.6236, "step": 196850 }, { "epoch": 0.3674372928325002, "grad_norm": 0.910994827747345, "learning_rate": 0.00019879572154768367, "loss": 4.5114, "step": 196900 }, { "epoch": 0.3675305983918787, "grad_norm": 0.9143189787864685, "learning_rate": 0.00019879507145403365, "loss": 4.4976, "step": 196950 }, { "epoch": 0.36762390395125716, "grad_norm": 1.1715046167373657, "learning_rate": 0.0001987944211860277, "loss": 4.6836, "step": 197000 }, { "epoch": 0.3677172095106357, "grad_norm": 1.333890438079834, "learning_rate": 0.00019879377074366703, "loss": 4.7598, "step": 197050 }, { "epoch": 0.36781051507001417, "grad_norm": 1.0626753568649292, "learning_rate": 0.00019879312012695272, "loss": 4.7344, "step": 197100 }, { "epoch": 0.36790382062939264, "grad_norm": 0.8466847538948059, "learning_rate": 0.0001987924693358859, "loss": 4.6338, "step": 197150 }, { "epoch": 0.3679971261887711, "grad_norm": 1.0398815870285034, "learning_rate": 0.00019879181837046777, "loss": 4.5708, "step": 197200 }, { "epoch": 0.36809043174814965, "grad_norm": 1.2056119441986084, "learning_rate": 0.00019879116723069948, "loss": 4.8655, "step": 197250 }, { "epoch": 0.3681837373075281, "grad_norm": 1.0409513711929321, "learning_rate": 0.00019879051591658215, "loss": 4.6327, "step": 197300 }, { "epoch": 0.3682770428669066, "grad_norm": 1.0986347198486328, "learning_rate": 0.00019878986442811695, "loss": 4.6958, "step": 197350 }, { "epoch": 0.36837034842628513, "grad_norm": 1.093003749847412, "learning_rate": 0.00019878921276530502, "loss": 4.6851, "step": 197400 }, { "epoch": 0.3684636539856636, "grad_norm": 1.156713843345642, "learning_rate": 0.0001987885609281475, "loss": 4.7068, "step": 197450 }, { "epoch": 0.3685569595450421, "grad_norm": 1.0253217220306396, "learning_rate": 0.00019878790891664555, "loss": 4.5718, "step": 197500 }, { "epoch": 0.36865026510442056, "grad_norm": 0.8153467774391174, "learning_rate": 0.00019878725673080038, "loss": 4.6435, "step": 197550 }, { "epoch": 0.3687435706637991, "grad_norm": 0.8387562036514282, "learning_rate": 0.00019878660437061302, "loss": 4.733, "step": 197600 }, { "epoch": 0.36883687622317757, "grad_norm": 1.011487603187561, "learning_rate": 0.0001987859518360847, "loss": 4.772, "step": 197650 }, { "epoch": 0.36893018178255604, "grad_norm": 1.247372031211853, "learning_rate": 0.00019878529912721658, "loss": 4.6005, "step": 197700 }, { "epoch": 0.3690234873419346, "grad_norm": 0.8897690176963806, "learning_rate": 0.00019878464624400978, "loss": 4.6325, "step": 197750 }, { "epoch": 0.36911679290131305, "grad_norm": 1.1363788843154907, "learning_rate": 0.00019878399318646543, "loss": 4.7405, "step": 197800 }, { "epoch": 0.3692100984606915, "grad_norm": 0.8495743870735168, "learning_rate": 0.00019878333995458473, "loss": 4.4341, "step": 197850 }, { "epoch": 0.36930340402007, "grad_norm": 0.9996325969696045, "learning_rate": 0.00019878268654836882, "loss": 4.6559, "step": 197900 }, { "epoch": 0.36939670957944853, "grad_norm": 0.7758403420448303, "learning_rate": 0.00019878203296781885, "loss": 5.0731, "step": 197950 }, { "epoch": 0.369490015138827, "grad_norm": 0.9682093262672424, "learning_rate": 0.000198781379212936, "loss": 4.4734, "step": 198000 }, { "epoch": 0.369490015138827, "eval_loss": 4.794241905212402, "eval_runtime": 232.0461, "eval_samples_per_second": 11.239, "eval_steps_per_second": 11.239, "eval_tts_loss": 7.463459329881014, "step": 198000 }, { "epoch": 0.3695833206982055, "grad_norm": 1.051621675491333, "learning_rate": 0.00019878072528372136, "loss": 4.6863, "step": 198050 }, { "epoch": 0.369676626257584, "grad_norm": 1.0777297019958496, "learning_rate": 0.0001987800711801761, "loss": 4.5781, "step": 198100 }, { "epoch": 0.3697699318169625, "grad_norm": 0.8551693558692932, "learning_rate": 0.00019877941690230145, "loss": 4.6632, "step": 198150 }, { "epoch": 0.36986323737634097, "grad_norm": 0.8395803570747375, "learning_rate": 0.00019877876245009849, "loss": 4.6038, "step": 198200 }, { "epoch": 0.36995654293571945, "grad_norm": 0.860999584197998, "learning_rate": 0.0001987781078235684, "loss": 4.7329, "step": 198250 }, { "epoch": 0.370049848495098, "grad_norm": 1.099967360496521, "learning_rate": 0.0001987774530227123, "loss": 4.7567, "step": 198300 }, { "epoch": 0.37014315405447645, "grad_norm": 0.6936938166618347, "learning_rate": 0.00019877679804753141, "loss": 4.673, "step": 198350 }, { "epoch": 0.37023645961385493, "grad_norm": 0.959591805934906, "learning_rate": 0.00019877614289802686, "loss": 4.6583, "step": 198400 }, { "epoch": 0.37032976517323346, "grad_norm": 0.7232271432876587, "learning_rate": 0.00019877548757419976, "loss": 4.5433, "step": 198450 }, { "epoch": 0.37042307073261194, "grad_norm": 1.075395941734314, "learning_rate": 0.00019877483207605132, "loss": 4.5049, "step": 198500 }, { "epoch": 0.3705163762919904, "grad_norm": 1.0577384233474731, "learning_rate": 0.0001987741764035827, "loss": 4.642, "step": 198550 }, { "epoch": 0.3706096818513689, "grad_norm": 1.5923864841461182, "learning_rate": 0.000198773520556795, "loss": 4.7332, "step": 198600 }, { "epoch": 0.3707029874107474, "grad_norm": 0.9335559010505676, "learning_rate": 0.00019877286453568943, "loss": 4.4329, "step": 198650 }, { "epoch": 0.3707962929701259, "grad_norm": 1.0967572927474976, "learning_rate": 0.00019877220834026718, "loss": 4.4424, "step": 198700 }, { "epoch": 0.37088959852950437, "grad_norm": 1.1251230239868164, "learning_rate": 0.00019877155197052933, "loss": 4.5098, "step": 198750 }, { "epoch": 0.3709829040888829, "grad_norm": 1.2461490631103516, "learning_rate": 0.00019877089542647705, "loss": 4.636, "step": 198800 }, { "epoch": 0.3710762096482614, "grad_norm": 0.9544318914413452, "learning_rate": 0.00019877023870811154, "loss": 4.6384, "step": 198850 }, { "epoch": 0.37116951520763986, "grad_norm": 1.099353313446045, "learning_rate": 0.00019876958181543392, "loss": 4.7218, "step": 198900 }, { "epoch": 0.37126282076701833, "grad_norm": 0.7116580009460449, "learning_rate": 0.0001987689247484454, "loss": 4.5868, "step": 198950 }, { "epoch": 0.37135612632639686, "grad_norm": 0.9305589199066162, "learning_rate": 0.00019876826750714707, "loss": 4.5696, "step": 199000 }, { "epoch": 0.37144943188577534, "grad_norm": 1.0949960947036743, "learning_rate": 0.00019876761009154014, "loss": 4.4849, "step": 199050 }, { "epoch": 0.3715427374451538, "grad_norm": 0.9973856210708618, "learning_rate": 0.00019876695250162576, "loss": 4.6221, "step": 199100 }, { "epoch": 0.3716360430045323, "grad_norm": 0.7113000154495239, "learning_rate": 0.00019876629473740507, "loss": 4.6113, "step": 199150 }, { "epoch": 0.3717293485639108, "grad_norm": 0.8312941789627075, "learning_rate": 0.00019876563679887928, "loss": 4.6111, "step": 199200 }, { "epoch": 0.3718226541232893, "grad_norm": 1.2139105796813965, "learning_rate": 0.00019876497868604948, "loss": 4.5554, "step": 199250 }, { "epoch": 0.3719159596826678, "grad_norm": 1.0658022165298462, "learning_rate": 0.0001987643203989169, "loss": 4.6281, "step": 199300 }, { "epoch": 0.3720092652420463, "grad_norm": 0.9115583300590515, "learning_rate": 0.00019876366193748269, "loss": 4.8648, "step": 199350 }, { "epoch": 0.3721025708014248, "grad_norm": 1.0521410703659058, "learning_rate": 0.00019876300330174798, "loss": 4.5097, "step": 199400 }, { "epoch": 0.37219587636080326, "grad_norm": 0.9952971935272217, "learning_rate": 0.00019876234449171391, "loss": 4.7379, "step": 199450 }, { "epoch": 0.37228918192018173, "grad_norm": 0.8925333023071289, "learning_rate": 0.00019876168550738172, "loss": 4.5746, "step": 199500 }, { "epoch": 0.37238248747956026, "grad_norm": 1.0621514320373535, "learning_rate": 0.00019876102634875254, "loss": 4.7674, "step": 199550 }, { "epoch": 0.37247579303893874, "grad_norm": 1.0929023027420044, "learning_rate": 0.00019876036701582748, "loss": 4.7718, "step": 199600 }, { "epoch": 0.3725690985983172, "grad_norm": 0.9549880027770996, "learning_rate": 0.00019875970750860778, "loss": 4.7339, "step": 199650 }, { "epoch": 0.37266240415769575, "grad_norm": 0.9763375520706177, "learning_rate": 0.0001987590478270946, "loss": 4.6824, "step": 199700 }, { "epoch": 0.3727557097170742, "grad_norm": 0.9068183898925781, "learning_rate": 0.00019875838797128902, "loss": 4.7193, "step": 199750 }, { "epoch": 0.3728490152764527, "grad_norm": 0.9717890024185181, "learning_rate": 0.0001987577279411923, "loss": 4.5489, "step": 199800 }, { "epoch": 0.3729423208358312, "grad_norm": 1.2245581150054932, "learning_rate": 0.00019875706773680554, "loss": 4.705, "step": 199850 }, { "epoch": 0.3730356263952097, "grad_norm": 0.7146151065826416, "learning_rate": 0.00019875640735812997, "loss": 4.5574, "step": 199900 }, { "epoch": 0.3731289319545882, "grad_norm": 0.8383366465568542, "learning_rate": 0.00019875574680516667, "loss": 4.5159, "step": 199950 }, { "epoch": 0.37322223751396666, "grad_norm": 1.15836501121521, "learning_rate": 0.0001987550860779169, "loss": 4.8317, "step": 200000 }, { "epoch": 0.3733155430733452, "grad_norm": 0.8754977583885193, "learning_rate": 0.00019875442517638177, "loss": 4.7887, "step": 200050 }, { "epoch": 0.37340884863272367, "grad_norm": 0.8341622352600098, "learning_rate": 0.00019875376410056244, "loss": 4.5602, "step": 200100 }, { "epoch": 0.37350215419210214, "grad_norm": 0.9436699151992798, "learning_rate": 0.00019875310285046008, "loss": 4.5878, "step": 200150 }, { "epoch": 0.3735954597514806, "grad_norm": 1.203100562095642, "learning_rate": 0.0001987524414260759, "loss": 4.5758, "step": 200200 }, { "epoch": 0.37368876531085915, "grad_norm": 0.9803832173347473, "learning_rate": 0.00019875177982741106, "loss": 4.5436, "step": 200250 }, { "epoch": 0.3737820708702376, "grad_norm": 0.9167838096618652, "learning_rate": 0.00019875111805446664, "loss": 4.7414, "step": 200300 }, { "epoch": 0.3738753764296161, "grad_norm": 1.015909194946289, "learning_rate": 0.0001987504561072439, "loss": 4.7295, "step": 200350 }, { "epoch": 0.37396868198899463, "grad_norm": 1.1135481595993042, "learning_rate": 0.00019874979398574398, "loss": 4.4958, "step": 200400 }, { "epoch": 0.3740619875483731, "grad_norm": 0.9369817972183228, "learning_rate": 0.00019874913168996808, "loss": 4.5166, "step": 200450 }, { "epoch": 0.3741552931077516, "grad_norm": 1.2627172470092773, "learning_rate": 0.0001987484692199173, "loss": 4.661, "step": 200500 }, { "epoch": 0.37424859866713006, "grad_norm": 1.0819026231765747, "learning_rate": 0.00019874780657559287, "loss": 4.4839, "step": 200550 }, { "epoch": 0.3743419042265086, "grad_norm": 1.0878807306289673, "learning_rate": 0.0001987471437569959, "loss": 4.5501, "step": 200600 }, { "epoch": 0.37443520978588707, "grad_norm": 1.0913496017456055, "learning_rate": 0.00019874648076412762, "loss": 4.9156, "step": 200650 }, { "epoch": 0.37452851534526554, "grad_norm": 0.9594210386276245, "learning_rate": 0.00019874581759698917, "loss": 4.5145, "step": 200700 }, { "epoch": 0.3746218209046441, "grad_norm": 0.8627813458442688, "learning_rate": 0.00019874515425558173, "loss": 4.5963, "step": 200750 }, { "epoch": 0.37471512646402255, "grad_norm": 0.9256874918937683, "learning_rate": 0.0001987444907399065, "loss": 4.671, "step": 200800 }, { "epoch": 0.37480843202340103, "grad_norm": 1.5067219734191895, "learning_rate": 0.0001987438270499646, "loss": 4.672, "step": 200850 }, { "epoch": 0.3749017375827795, "grad_norm": 1.0874707698822021, "learning_rate": 0.00019874316318575718, "loss": 4.6663, "step": 200900 }, { "epoch": 0.37499504314215804, "grad_norm": 1.2468639612197876, "learning_rate": 0.00019874249914728548, "loss": 4.6853, "step": 200950 }, { "epoch": 0.3750883487015365, "grad_norm": 0.7165033221244812, "learning_rate": 0.00019874183493455064, "loss": 4.6766, "step": 201000 }, { "epoch": 0.3750883487015365, "eval_loss": 4.802398681640625, "eval_runtime": 230.2804, "eval_samples_per_second": 11.325, "eval_steps_per_second": 11.325, "eval_tts_loss": 7.475072805292653, "step": 201000 }, { "epoch": 0.375181654260915, "grad_norm": 1.1223444938659668, "learning_rate": 0.00019874117054755386, "loss": 4.6664, "step": 201050 }, { "epoch": 0.3752749598202935, "grad_norm": 1.1241016387939453, "learning_rate": 0.00019874050598629625, "loss": 4.5235, "step": 201100 }, { "epoch": 0.375368265379672, "grad_norm": 1.3051515817642212, "learning_rate": 0.00019873984125077905, "loss": 4.4631, "step": 201150 }, { "epoch": 0.37546157093905047, "grad_norm": 1.0892348289489746, "learning_rate": 0.0001987391763410034, "loss": 4.4371, "step": 201200 }, { "epoch": 0.37555487649842895, "grad_norm": 0.9838119745254517, "learning_rate": 0.00019873851125697044, "loss": 4.723, "step": 201250 }, { "epoch": 0.3756481820578075, "grad_norm": 1.3931784629821777, "learning_rate": 0.00019873784599868145, "loss": 4.7364, "step": 201300 }, { "epoch": 0.37574148761718595, "grad_norm": 0.9986215829849243, "learning_rate": 0.00019873718056613747, "loss": 4.4411, "step": 201350 }, { "epoch": 0.37583479317656443, "grad_norm": 1.0121062994003296, "learning_rate": 0.0001987365149593398, "loss": 4.3812, "step": 201400 }, { "epoch": 0.37592809873594296, "grad_norm": 1.1041994094848633, "learning_rate": 0.0001987358491782895, "loss": 4.4639, "step": 201450 }, { "epoch": 0.37602140429532144, "grad_norm": 0.7851685285568237, "learning_rate": 0.00019873518322298782, "loss": 4.4774, "step": 201500 }, { "epoch": 0.3761147098546999, "grad_norm": 1.1037462949752808, "learning_rate": 0.0001987345170934359, "loss": 4.6946, "step": 201550 }, { "epoch": 0.3762080154140784, "grad_norm": 0.9130716323852539, "learning_rate": 0.00019873385078963496, "loss": 4.6075, "step": 201600 }, { "epoch": 0.3763013209734569, "grad_norm": 1.2021543979644775, "learning_rate": 0.00019873318431158613, "loss": 4.5221, "step": 201650 }, { "epoch": 0.3763946265328354, "grad_norm": 0.9670273065567017, "learning_rate": 0.0001987325176592906, "loss": 4.8346, "step": 201700 }, { "epoch": 0.3764879320922139, "grad_norm": 0.8241245746612549, "learning_rate": 0.00019873185083274956, "loss": 4.487, "step": 201750 }, { "epoch": 0.37658123765159235, "grad_norm": 1.071000337600708, "learning_rate": 0.00019873118383196417, "loss": 4.5168, "step": 201800 }, { "epoch": 0.3766745432109709, "grad_norm": 0.627001166343689, "learning_rate": 0.0001987305166569356, "loss": 4.8239, "step": 201850 }, { "epoch": 0.37676784877034936, "grad_norm": 0.781301736831665, "learning_rate": 0.00019872984930766505, "loss": 4.4841, "step": 201900 }, { "epoch": 0.37686115432972783, "grad_norm": 0.7564511895179749, "learning_rate": 0.0001987291817841537, "loss": 4.4971, "step": 201950 }, { "epoch": 0.37695445988910636, "grad_norm": 0.9705767035484314, "learning_rate": 0.0001987285140864027, "loss": 4.4673, "step": 202000 }, { "epoch": 0.37704776544848484, "grad_norm": 0.9428151845932007, "learning_rate": 0.00019872784621441326, "loss": 4.8026, "step": 202050 }, { "epoch": 0.3771410710078633, "grad_norm": 1.1641000509262085, "learning_rate": 0.00019872717816818653, "loss": 4.7319, "step": 202100 }, { "epoch": 0.3772343765672418, "grad_norm": 1.1046082973480225, "learning_rate": 0.0001987265099477237, "loss": 4.8214, "step": 202150 }, { "epoch": 0.3773276821266203, "grad_norm": 1.2519506216049194, "learning_rate": 0.00019872584155302593, "loss": 4.6536, "step": 202200 }, { "epoch": 0.3774209876859988, "grad_norm": 1.1127455234527588, "learning_rate": 0.0001987251729840945, "loss": 4.5374, "step": 202250 }, { "epoch": 0.3775142932453773, "grad_norm": 1.0181680917739868, "learning_rate": 0.00019872450424093043, "loss": 4.4758, "step": 202300 }, { "epoch": 0.3776075988047558, "grad_norm": 0.7517116069793701, "learning_rate": 0.000198723835323535, "loss": 4.5575, "step": 202350 }, { "epoch": 0.3777009043641343, "grad_norm": 0.9754074215888977, "learning_rate": 0.00019872316623190938, "loss": 4.5489, "step": 202400 }, { "epoch": 0.37779420992351276, "grad_norm": 0.9626929759979248, "learning_rate": 0.00019872249696605473, "loss": 4.7142, "step": 202450 }, { "epoch": 0.37788751548289123, "grad_norm": 0.8055295944213867, "learning_rate": 0.0001987218275259723, "loss": 4.591, "step": 202500 }, { "epoch": 0.37798082104226977, "grad_norm": 0.821767270565033, "learning_rate": 0.00019872115791166314, "loss": 4.5601, "step": 202550 }, { "epoch": 0.37807412660164824, "grad_norm": 1.3287692070007324, "learning_rate": 0.00019872048812312855, "loss": 4.6671, "step": 202600 }, { "epoch": 0.3781674321610267, "grad_norm": 0.9818217158317566, "learning_rate": 0.00019871981816036965, "loss": 4.638, "step": 202650 }, { "epoch": 0.37826073772040525, "grad_norm": 1.010910987854004, "learning_rate": 0.00019871914802338763, "loss": 4.742, "step": 202700 }, { "epoch": 0.3783540432797837, "grad_norm": 1.0474064350128174, "learning_rate": 0.00019871847771218374, "loss": 4.7676, "step": 202750 }, { "epoch": 0.3784473488391622, "grad_norm": 0.8638759255409241, "learning_rate": 0.00019871780722675904, "loss": 4.7774, "step": 202800 }, { "epoch": 0.3785406543985407, "grad_norm": 1.0039918422698975, "learning_rate": 0.00019871713656711482, "loss": 4.5449, "step": 202850 }, { "epoch": 0.3786339599579192, "grad_norm": 1.126681923866272, "learning_rate": 0.00019871646573325222, "loss": 4.3922, "step": 202900 }, { "epoch": 0.3787272655172977, "grad_norm": 1.3228673934936523, "learning_rate": 0.00019871579472517241, "loss": 4.6475, "step": 202950 }, { "epoch": 0.37882057107667616, "grad_norm": 0.9036591649055481, "learning_rate": 0.00019871512354287664, "loss": 4.7047, "step": 203000 }, { "epoch": 0.3789138766360547, "grad_norm": 0.9258872866630554, "learning_rate": 0.00019871445218636597, "loss": 4.6349, "step": 203050 }, { "epoch": 0.37900718219543317, "grad_norm": 2.290501356124878, "learning_rate": 0.0001987137806556417, "loss": 4.4937, "step": 203100 }, { "epoch": 0.37910048775481164, "grad_norm": 0.8478452563285828, "learning_rate": 0.000198713108950705, "loss": 4.5709, "step": 203150 }, { "epoch": 0.3791937933141901, "grad_norm": 1.1010254621505737, "learning_rate": 0.000198712437071557, "loss": 4.7398, "step": 203200 }, { "epoch": 0.37928709887356865, "grad_norm": 1.4688420295715332, "learning_rate": 0.00019871176501819892, "loss": 4.6441, "step": 203250 }, { "epoch": 0.3793804044329471, "grad_norm": 1.065325379371643, "learning_rate": 0.00019871109279063196, "loss": 4.4881, "step": 203300 }, { "epoch": 0.3794737099923256, "grad_norm": 0.981877863407135, "learning_rate": 0.00019871042038885727, "loss": 4.5731, "step": 203350 }, { "epoch": 0.37956701555170413, "grad_norm": 1.0577549934387207, "learning_rate": 0.0001987097478128761, "loss": 4.7339, "step": 203400 }, { "epoch": 0.3796603211110826, "grad_norm": 0.9602108597755432, "learning_rate": 0.00019870907506268953, "loss": 4.5568, "step": 203450 }, { "epoch": 0.3797536266704611, "grad_norm": 1.1359353065490723, "learning_rate": 0.00019870840213829884, "loss": 4.6005, "step": 203500 }, { "epoch": 0.37984693222983956, "grad_norm": 0.9004436731338501, "learning_rate": 0.0001987077290397052, "loss": 4.7455, "step": 203550 }, { "epoch": 0.3799402377892181, "grad_norm": 1.0926775932312012, "learning_rate": 0.0001987070557669098, "loss": 4.4615, "step": 203600 }, { "epoch": 0.38003354334859657, "grad_norm": 0.8713423609733582, "learning_rate": 0.00019870638231991376, "loss": 4.4985, "step": 203650 }, { "epoch": 0.38012684890797505, "grad_norm": 0.6343261003494263, "learning_rate": 0.00019870570869871837, "loss": 4.437, "step": 203700 }, { "epoch": 0.3802201544673536, "grad_norm": 0.7554951310157776, "learning_rate": 0.00019870503490332475, "loss": 4.3759, "step": 203750 }, { "epoch": 0.38031346002673205, "grad_norm": 0.9022654891014099, "learning_rate": 0.0001987043609337341, "loss": 4.5305, "step": 203800 }, { "epoch": 0.38040676558611053, "grad_norm": 1.4872987270355225, "learning_rate": 0.00019870368678994766, "loss": 4.659, "step": 203850 }, { "epoch": 0.380500071145489, "grad_norm": 0.8781129121780396, "learning_rate": 0.00019870301247196655, "loss": 4.5687, "step": 203900 }, { "epoch": 0.38059337670486754, "grad_norm": 0.9872531294822693, "learning_rate": 0.000198702337979792, "loss": 4.4754, "step": 203950 }, { "epoch": 0.380686682264246, "grad_norm": 1.1369433403015137, "learning_rate": 0.0001987016633134252, "loss": 4.6713, "step": 204000 }, { "epoch": 0.380686682264246, "eval_loss": 4.801127910614014, "eval_runtime": 232.9265, "eval_samples_per_second": 11.197, "eval_steps_per_second": 11.197, "eval_tts_loss": 7.454781288711034, "step": 204000 }, { "epoch": 0.3807799878236245, "grad_norm": 1.0344752073287964, "learning_rate": 0.0001987009884728673, "loss": 4.5242, "step": 204050 }, { "epoch": 0.380873293383003, "grad_norm": 0.8404310345649719, "learning_rate": 0.00019870031345811955, "loss": 4.6894, "step": 204100 }, { "epoch": 0.3809665989423815, "grad_norm": 1.081472635269165, "learning_rate": 0.00019869963826918313, "loss": 4.5426, "step": 204150 }, { "epoch": 0.38105990450175997, "grad_norm": 0.9155741333961487, "learning_rate": 0.00019869896290605917, "loss": 4.781, "step": 204200 }, { "epoch": 0.38115321006113845, "grad_norm": 1.0043416023254395, "learning_rate": 0.00019869828736874894, "loss": 4.6185, "step": 204250 }, { "epoch": 0.381246515620517, "grad_norm": 0.6685605049133301, "learning_rate": 0.00019869761165725356, "loss": 4.7304, "step": 204300 }, { "epoch": 0.38133982117989546, "grad_norm": 0.9836552143096924, "learning_rate": 0.00019869693577157432, "loss": 4.6497, "step": 204350 }, { "epoch": 0.38143312673927393, "grad_norm": 1.1131340265274048, "learning_rate": 0.0001986962597117123, "loss": 4.5424, "step": 204400 }, { "epoch": 0.3815264322986524, "grad_norm": 0.8396437168121338, "learning_rate": 0.00019869558347766878, "loss": 4.6054, "step": 204450 }, { "epoch": 0.38161973785803094, "grad_norm": 0.8572611808776855, "learning_rate": 0.0001986949070694449, "loss": 4.5767, "step": 204500 }, { "epoch": 0.3817130434174094, "grad_norm": 0.9373936653137207, "learning_rate": 0.0001986942304870419, "loss": 4.5014, "step": 204550 }, { "epoch": 0.3818063489767879, "grad_norm": 1.0423073768615723, "learning_rate": 0.00019869355373046093, "loss": 4.3117, "step": 204600 }, { "epoch": 0.3818996545361664, "grad_norm": 0.8610841631889343, "learning_rate": 0.0001986928767997032, "loss": 4.2993, "step": 204650 }, { "epoch": 0.3819929600955449, "grad_norm": 1.0650521516799927, "learning_rate": 0.00019869219969476993, "loss": 4.7043, "step": 204700 }, { "epoch": 0.3820862656549234, "grad_norm": 1.2617119550704956, "learning_rate": 0.0001986915224156623, "loss": 4.7328, "step": 204750 }, { "epoch": 0.38217957121430185, "grad_norm": 1.0750634670257568, "learning_rate": 0.00019869084496238146, "loss": 4.6412, "step": 204800 }, { "epoch": 0.3822728767736804, "grad_norm": 1.0128196477890015, "learning_rate": 0.00019869016733492867, "loss": 4.7131, "step": 204850 }, { "epoch": 0.38236618233305886, "grad_norm": 0.9253135919570923, "learning_rate": 0.00019868948953330508, "loss": 4.5573, "step": 204900 }, { "epoch": 0.38245948789243733, "grad_norm": 1.0896155834197998, "learning_rate": 0.00019868881155751193, "loss": 4.5266, "step": 204950 }, { "epoch": 0.38255279345181586, "grad_norm": 1.0464351177215576, "learning_rate": 0.00019868813340755037, "loss": 4.5613, "step": 205000 }, { "epoch": 0.38264609901119434, "grad_norm": 0.8733029365539551, "learning_rate": 0.00019868745508342165, "loss": 4.4978, "step": 205050 }, { "epoch": 0.3827394045705728, "grad_norm": 1.1396583318710327, "learning_rate": 0.0001986867765851269, "loss": 4.5443, "step": 205100 }, { "epoch": 0.3828327101299513, "grad_norm": 1.0709539651870728, "learning_rate": 0.00019868609791266737, "loss": 4.7764, "step": 205150 }, { "epoch": 0.3829260156893298, "grad_norm": 0.9949414134025574, "learning_rate": 0.00019868541906604425, "loss": 4.5682, "step": 205200 }, { "epoch": 0.3830193212487083, "grad_norm": 0.9429978132247925, "learning_rate": 0.0001986847400452587, "loss": 4.5577, "step": 205250 }, { "epoch": 0.3831126268080868, "grad_norm": 1.0749703645706177, "learning_rate": 0.00019868406085031198, "loss": 4.6041, "step": 205300 }, { "epoch": 0.3832059323674653, "grad_norm": 0.7519810795783997, "learning_rate": 0.00019868338148120525, "loss": 4.5119, "step": 205350 }, { "epoch": 0.3832992379268438, "grad_norm": 1.1606097221374512, "learning_rate": 0.00019868270193793973, "loss": 4.6795, "step": 205400 }, { "epoch": 0.38339254348622226, "grad_norm": 0.6909898519515991, "learning_rate": 0.00019868202222051657, "loss": 4.7247, "step": 205450 }, { "epoch": 0.38348584904560074, "grad_norm": 1.001676321029663, "learning_rate": 0.00019868134232893704, "loss": 4.6081, "step": 205500 }, { "epoch": 0.38357915460497927, "grad_norm": 1.1560254096984863, "learning_rate": 0.00019868066226320227, "loss": 4.6179, "step": 205550 }, { "epoch": 0.38367246016435774, "grad_norm": 0.7835885882377625, "learning_rate": 0.00019867998202331352, "loss": 4.5546, "step": 205600 }, { "epoch": 0.3837657657237362, "grad_norm": 0.5636594295501709, "learning_rate": 0.00019867930160927195, "loss": 4.508, "step": 205650 }, { "epoch": 0.38385907128311475, "grad_norm": 1.0891164541244507, "learning_rate": 0.00019867862102107878, "loss": 4.6448, "step": 205700 }, { "epoch": 0.3839523768424932, "grad_norm": 0.8994442820549011, "learning_rate": 0.0001986779402587352, "loss": 4.4305, "step": 205750 }, { "epoch": 0.3840456824018717, "grad_norm": 1.0475820302963257, "learning_rate": 0.00019867725932224244, "loss": 4.5979, "step": 205800 }, { "epoch": 0.3841389879612502, "grad_norm": 1.0817757844924927, "learning_rate": 0.00019867657821160165, "loss": 4.4495, "step": 205850 }, { "epoch": 0.3842322935206287, "grad_norm": 1.2010945081710815, "learning_rate": 0.0001986758969268141, "loss": 4.5591, "step": 205900 }, { "epoch": 0.3843255990800072, "grad_norm": 0.9652662873268127, "learning_rate": 0.00019867521546788091, "loss": 4.5538, "step": 205950 }, { "epoch": 0.38441890463938566, "grad_norm": 1.1052769422531128, "learning_rate": 0.00019867453383480338, "loss": 4.6479, "step": 206000 }, { "epoch": 0.3845122101987642, "grad_norm": 1.1473944187164307, "learning_rate": 0.00019867385202758262, "loss": 4.5248, "step": 206050 }, { "epoch": 0.38460551575814267, "grad_norm": 1.0132917165756226, "learning_rate": 0.00019867317004621988, "loss": 4.6242, "step": 206100 }, { "epoch": 0.38469882131752114, "grad_norm": 1.299413800239563, "learning_rate": 0.00019867248789071638, "loss": 4.6021, "step": 206150 }, { "epoch": 0.3847921268768996, "grad_norm": 1.0853779315948486, "learning_rate": 0.0001986718055610733, "loss": 4.6003, "step": 206200 }, { "epoch": 0.38488543243627815, "grad_norm": 0.9251170754432678, "learning_rate": 0.0001986711230572918, "loss": 4.5685, "step": 206250 }, { "epoch": 0.38497873799565663, "grad_norm": 1.1247254610061646, "learning_rate": 0.00019867044037937315, "loss": 4.5821, "step": 206300 }, { "epoch": 0.3850720435550351, "grad_norm": 1.1961967945098877, "learning_rate": 0.00019866975752731855, "loss": 4.7458, "step": 206350 }, { "epoch": 0.38516534911441364, "grad_norm": 0.7294309735298157, "learning_rate": 0.00019866907450112917, "loss": 4.5956, "step": 206400 }, { "epoch": 0.3852586546737921, "grad_norm": 0.7703094482421875, "learning_rate": 0.00019866839130080623, "loss": 4.6983, "step": 206450 }, { "epoch": 0.3853519602331706, "grad_norm": 0.9793773293495178, "learning_rate": 0.00019866770792635098, "loss": 4.4386, "step": 206500 }, { "epoch": 0.38544526579254906, "grad_norm": 0.9847137928009033, "learning_rate": 0.00019866702437776454, "loss": 4.6341, "step": 206550 }, { "epoch": 0.3855385713519276, "grad_norm": 1.0143849849700928, "learning_rate": 0.00019866634065504818, "loss": 4.816, "step": 206600 }, { "epoch": 0.38563187691130607, "grad_norm": 0.8812471628189087, "learning_rate": 0.0001986656567582031, "loss": 4.4307, "step": 206650 }, { "epoch": 0.38572518247068455, "grad_norm": 0.9908750057220459, "learning_rate": 0.0001986649726872305, "loss": 4.7599, "step": 206700 }, { "epoch": 0.3858184880300631, "grad_norm": 1.2358583211898804, "learning_rate": 0.00019866428844213157, "loss": 4.7192, "step": 206750 }, { "epoch": 0.38591179358944155, "grad_norm": 0.8084837794303894, "learning_rate": 0.00019866360402290753, "loss": 4.6309, "step": 206800 }, { "epoch": 0.38600509914882003, "grad_norm": 0.917667806148529, "learning_rate": 0.0001986629194295596, "loss": 4.5326, "step": 206850 }, { "epoch": 0.3860984047081985, "grad_norm": 1.1653844118118286, "learning_rate": 0.00019866223466208897, "loss": 4.5962, "step": 206900 }, { "epoch": 0.38619171026757704, "grad_norm": 0.9410626888275146, "learning_rate": 0.00019866154972049687, "loss": 4.6704, "step": 206950 }, { "epoch": 0.3862850158269555, "grad_norm": 0.8716585040092468, "learning_rate": 0.00019866086460478447, "loss": 4.6156, "step": 207000 }, { "epoch": 0.3862850158269555, "eval_loss": 4.8045244216918945, "eval_runtime": 231.8143, "eval_samples_per_second": 11.25, "eval_steps_per_second": 11.25, "eval_tts_loss": 7.503764759250659, "step": 207000 }, { "epoch": 0.386378321386334, "grad_norm": 0.9506188631057739, "learning_rate": 0.00019866017931495302, "loss": 4.6642, "step": 207050 }, { "epoch": 0.38647162694571247, "grad_norm": 1.3123427629470825, "learning_rate": 0.00019865949385100372, "loss": 4.6589, "step": 207100 }, { "epoch": 0.386564932505091, "grad_norm": 1.0116405487060547, "learning_rate": 0.00019865880821293776, "loss": 4.6218, "step": 207150 }, { "epoch": 0.3866582380644695, "grad_norm": 1.0257128477096558, "learning_rate": 0.00019865812240075635, "loss": 4.8113, "step": 207200 }, { "epoch": 0.38675154362384795, "grad_norm": 1.1300727128982544, "learning_rate": 0.00019865743641446075, "loss": 4.6969, "step": 207250 }, { "epoch": 0.3868448491832265, "grad_norm": 0.8774489164352417, "learning_rate": 0.00019865675025405213, "loss": 4.6738, "step": 207300 }, { "epoch": 0.38693815474260496, "grad_norm": 1.0406553745269775, "learning_rate": 0.00019865606391953166, "loss": 4.7365, "step": 207350 }, { "epoch": 0.38703146030198343, "grad_norm": 0.9450544118881226, "learning_rate": 0.00019865537741090063, "loss": 4.454, "step": 207400 }, { "epoch": 0.3871247658613619, "grad_norm": 0.9951105117797852, "learning_rate": 0.0001986546907281602, "loss": 4.538, "step": 207450 }, { "epoch": 0.38721807142074044, "grad_norm": 0.7700023651123047, "learning_rate": 0.00019865400387131162, "loss": 4.4626, "step": 207500 }, { "epoch": 0.3873113769801189, "grad_norm": 0.8401572108268738, "learning_rate": 0.00019865331684035609, "loss": 4.7251, "step": 207550 }, { "epoch": 0.3874046825394974, "grad_norm": 1.3303920030593872, "learning_rate": 0.0001986526296352948, "loss": 4.6399, "step": 207600 }, { "epoch": 0.3874979880988759, "grad_norm": 0.9098189473152161, "learning_rate": 0.00019865194225612896, "loss": 4.6386, "step": 207650 }, { "epoch": 0.3875912936582544, "grad_norm": 1.0674896240234375, "learning_rate": 0.00019865125470285984, "loss": 4.652, "step": 207700 }, { "epoch": 0.3876845992176329, "grad_norm": 1.1178187131881714, "learning_rate": 0.0001986505669754886, "loss": 4.5139, "step": 207750 }, { "epoch": 0.38777790477701135, "grad_norm": 0.8213180303573608, "learning_rate": 0.00019864987907401645, "loss": 4.734, "step": 207800 }, { "epoch": 0.3878712103363899, "grad_norm": 1.1217759847640991, "learning_rate": 0.00019864919099844461, "loss": 4.5176, "step": 207850 }, { "epoch": 0.38796451589576836, "grad_norm": 0.978659451007843, "learning_rate": 0.00019864850274877433, "loss": 4.6919, "step": 207900 }, { "epoch": 0.38805782145514683, "grad_norm": 1.1488454341888428, "learning_rate": 0.0001986478143250068, "loss": 4.6732, "step": 207950 }, { "epoch": 0.38815112701452537, "grad_norm": 0.9637132287025452, "learning_rate": 0.00019864712572714323, "loss": 4.6186, "step": 208000 }, { "epoch": 0.38824443257390384, "grad_norm": 1.1298731565475464, "learning_rate": 0.00019864643695518487, "loss": 4.3077, "step": 208050 }, { "epoch": 0.3883377381332823, "grad_norm": 0.889434814453125, "learning_rate": 0.00019864574800913286, "loss": 4.4539, "step": 208100 }, { "epoch": 0.3884310436926608, "grad_norm": 1.0395722389221191, "learning_rate": 0.00019864505888898847, "loss": 4.6617, "step": 208150 }, { "epoch": 0.3885243492520393, "grad_norm": 0.8869847059249878, "learning_rate": 0.00019864436959475293, "loss": 4.5674, "step": 208200 }, { "epoch": 0.3886176548114178, "grad_norm": 1.071192979812622, "learning_rate": 0.0001986436801264274, "loss": 4.7536, "step": 208250 }, { "epoch": 0.3887109603707963, "grad_norm": 1.0335016250610352, "learning_rate": 0.00019864299048401315, "loss": 4.5433, "step": 208300 }, { "epoch": 0.3888042659301748, "grad_norm": 1.0146387815475464, "learning_rate": 0.00019864230066751138, "loss": 4.6243, "step": 208350 }, { "epoch": 0.3888975714895533, "grad_norm": 1.2553081512451172, "learning_rate": 0.00019864161067692333, "loss": 4.3292, "step": 208400 }, { "epoch": 0.38899087704893176, "grad_norm": 0.8601014018058777, "learning_rate": 0.00019864092051225017, "loss": 4.6767, "step": 208450 }, { "epoch": 0.38908418260831024, "grad_norm": 0.8037707209587097, "learning_rate": 0.00019864023017349315, "loss": 4.5461, "step": 208500 }, { "epoch": 0.38917748816768877, "grad_norm": 0.9061649441719055, "learning_rate": 0.00019863953966065348, "loss": 4.8048, "step": 208550 }, { "epoch": 0.38927079372706724, "grad_norm": 1.0509523153305054, "learning_rate": 0.00019863884897373236, "loss": 4.458, "step": 208600 }, { "epoch": 0.3893640992864457, "grad_norm": 1.026106834411621, "learning_rate": 0.00019863815811273106, "loss": 4.7856, "step": 208650 }, { "epoch": 0.38945740484582425, "grad_norm": 0.9003967642784119, "learning_rate": 0.00019863746707765074, "loss": 4.7185, "step": 208700 }, { "epoch": 0.3895507104052027, "grad_norm": 1.1053556203842163, "learning_rate": 0.00019863677586849266, "loss": 4.5309, "step": 208750 }, { "epoch": 0.3896440159645812, "grad_norm": 0.8523988127708435, "learning_rate": 0.00019863608448525798, "loss": 4.7219, "step": 208800 }, { "epoch": 0.3897373215239597, "grad_norm": 0.9140484929084778, "learning_rate": 0.00019863539292794803, "loss": 4.5942, "step": 208850 }, { "epoch": 0.3898306270833382, "grad_norm": 0.9049879312515259, "learning_rate": 0.00019863470119656394, "loss": 4.7273, "step": 208900 }, { "epoch": 0.3899239326427167, "grad_norm": 0.8477234840393066, "learning_rate": 0.00019863400929110697, "loss": 4.4955, "step": 208950 }, { "epoch": 0.39001723820209516, "grad_norm": 1.2483469247817993, "learning_rate": 0.00019863331721157833, "loss": 4.8071, "step": 209000 }, { "epoch": 0.3901105437614737, "grad_norm": 1.118116021156311, "learning_rate": 0.0001986326249579792, "loss": 4.6028, "step": 209050 }, { "epoch": 0.39020384932085217, "grad_norm": 1.069517731666565, "learning_rate": 0.0001986319325303109, "loss": 4.761, "step": 209100 }, { "epoch": 0.39029715488023065, "grad_norm": 1.1618387699127197, "learning_rate": 0.00019863123992857454, "loss": 4.5113, "step": 209150 }, { "epoch": 0.3903904604396091, "grad_norm": 0.905892014503479, "learning_rate": 0.00019863054715277143, "loss": 4.6528, "step": 209200 }, { "epoch": 0.39048376599898765, "grad_norm": 1.0331082344055176, "learning_rate": 0.00019862985420290277, "loss": 4.8151, "step": 209250 }, { "epoch": 0.39057707155836613, "grad_norm": 0.9063888788223267, "learning_rate": 0.00019862916107896971, "loss": 4.2769, "step": 209300 }, { "epoch": 0.3906703771177446, "grad_norm": 0.8382515907287598, "learning_rate": 0.00019862846778097358, "loss": 4.3706, "step": 209350 }, { "epoch": 0.39076368267712314, "grad_norm": 0.9924123883247375, "learning_rate": 0.00019862777430891556, "loss": 4.7743, "step": 209400 }, { "epoch": 0.3908569882365016, "grad_norm": 1.145404577255249, "learning_rate": 0.00019862708066279688, "loss": 4.5281, "step": 209450 }, { "epoch": 0.3909502937958801, "grad_norm": 1.0358670949935913, "learning_rate": 0.00019862638684261876, "loss": 4.6629, "step": 209500 }, { "epoch": 0.39104359935525856, "grad_norm": 1.2141669988632202, "learning_rate": 0.0001986256928483824, "loss": 4.5794, "step": 209550 }, { "epoch": 0.3911369049146371, "grad_norm": 0.9178851246833801, "learning_rate": 0.00019862499868008904, "loss": 4.6268, "step": 209600 }, { "epoch": 0.39123021047401557, "grad_norm": 1.0358548164367676, "learning_rate": 0.00019862430433773993, "loss": 4.7368, "step": 209650 }, { "epoch": 0.39132351603339405, "grad_norm": 0.8665915131568909, "learning_rate": 0.00019862360982133626, "loss": 4.3792, "step": 209700 }, { "epoch": 0.3914168215927725, "grad_norm": 1.0212570428848267, "learning_rate": 0.0001986229151308793, "loss": 4.8161, "step": 209750 }, { "epoch": 0.39151012715215106, "grad_norm": 1.306823968887329, "learning_rate": 0.00019862222026637022, "loss": 4.4615, "step": 209800 }, { "epoch": 0.39160343271152953, "grad_norm": 1.1909260749816895, "learning_rate": 0.00019862152522781028, "loss": 4.6705, "step": 209850 }, { "epoch": 0.391696738270908, "grad_norm": 0.8337374925613403, "learning_rate": 0.00019862083001520071, "loss": 4.5671, "step": 209900 }, { "epoch": 0.39179004383028654, "grad_norm": 1.068738341331482, "learning_rate": 0.0001986201346285427, "loss": 4.5786, "step": 209950 }, { "epoch": 0.391883349389665, "grad_norm": 0.9537367820739746, "learning_rate": 0.00019861943906783757, "loss": 4.6939, "step": 210000 }, { "epoch": 0.391883349389665, "eval_loss": 4.793107509613037, "eval_runtime": 232.1494, "eval_samples_per_second": 11.234, "eval_steps_per_second": 11.234, "eval_tts_loss": 7.506199054619971, "step": 210000 }, { "epoch": 0.3919766549490435, "grad_norm": 1.085239052772522, "learning_rate": 0.00019861874333308642, "loss": 4.599, "step": 210050 }, { "epoch": 0.39206996050842197, "grad_norm": 0.9580297470092773, "learning_rate": 0.00019861804742429056, "loss": 4.6775, "step": 210100 }, { "epoch": 0.3921632660678005, "grad_norm": 1.1435822248458862, "learning_rate": 0.00019861735134145118, "loss": 4.6902, "step": 210150 }, { "epoch": 0.392256571627179, "grad_norm": 1.1688177585601807, "learning_rate": 0.00019861665508456958, "loss": 4.63, "step": 210200 }, { "epoch": 0.39234987718655745, "grad_norm": 0.7775763273239136, "learning_rate": 0.0001986159586536469, "loss": 4.726, "step": 210250 }, { "epoch": 0.392443182745936, "grad_norm": 1.0630296468734741, "learning_rate": 0.0001986152620486844, "loss": 4.6786, "step": 210300 }, { "epoch": 0.39253648830531446, "grad_norm": 1.1185650825500488, "learning_rate": 0.0001986145652696833, "loss": 4.7136, "step": 210350 }, { "epoch": 0.39262979386469293, "grad_norm": 1.1582688093185425, "learning_rate": 0.00019861386831664487, "loss": 4.5912, "step": 210400 }, { "epoch": 0.3927230994240714, "grad_norm": 1.0869005918502808, "learning_rate": 0.00019861317118957032, "loss": 4.5373, "step": 210450 }, { "epoch": 0.39281640498344994, "grad_norm": 1.0537189245224, "learning_rate": 0.00019861247388846086, "loss": 4.7754, "step": 210500 }, { "epoch": 0.3929097105428284, "grad_norm": 1.4728331565856934, "learning_rate": 0.00019861177641331774, "loss": 4.7525, "step": 210550 }, { "epoch": 0.3930030161022069, "grad_norm": 1.0283353328704834, "learning_rate": 0.0001986110787641422, "loss": 4.4389, "step": 210600 }, { "epoch": 0.3930963216615854, "grad_norm": 1.0805425643920898, "learning_rate": 0.00019861038094093542, "loss": 4.6202, "step": 210650 }, { "epoch": 0.3931896272209639, "grad_norm": 0.8940476179122925, "learning_rate": 0.0001986096829436987, "loss": 4.6031, "step": 210700 }, { "epoch": 0.3932829327803424, "grad_norm": 0.9329911470413208, "learning_rate": 0.00019860898477243324, "loss": 4.4502, "step": 210750 }, { "epoch": 0.39337623833972085, "grad_norm": 0.9388894438743591, "learning_rate": 0.00019860828642714027, "loss": 4.675, "step": 210800 }, { "epoch": 0.3934695438990994, "grad_norm": 0.6746310591697693, "learning_rate": 0.000198607587907821, "loss": 4.8536, "step": 210850 }, { "epoch": 0.39356284945847786, "grad_norm": 0.9739928245544434, "learning_rate": 0.00019860688921447673, "loss": 4.6031, "step": 210900 }, { "epoch": 0.39365615501785634, "grad_norm": 1.030884027481079, "learning_rate": 0.00019860619034710864, "loss": 4.6616, "step": 210950 }, { "epoch": 0.39374946057723487, "grad_norm": 1.0043467283248901, "learning_rate": 0.00019860549130571795, "loss": 4.631, "step": 211000 }, { "epoch": 0.39384276613661334, "grad_norm": 1.2081234455108643, "learning_rate": 0.00019860479209030593, "loss": 4.5703, "step": 211050 }, { "epoch": 0.3939360716959918, "grad_norm": 0.9312835335731506, "learning_rate": 0.0001986040927008738, "loss": 4.6645, "step": 211100 }, { "epoch": 0.3940293772553703, "grad_norm": 0.9531163573265076, "learning_rate": 0.0001986033931374228, "loss": 4.6137, "step": 211150 }, { "epoch": 0.3941226828147488, "grad_norm": 1.0482721328735352, "learning_rate": 0.00019860269339995416, "loss": 4.3711, "step": 211200 }, { "epoch": 0.3942159883741273, "grad_norm": 0.7411187887191772, "learning_rate": 0.0001986019934884691, "loss": 4.4535, "step": 211250 }, { "epoch": 0.3943092939335058, "grad_norm": 1.1720820665359497, "learning_rate": 0.0001986012934029689, "loss": 4.6375, "step": 211300 }, { "epoch": 0.3944025994928843, "grad_norm": 0.9703693389892578, "learning_rate": 0.00019860059314345474, "loss": 4.4733, "step": 211350 }, { "epoch": 0.3944959050522628, "grad_norm": 1.1301404237747192, "learning_rate": 0.0001985998927099279, "loss": 4.6341, "step": 211400 }, { "epoch": 0.39458921061164126, "grad_norm": 0.7381196618080139, "learning_rate": 0.0001985991921023896, "loss": 4.675, "step": 211450 }, { "epoch": 0.39468251617101974, "grad_norm": 1.1832466125488281, "learning_rate": 0.00019859849132084108, "loss": 4.6554, "step": 211500 }, { "epoch": 0.39477582173039827, "grad_norm": 1.0502104759216309, "learning_rate": 0.00019859779036528354, "loss": 4.663, "step": 211550 }, { "epoch": 0.39486912728977674, "grad_norm": 1.127864122390747, "learning_rate": 0.0001985970892357183, "loss": 4.6673, "step": 211600 }, { "epoch": 0.3949624328491552, "grad_norm": 1.035516619682312, "learning_rate": 0.0001985963879321465, "loss": 4.623, "step": 211650 }, { "epoch": 0.39505573840853375, "grad_norm": 0.8197730779647827, "learning_rate": 0.00019859568645456946, "loss": 4.3858, "step": 211700 }, { "epoch": 0.39514904396791223, "grad_norm": 0.8772891163825989, "learning_rate": 0.00019859498480298836, "loss": 4.6063, "step": 211750 }, { "epoch": 0.3952423495272907, "grad_norm": 0.8859339952468872, "learning_rate": 0.00019859428297740444, "loss": 4.5472, "step": 211800 }, { "epoch": 0.3953356550866692, "grad_norm": 1.1464184522628784, "learning_rate": 0.000198593580977819, "loss": 4.6492, "step": 211850 }, { "epoch": 0.3954289606460477, "grad_norm": 1.1008684635162354, "learning_rate": 0.00019859287880423317, "loss": 4.5615, "step": 211900 }, { "epoch": 0.3955222662054262, "grad_norm": 1.2038742303848267, "learning_rate": 0.00019859217645664835, "loss": 4.6712, "step": 211950 }, { "epoch": 0.39561557176480466, "grad_norm": 1.1197429895401, "learning_rate": 0.00019859147393506564, "loss": 4.4475, "step": 212000 }, { "epoch": 0.3957088773241832, "grad_norm": 0.9153450727462769, "learning_rate": 0.00019859077123948632, "loss": 4.4927, "step": 212050 }, { "epoch": 0.39580218288356167, "grad_norm": 0.9062692523002625, "learning_rate": 0.00019859006836991166, "loss": 4.513, "step": 212100 }, { "epoch": 0.39589548844294015, "grad_norm": 1.01179838180542, "learning_rate": 0.00019858936532634282, "loss": 4.4879, "step": 212150 }, { "epoch": 0.3959887940023186, "grad_norm": 1.3619974851608276, "learning_rate": 0.00019858866210878116, "loss": 4.5294, "step": 212200 }, { "epoch": 0.39608209956169715, "grad_norm": 1.009998083114624, "learning_rate": 0.00019858795871722783, "loss": 4.6197, "step": 212250 }, { "epoch": 0.39617540512107563, "grad_norm": 1.0456039905548096, "learning_rate": 0.00019858725515168414, "loss": 4.3916, "step": 212300 }, { "epoch": 0.3962687106804541, "grad_norm": 1.0402274131774902, "learning_rate": 0.00019858655141215124, "loss": 4.6354, "step": 212350 }, { "epoch": 0.39636201623983264, "grad_norm": 1.4560786485671997, "learning_rate": 0.00019858584749863042, "loss": 4.6766, "step": 212400 }, { "epoch": 0.3964553217992111, "grad_norm": 0.6697715520858765, "learning_rate": 0.00019858514341112297, "loss": 4.5072, "step": 212450 }, { "epoch": 0.3965486273585896, "grad_norm": 0.9932947158813477, "learning_rate": 0.00019858443914963004, "loss": 4.7205, "step": 212500 }, { "epoch": 0.39664193291796807, "grad_norm": 1.2164796590805054, "learning_rate": 0.00019858373471415297, "loss": 4.5836, "step": 212550 }, { "epoch": 0.3967352384773466, "grad_norm": 1.0917052030563354, "learning_rate": 0.00019858303010469291, "loss": 4.5275, "step": 212600 }, { "epoch": 0.3968285440367251, "grad_norm": 0.9048320055007935, "learning_rate": 0.00019858232532125118, "loss": 4.709, "step": 212650 }, { "epoch": 0.39692184959610355, "grad_norm": 0.568315327167511, "learning_rate": 0.000198581620363829, "loss": 4.5221, "step": 212700 }, { "epoch": 0.397015155155482, "grad_norm": 1.132904052734375, "learning_rate": 0.00019858091523242756, "loss": 4.6825, "step": 212750 }, { "epoch": 0.39710846071486056, "grad_norm": 0.9325287342071533, "learning_rate": 0.0001985802099270482, "loss": 4.6598, "step": 212800 }, { "epoch": 0.39720176627423903, "grad_norm": 1.0554152727127075, "learning_rate": 0.00019857950444769208, "loss": 4.7556, "step": 212850 }, { "epoch": 0.3972950718336175, "grad_norm": 0.9634024500846863, "learning_rate": 0.0001985787987943605, "loss": 4.4973, "step": 212900 }, { "epoch": 0.39738837739299604, "grad_norm": 1.161877989768982, "learning_rate": 0.0001985780929670547, "loss": 4.6044, "step": 212950 }, { "epoch": 0.3974816829523745, "grad_norm": 0.954832136631012, "learning_rate": 0.00019857738696577588, "loss": 4.4247, "step": 213000 }, { "epoch": 0.3974816829523745, "eval_loss": 4.804925918579102, "eval_runtime": 229.7393, "eval_samples_per_second": 11.352, "eval_steps_per_second": 11.352, "eval_tts_loss": 7.437013648446089, "step": 213000 }, { "epoch": 0.397574988511753, "grad_norm": 0.9206513166427612, "learning_rate": 0.00019857668079052533, "loss": 4.6913, "step": 213050 }, { "epoch": 0.39766829407113147, "grad_norm": 1.1719000339508057, "learning_rate": 0.00019857597444130427, "loss": 4.6099, "step": 213100 }, { "epoch": 0.39776159963051, "grad_norm": 0.8833860158920288, "learning_rate": 0.00019857526791811398, "loss": 4.4701, "step": 213150 }, { "epoch": 0.3978549051898885, "grad_norm": 0.8113496899604797, "learning_rate": 0.00019857456122095567, "loss": 4.5576, "step": 213200 }, { "epoch": 0.39794821074926695, "grad_norm": 0.7514514327049255, "learning_rate": 0.0001985738543498306, "loss": 4.6348, "step": 213250 }, { "epoch": 0.3980415163086455, "grad_norm": 0.8051292896270752, "learning_rate": 0.00019857314730474006, "loss": 4.6339, "step": 213300 }, { "epoch": 0.39813482186802396, "grad_norm": 0.8697991967201233, "learning_rate": 0.00019857244008568523, "loss": 4.5064, "step": 213350 }, { "epoch": 0.39822812742740243, "grad_norm": 1.3794461488723755, "learning_rate": 0.0001985717326926674, "loss": 4.6744, "step": 213400 }, { "epoch": 0.3983214329867809, "grad_norm": 1.0612692832946777, "learning_rate": 0.0001985710251256878, "loss": 4.8218, "step": 213450 }, { "epoch": 0.39841473854615944, "grad_norm": 0.7990536689758301, "learning_rate": 0.00019857031738474768, "loss": 4.8063, "step": 213500 }, { "epoch": 0.3985080441055379, "grad_norm": 0.9997884035110474, "learning_rate": 0.0001985696094698483, "loss": 4.6703, "step": 213550 }, { "epoch": 0.3986013496649164, "grad_norm": 1.0094455480575562, "learning_rate": 0.00019856890138099088, "loss": 4.5359, "step": 213600 }, { "epoch": 0.3986946552242949, "grad_norm": 0.9724659323692322, "learning_rate": 0.00019856819311817674, "loss": 4.7684, "step": 213650 }, { "epoch": 0.3987879607836734, "grad_norm": 0.9077069759368896, "learning_rate": 0.00019856748468140704, "loss": 4.4383, "step": 213700 }, { "epoch": 0.3988812663430519, "grad_norm": 0.8644628524780273, "learning_rate": 0.00019856677607068307, "loss": 4.6682, "step": 213750 }, { "epoch": 0.39897457190243035, "grad_norm": 1.038778305053711, "learning_rate": 0.00019856606728600612, "loss": 4.6176, "step": 213800 }, { "epoch": 0.3990678774618089, "grad_norm": 0.8550258278846741, "learning_rate": 0.0001985653583273774, "loss": 4.3842, "step": 213850 }, { "epoch": 0.39916118302118736, "grad_norm": 0.929908275604248, "learning_rate": 0.00019856464919479813, "loss": 4.7997, "step": 213900 }, { "epoch": 0.39925448858056584, "grad_norm": 0.948686957359314, "learning_rate": 0.00019856393988826963, "loss": 4.608, "step": 213950 }, { "epoch": 0.39934779413994437, "grad_norm": 1.081762433052063, "learning_rate": 0.00019856323040779312, "loss": 4.5192, "step": 214000 }, { "epoch": 0.39944109969932284, "grad_norm": 1.0186339616775513, "learning_rate": 0.00019856252075336984, "loss": 4.5424, "step": 214050 }, { "epoch": 0.3995344052587013, "grad_norm": 1.243036150932312, "learning_rate": 0.00019856181092500104, "loss": 4.6291, "step": 214100 }, { "epoch": 0.3996277108180798, "grad_norm": 0.9121257066726685, "learning_rate": 0.000198561100922688, "loss": 4.4488, "step": 214150 }, { "epoch": 0.3997210163774583, "grad_norm": 1.058773159980774, "learning_rate": 0.00019856039074643195, "loss": 4.5231, "step": 214200 }, { "epoch": 0.3998143219368368, "grad_norm": 1.0914162397384644, "learning_rate": 0.00019855968039623418, "loss": 4.5719, "step": 214250 }, { "epoch": 0.3999076274962153, "grad_norm": 1.0850412845611572, "learning_rate": 0.0001985589698720959, "loss": 4.6784, "step": 214300 }, { "epoch": 0.4000009330555938, "grad_norm": 0.8191171884536743, "learning_rate": 0.00019855825917401837, "loss": 4.6503, "step": 214350 }, { "epoch": 0.4000942386149723, "grad_norm": 1.133804440498352, "learning_rate": 0.00019855754830200284, "loss": 4.6447, "step": 214400 }, { "epoch": 0.40018754417435076, "grad_norm": 0.927496612071991, "learning_rate": 0.0001985568372560506, "loss": 4.6549, "step": 214450 }, { "epoch": 0.40028084973372924, "grad_norm": 0.9108291864395142, "learning_rate": 0.00019855612603616288, "loss": 4.6962, "step": 214500 }, { "epoch": 0.40037415529310777, "grad_norm": 1.0643376111984253, "learning_rate": 0.00019855541464234094, "loss": 4.6508, "step": 214550 }, { "epoch": 0.40046746085248625, "grad_norm": 1.0396971702575684, "learning_rate": 0.00019855470307458606, "loss": 4.6046, "step": 214600 }, { "epoch": 0.4005607664118647, "grad_norm": 0.9104790091514587, "learning_rate": 0.00019855399133289944, "loss": 4.3958, "step": 214650 }, { "epoch": 0.40065407197124325, "grad_norm": 0.8371784687042236, "learning_rate": 0.00019855327941728238, "loss": 4.6615, "step": 214700 }, { "epoch": 0.40074737753062173, "grad_norm": 0.8036884665489197, "learning_rate": 0.0001985525673277361, "loss": 4.4121, "step": 214750 }, { "epoch": 0.4008406830900002, "grad_norm": 1.0669821500778198, "learning_rate": 0.0001985518550642619, "loss": 4.4635, "step": 214800 }, { "epoch": 0.4009339886493787, "grad_norm": 1.0389690399169922, "learning_rate": 0.00019855114262686102, "loss": 4.6575, "step": 214850 }, { "epoch": 0.4010272942087572, "grad_norm": 0.661183774471283, "learning_rate": 0.0001985504300155347, "loss": 4.6654, "step": 214900 }, { "epoch": 0.4011205997681357, "grad_norm": 1.4455314874649048, "learning_rate": 0.00019854971723028424, "loss": 4.7863, "step": 214950 }, { "epoch": 0.40121390532751416, "grad_norm": 0.9810767769813538, "learning_rate": 0.00019854900427111085, "loss": 4.4652, "step": 215000 }, { "epoch": 0.4013072108868927, "grad_norm": 0.9444175362586975, "learning_rate": 0.00019854829113801582, "loss": 4.6636, "step": 215050 }, { "epoch": 0.40140051644627117, "grad_norm": 0.8920708894729614, "learning_rate": 0.00019854757783100037, "loss": 4.5821, "step": 215100 }, { "epoch": 0.40149382200564965, "grad_norm": 0.9815887808799744, "learning_rate": 0.0001985468643500658, "loss": 4.6339, "step": 215150 }, { "epoch": 0.4015871275650281, "grad_norm": 0.9865546226501465, "learning_rate": 0.00019854615069521333, "loss": 4.5329, "step": 215200 }, { "epoch": 0.40168043312440666, "grad_norm": 1.2863136529922485, "learning_rate": 0.00019854543686644428, "loss": 4.7172, "step": 215250 }, { "epoch": 0.40177373868378513, "grad_norm": 1.207463264465332, "learning_rate": 0.00019854472286375986, "loss": 4.4284, "step": 215300 }, { "epoch": 0.4018670442431636, "grad_norm": 0.9688467383384705, "learning_rate": 0.00019854400868716133, "loss": 4.6533, "step": 215350 }, { "epoch": 0.4019603498025421, "grad_norm": 0.767192542552948, "learning_rate": 0.00019854329433665, "loss": 4.4779, "step": 215400 }, { "epoch": 0.4020536553619206, "grad_norm": 1.0222392082214355, "learning_rate": 0.00019854257981222705, "loss": 4.6248, "step": 215450 }, { "epoch": 0.4021469609212991, "grad_norm": 1.1537096500396729, "learning_rate": 0.00019854186511389382, "loss": 4.5435, "step": 215500 }, { "epoch": 0.40224026648067757, "grad_norm": 0.7932273745536804, "learning_rate": 0.00019854115024165153, "loss": 4.7153, "step": 215550 }, { "epoch": 0.4023335720400561, "grad_norm": 0.8604770302772522, "learning_rate": 0.00019854043519550144, "loss": 4.7003, "step": 215600 }, { "epoch": 0.4024268775994346, "grad_norm": 1.160075306892395, "learning_rate": 0.00019853971997544481, "loss": 4.4709, "step": 215650 }, { "epoch": 0.40252018315881305, "grad_norm": 0.9051239490509033, "learning_rate": 0.00019853900458148294, "loss": 4.7949, "step": 215700 }, { "epoch": 0.4026134887181915, "grad_norm": 0.8977974653244019, "learning_rate": 0.00019853828901361705, "loss": 4.5747, "step": 215750 }, { "epoch": 0.40270679427757006, "grad_norm": 1.0881613492965698, "learning_rate": 0.0001985375732718484, "loss": 4.6564, "step": 215800 }, { "epoch": 0.40280009983694853, "grad_norm": 1.3626446723937988, "learning_rate": 0.00019853685735617832, "loss": 4.613, "step": 215850 }, { "epoch": 0.402893405396327, "grad_norm": 1.0976074934005737, "learning_rate": 0.00019853614126660797, "loss": 4.6058, "step": 215900 }, { "epoch": 0.40298671095570554, "grad_norm": 1.0088049173355103, "learning_rate": 0.0001985354250031387, "loss": 4.5451, "step": 215950 }, { "epoch": 0.403080016515084, "grad_norm": 0.7824239730834961, "learning_rate": 0.00019853470856577175, "loss": 4.5725, "step": 216000 }, { "epoch": 0.403080016515084, "eval_loss": 4.79715633392334, "eval_runtime": 235.0521, "eval_samples_per_second": 11.095, "eval_steps_per_second": 11.095, "eval_tts_loss": 7.512439045195176, "step": 216000 }, { "epoch": 0.4031733220744625, "grad_norm": 0.9689962863922119, "learning_rate": 0.00019853399195450834, "loss": 4.4423, "step": 216050 }, { "epoch": 0.40326662763384097, "grad_norm": 1.1092582941055298, "learning_rate": 0.00019853327516934982, "loss": 4.5845, "step": 216100 }, { "epoch": 0.4033599331932195, "grad_norm": 1.217396855354309, "learning_rate": 0.0001985325582102974, "loss": 4.634, "step": 216150 }, { "epoch": 0.403453238752598, "grad_norm": 1.0966297388076782, "learning_rate": 0.00019853184107735232, "loss": 4.6351, "step": 216200 }, { "epoch": 0.40354654431197645, "grad_norm": 0.943762481212616, "learning_rate": 0.0001985311237705159, "loss": 4.7421, "step": 216250 }, { "epoch": 0.403639849871355, "grad_norm": 1.03568434715271, "learning_rate": 0.00019853040628978937, "loss": 4.6778, "step": 216300 }, { "epoch": 0.40373315543073346, "grad_norm": 1.1071677207946777, "learning_rate": 0.000198529688635174, "loss": 4.6034, "step": 216350 }, { "epoch": 0.40382646099011194, "grad_norm": 0.9357593059539795, "learning_rate": 0.0001985289708066711, "loss": 4.7582, "step": 216400 }, { "epoch": 0.4039197665494904, "grad_norm": 1.0563290119171143, "learning_rate": 0.00019852825280428186, "loss": 4.552, "step": 216450 }, { "epoch": 0.40401307210886894, "grad_norm": 1.0622097253799438, "learning_rate": 0.00019852753462800763, "loss": 4.5838, "step": 216500 }, { "epoch": 0.4041063776682474, "grad_norm": 1.0469093322753906, "learning_rate": 0.00019852681627784962, "loss": 4.5972, "step": 216550 }, { "epoch": 0.4041996832276259, "grad_norm": 0.950249969959259, "learning_rate": 0.00019852609775380912, "loss": 4.5714, "step": 216600 }, { "epoch": 0.4042929887870044, "grad_norm": 0.9721581935882568, "learning_rate": 0.00019852537905588741, "loss": 4.5639, "step": 216650 }, { "epoch": 0.4043862943463829, "grad_norm": 1.0957340002059937, "learning_rate": 0.00019852466018408573, "loss": 4.4685, "step": 216700 }, { "epoch": 0.4044795999057614, "grad_norm": 1.099837064743042, "learning_rate": 0.00019852394113840536, "loss": 4.6103, "step": 216750 }, { "epoch": 0.40457290546513985, "grad_norm": 1.1709736585617065, "learning_rate": 0.00019852322191884757, "loss": 4.6009, "step": 216800 }, { "epoch": 0.4046662110245184, "grad_norm": 0.804123044013977, "learning_rate": 0.0001985225025254136, "loss": 4.614, "step": 216850 }, { "epoch": 0.40475951658389686, "grad_norm": 0.7690798044204712, "learning_rate": 0.0001985217829581048, "loss": 4.5003, "step": 216900 }, { "epoch": 0.40485282214327534, "grad_norm": 1.1280100345611572, "learning_rate": 0.00019852106321692234, "loss": 4.3368, "step": 216950 }, { "epoch": 0.40494612770265387, "grad_norm": 0.9494564533233643, "learning_rate": 0.00019852034330186755, "loss": 4.5604, "step": 217000 }, { "epoch": 0.40503943326203234, "grad_norm": 0.8637983798980713, "learning_rate": 0.00019851962321294173, "loss": 4.6788, "step": 217050 }, { "epoch": 0.4051327388214108, "grad_norm": 1.0316267013549805, "learning_rate": 0.00019851890295014606, "loss": 4.6512, "step": 217100 }, { "epoch": 0.4052260443807893, "grad_norm": 1.0898293256759644, "learning_rate": 0.00019851818251348188, "loss": 4.5092, "step": 217150 }, { "epoch": 0.40531934994016783, "grad_norm": 0.9936012029647827, "learning_rate": 0.00019851746190295043, "loss": 4.5646, "step": 217200 }, { "epoch": 0.4054126554995463, "grad_norm": 1.9127103090286255, "learning_rate": 0.00019851674111855304, "loss": 4.6748, "step": 217250 }, { "epoch": 0.4055059610589248, "grad_norm": 1.153311014175415, "learning_rate": 0.0001985160201602909, "loss": 4.6777, "step": 217300 }, { "epoch": 0.4055992666183033, "grad_norm": 0.802513599395752, "learning_rate": 0.0001985152990281653, "loss": 4.7041, "step": 217350 }, { "epoch": 0.4056925721776818, "grad_norm": 0.7697170972824097, "learning_rate": 0.00019851457772217756, "loss": 4.6375, "step": 217400 }, { "epoch": 0.40578587773706026, "grad_norm": 1.0160504579544067, "learning_rate": 0.00019851385624232895, "loss": 4.5718, "step": 217450 }, { "epoch": 0.40587918329643874, "grad_norm": 1.2011131048202515, "learning_rate": 0.0001985131345886207, "loss": 4.5841, "step": 217500 }, { "epoch": 0.40597248885581727, "grad_norm": 0.7780730724334717, "learning_rate": 0.00019851241276105407, "loss": 4.6344, "step": 217550 }, { "epoch": 0.40606579441519575, "grad_norm": 1.0549157857894897, "learning_rate": 0.00019851169075963036, "loss": 4.7876, "step": 217600 }, { "epoch": 0.4061590999745742, "grad_norm": 0.9472026228904724, "learning_rate": 0.00019851096858435088, "loss": 4.6313, "step": 217650 }, { "epoch": 0.40625240553395275, "grad_norm": 0.9771903157234192, "learning_rate": 0.0001985102462352169, "loss": 4.5077, "step": 217700 }, { "epoch": 0.40634571109333123, "grad_norm": 1.2292876243591309, "learning_rate": 0.00019850952371222962, "loss": 4.5595, "step": 217750 }, { "epoch": 0.4064390166527097, "grad_norm": 1.2494375705718994, "learning_rate": 0.00019850880101539037, "loss": 4.713, "step": 217800 }, { "epoch": 0.4065323222120882, "grad_norm": 1.2384469509124756, "learning_rate": 0.00019850807814470045, "loss": 4.5567, "step": 217850 }, { "epoch": 0.4066256277714667, "grad_norm": 1.1389862298965454, "learning_rate": 0.00019850735510016108, "loss": 4.5898, "step": 217900 }, { "epoch": 0.4067189333308452, "grad_norm": 1.0417871475219727, "learning_rate": 0.00019850663188177356, "loss": 4.6815, "step": 217950 }, { "epoch": 0.40681223889022367, "grad_norm": 1.019178867340088, "learning_rate": 0.00019850590848953918, "loss": 4.6889, "step": 218000 }, { "epoch": 0.40690554444960214, "grad_norm": 0.7551344633102417, "learning_rate": 0.0001985051849234592, "loss": 4.6242, "step": 218050 }, { "epoch": 0.4069988500089807, "grad_norm": 1.3654025793075562, "learning_rate": 0.0001985044611835349, "loss": 4.7712, "step": 218100 }, { "epoch": 0.40709215556835915, "grad_norm": 0.6821979284286499, "learning_rate": 0.00019850373726976756, "loss": 4.6176, "step": 218150 }, { "epoch": 0.4071854611277376, "grad_norm": 1.0702706575393677, "learning_rate": 0.00019850301318215846, "loss": 4.4522, "step": 218200 }, { "epoch": 0.40727876668711616, "grad_norm": 1.0518271923065186, "learning_rate": 0.00019850228892070884, "loss": 4.7112, "step": 218250 }, { "epoch": 0.40737207224649463, "grad_norm": 1.0100867748260498, "learning_rate": 0.00019850156448542007, "loss": 4.7564, "step": 218300 }, { "epoch": 0.4074653778058731, "grad_norm": 1.5200986862182617, "learning_rate": 0.00019850083987629332, "loss": 4.5178, "step": 218350 }, { "epoch": 0.4075586833652516, "grad_norm": 1.1372010707855225, "learning_rate": 0.00019850011509332994, "loss": 4.5793, "step": 218400 }, { "epoch": 0.4076519889246301, "grad_norm": 1.1734660863876343, "learning_rate": 0.0001984993901365312, "loss": 4.7042, "step": 218450 }, { "epoch": 0.4077452944840086, "grad_norm": 0.9464728236198425, "learning_rate": 0.00019849866500589833, "loss": 4.5238, "step": 218500 }, { "epoch": 0.40783860004338707, "grad_norm": 1.0089821815490723, "learning_rate": 0.0001984979397014327, "loss": 4.4707, "step": 218550 }, { "epoch": 0.4079319056027656, "grad_norm": 0.7066876292228699, "learning_rate": 0.0001984972142231355, "loss": 4.556, "step": 218600 }, { "epoch": 0.4080252111621441, "grad_norm": 1.1190834045410156, "learning_rate": 0.00019849648857100808, "loss": 4.4077, "step": 218650 }, { "epoch": 0.40811851672152255, "grad_norm": 0.7868063449859619, "learning_rate": 0.00019849576274505164, "loss": 4.6604, "step": 218700 }, { "epoch": 0.408211822280901, "grad_norm": 1.2350362539291382, "learning_rate": 0.00019849503674526756, "loss": 4.5138, "step": 218750 }, { "epoch": 0.40830512784027956, "grad_norm": 1.3599216938018799, "learning_rate": 0.00019849431057165703, "loss": 4.6163, "step": 218800 }, { "epoch": 0.40839843339965803, "grad_norm": 0.9188289642333984, "learning_rate": 0.0001984935842242214, "loss": 4.6306, "step": 218850 }, { "epoch": 0.4084917389590365, "grad_norm": 0.9845192432403564, "learning_rate": 0.00019849285770296189, "loss": 4.656, "step": 218900 }, { "epoch": 0.40858504451841504, "grad_norm": 1.1785905361175537, "learning_rate": 0.00019849213100787987, "loss": 4.6391, "step": 218950 }, { "epoch": 0.4086783500777935, "grad_norm": 1.0496912002563477, "learning_rate": 0.00019849140413897654, "loss": 4.4645, "step": 219000 }, { "epoch": 0.4086783500777935, "eval_loss": 4.787511825561523, "eval_runtime": 234.8152, "eval_samples_per_second": 11.107, "eval_steps_per_second": 11.107, "eval_tts_loss": 7.520304385868353, "step": 219000 }, { "epoch": 0.408771655637172, "grad_norm": 1.1836047172546387, "learning_rate": 0.0001984906770962532, "loss": 4.5759, "step": 219050 }, { "epoch": 0.40886496119655047, "grad_norm": 0.866492748260498, "learning_rate": 0.00019848994987971117, "loss": 4.7295, "step": 219100 }, { "epoch": 0.408958266755929, "grad_norm": 0.7800747752189636, "learning_rate": 0.00019848922248935167, "loss": 4.6962, "step": 219150 }, { "epoch": 0.4090515723153075, "grad_norm": 1.0174862146377563, "learning_rate": 0.00019848849492517604, "loss": 4.4633, "step": 219200 }, { "epoch": 0.40914487787468595, "grad_norm": 1.1767679452896118, "learning_rate": 0.00019848776718718556, "loss": 4.6195, "step": 219250 }, { "epoch": 0.4092381834340645, "grad_norm": 1.1328825950622559, "learning_rate": 0.00019848703927538152, "loss": 4.518, "step": 219300 }, { "epoch": 0.40933148899344296, "grad_norm": 0.7945483326911926, "learning_rate": 0.00019848631118976514, "loss": 4.7111, "step": 219350 }, { "epoch": 0.40942479455282144, "grad_norm": 1.0717135667800903, "learning_rate": 0.00019848558293033777, "loss": 4.6445, "step": 219400 }, { "epoch": 0.4095181001121999, "grad_norm": 0.9013932347297668, "learning_rate": 0.00019848485449710067, "loss": 4.511, "step": 219450 }, { "epoch": 0.40961140567157844, "grad_norm": 1.1466339826583862, "learning_rate": 0.00019848412589005513, "loss": 4.4065, "step": 219500 }, { "epoch": 0.4097047112309569, "grad_norm": 1.0499449968338013, "learning_rate": 0.00019848339710920246, "loss": 4.489, "step": 219550 }, { "epoch": 0.4097980167903354, "grad_norm": 0.9530263543128967, "learning_rate": 0.0001984826681545439, "loss": 4.8684, "step": 219600 }, { "epoch": 0.4098913223497139, "grad_norm": 0.9534028768539429, "learning_rate": 0.00019848193902608077, "loss": 4.5792, "step": 219650 }, { "epoch": 0.4099846279090924, "grad_norm": 1.157947063446045, "learning_rate": 0.00019848120972381432, "loss": 4.5243, "step": 219700 }, { "epoch": 0.4100779334684709, "grad_norm": 1.022988200187683, "learning_rate": 0.00019848048024774588, "loss": 4.7144, "step": 219750 }, { "epoch": 0.41017123902784935, "grad_norm": 1.097163438796997, "learning_rate": 0.00019847975059787674, "loss": 4.8489, "step": 219800 }, { "epoch": 0.4102645445872279, "grad_norm": 1.2663757801055908, "learning_rate": 0.00019847902077420814, "loss": 4.629, "step": 219850 }, { "epoch": 0.41035785014660636, "grad_norm": 1.103257656097412, "learning_rate": 0.00019847829077674142, "loss": 4.6784, "step": 219900 }, { "epoch": 0.41045115570598484, "grad_norm": 1.0042039155960083, "learning_rate": 0.00019847756060547783, "loss": 4.6846, "step": 219950 }, { "epoch": 0.41054446126536337, "grad_norm": 0.9465295076370239, "learning_rate": 0.00019847683026041869, "loss": 4.3015, "step": 220000 }, { "epoch": 0.41063776682474185, "grad_norm": 0.9060609936714172, "learning_rate": 0.00019847609974156524, "loss": 4.3934, "step": 220050 }, { "epoch": 0.4107310723841203, "grad_norm": 0.7724514603614807, "learning_rate": 0.00019847536904891882, "loss": 4.617, "step": 220100 }, { "epoch": 0.4108243779434988, "grad_norm": 1.2579684257507324, "learning_rate": 0.00019847463818248068, "loss": 4.756, "step": 220150 }, { "epoch": 0.41091768350287733, "grad_norm": 0.7629663944244385, "learning_rate": 0.00019847390714225217, "loss": 4.5186, "step": 220200 }, { "epoch": 0.4110109890622558, "grad_norm": 0.8361650109291077, "learning_rate": 0.0001984731759282345, "loss": 4.6316, "step": 220250 }, { "epoch": 0.4111042946216343, "grad_norm": 1.2181036472320557, "learning_rate": 0.00019847244454042904, "loss": 4.7891, "step": 220300 }, { "epoch": 0.4111976001810128, "grad_norm": 0.9293192028999329, "learning_rate": 0.00019847171297883703, "loss": 4.4564, "step": 220350 }, { "epoch": 0.4112909057403913, "grad_norm": 0.9219282269477844, "learning_rate": 0.00019847098124345977, "loss": 4.5263, "step": 220400 }, { "epoch": 0.41138421129976976, "grad_norm": 1.1117901802062988, "learning_rate": 0.00019847024933429856, "loss": 4.6565, "step": 220450 }, { "epoch": 0.41147751685914824, "grad_norm": 1.5147013664245605, "learning_rate": 0.00019846951725135468, "loss": 4.8577, "step": 220500 }, { "epoch": 0.41157082241852677, "grad_norm": 1.095376968383789, "learning_rate": 0.00019846878499462942, "loss": 4.4866, "step": 220550 }, { "epoch": 0.41166412797790525, "grad_norm": 0.8250828385353088, "learning_rate": 0.00019846805256412407, "loss": 4.6568, "step": 220600 }, { "epoch": 0.4117574335372837, "grad_norm": 1.0293447971343994, "learning_rate": 0.00019846731995983995, "loss": 4.8399, "step": 220650 }, { "epoch": 0.4118507390966622, "grad_norm": 1.0483900308609009, "learning_rate": 0.00019846658718177836, "loss": 4.7965, "step": 220700 }, { "epoch": 0.41194404465604073, "grad_norm": 0.627288281917572, "learning_rate": 0.00019846585422994055, "loss": 4.1862, "step": 220750 }, { "epoch": 0.4120373502154192, "grad_norm": 1.12353515625, "learning_rate": 0.0001984651211043278, "loss": 4.7008, "step": 220800 }, { "epoch": 0.4121306557747977, "grad_norm": 1.2579227685928345, "learning_rate": 0.00019846438780494148, "loss": 4.5624, "step": 220850 }, { "epoch": 0.4122239613341762, "grad_norm": 1.1707371473312378, "learning_rate": 0.00019846365433178284, "loss": 4.6096, "step": 220900 }, { "epoch": 0.4123172668935547, "grad_norm": 0.7681998610496521, "learning_rate": 0.00019846292068485317, "loss": 4.6948, "step": 220950 }, { "epoch": 0.41241057245293317, "grad_norm": 0.5637853741645813, "learning_rate": 0.00019846218686415376, "loss": 4.5245, "step": 221000 }, { "epoch": 0.41250387801231164, "grad_norm": 0.9706264734268188, "learning_rate": 0.0001984614528696859, "loss": 4.6419, "step": 221050 }, { "epoch": 0.4125971835716902, "grad_norm": 1.179498553276062, "learning_rate": 0.00019846071870145094, "loss": 4.4572, "step": 221100 }, { "epoch": 0.41269048913106865, "grad_norm": 1.05915105342865, "learning_rate": 0.0001984599843594501, "loss": 4.7192, "step": 221150 }, { "epoch": 0.4127837946904471, "grad_norm": 1.2832460403442383, "learning_rate": 0.00019845924984368473, "loss": 4.632, "step": 221200 }, { "epoch": 0.41287710024982566, "grad_norm": 0.977168083190918, "learning_rate": 0.0001984585151541561, "loss": 4.6565, "step": 221250 }, { "epoch": 0.41297040580920413, "grad_norm": 0.9659879207611084, "learning_rate": 0.00019845778029086553, "loss": 4.5208, "step": 221300 }, { "epoch": 0.4130637113685826, "grad_norm": 1.013081669807434, "learning_rate": 0.0001984570452538143, "loss": 4.4713, "step": 221350 }, { "epoch": 0.4131570169279611, "grad_norm": 0.9163434505462646, "learning_rate": 0.00019845631004300368, "loss": 4.6295, "step": 221400 }, { "epoch": 0.4132503224873396, "grad_norm": 0.9959138631820679, "learning_rate": 0.00019845557465843504, "loss": 4.497, "step": 221450 }, { "epoch": 0.4133436280467181, "grad_norm": 1.0366419553756714, "learning_rate": 0.0001984548391001096, "loss": 4.6001, "step": 221500 }, { "epoch": 0.41343693360609657, "grad_norm": 0.8534669280052185, "learning_rate": 0.00019845410336802872, "loss": 4.6079, "step": 221550 }, { "epoch": 0.4135302391654751, "grad_norm": 1.0028185844421387, "learning_rate": 0.00019845336746219364, "loss": 4.5135, "step": 221600 }, { "epoch": 0.4136235447248536, "grad_norm": 0.957848846912384, "learning_rate": 0.0001984526313826057, "loss": 4.5515, "step": 221650 }, { "epoch": 0.41371685028423205, "grad_norm": 0.7522804141044617, "learning_rate": 0.00019845189512926622, "loss": 4.679, "step": 221700 }, { "epoch": 0.4138101558436105, "grad_norm": 0.7022336721420288, "learning_rate": 0.00019845115870217644, "loss": 4.642, "step": 221750 }, { "epoch": 0.41390346140298906, "grad_norm": 1.0571622848510742, "learning_rate": 0.0001984504221013377, "loss": 4.6447, "step": 221800 }, { "epoch": 0.41399676696236754, "grad_norm": 0.7391085624694824, "learning_rate": 0.00019844968532675125, "loss": 4.3801, "step": 221850 }, { "epoch": 0.414090072521746, "grad_norm": 0.9244349598884583, "learning_rate": 0.00019844894837841846, "loss": 4.5394, "step": 221900 }, { "epoch": 0.41418337808112454, "grad_norm": 0.9880529642105103, "learning_rate": 0.00019844821125634062, "loss": 4.4614, "step": 221950 }, { "epoch": 0.414276683640503, "grad_norm": 1.040346622467041, "learning_rate": 0.00019844747396051898, "loss": 4.574, "step": 222000 }, { "epoch": 0.414276683640503, "eval_loss": 4.79351282119751, "eval_runtime": 232.2857, "eval_samples_per_second": 11.228, "eval_steps_per_second": 11.228, "eval_tts_loss": 7.543720753433503, "step": 222000 }, { "epoch": 0.4143699891998815, "grad_norm": 0.9894281029701233, "learning_rate": 0.00019844673649095486, "loss": 4.5986, "step": 222050 }, { "epoch": 0.41446329475925997, "grad_norm": 0.9158909320831299, "learning_rate": 0.0001984459988476496, "loss": 4.7346, "step": 222100 }, { "epoch": 0.4145566003186385, "grad_norm": 0.9308950901031494, "learning_rate": 0.00019844526103060447, "loss": 4.5288, "step": 222150 }, { "epoch": 0.414649905878017, "grad_norm": 0.967197597026825, "learning_rate": 0.00019844452303982075, "loss": 4.6904, "step": 222200 }, { "epoch": 0.41474321143739545, "grad_norm": 0.8916376233100891, "learning_rate": 0.0001984437848752998, "loss": 4.4642, "step": 222250 }, { "epoch": 0.414836516996774, "grad_norm": 1.0411932468414307, "learning_rate": 0.00019844304653704285, "loss": 4.6579, "step": 222300 }, { "epoch": 0.41492982255615246, "grad_norm": 0.9220390915870667, "learning_rate": 0.0001984423080250513, "loss": 4.6703, "step": 222350 }, { "epoch": 0.41502312811553094, "grad_norm": 1.0786405801773071, "learning_rate": 0.00019844156933932635, "loss": 4.5595, "step": 222400 }, { "epoch": 0.4151164336749094, "grad_norm": 1.1047947406768799, "learning_rate": 0.0001984408304798694, "loss": 4.6576, "step": 222450 }, { "epoch": 0.41520973923428794, "grad_norm": 0.8863925933837891, "learning_rate": 0.00019844009144668167, "loss": 4.6185, "step": 222500 }, { "epoch": 0.4153030447936664, "grad_norm": 0.9184449911117554, "learning_rate": 0.0001984393522397645, "loss": 4.7059, "step": 222550 }, { "epoch": 0.4153963503530449, "grad_norm": 0.9710171818733215, "learning_rate": 0.00019843861285911922, "loss": 4.5444, "step": 222600 }, { "epoch": 0.41548965591242343, "grad_norm": 1.0862765312194824, "learning_rate": 0.00019843787330474708, "loss": 4.6252, "step": 222650 }, { "epoch": 0.4155829614718019, "grad_norm": 0.7898451685905457, "learning_rate": 0.00019843713357664943, "loss": 4.8069, "step": 222700 }, { "epoch": 0.4156762670311804, "grad_norm": 0.9102611541748047, "learning_rate": 0.00019843639367482757, "loss": 4.5326, "step": 222750 }, { "epoch": 0.41576957259055886, "grad_norm": 0.8616012334823608, "learning_rate": 0.00019843565359928275, "loss": 4.4656, "step": 222800 }, { "epoch": 0.4158628781499374, "grad_norm": 1.0981239080429077, "learning_rate": 0.00019843491335001638, "loss": 4.4092, "step": 222850 }, { "epoch": 0.41595618370931586, "grad_norm": 1.237775444984436, "learning_rate": 0.00019843417292702967, "loss": 4.5823, "step": 222900 }, { "epoch": 0.41604948926869434, "grad_norm": 1.199179768562317, "learning_rate": 0.000198433432330324, "loss": 4.7258, "step": 222950 }, { "epoch": 0.41614279482807287, "grad_norm": 1.1261119842529297, "learning_rate": 0.00019843269155990059, "loss": 4.5042, "step": 223000 }, { "epoch": 0.41623610038745135, "grad_norm": 1.0203489065170288, "learning_rate": 0.00019843195061576083, "loss": 4.49, "step": 223050 }, { "epoch": 0.4163294059468298, "grad_norm": 0.8318331837654114, "learning_rate": 0.000198431209497906, "loss": 4.5679, "step": 223100 }, { "epoch": 0.4164227115062083, "grad_norm": 1.0741044282913208, "learning_rate": 0.0001984304682063374, "loss": 4.6131, "step": 223150 }, { "epoch": 0.41651601706558683, "grad_norm": 1.0485063791275024, "learning_rate": 0.00019842972674105632, "loss": 4.4938, "step": 223200 }, { "epoch": 0.4166093226249653, "grad_norm": 0.9604942202568054, "learning_rate": 0.0001984289851020641, "loss": 4.536, "step": 223250 }, { "epoch": 0.4167026281843438, "grad_norm": 1.0192519426345825, "learning_rate": 0.0001984282432893621, "loss": 4.5778, "step": 223300 }, { "epoch": 0.41679593374372226, "grad_norm": 1.1776680946350098, "learning_rate": 0.0001984275013029515, "loss": 4.359, "step": 223350 }, { "epoch": 0.4168892393031008, "grad_norm": 0.999553382396698, "learning_rate": 0.00019842675914283372, "loss": 4.5106, "step": 223400 }, { "epoch": 0.41698254486247927, "grad_norm": 0.972808301448822, "learning_rate": 0.00019842601680901, "loss": 4.5058, "step": 223450 }, { "epoch": 0.41707585042185774, "grad_norm": 0.6613345146179199, "learning_rate": 0.0001984252743014817, "loss": 4.5789, "step": 223500 }, { "epoch": 0.4171691559812363, "grad_norm": 0.9558602571487427, "learning_rate": 0.00019842453162025007, "loss": 4.7024, "step": 223550 }, { "epoch": 0.41726246154061475, "grad_norm": 1.245826244354248, "learning_rate": 0.0001984237887653165, "loss": 4.5732, "step": 223600 }, { "epoch": 0.4173557670999932, "grad_norm": 1.244626760482788, "learning_rate": 0.00019842304573668224, "loss": 4.5236, "step": 223650 }, { "epoch": 0.4174490726593717, "grad_norm": 1.0705629587173462, "learning_rate": 0.00019842230253434864, "loss": 4.6263, "step": 223700 }, { "epoch": 0.41754237821875023, "grad_norm": 1.0318876504898071, "learning_rate": 0.00019842155915831698, "loss": 4.795, "step": 223750 }, { "epoch": 0.4176356837781287, "grad_norm": 1.3221536874771118, "learning_rate": 0.00019842081560858858, "loss": 4.7132, "step": 223800 }, { "epoch": 0.4177289893375072, "grad_norm": 0.8112679719924927, "learning_rate": 0.0001984200718851648, "loss": 4.4579, "step": 223850 }, { "epoch": 0.4178222948968857, "grad_norm": 1.2415705919265747, "learning_rate": 0.00019841932798804685, "loss": 4.7319, "step": 223900 }, { "epoch": 0.4179156004562642, "grad_norm": 1.0489301681518555, "learning_rate": 0.00019841858391723616, "loss": 4.7578, "step": 223950 }, { "epoch": 0.41800890601564267, "grad_norm": 1.1561607122421265, "learning_rate": 0.00019841783967273395, "loss": 4.5376, "step": 224000 }, { "epoch": 0.41810221157502114, "grad_norm": 1.2369205951690674, "learning_rate": 0.00019841709525454158, "loss": 4.506, "step": 224050 }, { "epoch": 0.4181955171343997, "grad_norm": 1.139877438545227, "learning_rate": 0.00019841635066266035, "loss": 4.7282, "step": 224100 }, { "epoch": 0.41828882269377815, "grad_norm": 0.8672619462013245, "learning_rate": 0.00019841560589709158, "loss": 4.7663, "step": 224150 }, { "epoch": 0.4183821282531566, "grad_norm": 1.0294967889785767, "learning_rate": 0.0001984148609578366, "loss": 4.7002, "step": 224200 }, { "epoch": 0.41847543381253516, "grad_norm": 1.0474852323532104, "learning_rate": 0.00019841411584489672, "loss": 4.4804, "step": 224250 }, { "epoch": 0.41856873937191363, "grad_norm": 1.1958084106445312, "learning_rate": 0.00019841337055827318, "loss": 4.4729, "step": 224300 }, { "epoch": 0.4186620449312921, "grad_norm": 0.6683170795440674, "learning_rate": 0.0001984126250979674, "loss": 4.5723, "step": 224350 }, { "epoch": 0.4187553504906706, "grad_norm": 1.3585337400436401, "learning_rate": 0.00019841187946398065, "loss": 4.6204, "step": 224400 }, { "epoch": 0.4188486560500491, "grad_norm": 0.8530429005622864, "learning_rate": 0.00019841113365631424, "loss": 4.7279, "step": 224450 }, { "epoch": 0.4189419616094276, "grad_norm": 0.943585216999054, "learning_rate": 0.00019841038767496948, "loss": 4.5772, "step": 224500 }, { "epoch": 0.41903526716880607, "grad_norm": 1.0253528356552124, "learning_rate": 0.00019840964151994774, "loss": 4.6783, "step": 224550 }, { "epoch": 0.4191285727281846, "grad_norm": 1.0521509647369385, "learning_rate": 0.00019840889519125027, "loss": 4.6578, "step": 224600 }, { "epoch": 0.4192218782875631, "grad_norm": 0.9954854249954224, "learning_rate": 0.00019840814868887842, "loss": 4.7111, "step": 224650 }, { "epoch": 0.41931518384694155, "grad_norm": 1.1058915853500366, "learning_rate": 0.00019840740201283353, "loss": 4.4783, "step": 224700 }, { "epoch": 0.41940848940632003, "grad_norm": 1.0141825675964355, "learning_rate": 0.00019840665516311688, "loss": 4.603, "step": 224750 }, { "epoch": 0.41950179496569856, "grad_norm": 0.7934043407440186, "learning_rate": 0.00019840590813972978, "loss": 4.6601, "step": 224800 }, { "epoch": 0.41959510052507704, "grad_norm": 1.1680604219436646, "learning_rate": 0.00019840516094267357, "loss": 4.5609, "step": 224850 }, { "epoch": 0.4196884060844555, "grad_norm": 0.995583713054657, "learning_rate": 0.00019840441357194956, "loss": 4.5134, "step": 224900 }, { "epoch": 0.41978171164383404, "grad_norm": 0.9852712750434875, "learning_rate": 0.00019840366602755908, "loss": 4.5728, "step": 224950 }, { "epoch": 0.4198750172032125, "grad_norm": 1.1983600854873657, "learning_rate": 0.00019840291830950345, "loss": 4.5515, "step": 225000 }, { "epoch": 0.4198750172032125, "eval_loss": 4.793398380279541, "eval_runtime": 231.8067, "eval_samples_per_second": 11.251, "eval_steps_per_second": 11.251, "eval_tts_loss": 7.515776119422042, "step": 225000 }, { "epoch": 0.419968322762591, "grad_norm": 1.0381780862808228, "learning_rate": 0.000198402170417784, "loss": 4.4676, "step": 225050 }, { "epoch": 0.42006162832196947, "grad_norm": 1.10462486743927, "learning_rate": 0.000198401422352402, "loss": 4.3191, "step": 225100 }, { "epoch": 0.420154933881348, "grad_norm": 1.2322332859039307, "learning_rate": 0.00019840067411335883, "loss": 4.7742, "step": 225150 }, { "epoch": 0.4202482394407265, "grad_norm": 0.9938576817512512, "learning_rate": 0.0001983999257006558, "loss": 4.5088, "step": 225200 }, { "epoch": 0.42034154500010495, "grad_norm": 0.9337325692176819, "learning_rate": 0.00019839917711429416, "loss": 4.6499, "step": 225250 }, { "epoch": 0.4204348505594835, "grad_norm": 0.9972164630889893, "learning_rate": 0.0001983984283542753, "loss": 4.5516, "step": 225300 }, { "epoch": 0.42052815611886196, "grad_norm": 1.3963088989257812, "learning_rate": 0.00019839767942060057, "loss": 4.5577, "step": 225350 }, { "epoch": 0.42062146167824044, "grad_norm": 1.1737512350082397, "learning_rate": 0.0001983969303132712, "loss": 4.4765, "step": 225400 }, { "epoch": 0.4207147672376189, "grad_norm": 0.8017054796218872, "learning_rate": 0.0001983961810322886, "loss": 4.524, "step": 225450 }, { "epoch": 0.42080807279699745, "grad_norm": 1.0323697328567505, "learning_rate": 0.00019839543157765404, "loss": 4.3256, "step": 225500 }, { "epoch": 0.4209013783563759, "grad_norm": 1.0249814987182617, "learning_rate": 0.00019839468194936887, "loss": 4.7338, "step": 225550 }, { "epoch": 0.4209946839157544, "grad_norm": 1.0047541856765747, "learning_rate": 0.00019839393214743435, "loss": 4.4109, "step": 225600 }, { "epoch": 0.42108798947513293, "grad_norm": 1.2357292175292969, "learning_rate": 0.0001983931821718519, "loss": 4.683, "step": 225650 }, { "epoch": 0.4211812950345114, "grad_norm": 1.0345031023025513, "learning_rate": 0.0001983924320226228, "loss": 4.5459, "step": 225700 }, { "epoch": 0.4212746005938899, "grad_norm": 0.9663260579109192, "learning_rate": 0.00019839168169974835, "loss": 4.7504, "step": 225750 }, { "epoch": 0.42136790615326836, "grad_norm": 1.1099520921707153, "learning_rate": 0.0001983909312032299, "loss": 4.435, "step": 225800 }, { "epoch": 0.4214612117126469, "grad_norm": 0.9571065902709961, "learning_rate": 0.00019839018053306875, "loss": 4.7323, "step": 225850 }, { "epoch": 0.42155451727202536, "grad_norm": 1.3123244047164917, "learning_rate": 0.00019838942968926625, "loss": 4.6093, "step": 225900 }, { "epoch": 0.42164782283140384, "grad_norm": 0.9905040860176086, "learning_rate": 0.00019838867867182373, "loss": 4.6339, "step": 225950 }, { "epoch": 0.4217411283907823, "grad_norm": 0.8658848404884338, "learning_rate": 0.0001983879274807425, "loss": 4.7805, "step": 226000 }, { "epoch": 0.42183443395016085, "grad_norm": 1.1252857446670532, "learning_rate": 0.0001983871761160239, "loss": 4.3846, "step": 226050 }, { "epoch": 0.4219277395095393, "grad_norm": 0.9474992752075195, "learning_rate": 0.00019838642457766924, "loss": 4.5989, "step": 226100 }, { "epoch": 0.4220210450689178, "grad_norm": 1.1194905042648315, "learning_rate": 0.00019838567286567985, "loss": 4.5753, "step": 226150 }, { "epoch": 0.42211435062829633, "grad_norm": 1.173120379447937, "learning_rate": 0.00019838492098005705, "loss": 4.6295, "step": 226200 }, { "epoch": 0.4222076561876748, "grad_norm": 0.8689015507698059, "learning_rate": 0.0001983841689208022, "loss": 4.6237, "step": 226250 }, { "epoch": 0.4223009617470533, "grad_norm": 1.1971979141235352, "learning_rate": 0.0001983834166879166, "loss": 4.5907, "step": 226300 }, { "epoch": 0.42239426730643176, "grad_norm": 1.1240501403808594, "learning_rate": 0.00019838266428140155, "loss": 4.6409, "step": 226350 }, { "epoch": 0.4224875728658103, "grad_norm": 0.8118560314178467, "learning_rate": 0.00019838191170125843, "loss": 4.4653, "step": 226400 }, { "epoch": 0.42258087842518877, "grad_norm": 0.9946422576904297, "learning_rate": 0.00019838115894748858, "loss": 4.542, "step": 226450 }, { "epoch": 0.42267418398456724, "grad_norm": 0.7898224592208862, "learning_rate": 0.00019838040602009323, "loss": 4.5506, "step": 226500 }, { "epoch": 0.4227674895439458, "grad_norm": 0.921465277671814, "learning_rate": 0.00019837965291907378, "loss": 4.6318, "step": 226550 }, { "epoch": 0.42286079510332425, "grad_norm": 1.0502792596817017, "learning_rate": 0.00019837889964443163, "loss": 4.5862, "step": 226600 }, { "epoch": 0.4229541006627027, "grad_norm": 0.8788375854492188, "learning_rate": 0.00019837814619616794, "loss": 4.6376, "step": 226650 }, { "epoch": 0.4230474062220812, "grad_norm": 0.8763489127159119, "learning_rate": 0.00019837739257428418, "loss": 4.8787, "step": 226700 }, { "epoch": 0.42314071178145973, "grad_norm": 0.9592618346214294, "learning_rate": 0.00019837663877878163, "loss": 4.5449, "step": 226750 }, { "epoch": 0.4232340173408382, "grad_norm": 0.5952169895172119, "learning_rate": 0.0001983758848096616, "loss": 4.506, "step": 226800 }, { "epoch": 0.4233273229002167, "grad_norm": 1.1658716201782227, "learning_rate": 0.00019837513066692548, "loss": 4.5342, "step": 226850 }, { "epoch": 0.4234206284595952, "grad_norm": 1.1232601404190063, "learning_rate": 0.00019837437635057453, "loss": 4.5859, "step": 226900 }, { "epoch": 0.4235139340189737, "grad_norm": 1.1959850788116455, "learning_rate": 0.00019837362186061013, "loss": 4.5792, "step": 226950 }, { "epoch": 0.42360723957835217, "grad_norm": 0.9566145539283752, "learning_rate": 0.0001983728671970336, "loss": 4.4619, "step": 227000 }, { "epoch": 0.42370054513773064, "grad_norm": 1.128745675086975, "learning_rate": 0.00019837211235984627, "loss": 4.6829, "step": 227050 }, { "epoch": 0.4237938506971092, "grad_norm": 1.116195559501648, "learning_rate": 0.00019837135734904947, "loss": 4.7385, "step": 227100 }, { "epoch": 0.42388715625648765, "grad_norm": 0.8514589667320251, "learning_rate": 0.00019837060216464452, "loss": 4.5505, "step": 227150 }, { "epoch": 0.4239804618158661, "grad_norm": 1.0022069215774536, "learning_rate": 0.00019836984680663276, "loss": 4.6128, "step": 227200 }, { "epoch": 0.42407376737524466, "grad_norm": 0.8854489326477051, "learning_rate": 0.00019836909127501554, "loss": 4.198, "step": 227250 }, { "epoch": 0.42416707293462313, "grad_norm": 0.9211372137069702, "learning_rate": 0.00019836833556979418, "loss": 4.5983, "step": 227300 }, { "epoch": 0.4242603784940016, "grad_norm": 0.9548203349113464, "learning_rate": 0.00019836757969097005, "loss": 4.5211, "step": 227350 }, { "epoch": 0.4243536840533801, "grad_norm": 0.9231182336807251, "learning_rate": 0.0001983668236385444, "loss": 4.5906, "step": 227400 }, { "epoch": 0.4244469896127586, "grad_norm": 1.0974338054656982, "learning_rate": 0.00019836606741251863, "loss": 4.8236, "step": 227450 }, { "epoch": 0.4245402951721371, "grad_norm": 1.2579002380371094, "learning_rate": 0.00019836531101289408, "loss": 4.7423, "step": 227500 }, { "epoch": 0.42463360073151557, "grad_norm": 0.9564178586006165, "learning_rate": 0.00019836455443967204, "loss": 4.5849, "step": 227550 }, { "epoch": 0.4247269062908941, "grad_norm": 0.7790428996086121, "learning_rate": 0.0001983637976928539, "loss": 4.4314, "step": 227600 }, { "epoch": 0.4248202118502726, "grad_norm": 1.0943279266357422, "learning_rate": 0.0001983630407724409, "loss": 4.4287, "step": 227650 }, { "epoch": 0.42491351740965105, "grad_norm": 0.8651321530342102, "learning_rate": 0.0001983622836784345, "loss": 4.6651, "step": 227700 }, { "epoch": 0.42500682296902953, "grad_norm": 1.2312800884246826, "learning_rate": 0.00019836152641083596, "loss": 4.6271, "step": 227750 }, { "epoch": 0.42510012852840806, "grad_norm": 0.8795189261436462, "learning_rate": 0.00019836076896964663, "loss": 4.7112, "step": 227800 }, { "epoch": 0.42519343408778654, "grad_norm": 0.861788809299469, "learning_rate": 0.00019836001135486786, "loss": 4.5759, "step": 227850 }, { "epoch": 0.425286739647165, "grad_norm": 0.9528782963752747, "learning_rate": 0.00019835925356650097, "loss": 4.6581, "step": 227900 }, { "epoch": 0.42538004520654354, "grad_norm": 0.9587149024009705, "learning_rate": 0.00019835849560454732, "loss": 4.3163, "step": 227950 }, { "epoch": 0.425473350765922, "grad_norm": 1.0508127212524414, "learning_rate": 0.00019835773746900822, "loss": 4.4864, "step": 228000 }, { "epoch": 0.425473350765922, "eval_loss": 4.791882514953613, "eval_runtime": 232.0509, "eval_samples_per_second": 11.239, "eval_steps_per_second": 11.239, "eval_tts_loss": 7.540685358295857, "step": 228000 }, { "epoch": 0.4255666563253005, "grad_norm": 1.0788441896438599, "learning_rate": 0.00019835697915988502, "loss": 4.6176, "step": 228050 }, { "epoch": 0.425659961884679, "grad_norm": 0.8570576310157776, "learning_rate": 0.00019835622067717904, "loss": 4.8553, "step": 228100 }, { "epoch": 0.4257532674440575, "grad_norm": 0.7701833844184875, "learning_rate": 0.00019835546202089165, "loss": 4.5462, "step": 228150 }, { "epoch": 0.425846573003436, "grad_norm": 0.9959319233894348, "learning_rate": 0.0001983547031910242, "loss": 4.4326, "step": 228200 }, { "epoch": 0.42593987856281446, "grad_norm": 0.9818710088729858, "learning_rate": 0.00019835394418757797, "loss": 4.7489, "step": 228250 }, { "epoch": 0.426033184122193, "grad_norm": 0.932654082775116, "learning_rate": 0.00019835318501055437, "loss": 4.571, "step": 228300 }, { "epoch": 0.42612648968157146, "grad_norm": 1.0089269876480103, "learning_rate": 0.00019835242565995466, "loss": 4.6125, "step": 228350 }, { "epoch": 0.42621979524094994, "grad_norm": 0.9701918363571167, "learning_rate": 0.00019835166613578026, "loss": 4.4908, "step": 228400 }, { "epoch": 0.4263131008003284, "grad_norm": 0.9656922221183777, "learning_rate": 0.00019835090643803246, "loss": 4.7232, "step": 228450 }, { "epoch": 0.42640640635970695, "grad_norm": 0.9137131571769714, "learning_rate": 0.0001983501465667126, "loss": 4.6645, "step": 228500 }, { "epoch": 0.4264997119190854, "grad_norm": 0.9897300601005554, "learning_rate": 0.00019834938652182204, "loss": 4.6834, "step": 228550 }, { "epoch": 0.4265930174784639, "grad_norm": 0.9887792468070984, "learning_rate": 0.00019834862630336214, "loss": 4.5679, "step": 228600 }, { "epoch": 0.42668632303784243, "grad_norm": 0.6781644821166992, "learning_rate": 0.0001983478659113342, "loss": 4.5704, "step": 228650 }, { "epoch": 0.4267796285972209, "grad_norm": 0.9480149745941162, "learning_rate": 0.00019834710534573962, "loss": 4.5627, "step": 228700 }, { "epoch": 0.4268729341565994, "grad_norm": 0.8901644945144653, "learning_rate": 0.00019834634460657964, "loss": 4.6213, "step": 228750 }, { "epoch": 0.42696623971597786, "grad_norm": 0.7598466873168945, "learning_rate": 0.00019834558369385573, "loss": 4.4153, "step": 228800 }, { "epoch": 0.4270595452753564, "grad_norm": 1.0075832605361938, "learning_rate": 0.00019834482260756912, "loss": 4.6078, "step": 228850 }, { "epoch": 0.42715285083473487, "grad_norm": 0.9788805842399597, "learning_rate": 0.00019834406134772124, "loss": 4.5693, "step": 228900 }, { "epoch": 0.42724615639411334, "grad_norm": 1.0992543697357178, "learning_rate": 0.00019834329991431338, "loss": 4.4368, "step": 228950 }, { "epoch": 0.4273394619534918, "grad_norm": 1.0395466089248657, "learning_rate": 0.0001983425383073469, "loss": 4.6703, "step": 229000 }, { "epoch": 0.42743276751287035, "grad_norm": 0.8829817771911621, "learning_rate": 0.00019834177652682314, "loss": 4.6088, "step": 229050 }, { "epoch": 0.4275260730722488, "grad_norm": 1.060113787651062, "learning_rate": 0.00019834101457274342, "loss": 4.5978, "step": 229100 }, { "epoch": 0.4276193786316273, "grad_norm": 1.1380964517593384, "learning_rate": 0.00019834025244510917, "loss": 4.5446, "step": 229150 }, { "epoch": 0.42771268419100583, "grad_norm": 1.2462329864501953, "learning_rate": 0.00019833949014392164, "loss": 4.3495, "step": 229200 }, { "epoch": 0.4278059897503843, "grad_norm": 0.9149359464645386, "learning_rate": 0.00019833872766918223, "loss": 4.5514, "step": 229250 }, { "epoch": 0.4278992953097628, "grad_norm": 0.770584762096405, "learning_rate": 0.00019833796502089225, "loss": 4.6721, "step": 229300 }, { "epoch": 0.42799260086914126, "grad_norm": 0.7645120024681091, "learning_rate": 0.00019833720219905308, "loss": 4.5611, "step": 229350 }, { "epoch": 0.4280859064285198, "grad_norm": 0.9542115330696106, "learning_rate": 0.00019833643920366605, "loss": 4.5287, "step": 229400 }, { "epoch": 0.42817921198789827, "grad_norm": 1.430662989616394, "learning_rate": 0.0001983356760347325, "loss": 4.4284, "step": 229450 }, { "epoch": 0.42827251754727674, "grad_norm": 0.8985748291015625, "learning_rate": 0.0001983349126922538, "loss": 4.5965, "step": 229500 }, { "epoch": 0.4283658231066553, "grad_norm": 0.901803195476532, "learning_rate": 0.00019833414917623128, "loss": 4.7505, "step": 229550 }, { "epoch": 0.42845912866603375, "grad_norm": 1.0440247058868408, "learning_rate": 0.0001983333854866663, "loss": 4.5123, "step": 229600 }, { "epoch": 0.4285524342254122, "grad_norm": 0.8303043842315674, "learning_rate": 0.00019833262162356015, "loss": 4.4886, "step": 229650 }, { "epoch": 0.4286457397847907, "grad_norm": 1.0328497886657715, "learning_rate": 0.00019833185758691427, "loss": 4.594, "step": 229700 }, { "epoch": 0.42873904534416923, "grad_norm": 1.0960317850112915, "learning_rate": 0.00019833109337672995, "loss": 4.5405, "step": 229750 }, { "epoch": 0.4288323509035477, "grad_norm": 0.9300960898399353, "learning_rate": 0.00019833032899300854, "loss": 4.7002, "step": 229800 }, { "epoch": 0.4289256564629262, "grad_norm": 1.045456051826477, "learning_rate": 0.0001983295644357514, "loss": 4.5931, "step": 229850 }, { "epoch": 0.4290189620223047, "grad_norm": 1.1042137145996094, "learning_rate": 0.0001983287997049599, "loss": 4.7307, "step": 229900 }, { "epoch": 0.4291122675816832, "grad_norm": 1.2343262434005737, "learning_rate": 0.00019832803480063538, "loss": 4.621, "step": 229950 }, { "epoch": 0.42920557314106167, "grad_norm": 1.0615925788879395, "learning_rate": 0.00019832726972277915, "loss": 4.6954, "step": 230000 }, { "epoch": 0.42929887870044015, "grad_norm": 1.0219675302505493, "learning_rate": 0.00019832650447139258, "loss": 4.6932, "step": 230050 }, { "epoch": 0.4293921842598187, "grad_norm": 0.7742038369178772, "learning_rate": 0.00019832573904647707, "loss": 4.649, "step": 230100 }, { "epoch": 0.42948548981919715, "grad_norm": 0.9617169499397278, "learning_rate": 0.00019832497344803393, "loss": 4.5471, "step": 230150 }, { "epoch": 0.42957879537857563, "grad_norm": 1.044552206993103, "learning_rate": 0.0001983242076760645, "loss": 4.4572, "step": 230200 }, { "epoch": 0.42967210093795416, "grad_norm": 0.9525365829467773, "learning_rate": 0.00019832344173057012, "loss": 4.6309, "step": 230250 }, { "epoch": 0.42976540649733264, "grad_norm": 0.9384802579879761, "learning_rate": 0.0001983226756115522, "loss": 4.6633, "step": 230300 }, { "epoch": 0.4298587120567111, "grad_norm": 0.7917965054512024, "learning_rate": 0.00019832190931901202, "loss": 4.6365, "step": 230350 }, { "epoch": 0.4299520176160896, "grad_norm": 0.9427030086517334, "learning_rate": 0.000198321142852951, "loss": 4.6762, "step": 230400 }, { "epoch": 0.4300453231754681, "grad_norm": 1.0474097728729248, "learning_rate": 0.00019832037621337047, "loss": 4.574, "step": 230450 }, { "epoch": 0.4301386287348466, "grad_norm": 1.0150837898254395, "learning_rate": 0.00019831960940027178, "loss": 4.6684, "step": 230500 }, { "epoch": 0.43023193429422507, "grad_norm": 1.0305523872375488, "learning_rate": 0.00019831884241365623, "loss": 4.4752, "step": 230550 }, { "epoch": 0.4303252398536036, "grad_norm": 1.2383668422698975, "learning_rate": 0.00019831807525352526, "loss": 4.4957, "step": 230600 }, { "epoch": 0.4304185454129821, "grad_norm": 1.0035946369171143, "learning_rate": 0.00019831730791988018, "loss": 4.5095, "step": 230650 }, { "epoch": 0.43051185097236055, "grad_norm": 0.8765376210212708, "learning_rate": 0.00019831654041272236, "loss": 4.3764, "step": 230700 }, { "epoch": 0.43060515653173903, "grad_norm": 1.094586730003357, "learning_rate": 0.00019831577273205314, "loss": 4.5836, "step": 230750 }, { "epoch": 0.43069846209111756, "grad_norm": 1.0345427989959717, "learning_rate": 0.00019831500487787386, "loss": 4.5329, "step": 230800 }, { "epoch": 0.43079176765049604, "grad_norm": 0.7867444157600403, "learning_rate": 0.00019831423685018592, "loss": 4.45, "step": 230850 }, { "epoch": 0.4308850732098745, "grad_norm": 1.01674222946167, "learning_rate": 0.00019831346864899066, "loss": 4.3076, "step": 230900 }, { "epoch": 0.43097837876925305, "grad_norm": 0.9102169871330261, "learning_rate": 0.0001983127002742894, "loss": 4.7141, "step": 230950 }, { "epoch": 0.4310716843286315, "grad_norm": 0.9253519773483276, "learning_rate": 0.00019831193172608353, "loss": 4.6601, "step": 231000 }, { "epoch": 0.4310716843286315, "eval_loss": 4.782818794250488, "eval_runtime": 230.7143, "eval_samples_per_second": 11.304, "eval_steps_per_second": 11.304, "eval_tts_loss": 7.499476979419509, "step": 231000 }, { "epoch": 0.43116498988801, "grad_norm": 0.9714641571044922, "learning_rate": 0.0001983111630043744, "loss": 4.826, "step": 231050 }, { "epoch": 0.4312582954473885, "grad_norm": 1.0827927589416504, "learning_rate": 0.0001983103941091634, "loss": 4.5804, "step": 231100 }, { "epoch": 0.431351601006767, "grad_norm": 0.9766815304756165, "learning_rate": 0.0001983096250404518, "loss": 4.4809, "step": 231150 }, { "epoch": 0.4314449065661455, "grad_norm": 1.1361207962036133, "learning_rate": 0.00019830885579824104, "loss": 4.7216, "step": 231200 }, { "epoch": 0.43153821212552396, "grad_norm": 0.9344854950904846, "learning_rate": 0.00019830808638253243, "loss": 4.4767, "step": 231250 }, { "epoch": 0.4316315176849025, "grad_norm": 0.6080878376960754, "learning_rate": 0.00019830731679332738, "loss": 4.582, "step": 231300 }, { "epoch": 0.43172482324428096, "grad_norm": 1.0502957105636597, "learning_rate": 0.00019830654703062716, "loss": 4.5399, "step": 231350 }, { "epoch": 0.43181812880365944, "grad_norm": 1.0605568885803223, "learning_rate": 0.0001983057770944332, "loss": 4.5047, "step": 231400 }, { "epoch": 0.4319114343630379, "grad_norm": 1.162824273109436, "learning_rate": 0.00019830500698474687, "loss": 4.6726, "step": 231450 }, { "epoch": 0.43200473992241645, "grad_norm": 0.957460343837738, "learning_rate": 0.0001983042367015695, "loss": 4.6275, "step": 231500 }, { "epoch": 0.4320980454817949, "grad_norm": 0.8100277185440063, "learning_rate": 0.0001983034662449024, "loss": 4.5962, "step": 231550 }, { "epoch": 0.4321913510411734, "grad_norm": 1.1255825757980347, "learning_rate": 0.00019830269561474704, "loss": 4.6635, "step": 231600 }, { "epoch": 0.4322846566005519, "grad_norm": 1.0411120653152466, "learning_rate": 0.0001983019248111047, "loss": 4.4638, "step": 231650 }, { "epoch": 0.4323779621599304, "grad_norm": 0.6740716099739075, "learning_rate": 0.00019830115383397674, "loss": 4.5441, "step": 231700 }, { "epoch": 0.4324712677193089, "grad_norm": 1.0795879364013672, "learning_rate": 0.00019830038268336454, "loss": 4.8265, "step": 231750 }, { "epoch": 0.43256457327868736, "grad_norm": 1.218779444694519, "learning_rate": 0.00019829961135926948, "loss": 4.7943, "step": 231800 }, { "epoch": 0.4326578788380659, "grad_norm": 0.9649580717086792, "learning_rate": 0.00019829883986169285, "loss": 4.5642, "step": 231850 }, { "epoch": 0.43275118439744437, "grad_norm": 0.9463967084884644, "learning_rate": 0.00019829806819063613, "loss": 4.5809, "step": 231900 }, { "epoch": 0.43284448995682284, "grad_norm": 1.0788993835449219, "learning_rate": 0.0001982972963461006, "loss": 4.5609, "step": 231950 }, { "epoch": 0.4329377955162013, "grad_norm": 1.184398889541626, "learning_rate": 0.0001982965243280876, "loss": 4.6776, "step": 232000 }, { "epoch": 0.43303110107557985, "grad_norm": 0.9360151886940002, "learning_rate": 0.00019829575213659858, "loss": 4.5851, "step": 232050 }, { "epoch": 0.4331244066349583, "grad_norm": 1.1861097812652588, "learning_rate": 0.00019829497977163482, "loss": 4.5354, "step": 232100 }, { "epoch": 0.4332177121943368, "grad_norm": 0.9626287221908569, "learning_rate": 0.00019829420723319773, "loss": 4.8662, "step": 232150 }, { "epoch": 0.43331101775371533, "grad_norm": 0.9649963974952698, "learning_rate": 0.00019829343452128865, "loss": 4.5901, "step": 232200 }, { "epoch": 0.4334043233130938, "grad_norm": 1.0452684164047241, "learning_rate": 0.000198292661635909, "loss": 4.7717, "step": 232250 }, { "epoch": 0.4334976288724723, "grad_norm": 0.963649332523346, "learning_rate": 0.00019829188857706004, "loss": 4.6523, "step": 232300 }, { "epoch": 0.43359093443185076, "grad_norm": 1.0967086553573608, "learning_rate": 0.00019829111534474324, "loss": 4.7513, "step": 232350 }, { "epoch": 0.4336842399912293, "grad_norm": 1.1611350774765015, "learning_rate": 0.00019829034193895987, "loss": 4.6203, "step": 232400 }, { "epoch": 0.43377754555060777, "grad_norm": 1.3620021343231201, "learning_rate": 0.00019828956835971136, "loss": 4.5346, "step": 232450 }, { "epoch": 0.43387085110998624, "grad_norm": 1.0658037662506104, "learning_rate": 0.00019828879460699904, "loss": 4.6917, "step": 232500 }, { "epoch": 0.4339641566693648, "grad_norm": 1.0672107934951782, "learning_rate": 0.00019828802068082432, "loss": 4.7152, "step": 232550 }, { "epoch": 0.43405746222874325, "grad_norm": 0.8459929823875427, "learning_rate": 0.0001982872465811885, "loss": 4.4931, "step": 232600 }, { "epoch": 0.4341507677881217, "grad_norm": 0.9519999027252197, "learning_rate": 0.00019828647230809302, "loss": 4.6345, "step": 232650 }, { "epoch": 0.4342440733475002, "grad_norm": 0.7836452126502991, "learning_rate": 0.00019828569786153918, "loss": 4.6803, "step": 232700 }, { "epoch": 0.43433737890687873, "grad_norm": 1.1806153059005737, "learning_rate": 0.00019828492324152844, "loss": 4.5936, "step": 232750 }, { "epoch": 0.4344306844662572, "grad_norm": 1.0644776821136475, "learning_rate": 0.00019828414844806204, "loss": 4.5906, "step": 232800 }, { "epoch": 0.4345239900256357, "grad_norm": 1.392525315284729, "learning_rate": 0.00019828337348114144, "loss": 4.6483, "step": 232850 }, { "epoch": 0.4346172955850142, "grad_norm": 1.0627968311309814, "learning_rate": 0.00019828259834076797, "loss": 4.6276, "step": 232900 }, { "epoch": 0.4347106011443927, "grad_norm": 0.71669602394104, "learning_rate": 0.000198281823026943, "loss": 4.5737, "step": 232950 }, { "epoch": 0.43480390670377117, "grad_norm": 1.0099847316741943, "learning_rate": 0.00019828104753966792, "loss": 4.491, "step": 233000 }, { "epoch": 0.43489721226314965, "grad_norm": 1.1775974035263062, "learning_rate": 0.0001982802718789441, "loss": 4.4005, "step": 233050 }, { "epoch": 0.4349905178225282, "grad_norm": 1.0557302236557007, "learning_rate": 0.00019827949604477285, "loss": 4.5978, "step": 233100 }, { "epoch": 0.43508382338190665, "grad_norm": 1.1460533142089844, "learning_rate": 0.0001982787200371556, "loss": 4.6675, "step": 233150 }, { "epoch": 0.43517712894128513, "grad_norm": 1.1376761198043823, "learning_rate": 0.00019827794385609372, "loss": 4.5659, "step": 233200 }, { "epoch": 0.43527043450066366, "grad_norm": 0.9409826993942261, "learning_rate": 0.00019827716750158854, "loss": 4.6083, "step": 233250 }, { "epoch": 0.43536374006004214, "grad_norm": 0.7400208711624146, "learning_rate": 0.00019827639097364145, "loss": 4.7026, "step": 233300 }, { "epoch": 0.4354570456194206, "grad_norm": 1.1714075803756714, "learning_rate": 0.00019827561427225384, "loss": 4.6554, "step": 233350 }, { "epoch": 0.4355503511787991, "grad_norm": 0.8486500382423401, "learning_rate": 0.00019827483739742705, "loss": 4.6074, "step": 233400 }, { "epoch": 0.4356436567381776, "grad_norm": 1.0664108991622925, "learning_rate": 0.00019827406034916248, "loss": 4.5721, "step": 233450 }, { "epoch": 0.4357369622975561, "grad_norm": 0.9367774128913879, "learning_rate": 0.00019827328312746145, "loss": 4.5607, "step": 233500 }, { "epoch": 0.4358302678569346, "grad_norm": 1.0547155141830444, "learning_rate": 0.00019827250573232542, "loss": 4.5985, "step": 233550 }, { "epoch": 0.4359235734163131, "grad_norm": 0.9410718679428101, "learning_rate": 0.00019827172816375566, "loss": 4.5643, "step": 233600 }, { "epoch": 0.4360168789756916, "grad_norm": 0.9093536734580994, "learning_rate": 0.0001982709504217536, "loss": 4.6474, "step": 233650 }, { "epoch": 0.43611018453507006, "grad_norm": 0.9456886053085327, "learning_rate": 0.00019827017250632063, "loss": 4.7075, "step": 233700 }, { "epoch": 0.43620349009444853, "grad_norm": 1.0371719598770142, "learning_rate": 0.00019826939441745805, "loss": 4.5081, "step": 233750 }, { "epoch": 0.43629679565382706, "grad_norm": 0.7762008905410767, "learning_rate": 0.00019826861615516734, "loss": 4.5744, "step": 233800 }, { "epoch": 0.43639010121320554, "grad_norm": 0.8495267629623413, "learning_rate": 0.0001982678377194498, "loss": 4.6972, "step": 233850 }, { "epoch": 0.436483406772584, "grad_norm": 0.9476446509361267, "learning_rate": 0.00019826705911030678, "loss": 4.6825, "step": 233900 }, { "epoch": 0.43657671233196255, "grad_norm": 0.92109614610672, "learning_rate": 0.0001982662803277397, "loss": 4.5082, "step": 233950 }, { "epoch": 0.436670017891341, "grad_norm": 1.5745224952697754, "learning_rate": 0.00019826550137174996, "loss": 4.702, "step": 234000 }, { "epoch": 0.436670017891341, "eval_loss": 4.786718368530273, "eval_runtime": 234.4074, "eval_samples_per_second": 11.126, "eval_steps_per_second": 11.126, "eval_tts_loss": 7.521374580757464, "step": 234000 }, { "epoch": 0.4367633234507195, "grad_norm": 0.6546002626419067, "learning_rate": 0.00019826472224233888, "loss": 4.6363, "step": 234050 }, { "epoch": 0.436856629010098, "grad_norm": 1.0382810831069946, "learning_rate": 0.00019826394293950784, "loss": 4.6492, "step": 234100 }, { "epoch": 0.4369499345694765, "grad_norm": 1.125950574874878, "learning_rate": 0.00019826316346325826, "loss": 4.6651, "step": 234150 }, { "epoch": 0.437043240128855, "grad_norm": 0.9821774959564209, "learning_rate": 0.00019826238381359147, "loss": 4.4427, "step": 234200 }, { "epoch": 0.43713654568823346, "grad_norm": 0.9345731735229492, "learning_rate": 0.00019826160399050887, "loss": 4.6285, "step": 234250 }, { "epoch": 0.43722985124761193, "grad_norm": 1.1006258726119995, "learning_rate": 0.0001982608239940118, "loss": 4.6635, "step": 234300 }, { "epoch": 0.43732315680699047, "grad_norm": 0.8882812261581421, "learning_rate": 0.0001982600438241017, "loss": 4.5461, "step": 234350 }, { "epoch": 0.43741646236636894, "grad_norm": 0.8373991847038269, "learning_rate": 0.0001982592634807799, "loss": 4.6949, "step": 234400 }, { "epoch": 0.4375097679257474, "grad_norm": 0.9993740916252136, "learning_rate": 0.00019825848296404778, "loss": 4.6201, "step": 234450 }, { "epoch": 0.43760307348512595, "grad_norm": 1.093626856803894, "learning_rate": 0.00019825770227390675, "loss": 4.6239, "step": 234500 }, { "epoch": 0.4376963790445044, "grad_norm": 1.159334421157837, "learning_rate": 0.00019825692141035815, "loss": 4.4909, "step": 234550 }, { "epoch": 0.4377896846038829, "grad_norm": 0.8569579124450684, "learning_rate": 0.0001982561403734034, "loss": 4.54, "step": 234600 }, { "epoch": 0.4378829901632614, "grad_norm": 1.0872637033462524, "learning_rate": 0.00019825535916304378, "loss": 4.5892, "step": 234650 }, { "epoch": 0.4379762957226399, "grad_norm": 1.0390820503234863, "learning_rate": 0.0001982545777792808, "loss": 4.4236, "step": 234700 }, { "epoch": 0.4380696012820184, "grad_norm": 0.8615942597389221, "learning_rate": 0.00019825379622211576, "loss": 4.7319, "step": 234750 }, { "epoch": 0.43816290684139686, "grad_norm": 0.9036928415298462, "learning_rate": 0.00019825301449155006, "loss": 4.6797, "step": 234800 }, { "epoch": 0.4382562124007754, "grad_norm": 0.8488964438438416, "learning_rate": 0.00019825223258758507, "loss": 4.6368, "step": 234850 }, { "epoch": 0.43834951796015387, "grad_norm": 0.7917913198471069, "learning_rate": 0.0001982514505102222, "loss": 4.5903, "step": 234900 }, { "epoch": 0.43844282351953234, "grad_norm": 1.0873136520385742, "learning_rate": 0.0001982506682594628, "loss": 4.5301, "step": 234950 }, { "epoch": 0.4385361290789108, "grad_norm": 1.2058392763137817, "learning_rate": 0.00019824988583530824, "loss": 4.3379, "step": 235000 }, { "epoch": 0.43862943463828935, "grad_norm": 1.043069839477539, "learning_rate": 0.00019824910323775993, "loss": 4.9557, "step": 235050 }, { "epoch": 0.4387227401976678, "grad_norm": 0.8327590823173523, "learning_rate": 0.00019824832046681925, "loss": 4.5355, "step": 235100 }, { "epoch": 0.4388160457570463, "grad_norm": 1.312573790550232, "learning_rate": 0.00019824753752248754, "loss": 4.7489, "step": 235150 }, { "epoch": 0.43890935131642483, "grad_norm": 1.0750068426132202, "learning_rate": 0.00019824675440476624, "loss": 4.61, "step": 235200 }, { "epoch": 0.4390026568758033, "grad_norm": 1.0304391384124756, "learning_rate": 0.0001982459711136567, "loss": 4.5803, "step": 235250 }, { "epoch": 0.4390959624351818, "grad_norm": 1.0730409622192383, "learning_rate": 0.0001982451876491603, "loss": 4.661, "step": 235300 }, { "epoch": 0.43918926799456026, "grad_norm": 1.1401572227478027, "learning_rate": 0.00019824440401127846, "loss": 4.5335, "step": 235350 }, { "epoch": 0.4392825735539388, "grad_norm": 1.187408685684204, "learning_rate": 0.00019824362020001251, "loss": 4.6488, "step": 235400 }, { "epoch": 0.43937587911331727, "grad_norm": 0.830162763595581, "learning_rate": 0.00019824283621536386, "loss": 4.716, "step": 235450 }, { "epoch": 0.43946918467269575, "grad_norm": 1.0726609230041504, "learning_rate": 0.0001982420520573339, "loss": 4.5043, "step": 235500 }, { "epoch": 0.4395624902320743, "grad_norm": 0.8797732591629028, "learning_rate": 0.000198241267725924, "loss": 4.5403, "step": 235550 }, { "epoch": 0.43965579579145275, "grad_norm": 1.1645294427871704, "learning_rate": 0.00019824048322113552, "loss": 4.5505, "step": 235600 }, { "epoch": 0.43974910135083123, "grad_norm": 1.0088536739349365, "learning_rate": 0.0001982396985429699, "loss": 4.5699, "step": 235650 }, { "epoch": 0.4398424069102097, "grad_norm": 1.086678385734558, "learning_rate": 0.00019823891369142847, "loss": 4.5694, "step": 235700 }, { "epoch": 0.43993571246958824, "grad_norm": 1.2575753927230835, "learning_rate": 0.00019823812866651271, "loss": 4.6272, "step": 235750 }, { "epoch": 0.4400290180289667, "grad_norm": 0.8871008157730103, "learning_rate": 0.00019823734346822387, "loss": 4.5924, "step": 235800 }, { "epoch": 0.4401223235883452, "grad_norm": 0.8986257314682007, "learning_rate": 0.00019823655809656345, "loss": 4.7974, "step": 235850 }, { "epoch": 0.4402156291477237, "grad_norm": 0.5982351899147034, "learning_rate": 0.00019823577255153276, "loss": 4.4457, "step": 235900 }, { "epoch": 0.4403089347071022, "grad_norm": 1.1470563411712646, "learning_rate": 0.00019823498683313324, "loss": 4.5922, "step": 235950 }, { "epoch": 0.44040224026648067, "grad_norm": 0.9602048397064209, "learning_rate": 0.00019823420094136623, "loss": 4.5407, "step": 236000 }, { "epoch": 0.44049554582585915, "grad_norm": 1.228560209274292, "learning_rate": 0.00019823341487623314, "loss": 4.6289, "step": 236050 }, { "epoch": 0.4405888513852377, "grad_norm": 1.1307828426361084, "learning_rate": 0.00019823262863773538, "loss": 4.5017, "step": 236100 }, { "epoch": 0.44068215694461615, "grad_norm": 0.942937970161438, "learning_rate": 0.0001982318422258743, "loss": 4.6442, "step": 236150 }, { "epoch": 0.44077546250399463, "grad_norm": 0.9702286720275879, "learning_rate": 0.00019823105564065129, "loss": 4.5807, "step": 236200 }, { "epoch": 0.44086876806337316, "grad_norm": 0.9227506518363953, "learning_rate": 0.00019823026888206777, "loss": 4.5915, "step": 236250 }, { "epoch": 0.44096207362275164, "grad_norm": 0.7443217635154724, "learning_rate": 0.00019822948195012512, "loss": 4.5009, "step": 236300 }, { "epoch": 0.4410553791821301, "grad_norm": 0.9936513304710388, "learning_rate": 0.0001982286948448247, "loss": 4.7613, "step": 236350 }, { "epoch": 0.4411486847415086, "grad_norm": 0.7316447496414185, "learning_rate": 0.00019822790756616793, "loss": 4.4194, "step": 236400 }, { "epoch": 0.4412419903008871, "grad_norm": 0.6408354640007019, "learning_rate": 0.00019822712011415615, "loss": 4.5336, "step": 236450 }, { "epoch": 0.4413352958602656, "grad_norm": 1.0418235063552856, "learning_rate": 0.00019822633248879084, "loss": 4.4547, "step": 236500 }, { "epoch": 0.4414286014196441, "grad_norm": 1.010443925857544, "learning_rate": 0.0001982255446900733, "loss": 4.5134, "step": 236550 }, { "epoch": 0.4415219069790226, "grad_norm": 1.3255016803741455, "learning_rate": 0.00019822475671800498, "loss": 4.7933, "step": 236600 }, { "epoch": 0.4416152125384011, "grad_norm": 0.9770745038986206, "learning_rate": 0.00019822396857258724, "loss": 4.7556, "step": 236650 }, { "epoch": 0.44170851809777956, "grad_norm": 1.0298558473587036, "learning_rate": 0.00019822318025382147, "loss": 4.6711, "step": 236700 }, { "epoch": 0.44180182365715803, "grad_norm": 0.7105265259742737, "learning_rate": 0.00019822239176170908, "loss": 4.5719, "step": 236750 }, { "epoch": 0.44189512921653656, "grad_norm": 0.9798986315727234, "learning_rate": 0.00019822160309625143, "loss": 4.5355, "step": 236800 }, { "epoch": 0.44198843477591504, "grad_norm": 1.1659282445907593, "learning_rate": 0.00019822081425744997, "loss": 4.6384, "step": 236850 }, { "epoch": 0.4420817403352935, "grad_norm": 0.9095149636268616, "learning_rate": 0.00019822002524530602, "loss": 4.5297, "step": 236900 }, { "epoch": 0.442175045894672, "grad_norm": 0.8372174501419067, "learning_rate": 0.00019821923605982104, "loss": 4.6121, "step": 236950 }, { "epoch": 0.4422683514540505, "grad_norm": 0.9444534182548523, "learning_rate": 0.00019821844670099637, "loss": 4.5425, "step": 237000 }, { "epoch": 0.4422683514540505, "eval_loss": 4.778234958648682, "eval_runtime": 232.886, "eval_samples_per_second": 11.199, "eval_steps_per_second": 11.199, "eval_tts_loss": 7.544113754277672, "step": 237000 }, { "epoch": 0.442361657013429, "grad_norm": 1.01829195022583, "learning_rate": 0.0001982176571688334, "loss": 4.5883, "step": 237050 }, { "epoch": 0.4424549625728075, "grad_norm": 0.8927636742591858, "learning_rate": 0.0001982168674633336, "loss": 4.6002, "step": 237100 }, { "epoch": 0.442548268132186, "grad_norm": 0.9533956050872803, "learning_rate": 0.00019821607758449826, "loss": 4.4706, "step": 237150 }, { "epoch": 0.4426415736915645, "grad_norm": 1.0447924137115479, "learning_rate": 0.00019821528753232888, "loss": 4.7974, "step": 237200 }, { "epoch": 0.44273487925094296, "grad_norm": 0.9149044156074524, "learning_rate": 0.00019821449730682676, "loss": 4.4807, "step": 237250 }, { "epoch": 0.44282818481032143, "grad_norm": 0.6954190135002136, "learning_rate": 0.00019821370690799333, "loss": 4.3259, "step": 237300 }, { "epoch": 0.44292149036969997, "grad_norm": 0.8902153372764587, "learning_rate": 0.00019821291633583001, "loss": 4.6436, "step": 237350 }, { "epoch": 0.44301479592907844, "grad_norm": 0.9041743874549866, "learning_rate": 0.00019821212559033814, "loss": 4.5148, "step": 237400 }, { "epoch": 0.4431081014884569, "grad_norm": 1.206248164176941, "learning_rate": 0.0001982113346715192, "loss": 4.5886, "step": 237450 }, { "epoch": 0.44320140704783545, "grad_norm": 0.9443100690841675, "learning_rate": 0.0001982105435793745, "loss": 4.5412, "step": 237500 }, { "epoch": 0.4432947126072139, "grad_norm": 1.071516990661621, "learning_rate": 0.00019820975231390548, "loss": 4.5386, "step": 237550 }, { "epoch": 0.4433880181665924, "grad_norm": 1.0057834386825562, "learning_rate": 0.0001982089608751135, "loss": 4.6202, "step": 237600 }, { "epoch": 0.4434813237259709, "grad_norm": 1.0348365306854248, "learning_rate": 0.00019820816926300006, "loss": 4.4221, "step": 237650 }, { "epoch": 0.4435746292853494, "grad_norm": 0.8951256275177002, "learning_rate": 0.00019820737747756642, "loss": 4.5601, "step": 237700 }, { "epoch": 0.4436679348447279, "grad_norm": 0.783212423324585, "learning_rate": 0.00019820658551881405, "loss": 4.5649, "step": 237750 }, { "epoch": 0.44376124040410636, "grad_norm": 1.0642147064208984, "learning_rate": 0.0001982057933867443, "loss": 4.8408, "step": 237800 }, { "epoch": 0.4438545459634849, "grad_norm": 0.9804556965827942, "learning_rate": 0.00019820500108135865, "loss": 4.7134, "step": 237850 }, { "epoch": 0.44394785152286337, "grad_norm": 0.9035722017288208, "learning_rate": 0.0001982042086026585, "loss": 4.6311, "step": 237900 }, { "epoch": 0.44404115708224184, "grad_norm": 1.0001559257507324, "learning_rate": 0.00019820341595064514, "loss": 4.804, "step": 237950 }, { "epoch": 0.4441344626416203, "grad_norm": 1.2685824632644653, "learning_rate": 0.00019820262312532, "loss": 4.6524, "step": 238000 }, { "epoch": 0.44422776820099885, "grad_norm": 1.238340973854065, "learning_rate": 0.00019820183012668458, "loss": 4.6042, "step": 238050 }, { "epoch": 0.4443210737603773, "grad_norm": 1.1636897325515747, "learning_rate": 0.00019820103695474014, "loss": 4.562, "step": 238100 }, { "epoch": 0.4444143793197558, "grad_norm": 1.0242795944213867, "learning_rate": 0.0001982002436094882, "loss": 4.3683, "step": 238150 }, { "epoch": 0.44450768487913433, "grad_norm": 0.8869125843048096, "learning_rate": 0.0001981994500909301, "loss": 4.6679, "step": 238200 }, { "epoch": 0.4446009904385128, "grad_norm": 0.8452954292297363, "learning_rate": 0.00019819865639906725, "loss": 4.6062, "step": 238250 }, { "epoch": 0.4446942959978913, "grad_norm": 1.0370817184448242, "learning_rate": 0.00019819786253390102, "loss": 4.5246, "step": 238300 }, { "epoch": 0.44478760155726976, "grad_norm": 1.0226502418518066, "learning_rate": 0.0001981970684954329, "loss": 4.7068, "step": 238350 }, { "epoch": 0.4448809071166483, "grad_norm": 0.9036027789115906, "learning_rate": 0.00019819627428366418, "loss": 4.5923, "step": 238400 }, { "epoch": 0.44497421267602677, "grad_norm": 1.0699036121368408, "learning_rate": 0.00019819547989859633, "loss": 4.8253, "step": 238450 }, { "epoch": 0.44506751823540525, "grad_norm": 0.9457048177719116, "learning_rate": 0.00019819468534023073, "loss": 4.3733, "step": 238500 }, { "epoch": 0.4451608237947838, "grad_norm": 1.337355375289917, "learning_rate": 0.0001981938906085688, "loss": 4.4989, "step": 238550 }, { "epoch": 0.44525412935416225, "grad_norm": 1.0064990520477295, "learning_rate": 0.00019819309570361194, "loss": 4.7772, "step": 238600 }, { "epoch": 0.44534743491354073, "grad_norm": 0.9159983396530151, "learning_rate": 0.00019819230062536152, "loss": 4.6947, "step": 238650 }, { "epoch": 0.4454407404729192, "grad_norm": 1.0956370830535889, "learning_rate": 0.00019819150537381898, "loss": 4.6732, "step": 238700 }, { "epoch": 0.44553404603229774, "grad_norm": 0.8136811852455139, "learning_rate": 0.0001981907099489857, "loss": 4.5511, "step": 238750 }, { "epoch": 0.4456273515916762, "grad_norm": 0.7191414833068848, "learning_rate": 0.00019818991435086313, "loss": 4.8481, "step": 238800 }, { "epoch": 0.4457206571510547, "grad_norm": 1.0575743913650513, "learning_rate": 0.00019818911857945258, "loss": 4.4962, "step": 238850 }, { "epoch": 0.4458139627104332, "grad_norm": 0.9950775504112244, "learning_rate": 0.00019818832263475552, "loss": 4.44, "step": 238900 }, { "epoch": 0.4459072682698117, "grad_norm": 1.380645990371704, "learning_rate": 0.00019818752651677338, "loss": 4.5819, "step": 238950 }, { "epoch": 0.44600057382919017, "grad_norm": 1.2283092737197876, "learning_rate": 0.00019818673022550755, "loss": 4.6796, "step": 239000 }, { "epoch": 0.44609387938856865, "grad_norm": 1.224413275718689, "learning_rate": 0.00019818593376095937, "loss": 4.6762, "step": 239050 }, { "epoch": 0.4461871849479472, "grad_norm": 1.0246546268463135, "learning_rate": 0.00019818513712313032, "loss": 4.6132, "step": 239100 }, { "epoch": 0.44628049050732566, "grad_norm": 1.1078890562057495, "learning_rate": 0.00019818434031202177, "loss": 4.5609, "step": 239150 }, { "epoch": 0.44637379606670413, "grad_norm": 1.020003080368042, "learning_rate": 0.00019818354332763513, "loss": 4.5842, "step": 239200 }, { "epoch": 0.44646710162608266, "grad_norm": 0.9246664643287659, "learning_rate": 0.00019818274616997182, "loss": 4.7182, "step": 239250 }, { "epoch": 0.44656040718546114, "grad_norm": 0.9110824465751648, "learning_rate": 0.00019818194883903324, "loss": 4.4936, "step": 239300 }, { "epoch": 0.4466537127448396, "grad_norm": 1.1827949285507202, "learning_rate": 0.0001981811513348208, "loss": 4.3821, "step": 239350 }, { "epoch": 0.4467470183042181, "grad_norm": 0.9574491381645203, "learning_rate": 0.0001981803536573359, "loss": 4.8013, "step": 239400 }, { "epoch": 0.4468403238635966, "grad_norm": 1.1559712886810303, "learning_rate": 0.00019817955580657995, "loss": 4.4273, "step": 239450 }, { "epoch": 0.4469336294229751, "grad_norm": 0.8281153440475464, "learning_rate": 0.00019817875778255434, "loss": 4.6266, "step": 239500 }, { "epoch": 0.4470269349823536, "grad_norm": 1.210089087486267, "learning_rate": 0.00019817795958526052, "loss": 4.5594, "step": 239550 }, { "epoch": 0.44712024054173205, "grad_norm": 1.2433289289474487, "learning_rate": 0.00019817716121469986, "loss": 4.5299, "step": 239600 }, { "epoch": 0.4472135461011106, "grad_norm": 1.1075997352600098, "learning_rate": 0.00019817636267087382, "loss": 4.3679, "step": 239650 }, { "epoch": 0.44730685166048906, "grad_norm": 1.0246319770812988, "learning_rate": 0.00019817556395378374, "loss": 4.8387, "step": 239700 }, { "epoch": 0.44740015721986753, "grad_norm": 1.2445541620254517, "learning_rate": 0.00019817476506343108, "loss": 4.6812, "step": 239750 }, { "epoch": 0.44749346277924607, "grad_norm": 0.791220486164093, "learning_rate": 0.00019817396599981725, "loss": 4.6479, "step": 239800 }, { "epoch": 0.44758676833862454, "grad_norm": 1.1053506135940552, "learning_rate": 0.0001981731667629436, "loss": 4.5058, "step": 239850 }, { "epoch": 0.447680073898003, "grad_norm": 0.7307910919189453, "learning_rate": 0.00019817236735281162, "loss": 4.4622, "step": 239900 }, { "epoch": 0.4477733794573815, "grad_norm": 0.864814281463623, "learning_rate": 0.00019817156776942265, "loss": 4.5375, "step": 239950 }, { "epoch": 0.44786668501676, "grad_norm": 1.1574825048446655, "learning_rate": 0.0001981707680127782, "loss": 4.661, "step": 240000 }, { "epoch": 0.44786668501676, "eval_loss": 4.771146774291992, "eval_runtime": 230.5723, "eval_samples_per_second": 11.311, "eval_steps_per_second": 11.311, "eval_tts_loss": 7.4719085190530885, "step": 240000 }, { "epoch": 0.4479599905761385, "grad_norm": 1.3045135736465454, "learning_rate": 0.00019816996808287956, "loss": 4.6628, "step": 240050 }, { "epoch": 0.448053296135517, "grad_norm": 1.4040054082870483, "learning_rate": 0.0001981691679797282, "loss": 4.763, "step": 240100 }, { "epoch": 0.4481466016948955, "grad_norm": 0.8123361468315125, "learning_rate": 0.00019816836770332556, "loss": 4.6492, "step": 240150 }, { "epoch": 0.448239907254274, "grad_norm": 0.8047076463699341, "learning_rate": 0.000198167567253673, "loss": 4.7971, "step": 240200 }, { "epoch": 0.44833321281365246, "grad_norm": 0.7521131634712219, "learning_rate": 0.000198166766630772, "loss": 4.7822, "step": 240250 }, { "epoch": 0.44842651837303094, "grad_norm": 0.808711051940918, "learning_rate": 0.0001981659658346239, "loss": 4.5455, "step": 240300 }, { "epoch": 0.44851982393240947, "grad_norm": 0.6877180933952332, "learning_rate": 0.00019816516486523015, "loss": 4.5178, "step": 240350 }, { "epoch": 0.44861312949178794, "grad_norm": 1.0691016912460327, "learning_rate": 0.00019816436372259216, "loss": 4.6918, "step": 240400 }, { "epoch": 0.4487064350511664, "grad_norm": 1.0807589292526245, "learning_rate": 0.0001981635624067113, "loss": 4.7728, "step": 240450 }, { "epoch": 0.44879974061054495, "grad_norm": 0.9326990842819214, "learning_rate": 0.00019816276091758907, "loss": 4.5443, "step": 240500 }, { "epoch": 0.4488930461699234, "grad_norm": 0.9165334105491638, "learning_rate": 0.00019816195925522682, "loss": 4.6545, "step": 240550 }, { "epoch": 0.4489863517293019, "grad_norm": 0.8920701146125793, "learning_rate": 0.000198161157419626, "loss": 4.7592, "step": 240600 }, { "epoch": 0.4490796572886804, "grad_norm": 1.1457407474517822, "learning_rate": 0.000198160355410788, "loss": 4.5365, "step": 240650 }, { "epoch": 0.4491729628480589, "grad_norm": 0.9040488004684448, "learning_rate": 0.00019815955322871425, "loss": 4.6286, "step": 240700 }, { "epoch": 0.4492662684074374, "grad_norm": 0.8481893539428711, "learning_rate": 0.00019815875087340613, "loss": 4.5136, "step": 240750 }, { "epoch": 0.44935957396681586, "grad_norm": 1.0525832176208496, "learning_rate": 0.00019815794834486511, "loss": 4.791, "step": 240800 }, { "epoch": 0.4494528795261944, "grad_norm": 1.0377291440963745, "learning_rate": 0.00019815714564309258, "loss": 4.6304, "step": 240850 }, { "epoch": 0.44954618508557287, "grad_norm": 0.7769307494163513, "learning_rate": 0.00019815634276808996, "loss": 4.4798, "step": 240900 }, { "epoch": 0.44963949064495135, "grad_norm": 0.9710248112678528, "learning_rate": 0.00019815553971985867, "loss": 4.4885, "step": 240950 }, { "epoch": 0.4497327962043298, "grad_norm": 1.0528430938720703, "learning_rate": 0.00019815473649840014, "loss": 4.6653, "step": 241000 }, { "epoch": 0.44982610176370835, "grad_norm": 0.9350594282150269, "learning_rate": 0.00019815393310371575, "loss": 4.6053, "step": 241050 }, { "epoch": 0.44991940732308683, "grad_norm": 0.9134218692779541, "learning_rate": 0.00019815312953580692, "loss": 4.6681, "step": 241100 }, { "epoch": 0.4500127128824653, "grad_norm": 1.070358395576477, "learning_rate": 0.00019815232579467514, "loss": 4.4734, "step": 241150 }, { "epoch": 0.45010601844184384, "grad_norm": 1.081626296043396, "learning_rate": 0.00019815152188032173, "loss": 4.5605, "step": 241200 }, { "epoch": 0.4501993240012223, "grad_norm": 1.0206654071807861, "learning_rate": 0.00019815071779274816, "loss": 4.7068, "step": 241250 }, { "epoch": 0.4502926295606008, "grad_norm": 0.9931300282478333, "learning_rate": 0.00019814991353195587, "loss": 4.5779, "step": 241300 }, { "epoch": 0.45038593511997926, "grad_norm": 1.139255166053772, "learning_rate": 0.00019814910909794623, "loss": 4.5736, "step": 241350 }, { "epoch": 0.4504792406793578, "grad_norm": 1.1006131172180176, "learning_rate": 0.00019814830449072068, "loss": 4.7273, "step": 241400 }, { "epoch": 0.45057254623873627, "grad_norm": 0.8194779753684998, "learning_rate": 0.00019814749971028067, "loss": 4.7137, "step": 241450 }, { "epoch": 0.45066585179811475, "grad_norm": 1.067018747329712, "learning_rate": 0.00019814669475662754, "loss": 4.5206, "step": 241500 }, { "epoch": 0.4507591573574933, "grad_norm": 1.050444483757019, "learning_rate": 0.0001981458896297628, "loss": 4.6427, "step": 241550 }, { "epoch": 0.45085246291687175, "grad_norm": 1.1245814561843872, "learning_rate": 0.00019814508432968784, "loss": 4.4676, "step": 241600 }, { "epoch": 0.45094576847625023, "grad_norm": 1.1817681789398193, "learning_rate": 0.00019814427885640407, "loss": 4.8481, "step": 241650 }, { "epoch": 0.4510390740356287, "grad_norm": 1.0630196332931519, "learning_rate": 0.00019814347320991292, "loss": 4.4368, "step": 241700 }, { "epoch": 0.45113237959500724, "grad_norm": 1.3640244007110596, "learning_rate": 0.00019814266739021577, "loss": 4.5863, "step": 241750 }, { "epoch": 0.4512256851543857, "grad_norm": 0.9691682457923889, "learning_rate": 0.00019814186139731409, "loss": 4.6831, "step": 241800 }, { "epoch": 0.4513189907137642, "grad_norm": 0.8320736289024353, "learning_rate": 0.00019814105523120935, "loss": 4.4026, "step": 241850 }, { "epoch": 0.4514122962731427, "grad_norm": 0.7617359757423401, "learning_rate": 0.00019814024889190285, "loss": 4.591, "step": 241900 }, { "epoch": 0.4515056018325212, "grad_norm": 0.9816135168075562, "learning_rate": 0.00019813944237939612, "loss": 4.7765, "step": 241950 }, { "epoch": 0.4515989073918997, "grad_norm": 0.8725814819335938, "learning_rate": 0.00019813863569369052, "loss": 4.5988, "step": 242000 }, { "epoch": 0.45169221295127815, "grad_norm": 0.9102336764335632, "learning_rate": 0.00019813782883478754, "loss": 4.6437, "step": 242050 }, { "epoch": 0.4517855185106567, "grad_norm": 1.0546931028366089, "learning_rate": 0.0001981370218026885, "loss": 4.5362, "step": 242100 }, { "epoch": 0.45187882407003516, "grad_norm": 1.28167724609375, "learning_rate": 0.00019813621459739494, "loss": 4.7866, "step": 242150 }, { "epoch": 0.45197212962941363, "grad_norm": 1.2853349447250366, "learning_rate": 0.00019813540721890817, "loss": 4.6789, "step": 242200 }, { "epoch": 0.4520654351887921, "grad_norm": 1.1249480247497559, "learning_rate": 0.0001981345996672297, "loss": 4.4926, "step": 242250 }, { "epoch": 0.45215874074817064, "grad_norm": 0.5457555651664734, "learning_rate": 0.00019813379194236094, "loss": 4.4383, "step": 242300 }, { "epoch": 0.4522520463075491, "grad_norm": 1.1314606666564941, "learning_rate": 0.00019813298404430332, "loss": 4.7619, "step": 242350 }, { "epoch": 0.4523453518669276, "grad_norm": 1.0931801795959473, "learning_rate": 0.0001981321759730582, "loss": 4.7695, "step": 242400 }, { "epoch": 0.4524386574263061, "grad_norm": 1.2083085775375366, "learning_rate": 0.00019813136772862708, "loss": 4.5025, "step": 242450 }, { "epoch": 0.4525319629856846, "grad_norm": 1.1209338903427124, "learning_rate": 0.00019813055931101138, "loss": 4.649, "step": 242500 }, { "epoch": 0.4526252685450631, "grad_norm": 1.0771358013153076, "learning_rate": 0.0001981297507202125, "loss": 4.5248, "step": 242550 }, { "epoch": 0.45271857410444155, "grad_norm": 1.0569473505020142, "learning_rate": 0.00019812894195623187, "loss": 4.5961, "step": 242600 }, { "epoch": 0.4528118796638201, "grad_norm": 0.7088388204574585, "learning_rate": 0.0001981281330190709, "loss": 4.6979, "step": 242650 }, { "epoch": 0.45290518522319856, "grad_norm": 0.8838179707527161, "learning_rate": 0.0001981273239087311, "loss": 4.7185, "step": 242700 }, { "epoch": 0.45299849078257703, "grad_norm": 1.145406723022461, "learning_rate": 0.0001981265146252138, "loss": 4.6553, "step": 242750 }, { "epoch": 0.45309179634195557, "grad_norm": 0.8475630879402161, "learning_rate": 0.00019812570516852048, "loss": 4.6364, "step": 242800 }, { "epoch": 0.45318510190133404, "grad_norm": 0.8554997444152832, "learning_rate": 0.00019812489553865255, "loss": 4.3752, "step": 242850 }, { "epoch": 0.4532784074607125, "grad_norm": 1.0469423532485962, "learning_rate": 0.00019812408573561144, "loss": 4.6012, "step": 242900 }, { "epoch": 0.453371713020091, "grad_norm": 0.9449438452720642, "learning_rate": 0.00019812327575939858, "loss": 4.4068, "step": 242950 }, { "epoch": 0.4534650185794695, "grad_norm": 0.8156939744949341, "learning_rate": 0.00019812246561001543, "loss": 4.4701, "step": 243000 }, { "epoch": 0.4534650185794695, "eval_loss": 4.782342910766602, "eval_runtime": 233.441, "eval_samples_per_second": 11.172, "eval_steps_per_second": 11.172, "eval_tts_loss": 7.457396795253263, "step": 243000 }, { "epoch": 0.453558324138848, "grad_norm": 1.1916180849075317, "learning_rate": 0.00019812165528746337, "loss": 4.5937, "step": 243050 }, { "epoch": 0.4536516296982265, "grad_norm": 1.1467726230621338, "learning_rate": 0.00019812084479174384, "loss": 4.737, "step": 243100 }, { "epoch": 0.453744935257605, "grad_norm": 1.2688440084457397, "learning_rate": 0.0001981200341228583, "loss": 4.7588, "step": 243150 }, { "epoch": 0.4538382408169835, "grad_norm": 0.977979302406311, "learning_rate": 0.0001981192232808082, "loss": 4.5473, "step": 243200 }, { "epoch": 0.45393154637636196, "grad_norm": 0.8517860174179077, "learning_rate": 0.00019811841226559487, "loss": 4.6273, "step": 243250 }, { "epoch": 0.45402485193574044, "grad_norm": 1.3185665607452393, "learning_rate": 0.00019811760107721983, "loss": 4.6099, "step": 243300 }, { "epoch": 0.45411815749511897, "grad_norm": 1.0361919403076172, "learning_rate": 0.00019811678971568453, "loss": 4.5083, "step": 243350 }, { "epoch": 0.45421146305449744, "grad_norm": 1.0736933946609497, "learning_rate": 0.00019811597818099032, "loss": 4.5282, "step": 243400 }, { "epoch": 0.4543047686138759, "grad_norm": 0.9656780362129211, "learning_rate": 0.00019811516647313867, "loss": 4.7517, "step": 243450 }, { "epoch": 0.45439807417325445, "grad_norm": 1.222460389137268, "learning_rate": 0.00019811435459213104, "loss": 4.5887, "step": 243500 }, { "epoch": 0.4544913797326329, "grad_norm": 1.2716476917266846, "learning_rate": 0.00019811354253796882, "loss": 4.6937, "step": 243550 }, { "epoch": 0.4545846852920114, "grad_norm": 0.9141228199005127, "learning_rate": 0.00019811273031065347, "loss": 4.5894, "step": 243600 }, { "epoch": 0.4546779908513899, "grad_norm": 1.0707182884216309, "learning_rate": 0.00019811191791018642, "loss": 4.6077, "step": 243650 }, { "epoch": 0.4547712964107684, "grad_norm": 1.1825801134109497, "learning_rate": 0.00019811110533656909, "loss": 4.5034, "step": 243700 }, { "epoch": 0.4548646019701469, "grad_norm": 1.3206427097320557, "learning_rate": 0.0001981102925898029, "loss": 4.7394, "step": 243750 }, { "epoch": 0.45495790752952536, "grad_norm": 0.9127716422080994, "learning_rate": 0.00019810947966988934, "loss": 4.5219, "step": 243800 }, { "epoch": 0.4550512130889039, "grad_norm": 0.9963350296020508, "learning_rate": 0.0001981086665768298, "loss": 4.5856, "step": 243850 }, { "epoch": 0.45514451864828237, "grad_norm": 0.8234110474586487, "learning_rate": 0.0001981078533106257, "loss": 4.7194, "step": 243900 }, { "epoch": 0.45523782420766085, "grad_norm": 0.8868279457092285, "learning_rate": 0.00019810703987127855, "loss": 4.5398, "step": 243950 }, { "epoch": 0.4553311297670393, "grad_norm": 0.8374065160751343, "learning_rate": 0.0001981062262587897, "loss": 4.5695, "step": 244000 }, { "epoch": 0.45542443532641785, "grad_norm": 0.884613573551178, "learning_rate": 0.00019810541247316062, "loss": 4.615, "step": 244050 }, { "epoch": 0.45551774088579633, "grad_norm": 0.8596709966659546, "learning_rate": 0.00019810459851439278, "loss": 4.4475, "step": 244100 }, { "epoch": 0.4556110464451748, "grad_norm": 0.7834071516990662, "learning_rate": 0.00019810378438248758, "loss": 4.4734, "step": 244150 }, { "epoch": 0.45570435200455334, "grad_norm": 0.9869714975357056, "learning_rate": 0.00019810297007744645, "loss": 4.6483, "step": 244200 }, { "epoch": 0.4557976575639318, "grad_norm": 1.230589509010315, "learning_rate": 0.00019810215559927085, "loss": 4.5132, "step": 244250 }, { "epoch": 0.4558909631233103, "grad_norm": 1.0174620151519775, "learning_rate": 0.0001981013409479622, "loss": 4.6885, "step": 244300 }, { "epoch": 0.45598426868268876, "grad_norm": 1.0130627155303955, "learning_rate": 0.00019810052612352194, "loss": 4.4784, "step": 244350 }, { "epoch": 0.4560775742420673, "grad_norm": 0.9201186895370483, "learning_rate": 0.0001980997111259515, "loss": 4.5096, "step": 244400 }, { "epoch": 0.45617087980144577, "grad_norm": 1.1127043962478638, "learning_rate": 0.00019809889595525239, "loss": 4.5587, "step": 244450 }, { "epoch": 0.45626418536082425, "grad_norm": 1.2005751132965088, "learning_rate": 0.0001980980806114259, "loss": 4.6419, "step": 244500 }, { "epoch": 0.4563574909202028, "grad_norm": 1.1009554862976074, "learning_rate": 0.00019809726509447363, "loss": 4.5501, "step": 244550 }, { "epoch": 0.45645079647958126, "grad_norm": 1.0324193239212036, "learning_rate": 0.00019809644940439692, "loss": 4.3942, "step": 244600 }, { "epoch": 0.45654410203895973, "grad_norm": 1.0688228607177734, "learning_rate": 0.00019809563354119725, "loss": 4.5763, "step": 244650 }, { "epoch": 0.4566374075983382, "grad_norm": 0.9501363039016724, "learning_rate": 0.00019809481750487603, "loss": 4.3471, "step": 244700 }, { "epoch": 0.45673071315771674, "grad_norm": 1.0971622467041016, "learning_rate": 0.00019809400129543474, "loss": 4.7443, "step": 244750 }, { "epoch": 0.4568240187170952, "grad_norm": 1.5378892421722412, "learning_rate": 0.0001980931849128748, "loss": 4.4834, "step": 244800 }, { "epoch": 0.4569173242764737, "grad_norm": 1.0331060886383057, "learning_rate": 0.0001980923683571976, "loss": 4.5969, "step": 244850 }, { "epoch": 0.4570106298358522, "grad_norm": 0.9876800775527954, "learning_rate": 0.00019809155162840467, "loss": 4.6108, "step": 244900 }, { "epoch": 0.4571039353952307, "grad_norm": 1.113776683807373, "learning_rate": 0.00019809073472649738, "loss": 4.7422, "step": 244950 }, { "epoch": 0.4571972409546092, "grad_norm": 0.8026116490364075, "learning_rate": 0.00019808991765147727, "loss": 4.6976, "step": 245000 }, { "epoch": 0.45729054651398765, "grad_norm": 1.0806735754013062, "learning_rate": 0.00019808910040334564, "loss": 4.5532, "step": 245050 }, { "epoch": 0.4573838520733662, "grad_norm": 1.0394002199172974, "learning_rate": 0.00019808828298210408, "loss": 4.6943, "step": 245100 }, { "epoch": 0.45747715763274466, "grad_norm": 1.022925853729248, "learning_rate": 0.0001980874653877539, "loss": 4.5151, "step": 245150 }, { "epoch": 0.45757046319212313, "grad_norm": 0.7271643280982971, "learning_rate": 0.00019808664762029662, "loss": 4.4889, "step": 245200 }, { "epoch": 0.4576637687515016, "grad_norm": 0.8295145630836487, "learning_rate": 0.00019808582967973366, "loss": 4.5364, "step": 245250 }, { "epoch": 0.45775707431088014, "grad_norm": 1.0905122756958008, "learning_rate": 0.0001980850115660665, "loss": 4.7712, "step": 245300 }, { "epoch": 0.4578503798702586, "grad_norm": 1.2582670450210571, "learning_rate": 0.0001980841932792965, "loss": 4.557, "step": 245350 }, { "epoch": 0.4579436854296371, "grad_norm": 0.8376700282096863, "learning_rate": 0.0001980833748194252, "loss": 4.5676, "step": 245400 }, { "epoch": 0.4580369909890156, "grad_norm": 0.8749433159828186, "learning_rate": 0.00019808255618645397, "loss": 4.729, "step": 245450 }, { "epoch": 0.4581302965483941, "grad_norm": 1.354799747467041, "learning_rate": 0.0001980817373803843, "loss": 4.5384, "step": 245500 }, { "epoch": 0.4582236021077726, "grad_norm": 1.1467299461364746, "learning_rate": 0.00019808091840121765, "loss": 4.6499, "step": 245550 }, { "epoch": 0.45831690766715105, "grad_norm": 1.0075682401657104, "learning_rate": 0.0001980800992489554, "loss": 4.643, "step": 245600 }, { "epoch": 0.4584102132265296, "grad_norm": 1.1275964975357056, "learning_rate": 0.00019807927992359904, "loss": 4.7408, "step": 245650 }, { "epoch": 0.45850351878590806, "grad_norm": 0.9080273509025574, "learning_rate": 0.00019807846042515004, "loss": 4.8073, "step": 245700 }, { "epoch": 0.45859682434528654, "grad_norm": 0.7634953856468201, "learning_rate": 0.00019807764075360977, "loss": 4.6437, "step": 245750 }, { "epoch": 0.45869012990466507, "grad_norm": 0.9849189519882202, "learning_rate": 0.00019807682090897976, "loss": 4.6905, "step": 245800 }, { "epoch": 0.45878343546404354, "grad_norm": 0.7696005702018738, "learning_rate": 0.0001980760008912614, "loss": 4.5023, "step": 245850 }, { "epoch": 0.458876741023422, "grad_norm": 1.1144013404846191, "learning_rate": 0.00019807518070045615, "loss": 4.7656, "step": 245900 }, { "epoch": 0.4589700465828005, "grad_norm": 1.2693634033203125, "learning_rate": 0.00019807436033656545, "loss": 4.572, "step": 245950 }, { "epoch": 0.459063352142179, "grad_norm": 1.1602495908737183, "learning_rate": 0.0001980735397995908, "loss": 4.5726, "step": 246000 }, { "epoch": 0.459063352142179, "eval_loss": 4.774869441986084, "eval_runtime": 233.5371, "eval_samples_per_second": 11.167, "eval_steps_per_second": 11.167, "eval_tts_loss": 7.48305198906898, "step": 246000 }, { "epoch": 0.4591566577015575, "grad_norm": 0.9617483615875244, "learning_rate": 0.0001980727190895336, "loss": 4.4622, "step": 246050 }, { "epoch": 0.459249963260936, "grad_norm": 0.9876244068145752, "learning_rate": 0.00019807189820639527, "loss": 4.7104, "step": 246100 }, { "epoch": 0.4593432688203145, "grad_norm": 1.061923861503601, "learning_rate": 0.00019807107715017733, "loss": 4.6161, "step": 246150 }, { "epoch": 0.459436574379693, "grad_norm": 1.3784229755401611, "learning_rate": 0.00019807025592088117, "loss": 4.5802, "step": 246200 }, { "epoch": 0.45952987993907146, "grad_norm": 1.1820831298828125, "learning_rate": 0.00019806943451850827, "loss": 4.5958, "step": 246250 }, { "epoch": 0.45962318549844994, "grad_norm": 0.9838506579399109, "learning_rate": 0.0001980686129430601, "loss": 4.7938, "step": 246300 }, { "epoch": 0.45971649105782847, "grad_norm": 1.1807385683059692, "learning_rate": 0.00019806779119453807, "loss": 4.5621, "step": 246350 }, { "epoch": 0.45980979661720695, "grad_norm": 1.1486557722091675, "learning_rate": 0.00019806696927294364, "loss": 4.5431, "step": 246400 }, { "epoch": 0.4599031021765854, "grad_norm": 1.1213480234146118, "learning_rate": 0.00019806614717827827, "loss": 4.5494, "step": 246450 }, { "epoch": 0.45999640773596395, "grad_norm": 1.0412524938583374, "learning_rate": 0.00019806532491054338, "loss": 4.6432, "step": 246500 }, { "epoch": 0.46008971329534243, "grad_norm": 1.125463843345642, "learning_rate": 0.00019806450246974048, "loss": 4.6128, "step": 246550 }, { "epoch": 0.4601830188547209, "grad_norm": 1.0406776666641235, "learning_rate": 0.00019806367985587099, "loss": 4.7215, "step": 246600 }, { "epoch": 0.4602763244140994, "grad_norm": 1.2460758686065674, "learning_rate": 0.00019806285706893632, "loss": 4.6226, "step": 246650 }, { "epoch": 0.4603696299734779, "grad_norm": 0.9566603302955627, "learning_rate": 0.000198062034108938, "loss": 4.4905, "step": 246700 }, { "epoch": 0.4604629355328564, "grad_norm": 1.013724684715271, "learning_rate": 0.00019806121097587745, "loss": 4.507, "step": 246750 }, { "epoch": 0.46055624109223486, "grad_norm": 0.9310964941978455, "learning_rate": 0.00019806038766975606, "loss": 4.3787, "step": 246800 }, { "epoch": 0.4606495466516134, "grad_norm": 1.0492833852767944, "learning_rate": 0.0001980595641905754, "loss": 4.5084, "step": 246850 }, { "epoch": 0.46074285221099187, "grad_norm": 0.5750857591629028, "learning_rate": 0.00019805874053833685, "loss": 4.5357, "step": 246900 }, { "epoch": 0.46083615777037035, "grad_norm": 1.0940406322479248, "learning_rate": 0.00019805791671304187, "loss": 4.4754, "step": 246950 }, { "epoch": 0.4609294633297488, "grad_norm": 1.030585527420044, "learning_rate": 0.0001980570927146919, "loss": 4.6358, "step": 247000 }, { "epoch": 0.46102276888912735, "grad_norm": 1.0150128602981567, "learning_rate": 0.00019805626854328844, "loss": 4.7402, "step": 247050 }, { "epoch": 0.46111607444850583, "grad_norm": 2.006843090057373, "learning_rate": 0.00019805544419883292, "loss": 4.5641, "step": 247100 }, { "epoch": 0.4612093800078843, "grad_norm": 1.1281328201293945, "learning_rate": 0.00019805461968132679, "loss": 4.5954, "step": 247150 }, { "epoch": 0.46130268556726284, "grad_norm": 1.0709025859832764, "learning_rate": 0.00019805379499077152, "loss": 4.7013, "step": 247200 }, { "epoch": 0.4613959911266413, "grad_norm": 1.0696786642074585, "learning_rate": 0.00019805297012716856, "loss": 4.6369, "step": 247250 }, { "epoch": 0.4614892966860198, "grad_norm": 1.0185669660568237, "learning_rate": 0.00019805214509051932, "loss": 4.685, "step": 247300 }, { "epoch": 0.46158260224539827, "grad_norm": 0.921750009059906, "learning_rate": 0.00019805131988082534, "loss": 4.6481, "step": 247350 }, { "epoch": 0.4616759078047768, "grad_norm": 0.625173032283783, "learning_rate": 0.000198050494498088, "loss": 4.4107, "step": 247400 }, { "epoch": 0.4617692133641553, "grad_norm": 0.8325273990631104, "learning_rate": 0.0001980496689423088, "loss": 4.7302, "step": 247450 }, { "epoch": 0.46186251892353375, "grad_norm": 1.2000508308410645, "learning_rate": 0.00019804884321348917, "loss": 4.7503, "step": 247500 }, { "epoch": 0.4619558244829123, "grad_norm": 1.06386137008667, "learning_rate": 0.00019804801731163063, "loss": 4.5668, "step": 247550 }, { "epoch": 0.46204913004229076, "grad_norm": 1.1253514289855957, "learning_rate": 0.00019804719123673456, "loss": 4.4186, "step": 247600 }, { "epoch": 0.46214243560166923, "grad_norm": 0.9880508780479431, "learning_rate": 0.00019804636498880248, "loss": 4.6186, "step": 247650 }, { "epoch": 0.4622357411610477, "grad_norm": 1.0324673652648926, "learning_rate": 0.00019804553856783577, "loss": 4.4925, "step": 247700 }, { "epoch": 0.46232904672042624, "grad_norm": 0.7357877492904663, "learning_rate": 0.00019804471197383596, "loss": 4.6977, "step": 247750 }, { "epoch": 0.4624223522798047, "grad_norm": 0.9549800753593445, "learning_rate": 0.0001980438852068045, "loss": 4.4168, "step": 247800 }, { "epoch": 0.4625156578391832, "grad_norm": 0.9433833956718445, "learning_rate": 0.0001980430582667428, "loss": 4.6086, "step": 247850 }, { "epoch": 0.46260896339856167, "grad_norm": 1.2438510656356812, "learning_rate": 0.00019804223115365237, "loss": 4.6693, "step": 247900 }, { "epoch": 0.4627022689579402, "grad_norm": 0.8715135455131531, "learning_rate": 0.00019804140386753467, "loss": 4.6429, "step": 247950 }, { "epoch": 0.4627955745173187, "grad_norm": 0.9428792595863342, "learning_rate": 0.00019804057640839112, "loss": 4.5094, "step": 248000 }, { "epoch": 0.46288888007669715, "grad_norm": 0.8112114071846008, "learning_rate": 0.0001980397487762232, "loss": 4.538, "step": 248050 }, { "epoch": 0.4629821856360757, "grad_norm": 0.899664044380188, "learning_rate": 0.0001980389209710324, "loss": 4.4226, "step": 248100 }, { "epoch": 0.46307549119545416, "grad_norm": 1.0924978256225586, "learning_rate": 0.00019803809299282014, "loss": 4.6605, "step": 248150 }, { "epoch": 0.46316879675483263, "grad_norm": 0.9904999136924744, "learning_rate": 0.0001980372648415879, "loss": 4.5332, "step": 248200 }, { "epoch": 0.4632621023142111, "grad_norm": 1.1240946054458618, "learning_rate": 0.00019803643651733713, "loss": 4.6454, "step": 248250 }, { "epoch": 0.46335540787358964, "grad_norm": 1.0554358959197998, "learning_rate": 0.0001980356080200693, "loss": 4.757, "step": 248300 }, { "epoch": 0.4634487134329681, "grad_norm": 1.0673205852508545, "learning_rate": 0.00019803477934978586, "loss": 4.3607, "step": 248350 }, { "epoch": 0.4635420189923466, "grad_norm": 1.0549492835998535, "learning_rate": 0.00019803395050648832, "loss": 4.4935, "step": 248400 }, { "epoch": 0.4636353245517251, "grad_norm": 1.18800687789917, "learning_rate": 0.00019803312149017807, "loss": 4.7759, "step": 248450 }, { "epoch": 0.4637286301111036, "grad_norm": 1.0293747186660767, "learning_rate": 0.00019803229230085662, "loss": 4.5521, "step": 248500 }, { "epoch": 0.4638219356704821, "grad_norm": 1.0756644010543823, "learning_rate": 0.00019803146293852545, "loss": 4.5545, "step": 248550 }, { "epoch": 0.46391524122986055, "grad_norm": 0.8822924494743347, "learning_rate": 0.00019803063340318597, "loss": 4.4807, "step": 248600 }, { "epoch": 0.4640085467892391, "grad_norm": 1.109789490699768, "learning_rate": 0.00019802980369483967, "loss": 4.5922, "step": 248650 }, { "epoch": 0.46410185234861756, "grad_norm": 1.0648267269134521, "learning_rate": 0.00019802897381348804, "loss": 4.5294, "step": 248700 }, { "epoch": 0.46419515790799604, "grad_norm": 0.9832372069358826, "learning_rate": 0.00019802814375913247, "loss": 4.6105, "step": 248750 }, { "epoch": 0.46428846346737457, "grad_norm": 1.0333669185638428, "learning_rate": 0.00019802731353177452, "loss": 4.6382, "step": 248800 }, { "epoch": 0.46438176902675304, "grad_norm": 1.1665089130401611, "learning_rate": 0.00019802648313141562, "loss": 4.5055, "step": 248850 }, { "epoch": 0.4644750745861315, "grad_norm": 0.7576875686645508, "learning_rate": 0.00019802565255805718, "loss": 4.555, "step": 248900 }, { "epoch": 0.46456838014551, "grad_norm": 0.8509981036186218, "learning_rate": 0.00019802482181170074, "loss": 4.6874, "step": 248950 }, { "epoch": 0.4646616857048885, "grad_norm": 1.0222673416137695, "learning_rate": 0.00019802399089234775, "loss": 4.7886, "step": 249000 }, { "epoch": 0.4646616857048885, "eval_loss": 4.778683662414551, "eval_runtime": 230.8207, "eval_samples_per_second": 11.299, "eval_steps_per_second": 11.299, "eval_tts_loss": 7.546051602269254, "step": 249000 }, { "epoch": 0.464754991264267, "grad_norm": 0.9813963770866394, "learning_rate": 0.00019802315979999965, "loss": 4.4973, "step": 249050 }, { "epoch": 0.4648482968236455, "grad_norm": 1.1679600477218628, "learning_rate": 0.0001980223285346579, "loss": 4.716, "step": 249100 }, { "epoch": 0.464941602383024, "grad_norm": 1.014891505241394, "learning_rate": 0.000198021497096324, "loss": 4.641, "step": 249150 }, { "epoch": 0.4650349079424025, "grad_norm": 1.1043691635131836, "learning_rate": 0.00019802066548499941, "loss": 4.712, "step": 249200 }, { "epoch": 0.46512821350178096, "grad_norm": 1.0716991424560547, "learning_rate": 0.0001980198337006856, "loss": 4.6739, "step": 249250 }, { "epoch": 0.46522151906115944, "grad_norm": 1.2139348983764648, "learning_rate": 0.00019801900174338402, "loss": 4.6826, "step": 249300 }, { "epoch": 0.46531482462053797, "grad_norm": 0.8604678511619568, "learning_rate": 0.00019801816961309617, "loss": 4.6827, "step": 249350 }, { "epoch": 0.46540813017991645, "grad_norm": 1.0838696956634521, "learning_rate": 0.00019801733730982348, "loss": 4.6474, "step": 249400 }, { "epoch": 0.4655014357392949, "grad_norm": 1.5749475955963135, "learning_rate": 0.00019801650483356742, "loss": 4.4896, "step": 249450 }, { "epoch": 0.46559474129867345, "grad_norm": 0.7585879564285278, "learning_rate": 0.0001980156721843295, "loss": 4.3649, "step": 249500 }, { "epoch": 0.46568804685805193, "grad_norm": 1.1551792621612549, "learning_rate": 0.00019801483936211117, "loss": 4.6776, "step": 249550 }, { "epoch": 0.4657813524174304, "grad_norm": 0.8028654456138611, "learning_rate": 0.00019801400636691385, "loss": 4.4243, "step": 249600 }, { "epoch": 0.4658746579768089, "grad_norm": 1.0627497434616089, "learning_rate": 0.0001980131731987391, "loss": 4.3195, "step": 249650 }, { "epoch": 0.4659679635361874, "grad_norm": 1.1818534135818481, "learning_rate": 0.00019801233985758835, "loss": 4.5887, "step": 249700 }, { "epoch": 0.4660612690955659, "grad_norm": 1.0179072618484497, "learning_rate": 0.000198011506343463, "loss": 4.5652, "step": 249750 }, { "epoch": 0.46615457465494436, "grad_norm": 1.0190846920013428, "learning_rate": 0.00019801067265636465, "loss": 4.7446, "step": 249800 }, { "epoch": 0.4662478802143229, "grad_norm": 0.9870844483375549, "learning_rate": 0.0001980098387962947, "loss": 4.6793, "step": 249850 }, { "epoch": 0.46634118577370137, "grad_norm": 0.8866018652915955, "learning_rate": 0.00019800900476325462, "loss": 4.7872, "step": 249900 }, { "epoch": 0.46643449133307985, "grad_norm": 1.2715367078781128, "learning_rate": 0.0001980081705572459, "loss": 4.7906, "step": 249950 }, { "epoch": 0.4665277968924583, "grad_norm": 1.1326353549957275, "learning_rate": 0.00019800733617827, "loss": 4.7138, "step": 250000 }, { "epoch": 0.46662110245183686, "grad_norm": 1.0294839143753052, "learning_rate": 0.0001980065016263284, "loss": 4.5253, "step": 250050 }, { "epoch": 0.46671440801121533, "grad_norm": 1.1062980890274048, "learning_rate": 0.00019800566690142258, "loss": 4.6491, "step": 250100 }, { "epoch": 0.4668077135705938, "grad_norm": 0.8462019562721252, "learning_rate": 0.00019800483200355394, "loss": 4.6245, "step": 250150 }, { "epoch": 0.46690101912997234, "grad_norm": 0.8871804475784302, "learning_rate": 0.0001980039969327241, "loss": 4.7166, "step": 250200 }, { "epoch": 0.4669943246893508, "grad_norm": 1.0749971866607666, "learning_rate": 0.0001980031616889344, "loss": 4.4303, "step": 250250 }, { "epoch": 0.4670876302487293, "grad_norm": 1.1696592569351196, "learning_rate": 0.00019800232627218637, "loss": 4.6637, "step": 250300 }, { "epoch": 0.46718093580810777, "grad_norm": 1.017404317855835, "learning_rate": 0.0001980014906824815, "loss": 4.6735, "step": 250350 }, { "epoch": 0.4672742413674863, "grad_norm": 0.9707728624343872, "learning_rate": 0.00019800065491982122, "loss": 4.5803, "step": 250400 }, { "epoch": 0.4673675469268648, "grad_norm": 1.125417709350586, "learning_rate": 0.00019799981898420703, "loss": 4.6582, "step": 250450 }, { "epoch": 0.46746085248624325, "grad_norm": 0.9616281986236572, "learning_rate": 0.00019799898287564042, "loss": 4.545, "step": 250500 }, { "epoch": 0.4675541580456217, "grad_norm": 0.8503236770629883, "learning_rate": 0.00019799814659412284, "loss": 4.6893, "step": 250550 }, { "epoch": 0.46764746360500026, "grad_norm": 1.1284072399139404, "learning_rate": 0.00019799731013965576, "loss": 4.6788, "step": 250600 }, { "epoch": 0.46774076916437873, "grad_norm": 0.9626160264015198, "learning_rate": 0.00019799647351224068, "loss": 4.577, "step": 250650 }, { "epoch": 0.4678340747237572, "grad_norm": 0.9178403615951538, "learning_rate": 0.00019799563671187907, "loss": 4.3979, "step": 250700 }, { "epoch": 0.46792738028313574, "grad_norm": 0.9638570547103882, "learning_rate": 0.0001979947997385724, "loss": 4.4919, "step": 250750 }, { "epoch": 0.4680206858425142, "grad_norm": 0.9107949733734131, "learning_rate": 0.00019799396259232216, "loss": 4.5359, "step": 250800 }, { "epoch": 0.4681139914018927, "grad_norm": 1.2097649574279785, "learning_rate": 0.00019799312527312983, "loss": 4.5872, "step": 250850 }, { "epoch": 0.46820729696127117, "grad_norm": 0.9304313063621521, "learning_rate": 0.00019799228778099687, "loss": 4.6143, "step": 250900 }, { "epoch": 0.4683006025206497, "grad_norm": 0.9003384113311768, "learning_rate": 0.00019799145011592474, "loss": 4.6902, "step": 250950 }, { "epoch": 0.4683939080800282, "grad_norm": 0.825849711894989, "learning_rate": 0.00019799061227791494, "loss": 4.5176, "step": 251000 }, { "epoch": 0.46848721363940665, "grad_norm": 0.8515427708625793, "learning_rate": 0.00019798977426696896, "loss": 4.5787, "step": 251050 }, { "epoch": 0.4685805191987852, "grad_norm": 0.9630306959152222, "learning_rate": 0.0001979889360830883, "loss": 4.4875, "step": 251100 }, { "epoch": 0.46867382475816366, "grad_norm": 0.694108784198761, "learning_rate": 0.00019798809772627438, "loss": 4.2851, "step": 251150 }, { "epoch": 0.46876713031754214, "grad_norm": 0.8134151697158813, "learning_rate": 0.00019798725919652872, "loss": 4.6372, "step": 251200 }, { "epoch": 0.4688604358769206, "grad_norm": 0.8471952080726624, "learning_rate": 0.00019798642049385278, "loss": 4.3836, "step": 251250 }, { "epoch": 0.46895374143629914, "grad_norm": 1.0113226175308228, "learning_rate": 0.00019798558161824805, "loss": 4.3661, "step": 251300 }, { "epoch": 0.4690470469956776, "grad_norm": 0.840195894241333, "learning_rate": 0.000197984742569716, "loss": 4.5986, "step": 251350 }, { "epoch": 0.4691403525550561, "grad_norm": 0.9539893865585327, "learning_rate": 0.00019798390334825817, "loss": 4.6197, "step": 251400 }, { "epoch": 0.4692336581144346, "grad_norm": 0.8172125816345215, "learning_rate": 0.00019798306395387596, "loss": 4.6553, "step": 251450 }, { "epoch": 0.4693269636738131, "grad_norm": 1.068747878074646, "learning_rate": 0.00019798222438657085, "loss": 4.5775, "step": 251500 }, { "epoch": 0.4694202692331916, "grad_norm": 1.1480756998062134, "learning_rate": 0.00019798138464634439, "loss": 4.4341, "step": 251550 }, { "epoch": 0.46951357479257005, "grad_norm": 1.1255130767822266, "learning_rate": 0.00019798054473319803, "loss": 4.4635, "step": 251600 }, { "epoch": 0.4696068803519486, "grad_norm": 0.9743191599845886, "learning_rate": 0.00019797970464713321, "loss": 4.8709, "step": 251650 }, { "epoch": 0.46970018591132706, "grad_norm": 1.054312825202942, "learning_rate": 0.0001979788643881515, "loss": 4.7223, "step": 251700 }, { "epoch": 0.46979349147070554, "grad_norm": 0.6565726399421692, "learning_rate": 0.00019797802395625432, "loss": 4.6015, "step": 251750 }, { "epoch": 0.46988679703008407, "grad_norm": 1.1399985551834106, "learning_rate": 0.00019797718335144316, "loss": 4.7225, "step": 251800 }, { "epoch": 0.46998010258946255, "grad_norm": 0.9679462909698486, "learning_rate": 0.00019797634257371952, "loss": 4.6508, "step": 251850 }, { "epoch": 0.470073408148841, "grad_norm": 1.1389963626861572, "learning_rate": 0.00019797550162308487, "loss": 4.5277, "step": 251900 }, { "epoch": 0.4701667137082195, "grad_norm": 0.8405935764312744, "learning_rate": 0.00019797466049954068, "loss": 4.4807, "step": 251950 }, { "epoch": 0.47026001926759803, "grad_norm": 0.9860919713973999, "learning_rate": 0.00019797381920308847, "loss": 4.5945, "step": 252000 }, { "epoch": 0.47026001926759803, "eval_loss": 4.771847724914551, "eval_runtime": 228.5901, "eval_samples_per_second": 11.409, "eval_steps_per_second": 11.409, "eval_tts_loss": 7.527745360290014, "step": 252000 }, { "epoch": 0.4703533248269765, "grad_norm": 0.8719726204872131, "learning_rate": 0.0001979729777337297, "loss": 4.5289, "step": 252050 }, { "epoch": 0.470446630386355, "grad_norm": 1.0118962526321411, "learning_rate": 0.00019797213609146588, "loss": 4.6344, "step": 252100 }, { "epoch": 0.4705399359457335, "grad_norm": 1.014058232307434, "learning_rate": 0.00019797129427629847, "loss": 4.6503, "step": 252150 }, { "epoch": 0.470633241505112, "grad_norm": 0.9134781360626221, "learning_rate": 0.00019797045228822898, "loss": 4.538, "step": 252200 }, { "epoch": 0.47072654706449046, "grad_norm": 1.0648183822631836, "learning_rate": 0.00019796961012725884, "loss": 4.5453, "step": 252250 }, { "epoch": 0.47081985262386894, "grad_norm": 1.1719633340835571, "learning_rate": 0.0001979687677933896, "loss": 4.7073, "step": 252300 }, { "epoch": 0.47091315818324747, "grad_norm": 1.206669569015503, "learning_rate": 0.00019796792528662272, "loss": 4.708, "step": 252350 }, { "epoch": 0.47100646374262595, "grad_norm": 1.1847809553146362, "learning_rate": 0.00019796708260695972, "loss": 4.5306, "step": 252400 }, { "epoch": 0.4710997693020044, "grad_norm": 0.8223580718040466, "learning_rate": 0.00019796623975440204, "loss": 4.5695, "step": 252450 }, { "epoch": 0.47119307486138295, "grad_norm": 0.9964529871940613, "learning_rate": 0.00019796539672895116, "loss": 4.6491, "step": 252500 }, { "epoch": 0.47128638042076143, "grad_norm": 0.9831516146659851, "learning_rate": 0.0001979645535306086, "loss": 4.6033, "step": 252550 }, { "epoch": 0.4713796859801399, "grad_norm": 0.7860525250434875, "learning_rate": 0.00019796371015937586, "loss": 4.4849, "step": 252600 }, { "epoch": 0.4714729915395184, "grad_norm": 1.078437328338623, "learning_rate": 0.00019796286661525437, "loss": 4.7393, "step": 252650 }, { "epoch": 0.4715662970988969, "grad_norm": 0.8938062191009521, "learning_rate": 0.0001979620228982457, "loss": 4.4925, "step": 252700 }, { "epoch": 0.4716596026582754, "grad_norm": 1.3449671268463135, "learning_rate": 0.00019796117900835127, "loss": 4.5993, "step": 252750 }, { "epoch": 0.47175290821765387, "grad_norm": 1.0457773208618164, "learning_rate": 0.00019796033494557262, "loss": 4.507, "step": 252800 }, { "epoch": 0.4718462137770324, "grad_norm": 1.118150234222412, "learning_rate": 0.0001979594907099112, "loss": 4.5657, "step": 252850 }, { "epoch": 0.4719395193364109, "grad_norm": 1.0260703563690186, "learning_rate": 0.00019795864630136852, "loss": 4.5898, "step": 252900 }, { "epoch": 0.47203282489578935, "grad_norm": 0.9269238114356995, "learning_rate": 0.00019795780171994606, "loss": 4.808, "step": 252950 }, { "epoch": 0.4721261304551678, "grad_norm": 1.0303397178649902, "learning_rate": 0.00019795695696564534, "loss": 4.4025, "step": 253000 }, { "epoch": 0.47221943601454636, "grad_norm": 0.9815906882286072, "learning_rate": 0.00019795611203846778, "loss": 4.6513, "step": 253050 }, { "epoch": 0.47231274157392483, "grad_norm": 1.0139660835266113, "learning_rate": 0.00019795526693841496, "loss": 4.5168, "step": 253100 }, { "epoch": 0.4724060471333033, "grad_norm": 1.0596623420715332, "learning_rate": 0.0001979544216654883, "loss": 4.6385, "step": 253150 }, { "epoch": 0.4724993526926818, "grad_norm": 1.116756796836853, "learning_rate": 0.00019795357621968935, "loss": 4.6691, "step": 253200 }, { "epoch": 0.4725926582520603, "grad_norm": 0.955316960811615, "learning_rate": 0.00019795273060101956, "loss": 4.6902, "step": 253250 }, { "epoch": 0.4726859638114388, "grad_norm": 0.9806995987892151, "learning_rate": 0.00019795188480948043, "loss": 4.6256, "step": 253300 }, { "epoch": 0.47277926937081727, "grad_norm": 1.410809874534607, "learning_rate": 0.00019795103884507345, "loss": 4.5668, "step": 253350 }, { "epoch": 0.4728725749301958, "grad_norm": 0.9873063564300537, "learning_rate": 0.00019795019270780014, "loss": 4.5323, "step": 253400 }, { "epoch": 0.4729658804895743, "grad_norm": 0.8761279582977295, "learning_rate": 0.00019794934639766196, "loss": 4.5232, "step": 253450 }, { "epoch": 0.47305918604895275, "grad_norm": 0.7007517218589783, "learning_rate": 0.00019794849991466046, "loss": 4.4708, "step": 253500 }, { "epoch": 0.4731524916083312, "grad_norm": 0.8728896379470825, "learning_rate": 0.00019794765325879704, "loss": 4.5846, "step": 253550 }, { "epoch": 0.47324579716770976, "grad_norm": 1.0133166313171387, "learning_rate": 0.00019794680643007325, "loss": 4.5587, "step": 253600 }, { "epoch": 0.47333910272708823, "grad_norm": 1.3427821397781372, "learning_rate": 0.0001979459594284906, "loss": 4.663, "step": 253650 }, { "epoch": 0.4734324082864667, "grad_norm": 1.2627538442611694, "learning_rate": 0.00019794511225405058, "loss": 4.6541, "step": 253700 }, { "epoch": 0.47352571384584524, "grad_norm": 1.0716121196746826, "learning_rate": 0.00019794426490675463, "loss": 4.5518, "step": 253750 }, { "epoch": 0.4736190194052237, "grad_norm": 1.0432318449020386, "learning_rate": 0.00019794341738660432, "loss": 4.5539, "step": 253800 }, { "epoch": 0.4737123249646022, "grad_norm": 1.083561658859253, "learning_rate": 0.00019794256969360113, "loss": 4.7249, "step": 253850 }, { "epoch": 0.47380563052398067, "grad_norm": 1.0139247179031372, "learning_rate": 0.00019794172182774648, "loss": 4.5797, "step": 253900 }, { "epoch": 0.4738989360833592, "grad_norm": 1.054151177406311, "learning_rate": 0.00019794087378904195, "loss": 4.6604, "step": 253950 }, { "epoch": 0.4739922416427377, "grad_norm": 1.1216671466827393, "learning_rate": 0.00019794002557748904, "loss": 4.6178, "step": 254000 }, { "epoch": 0.47408554720211615, "grad_norm": 1.1557576656341553, "learning_rate": 0.0001979391771930892, "loss": 4.5867, "step": 254050 }, { "epoch": 0.4741788527614947, "grad_norm": 1.1198077201843262, "learning_rate": 0.00019793832863584392, "loss": 4.6404, "step": 254100 }, { "epoch": 0.47427215832087316, "grad_norm": 0.9180806279182434, "learning_rate": 0.00019793747990575473, "loss": 4.4542, "step": 254150 }, { "epoch": 0.47436546388025164, "grad_norm": 0.9826111793518066, "learning_rate": 0.00019793663100282314, "loss": 4.6157, "step": 254200 }, { "epoch": 0.4744587694396301, "grad_norm": 1.0048807859420776, "learning_rate": 0.00019793578192705064, "loss": 4.4314, "step": 254250 }, { "epoch": 0.47455207499900864, "grad_norm": 1.0159192085266113, "learning_rate": 0.0001979349326784387, "loss": 4.6976, "step": 254300 }, { "epoch": 0.4746453805583871, "grad_norm": 1.0906916856765747, "learning_rate": 0.00019793408325698882, "loss": 4.7204, "step": 254350 }, { "epoch": 0.4747386861177656, "grad_norm": 1.0381954908370972, "learning_rate": 0.00019793323366270254, "loss": 4.4653, "step": 254400 }, { "epoch": 0.4748319916771441, "grad_norm": 1.0073879957199097, "learning_rate": 0.0001979323838955813, "loss": 4.5649, "step": 254450 }, { "epoch": 0.4749252972365226, "grad_norm": 1.224362850189209, "learning_rate": 0.00019793153395562663, "loss": 4.7343, "step": 254500 }, { "epoch": 0.4750186027959011, "grad_norm": 0.9153668880462646, "learning_rate": 0.00019793068384284007, "loss": 4.6776, "step": 254550 }, { "epoch": 0.47511190835527956, "grad_norm": 1.0687617063522339, "learning_rate": 0.0001979298335572231, "loss": 4.4508, "step": 254600 }, { "epoch": 0.4752052139146581, "grad_norm": 1.1399562358856201, "learning_rate": 0.00019792898309877716, "loss": 4.5694, "step": 254650 }, { "epoch": 0.47529851947403656, "grad_norm": 1.0620476007461548, "learning_rate": 0.00019792813246750384, "loss": 4.4951, "step": 254700 }, { "epoch": 0.47539182503341504, "grad_norm": 0.987824559211731, "learning_rate": 0.00019792728166340457, "loss": 4.5894, "step": 254750 }, { "epoch": 0.47548513059279357, "grad_norm": 0.9981347322463989, "learning_rate": 0.00019792643068648088, "loss": 4.5573, "step": 254800 }, { "epoch": 0.47557843615217205, "grad_norm": 1.3554757833480835, "learning_rate": 0.00019792557953673428, "loss": 4.4722, "step": 254850 }, { "epoch": 0.4756717417115505, "grad_norm": 1.2664356231689453, "learning_rate": 0.00019792472821416625, "loss": 4.467, "step": 254900 }, { "epoch": 0.475765047270929, "grad_norm": 1.125602126121521, "learning_rate": 0.00019792387671877833, "loss": 4.5172, "step": 254950 }, { "epoch": 0.47585835283030753, "grad_norm": 0.9335258603096008, "learning_rate": 0.00019792302505057196, "loss": 4.5266, "step": 255000 }, { "epoch": 0.47585835283030753, "eval_loss": 4.763278484344482, "eval_runtime": 230.157, "eval_samples_per_second": 11.331, "eval_steps_per_second": 11.331, "eval_tts_loss": 7.542246853862746, "step": 255000 }, { "epoch": 0.475951658389686, "grad_norm": 1.0923206806182861, "learning_rate": 0.00019792217320954868, "loss": 4.7171, "step": 255050 }, { "epoch": 0.4760449639490645, "grad_norm": 1.037209391593933, "learning_rate": 0.00019792132119571002, "loss": 4.4465, "step": 255100 }, { "epoch": 0.476138269508443, "grad_norm": 1.2115193605422974, "learning_rate": 0.00019792046900905748, "loss": 4.4359, "step": 255150 }, { "epoch": 0.4762315750678215, "grad_norm": 1.0448733568191528, "learning_rate": 0.0001979196166495925, "loss": 4.5225, "step": 255200 }, { "epoch": 0.47632488062719996, "grad_norm": 1.1041306257247925, "learning_rate": 0.00019791876411731665, "loss": 4.7055, "step": 255250 }, { "epoch": 0.47641818618657844, "grad_norm": 0.9525231122970581, "learning_rate": 0.0001979179114122314, "loss": 4.6905, "step": 255300 }, { "epoch": 0.47651149174595697, "grad_norm": 1.0401257276535034, "learning_rate": 0.00019791705853433827, "loss": 4.5725, "step": 255350 }, { "epoch": 0.47660479730533545, "grad_norm": 0.9512485265731812, "learning_rate": 0.00019791620548363875, "loss": 4.6771, "step": 255400 }, { "epoch": 0.4766981028647139, "grad_norm": 0.9402346611022949, "learning_rate": 0.00019791535226013437, "loss": 4.6207, "step": 255450 }, { "epoch": 0.47679140842409246, "grad_norm": 1.0512655973434448, "learning_rate": 0.0001979144988638266, "loss": 4.5787, "step": 255500 }, { "epoch": 0.47688471398347093, "grad_norm": 0.9695249795913696, "learning_rate": 0.000197913645294717, "loss": 4.5116, "step": 255550 }, { "epoch": 0.4769780195428494, "grad_norm": 0.9149754643440247, "learning_rate": 0.000197912791552807, "loss": 4.6068, "step": 255600 }, { "epoch": 0.4770713251022279, "grad_norm": 1.0737544298171997, "learning_rate": 0.00019791193763809818, "loss": 4.6188, "step": 255650 }, { "epoch": 0.4771646306616064, "grad_norm": 0.9186537861824036, "learning_rate": 0.00019791108355059203, "loss": 4.446, "step": 255700 }, { "epoch": 0.4772579362209849, "grad_norm": 1.1221696138381958, "learning_rate": 0.00019791022929029005, "loss": 4.5936, "step": 255750 }, { "epoch": 0.47735124178036337, "grad_norm": 1.1481096744537354, "learning_rate": 0.00019790937485719373, "loss": 4.5989, "step": 255800 }, { "epoch": 0.47744454733974184, "grad_norm": 0.8442092537879944, "learning_rate": 0.00019790852025130455, "loss": 4.5942, "step": 255850 }, { "epoch": 0.4775378528991204, "grad_norm": 0.8604154586791992, "learning_rate": 0.00019790766547262412, "loss": 4.5971, "step": 255900 }, { "epoch": 0.47763115845849885, "grad_norm": 0.8299677968025208, "learning_rate": 0.00019790681052115385, "loss": 4.4849, "step": 255950 }, { "epoch": 0.4777244640178773, "grad_norm": 0.9092508554458618, "learning_rate": 0.0001979059553968953, "loss": 4.5899, "step": 256000 }, { "epoch": 0.47781776957725586, "grad_norm": 0.8719881176948547, "learning_rate": 0.00019790510009984997, "loss": 4.6535, "step": 256050 }, { "epoch": 0.47791107513663433, "grad_norm": 1.0991895198822021, "learning_rate": 0.00019790424463001936, "loss": 4.6057, "step": 256100 }, { "epoch": 0.4780043806960128, "grad_norm": 0.9137606620788574, "learning_rate": 0.00019790338898740498, "loss": 4.6306, "step": 256150 }, { "epoch": 0.4780976862553913, "grad_norm": 0.8069062829017639, "learning_rate": 0.00019790253317200835, "loss": 4.6064, "step": 256200 }, { "epoch": 0.4781909918147698, "grad_norm": 1.0056474208831787, "learning_rate": 0.00019790167718383098, "loss": 4.4567, "step": 256250 }, { "epoch": 0.4782842973741483, "grad_norm": 1.1140251159667969, "learning_rate": 0.00019790082102287437, "loss": 4.6005, "step": 256300 }, { "epoch": 0.47837760293352677, "grad_norm": 1.049212098121643, "learning_rate": 0.00019789996468914004, "loss": 4.5948, "step": 256350 }, { "epoch": 0.4784709084929053, "grad_norm": 1.2043838500976562, "learning_rate": 0.00019789910818262952, "loss": 4.7654, "step": 256400 }, { "epoch": 0.4785642140522838, "grad_norm": 1.0834003686904907, "learning_rate": 0.00019789825150334426, "loss": 4.4575, "step": 256450 }, { "epoch": 0.47865751961166225, "grad_norm": 1.3048619031906128, "learning_rate": 0.00019789739465128582, "loss": 4.628, "step": 256500 }, { "epoch": 0.47875082517104073, "grad_norm": 0.926943302154541, "learning_rate": 0.00019789653762645576, "loss": 4.6248, "step": 256550 }, { "epoch": 0.47884413073041926, "grad_norm": 1.0053600072860718, "learning_rate": 0.00019789568042885547, "loss": 4.4205, "step": 256600 }, { "epoch": 0.47893743628979774, "grad_norm": 0.9861886501312256, "learning_rate": 0.00019789482305848655, "loss": 4.7163, "step": 256650 }, { "epoch": 0.4790307418491762, "grad_norm": 1.014960527420044, "learning_rate": 0.0001978939655153505, "loss": 4.679, "step": 256700 }, { "epoch": 0.47912404740855474, "grad_norm": 0.6750307679176331, "learning_rate": 0.0001978931077994488, "loss": 4.5416, "step": 256750 }, { "epoch": 0.4792173529679332, "grad_norm": 0.891221284866333, "learning_rate": 0.000197892249910783, "loss": 4.2551, "step": 256800 }, { "epoch": 0.4793106585273117, "grad_norm": 0.7130653262138367, "learning_rate": 0.00019789139184935462, "loss": 4.6392, "step": 256850 }, { "epoch": 0.47940396408669017, "grad_norm": 0.7216616272926331, "learning_rate": 0.00019789053361516517, "loss": 4.5822, "step": 256900 }, { "epoch": 0.4794972696460687, "grad_norm": 1.1904911994934082, "learning_rate": 0.00019788967520821614, "loss": 4.7572, "step": 256950 }, { "epoch": 0.4795905752054472, "grad_norm": 0.8761834502220154, "learning_rate": 0.00019788881662850902, "loss": 4.6696, "step": 257000 }, { "epoch": 0.47968388076482565, "grad_norm": 0.9479681849479675, "learning_rate": 0.00019788795787604538, "loss": 4.6753, "step": 257050 }, { "epoch": 0.4797771863242042, "grad_norm": 1.1873890161514282, "learning_rate": 0.00019788709895082675, "loss": 4.4175, "step": 257100 }, { "epoch": 0.47987049188358266, "grad_norm": 1.023085594177246, "learning_rate": 0.0001978862398528546, "loss": 4.5694, "step": 257150 }, { "epoch": 0.47996379744296114, "grad_norm": 0.9726206064224243, "learning_rate": 0.00019788538058213045, "loss": 4.5753, "step": 257200 }, { "epoch": 0.4800571030023396, "grad_norm": 0.7751583456993103, "learning_rate": 0.00019788452113865583, "loss": 4.4491, "step": 257250 }, { "epoch": 0.48015040856171814, "grad_norm": 1.2681344747543335, "learning_rate": 0.00019788366152243225, "loss": 4.5964, "step": 257300 }, { "epoch": 0.4802437141210966, "grad_norm": 1.1450700759887695, "learning_rate": 0.00019788280173346125, "loss": 4.6184, "step": 257350 }, { "epoch": 0.4803370196804751, "grad_norm": 0.9162212610244751, "learning_rate": 0.00019788194177174431, "loss": 4.4201, "step": 257400 }, { "epoch": 0.48043032523985363, "grad_norm": 1.0206190347671509, "learning_rate": 0.00019788108163728296, "loss": 4.6341, "step": 257450 }, { "epoch": 0.4805236307992321, "grad_norm": 0.9956310391426086, "learning_rate": 0.00019788022133007872, "loss": 4.5461, "step": 257500 }, { "epoch": 0.4806169363586106, "grad_norm": 0.968657374382019, "learning_rate": 0.00019787936085013315, "loss": 4.3365, "step": 257550 }, { "epoch": 0.48071024191798906, "grad_norm": 0.9663876891136169, "learning_rate": 0.0001978785001974477, "loss": 4.6494, "step": 257600 }, { "epoch": 0.4808035474773676, "grad_norm": 0.8891162276268005, "learning_rate": 0.00019787763937202392, "loss": 4.6504, "step": 257650 }, { "epoch": 0.48089685303674606, "grad_norm": 1.1420527696609497, "learning_rate": 0.00019787677837386334, "loss": 4.5885, "step": 257700 }, { "epoch": 0.48099015859612454, "grad_norm": 1.198561429977417, "learning_rate": 0.00019787591720296747, "loss": 4.6936, "step": 257750 }, { "epoch": 0.48108346415550307, "grad_norm": 0.9371161460876465, "learning_rate": 0.0001978750558593378, "loss": 4.5657, "step": 257800 }, { "epoch": 0.48117676971488155, "grad_norm": 0.8604893684387207, "learning_rate": 0.0001978741943429759, "loss": 4.4049, "step": 257850 }, { "epoch": 0.48127007527426, "grad_norm": 0.8791077733039856, "learning_rate": 0.00019787333265388325, "loss": 4.3852, "step": 257900 }, { "epoch": 0.4813633808336385, "grad_norm": 1.054795265197754, "learning_rate": 0.00019787247079206143, "loss": 4.6211, "step": 257950 }, { "epoch": 0.48145668639301703, "grad_norm": 1.1401352882385254, "learning_rate": 0.0001978716087575119, "loss": 4.6098, "step": 258000 }, { "epoch": 0.48145668639301703, "eval_loss": 4.777442932128906, "eval_runtime": 230.5421, "eval_samples_per_second": 11.312, "eval_steps_per_second": 11.312, "eval_tts_loss": 7.488510146787093, "step": 258000 }, { "epoch": 0.4815499919523955, "grad_norm": 1.9049067497253418, "learning_rate": 0.00019787074655023618, "loss": 4.5569, "step": 258050 }, { "epoch": 0.481643297511774, "grad_norm": 1.1054402589797974, "learning_rate": 0.00019786988417023585, "loss": 4.6882, "step": 258100 }, { "epoch": 0.4817366030711525, "grad_norm": 0.8335074186325073, "learning_rate": 0.00019786902161751234, "loss": 4.6304, "step": 258150 }, { "epoch": 0.481829908630531, "grad_norm": 0.9190850257873535, "learning_rate": 0.00019786815889206728, "loss": 4.5275, "step": 258200 }, { "epoch": 0.48192321418990947, "grad_norm": 0.9075625538825989, "learning_rate": 0.0001978672959939021, "loss": 4.5176, "step": 258250 }, { "epoch": 0.48201651974928794, "grad_norm": 0.9331411719322205, "learning_rate": 0.00019786643292301842, "loss": 4.5701, "step": 258300 }, { "epoch": 0.4821098253086665, "grad_norm": 0.732326865196228, "learning_rate": 0.00019786556967941768, "loss": 4.6525, "step": 258350 }, { "epoch": 0.48220313086804495, "grad_norm": 1.284225583076477, "learning_rate": 0.0001978647062631014, "loss": 4.8359, "step": 258400 }, { "epoch": 0.4822964364274234, "grad_norm": 1.2217979431152344, "learning_rate": 0.00019786384267407117, "loss": 4.6221, "step": 258450 }, { "epoch": 0.4823897419868019, "grad_norm": 0.6436966061592102, "learning_rate": 0.00019786297891232846, "loss": 4.5166, "step": 258500 }, { "epoch": 0.48248304754618043, "grad_norm": 0.9446617364883423, "learning_rate": 0.0001978621149778748, "loss": 4.689, "step": 258550 }, { "epoch": 0.4825763531055589, "grad_norm": 0.9324028491973877, "learning_rate": 0.00019786125087071173, "loss": 4.5694, "step": 258600 }, { "epoch": 0.4826696586649374, "grad_norm": 0.9746628999710083, "learning_rate": 0.00019786038659084078, "loss": 4.6517, "step": 258650 }, { "epoch": 0.4827629642243159, "grad_norm": 0.9371702671051025, "learning_rate": 0.0001978595221382635, "loss": 4.6591, "step": 258700 }, { "epoch": 0.4828562697836944, "grad_norm": 0.8363770246505737, "learning_rate": 0.00019785865751298135, "loss": 4.4523, "step": 258750 }, { "epoch": 0.48294957534307287, "grad_norm": 0.8663115501403809, "learning_rate": 0.00019785779271499587, "loss": 4.5986, "step": 258800 }, { "epoch": 0.48304288090245134, "grad_norm": 0.9521824717521667, "learning_rate": 0.00019785692774430862, "loss": 4.6115, "step": 258850 }, { "epoch": 0.4831361864618299, "grad_norm": 1.0983976125717163, "learning_rate": 0.0001978560626009211, "loss": 4.5424, "step": 258900 }, { "epoch": 0.48322949202120835, "grad_norm": 0.9267615675926208, "learning_rate": 0.0001978551972848349, "loss": 4.5658, "step": 258950 }, { "epoch": 0.4833227975805868, "grad_norm": 1.1117475032806396, "learning_rate": 0.00019785433179605143, "loss": 4.6279, "step": 259000 }, { "epoch": 0.48341610313996536, "grad_norm": 0.9783563613891602, "learning_rate": 0.00019785346613457235, "loss": 4.5828, "step": 259050 }, { "epoch": 0.48350940869934383, "grad_norm": 1.0470584630966187, "learning_rate": 0.00019785260030039906, "loss": 4.5992, "step": 259100 }, { "epoch": 0.4836027142587223, "grad_norm": 1.1836624145507812, "learning_rate": 0.00019785173429353318, "loss": 4.6148, "step": 259150 }, { "epoch": 0.4836960198181008, "grad_norm": 1.0793054103851318, "learning_rate": 0.00019785086811397617, "loss": 4.4822, "step": 259200 }, { "epoch": 0.4837893253774793, "grad_norm": 0.7679390907287598, "learning_rate": 0.0001978500017617296, "loss": 4.6483, "step": 259250 }, { "epoch": 0.4838826309368578, "grad_norm": 0.987201988697052, "learning_rate": 0.00019784913523679504, "loss": 4.5346, "step": 259300 }, { "epoch": 0.48397593649623627, "grad_norm": 1.2286121845245361, "learning_rate": 0.00019784826853917395, "loss": 4.751, "step": 259350 }, { "epoch": 0.4840692420556148, "grad_norm": 1.232087254524231, "learning_rate": 0.0001978474016688679, "loss": 4.6215, "step": 259400 }, { "epoch": 0.4841625476149933, "grad_norm": 0.9811508059501648, "learning_rate": 0.00019784653462587835, "loss": 4.3616, "step": 259450 }, { "epoch": 0.48425585317437175, "grad_norm": 1.0040799379348755, "learning_rate": 0.00019784566741020695, "loss": 4.4287, "step": 259500 }, { "epoch": 0.48434915873375023, "grad_norm": 0.8686191439628601, "learning_rate": 0.0001978448000218551, "loss": 4.468, "step": 259550 }, { "epoch": 0.48444246429312876, "grad_norm": 0.8420849442481995, "learning_rate": 0.0001978439324608244, "loss": 4.6569, "step": 259600 }, { "epoch": 0.48453576985250724, "grad_norm": 0.7843796610832214, "learning_rate": 0.00019784306472711642, "loss": 4.4424, "step": 259650 }, { "epoch": 0.4846290754118857, "grad_norm": 1.2961535453796387, "learning_rate": 0.00019784219682073264, "loss": 4.6343, "step": 259700 }, { "epoch": 0.48472238097126424, "grad_norm": 1.09566330909729, "learning_rate": 0.0001978413287416746, "loss": 4.4992, "step": 259750 }, { "epoch": 0.4848156865306427, "grad_norm": 0.9257076978683472, "learning_rate": 0.0001978404604899438, "loss": 4.6446, "step": 259800 }, { "epoch": 0.4849089920900212, "grad_norm": 0.9009526371955872, "learning_rate": 0.0001978395920655418, "loss": 4.4916, "step": 259850 }, { "epoch": 0.48500229764939967, "grad_norm": 1.0999066829681396, "learning_rate": 0.00019783872346847018, "loss": 4.4681, "step": 259900 }, { "epoch": 0.4850956032087782, "grad_norm": 0.8918809294700623, "learning_rate": 0.00019783785469873038, "loss": 4.5746, "step": 259950 }, { "epoch": 0.4851889087681567, "grad_norm": 1.1592248678207397, "learning_rate": 0.000197836985756324, "loss": 4.6462, "step": 260000 }, { "epoch": 0.48528221432753516, "grad_norm": 1.0789110660552979, "learning_rate": 0.00019783611664125257, "loss": 4.6372, "step": 260050 }, { "epoch": 0.4853755198869137, "grad_norm": 0.7961848974227905, "learning_rate": 0.0001978352473535176, "loss": 4.7098, "step": 260100 }, { "epoch": 0.48546882544629216, "grad_norm": 1.1101049184799194, "learning_rate": 0.0001978343778931206, "loss": 4.4104, "step": 260150 }, { "epoch": 0.48556213100567064, "grad_norm": 1.1163572072982788, "learning_rate": 0.00019783350826006317, "loss": 4.6022, "step": 260200 }, { "epoch": 0.4856554365650491, "grad_norm": 0.8605142831802368, "learning_rate": 0.00019783263845434682, "loss": 4.6625, "step": 260250 }, { "epoch": 0.48574874212442765, "grad_norm": 0.9262938499450684, "learning_rate": 0.00019783176847597308, "loss": 4.6614, "step": 260300 }, { "epoch": 0.4858420476838061, "grad_norm": 1.254042387008667, "learning_rate": 0.00019783089832494345, "loss": 4.5324, "step": 260350 }, { "epoch": 0.4859353532431846, "grad_norm": 1.2252981662750244, "learning_rate": 0.00019783002800125954, "loss": 4.541, "step": 260400 }, { "epoch": 0.48602865880256313, "grad_norm": 0.9264373779296875, "learning_rate": 0.00019782915750492282, "loss": 4.6012, "step": 260450 }, { "epoch": 0.4861219643619416, "grad_norm": 1.1347676515579224, "learning_rate": 0.00019782828683593485, "loss": 4.7729, "step": 260500 }, { "epoch": 0.4862152699213201, "grad_norm": 0.581415593624115, "learning_rate": 0.00019782741599429717, "loss": 4.5686, "step": 260550 }, { "epoch": 0.48630857548069856, "grad_norm": 1.1789880990982056, "learning_rate": 0.00019782654498001133, "loss": 4.5014, "step": 260600 }, { "epoch": 0.4864018810400771, "grad_norm": 1.1918727159500122, "learning_rate": 0.0001978256737930788, "loss": 4.5867, "step": 260650 }, { "epoch": 0.48649518659945556, "grad_norm": 1.2388694286346436, "learning_rate": 0.00019782480243350122, "loss": 4.4532, "step": 260700 }, { "epoch": 0.48658849215883404, "grad_norm": 0.9649792313575745, "learning_rate": 0.00019782393090128007, "loss": 4.6906, "step": 260750 }, { "epoch": 0.48668179771821257, "grad_norm": 0.8217645883560181, "learning_rate": 0.00019782305919641689, "loss": 4.4997, "step": 260800 }, { "epoch": 0.48677510327759105, "grad_norm": 1.2379895448684692, "learning_rate": 0.0001978221873189132, "loss": 4.7031, "step": 260850 }, { "epoch": 0.4868684088369695, "grad_norm": 0.9519298672676086, "learning_rate": 0.00019782131526877058, "loss": 4.7055, "step": 260900 }, { "epoch": 0.486961714396348, "grad_norm": 0.8161171674728394, "learning_rate": 0.00019782044304599055, "loss": 4.826, "step": 260950 }, { "epoch": 0.48705501995572653, "grad_norm": 1.0965524911880493, "learning_rate": 0.00019781957065057463, "loss": 4.6886, "step": 261000 }, { "epoch": 0.48705501995572653, "eval_loss": 4.7727813720703125, "eval_runtime": 228.7535, "eval_samples_per_second": 11.401, "eval_steps_per_second": 11.401, "eval_tts_loss": 7.503641583035455, "step": 261000 }, { "epoch": 0.487148325515105, "grad_norm": 1.025140643119812, "learning_rate": 0.00019781869808252443, "loss": 4.613, "step": 261050 }, { "epoch": 0.4872416310744835, "grad_norm": 1.0564911365509033, "learning_rate": 0.0001978178253418414, "loss": 4.5312, "step": 261100 }, { "epoch": 0.487334936633862, "grad_norm": 0.8013015985488892, "learning_rate": 0.00019781695242852712, "loss": 4.5406, "step": 261150 }, { "epoch": 0.4874282421932405, "grad_norm": 0.806037425994873, "learning_rate": 0.00019781607934258314, "loss": 4.6685, "step": 261200 }, { "epoch": 0.48752154775261897, "grad_norm": 1.0473806858062744, "learning_rate": 0.000197815206084011, "loss": 4.5322, "step": 261250 }, { "epoch": 0.48761485331199744, "grad_norm": 1.0605626106262207, "learning_rate": 0.0001978143326528122, "loss": 4.4764, "step": 261300 }, { "epoch": 0.487708158871376, "grad_norm": 0.7920166254043579, "learning_rate": 0.00019781345904898835, "loss": 4.6327, "step": 261350 }, { "epoch": 0.48780146443075445, "grad_norm": 0.9014812111854553, "learning_rate": 0.00019781258527254097, "loss": 4.6423, "step": 261400 }, { "epoch": 0.4878947699901329, "grad_norm": 0.8125473856925964, "learning_rate": 0.00019781171132347155, "loss": 4.6761, "step": 261450 }, { "epoch": 0.4879880755495114, "grad_norm": 1.018195390701294, "learning_rate": 0.00019781083720178167, "loss": 4.6347, "step": 261500 }, { "epoch": 0.48808138110888993, "grad_norm": 1.0549075603485107, "learning_rate": 0.0001978099629074729, "loss": 4.633, "step": 261550 }, { "epoch": 0.4881746866682684, "grad_norm": 1.154897689819336, "learning_rate": 0.00019780908844054673, "loss": 4.6492, "step": 261600 }, { "epoch": 0.4882679922276469, "grad_norm": 0.9253442883491516, "learning_rate": 0.00019780821380100473, "loss": 4.6132, "step": 261650 }, { "epoch": 0.4883612977870254, "grad_norm": 1.029456615447998, "learning_rate": 0.00019780733898884848, "loss": 4.4266, "step": 261700 }, { "epoch": 0.4884546033464039, "grad_norm": 1.1559395790100098, "learning_rate": 0.00019780646400407946, "loss": 4.4551, "step": 261750 }, { "epoch": 0.48854790890578237, "grad_norm": 1.0872691869735718, "learning_rate": 0.00019780558884669922, "loss": 4.4269, "step": 261800 }, { "epoch": 0.48864121446516084, "grad_norm": 0.7114251852035522, "learning_rate": 0.00019780471351670933, "loss": 4.405, "step": 261850 }, { "epoch": 0.4887345200245394, "grad_norm": 0.846502423286438, "learning_rate": 0.00019780383801411134, "loss": 4.4984, "step": 261900 }, { "epoch": 0.48882782558391785, "grad_norm": 1.0822960138320923, "learning_rate": 0.0001978029623389068, "loss": 4.5524, "step": 261950 }, { "epoch": 0.48892113114329633, "grad_norm": 1.0587904453277588, "learning_rate": 0.0001978020864910972, "loss": 4.5385, "step": 262000 }, { "epoch": 0.48901443670267486, "grad_norm": 0.9366457462310791, "learning_rate": 0.00019780121047068417, "loss": 4.7893, "step": 262050 }, { "epoch": 0.48910774226205334, "grad_norm": 1.1019724607467651, "learning_rate": 0.00019780033427766922, "loss": 4.5838, "step": 262100 }, { "epoch": 0.4892010478214318, "grad_norm": 1.3364999294281006, "learning_rate": 0.00019779945791205386, "loss": 4.5512, "step": 262150 }, { "epoch": 0.4892943533808103, "grad_norm": 0.7997466921806335, "learning_rate": 0.00019779858137383967, "loss": 4.4808, "step": 262200 }, { "epoch": 0.4893876589401888, "grad_norm": 1.033316731452942, "learning_rate": 0.0001977977046630282, "loss": 4.5601, "step": 262250 }, { "epoch": 0.4894809644995673, "grad_norm": 1.1399928331375122, "learning_rate": 0.00019779682777962097, "loss": 4.5472, "step": 262300 }, { "epoch": 0.48957427005894577, "grad_norm": 0.9686580896377563, "learning_rate": 0.00019779595072361955, "loss": 4.8388, "step": 262350 }, { "epoch": 0.4896675756183243, "grad_norm": 0.9062378406524658, "learning_rate": 0.00019779507349502548, "loss": 4.5663, "step": 262400 }, { "epoch": 0.4897608811777028, "grad_norm": 1.001134991645813, "learning_rate": 0.00019779419609384035, "loss": 4.6099, "step": 262450 }, { "epoch": 0.48985418673708125, "grad_norm": 1.1371427774429321, "learning_rate": 0.00019779331852006564, "loss": 4.5837, "step": 262500 }, { "epoch": 0.48994749229645973, "grad_norm": 0.9281883835792542, "learning_rate": 0.00019779244077370293, "loss": 4.7212, "step": 262550 }, { "epoch": 0.49004079785583826, "grad_norm": 1.3296451568603516, "learning_rate": 0.0001977915628547538, "loss": 4.8147, "step": 262600 }, { "epoch": 0.49013410341521674, "grad_norm": 1.0357704162597656, "learning_rate": 0.00019779068476321975, "loss": 4.7545, "step": 262650 }, { "epoch": 0.4902274089745952, "grad_norm": 1.1672788858413696, "learning_rate": 0.00019778980649910234, "loss": 4.6424, "step": 262700 }, { "epoch": 0.49032071453397374, "grad_norm": 1.162309169769287, "learning_rate": 0.0001977889280624031, "loss": 4.7544, "step": 262750 }, { "epoch": 0.4904140200933522, "grad_norm": 1.0663206577301025, "learning_rate": 0.00019778804945312365, "loss": 4.3885, "step": 262800 }, { "epoch": 0.4905073256527307, "grad_norm": 0.9657242298126221, "learning_rate": 0.0001977871706712655, "loss": 4.7163, "step": 262850 }, { "epoch": 0.4906006312121092, "grad_norm": 1.080690860748291, "learning_rate": 0.0001977862917168302, "loss": 4.5714, "step": 262900 }, { "epoch": 0.4906939367714877, "grad_norm": 1.0465108156204224, "learning_rate": 0.00019778541258981926, "loss": 4.511, "step": 262950 }, { "epoch": 0.4907872423308662, "grad_norm": 1.0422312021255493, "learning_rate": 0.00019778453329023432, "loss": 4.4401, "step": 263000 }, { "epoch": 0.49088054789024466, "grad_norm": 1.068161129951477, "learning_rate": 0.00019778365381807687, "loss": 4.5966, "step": 263050 }, { "epoch": 0.4909738534496232, "grad_norm": 1.1453481912612915, "learning_rate": 0.00019778277417334845, "loss": 4.5744, "step": 263100 }, { "epoch": 0.49106715900900166, "grad_norm": 0.9773776531219482, "learning_rate": 0.00019778189435605068, "loss": 4.7388, "step": 263150 }, { "epoch": 0.49116046456838014, "grad_norm": 0.8863257169723511, "learning_rate": 0.00019778101436618506, "loss": 4.4572, "step": 263200 }, { "epoch": 0.4912537701277586, "grad_norm": 0.8632976412773132, "learning_rate": 0.00019778013420375314, "loss": 4.7325, "step": 263250 }, { "epoch": 0.49134707568713715, "grad_norm": 1.222376823425293, "learning_rate": 0.00019777925386875648, "loss": 4.4925, "step": 263300 }, { "epoch": 0.4914403812465156, "grad_norm": 1.2068899869918823, "learning_rate": 0.00019777837336119668, "loss": 4.6069, "step": 263350 }, { "epoch": 0.4915336868058941, "grad_norm": 1.1874123811721802, "learning_rate": 0.00019777749268107525, "loss": 4.5623, "step": 263400 }, { "epoch": 0.49162699236527263, "grad_norm": 0.8214643001556396, "learning_rate": 0.00019777661182839373, "loss": 4.5084, "step": 263450 }, { "epoch": 0.4917202979246511, "grad_norm": 1.1660773754119873, "learning_rate": 0.0001977757308031537, "loss": 4.5678, "step": 263500 }, { "epoch": 0.4918136034840296, "grad_norm": 0.8827805519104004, "learning_rate": 0.00019777484960535667, "loss": 4.7326, "step": 263550 }, { "epoch": 0.49190690904340806, "grad_norm": 0.9647439122200012, "learning_rate": 0.0001977739682350043, "loss": 4.7222, "step": 263600 }, { "epoch": 0.4920002146027866, "grad_norm": 0.7739233374595642, "learning_rate": 0.00019777308669209807, "loss": 4.4787, "step": 263650 }, { "epoch": 0.49209352016216507, "grad_norm": 0.7975419759750366, "learning_rate": 0.0001977722049766395, "loss": 4.6783, "step": 263700 }, { "epoch": 0.49218682572154354, "grad_norm": 1.198824167251587, "learning_rate": 0.00019777132308863022, "loss": 4.5501, "step": 263750 }, { "epoch": 0.4922801312809221, "grad_norm": 1.0451983213424683, "learning_rate": 0.00019777044102807175, "loss": 4.63, "step": 263800 }, { "epoch": 0.49237343684030055, "grad_norm": 1.064078450202942, "learning_rate": 0.00019776955879496569, "loss": 4.5926, "step": 263850 }, { "epoch": 0.492466742399679, "grad_norm": 0.6753122806549072, "learning_rate": 0.00019776867638931353, "loss": 4.5474, "step": 263900 }, { "epoch": 0.4925600479590575, "grad_norm": 1.0703619718551636, "learning_rate": 0.00019776779381111686, "loss": 4.6989, "step": 263950 }, { "epoch": 0.49265335351843603, "grad_norm": 0.9324034452438354, "learning_rate": 0.00019776691106037722, "loss": 4.6068, "step": 264000 }, { "epoch": 0.49265335351843603, "eval_loss": 4.763203144073486, "eval_runtime": 228.6077, "eval_samples_per_second": 11.408, "eval_steps_per_second": 11.408, "eval_tts_loss": 7.533669790340176, "step": 264000 }, { "epoch": 0.4927466590778145, "grad_norm": 1.1262904405593872, "learning_rate": 0.00019776602813709624, "loss": 4.6063, "step": 264050 }, { "epoch": 0.492839964637193, "grad_norm": 0.7986791133880615, "learning_rate": 0.00019776514504127538, "loss": 4.5345, "step": 264100 }, { "epoch": 0.49293327019657146, "grad_norm": 1.026489496231079, "learning_rate": 0.00019776426177291624, "loss": 4.638, "step": 264150 }, { "epoch": 0.49302657575595, "grad_norm": 1.1013755798339844, "learning_rate": 0.00019776337833202039, "loss": 4.5174, "step": 264200 }, { "epoch": 0.49311988131532847, "grad_norm": 0.9843734502792358, "learning_rate": 0.0001977624947185894, "loss": 4.3527, "step": 264250 }, { "epoch": 0.49321318687470694, "grad_norm": 1.009853720664978, "learning_rate": 0.00019776161093262474, "loss": 4.5395, "step": 264300 }, { "epoch": 0.4933064924340855, "grad_norm": 1.039201021194458, "learning_rate": 0.00019776072697412812, "loss": 4.5555, "step": 264350 }, { "epoch": 0.49339979799346395, "grad_norm": 1.022912859916687, "learning_rate": 0.000197759842843101, "loss": 4.6428, "step": 264400 }, { "epoch": 0.4934931035528424, "grad_norm": 1.0572845935821533, "learning_rate": 0.0001977589585395449, "loss": 4.4889, "step": 264450 }, { "epoch": 0.4935864091122209, "grad_norm": 1.2854913473129272, "learning_rate": 0.0001977580740634615, "loss": 4.695, "step": 264500 }, { "epoch": 0.49367971467159943, "grad_norm": 1.0805784463882446, "learning_rate": 0.00019775718941485228, "loss": 4.6472, "step": 264550 }, { "epoch": 0.4937730202309779, "grad_norm": 1.0864139795303345, "learning_rate": 0.00019775630459371884, "loss": 4.5395, "step": 264600 }, { "epoch": 0.4938663257903564, "grad_norm": 0.8742742538452148, "learning_rate": 0.0001977554196000627, "loss": 4.5496, "step": 264650 }, { "epoch": 0.4939596313497349, "grad_norm": 1.1790704727172852, "learning_rate": 0.00019775453443388546, "loss": 4.5932, "step": 264700 }, { "epoch": 0.4940529369091134, "grad_norm": 0.8600195050239563, "learning_rate": 0.00019775364909518865, "loss": 4.3761, "step": 264750 }, { "epoch": 0.49414624246849187, "grad_norm": 1.3297063112258911, "learning_rate": 0.00019775276358397384, "loss": 4.7526, "step": 264800 }, { "epoch": 0.49423954802787035, "grad_norm": 0.9508358240127563, "learning_rate": 0.00019775187790024263, "loss": 4.4664, "step": 264850 }, { "epoch": 0.4943328535872489, "grad_norm": 1.2345694303512573, "learning_rate": 0.00019775099204399657, "loss": 4.6711, "step": 264900 }, { "epoch": 0.49442615914662735, "grad_norm": 1.387581467628479, "learning_rate": 0.00019775010601523718, "loss": 4.5366, "step": 264950 }, { "epoch": 0.49451946470600583, "grad_norm": 0.9651834964752197, "learning_rate": 0.00019774921981396607, "loss": 4.5564, "step": 265000 }, { "epoch": 0.49461277026538436, "grad_norm": 1.261367917060852, "learning_rate": 0.00019774833344018476, "loss": 4.4319, "step": 265050 }, { "epoch": 0.49470607582476284, "grad_norm": 0.7602758407592773, "learning_rate": 0.00019774744689389487, "loss": 4.737, "step": 265100 }, { "epoch": 0.4947993813841413, "grad_norm": 1.285857081413269, "learning_rate": 0.0001977465601750979, "loss": 4.6235, "step": 265150 }, { "epoch": 0.4948926869435198, "grad_norm": 1.1164087057113647, "learning_rate": 0.00019774567328379548, "loss": 4.4363, "step": 265200 }, { "epoch": 0.4949859925028983, "grad_norm": 1.1866155862808228, "learning_rate": 0.00019774478621998912, "loss": 4.5297, "step": 265250 }, { "epoch": 0.4950792980622768, "grad_norm": 1.0542068481445312, "learning_rate": 0.00019774389898368045, "loss": 4.6326, "step": 265300 }, { "epoch": 0.49517260362165527, "grad_norm": 1.035804033279419, "learning_rate": 0.00019774301157487095, "loss": 4.7021, "step": 265350 }, { "epoch": 0.4952659091810338, "grad_norm": 1.0542982816696167, "learning_rate": 0.00019774212399356226, "loss": 4.7791, "step": 265400 }, { "epoch": 0.4953592147404123, "grad_norm": 1.058598279953003, "learning_rate": 0.00019774123623975592, "loss": 4.4212, "step": 265450 }, { "epoch": 0.49545252029979076, "grad_norm": 1.0170423984527588, "learning_rate": 0.0001977403483134535, "loss": 4.3866, "step": 265500 }, { "epoch": 0.49554582585916923, "grad_norm": 0.9709385633468628, "learning_rate": 0.00019773946021465656, "loss": 4.5444, "step": 265550 }, { "epoch": 0.49563913141854776, "grad_norm": 1.150532603263855, "learning_rate": 0.00019773857194336667, "loss": 4.6281, "step": 265600 }, { "epoch": 0.49573243697792624, "grad_norm": 0.9875307679176331, "learning_rate": 0.0001977376834995854, "loss": 4.6653, "step": 265650 }, { "epoch": 0.4958257425373047, "grad_norm": 1.2764935493469238, "learning_rate": 0.00019773679488331429, "loss": 4.4393, "step": 265700 }, { "epoch": 0.49591904809668325, "grad_norm": 0.9200681447982788, "learning_rate": 0.00019773590609455498, "loss": 4.7367, "step": 265750 }, { "epoch": 0.4960123536560617, "grad_norm": 0.8322024941444397, "learning_rate": 0.00019773501713330895, "loss": 4.5703, "step": 265800 }, { "epoch": 0.4961056592154402, "grad_norm": 0.9457756876945496, "learning_rate": 0.00019773412799957784, "loss": 4.6463, "step": 265850 }, { "epoch": 0.4961989647748187, "grad_norm": 1.083602786064148, "learning_rate": 0.00019773323869336318, "loss": 4.728, "step": 265900 }, { "epoch": 0.4962922703341972, "grad_norm": 0.7881966829299927, "learning_rate": 0.00019773234921466654, "loss": 4.4035, "step": 265950 }, { "epoch": 0.4963855758935757, "grad_norm": 0.9676995873451233, "learning_rate": 0.00019773145956348954, "loss": 4.542, "step": 266000 }, { "epoch": 0.49647888145295416, "grad_norm": 1.1931356191635132, "learning_rate": 0.00019773056973983367, "loss": 4.7262, "step": 266050 }, { "epoch": 0.4965721870123327, "grad_norm": 1.042542815208435, "learning_rate": 0.00019772967974370054, "loss": 4.7454, "step": 266100 }, { "epoch": 0.49666549257171116, "grad_norm": 0.9951144456863403, "learning_rate": 0.00019772878957509177, "loss": 4.4392, "step": 266150 }, { "epoch": 0.49675879813108964, "grad_norm": 1.0997178554534912, "learning_rate": 0.00019772789923400882, "loss": 4.454, "step": 266200 }, { "epoch": 0.4968521036904681, "grad_norm": 1.0528903007507324, "learning_rate": 0.00019772700872045337, "loss": 4.6066, "step": 266250 }, { "epoch": 0.49694540924984665, "grad_norm": 1.052966833114624, "learning_rate": 0.00019772611803442692, "loss": 4.5549, "step": 266300 }, { "epoch": 0.4970387148092251, "grad_norm": 0.7441802620887756, "learning_rate": 0.00019772522717593108, "loss": 4.6553, "step": 266350 }, { "epoch": 0.4971320203686036, "grad_norm": 1.0470805168151855, "learning_rate": 0.0001977243361449674, "loss": 4.5749, "step": 266400 }, { "epoch": 0.49722532592798213, "grad_norm": 0.9941937327384949, "learning_rate": 0.00019772344494153746, "loss": 4.6922, "step": 266450 }, { "epoch": 0.4973186314873606, "grad_norm": 0.9114199280738831, "learning_rate": 0.00019772255356564282, "loss": 4.532, "step": 266500 }, { "epoch": 0.4974119370467391, "grad_norm": 1.1066757440567017, "learning_rate": 0.00019772166201728508, "loss": 4.7109, "step": 266550 }, { "epoch": 0.49750524260611756, "grad_norm": 0.7902519106864929, "learning_rate": 0.0001977207702964658, "loss": 4.7002, "step": 266600 }, { "epoch": 0.4975985481654961, "grad_norm": 0.84111088514328, "learning_rate": 0.00019771987840318657, "loss": 4.6185, "step": 266650 }, { "epoch": 0.49769185372487457, "grad_norm": 0.8162730932235718, "learning_rate": 0.00019771898633744894, "loss": 4.6529, "step": 266700 }, { "epoch": 0.49778515928425304, "grad_norm": 1.0409742593765259, "learning_rate": 0.0001977180940992545, "loss": 4.5277, "step": 266750 }, { "epoch": 0.4978784648436315, "grad_norm": 0.9645498394966125, "learning_rate": 0.00019771720168860481, "loss": 4.6303, "step": 266800 }, { "epoch": 0.49797177040301005, "grad_norm": 1.0645626783370972, "learning_rate": 0.00019771630910550146, "loss": 4.7958, "step": 266850 }, { "epoch": 0.4980650759623885, "grad_norm": 1.721364974975586, "learning_rate": 0.000197715416349946, "loss": 4.5781, "step": 266900 }, { "epoch": 0.498158381521767, "grad_norm": 0.9426038265228271, "learning_rate": 0.00019771452342194005, "loss": 4.5271, "step": 266950 }, { "epoch": 0.49825168708114553, "grad_norm": 1.0938518047332764, "learning_rate": 0.00019771363032148516, "loss": 4.5928, "step": 267000 }, { "epoch": 0.49825168708114553, "eval_loss": 4.766524314880371, "eval_runtime": 229.0764, "eval_samples_per_second": 11.385, "eval_steps_per_second": 11.385, "eval_tts_loss": 7.488897705084981, "step": 267000 }, { "epoch": 0.498344992640524, "grad_norm": 1.268874168395996, "learning_rate": 0.00019771273704858287, "loss": 4.4298, "step": 267050 }, { "epoch": 0.4984382981999025, "grad_norm": 1.0305707454681396, "learning_rate": 0.0001977118436032348, "loss": 4.657, "step": 267100 }, { "epoch": 0.49853160375928096, "grad_norm": 0.9082435369491577, "learning_rate": 0.00019771094998544255, "loss": 4.6786, "step": 267150 }, { "epoch": 0.4986249093186595, "grad_norm": 0.9867738485336304, "learning_rate": 0.00019771005619520767, "loss": 4.6024, "step": 267200 }, { "epoch": 0.49871821487803797, "grad_norm": 1.0489767789840698, "learning_rate": 0.00019770916223253168, "loss": 4.5919, "step": 267250 }, { "epoch": 0.49881152043741644, "grad_norm": 1.0940841436386108, "learning_rate": 0.00019770826809741628, "loss": 4.6786, "step": 267300 }, { "epoch": 0.498904825996795, "grad_norm": 1.0874834060668945, "learning_rate": 0.00019770737378986292, "loss": 4.8278, "step": 267350 }, { "epoch": 0.49899813155617345, "grad_norm": 1.1011561155319214, "learning_rate": 0.00019770647930987328, "loss": 4.3219, "step": 267400 }, { "epoch": 0.49909143711555193, "grad_norm": 1.0629221200942993, "learning_rate": 0.00019770558465744887, "loss": 4.6627, "step": 267450 }, { "epoch": 0.4991847426749304, "grad_norm": 0.8989550471305847, "learning_rate": 0.0001977046898325913, "loss": 4.7171, "step": 267500 }, { "epoch": 0.49927804823430894, "grad_norm": 1.0148009061813354, "learning_rate": 0.00019770379483530215, "loss": 4.6872, "step": 267550 }, { "epoch": 0.4993713537936874, "grad_norm": 1.0268181562423706, "learning_rate": 0.00019770289966558298, "loss": 4.5703, "step": 267600 }, { "epoch": 0.4994646593530659, "grad_norm": 0.9970580339431763, "learning_rate": 0.0001977020043234354, "loss": 4.4933, "step": 267650 }, { "epoch": 0.4995579649124444, "grad_norm": 1.5218355655670166, "learning_rate": 0.00019770110880886095, "loss": 4.6482, "step": 267700 }, { "epoch": 0.4996512704718229, "grad_norm": 0.9240729212760925, "learning_rate": 0.00019770021312186126, "loss": 4.5817, "step": 267750 }, { "epoch": 0.49974457603120137, "grad_norm": 1.1267787218093872, "learning_rate": 0.00019769931726243786, "loss": 4.6775, "step": 267800 }, { "epoch": 0.49983788159057985, "grad_norm": 0.8717399835586548, "learning_rate": 0.0001976984212305924, "loss": 4.7974, "step": 267850 }, { "epoch": 0.4999311871499584, "grad_norm": 0.7997406125068665, "learning_rate": 0.00019769752502632638, "loss": 4.7391, "step": 267900 }, { "epoch": 0.5000244927093368, "grad_norm": 1.135779619216919, "learning_rate": 0.00019769662864964143, "loss": 4.5182, "step": 267950 }, { "epoch": 0.5001177982687154, "grad_norm": 1.0822643041610718, "learning_rate": 0.0001976957321005391, "loss": 4.5152, "step": 268000 }, { "epoch": 0.5002111038280939, "grad_norm": 0.8304023146629333, "learning_rate": 0.00019769483537902103, "loss": 4.5748, "step": 268050 }, { "epoch": 0.5003044093874723, "grad_norm": 0.9894828200340271, "learning_rate": 0.00019769393848508875, "loss": 4.5601, "step": 268100 }, { "epoch": 0.5003977149468508, "grad_norm": 1.1318585872650146, "learning_rate": 0.00019769304141874385, "loss": 4.4261, "step": 268150 }, { "epoch": 0.5004910205062293, "grad_norm": 1.3329347372055054, "learning_rate": 0.00019769214417998793, "loss": 4.7159, "step": 268200 }, { "epoch": 0.5005843260656078, "grad_norm": 1.1386692523956299, "learning_rate": 0.00019769124676882255, "loss": 4.374, "step": 268250 }, { "epoch": 0.5006776316249862, "grad_norm": 1.0569969415664673, "learning_rate": 0.0001976903491852493, "loss": 4.7144, "step": 268300 }, { "epoch": 0.5007709371843648, "grad_norm": 0.9701844453811646, "learning_rate": 0.0001976894514292698, "loss": 4.4343, "step": 268350 }, { "epoch": 0.5008642427437433, "grad_norm": 0.9479658007621765, "learning_rate": 0.00019768855350088562, "loss": 4.3689, "step": 268400 }, { "epoch": 0.5009575483031218, "grad_norm": 1.123121976852417, "learning_rate": 0.0001976876554000983, "loss": 4.3821, "step": 268450 }, { "epoch": 0.5010508538625003, "grad_norm": 1.1315399408340454, "learning_rate": 0.00019768675712690948, "loss": 4.7261, "step": 268500 }, { "epoch": 0.5011441594218787, "grad_norm": 0.9239130616188049, "learning_rate": 0.0001976858586813207, "loss": 4.7715, "step": 268550 }, { "epoch": 0.5012374649812572, "grad_norm": 1.1073530912399292, "learning_rate": 0.00019768496006333357, "loss": 4.5403, "step": 268600 }, { "epoch": 0.5013307705406357, "grad_norm": 0.8722808957099915, "learning_rate": 0.00019768406127294969, "loss": 4.6618, "step": 268650 }, { "epoch": 0.5014240761000143, "grad_norm": 1.0455543994903564, "learning_rate": 0.0001976831623101706, "loss": 4.6246, "step": 268700 }, { "epoch": 0.5015173816593927, "grad_norm": 1.1023845672607422, "learning_rate": 0.00019768226317499796, "loss": 4.6571, "step": 268750 }, { "epoch": 0.5016106872187712, "grad_norm": 1.113635778427124, "learning_rate": 0.0001976813638674333, "loss": 4.6877, "step": 268800 }, { "epoch": 0.5017039927781497, "grad_norm": 1.1741982698440552, "learning_rate": 0.0001976804643874782, "loss": 4.5927, "step": 268850 }, { "epoch": 0.5017972983375282, "grad_norm": 1.0700513124465942, "learning_rate": 0.00019767956473513428, "loss": 4.5712, "step": 268900 }, { "epoch": 0.5018906038969067, "grad_norm": 0.9588819146156311, "learning_rate": 0.00019767866491040312, "loss": 4.6768, "step": 268950 }, { "epoch": 0.5019839094562851, "grad_norm": 1.2364482879638672, "learning_rate": 0.0001976777649132863, "loss": 4.643, "step": 269000 }, { "epoch": 0.5020772150156637, "grad_norm": 1.096693754196167, "learning_rate": 0.0001976768647437854, "loss": 4.4986, "step": 269050 }, { "epoch": 0.5021705205750422, "grad_norm": 0.6519501209259033, "learning_rate": 0.00019767596440190203, "loss": 4.5439, "step": 269100 }, { "epoch": 0.5022638261344207, "grad_norm": 1.2315655946731567, "learning_rate": 0.00019767506388763777, "loss": 4.7023, "step": 269150 }, { "epoch": 0.5023571316937991, "grad_norm": 1.1877769231796265, "learning_rate": 0.00019767416320099423, "loss": 4.4986, "step": 269200 }, { "epoch": 0.5024504372531776, "grad_norm": 1.1195615530014038, "learning_rate": 0.00019767326234197295, "loss": 4.6128, "step": 269250 }, { "epoch": 0.5025437428125561, "grad_norm": 0.8910592794418335, "learning_rate": 0.00019767236131057553, "loss": 4.4873, "step": 269300 }, { "epoch": 0.5026370483719346, "grad_norm": 0.8761081099510193, "learning_rate": 0.0001976714601068036, "loss": 4.8037, "step": 269350 }, { "epoch": 0.5027303539313132, "grad_norm": 1.1296546459197998, "learning_rate": 0.00019767055873065877, "loss": 4.6226, "step": 269400 }, { "epoch": 0.5028236594906916, "grad_norm": 0.8964772820472717, "learning_rate": 0.00019766965718214253, "loss": 4.4358, "step": 269450 }, { "epoch": 0.5029169650500701, "grad_norm": 1.181470274925232, "learning_rate": 0.00019766875546125656, "loss": 4.7548, "step": 269500 }, { "epoch": 0.5030102706094486, "grad_norm": 1.027578353881836, "learning_rate": 0.00019766785356800242, "loss": 4.6212, "step": 269550 }, { "epoch": 0.5031035761688271, "grad_norm": 0.7062067985534668, "learning_rate": 0.00019766695150238167, "loss": 4.5479, "step": 269600 }, { "epoch": 0.5031968817282055, "grad_norm": 0.781272828578949, "learning_rate": 0.000197666049264396, "loss": 4.4475, "step": 269650 }, { "epoch": 0.503290187287584, "grad_norm": 0.8852269053459167, "learning_rate": 0.00019766514685404688, "loss": 4.7096, "step": 269700 }, { "epoch": 0.5033834928469626, "grad_norm": 0.8530430197715759, "learning_rate": 0.000197664244271336, "loss": 4.3999, "step": 269750 }, { "epoch": 0.5034767984063411, "grad_norm": 0.670219898223877, "learning_rate": 0.0001976633415162649, "loss": 4.7789, "step": 269800 }, { "epoch": 0.5035701039657196, "grad_norm": 0.8426247239112854, "learning_rate": 0.00019766243858883518, "loss": 4.5608, "step": 269850 }, { "epoch": 0.503663409525098, "grad_norm": 1.1550434827804565, "learning_rate": 0.00019766153548904844, "loss": 4.605, "step": 269900 }, { "epoch": 0.5037567150844765, "grad_norm": 0.9927839636802673, "learning_rate": 0.00019766063221690627, "loss": 4.41, "step": 269950 }, { "epoch": 0.503850020643855, "grad_norm": 0.9224207401275635, "learning_rate": 0.00019765972877241027, "loss": 4.5662, "step": 270000 }, { "epoch": 0.503850020643855, "eval_loss": 4.762325286865234, "eval_runtime": 230.697, "eval_samples_per_second": 11.305, "eval_steps_per_second": 11.305, "eval_tts_loss": 7.51998570571265, "step": 270000 }, { "epoch": 0.5039433262032335, "grad_norm": 1.099448561668396, "learning_rate": 0.00019765882515556204, "loss": 4.7066, "step": 270050 }, { "epoch": 0.504036631762612, "grad_norm": 0.9349024891853333, "learning_rate": 0.00019765792136636315, "loss": 4.5372, "step": 270100 }, { "epoch": 0.5041299373219905, "grad_norm": 0.8908520936965942, "learning_rate": 0.00019765701740481523, "loss": 4.4929, "step": 270150 }, { "epoch": 0.504223242881369, "grad_norm": 0.9843934774398804, "learning_rate": 0.00019765611327091987, "loss": 4.6941, "step": 270200 }, { "epoch": 0.5043165484407475, "grad_norm": 1.1821966171264648, "learning_rate": 0.00019765520896467863, "loss": 4.3483, "step": 270250 }, { "epoch": 0.504409854000126, "grad_norm": 1.0089640617370605, "learning_rate": 0.00019765430448609315, "loss": 4.6964, "step": 270300 }, { "epoch": 0.5045031595595044, "grad_norm": 0.9127652645111084, "learning_rate": 0.00019765339983516498, "loss": 4.4906, "step": 270350 }, { "epoch": 0.5045964651188829, "grad_norm": 1.262275218963623, "learning_rate": 0.00019765249501189574, "loss": 4.8157, "step": 270400 }, { "epoch": 0.5046897706782615, "grad_norm": 0.7381452322006226, "learning_rate": 0.00019765159001628707, "loss": 4.4541, "step": 270450 }, { "epoch": 0.50478307623764, "grad_norm": 1.1626837253570557, "learning_rate": 0.00019765068484834048, "loss": 4.405, "step": 270500 }, { "epoch": 0.5048763817970184, "grad_norm": 1.3286311626434326, "learning_rate": 0.00019764977950805762, "loss": 4.6207, "step": 270550 }, { "epoch": 0.5049696873563969, "grad_norm": 0.9503090977668762, "learning_rate": 0.00019764887399544012, "loss": 4.4284, "step": 270600 }, { "epoch": 0.5050629929157754, "grad_norm": 1.025394082069397, "learning_rate": 0.00019764796831048948, "loss": 4.5853, "step": 270650 }, { "epoch": 0.5051562984751539, "grad_norm": 1.0864533185958862, "learning_rate": 0.00019764706245320741, "loss": 4.6884, "step": 270700 }, { "epoch": 0.5052496040345323, "grad_norm": 1.156358242034912, "learning_rate": 0.00019764615642359543, "loss": 4.5813, "step": 270750 }, { "epoch": 0.5053429095939109, "grad_norm": 1.1821043491363525, "learning_rate": 0.00019764525022165517, "loss": 4.6636, "step": 270800 }, { "epoch": 0.5054362151532894, "grad_norm": 1.121275782585144, "learning_rate": 0.00019764434384738825, "loss": 4.669, "step": 270850 }, { "epoch": 0.5055295207126679, "grad_norm": 0.9881389141082764, "learning_rate": 0.0001976434373007962, "loss": 4.6018, "step": 270900 }, { "epoch": 0.5056228262720464, "grad_norm": 1.1502811908721924, "learning_rate": 0.0001976425305818807, "loss": 4.6788, "step": 270950 }, { "epoch": 0.5057161318314248, "grad_norm": 0.8529852032661438, "learning_rate": 0.00019764162369064327, "loss": 4.6025, "step": 271000 }, { "epoch": 0.5058094373908033, "grad_norm": 0.8883637189865112, "learning_rate": 0.0001976407166270856, "loss": 4.5644, "step": 271050 }, { "epoch": 0.5059027429501818, "grad_norm": 1.2354636192321777, "learning_rate": 0.00019763980939120922, "loss": 4.4785, "step": 271100 }, { "epoch": 0.5059960485095604, "grad_norm": 0.7551540732383728, "learning_rate": 0.0001976389019830158, "loss": 4.4654, "step": 271150 }, { "epoch": 0.5060893540689388, "grad_norm": 1.2407974004745483, "learning_rate": 0.00019763799440250684, "loss": 4.5676, "step": 271200 }, { "epoch": 0.5061826596283173, "grad_norm": 1.2217525243759155, "learning_rate": 0.00019763708664968403, "loss": 4.6668, "step": 271250 }, { "epoch": 0.5062759651876958, "grad_norm": 1.1022509336471558, "learning_rate": 0.00019763617872454892, "loss": 4.5131, "step": 271300 }, { "epoch": 0.5063692707470743, "grad_norm": 0.9949933290481567, "learning_rate": 0.00019763527062710315, "loss": 4.6129, "step": 271350 }, { "epoch": 0.5064625763064527, "grad_norm": 0.9555957913398743, "learning_rate": 0.0001976343623573483, "loss": 4.4298, "step": 271400 }, { "epoch": 0.5065558818658312, "grad_norm": 0.9663435220718384, "learning_rate": 0.00019763345391528598, "loss": 4.5994, "step": 271450 }, { "epoch": 0.5066491874252098, "grad_norm": 1.056382417678833, "learning_rate": 0.0001976325453009178, "loss": 4.5895, "step": 271500 }, { "epoch": 0.5067424929845883, "grad_norm": 1.0316522121429443, "learning_rate": 0.00019763163651424537, "loss": 4.67, "step": 271550 }, { "epoch": 0.5068357985439668, "grad_norm": 1.1362007856369019, "learning_rate": 0.00019763072755527028, "loss": 4.7587, "step": 271600 }, { "epoch": 0.5069291041033452, "grad_norm": 1.2041569948196411, "learning_rate": 0.0001976298184239941, "loss": 4.6065, "step": 271650 }, { "epoch": 0.5070224096627237, "grad_norm": 0.9852795600891113, "learning_rate": 0.0001976289091204185, "loss": 4.4963, "step": 271700 }, { "epoch": 0.5071157152221022, "grad_norm": 0.8907833695411682, "learning_rate": 0.00019762799964454504, "loss": 4.6828, "step": 271750 }, { "epoch": 0.5072090207814807, "grad_norm": 1.104365348815918, "learning_rate": 0.00019762708999637534, "loss": 4.8107, "step": 271800 }, { "epoch": 0.5073023263408593, "grad_norm": 1.104337215423584, "learning_rate": 0.00019762618017591098, "loss": 4.536, "step": 271850 }, { "epoch": 0.5073956319002377, "grad_norm": 1.1071338653564453, "learning_rate": 0.00019762527018315365, "loss": 4.7241, "step": 271900 }, { "epoch": 0.5074889374596162, "grad_norm": 1.119663119316101, "learning_rate": 0.00019762436001810485, "loss": 4.5535, "step": 271950 }, { "epoch": 0.5075822430189947, "grad_norm": 0.6204298734664917, "learning_rate": 0.00019762344968076624, "loss": 4.4117, "step": 272000 }, { "epoch": 0.5076755485783732, "grad_norm": 1.1064519882202148, "learning_rate": 0.00019762253917113943, "loss": 4.7273, "step": 272050 }, { "epoch": 0.5077688541377516, "grad_norm": 0.985309362411499, "learning_rate": 0.000197621628489226, "loss": 4.5156, "step": 272100 }, { "epoch": 0.5078621596971301, "grad_norm": 1.3585158586502075, "learning_rate": 0.0001976207176350276, "loss": 4.5401, "step": 272150 }, { "epoch": 0.5079554652565087, "grad_norm": 1.3111861944198608, "learning_rate": 0.00019761980660854578, "loss": 4.6386, "step": 272200 }, { "epoch": 0.5080487708158872, "grad_norm": 1.2555516958236694, "learning_rate": 0.00019761889540978217, "loss": 4.6786, "step": 272250 }, { "epoch": 0.5081420763752656, "grad_norm": 1.0087488889694214, "learning_rate": 0.00019761798403873844, "loss": 4.5918, "step": 272300 }, { "epoch": 0.5082353819346441, "grad_norm": 1.2685153484344482, "learning_rate": 0.0001976170724954161, "loss": 4.7424, "step": 272350 }, { "epoch": 0.5083286874940226, "grad_norm": 0.9734815955162048, "learning_rate": 0.00019761616077981683, "loss": 4.81, "step": 272400 }, { "epoch": 0.5084219930534011, "grad_norm": 1.008641242980957, "learning_rate": 0.00019761524889194218, "loss": 4.3828, "step": 272450 }, { "epoch": 0.5085152986127796, "grad_norm": 0.7137326598167419, "learning_rate": 0.0001976143368317938, "loss": 4.5799, "step": 272500 }, { "epoch": 0.5086086041721581, "grad_norm": 1.0536925792694092, "learning_rate": 0.00019761342459937331, "loss": 4.4982, "step": 272550 }, { "epoch": 0.5087019097315366, "grad_norm": 1.1726641654968262, "learning_rate": 0.00019761251219468226, "loss": 4.5372, "step": 272600 }, { "epoch": 0.5087952152909151, "grad_norm": 0.9436787962913513, "learning_rate": 0.00019761159961772235, "loss": 4.7957, "step": 272650 }, { "epoch": 0.5088885208502936, "grad_norm": 1.3651800155639648, "learning_rate": 0.00019761068686849513, "loss": 4.5775, "step": 272700 }, { "epoch": 0.508981826409672, "grad_norm": 1.1570991277694702, "learning_rate": 0.0001976097739470022, "loss": 4.6765, "step": 272750 }, { "epoch": 0.5090751319690505, "grad_norm": 1.0292539596557617, "learning_rate": 0.0001976088608532452, "loss": 4.6702, "step": 272800 }, { "epoch": 0.509168437528429, "grad_norm": 1.0812225341796875, "learning_rate": 0.00019760794758722576, "loss": 4.486, "step": 272850 }, { "epoch": 0.5092617430878076, "grad_norm": 0.784271240234375, "learning_rate": 0.00019760703414894545, "loss": 4.4899, "step": 272900 }, { "epoch": 0.5093550486471861, "grad_norm": 1.1616487503051758, "learning_rate": 0.0001976061205384059, "loss": 4.909, "step": 272950 }, { "epoch": 0.5094483542065645, "grad_norm": 0.8870630860328674, "learning_rate": 0.00019760520675560873, "loss": 4.613, "step": 273000 }, { "epoch": 0.5094483542065645, "eval_loss": 4.762674808502197, "eval_runtime": 230.6095, "eval_samples_per_second": 11.309, "eval_steps_per_second": 11.309, "eval_tts_loss": 7.561236815871328, "step": 273000 }, { "epoch": 0.509541659765943, "grad_norm": 1.079444169998169, "learning_rate": 0.00019760429280055552, "loss": 4.5656, "step": 273050 }, { "epoch": 0.5096349653253215, "grad_norm": 1.0470081567764282, "learning_rate": 0.00019760337867324795, "loss": 4.4551, "step": 273100 }, { "epoch": 0.5097282708847, "grad_norm": 1.2054589986801147, "learning_rate": 0.00019760246437368757, "loss": 4.5344, "step": 273150 }, { "epoch": 0.5098215764440784, "grad_norm": 1.3342701196670532, "learning_rate": 0.000197601549901876, "loss": 4.4564, "step": 273200 }, { "epoch": 0.509914882003457, "grad_norm": 0.9556697010993958, "learning_rate": 0.00019760063525781486, "loss": 4.6027, "step": 273250 }, { "epoch": 0.5100081875628355, "grad_norm": 1.4421682357788086, "learning_rate": 0.00019759972044150579, "loss": 4.7149, "step": 273300 }, { "epoch": 0.510101493122214, "grad_norm": 1.053886890411377, "learning_rate": 0.0001975988054529504, "loss": 4.4437, "step": 273350 }, { "epoch": 0.5101947986815925, "grad_norm": 0.9334540963172913, "learning_rate": 0.00019759789029215026, "loss": 4.4206, "step": 273400 }, { "epoch": 0.5102881042409709, "grad_norm": 0.9920104146003723, "learning_rate": 0.00019759697495910702, "loss": 4.4509, "step": 273450 }, { "epoch": 0.5103814098003494, "grad_norm": 1.2183308601379395, "learning_rate": 0.00019759605945382234, "loss": 4.5994, "step": 273500 }, { "epoch": 0.5104747153597279, "grad_norm": 1.0757365226745605, "learning_rate": 0.00019759514377629773, "loss": 4.5886, "step": 273550 }, { "epoch": 0.5105680209191064, "grad_norm": 0.8739063143730164, "learning_rate": 0.0001975942279265349, "loss": 4.6799, "step": 273600 }, { "epoch": 0.510661326478485, "grad_norm": 1.4280765056610107, "learning_rate": 0.0001975933119045354, "loss": 4.6141, "step": 273650 }, { "epoch": 0.5107546320378634, "grad_norm": 0.8664491176605225, "learning_rate": 0.0001975923957103009, "loss": 4.5229, "step": 273700 }, { "epoch": 0.5108479375972419, "grad_norm": 1.3105825185775757, "learning_rate": 0.000197591479343833, "loss": 4.61, "step": 273750 }, { "epoch": 0.5109412431566204, "grad_norm": 1.0648022890090942, "learning_rate": 0.0001975905628051333, "loss": 4.5341, "step": 273800 }, { "epoch": 0.5110345487159988, "grad_norm": 1.1797337532043457, "learning_rate": 0.00019758964609420342, "loss": 4.6844, "step": 273850 }, { "epoch": 0.5111278542753773, "grad_norm": 0.919456958770752, "learning_rate": 0.000197588729211045, "loss": 4.6024, "step": 273900 }, { "epoch": 0.5112211598347558, "grad_norm": 1.0891369581222534, "learning_rate": 0.00019758781215565964, "loss": 4.6621, "step": 273950 }, { "epoch": 0.5113144653941344, "grad_norm": 1.0899498462677002, "learning_rate": 0.00019758689492804896, "loss": 4.6085, "step": 274000 }, { "epoch": 0.5114077709535129, "grad_norm": 0.9327117204666138, "learning_rate": 0.00019758597752821458, "loss": 4.4207, "step": 274050 }, { "epoch": 0.5115010765128913, "grad_norm": 1.0181550979614258, "learning_rate": 0.0001975850599561581, "loss": 4.5633, "step": 274100 }, { "epoch": 0.5115943820722698, "grad_norm": 1.0915321111679077, "learning_rate": 0.00019758414221188118, "loss": 4.7161, "step": 274150 }, { "epoch": 0.5116876876316483, "grad_norm": 1.1579314470291138, "learning_rate": 0.00019758322429538541, "loss": 4.6698, "step": 274200 }, { "epoch": 0.5117809931910268, "grad_norm": 1.0257996320724487, "learning_rate": 0.00019758230620667245, "loss": 4.762, "step": 274250 }, { "epoch": 0.5118742987504052, "grad_norm": 1.128806710243225, "learning_rate": 0.00019758138794574387, "loss": 4.7106, "step": 274300 }, { "epoch": 0.5119676043097838, "grad_norm": 1.203101396560669, "learning_rate": 0.00019758046951260129, "loss": 4.5531, "step": 274350 }, { "epoch": 0.5120609098691623, "grad_norm": 1.0775619745254517, "learning_rate": 0.0001975795509072464, "loss": 4.6696, "step": 274400 }, { "epoch": 0.5121542154285408, "grad_norm": 0.9243199825286865, "learning_rate": 0.00019757863212968072, "loss": 4.3668, "step": 274450 }, { "epoch": 0.5122475209879193, "grad_norm": 0.9755812883377075, "learning_rate": 0.00019757771317990597, "loss": 4.3343, "step": 274500 }, { "epoch": 0.5123408265472977, "grad_norm": 0.8233321309089661, "learning_rate": 0.0001975767940579237, "loss": 4.6018, "step": 274550 }, { "epoch": 0.5124341321066762, "grad_norm": 0.9258056879043579, "learning_rate": 0.00019757587476373556, "loss": 4.528, "step": 274600 }, { "epoch": 0.5125274376660547, "grad_norm": 0.988038957118988, "learning_rate": 0.00019757495529734316, "loss": 4.5291, "step": 274650 }, { "epoch": 0.5126207432254333, "grad_norm": 1.2431509494781494, "learning_rate": 0.00019757403565874815, "loss": 4.6244, "step": 274700 }, { "epoch": 0.5127140487848117, "grad_norm": 0.9292747974395752, "learning_rate": 0.00019757311584795212, "loss": 4.4623, "step": 274750 }, { "epoch": 0.5128073543441902, "grad_norm": 1.3109321594238281, "learning_rate": 0.0001975721958649567, "loss": 4.6401, "step": 274800 }, { "epoch": 0.5129006599035687, "grad_norm": 0.715238094329834, "learning_rate": 0.00019757127570976354, "loss": 4.8008, "step": 274850 }, { "epoch": 0.5129939654629472, "grad_norm": 1.2275428771972656, "learning_rate": 0.00019757035538237427, "loss": 4.5437, "step": 274900 }, { "epoch": 0.5130872710223257, "grad_norm": 0.7300710678100586, "learning_rate": 0.00019756943488279048, "loss": 4.3459, "step": 274950 }, { "epoch": 0.5131805765817041, "grad_norm": 1.039902925491333, "learning_rate": 0.00019756851421101377, "loss": 4.6869, "step": 275000 }, { "epoch": 0.5132738821410827, "grad_norm": 1.2344797849655151, "learning_rate": 0.00019756759336704582, "loss": 4.6589, "step": 275050 }, { "epoch": 0.5133671877004612, "grad_norm": 1.012023687362671, "learning_rate": 0.00019756667235088825, "loss": 4.6262, "step": 275100 }, { "epoch": 0.5134604932598397, "grad_norm": 0.8712413907051086, "learning_rate": 0.00019756575116254265, "loss": 4.4664, "step": 275150 }, { "epoch": 0.5135537988192181, "grad_norm": 1.4855979681015015, "learning_rate": 0.00019756482980201067, "loss": 4.4876, "step": 275200 }, { "epoch": 0.5136471043785966, "grad_norm": 0.8439715504646301, "learning_rate": 0.00019756390826929394, "loss": 4.5116, "step": 275250 }, { "epoch": 0.5137404099379751, "grad_norm": 1.170244574546814, "learning_rate": 0.00019756298656439407, "loss": 4.5469, "step": 275300 }, { "epoch": 0.5138337154973536, "grad_norm": 0.8794098496437073, "learning_rate": 0.00019756206468731272, "loss": 4.6529, "step": 275350 }, { "epoch": 0.5139270210567322, "grad_norm": 0.6970331072807312, "learning_rate": 0.00019756114263805148, "loss": 4.4499, "step": 275400 }, { "epoch": 0.5140203266161106, "grad_norm": 1.5795390605926514, "learning_rate": 0.00019756022041661196, "loss": 4.6471, "step": 275450 }, { "epoch": 0.5141136321754891, "grad_norm": 1.1724786758422852, "learning_rate": 0.00019755929802299583, "loss": 4.5844, "step": 275500 }, { "epoch": 0.5142069377348676, "grad_norm": 1.2464286088943481, "learning_rate": 0.00019755837545720474, "loss": 4.4931, "step": 275550 }, { "epoch": 0.5143002432942461, "grad_norm": 0.9109342694282532, "learning_rate": 0.00019755745271924028, "loss": 4.6377, "step": 275600 }, { "epoch": 0.5143935488536245, "grad_norm": 0.9874877333641052, "learning_rate": 0.00019755652980910405, "loss": 4.8261, "step": 275650 }, { "epoch": 0.514486854413003, "grad_norm": 1.1274828910827637, "learning_rate": 0.0001975556067267977, "loss": 4.6338, "step": 275700 }, { "epoch": 0.5145801599723816, "grad_norm": 1.0080583095550537, "learning_rate": 0.00019755468347232289, "loss": 4.4742, "step": 275750 }, { "epoch": 0.5146734655317601, "grad_norm": 0.6097507476806641, "learning_rate": 0.00019755376004568126, "loss": 4.5828, "step": 275800 }, { "epoch": 0.5147667710911386, "grad_norm": 1.115775465965271, "learning_rate": 0.00019755283644687434, "loss": 4.454, "step": 275850 }, { "epoch": 0.514860076650517, "grad_norm": 1.1590081453323364, "learning_rate": 0.00019755191267590387, "loss": 4.6587, "step": 275900 }, { "epoch": 0.5149533822098955, "grad_norm": 1.1706831455230713, "learning_rate": 0.00019755098873277144, "loss": 4.5509, "step": 275950 }, { "epoch": 0.515046687769274, "grad_norm": 1.1684876680374146, "learning_rate": 0.0001975500646174787, "loss": 4.4862, "step": 276000 }, { "epoch": 0.515046687769274, "eval_loss": 4.763826370239258, "eval_runtime": 231.8604, "eval_samples_per_second": 11.248, "eval_steps_per_second": 11.248, "eval_tts_loss": 7.552985722040529, "step": 276000 }, { "epoch": 0.5151399933286525, "grad_norm": 0.8337443470954895, "learning_rate": 0.00019754914033002724, "loss": 4.4825, "step": 276050 }, { "epoch": 0.515233298888031, "grad_norm": 0.8753382563591003, "learning_rate": 0.00019754821587041868, "loss": 4.5664, "step": 276100 }, { "epoch": 0.5153266044474095, "grad_norm": 0.86767578125, "learning_rate": 0.00019754729123865473, "loss": 4.2846, "step": 276150 }, { "epoch": 0.515419910006788, "grad_norm": 0.8605402708053589, "learning_rate": 0.00019754636643473693, "loss": 4.5609, "step": 276200 }, { "epoch": 0.5155132155661665, "grad_norm": 1.1940838098526, "learning_rate": 0.000197545441458667, "loss": 4.7445, "step": 276250 }, { "epoch": 0.515606521125545, "grad_norm": 0.7858842611312866, "learning_rate": 0.00019754451631044654, "loss": 4.6351, "step": 276300 }, { "epoch": 0.5156998266849234, "grad_norm": 1.128357172012329, "learning_rate": 0.0001975435909900771, "loss": 4.4674, "step": 276350 }, { "epoch": 0.5157931322443019, "grad_norm": 1.2002841234207153, "learning_rate": 0.00019754266549756045, "loss": 4.7851, "step": 276400 }, { "epoch": 0.5158864378036805, "grad_norm": 1.0065298080444336, "learning_rate": 0.00019754173983289812, "loss": 4.557, "step": 276450 }, { "epoch": 0.515979743363059, "grad_norm": 0.8537634015083313, "learning_rate": 0.0001975408139960918, "loss": 4.6682, "step": 276500 }, { "epoch": 0.5160730489224374, "grad_norm": 1.0204187631607056, "learning_rate": 0.00019753988798714312, "loss": 4.7636, "step": 276550 }, { "epoch": 0.5161663544818159, "grad_norm": 1.0077636241912842, "learning_rate": 0.0001975389618060537, "loss": 4.4058, "step": 276600 }, { "epoch": 0.5162596600411944, "grad_norm": 1.3681881427764893, "learning_rate": 0.00019753803545282517, "loss": 4.5586, "step": 276650 }, { "epoch": 0.5163529656005729, "grad_norm": 1.233417272567749, "learning_rate": 0.00019753710892745914, "loss": 4.7313, "step": 276700 }, { "epoch": 0.5164462711599513, "grad_norm": 1.226003885269165, "learning_rate": 0.00019753618222995728, "loss": 4.5046, "step": 276750 }, { "epoch": 0.5165395767193299, "grad_norm": 1.183275580406189, "learning_rate": 0.00019753525536032128, "loss": 4.6232, "step": 276800 }, { "epoch": 0.5166328822787084, "grad_norm": 1.11515212059021, "learning_rate": 0.00019753432831855265, "loss": 4.6647, "step": 276850 }, { "epoch": 0.5167261878380869, "grad_norm": 0.8872556686401367, "learning_rate": 0.0001975334011046531, "loss": 4.5093, "step": 276900 }, { "epoch": 0.5168194933974654, "grad_norm": 0.9870370626449585, "learning_rate": 0.0001975324737186243, "loss": 4.6153, "step": 276950 }, { "epoch": 0.5169127989568438, "grad_norm": 1.020976185798645, "learning_rate": 0.0001975315461604678, "loss": 4.4416, "step": 277000 }, { "epoch": 0.5170061045162223, "grad_norm": 0.7729629278182983, "learning_rate": 0.00019753061843018533, "loss": 4.5748, "step": 277050 }, { "epoch": 0.5170994100756008, "grad_norm": 1.0765986442565918, "learning_rate": 0.00019752969052777843, "loss": 4.6085, "step": 277100 }, { "epoch": 0.5171927156349794, "grad_norm": 1.0300875902175903, "learning_rate": 0.0001975287624532488, "loss": 4.5239, "step": 277150 }, { "epoch": 0.5172860211943578, "grad_norm": 0.9864205718040466, "learning_rate": 0.00019752783420659808, "loss": 4.6854, "step": 277200 }, { "epoch": 0.5173793267537363, "grad_norm": 0.8580088019371033, "learning_rate": 0.0001975269057878279, "loss": 4.55, "step": 277250 }, { "epoch": 0.5174726323131148, "grad_norm": 0.8540656566619873, "learning_rate": 0.00019752597719693987, "loss": 4.488, "step": 277300 }, { "epoch": 0.5175659378724933, "grad_norm": 0.6463267207145691, "learning_rate": 0.00019752504843393567, "loss": 4.5579, "step": 277350 }, { "epoch": 0.5176592434318718, "grad_norm": 1.2098805904388428, "learning_rate": 0.00019752411949881688, "loss": 4.7437, "step": 277400 }, { "epoch": 0.5177525489912502, "grad_norm": 0.8357241749763489, "learning_rate": 0.00019752319039158521, "loss": 4.5648, "step": 277450 }, { "epoch": 0.5178458545506288, "grad_norm": 0.9252105355262756, "learning_rate": 0.00019752226111224226, "loss": 4.6191, "step": 277500 }, { "epoch": 0.5179391601100073, "grad_norm": 0.9856184124946594, "learning_rate": 0.0001975213316607897, "loss": 4.6529, "step": 277550 }, { "epoch": 0.5180324656693858, "grad_norm": 1.0845974683761597, "learning_rate": 0.00019752040203722913, "loss": 4.4621, "step": 277600 }, { "epoch": 0.5181257712287642, "grad_norm": 0.7452815771102905, "learning_rate": 0.0001975194722415622, "loss": 4.5138, "step": 277650 }, { "epoch": 0.5182190767881427, "grad_norm": 1.0747730731964111, "learning_rate": 0.00019751854227379058, "loss": 4.678, "step": 277700 }, { "epoch": 0.5183123823475212, "grad_norm": 1.055133581161499, "learning_rate": 0.0001975176121339159, "loss": 4.653, "step": 277750 }, { "epoch": 0.5184056879068997, "grad_norm": 0.7984835505485535, "learning_rate": 0.0001975166818219398, "loss": 4.4666, "step": 277800 }, { "epoch": 0.5184989934662783, "grad_norm": 1.2567082643508911, "learning_rate": 0.00019751575133786385, "loss": 4.6693, "step": 277850 }, { "epoch": 0.5185922990256567, "grad_norm": 0.9567890167236328, "learning_rate": 0.00019751482068168982, "loss": 4.5737, "step": 277900 }, { "epoch": 0.5186856045850352, "grad_norm": 0.942671537399292, "learning_rate": 0.00019751388985341927, "loss": 4.7707, "step": 277950 }, { "epoch": 0.5187789101444137, "grad_norm": 1.07314932346344, "learning_rate": 0.00019751295885305384, "loss": 4.7499, "step": 278000 }, { "epoch": 0.5188722157037922, "grad_norm": 0.8657567501068115, "learning_rate": 0.00019751202768059526, "loss": 4.5627, "step": 278050 }, { "epoch": 0.5189655212631706, "grad_norm": 0.9359160661697388, "learning_rate": 0.00019751109633604506, "loss": 4.5867, "step": 278100 }, { "epoch": 0.5190588268225491, "grad_norm": 1.0121476650238037, "learning_rate": 0.00019751016481940492, "loss": 4.7104, "step": 278150 }, { "epoch": 0.5191521323819277, "grad_norm": 1.4921369552612305, "learning_rate": 0.00019750923313067653, "loss": 4.5998, "step": 278200 }, { "epoch": 0.5192454379413062, "grad_norm": 1.2651481628417969, "learning_rate": 0.00019750830126986148, "loss": 4.3789, "step": 278250 }, { "epoch": 0.5193387435006847, "grad_norm": 0.9953886866569519, "learning_rate": 0.00019750736923696146, "loss": 4.6677, "step": 278300 }, { "epoch": 0.5194320490600631, "grad_norm": 0.7058857083320618, "learning_rate": 0.00019750643703197806, "loss": 4.5534, "step": 278350 }, { "epoch": 0.5195253546194416, "grad_norm": 1.0814391374588013, "learning_rate": 0.00019750550465491297, "loss": 4.485, "step": 278400 }, { "epoch": 0.5196186601788201, "grad_norm": 0.9021806716918945, "learning_rate": 0.00019750457210576783, "loss": 4.485, "step": 278450 }, { "epoch": 0.5197119657381986, "grad_norm": 0.9229526519775391, "learning_rate": 0.00019750363938454425, "loss": 4.4707, "step": 278500 }, { "epoch": 0.5198052712975771, "grad_norm": 1.1465047597885132, "learning_rate": 0.00019750270649124393, "loss": 4.4243, "step": 278550 }, { "epoch": 0.5198985768569556, "grad_norm": 0.8875910639762878, "learning_rate": 0.00019750177342586848, "loss": 4.5245, "step": 278600 }, { "epoch": 0.5199918824163341, "grad_norm": 1.1508044004440308, "learning_rate": 0.00019750084018841955, "loss": 4.7269, "step": 278650 }, { "epoch": 0.5200851879757126, "grad_norm": 0.9982178211212158, "learning_rate": 0.00019749990677889878, "loss": 4.806, "step": 278700 }, { "epoch": 0.520178493535091, "grad_norm": 0.7332342267036438, "learning_rate": 0.00019749897319730787, "loss": 4.5537, "step": 278750 }, { "epoch": 0.5202717990944695, "grad_norm": 1.0460069179534912, "learning_rate": 0.0001974980394436484, "loss": 4.4889, "step": 278800 }, { "epoch": 0.520365104653848, "grad_norm": 1.1912788152694702, "learning_rate": 0.00019749710551792205, "loss": 4.7645, "step": 278850 }, { "epoch": 0.5204584102132265, "grad_norm": 1.0818390846252441, "learning_rate": 0.00019749617142013047, "loss": 4.7387, "step": 278900 }, { "epoch": 0.5205517157726051, "grad_norm": 1.156233549118042, "learning_rate": 0.00019749523715027528, "loss": 4.6504, "step": 278950 }, { "epoch": 0.5206450213319835, "grad_norm": 0.9997411966323853, "learning_rate": 0.00019749430270835818, "loss": 4.4489, "step": 279000 }, { "epoch": 0.5206450213319835, "eval_loss": 4.75526237487793, "eval_runtime": 229.0585, "eval_samples_per_second": 11.386, "eval_steps_per_second": 11.386, "eval_tts_loss": 7.540098349903738, "step": 279000 }, { "epoch": 0.520738326891362, "grad_norm": 0.8896755576133728, "learning_rate": 0.00019749336809438077, "loss": 4.7217, "step": 279050 }, { "epoch": 0.5208316324507405, "grad_norm": 0.9033337831497192, "learning_rate": 0.00019749243330834473, "loss": 4.6086, "step": 279100 }, { "epoch": 0.520924938010119, "grad_norm": 0.7716773152351379, "learning_rate": 0.0001974914983502517, "loss": 4.7269, "step": 279150 }, { "epoch": 0.5210182435694974, "grad_norm": 1.14053213596344, "learning_rate": 0.00019749056322010332, "loss": 4.6691, "step": 279200 }, { "epoch": 0.5211115491288759, "grad_norm": 1.0877877473831177, "learning_rate": 0.00019748962791790126, "loss": 4.4902, "step": 279250 }, { "epoch": 0.5212048546882545, "grad_norm": 0.9544693231582642, "learning_rate": 0.00019748869244364716, "loss": 4.5944, "step": 279300 }, { "epoch": 0.521298160247633, "grad_norm": 1.047554850578308, "learning_rate": 0.0001974877567973427, "loss": 4.7411, "step": 279350 }, { "epoch": 0.5213914658070115, "grad_norm": 0.8173419237136841, "learning_rate": 0.00019748682097898946, "loss": 4.4421, "step": 279400 }, { "epoch": 0.5214847713663899, "grad_norm": 1.2062064409255981, "learning_rate": 0.00019748588498858912, "loss": 4.6812, "step": 279450 }, { "epoch": 0.5215780769257684, "grad_norm": 1.0467079877853394, "learning_rate": 0.00019748494882614338, "loss": 4.5961, "step": 279500 }, { "epoch": 0.5216713824851469, "grad_norm": 0.8883413076400757, "learning_rate": 0.00019748401249165384, "loss": 4.6766, "step": 279550 }, { "epoch": 0.5217646880445254, "grad_norm": 0.8880034685134888, "learning_rate": 0.0001974830759851222, "loss": 4.2964, "step": 279600 }, { "epoch": 0.521857993603904, "grad_norm": 1.1952457427978516, "learning_rate": 0.00019748213930655006, "loss": 4.663, "step": 279650 }, { "epoch": 0.5219512991632824, "grad_norm": 0.9282484650611877, "learning_rate": 0.00019748120245593911, "loss": 4.6977, "step": 279700 }, { "epoch": 0.5220446047226609, "grad_norm": 1.0370795726776123, "learning_rate": 0.000197480265433291, "loss": 4.5207, "step": 279750 }, { "epoch": 0.5221379102820394, "grad_norm": 1.0368770360946655, "learning_rate": 0.00019747932823860733, "loss": 4.4436, "step": 279800 }, { "epoch": 0.5222312158414179, "grad_norm": 1.122261881828308, "learning_rate": 0.00019747839087188985, "loss": 4.6599, "step": 279850 }, { "epoch": 0.5223245214007963, "grad_norm": 1.1524388790130615, "learning_rate": 0.00019747745333314012, "loss": 4.4843, "step": 279900 }, { "epoch": 0.5224178269601748, "grad_norm": 1.1980907917022705, "learning_rate": 0.00019747651562235984, "loss": 4.6278, "step": 279950 }, { "epoch": 0.5225111325195534, "grad_norm": 1.0899757146835327, "learning_rate": 0.0001974755777395507, "loss": 4.5952, "step": 280000 }, { "epoch": 0.5226044380789319, "grad_norm": 1.0352386236190796, "learning_rate": 0.0001974746396847143, "loss": 4.5709, "step": 280050 }, { "epoch": 0.5226977436383103, "grad_norm": 1.149746298789978, "learning_rate": 0.00019747370145785226, "loss": 4.4802, "step": 280100 }, { "epoch": 0.5227910491976888, "grad_norm": 1.0010826587677002, "learning_rate": 0.00019747276305896632, "loss": 4.7044, "step": 280150 }, { "epoch": 0.5228843547570673, "grad_norm": 1.0683485269546509, "learning_rate": 0.00019747182448805816, "loss": 4.5503, "step": 280200 }, { "epoch": 0.5229776603164458, "grad_norm": 1.128632664680481, "learning_rate": 0.00019747088574512933, "loss": 4.5131, "step": 280250 }, { "epoch": 0.5230709658758242, "grad_norm": 1.312676191329956, "learning_rate": 0.00019746994683018154, "loss": 4.6438, "step": 280300 }, { "epoch": 0.5231642714352028, "grad_norm": 1.1339898109436035, "learning_rate": 0.00019746900774321644, "loss": 4.4215, "step": 280350 }, { "epoch": 0.5232575769945813, "grad_norm": 1.3833768367767334, "learning_rate": 0.00019746806848423568, "loss": 4.4671, "step": 280400 }, { "epoch": 0.5233508825539598, "grad_norm": 1.1775107383728027, "learning_rate": 0.00019746712905324097, "loss": 4.6874, "step": 280450 }, { "epoch": 0.5234441881133383, "grad_norm": 0.881015419960022, "learning_rate": 0.00019746618945023388, "loss": 4.6857, "step": 280500 }, { "epoch": 0.5235374936727167, "grad_norm": 1.1742804050445557, "learning_rate": 0.00019746524967521615, "loss": 4.5829, "step": 280550 }, { "epoch": 0.5236307992320952, "grad_norm": 1.0308231115341187, "learning_rate": 0.00019746430972818942, "loss": 4.4821, "step": 280600 }, { "epoch": 0.5237241047914737, "grad_norm": 0.9045799374580383, "learning_rate": 0.00019746336960915528, "loss": 4.6749, "step": 280650 }, { "epoch": 0.5238174103508523, "grad_norm": 1.095934510231018, "learning_rate": 0.00019746242931811548, "loss": 4.6171, "step": 280700 }, { "epoch": 0.5239107159102308, "grad_norm": 1.1799335479736328, "learning_rate": 0.00019746148885507162, "loss": 4.3884, "step": 280750 }, { "epoch": 0.5240040214696092, "grad_norm": 0.9894862771034241, "learning_rate": 0.0001974605482200254, "loss": 4.65, "step": 280800 }, { "epoch": 0.5240973270289877, "grad_norm": 0.993606448173523, "learning_rate": 0.00019745960741297843, "loss": 4.3739, "step": 280850 }, { "epoch": 0.5241906325883662, "grad_norm": 1.1276586055755615, "learning_rate": 0.00019745866643393245, "loss": 4.9291, "step": 280900 }, { "epoch": 0.5242839381477447, "grad_norm": 1.249640941619873, "learning_rate": 0.00019745772528288903, "loss": 4.5495, "step": 280950 }, { "epoch": 0.5243772437071231, "grad_norm": 0.9969247579574585, "learning_rate": 0.00019745678395984988, "loss": 4.7876, "step": 281000 }, { "epoch": 0.5244705492665017, "grad_norm": 0.9947354793548584, "learning_rate": 0.00019745584246481666, "loss": 4.6367, "step": 281050 }, { "epoch": 0.5245638548258802, "grad_norm": 1.1620121002197266, "learning_rate": 0.00019745490079779105, "loss": 4.6612, "step": 281100 }, { "epoch": 0.5246571603852587, "grad_norm": 0.8015241026878357, "learning_rate": 0.00019745395895877466, "loss": 4.5458, "step": 281150 }, { "epoch": 0.5247504659446371, "grad_norm": 1.141109585762024, "learning_rate": 0.00019745301694776916, "loss": 4.5927, "step": 281200 }, { "epoch": 0.5248437715040156, "grad_norm": 0.8895716071128845, "learning_rate": 0.0001974520747647763, "loss": 4.6357, "step": 281250 }, { "epoch": 0.5249370770633941, "grad_norm": 1.0158172845840454, "learning_rate": 0.00019745113240979763, "loss": 4.738, "step": 281300 }, { "epoch": 0.5250303826227726, "grad_norm": 1.0733691453933716, "learning_rate": 0.00019745018988283483, "loss": 4.5997, "step": 281350 }, { "epoch": 0.5251236881821512, "grad_norm": 1.1334929466247559, "learning_rate": 0.00019744924718388964, "loss": 4.6377, "step": 281400 }, { "epoch": 0.5252169937415296, "grad_norm": 1.1703345775604248, "learning_rate": 0.00019744830431296366, "loss": 4.6239, "step": 281450 }, { "epoch": 0.5253102993009081, "grad_norm": 1.0354031324386597, "learning_rate": 0.00019744736127005856, "loss": 4.5597, "step": 281500 }, { "epoch": 0.5254036048602866, "grad_norm": 1.0485466718673706, "learning_rate": 0.00019744641805517603, "loss": 4.5579, "step": 281550 }, { "epoch": 0.5254969104196651, "grad_norm": 1.360655665397644, "learning_rate": 0.0001974454746683177, "loss": 4.4888, "step": 281600 }, { "epoch": 0.5255902159790435, "grad_norm": 1.0404666662216187, "learning_rate": 0.0001974445311094853, "loss": 4.5823, "step": 281650 }, { "epoch": 0.525683521538422, "grad_norm": 1.0680186748504639, "learning_rate": 0.0001974435873786804, "loss": 4.5325, "step": 281700 }, { "epoch": 0.5257768270978006, "grad_norm": 1.1032402515411377, "learning_rate": 0.0001974426434759047, "loss": 4.5647, "step": 281750 }, { "epoch": 0.5258701326571791, "grad_norm": 0.9431357383728027, "learning_rate": 0.0001974416994011599, "loss": 4.5412, "step": 281800 }, { "epoch": 0.5259634382165576, "grad_norm": 1.0254426002502441, "learning_rate": 0.00019744075515444766, "loss": 4.4711, "step": 281850 }, { "epoch": 0.526056743775936, "grad_norm": 0.9435511231422424, "learning_rate": 0.00019743981073576966, "loss": 4.5718, "step": 281900 }, { "epoch": 0.5261500493353145, "grad_norm": 1.1502196788787842, "learning_rate": 0.00019743886614512747, "loss": 4.5823, "step": 281950 }, { "epoch": 0.526243354894693, "grad_norm": 1.0509811639785767, "learning_rate": 0.00019743792138252286, "loss": 4.7095, "step": 282000 }, { "epoch": 0.526243354894693, "eval_loss": 4.762018203735352, "eval_runtime": 231.817, "eval_samples_per_second": 11.25, "eval_steps_per_second": 11.25, "eval_tts_loss": 7.497885950622463, "step": 282000 }, { "epoch": 0.5263366604540715, "grad_norm": 0.8279403448104858, "learning_rate": 0.00019743697644795745, "loss": 4.4441, "step": 282050 }, { "epoch": 0.52642996601345, "grad_norm": 1.0076203346252441, "learning_rate": 0.00019743603134143295, "loss": 4.5973, "step": 282100 }, { "epoch": 0.5265232715728285, "grad_norm": 0.8520145416259766, "learning_rate": 0.000197435086062951, "loss": 4.7364, "step": 282150 }, { "epoch": 0.526616577132207, "grad_norm": 1.1191004514694214, "learning_rate": 0.00019743414061251324, "loss": 4.529, "step": 282200 }, { "epoch": 0.5267098826915855, "grad_norm": 1.0632494688034058, "learning_rate": 0.00019743319499012139, "loss": 4.4006, "step": 282250 }, { "epoch": 0.526803188250964, "grad_norm": 0.871149480342865, "learning_rate": 0.00019743224919577707, "loss": 4.6199, "step": 282300 }, { "epoch": 0.5268964938103424, "grad_norm": 1.0625333786010742, "learning_rate": 0.00019743130322948197, "loss": 4.5028, "step": 282350 }, { "epoch": 0.5269897993697209, "grad_norm": 1.019814133644104, "learning_rate": 0.00019743035709123778, "loss": 4.6573, "step": 282400 }, { "epoch": 0.5270831049290995, "grad_norm": 0.9306842088699341, "learning_rate": 0.00019742941078104616, "loss": 4.7412, "step": 282450 }, { "epoch": 0.527176410488478, "grad_norm": 1.135619878768921, "learning_rate": 0.0001974284642989088, "loss": 4.6812, "step": 282500 }, { "epoch": 0.5272697160478564, "grad_norm": 1.3619680404663086, "learning_rate": 0.0001974275176448273, "loss": 4.5811, "step": 282550 }, { "epoch": 0.5273630216072349, "grad_norm": 0.803629994392395, "learning_rate": 0.00019742657081880336, "loss": 4.776, "step": 282600 }, { "epoch": 0.5274563271666134, "grad_norm": 0.9954638481140137, "learning_rate": 0.0001974256238208387, "loss": 4.7264, "step": 282650 }, { "epoch": 0.5275496327259919, "grad_norm": 1.0035852193832397, "learning_rate": 0.00019742467665093493, "loss": 4.6593, "step": 282700 }, { "epoch": 0.5276429382853703, "grad_norm": 1.4352021217346191, "learning_rate": 0.00019742372930909379, "loss": 4.5499, "step": 282750 }, { "epoch": 0.5277362438447489, "grad_norm": 1.2543692588806152, "learning_rate": 0.00019742278179531687, "loss": 4.6667, "step": 282800 }, { "epoch": 0.5278295494041274, "grad_norm": 0.8242788314819336, "learning_rate": 0.0001974218341096059, "loss": 4.4721, "step": 282850 }, { "epoch": 0.5279228549635059, "grad_norm": 0.9672014117240906, "learning_rate": 0.00019742088625196251, "loss": 4.6143, "step": 282900 }, { "epoch": 0.5280161605228844, "grad_norm": 0.9670060873031616, "learning_rate": 0.00019741993822238843, "loss": 4.4835, "step": 282950 }, { "epoch": 0.5281094660822628, "grad_norm": 1.4916218519210815, "learning_rate": 0.0001974189900208853, "loss": 4.6265, "step": 283000 }, { "epoch": 0.5282027716416413, "grad_norm": 0.8706250786781311, "learning_rate": 0.00019741804164745476, "loss": 4.6028, "step": 283050 }, { "epoch": 0.5282960772010198, "grad_norm": 0.9748098254203796, "learning_rate": 0.00019741709310209854, "loss": 4.5401, "step": 283100 }, { "epoch": 0.5283893827603984, "grad_norm": 1.1996750831604004, "learning_rate": 0.0001974161443848183, "loss": 4.6417, "step": 283150 }, { "epoch": 0.5284826883197768, "grad_norm": 0.9673647880554199, "learning_rate": 0.0001974151954956157, "loss": 4.559, "step": 283200 }, { "epoch": 0.5285759938791553, "grad_norm": 0.6728837490081787, "learning_rate": 0.00019741424643449242, "loss": 4.497, "step": 283250 }, { "epoch": 0.5286692994385338, "grad_norm": 1.0573639869689941, "learning_rate": 0.00019741329720145015, "loss": 4.6833, "step": 283300 }, { "epoch": 0.5287626049979123, "grad_norm": 1.2526932954788208, "learning_rate": 0.00019741234779649053, "loss": 4.7223, "step": 283350 }, { "epoch": 0.5288559105572908, "grad_norm": 1.370250940322876, "learning_rate": 0.00019741139821961525, "loss": 4.5744, "step": 283400 }, { "epoch": 0.5289492161166692, "grad_norm": 1.1258753538131714, "learning_rate": 0.00019741044847082598, "loss": 4.734, "step": 283450 }, { "epoch": 0.5290425216760478, "grad_norm": 0.8937044143676758, "learning_rate": 0.00019740949855012445, "loss": 4.3539, "step": 283500 }, { "epoch": 0.5291358272354263, "grad_norm": 1.048108696937561, "learning_rate": 0.00019740854845751228, "loss": 4.4583, "step": 283550 }, { "epoch": 0.5292291327948048, "grad_norm": 0.9477057456970215, "learning_rate": 0.00019740759819299118, "loss": 4.553, "step": 283600 }, { "epoch": 0.5293224383541832, "grad_norm": 1.0268933773040771, "learning_rate": 0.00019740664775656276, "loss": 4.5117, "step": 283650 }, { "epoch": 0.5294157439135617, "grad_norm": 1.1293946504592896, "learning_rate": 0.00019740569714822876, "loss": 4.7847, "step": 283700 }, { "epoch": 0.5295090494729402, "grad_norm": 1.2568365335464478, "learning_rate": 0.00019740474636799089, "loss": 4.5597, "step": 283750 }, { "epoch": 0.5296023550323187, "grad_norm": 1.0086328983306885, "learning_rate": 0.00019740379541585075, "loss": 4.5726, "step": 283800 }, { "epoch": 0.5296956605916973, "grad_norm": 0.9730863571166992, "learning_rate": 0.00019740284429181003, "loss": 4.7144, "step": 283850 }, { "epoch": 0.5297889661510757, "grad_norm": 1.4535603523254395, "learning_rate": 0.00019740189299587044, "loss": 4.7222, "step": 283900 }, { "epoch": 0.5298822717104542, "grad_norm": 0.6582089066505432, "learning_rate": 0.00019740094152803365, "loss": 4.7405, "step": 283950 }, { "epoch": 0.5299755772698327, "grad_norm": 1.1787588596343994, "learning_rate": 0.00019739998988830136, "loss": 4.5965, "step": 284000 }, { "epoch": 0.5300688828292112, "grad_norm": 0.9802767038345337, "learning_rate": 0.00019739903807667518, "loss": 4.3648, "step": 284050 }, { "epoch": 0.5301621883885896, "grad_norm": 0.9081062078475952, "learning_rate": 0.00019739808609315688, "loss": 4.4767, "step": 284100 }, { "epoch": 0.5302554939479681, "grad_norm": 1.005858063697815, "learning_rate": 0.00019739713393774805, "loss": 4.5397, "step": 284150 }, { "epoch": 0.5303487995073466, "grad_norm": 0.6593231558799744, "learning_rate": 0.00019739618161045047, "loss": 4.5068, "step": 284200 }, { "epoch": 0.5304421050667252, "grad_norm": 0.8046698570251465, "learning_rate": 0.00019739522911126574, "loss": 4.6746, "step": 284250 }, { "epoch": 0.5305354106261037, "grad_norm": 1.3987061977386475, "learning_rate": 0.00019739427644019556, "loss": 4.469, "step": 284300 }, { "epoch": 0.5306287161854821, "grad_norm": 1.4002022743225098, "learning_rate": 0.00019739332359724164, "loss": 4.5972, "step": 284350 }, { "epoch": 0.5307220217448606, "grad_norm": 1.0923137664794922, "learning_rate": 0.00019739237058240562, "loss": 4.6223, "step": 284400 }, { "epoch": 0.5308153273042391, "grad_norm": 0.9077520370483398, "learning_rate": 0.00019739141739568921, "loss": 4.6044, "step": 284450 }, { "epoch": 0.5309086328636176, "grad_norm": 1.186133861541748, "learning_rate": 0.0001973904640370941, "loss": 4.8774, "step": 284500 }, { "epoch": 0.531001938422996, "grad_norm": 1.0076979398727417, "learning_rate": 0.0001973895105066219, "loss": 4.4713, "step": 284550 }, { "epoch": 0.5310952439823746, "grad_norm": 0.9802193641662598, "learning_rate": 0.00019738855680427443, "loss": 4.4804, "step": 284600 }, { "epoch": 0.5311885495417531, "grad_norm": 1.033267855644226, "learning_rate": 0.00019738760293005325, "loss": 4.6651, "step": 284650 }, { "epoch": 0.5312818551011316, "grad_norm": 0.9412552714347839, "learning_rate": 0.0001973866488839601, "loss": 4.8155, "step": 284700 }, { "epoch": 0.53137516066051, "grad_norm": 1.0584665536880493, "learning_rate": 0.00019738569466599664, "loss": 4.6357, "step": 284750 }, { "epoch": 0.5314684662198885, "grad_norm": 1.006569266319275, "learning_rate": 0.00019738474027616455, "loss": 4.4633, "step": 284800 }, { "epoch": 0.531561771779267, "grad_norm": 1.1155530214309692, "learning_rate": 0.00019738378571446555, "loss": 4.6122, "step": 284850 }, { "epoch": 0.5316550773386455, "grad_norm": 0.9938474297523499, "learning_rate": 0.0001973828309809013, "loss": 4.5603, "step": 284900 }, { "epoch": 0.5317483828980241, "grad_norm": 1.090856909751892, "learning_rate": 0.0001973818760754735, "loss": 4.4701, "step": 284950 }, { "epoch": 0.5318416884574025, "grad_norm": 1.146928310394287, "learning_rate": 0.0001973809209981838, "loss": 4.6746, "step": 285000 }, { "epoch": 0.5318416884574025, "eval_loss": 4.75473165512085, "eval_runtime": 230.3186, "eval_samples_per_second": 11.323, "eval_steps_per_second": 11.323, "eval_tts_loss": 7.53693535750549, "step": 285000 }, { "epoch": 0.531934994016781, "grad_norm": 0.9732300639152527, "learning_rate": 0.00019737996574903393, "loss": 4.5959, "step": 285050 }, { "epoch": 0.5320282995761595, "grad_norm": 1.0271157026290894, "learning_rate": 0.00019737901032802554, "loss": 4.5514, "step": 285100 }, { "epoch": 0.532121605135538, "grad_norm": 1.106013298034668, "learning_rate": 0.00019737805473516034, "loss": 4.7324, "step": 285150 }, { "epoch": 0.5322149106949164, "grad_norm": 0.9477126598358154, "learning_rate": 0.00019737709897044, "loss": 4.6364, "step": 285200 }, { "epoch": 0.5323082162542949, "grad_norm": 1.0403019189834595, "learning_rate": 0.00019737614303386622, "loss": 4.802, "step": 285250 }, { "epoch": 0.5324015218136735, "grad_norm": 1.1093212366104126, "learning_rate": 0.0001973751869254407, "loss": 4.4366, "step": 285300 }, { "epoch": 0.532494827373052, "grad_norm": 0.9458768367767334, "learning_rate": 0.00019737423064516507, "loss": 4.5341, "step": 285350 }, { "epoch": 0.5325881329324305, "grad_norm": 0.8419497609138489, "learning_rate": 0.00019737327419304106, "loss": 4.3237, "step": 285400 }, { "epoch": 0.5326814384918089, "grad_norm": 0.7255857586860657, "learning_rate": 0.0001973723175690704, "loss": 4.5856, "step": 285450 }, { "epoch": 0.5327747440511874, "grad_norm": 1.0696274042129517, "learning_rate": 0.0001973713607732547, "loss": 4.5296, "step": 285500 }, { "epoch": 0.5328680496105659, "grad_norm": 1.033064603805542, "learning_rate": 0.0001973704038055957, "loss": 4.64, "step": 285550 }, { "epoch": 0.5329613551699444, "grad_norm": 1.201291799545288, "learning_rate": 0.00019736944666609503, "loss": 4.6272, "step": 285600 }, { "epoch": 0.533054660729323, "grad_norm": 1.2754902839660645, "learning_rate": 0.00019736848935475446, "loss": 4.7426, "step": 285650 }, { "epoch": 0.5331479662887014, "grad_norm": 1.2380539178848267, "learning_rate": 0.0001973675318715756, "loss": 4.5593, "step": 285700 }, { "epoch": 0.5332412718480799, "grad_norm": 0.910490095615387, "learning_rate": 0.00019736657421656019, "loss": 4.4716, "step": 285750 }, { "epoch": 0.5333345774074584, "grad_norm": 0.6188794374465942, "learning_rate": 0.0001973656163897099, "loss": 4.5351, "step": 285800 }, { "epoch": 0.5334278829668369, "grad_norm": 1.1422240734100342, "learning_rate": 0.00019736465839102646, "loss": 4.5875, "step": 285850 }, { "epoch": 0.5335211885262153, "grad_norm": 1.2172739505767822, "learning_rate": 0.00019736370022051153, "loss": 4.5143, "step": 285900 }, { "epoch": 0.5336144940855938, "grad_norm": 1.0027328729629517, "learning_rate": 0.00019736274187816678, "loss": 4.6219, "step": 285950 }, { "epoch": 0.5337077996449724, "grad_norm": 0.7851713299751282, "learning_rate": 0.00019736178336399394, "loss": 4.5339, "step": 286000 }, { "epoch": 0.5338011052043509, "grad_norm": 0.8879555463790894, "learning_rate": 0.00019736082467799467, "loss": 4.5193, "step": 286050 }, { "epoch": 0.5338944107637293, "grad_norm": 0.8688876032829285, "learning_rate": 0.0001973598658201707, "loss": 4.5729, "step": 286100 }, { "epoch": 0.5339877163231078, "grad_norm": 1.100091814994812, "learning_rate": 0.00019735890679052368, "loss": 4.7068, "step": 286150 }, { "epoch": 0.5340810218824863, "grad_norm": 1.1109236478805542, "learning_rate": 0.0001973579475890553, "loss": 4.6203, "step": 286200 }, { "epoch": 0.5341743274418648, "grad_norm": 0.8393378853797913, "learning_rate": 0.0001973569882157673, "loss": 4.6772, "step": 286250 }, { "epoch": 0.5342676330012432, "grad_norm": 1.0183576345443726, "learning_rate": 0.00019735602867066134, "loss": 4.716, "step": 286300 }, { "epoch": 0.5343609385606218, "grad_norm": 1.0126652717590332, "learning_rate": 0.00019735506895373912, "loss": 4.6462, "step": 286350 }, { "epoch": 0.5344542441200003, "grad_norm": 0.8533194661140442, "learning_rate": 0.00019735410906500232, "loss": 4.542, "step": 286400 }, { "epoch": 0.5345475496793788, "grad_norm": 0.7949962019920349, "learning_rate": 0.00019735314900445268, "loss": 4.5383, "step": 286450 }, { "epoch": 0.5346408552387573, "grad_norm": 1.039992094039917, "learning_rate": 0.00019735218877209182, "loss": 4.4467, "step": 286500 }, { "epoch": 0.5347341607981357, "grad_norm": 0.6137852668762207, "learning_rate": 0.0001973512283679215, "loss": 4.7501, "step": 286550 }, { "epoch": 0.5348274663575142, "grad_norm": 1.16130793094635, "learning_rate": 0.00019735026779194338, "loss": 4.5853, "step": 286600 }, { "epoch": 0.5349207719168927, "grad_norm": 1.0255588293075562, "learning_rate": 0.0001973493070441592, "loss": 4.5439, "step": 286650 }, { "epoch": 0.5350140774762713, "grad_norm": 0.9397153258323669, "learning_rate": 0.00019734834612457057, "loss": 4.4539, "step": 286700 }, { "epoch": 0.5351073830356498, "grad_norm": 0.7592172622680664, "learning_rate": 0.00019734738503317928, "loss": 4.4875, "step": 286750 }, { "epoch": 0.5352006885950282, "grad_norm": 1.4141420125961304, "learning_rate": 0.00019734642376998695, "loss": 4.4221, "step": 286800 }, { "epoch": 0.5352939941544067, "grad_norm": 0.8302573561668396, "learning_rate": 0.00019734546233499532, "loss": 4.3809, "step": 286850 }, { "epoch": 0.5353872997137852, "grad_norm": 1.0930051803588867, "learning_rate": 0.00019734450072820612, "loss": 4.6485, "step": 286900 }, { "epoch": 0.5354806052731637, "grad_norm": 1.100099802017212, "learning_rate": 0.00019734353894962096, "loss": 4.5825, "step": 286950 }, { "epoch": 0.5355739108325421, "grad_norm": 1.1723564863204956, "learning_rate": 0.00019734257699924158, "loss": 4.3963, "step": 287000 }, { "epoch": 0.5356672163919207, "grad_norm": 1.0333951711654663, "learning_rate": 0.0001973416148770697, "loss": 4.4069, "step": 287050 }, { "epoch": 0.5357605219512992, "grad_norm": 1.0501785278320312, "learning_rate": 0.00019734065258310702, "loss": 4.5181, "step": 287100 }, { "epoch": 0.5358538275106777, "grad_norm": 1.1242258548736572, "learning_rate": 0.00019733969011735518, "loss": 4.3266, "step": 287150 }, { "epoch": 0.5359471330700561, "grad_norm": 1.1503238677978516, "learning_rate": 0.00019733872747981592, "loss": 4.5229, "step": 287200 }, { "epoch": 0.5360404386294346, "grad_norm": 0.9693351984024048, "learning_rate": 0.00019733776467049095, "loss": 4.5938, "step": 287250 }, { "epoch": 0.5361337441888131, "grad_norm": 0.7347463965415955, "learning_rate": 0.00019733680168938192, "loss": 4.4832, "step": 287300 }, { "epoch": 0.5362270497481916, "grad_norm": 1.387575626373291, "learning_rate": 0.00019733583853649058, "loss": 4.5632, "step": 287350 }, { "epoch": 0.5363203553075702, "grad_norm": 1.0806927680969238, "learning_rate": 0.00019733487521181862, "loss": 4.7334, "step": 287400 }, { "epoch": 0.5364136608669486, "grad_norm": 1.3513052463531494, "learning_rate": 0.00019733391171536773, "loss": 4.571, "step": 287450 }, { "epoch": 0.5365069664263271, "grad_norm": 1.1874183416366577, "learning_rate": 0.00019733294804713959, "loss": 4.5931, "step": 287500 }, { "epoch": 0.5366002719857056, "grad_norm": 1.212813138961792, "learning_rate": 0.00019733198420713595, "loss": 4.5991, "step": 287550 }, { "epoch": 0.5366935775450841, "grad_norm": 0.9041852355003357, "learning_rate": 0.00019733102019535846, "loss": 4.5846, "step": 287600 }, { "epoch": 0.5367868831044625, "grad_norm": 1.1560322046279907, "learning_rate": 0.00019733005601180886, "loss": 4.5619, "step": 287650 }, { "epoch": 0.536880188663841, "grad_norm": 0.8640111684799194, "learning_rate": 0.00019732909165648882, "loss": 4.7589, "step": 287700 }, { "epoch": 0.5369734942232196, "grad_norm": 1.0471079349517822, "learning_rate": 0.00019732812712940008, "loss": 4.6379, "step": 287750 }, { "epoch": 0.5370667997825981, "grad_norm": 1.2303370237350464, "learning_rate": 0.0001973271624305443, "loss": 4.5282, "step": 287800 }, { "epoch": 0.5371601053419766, "grad_norm": 0.6238991022109985, "learning_rate": 0.00019732619755992323, "loss": 4.533, "step": 287850 }, { "epoch": 0.537253410901355, "grad_norm": 1.0186882019042969, "learning_rate": 0.00019732523251753855, "loss": 4.4196, "step": 287900 }, { "epoch": 0.5373467164607335, "grad_norm": 0.9189663529396057, "learning_rate": 0.00019732426730339193, "loss": 4.8847, "step": 287950 }, { "epoch": 0.537440022020112, "grad_norm": 1.0216429233551025, "learning_rate": 0.0001973233019174851, "loss": 4.6982, "step": 288000 }, { "epoch": 0.537440022020112, "eval_loss": 4.758412837982178, "eval_runtime": 228.433, "eval_samples_per_second": 11.417, "eval_steps_per_second": 11.417, "eval_tts_loss": 7.586343273851861, "step": 288000 }, { "epoch": 0.5375333275794905, "grad_norm": 0.9328665137290955, "learning_rate": 0.0001973223363598198, "loss": 4.4907, "step": 288050 }, { "epoch": 0.537626633138869, "grad_norm": 0.9139004349708557, "learning_rate": 0.0001973213706303977, "loss": 4.4406, "step": 288100 }, { "epoch": 0.5377199386982475, "grad_norm": 0.7561057806015015, "learning_rate": 0.00019732040472922048, "loss": 4.6152, "step": 288150 }, { "epoch": 0.537813244257626, "grad_norm": 1.2258198261260986, "learning_rate": 0.00019731943865628987, "loss": 4.6745, "step": 288200 }, { "epoch": 0.5379065498170045, "grad_norm": 1.0092732906341553, "learning_rate": 0.0001973184724116076, "loss": 4.5054, "step": 288250 }, { "epoch": 0.537999855376383, "grad_norm": 1.1553289890289307, "learning_rate": 0.0001973175059951753, "loss": 4.6214, "step": 288300 }, { "epoch": 0.5380931609357614, "grad_norm": 0.8745859861373901, "learning_rate": 0.00019731653940699477, "loss": 4.4469, "step": 288350 }, { "epoch": 0.5381864664951399, "grad_norm": 1.3128819465637207, "learning_rate": 0.00019731557264706767, "loss": 4.8051, "step": 288400 }, { "epoch": 0.5382797720545185, "grad_norm": 0.9993950724601746, "learning_rate": 0.00019731460571539568, "loss": 4.6709, "step": 288450 }, { "epoch": 0.538373077613897, "grad_norm": 0.7750226259231567, "learning_rate": 0.00019731363861198056, "loss": 4.744, "step": 288500 }, { "epoch": 0.5384663831732754, "grad_norm": 1.1615644693374634, "learning_rate": 0.000197312671336824, "loss": 4.6706, "step": 288550 }, { "epoch": 0.5385596887326539, "grad_norm": 1.0617265701293945, "learning_rate": 0.00019731170388992766, "loss": 4.4617, "step": 288600 }, { "epoch": 0.5386529942920324, "grad_norm": 1.0319828987121582, "learning_rate": 0.0001973107362712933, "loss": 4.6425, "step": 288650 }, { "epoch": 0.5387462998514109, "grad_norm": 0.7510588765144348, "learning_rate": 0.00019730976848092265, "loss": 4.4306, "step": 288700 }, { "epoch": 0.5388396054107893, "grad_norm": 0.9257532358169556, "learning_rate": 0.00019730880051881732, "loss": 4.6251, "step": 288750 }, { "epoch": 0.5389329109701679, "grad_norm": 1.1619385480880737, "learning_rate": 0.0001973078323849791, "loss": 4.7905, "step": 288800 }, { "epoch": 0.5390262165295464, "grad_norm": 1.0947250127792358, "learning_rate": 0.0001973068640794097, "loss": 4.4057, "step": 288850 }, { "epoch": 0.5391195220889249, "grad_norm": 1.0521148443222046, "learning_rate": 0.00019730589560211078, "loss": 4.635, "step": 288900 }, { "epoch": 0.5392128276483034, "grad_norm": 0.9681885838508606, "learning_rate": 0.0001973049269530841, "loss": 4.4493, "step": 288950 }, { "epoch": 0.5393061332076818, "grad_norm": 0.8884767293930054, "learning_rate": 0.0001973039581323313, "loss": 4.5077, "step": 289000 }, { "epoch": 0.5393994387670603, "grad_norm": 1.1336555480957031, "learning_rate": 0.0001973029891398542, "loss": 4.536, "step": 289050 }, { "epoch": 0.5394927443264388, "grad_norm": 1.0752379894256592, "learning_rate": 0.0001973020199756544, "loss": 4.6676, "step": 289100 }, { "epoch": 0.5395860498858174, "grad_norm": 0.9591339230537415, "learning_rate": 0.00019730105063973369, "loss": 4.5266, "step": 289150 }, { "epoch": 0.5396793554451959, "grad_norm": 1.1089969873428345, "learning_rate": 0.00019730008113209372, "loss": 4.4951, "step": 289200 }, { "epoch": 0.5397726610045743, "grad_norm": 0.8516735434532166, "learning_rate": 0.00019729911145273624, "loss": 4.4916, "step": 289250 }, { "epoch": 0.5398659665639528, "grad_norm": 1.008996605873108, "learning_rate": 0.00019729814160166292, "loss": 4.6332, "step": 289300 }, { "epoch": 0.5399592721233313, "grad_norm": 0.9984986782073975, "learning_rate": 0.00019729717157887552, "loss": 4.6355, "step": 289350 }, { "epoch": 0.5400525776827098, "grad_norm": 0.8811632990837097, "learning_rate": 0.00019729620138437572, "loss": 4.3783, "step": 289400 }, { "epoch": 0.5401458832420882, "grad_norm": 1.0923969745635986, "learning_rate": 0.00019729523101816526, "loss": 4.771, "step": 289450 }, { "epoch": 0.5402391888014668, "grad_norm": 0.8841258883476257, "learning_rate": 0.00019729426048024586, "loss": 4.5809, "step": 289500 }, { "epoch": 0.5403324943608453, "grad_norm": 1.0029546022415161, "learning_rate": 0.00019729328977061917, "loss": 4.4276, "step": 289550 }, { "epoch": 0.5404257999202238, "grad_norm": 0.8529361486434937, "learning_rate": 0.00019729231888928697, "loss": 4.5002, "step": 289600 }, { "epoch": 0.5405191054796022, "grad_norm": 1.010185956954956, "learning_rate": 0.00019729134783625095, "loss": 4.6486, "step": 289650 }, { "epoch": 0.5406124110389807, "grad_norm": 0.8875432014465332, "learning_rate": 0.00019729037661151279, "loss": 4.5652, "step": 289700 }, { "epoch": 0.5407057165983592, "grad_norm": 1.2176717519760132, "learning_rate": 0.00019728940521507424, "loss": 4.4401, "step": 289750 }, { "epoch": 0.5407990221577377, "grad_norm": 0.6937578916549683, "learning_rate": 0.00019728843364693702, "loss": 4.5127, "step": 289800 }, { "epoch": 0.5408923277171162, "grad_norm": 1.0732295513153076, "learning_rate": 0.00019728746190710283, "loss": 4.3988, "step": 289850 }, { "epoch": 0.5409856332764947, "grad_norm": 1.096102237701416, "learning_rate": 0.00019728648999557338, "loss": 4.6227, "step": 289900 }, { "epoch": 0.5410789388358732, "grad_norm": 0.8393336534500122, "learning_rate": 0.00019728551791235038, "loss": 4.6566, "step": 289950 }, { "epoch": 0.5411722443952517, "grad_norm": 1.1003714799880981, "learning_rate": 0.0001972845456574356, "loss": 4.6147, "step": 290000 }, { "epoch": 0.5412655499546302, "grad_norm": 1.0207267999649048, "learning_rate": 0.0001972835732308307, "loss": 4.6324, "step": 290050 }, { "epoch": 0.5413588555140086, "grad_norm": 0.9939537644386292, "learning_rate": 0.0001972826006325374, "loss": 4.4746, "step": 290100 }, { "epoch": 0.5414521610733871, "grad_norm": 1.1141278743743896, "learning_rate": 0.0001972816278625574, "loss": 4.4754, "step": 290150 }, { "epoch": 0.5415454666327656, "grad_norm": 0.9295952916145325, "learning_rate": 0.00019728065492089248, "loss": 4.6563, "step": 290200 }, { "epoch": 0.5416387721921442, "grad_norm": 1.3614526987075806, "learning_rate": 0.00019727968180754433, "loss": 4.6514, "step": 290250 }, { "epoch": 0.5417320777515227, "grad_norm": 1.0252643823623657, "learning_rate": 0.00019727870852251464, "loss": 4.5044, "step": 290300 }, { "epoch": 0.5418253833109011, "grad_norm": 0.8520663380622864, "learning_rate": 0.00019727773506580517, "loss": 4.6316, "step": 290350 }, { "epoch": 0.5419186888702796, "grad_norm": 1.0140016078948975, "learning_rate": 0.00019727676143741757, "loss": 4.5094, "step": 290400 }, { "epoch": 0.5420119944296581, "grad_norm": 1.1733412742614746, "learning_rate": 0.00019727578763735365, "loss": 4.4411, "step": 290450 }, { "epoch": 0.5421052999890366, "grad_norm": 1.280668020248413, "learning_rate": 0.00019727481366561508, "loss": 4.5537, "step": 290500 }, { "epoch": 0.542198605548415, "grad_norm": 1.0060126781463623, "learning_rate": 0.00019727383952220353, "loss": 4.5231, "step": 290550 }, { "epoch": 0.5422919111077936, "grad_norm": 0.9130061268806458, "learning_rate": 0.0001972728652071208, "loss": 4.5402, "step": 290600 }, { "epoch": 0.5423852166671721, "grad_norm": 1.172722339630127, "learning_rate": 0.00019727189072036859, "loss": 4.6175, "step": 290650 }, { "epoch": 0.5424785222265506, "grad_norm": 0.979167640209198, "learning_rate": 0.0001972709160619486, "loss": 4.5181, "step": 290700 }, { "epoch": 0.542571827785929, "grad_norm": 0.9845288991928101, "learning_rate": 0.00019726994123186252, "loss": 4.5178, "step": 290750 }, { "epoch": 0.5426651333453075, "grad_norm": 0.8215778470039368, "learning_rate": 0.00019726896623011215, "loss": 4.6082, "step": 290800 }, { "epoch": 0.542758438904686, "grad_norm": 0.7376789450645447, "learning_rate": 0.00019726799105669918, "loss": 4.5592, "step": 290850 }, { "epoch": 0.5428517444640645, "grad_norm": 1.229047417640686, "learning_rate": 0.0001972670157116253, "loss": 4.6113, "step": 290900 }, { "epoch": 0.5429450500234431, "grad_norm": 1.0673273801803589, "learning_rate": 0.00019726604019489226, "loss": 4.5896, "step": 290950 }, { "epoch": 0.5430383555828215, "grad_norm": 1.2181791067123413, "learning_rate": 0.00019726506450650175, "loss": 4.6041, "step": 291000 }, { "epoch": 0.5430383555828215, "eval_loss": 4.755862236022949, "eval_runtime": 229.1406, "eval_samples_per_second": 11.382, "eval_steps_per_second": 11.382, "eval_tts_loss": 7.533074107711015, "step": 291000 }, { "epoch": 0.5431316611422, "grad_norm": 1.0295772552490234, "learning_rate": 0.00019726408864645555, "loss": 4.5021, "step": 291050 }, { "epoch": 0.5432249667015785, "grad_norm": 1.02863609790802, "learning_rate": 0.00019726311261475532, "loss": 4.7297, "step": 291100 }, { "epoch": 0.543318272260957, "grad_norm": 1.064806580543518, "learning_rate": 0.0001972621364114028, "loss": 4.7794, "step": 291150 }, { "epoch": 0.5434115778203354, "grad_norm": 1.2632663249969482, "learning_rate": 0.00019726116003639975, "loss": 4.5501, "step": 291200 }, { "epoch": 0.5435048833797139, "grad_norm": 0.9783966541290283, "learning_rate": 0.00019726018348974788, "loss": 4.6422, "step": 291250 }, { "epoch": 0.5435981889390925, "grad_norm": 1.000856637954712, "learning_rate": 0.0001972592067714489, "loss": 4.5299, "step": 291300 }, { "epoch": 0.543691494498471, "grad_norm": 1.0604976415634155, "learning_rate": 0.0001972582298815045, "loss": 4.2955, "step": 291350 }, { "epoch": 0.5437848000578495, "grad_norm": 1.129164695739746, "learning_rate": 0.00019725725281991645, "loss": 4.6644, "step": 291400 }, { "epoch": 0.5438781056172279, "grad_norm": 0.9613283276557922, "learning_rate": 0.00019725627558668646, "loss": 4.7118, "step": 291450 }, { "epoch": 0.5439714111766064, "grad_norm": 0.8535525798797607, "learning_rate": 0.00019725529818181626, "loss": 4.5516, "step": 291500 }, { "epoch": 0.5440647167359849, "grad_norm": 1.3331067562103271, "learning_rate": 0.00019725432060530755, "loss": 4.5355, "step": 291550 }, { "epoch": 0.5441580222953634, "grad_norm": 1.053837776184082, "learning_rate": 0.0001972533428571621, "loss": 4.4569, "step": 291600 }, { "epoch": 0.544251327854742, "grad_norm": 1.1131651401519775, "learning_rate": 0.0001972523649373816, "loss": 4.8562, "step": 291650 }, { "epoch": 0.5443446334141204, "grad_norm": 1.1066210269927979, "learning_rate": 0.00019725138684596782, "loss": 4.59, "step": 291700 }, { "epoch": 0.5444379389734989, "grad_norm": 0.9143478870391846, "learning_rate": 0.00019725040858292243, "loss": 4.5997, "step": 291750 }, { "epoch": 0.5445312445328774, "grad_norm": 0.9488169550895691, "learning_rate": 0.00019724943014824716, "loss": 4.4625, "step": 291800 }, { "epoch": 0.5446245500922559, "grad_norm": 1.1619845628738403, "learning_rate": 0.0001972484515419438, "loss": 4.5641, "step": 291850 }, { "epoch": 0.5447178556516343, "grad_norm": 0.7851349115371704, "learning_rate": 0.000197247472764014, "loss": 4.7522, "step": 291900 }, { "epoch": 0.5448111612110128, "grad_norm": 1.1872843503952026, "learning_rate": 0.00019724649381445955, "loss": 4.6147, "step": 291950 }, { "epoch": 0.5449044667703914, "grad_norm": 1.1382789611816406, "learning_rate": 0.0001972455146932821, "loss": 4.6354, "step": 292000 }, { "epoch": 0.5449977723297699, "grad_norm": 1.0083820819854736, "learning_rate": 0.00019724453540048348, "loss": 4.6133, "step": 292050 }, { "epoch": 0.5450910778891483, "grad_norm": 1.1564075946807861, "learning_rate": 0.00019724355593606535, "loss": 4.5128, "step": 292100 }, { "epoch": 0.5451843834485268, "grad_norm": 1.0278780460357666, "learning_rate": 0.00019724257630002946, "loss": 4.7728, "step": 292150 }, { "epoch": 0.5452776890079053, "grad_norm": 1.1431865692138672, "learning_rate": 0.00019724159649237754, "loss": 4.4999, "step": 292200 }, { "epoch": 0.5453709945672838, "grad_norm": 1.0673056840896606, "learning_rate": 0.0001972406165131113, "loss": 4.5965, "step": 292250 }, { "epoch": 0.5454643001266622, "grad_norm": 1.0915521383285522, "learning_rate": 0.00019723963636223245, "loss": 4.4474, "step": 292300 }, { "epoch": 0.5455576056860408, "grad_norm": 1.1890692710876465, "learning_rate": 0.0001972386560397428, "loss": 4.6021, "step": 292350 }, { "epoch": 0.5456509112454193, "grad_norm": 1.1801838874816895, "learning_rate": 0.000197237675545644, "loss": 4.4898, "step": 292400 }, { "epoch": 0.5457442168047978, "grad_norm": 1.9533765316009521, "learning_rate": 0.00019723669487993784, "loss": 4.5624, "step": 292450 }, { "epoch": 0.5458375223641763, "grad_norm": 1.0451918840408325, "learning_rate": 0.00019723571404262603, "loss": 4.6414, "step": 292500 }, { "epoch": 0.5459308279235547, "grad_norm": 0.9235672354698181, "learning_rate": 0.00019723473303371022, "loss": 4.4504, "step": 292550 }, { "epoch": 0.5460241334829332, "grad_norm": 0.9666757583618164, "learning_rate": 0.00019723375185319228, "loss": 4.5351, "step": 292600 }, { "epoch": 0.5461174390423117, "grad_norm": 0.9865828156471252, "learning_rate": 0.00019723277050107387, "loss": 4.6113, "step": 292650 }, { "epoch": 0.5462107446016903, "grad_norm": 0.8017159700393677, "learning_rate": 0.00019723178897735672, "loss": 4.6902, "step": 292700 }, { "epoch": 0.5463040501610688, "grad_norm": 0.959333598613739, "learning_rate": 0.00019723080728204257, "loss": 4.5285, "step": 292750 }, { "epoch": 0.5463973557204472, "grad_norm": 1.0091277360916138, "learning_rate": 0.00019722982541513318, "loss": 4.5637, "step": 292800 }, { "epoch": 0.5464906612798257, "grad_norm": 0.985924243927002, "learning_rate": 0.00019722884337663023, "loss": 4.686, "step": 292850 }, { "epoch": 0.5465839668392042, "grad_norm": 1.0899864435195923, "learning_rate": 0.00019722786116653548, "loss": 4.3514, "step": 292900 }, { "epoch": 0.5466772723985827, "grad_norm": 1.169295072555542, "learning_rate": 0.00019722687878485066, "loss": 4.5164, "step": 292950 }, { "epoch": 0.5467705779579611, "grad_norm": 1.144992709159851, "learning_rate": 0.00019722589623157751, "loss": 4.6521, "step": 293000 }, { "epoch": 0.5468638835173397, "grad_norm": 1.1579240560531616, "learning_rate": 0.00019722491350671777, "loss": 4.6411, "step": 293050 }, { "epoch": 0.5469571890767182, "grad_norm": 0.9811488389968872, "learning_rate": 0.00019722393061027315, "loss": 4.8124, "step": 293100 }, { "epoch": 0.5470504946360967, "grad_norm": 1.1563857793807983, "learning_rate": 0.0001972229475422454, "loss": 4.5458, "step": 293150 }, { "epoch": 0.5471438001954751, "grad_norm": 1.1877684593200684, "learning_rate": 0.00019722196430263628, "loss": 4.5323, "step": 293200 }, { "epoch": 0.5472371057548536, "grad_norm": 1.0926393270492554, "learning_rate": 0.00019722098089144748, "loss": 4.5031, "step": 293250 }, { "epoch": 0.5473304113142321, "grad_norm": 0.8998591303825378, "learning_rate": 0.00019721999730868076, "loss": 4.7053, "step": 293300 }, { "epoch": 0.5474237168736106, "grad_norm": 1.27166748046875, "learning_rate": 0.00019721901355433782, "loss": 4.553, "step": 293350 }, { "epoch": 0.5475170224329892, "grad_norm": 1.5627135038375854, "learning_rate": 0.00019721802962842047, "loss": 4.6294, "step": 293400 }, { "epoch": 0.5476103279923676, "grad_norm": 1.1435675621032715, "learning_rate": 0.0001972170455309304, "loss": 4.4011, "step": 293450 }, { "epoch": 0.5477036335517461, "grad_norm": 1.1832562685012817, "learning_rate": 0.00019721606126186933, "loss": 4.4907, "step": 293500 }, { "epoch": 0.5477969391111246, "grad_norm": 0.8468302488327026, "learning_rate": 0.00019721507682123903, "loss": 4.6218, "step": 293550 }, { "epoch": 0.5478902446705031, "grad_norm": 1.006789207458496, "learning_rate": 0.0001972140922090412, "loss": 4.5976, "step": 293600 }, { "epoch": 0.5479835502298815, "grad_norm": 0.8985641002655029, "learning_rate": 0.00019721310742527764, "loss": 4.5135, "step": 293650 }, { "epoch": 0.54807685578926, "grad_norm": 0.9235044121742249, "learning_rate": 0.00019721212246995006, "loss": 4.5865, "step": 293700 }, { "epoch": 0.5481701613486386, "grad_norm": 1.1763442754745483, "learning_rate": 0.00019721113734306014, "loss": 4.7422, "step": 293750 }, { "epoch": 0.5482634669080171, "grad_norm": 1.071958065032959, "learning_rate": 0.0001972101520446097, "loss": 4.7605, "step": 293800 }, { "epoch": 0.5483567724673956, "grad_norm": 0.9237344861030579, "learning_rate": 0.00019720916657460041, "loss": 4.6234, "step": 293850 }, { "epoch": 0.548450078026774, "grad_norm": 0.7987807393074036, "learning_rate": 0.00019720818093303408, "loss": 4.6693, "step": 293900 }, { "epoch": 0.5485433835861525, "grad_norm": 1.0376191139221191, "learning_rate": 0.0001972071951199124, "loss": 4.715, "step": 293950 }, { "epoch": 0.548636689145531, "grad_norm": 0.8929983377456665, "learning_rate": 0.00019720620913523716, "loss": 4.5809, "step": 294000 }, { "epoch": 0.548636689145531, "eval_loss": 4.764010429382324, "eval_runtime": 229.1673, "eval_samples_per_second": 11.38, "eval_steps_per_second": 11.38, "eval_tts_loss": 7.565109917429773, "step": 294000 }, { "epoch": 0.5487299947049095, "grad_norm": 0.9813073873519897, "learning_rate": 0.00019720522297901003, "loss": 4.6842, "step": 294050 }, { "epoch": 0.548823300264288, "grad_norm": 0.7194148302078247, "learning_rate": 0.00019720423665123277, "loss": 4.5848, "step": 294100 }, { "epoch": 0.5489166058236665, "grad_norm": 1.0934531688690186, "learning_rate": 0.00019720325015190718, "loss": 4.619, "step": 294150 }, { "epoch": 0.549009911383045, "grad_norm": 0.726915717124939, "learning_rate": 0.00019720226348103492, "loss": 4.5181, "step": 294200 }, { "epoch": 0.5491032169424235, "grad_norm": 0.8358513712882996, "learning_rate": 0.00019720127663861781, "loss": 4.3869, "step": 294250 }, { "epoch": 0.549196522501802, "grad_norm": 0.9418166875839233, "learning_rate": 0.0001972002896246575, "loss": 4.4952, "step": 294300 }, { "epoch": 0.5492898280611804, "grad_norm": 0.9657112956047058, "learning_rate": 0.0001971993024391558, "loss": 4.5174, "step": 294350 }, { "epoch": 0.5493831336205589, "grad_norm": 0.9693840146064758, "learning_rate": 0.00019719831508211447, "loss": 4.5318, "step": 294400 }, { "epoch": 0.5494764391799375, "grad_norm": 0.6684638261795044, "learning_rate": 0.00019719732755353516, "loss": 4.728, "step": 294450 }, { "epoch": 0.549569744739316, "grad_norm": 1.061577558517456, "learning_rate": 0.0001971963398534197, "loss": 4.7139, "step": 294500 }, { "epoch": 0.5496630502986944, "grad_norm": 1.12180495262146, "learning_rate": 0.0001971953519817698, "loss": 4.6283, "step": 294550 }, { "epoch": 0.5497563558580729, "grad_norm": 1.1466017961502075, "learning_rate": 0.0001971943639385872, "loss": 4.477, "step": 294600 }, { "epoch": 0.5498496614174514, "grad_norm": 1.0311684608459473, "learning_rate": 0.00019719337572387367, "loss": 4.5822, "step": 294650 }, { "epoch": 0.5499429669768299, "grad_norm": 0.9233488440513611, "learning_rate": 0.00019719238733763091, "loss": 4.5704, "step": 294700 }, { "epoch": 0.5500362725362083, "grad_norm": 1.1542483568191528, "learning_rate": 0.00019719139877986069, "loss": 4.62, "step": 294750 }, { "epoch": 0.5501295780955869, "grad_norm": 1.068114995956421, "learning_rate": 0.00019719041005056477, "loss": 4.6344, "step": 294800 }, { "epoch": 0.5502228836549654, "grad_norm": 1.212415099143982, "learning_rate": 0.00019718942114974484, "loss": 4.3569, "step": 294850 }, { "epoch": 0.5503161892143439, "grad_norm": 0.9979035258293152, "learning_rate": 0.00019718843207740272, "loss": 4.5337, "step": 294900 }, { "epoch": 0.5504094947737224, "grad_norm": 1.1693506240844727, "learning_rate": 0.0001971874428335401, "loss": 4.5714, "step": 294950 }, { "epoch": 0.5505028003331008, "grad_norm": 0.8749262690544128, "learning_rate": 0.00019718645341815875, "loss": 4.5207, "step": 295000 }, { "epoch": 0.5505961058924793, "grad_norm": 1.2598340511322021, "learning_rate": 0.00019718546383126043, "loss": 4.6344, "step": 295050 }, { "epoch": 0.5506894114518578, "grad_norm": 1.0515248775482178, "learning_rate": 0.00019718447407284683, "loss": 4.6527, "step": 295100 }, { "epoch": 0.5507827170112363, "grad_norm": 0.9423948526382446, "learning_rate": 0.00019718348414291974, "loss": 4.7807, "step": 295150 }, { "epoch": 0.5508760225706149, "grad_norm": 1.0501548051834106, "learning_rate": 0.0001971824940414809, "loss": 4.5323, "step": 295200 }, { "epoch": 0.5509693281299933, "grad_norm": 1.265120029449463, "learning_rate": 0.0001971815037685321, "loss": 4.7063, "step": 295250 }, { "epoch": 0.5510626336893718, "grad_norm": 1.0201817750930786, "learning_rate": 0.00019718051332407498, "loss": 4.6494, "step": 295300 }, { "epoch": 0.5511559392487503, "grad_norm": 0.9585539698600769, "learning_rate": 0.0001971795227081114, "loss": 4.5697, "step": 295350 }, { "epoch": 0.5512492448081288, "grad_norm": 0.8855648636817932, "learning_rate": 0.00019717853192064303, "loss": 4.69, "step": 295400 }, { "epoch": 0.5513425503675072, "grad_norm": 0.7688419818878174, "learning_rate": 0.00019717754096167166, "loss": 4.5284, "step": 295450 }, { "epoch": 0.5514358559268857, "grad_norm": 1.1366264820098877, "learning_rate": 0.00019717654983119907, "loss": 4.7289, "step": 295500 }, { "epoch": 0.5515291614862643, "grad_norm": 2.7926149368286133, "learning_rate": 0.00019717555852922695, "loss": 4.735, "step": 295550 }, { "epoch": 0.5516224670456428, "grad_norm": 1.3501007556915283, "learning_rate": 0.00019717456705575703, "loss": 4.4595, "step": 295600 }, { "epoch": 0.5517157726050212, "grad_norm": 0.9407560229301453, "learning_rate": 0.00019717357541079112, "loss": 4.703, "step": 295650 }, { "epoch": 0.5518090781643997, "grad_norm": 1.0523128509521484, "learning_rate": 0.00019717258359433094, "loss": 4.5778, "step": 295700 }, { "epoch": 0.5519023837237782, "grad_norm": 1.0216292142868042, "learning_rate": 0.00019717159160637828, "loss": 4.5917, "step": 295750 }, { "epoch": 0.5519956892831567, "grad_norm": 1.2850682735443115, "learning_rate": 0.00019717059944693484, "loss": 4.5741, "step": 295800 }, { "epoch": 0.5520889948425352, "grad_norm": 0.8473016023635864, "learning_rate": 0.00019716960711600235, "loss": 4.6217, "step": 295850 }, { "epoch": 0.5521823004019137, "grad_norm": 1.0704396963119507, "learning_rate": 0.00019716861461358266, "loss": 4.5319, "step": 295900 }, { "epoch": 0.5522756059612922, "grad_norm": 1.325152039527893, "learning_rate": 0.00019716762193967743, "loss": 4.6891, "step": 295950 }, { "epoch": 0.5523689115206707, "grad_norm": 0.8382663130760193, "learning_rate": 0.00019716662909428844, "loss": 4.4307, "step": 296000 }, { "epoch": 0.5524622170800492, "grad_norm": 0.8413145542144775, "learning_rate": 0.00019716563607741748, "loss": 4.7154, "step": 296050 }, { "epoch": 0.5525555226394276, "grad_norm": 0.9249318838119507, "learning_rate": 0.00019716464288906625, "loss": 4.6292, "step": 296100 }, { "epoch": 0.5526488281988061, "grad_norm": 0.9210410714149475, "learning_rate": 0.0001971636495292365, "loss": 4.6499, "step": 296150 }, { "epoch": 0.5527421337581846, "grad_norm": 0.881159245967865, "learning_rate": 0.00019716265599793004, "loss": 4.7111, "step": 296200 }, { "epoch": 0.5528354393175632, "grad_norm": 0.7526847720146179, "learning_rate": 0.00019716166229514855, "loss": 4.6092, "step": 296250 }, { "epoch": 0.5529287448769417, "grad_norm": 1.2674932479858398, "learning_rate": 0.00019716066842089387, "loss": 4.5312, "step": 296300 }, { "epoch": 0.5530220504363201, "grad_norm": 0.6895270347595215, "learning_rate": 0.00019715967437516766, "loss": 4.6097, "step": 296350 }, { "epoch": 0.5531153559956986, "grad_norm": 1.333682656288147, "learning_rate": 0.00019715868015797176, "loss": 4.6198, "step": 296400 }, { "epoch": 0.5532086615550771, "grad_norm": 1.0232683420181274, "learning_rate": 0.00019715768576930787, "loss": 4.6465, "step": 296450 }, { "epoch": 0.5533019671144556, "grad_norm": 0.875064492225647, "learning_rate": 0.00019715669120917774, "loss": 4.6424, "step": 296500 }, { "epoch": 0.553395272673834, "grad_norm": 1.2495737075805664, "learning_rate": 0.00019715569647758317, "loss": 4.694, "step": 296550 }, { "epoch": 0.5534885782332126, "grad_norm": 0.9244838953018188, "learning_rate": 0.0001971547015745259, "loss": 4.6244, "step": 296600 }, { "epoch": 0.5535818837925911, "grad_norm": 1.1226541996002197, "learning_rate": 0.00019715370650000765, "loss": 4.369, "step": 296650 }, { "epoch": 0.5536751893519696, "grad_norm": 1.1125637292861938, "learning_rate": 0.0001971527112540302, "loss": 4.8984, "step": 296700 }, { "epoch": 0.553768494911348, "grad_norm": 0.9522160887718201, "learning_rate": 0.00019715171583659536, "loss": 4.6531, "step": 296750 }, { "epoch": 0.5538618004707265, "grad_norm": 1.3746414184570312, "learning_rate": 0.0001971507202477048, "loss": 4.321, "step": 296800 }, { "epoch": 0.553955106030105, "grad_norm": 0.8997330069541931, "learning_rate": 0.0001971497244873603, "loss": 4.5425, "step": 296850 }, { "epoch": 0.5540484115894835, "grad_norm": 1.069753646850586, "learning_rate": 0.00019714872855556363, "loss": 4.4188, "step": 296900 }, { "epoch": 0.5541417171488621, "grad_norm": 1.023139476776123, "learning_rate": 0.00019714773245231657, "loss": 4.4299, "step": 296950 }, { "epoch": 0.5542350227082405, "grad_norm": 0.9713852405548096, "learning_rate": 0.00019714673617762086, "loss": 4.4206, "step": 297000 }, { "epoch": 0.5542350227082405, "eval_loss": 4.7526021003723145, "eval_runtime": 231.4288, "eval_samples_per_second": 11.269, "eval_steps_per_second": 11.269, "eval_tts_loss": 7.537269182848898, "step": 297000 }, { "epoch": 0.554328328267619, "grad_norm": 0.9608700275421143, "learning_rate": 0.00019714573973147822, "loss": 4.6437, "step": 297050 }, { "epoch": 0.5544216338269975, "grad_norm": 1.1964012384414673, "learning_rate": 0.00019714474311389047, "loss": 4.6537, "step": 297100 }, { "epoch": 0.554514939386376, "grad_norm": 1.1816233396530151, "learning_rate": 0.00019714374632485933, "loss": 4.4968, "step": 297150 }, { "epoch": 0.5546082449457544, "grad_norm": 0.9269901514053345, "learning_rate": 0.00019714274936438657, "loss": 4.6129, "step": 297200 }, { "epoch": 0.5547015505051329, "grad_norm": 1.0753899812698364, "learning_rate": 0.00019714175223247397, "loss": 4.6587, "step": 297250 }, { "epoch": 0.5547948560645115, "grad_norm": 1.0030367374420166, "learning_rate": 0.00019714075492912326, "loss": 4.6356, "step": 297300 }, { "epoch": 0.55488816162389, "grad_norm": 0.9227311015129089, "learning_rate": 0.0001971397574543362, "loss": 4.4709, "step": 297350 }, { "epoch": 0.5549814671832685, "grad_norm": 1.0956863164901733, "learning_rate": 0.00019713875980811456, "loss": 4.4858, "step": 297400 }, { "epoch": 0.5550747727426469, "grad_norm": 0.9432074427604675, "learning_rate": 0.0001971377619904601, "loss": 4.4785, "step": 297450 }, { "epoch": 0.5551680783020254, "grad_norm": 1.325101375579834, "learning_rate": 0.0001971367640013746, "loss": 4.4846, "step": 297500 }, { "epoch": 0.5552613838614039, "grad_norm": 1.0917229652404785, "learning_rate": 0.00019713576584085978, "loss": 4.465, "step": 297550 }, { "epoch": 0.5553546894207824, "grad_norm": 1.0985119342803955, "learning_rate": 0.00019713476750891744, "loss": 4.7151, "step": 297600 }, { "epoch": 0.555447994980161, "grad_norm": 0.9184651970863342, "learning_rate": 0.00019713376900554931, "loss": 4.5393, "step": 297650 }, { "epoch": 0.5555413005395394, "grad_norm": 1.0063183307647705, "learning_rate": 0.0001971327703307572, "loss": 4.5729, "step": 297700 }, { "epoch": 0.5556346060989179, "grad_norm": 0.9919562935829163, "learning_rate": 0.00019713177148454283, "loss": 4.7591, "step": 297750 }, { "epoch": 0.5557279116582964, "grad_norm": 0.8779546022415161, "learning_rate": 0.00019713077246690794, "loss": 4.4183, "step": 297800 }, { "epoch": 0.5558212172176749, "grad_norm": 0.8050033450126648, "learning_rate": 0.00019712977327785436, "loss": 4.463, "step": 297850 }, { "epoch": 0.5559145227770533, "grad_norm": 1.0061942338943481, "learning_rate": 0.00019712877391738382, "loss": 4.8949, "step": 297900 }, { "epoch": 0.5560078283364318, "grad_norm": 0.9918883442878723, "learning_rate": 0.0001971277743854981, "loss": 4.5452, "step": 297950 }, { "epoch": 0.5561011338958104, "grad_norm": 0.9152721762657166, "learning_rate": 0.00019712677468219892, "loss": 4.603, "step": 298000 }, { "epoch": 0.5561944394551889, "grad_norm": 0.9235228896141052, "learning_rate": 0.0001971257748074881, "loss": 4.4085, "step": 298050 }, { "epoch": 0.5562877450145673, "grad_norm": 1.0980701446533203, "learning_rate": 0.00019712477476136733, "loss": 4.4507, "step": 298100 }, { "epoch": 0.5563810505739458, "grad_norm": 0.8840431571006775, "learning_rate": 0.00019712377454383847, "loss": 4.7334, "step": 298150 }, { "epoch": 0.5564743561333243, "grad_norm": 0.9748885035514832, "learning_rate": 0.0001971227741549032, "loss": 4.8757, "step": 298200 }, { "epoch": 0.5565676616927028, "grad_norm": 1.0682768821716309, "learning_rate": 0.00019712177359456335, "loss": 4.6206, "step": 298250 }, { "epoch": 0.5566609672520813, "grad_norm": 1.0033636093139648, "learning_rate": 0.00019712077286282065, "loss": 4.4473, "step": 298300 }, { "epoch": 0.5567542728114598, "grad_norm": 0.8800263404846191, "learning_rate": 0.00019711977195967687, "loss": 4.5585, "step": 298350 }, { "epoch": 0.5568475783708383, "grad_norm": 1.004462718963623, "learning_rate": 0.00019711877088513378, "loss": 4.703, "step": 298400 }, { "epoch": 0.5569408839302168, "grad_norm": 0.99757981300354, "learning_rate": 0.00019711776963919316, "loss": 4.751, "step": 298450 }, { "epoch": 0.5570341894895953, "grad_norm": 0.8325973749160767, "learning_rate": 0.00019711676822185674, "loss": 4.7005, "step": 298500 }, { "epoch": 0.5571274950489737, "grad_norm": 0.8714603781700134, "learning_rate": 0.00019711576663312635, "loss": 4.7723, "step": 298550 }, { "epoch": 0.5572208006083522, "grad_norm": 1.3608721494674683, "learning_rate": 0.00019711476487300372, "loss": 4.6039, "step": 298600 }, { "epoch": 0.5573141061677307, "grad_norm": 1.2337629795074463, "learning_rate": 0.00019711376294149056, "loss": 4.4813, "step": 298650 }, { "epoch": 0.5574074117271093, "grad_norm": 1.042840600013733, "learning_rate": 0.00019711276083858877, "loss": 4.5122, "step": 298700 }, { "epoch": 0.5575007172864878, "grad_norm": 1.5062463283538818, "learning_rate": 0.0001971117585643, "loss": 4.6468, "step": 298750 }, { "epoch": 0.5575940228458662, "grad_norm": 1.0929310321807861, "learning_rate": 0.00019711075611862606, "loss": 4.6295, "step": 298800 }, { "epoch": 0.5576873284052447, "grad_norm": 0.9297996163368225, "learning_rate": 0.00019710975350156873, "loss": 4.448, "step": 298850 }, { "epoch": 0.5577806339646232, "grad_norm": 1.0459295511245728, "learning_rate": 0.00019710875071312977, "loss": 4.4881, "step": 298900 }, { "epoch": 0.5578739395240017, "grad_norm": 0.9883279800415039, "learning_rate": 0.00019710774775331098, "loss": 4.611, "step": 298950 }, { "epoch": 0.5579672450833801, "grad_norm": 0.9868349432945251, "learning_rate": 0.00019710674462211408, "loss": 4.525, "step": 299000 }, { "epoch": 0.5580605506427587, "grad_norm": 0.8064708113670349, "learning_rate": 0.00019710574131954087, "loss": 4.493, "step": 299050 }, { "epoch": 0.5581538562021372, "grad_norm": 1.1615211963653564, "learning_rate": 0.0001971047378455931, "loss": 4.6057, "step": 299100 }, { "epoch": 0.5582471617615157, "grad_norm": 1.1374075412750244, "learning_rate": 0.00019710373420027256, "loss": 4.5673, "step": 299150 }, { "epoch": 0.5583404673208942, "grad_norm": 0.9291498064994812, "learning_rate": 0.00019710273038358103, "loss": 4.4507, "step": 299200 }, { "epoch": 0.5584337728802726, "grad_norm": 1.0071158409118652, "learning_rate": 0.00019710172639552024, "loss": 4.7346, "step": 299250 }, { "epoch": 0.5585270784396511, "grad_norm": 0.8895450830459595, "learning_rate": 0.000197100722236092, "loss": 4.4564, "step": 299300 }, { "epoch": 0.5586203839990296, "grad_norm": 0.9139736890792847, "learning_rate": 0.00019709971790529807, "loss": 4.6733, "step": 299350 }, { "epoch": 0.5587136895584082, "grad_norm": 1.2478666305541992, "learning_rate": 0.0001970987134031402, "loss": 4.7858, "step": 299400 }, { "epoch": 0.5588069951177866, "grad_norm": 0.9026058912277222, "learning_rate": 0.00019709770872962024, "loss": 4.6203, "step": 299450 }, { "epoch": 0.5589003006771651, "grad_norm": 1.2664495706558228, "learning_rate": 0.00019709670388473987, "loss": 4.6035, "step": 299500 }, { "epoch": 0.5589936062365436, "grad_norm": 1.0902659893035889, "learning_rate": 0.00019709569886850093, "loss": 4.7008, "step": 299550 }, { "epoch": 0.5590869117959221, "grad_norm": 0.7903751134872437, "learning_rate": 0.00019709469368090515, "loss": 4.5173, "step": 299600 }, { "epoch": 0.5591802173553005, "grad_norm": 0.8331606984138489, "learning_rate": 0.0001970936883219543, "loss": 4.7115, "step": 299650 }, { "epoch": 0.559273522914679, "grad_norm": 1.0867011547088623, "learning_rate": 0.00019709268279165018, "loss": 4.6903, "step": 299700 }, { "epoch": 0.5593668284740576, "grad_norm": 1.0519412755966187, "learning_rate": 0.00019709167708999457, "loss": 4.4897, "step": 299750 }, { "epoch": 0.5594601340334361, "grad_norm": 1.4251899719238281, "learning_rate": 0.00019709067121698924, "loss": 4.6184, "step": 299800 }, { "epoch": 0.5595534395928146, "grad_norm": 1.2636020183563232, "learning_rate": 0.00019708966517263595, "loss": 4.4974, "step": 299850 }, { "epoch": 0.559646745152193, "grad_norm": 0.7390130758285522, "learning_rate": 0.00019708865895693648, "loss": 4.6559, "step": 299900 }, { "epoch": 0.5597400507115715, "grad_norm": 0.8000625967979431, "learning_rate": 0.00019708765256989263, "loss": 4.5117, "step": 299950 }, { "epoch": 0.55983335627095, "grad_norm": 1.0556968450546265, "learning_rate": 0.00019708664601150616, "loss": 4.4716, "step": 300000 }, { "epoch": 0.55983335627095, "eval_loss": 4.755502700805664, "eval_runtime": 232.7346, "eval_samples_per_second": 11.206, "eval_steps_per_second": 11.206, "eval_tts_loss": 7.561606850610771, "step": 300000 }, { "epoch": 0.5599266618303285, "grad_norm": 1.1514503955841064, "learning_rate": 0.0001970856392817788, "loss": 4.5287, "step": 300050 }, { "epoch": 0.560019967389707, "grad_norm": 1.0816736221313477, "learning_rate": 0.00019708463238071242, "loss": 4.6709, "step": 300100 }, { "epoch": 0.5601132729490855, "grad_norm": 0.9680812954902649, "learning_rate": 0.0001970836253083087, "loss": 4.5883, "step": 300150 }, { "epoch": 0.560206578508464, "grad_norm": 1.0057451725006104, "learning_rate": 0.0001970826180645695, "loss": 4.4862, "step": 300200 }, { "epoch": 0.5602998840678425, "grad_norm": 0.8072318434715271, "learning_rate": 0.00019708161064949656, "loss": 4.5203, "step": 300250 }, { "epoch": 0.560393189627221, "grad_norm": 0.8877402544021606, "learning_rate": 0.00019708060306309163, "loss": 4.6058, "step": 300300 }, { "epoch": 0.5604864951865994, "grad_norm": 0.8873353600502014, "learning_rate": 0.00019707959530535654, "loss": 4.8, "step": 300350 }, { "epoch": 0.5605798007459779, "grad_norm": 0.993877112865448, "learning_rate": 0.00019707858737629303, "loss": 4.4375, "step": 300400 }, { "epoch": 0.5606731063053564, "grad_norm": 0.8739638924598694, "learning_rate": 0.00019707757927590292, "loss": 4.4138, "step": 300450 }, { "epoch": 0.560766411864735, "grad_norm": 1.0241881608963013, "learning_rate": 0.00019707657100418794, "loss": 4.4477, "step": 300500 }, { "epoch": 0.5608597174241134, "grad_norm": 0.8494142293930054, "learning_rate": 0.0001970755625611499, "loss": 4.5795, "step": 300550 }, { "epoch": 0.5609530229834919, "grad_norm": 1.0680198669433594, "learning_rate": 0.00019707455394679058, "loss": 4.562, "step": 300600 }, { "epoch": 0.5610463285428704, "grad_norm": 0.9774929881095886, "learning_rate": 0.00019707354516111173, "loss": 4.6641, "step": 300650 }, { "epoch": 0.5611396341022489, "grad_norm": 1.0798308849334717, "learning_rate": 0.0001970725362041152, "loss": 4.433, "step": 300700 }, { "epoch": 0.5612329396616274, "grad_norm": 1.1354361772537231, "learning_rate": 0.0001970715270758027, "loss": 4.6163, "step": 300750 }, { "epoch": 0.5613262452210058, "grad_norm": 1.0032528638839722, "learning_rate": 0.00019707051777617603, "loss": 4.5807, "step": 300800 }, { "epoch": 0.5614195507803844, "grad_norm": 0.8227291107177734, "learning_rate": 0.00019706950830523695, "loss": 4.6253, "step": 300850 }, { "epoch": 0.5615128563397629, "grad_norm": 1.0578374862670898, "learning_rate": 0.00019706849866298732, "loss": 4.483, "step": 300900 }, { "epoch": 0.5616061618991414, "grad_norm": 0.8574528694152832, "learning_rate": 0.00019706748884942883, "loss": 4.7146, "step": 300950 }, { "epoch": 0.5616994674585198, "grad_norm": 1.0092363357543945, "learning_rate": 0.00019706647886456332, "loss": 4.4874, "step": 301000 }, { "epoch": 0.5617927730178983, "grad_norm": 1.14267098903656, "learning_rate": 0.00019706546870839256, "loss": 4.7584, "step": 301050 }, { "epoch": 0.5618860785772768, "grad_norm": 0.9092347621917725, "learning_rate": 0.0001970644583809183, "loss": 4.4427, "step": 301100 }, { "epoch": 0.5619793841366553, "grad_norm": 0.960862934589386, "learning_rate": 0.00019706344788214238, "loss": 4.6718, "step": 301150 }, { "epoch": 0.5620726896960339, "grad_norm": 0.8959313631057739, "learning_rate": 0.00019706243721206654, "loss": 4.5922, "step": 301200 }, { "epoch": 0.5621659952554123, "grad_norm": 0.8489114046096802, "learning_rate": 0.00019706142637069255, "loss": 4.7201, "step": 301250 }, { "epoch": 0.5622593008147908, "grad_norm": 1.1516261100769043, "learning_rate": 0.00019706041535802227, "loss": 4.5853, "step": 301300 }, { "epoch": 0.5623526063741693, "grad_norm": 1.163648009300232, "learning_rate": 0.00019705940417405742, "loss": 4.5929, "step": 301350 }, { "epoch": 0.5624459119335478, "grad_norm": 0.963965654373169, "learning_rate": 0.0001970583928187998, "loss": 4.5252, "step": 301400 }, { "epoch": 0.5625392174929262, "grad_norm": 1.6231391429901123, "learning_rate": 0.00019705738129225118, "loss": 4.5953, "step": 301450 }, { "epoch": 0.5626325230523047, "grad_norm": 0.8425366878509521, "learning_rate": 0.00019705636959441335, "loss": 4.7018, "step": 301500 }, { "epoch": 0.5627258286116833, "grad_norm": 1.016891598701477, "learning_rate": 0.00019705535772528813, "loss": 4.5982, "step": 301550 }, { "epoch": 0.5628191341710618, "grad_norm": 1.1888262033462524, "learning_rate": 0.0001970543456848773, "loss": 4.6413, "step": 301600 }, { "epoch": 0.5629124397304403, "grad_norm": 1.064733624458313, "learning_rate": 0.00019705333347318257, "loss": 4.5598, "step": 301650 }, { "epoch": 0.5630057452898187, "grad_norm": 0.8338524103164673, "learning_rate": 0.00019705232109020583, "loss": 4.4923, "step": 301700 }, { "epoch": 0.5630990508491972, "grad_norm": 1.349608302116394, "learning_rate": 0.0001970513085359488, "loss": 4.5175, "step": 301750 }, { "epoch": 0.5631923564085757, "grad_norm": 0.7603903412818909, "learning_rate": 0.00019705029581041327, "loss": 4.5393, "step": 301800 }, { "epoch": 0.5632856619679542, "grad_norm": 1.117794394493103, "learning_rate": 0.00019704928291360108, "loss": 4.4936, "step": 301850 }, { "epoch": 0.5633789675273327, "grad_norm": 1.124933123588562, "learning_rate": 0.00019704826984551398, "loss": 4.5207, "step": 301900 }, { "epoch": 0.5634722730867112, "grad_norm": 1.0115653276443481, "learning_rate": 0.00019704725660615372, "loss": 4.7858, "step": 301950 }, { "epoch": 0.5635655786460897, "grad_norm": 0.9481130838394165, "learning_rate": 0.0001970462431955222, "loss": 4.6487, "step": 302000 }, { "epoch": 0.5636588842054682, "grad_norm": 1.1163173913955688, "learning_rate": 0.00019704522961362105, "loss": 4.6169, "step": 302050 }, { "epoch": 0.5637521897648466, "grad_norm": 1.0677237510681152, "learning_rate": 0.0001970442158604522, "loss": 4.8105, "step": 302100 }, { "epoch": 0.5638454953242251, "grad_norm": 1.0545969009399414, "learning_rate": 0.00019704320193601738, "loss": 4.4465, "step": 302150 }, { "epoch": 0.5639388008836036, "grad_norm": 1.149702548980713, "learning_rate": 0.0001970421878403184, "loss": 4.4358, "step": 302200 }, { "epoch": 0.5640321064429822, "grad_norm": 1.150503158569336, "learning_rate": 0.000197041173573357, "loss": 4.7382, "step": 302250 }, { "epoch": 0.5641254120023607, "grad_norm": 0.8477134704589844, "learning_rate": 0.000197040159135135, "loss": 4.6298, "step": 302300 }, { "epoch": 0.5642187175617391, "grad_norm": 1.2105333805084229, "learning_rate": 0.00019703914452565423, "loss": 4.4616, "step": 302350 }, { "epoch": 0.5643120231211176, "grad_norm": 0.7254282832145691, "learning_rate": 0.0001970381297449164, "loss": 4.5311, "step": 302400 }, { "epoch": 0.5644053286804961, "grad_norm": 0.9602898359298706, "learning_rate": 0.00019703711479292338, "loss": 4.6499, "step": 302450 }, { "epoch": 0.5644986342398746, "grad_norm": 1.1957898139953613, "learning_rate": 0.0001970360996696769, "loss": 4.6785, "step": 302500 }, { "epoch": 0.564591939799253, "grad_norm": 1.1815950870513916, "learning_rate": 0.0001970350843751788, "loss": 4.526, "step": 302550 }, { "epoch": 0.5646852453586316, "grad_norm": 0.9010161757469177, "learning_rate": 0.00019703406890943085, "loss": 4.6529, "step": 302600 }, { "epoch": 0.5647785509180101, "grad_norm": 0.8498725891113281, "learning_rate": 0.00019703305327243485, "loss": 4.4077, "step": 302650 }, { "epoch": 0.5648718564773886, "grad_norm": 0.838350236415863, "learning_rate": 0.00019703203746419254, "loss": 4.5618, "step": 302700 }, { "epoch": 0.564965162036767, "grad_norm": 1.0802860260009766, "learning_rate": 0.0001970310214847058, "loss": 4.6637, "step": 302750 }, { "epoch": 0.5650584675961455, "grad_norm": 1.061564564704895, "learning_rate": 0.00019703000533397635, "loss": 4.5875, "step": 302800 }, { "epoch": 0.565151773155524, "grad_norm": 1.123292326927185, "learning_rate": 0.00019702898901200602, "loss": 4.7532, "step": 302850 }, { "epoch": 0.5652450787149025, "grad_norm": 0.901685357093811, "learning_rate": 0.00019702797251879662, "loss": 4.6985, "step": 302900 }, { "epoch": 0.5653383842742811, "grad_norm": 0.8660471439361572, "learning_rate": 0.0001970269558543499, "loss": 4.7795, "step": 302950 }, { "epoch": 0.5654316898336595, "grad_norm": 1.0963807106018066, "learning_rate": 0.00019702593901866768, "loss": 4.675, "step": 303000 }, { "epoch": 0.5654316898336595, "eval_loss": 4.748389720916748, "eval_runtime": 230.9537, "eval_samples_per_second": 11.292, "eval_steps_per_second": 11.292, "eval_tts_loss": 7.572273884011444, "step": 303000 }, { "epoch": 0.565524995393038, "grad_norm": 1.2750355005264282, "learning_rate": 0.00019702492201175177, "loss": 4.7116, "step": 303050 }, { "epoch": 0.5656183009524165, "grad_norm": 0.9990673661231995, "learning_rate": 0.0001970239048336039, "loss": 4.6189, "step": 303100 }, { "epoch": 0.565711606511795, "grad_norm": 1.0602651834487915, "learning_rate": 0.00019702288748422596, "loss": 4.8896, "step": 303150 }, { "epoch": 0.5658049120711734, "grad_norm": 0.9759448170661926, "learning_rate": 0.00019702186996361967, "loss": 4.6535, "step": 303200 }, { "epoch": 0.5658982176305519, "grad_norm": 0.7759549021720886, "learning_rate": 0.00019702085227178687, "loss": 4.63, "step": 303250 }, { "epoch": 0.5659915231899305, "grad_norm": 1.1305766105651855, "learning_rate": 0.0001970198344087293, "loss": 4.5686, "step": 303300 }, { "epoch": 0.566084828749309, "grad_norm": 1.027217149734497, "learning_rate": 0.0001970188163744488, "loss": 4.5261, "step": 303350 }, { "epoch": 0.5661781343086875, "grad_norm": 0.970268726348877, "learning_rate": 0.0001970177981689472, "loss": 4.6596, "step": 303400 }, { "epoch": 0.5662714398680659, "grad_norm": 1.0845476388931274, "learning_rate": 0.00019701677979222623, "loss": 4.645, "step": 303450 }, { "epoch": 0.5663647454274444, "grad_norm": 1.0397199392318726, "learning_rate": 0.00019701576124428772, "loss": 4.6715, "step": 303500 }, { "epoch": 0.5664580509868229, "grad_norm": 1.0684022903442383, "learning_rate": 0.00019701474252513347, "loss": 4.5604, "step": 303550 }, { "epoch": 0.5665513565462014, "grad_norm": 0.8587548732757568, "learning_rate": 0.00019701372363476523, "loss": 4.4554, "step": 303600 }, { "epoch": 0.56664466210558, "grad_norm": 1.0176758766174316, "learning_rate": 0.0001970127045731849, "loss": 4.6871, "step": 303650 }, { "epoch": 0.5667379676649584, "grad_norm": 0.9839355945587158, "learning_rate": 0.00019701168534039416, "loss": 4.4283, "step": 303700 }, { "epoch": 0.5668312732243369, "grad_norm": 1.0790977478027344, "learning_rate": 0.00019701066593639488, "loss": 4.5313, "step": 303750 }, { "epoch": 0.5669245787837154, "grad_norm": 1.2821701765060425, "learning_rate": 0.00019700964636118885, "loss": 4.6008, "step": 303800 }, { "epoch": 0.5670178843430939, "grad_norm": 0.9878072738647461, "learning_rate": 0.00019700862661477787, "loss": 4.3788, "step": 303850 }, { "epoch": 0.5671111899024723, "grad_norm": 1.4116833209991455, "learning_rate": 0.00019700760669716375, "loss": 4.5283, "step": 303900 }, { "epoch": 0.5672044954618508, "grad_norm": 0.9004996418952942, "learning_rate": 0.00019700658660834824, "loss": 4.7421, "step": 303950 }, { "epoch": 0.5672978010212294, "grad_norm": 1.0024306774139404, "learning_rate": 0.00019700556634833318, "loss": 4.7196, "step": 304000 }, { "epoch": 0.5673911065806079, "grad_norm": 1.0337103605270386, "learning_rate": 0.00019700454591712037, "loss": 4.5365, "step": 304050 }, { "epoch": 0.5674844121399863, "grad_norm": 1.1965677738189697, "learning_rate": 0.0001970035253147116, "loss": 4.4689, "step": 304100 }, { "epoch": 0.5675777176993648, "grad_norm": 0.9932730197906494, "learning_rate": 0.00019700250454110868, "loss": 4.7301, "step": 304150 }, { "epoch": 0.5676710232587433, "grad_norm": 0.9931434392929077, "learning_rate": 0.00019700148359631342, "loss": 4.4457, "step": 304200 }, { "epoch": 0.5677643288181218, "grad_norm": 0.8123148083686829, "learning_rate": 0.00019700046248032757, "loss": 4.4538, "step": 304250 }, { "epoch": 0.5678576343775003, "grad_norm": 1.0292284488677979, "learning_rate": 0.000196999441193153, "loss": 4.7729, "step": 304300 }, { "epoch": 0.5679509399368788, "grad_norm": 1.213063359260559, "learning_rate": 0.0001969984197347915, "loss": 4.584, "step": 304350 }, { "epoch": 0.5680442454962573, "grad_norm": 0.8655990362167358, "learning_rate": 0.0001969973981052448, "loss": 4.497, "step": 304400 }, { "epoch": 0.5681375510556358, "grad_norm": 1.1007928848266602, "learning_rate": 0.00019699637630451485, "loss": 4.4047, "step": 304450 }, { "epoch": 0.5682308566150143, "grad_norm": 1.1282696723937988, "learning_rate": 0.0001969953543326033, "loss": 4.575, "step": 304500 }, { "epoch": 0.5683241621743927, "grad_norm": 1.417205810546875, "learning_rate": 0.00019699433218951201, "loss": 4.6682, "step": 304550 }, { "epoch": 0.5684174677337712, "grad_norm": 1.053480863571167, "learning_rate": 0.00019699330987524283, "loss": 4.4982, "step": 304600 }, { "epoch": 0.5685107732931497, "grad_norm": 0.7026126980781555, "learning_rate": 0.00019699228738979748, "loss": 4.6107, "step": 304650 }, { "epoch": 0.5686040788525283, "grad_norm": 0.6384779214859009, "learning_rate": 0.00019699126473317783, "loss": 4.4579, "step": 304700 }, { "epoch": 0.5686973844119068, "grad_norm": 1.152971625328064, "learning_rate": 0.0001969902419053857, "loss": 4.6327, "step": 304750 }, { "epoch": 0.5687906899712852, "grad_norm": 1.2194347381591797, "learning_rate": 0.0001969892189064228, "loss": 4.7345, "step": 304800 }, { "epoch": 0.5688839955306637, "grad_norm": 0.9324080944061279, "learning_rate": 0.00019698819573629102, "loss": 4.6878, "step": 304850 }, { "epoch": 0.5689773010900422, "grad_norm": 1.0412653684616089, "learning_rate": 0.00019698717239499214, "loss": 4.6517, "step": 304900 }, { "epoch": 0.5690706066494207, "grad_norm": 1.0900930166244507, "learning_rate": 0.00019698614888252798, "loss": 4.4744, "step": 304950 }, { "epoch": 0.5691639122087991, "grad_norm": 1.0790338516235352, "learning_rate": 0.0001969851251989003, "loss": 4.5466, "step": 305000 }, { "epoch": 0.5692572177681777, "grad_norm": 1.126724123954773, "learning_rate": 0.00019698410134411096, "loss": 4.5577, "step": 305050 }, { "epoch": 0.5693505233275562, "grad_norm": 1.2070733308792114, "learning_rate": 0.00019698307731816174, "loss": 4.6863, "step": 305100 }, { "epoch": 0.5694438288869347, "grad_norm": 0.6844452023506165, "learning_rate": 0.00019698205312105444, "loss": 4.4842, "step": 305150 }, { "epoch": 0.5695371344463132, "grad_norm": 1.2279843091964722, "learning_rate": 0.0001969810287527909, "loss": 4.537, "step": 305200 }, { "epoch": 0.5696304400056916, "grad_norm": 0.9124521017074585, "learning_rate": 0.00019698000421337292, "loss": 4.3272, "step": 305250 }, { "epoch": 0.5697237455650701, "grad_norm": 1.001929521560669, "learning_rate": 0.00019697897950280227, "loss": 4.6252, "step": 305300 }, { "epoch": 0.5698170511244486, "grad_norm": 1.156294584274292, "learning_rate": 0.00019697795462108081, "loss": 4.6983, "step": 305350 }, { "epoch": 0.5699103566838272, "grad_norm": 1.0211153030395508, "learning_rate": 0.00019697692956821032, "loss": 4.4763, "step": 305400 }, { "epoch": 0.5700036622432056, "grad_norm": 1.119654893875122, "learning_rate": 0.0001969759043441926, "loss": 4.6318, "step": 305450 }, { "epoch": 0.5700969678025841, "grad_norm": 1.1528464555740356, "learning_rate": 0.00019697487894902946, "loss": 4.5226, "step": 305500 }, { "epoch": 0.5701902733619626, "grad_norm": 0.6716154217720032, "learning_rate": 0.00019697385338272275, "loss": 4.6249, "step": 305550 }, { "epoch": 0.5702835789213411, "grad_norm": 0.9842194318771362, "learning_rate": 0.00019697282764527424, "loss": 4.5753, "step": 305600 }, { "epoch": 0.5703768844807195, "grad_norm": 1.140363097190857, "learning_rate": 0.00019697180173668574, "loss": 4.5553, "step": 305650 }, { "epoch": 0.570470190040098, "grad_norm": 1.0138086080551147, "learning_rate": 0.0001969707756569591, "loss": 4.3587, "step": 305700 }, { "epoch": 0.5705634955994766, "grad_norm": 2.0012075901031494, "learning_rate": 0.00019696974940609606, "loss": 4.5402, "step": 305750 }, { "epoch": 0.5706568011588551, "grad_norm": 1.1730917692184448, "learning_rate": 0.0001969687229840985, "loss": 4.687, "step": 305800 }, { "epoch": 0.5707501067182336, "grad_norm": 1.1034109592437744, "learning_rate": 0.0001969676963909682, "loss": 4.6527, "step": 305850 }, { "epoch": 0.570843412277612, "grad_norm": 0.922483503818512, "learning_rate": 0.00019696666962670703, "loss": 4.4858, "step": 305900 }, { "epoch": 0.5709367178369905, "grad_norm": 1.0203330516815186, "learning_rate": 0.0001969656426913167, "loss": 4.7398, "step": 305950 }, { "epoch": 0.571030023396369, "grad_norm": 0.7679086923599243, "learning_rate": 0.00019696461558479906, "loss": 4.5363, "step": 306000 }, { "epoch": 0.571030023396369, "eval_loss": 4.761473178863525, "eval_runtime": 231.9504, "eval_samples_per_second": 11.244, "eval_steps_per_second": 11.244, "eval_tts_loss": 7.575939671752083, "step": 306000 }, { "epoch": 0.5711233289557475, "grad_norm": 1.031692385673523, "learning_rate": 0.00019696358830715598, "loss": 4.6841, "step": 306050 }, { "epoch": 0.5712166345151259, "grad_norm": 0.9715311527252197, "learning_rate": 0.0001969625608583892, "loss": 4.7471, "step": 306100 }, { "epoch": 0.5713099400745045, "grad_norm": 0.7275256514549255, "learning_rate": 0.00019696153323850056, "loss": 4.5227, "step": 306150 }, { "epoch": 0.571403245633883, "grad_norm": 0.9675799012184143, "learning_rate": 0.00019696050544749187, "loss": 4.3886, "step": 306200 }, { "epoch": 0.5714965511932615, "grad_norm": 0.7808157205581665, "learning_rate": 0.00019695947748536496, "loss": 4.3582, "step": 306250 }, { "epoch": 0.57158985675264, "grad_norm": 1.1058979034423828, "learning_rate": 0.00019695844935212163, "loss": 4.6779, "step": 306300 }, { "epoch": 0.5716831623120184, "grad_norm": 1.121512532234192, "learning_rate": 0.0001969574210477637, "loss": 4.8738, "step": 306350 }, { "epoch": 0.5717764678713969, "grad_norm": 0.8467114567756653, "learning_rate": 0.00019695639257229298, "loss": 4.4983, "step": 306400 }, { "epoch": 0.5718697734307754, "grad_norm": 0.9501798748970032, "learning_rate": 0.00019695536392571128, "loss": 4.6996, "step": 306450 }, { "epoch": 0.571963078990154, "grad_norm": 1.0179754495620728, "learning_rate": 0.00019695433510802043, "loss": 4.4817, "step": 306500 }, { "epoch": 0.5720563845495324, "grad_norm": 1.097285509109497, "learning_rate": 0.0001969533061192222, "loss": 4.5062, "step": 306550 }, { "epoch": 0.5721496901089109, "grad_norm": 1.1028809547424316, "learning_rate": 0.0001969522769593185, "loss": 4.4032, "step": 306600 }, { "epoch": 0.5722429956682894, "grad_norm": 0.931887149810791, "learning_rate": 0.00019695124762831106, "loss": 4.5055, "step": 306650 }, { "epoch": 0.5723363012276679, "grad_norm": 1.1663527488708496, "learning_rate": 0.00019695021812620177, "loss": 4.5626, "step": 306700 }, { "epoch": 0.5724296067870464, "grad_norm": 1.0238298177719116, "learning_rate": 0.00019694918845299236, "loss": 4.5973, "step": 306750 }, { "epoch": 0.5725229123464248, "grad_norm": 1.0825369358062744, "learning_rate": 0.0001969481586086847, "loss": 4.5194, "step": 306800 }, { "epoch": 0.5726162179058034, "grad_norm": 1.1342600584030151, "learning_rate": 0.0001969471285932806, "loss": 4.768, "step": 306850 }, { "epoch": 0.5727095234651819, "grad_norm": 1.0217125415802002, "learning_rate": 0.0001969460984067819, "loss": 4.7457, "step": 306900 }, { "epoch": 0.5728028290245604, "grad_norm": 1.3335061073303223, "learning_rate": 0.00019694506804919035, "loss": 4.6568, "step": 306950 }, { "epoch": 0.5728961345839388, "grad_norm": 0.8185517191886902, "learning_rate": 0.00019694403752050784, "loss": 4.6078, "step": 307000 }, { "epoch": 0.5729894401433173, "grad_norm": 1.478460669517517, "learning_rate": 0.00019694300682073616, "loss": 4.6365, "step": 307050 }, { "epoch": 0.5730827457026958, "grad_norm": 1.2282335758209229, "learning_rate": 0.00019694197594987713, "loss": 4.5315, "step": 307100 }, { "epoch": 0.5731760512620743, "grad_norm": 0.8165243864059448, "learning_rate": 0.00019694094490793255, "loss": 4.5083, "step": 307150 }, { "epoch": 0.5732693568214529, "grad_norm": 1.2821803092956543, "learning_rate": 0.0001969399136949043, "loss": 4.6808, "step": 307200 }, { "epoch": 0.5733626623808313, "grad_norm": 0.8003031015396118, "learning_rate": 0.00019693888231079413, "loss": 4.4266, "step": 307250 }, { "epoch": 0.5734559679402098, "grad_norm": 1.1053504943847656, "learning_rate": 0.00019693785075560388, "loss": 4.5509, "step": 307300 }, { "epoch": 0.5735492734995883, "grad_norm": 1.2611793279647827, "learning_rate": 0.00019693681902933541, "loss": 4.513, "step": 307350 }, { "epoch": 0.5736425790589668, "grad_norm": 0.9037719964981079, "learning_rate": 0.0001969357871319905, "loss": 4.367, "step": 307400 }, { "epoch": 0.5737358846183452, "grad_norm": 0.9977772235870361, "learning_rate": 0.00019693475506357098, "loss": 4.6679, "step": 307450 }, { "epoch": 0.5738291901777237, "grad_norm": 0.8833919167518616, "learning_rate": 0.00019693372282407866, "loss": 4.6624, "step": 307500 }, { "epoch": 0.5739224957371023, "grad_norm": 1.1849719285964966, "learning_rate": 0.0001969326904135154, "loss": 4.6164, "step": 307550 }, { "epoch": 0.5740158012964808, "grad_norm": 1.1160576343536377, "learning_rate": 0.00019693165783188298, "loss": 4.5444, "step": 307600 }, { "epoch": 0.5741091068558593, "grad_norm": 0.7999420166015625, "learning_rate": 0.00019693062507918324, "loss": 4.4761, "step": 307650 }, { "epoch": 0.5742024124152377, "grad_norm": 1.1961227655410767, "learning_rate": 0.000196929592155418, "loss": 4.5654, "step": 307700 }, { "epoch": 0.5742957179746162, "grad_norm": 1.1302611827850342, "learning_rate": 0.0001969285590605891, "loss": 4.532, "step": 307750 }, { "epoch": 0.5743890235339947, "grad_norm": 1.0442041158676147, "learning_rate": 0.00019692752579469834, "loss": 4.4396, "step": 307800 }, { "epoch": 0.5744823290933732, "grad_norm": 0.9419299364089966, "learning_rate": 0.00019692649235774755, "loss": 4.6001, "step": 307850 }, { "epoch": 0.5745756346527517, "grad_norm": 1.114232063293457, "learning_rate": 0.00019692545874973858, "loss": 4.5381, "step": 307900 }, { "epoch": 0.5746689402121302, "grad_norm": 1.1545286178588867, "learning_rate": 0.0001969244249706732, "loss": 4.5677, "step": 307950 }, { "epoch": 0.5747622457715087, "grad_norm": 1.051513433456421, "learning_rate": 0.00019692339102055328, "loss": 4.5663, "step": 308000 }, { "epoch": 0.5748555513308872, "grad_norm": 1.200150489807129, "learning_rate": 0.0001969223568993806, "loss": 4.5208, "step": 308050 }, { "epoch": 0.5749488568902656, "grad_norm": 0.8030568361282349, "learning_rate": 0.00019692132260715702, "loss": 4.344, "step": 308100 }, { "epoch": 0.5750421624496441, "grad_norm": 1.1953867673873901, "learning_rate": 0.00019692028814388438, "loss": 4.4139, "step": 308150 }, { "epoch": 0.5751354680090226, "grad_norm": 0.8794069886207581, "learning_rate": 0.00019691925350956449, "loss": 4.553, "step": 308200 }, { "epoch": 0.5752287735684012, "grad_norm": 1.015227198600769, "learning_rate": 0.00019691821870419917, "loss": 4.7194, "step": 308250 }, { "epoch": 0.5753220791277797, "grad_norm": 1.3813424110412598, "learning_rate": 0.00019691718372779024, "loss": 4.4443, "step": 308300 }, { "epoch": 0.5754153846871581, "grad_norm": 0.9208297729492188, "learning_rate": 0.00019691614858033954, "loss": 4.6475, "step": 308350 }, { "epoch": 0.5755086902465366, "grad_norm": 1.0939719676971436, "learning_rate": 0.00019691511326184888, "loss": 4.5593, "step": 308400 }, { "epoch": 0.5756019958059151, "grad_norm": 1.1301640272140503, "learning_rate": 0.0001969140777723201, "loss": 4.4633, "step": 308450 }, { "epoch": 0.5756953013652936, "grad_norm": 1.2512931823730469, "learning_rate": 0.000196913042111755, "loss": 4.6935, "step": 308500 }, { "epoch": 0.575788606924672, "grad_norm": 1.0407778024673462, "learning_rate": 0.0001969120062801555, "loss": 4.5598, "step": 308550 }, { "epoch": 0.5758819124840506, "grad_norm": 1.2233508825302124, "learning_rate": 0.0001969109702775233, "loss": 4.635, "step": 308600 }, { "epoch": 0.5759752180434291, "grad_norm": 1.032896637916565, "learning_rate": 0.0001969099341038603, "loss": 4.7609, "step": 308650 }, { "epoch": 0.5760685236028076, "grad_norm": 0.5946270227432251, "learning_rate": 0.00019690889775916834, "loss": 4.6175, "step": 308700 }, { "epoch": 0.5761618291621861, "grad_norm": 0.6330205798149109, "learning_rate": 0.00019690786124344921, "loss": 4.5726, "step": 308750 }, { "epoch": 0.5762551347215645, "grad_norm": 1.0800895690917969, "learning_rate": 0.00019690682455670476, "loss": 4.4216, "step": 308800 }, { "epoch": 0.576348440280943, "grad_norm": 0.8748204708099365, "learning_rate": 0.00019690578769893682, "loss": 4.668, "step": 308850 }, { "epoch": 0.5764417458403215, "grad_norm": 0.9928598403930664, "learning_rate": 0.00019690475067014721, "loss": 4.5196, "step": 308900 }, { "epoch": 0.5765350513997001, "grad_norm": 1.1503949165344238, "learning_rate": 0.00019690371347033775, "loss": 4.5349, "step": 308950 }, { "epoch": 0.5766283569590785, "grad_norm": 0.8418891429901123, "learning_rate": 0.00019690267609951035, "loss": 4.5646, "step": 309000 }, { "epoch": 0.5766283569590785, "eval_loss": 4.752696990966797, "eval_runtime": 228.2633, "eval_samples_per_second": 11.425, "eval_steps_per_second": 11.425, "eval_tts_loss": 7.530654103705157, "step": 309000 }, { "epoch": 0.576721662518457, "grad_norm": 0.988182544708252, "learning_rate": 0.0001969016385576667, "loss": 4.5647, "step": 309050 }, { "epoch": 0.5768149680778355, "grad_norm": 1.031363606452942, "learning_rate": 0.00019690060084480874, "loss": 4.3385, "step": 309100 }, { "epoch": 0.576908273637214, "grad_norm": 0.8367418646812439, "learning_rate": 0.00019689956296093828, "loss": 4.6036, "step": 309150 }, { "epoch": 0.5770015791965925, "grad_norm": 1.2426990270614624, "learning_rate": 0.00019689852490605713, "loss": 4.3699, "step": 309200 }, { "epoch": 0.5770948847559709, "grad_norm": 1.0288184881210327, "learning_rate": 0.0001968974866801671, "loss": 4.5097, "step": 309250 }, { "epoch": 0.5771881903153495, "grad_norm": 1.1623977422714233, "learning_rate": 0.00019689644828327013, "loss": 4.5612, "step": 309300 }, { "epoch": 0.577281495874728, "grad_norm": 1.1653733253479004, "learning_rate": 0.00019689540971536796, "loss": 4.5268, "step": 309350 }, { "epoch": 0.5773748014341065, "grad_norm": 1.2442301511764526, "learning_rate": 0.0001968943709764624, "loss": 4.5712, "step": 309400 }, { "epoch": 0.5774681069934849, "grad_norm": 1.1720430850982666, "learning_rate": 0.00019689333206655536, "loss": 4.6003, "step": 309450 }, { "epoch": 0.5775614125528634, "grad_norm": 1.1067763566970825, "learning_rate": 0.00019689229298564863, "loss": 4.699, "step": 309500 }, { "epoch": 0.5776547181122419, "grad_norm": 0.7484050989151001, "learning_rate": 0.000196891253733744, "loss": 4.6018, "step": 309550 }, { "epoch": 0.5777480236716204, "grad_norm": 1.0562546253204346, "learning_rate": 0.00019689021431084343, "loss": 4.714, "step": 309600 }, { "epoch": 0.577841329230999, "grad_norm": 1.1429967880249023, "learning_rate": 0.00019688917471694865, "loss": 4.6596, "step": 309650 }, { "epoch": 0.5779346347903774, "grad_norm": 1.0727895498275757, "learning_rate": 0.0001968881349520615, "loss": 4.6577, "step": 309700 }, { "epoch": 0.5780279403497559, "grad_norm": 0.8976140022277832, "learning_rate": 0.00019688709501618389, "loss": 4.6248, "step": 309750 }, { "epoch": 0.5781212459091344, "grad_norm": 0.8005908727645874, "learning_rate": 0.00019688605490931758, "loss": 4.6077, "step": 309800 }, { "epoch": 0.5782145514685129, "grad_norm": 1.1622130870819092, "learning_rate": 0.00019688501463146444, "loss": 4.6376, "step": 309850 }, { "epoch": 0.5783078570278913, "grad_norm": 0.9488914608955383, "learning_rate": 0.00019688397418262627, "loss": 4.384, "step": 309900 }, { "epoch": 0.5784011625872698, "grad_norm": 1.159583330154419, "learning_rate": 0.00019688293356280495, "loss": 4.7019, "step": 309950 }, { "epoch": 0.5784944681466484, "grad_norm": 0.9173402190208435, "learning_rate": 0.00019688189277200228, "loss": 4.5651, "step": 310000 }, { "epoch": 0.5785877737060269, "grad_norm": 1.0882700681686401, "learning_rate": 0.00019688085181022012, "loss": 4.3936, "step": 310050 }, { "epoch": 0.5786810792654054, "grad_norm": 1.401253581047058, "learning_rate": 0.00019687981067746037, "loss": 4.5729, "step": 310100 }, { "epoch": 0.5787743848247838, "grad_norm": 1.2113391160964966, "learning_rate": 0.00019687876937372472, "loss": 4.4746, "step": 310150 }, { "epoch": 0.5788676903841623, "grad_norm": 0.9598744511604309, "learning_rate": 0.00019687772789901513, "loss": 4.6706, "step": 310200 }, { "epoch": 0.5789609959435408, "grad_norm": 1.0837336778640747, "learning_rate": 0.00019687668625333337, "loss": 4.5265, "step": 310250 }, { "epoch": 0.5790543015029193, "grad_norm": 1.3954722881317139, "learning_rate": 0.00019687564443668132, "loss": 4.3927, "step": 310300 }, { "epoch": 0.5791476070622978, "grad_norm": 0.8956559896469116, "learning_rate": 0.0001968746024490608, "loss": 4.6382, "step": 310350 }, { "epoch": 0.5792409126216763, "grad_norm": 1.6203763484954834, "learning_rate": 0.00019687356029047365, "loss": 4.6029, "step": 310400 }, { "epoch": 0.5793342181810548, "grad_norm": 0.983693540096283, "learning_rate": 0.00019687251796092175, "loss": 4.5395, "step": 310450 }, { "epoch": 0.5794275237404333, "grad_norm": 1.05988347530365, "learning_rate": 0.00019687147546040685, "loss": 4.3837, "step": 310500 }, { "epoch": 0.5795208292998117, "grad_norm": 1.071925163269043, "learning_rate": 0.00019687043278893084, "loss": 4.553, "step": 310550 }, { "epoch": 0.5796141348591902, "grad_norm": 1.0153021812438965, "learning_rate": 0.0001968693899464956, "loss": 4.3998, "step": 310600 }, { "epoch": 0.5797074404185687, "grad_norm": 1.4140474796295166, "learning_rate": 0.00019686834693310293, "loss": 4.5908, "step": 310650 }, { "epoch": 0.5798007459779473, "grad_norm": 1.0499241352081299, "learning_rate": 0.00019686730374875465, "loss": 4.6639, "step": 310700 }, { "epoch": 0.5798940515373258, "grad_norm": 0.9955544471740723, "learning_rate": 0.00019686626039345262, "loss": 4.8298, "step": 310750 }, { "epoch": 0.5799873570967042, "grad_norm": 1.2786093950271606, "learning_rate": 0.00019686521686719868, "loss": 4.4873, "step": 310800 }, { "epoch": 0.5800806626560827, "grad_norm": 0.96354740858078, "learning_rate": 0.0001968641731699947, "loss": 4.5876, "step": 310850 }, { "epoch": 0.5801739682154612, "grad_norm": 1.0890089273452759, "learning_rate": 0.00019686312930184246, "loss": 4.6488, "step": 310900 }, { "epoch": 0.5802672737748397, "grad_norm": 1.1017330884933472, "learning_rate": 0.00019686208526274385, "loss": 4.6347, "step": 310950 }, { "epoch": 0.5803605793342181, "grad_norm": 1.126384973526001, "learning_rate": 0.00019686104105270073, "loss": 4.5384, "step": 311000 }, { "epoch": 0.5804538848935967, "grad_norm": 0.9543178081512451, "learning_rate": 0.0001968599966717149, "loss": 4.4462, "step": 311050 }, { "epoch": 0.5805471904529752, "grad_norm": 1.119261622428894, "learning_rate": 0.00019685895211978824, "loss": 4.7619, "step": 311100 }, { "epoch": 0.5806404960123537, "grad_norm": 0.9374808073043823, "learning_rate": 0.00019685790739692256, "loss": 4.5444, "step": 311150 }, { "epoch": 0.5807338015717322, "grad_norm": 0.8393676280975342, "learning_rate": 0.0001968568625031197, "loss": 4.7381, "step": 311200 }, { "epoch": 0.5808271071311106, "grad_norm": 0.9694864153862, "learning_rate": 0.00019685581743838154, "loss": 4.584, "step": 311250 }, { "epoch": 0.5809204126904891, "grad_norm": 0.9850889444351196, "learning_rate": 0.00019685477220270987, "loss": 4.578, "step": 311300 }, { "epoch": 0.5810137182498676, "grad_norm": 1.149882197380066, "learning_rate": 0.0001968537267961066, "loss": 4.6047, "step": 311350 }, { "epoch": 0.5811070238092461, "grad_norm": 0.8860505819320679, "learning_rate": 0.00019685268121857354, "loss": 4.6453, "step": 311400 }, { "epoch": 0.5812003293686246, "grad_norm": 0.9555177092552185, "learning_rate": 0.00019685163547011257, "loss": 4.5644, "step": 311450 }, { "epoch": 0.5812936349280031, "grad_norm": 0.7401108145713806, "learning_rate": 0.00019685058955072547, "loss": 4.5516, "step": 311500 }, { "epoch": 0.5813869404873816, "grad_norm": 0.8870165944099426, "learning_rate": 0.00019684954346041412, "loss": 4.4763, "step": 311550 }, { "epoch": 0.5814802460467601, "grad_norm": 1.1874957084655762, "learning_rate": 0.00019684849719918038, "loss": 4.3779, "step": 311600 }, { "epoch": 0.5815735516061386, "grad_norm": 0.9011445045471191, "learning_rate": 0.00019684745076702608, "loss": 4.7085, "step": 311650 }, { "epoch": 0.581666857165517, "grad_norm": 1.0121692419052124, "learning_rate": 0.00019684640416395304, "loss": 4.6841, "step": 311700 }, { "epoch": 0.5817601627248955, "grad_norm": 1.0623122453689575, "learning_rate": 0.00019684535738996318, "loss": 4.337, "step": 311750 }, { "epoch": 0.5818534682842741, "grad_norm": 1.0233830213546753, "learning_rate": 0.00019684431044505828, "loss": 4.4701, "step": 311800 }, { "epoch": 0.5819467738436526, "grad_norm": 0.7241932153701782, "learning_rate": 0.0001968432633292402, "loss": 4.3979, "step": 311850 }, { "epoch": 0.582040079403031, "grad_norm": 1.0669363737106323, "learning_rate": 0.00019684221604251084, "loss": 4.6099, "step": 311900 }, { "epoch": 0.5821333849624095, "grad_norm": 1.0737048387527466, "learning_rate": 0.00019684116858487198, "loss": 4.6644, "step": 311950 }, { "epoch": 0.582226690521788, "grad_norm": 1.3133255243301392, "learning_rate": 0.0001968401209563255, "loss": 4.4478, "step": 312000 }, { "epoch": 0.582226690521788, "eval_loss": 4.749062538146973, "eval_runtime": 231.04, "eval_samples_per_second": 11.288, "eval_steps_per_second": 11.288, "eval_tts_loss": 7.521649108948243, "step": 312000 }, { "epoch": 0.5823199960811665, "grad_norm": 1.216781497001648, "learning_rate": 0.00019683907315687328, "loss": 4.5382, "step": 312050 }, { "epoch": 0.582413301640545, "grad_norm": 1.1261478662490845, "learning_rate": 0.00019683802518651709, "loss": 4.5286, "step": 312100 }, { "epoch": 0.5825066071999235, "grad_norm": 0.9089665412902832, "learning_rate": 0.00019683697704525884, "loss": 4.5094, "step": 312150 }, { "epoch": 0.582599912759302, "grad_norm": 0.9918016791343689, "learning_rate": 0.00019683592873310035, "loss": 4.7113, "step": 312200 }, { "epoch": 0.5826932183186805, "grad_norm": 1.0284326076507568, "learning_rate": 0.00019683488025004352, "loss": 4.6837, "step": 312250 }, { "epoch": 0.582786523878059, "grad_norm": 0.9192066192626953, "learning_rate": 0.00019683383159609013, "loss": 4.6697, "step": 312300 }, { "epoch": 0.5828798294374374, "grad_norm": 0.8472697138786316, "learning_rate": 0.0001968327827712421, "loss": 4.627, "step": 312350 }, { "epoch": 0.5829731349968159, "grad_norm": 1.0066317319869995, "learning_rate": 0.0001968317337755012, "loss": 4.6281, "step": 312400 }, { "epoch": 0.5830664405561944, "grad_norm": 0.9507394433021545, "learning_rate": 0.00019683068460886935, "loss": 4.6588, "step": 312450 }, { "epoch": 0.583159746115573, "grad_norm": 1.2607237100601196, "learning_rate": 0.0001968296352713484, "loss": 4.4912, "step": 312500 }, { "epoch": 0.5832530516749515, "grad_norm": 1.0062910318374634, "learning_rate": 0.00019682858576294015, "loss": 4.6857, "step": 312550 }, { "epoch": 0.5833463572343299, "grad_norm": 0.8089048862457275, "learning_rate": 0.0001968275360836465, "loss": 4.6043, "step": 312600 }, { "epoch": 0.5834396627937084, "grad_norm": 0.5660948753356934, "learning_rate": 0.0001968264862334693, "loss": 4.4691, "step": 312650 }, { "epoch": 0.5835329683530869, "grad_norm": 1.0028718709945679, "learning_rate": 0.00019682543621241035, "loss": 4.6358, "step": 312700 }, { "epoch": 0.5836262739124654, "grad_norm": 1.0709588527679443, "learning_rate": 0.0001968243860204716, "loss": 4.5166, "step": 312750 }, { "epoch": 0.5837195794718438, "grad_norm": 0.80045086145401, "learning_rate": 0.0001968233356576548, "loss": 4.611, "step": 312800 }, { "epoch": 0.5838128850312224, "grad_norm": 1.1337125301361084, "learning_rate": 0.00019682228512396188, "loss": 4.5636, "step": 312850 }, { "epoch": 0.5839061905906009, "grad_norm": 1.2380632162094116, "learning_rate": 0.00019682123441939464, "loss": 4.5597, "step": 312900 }, { "epoch": 0.5839994961499794, "grad_norm": 1.3010809421539307, "learning_rate": 0.00019682018354395497, "loss": 4.5734, "step": 312950 }, { "epoch": 0.5840928017093578, "grad_norm": 1.0648142099380493, "learning_rate": 0.0001968191324976447, "loss": 4.5486, "step": 313000 }, { "epoch": 0.5841861072687363, "grad_norm": 0.9465823173522949, "learning_rate": 0.00019681808128046567, "loss": 4.5559, "step": 313050 }, { "epoch": 0.5842794128281148, "grad_norm": 1.0052610635757446, "learning_rate": 0.00019681702989241982, "loss": 4.7299, "step": 313100 }, { "epoch": 0.5843727183874933, "grad_norm": 1.0187366008758545, "learning_rate": 0.00019681597833350894, "loss": 4.6535, "step": 313150 }, { "epoch": 0.5844660239468719, "grad_norm": 0.8939917087554932, "learning_rate": 0.00019681492660373487, "loss": 4.6603, "step": 313200 }, { "epoch": 0.5845593295062503, "grad_norm": 1.0251716375350952, "learning_rate": 0.00019681387470309948, "loss": 4.7555, "step": 313250 }, { "epoch": 0.5846526350656288, "grad_norm": 0.9094771146774292, "learning_rate": 0.00019681282263160467, "loss": 4.6101, "step": 313300 }, { "epoch": 0.5847459406250073, "grad_norm": 0.9584060907363892, "learning_rate": 0.00019681177038925223, "loss": 4.7022, "step": 313350 }, { "epoch": 0.5848392461843858, "grad_norm": 0.9828512668609619, "learning_rate": 0.00019681071797604408, "loss": 4.6973, "step": 313400 }, { "epoch": 0.5849325517437642, "grad_norm": 1.157536506652832, "learning_rate": 0.000196809665391982, "loss": 4.3445, "step": 313450 }, { "epoch": 0.5850258573031427, "grad_norm": 1.133435606956482, "learning_rate": 0.00019680861263706792, "loss": 4.5914, "step": 313500 }, { "epoch": 0.5851191628625213, "grad_norm": 0.9810517430305481, "learning_rate": 0.00019680755971130371, "loss": 4.4814, "step": 313550 }, { "epoch": 0.5852124684218998, "grad_norm": 0.7324022650718689, "learning_rate": 0.00019680650661469114, "loss": 4.5606, "step": 313600 }, { "epoch": 0.5853057739812783, "grad_norm": 1.1952625513076782, "learning_rate": 0.00019680545334723216, "loss": 4.6152, "step": 313650 }, { "epoch": 0.5853990795406567, "grad_norm": 1.237855315208435, "learning_rate": 0.00019680439990892853, "loss": 4.4655, "step": 313700 }, { "epoch": 0.5854923851000352, "grad_norm": 1.2508081197738647, "learning_rate": 0.0001968033462997822, "loss": 4.5237, "step": 313750 }, { "epoch": 0.5855856906594137, "grad_norm": 0.8722935318946838, "learning_rate": 0.00019680229251979503, "loss": 4.6543, "step": 313800 }, { "epoch": 0.5856789962187922, "grad_norm": 1.2657021284103394, "learning_rate": 0.0001968012385689688, "loss": 4.6198, "step": 313850 }, { "epoch": 0.5857723017781707, "grad_norm": 1.203904151916504, "learning_rate": 0.00019680018444730542, "loss": 4.5212, "step": 313900 }, { "epoch": 0.5858656073375492, "grad_norm": 1.0725904703140259, "learning_rate": 0.0001967991301548068, "loss": 4.5757, "step": 313950 }, { "epoch": 0.5859589128969277, "grad_norm": 1.0794929265975952, "learning_rate": 0.00019679807569147472, "loss": 4.3468, "step": 314000 }, { "epoch": 0.5860522184563062, "grad_norm": 1.049497127532959, "learning_rate": 0.00019679702105731103, "loss": 4.7022, "step": 314050 }, { "epoch": 0.5861455240156846, "grad_norm": 1.3520044088363647, "learning_rate": 0.00019679596625231768, "loss": 4.418, "step": 314100 }, { "epoch": 0.5862388295750631, "grad_norm": 1.1205198764801025, "learning_rate": 0.00019679491127649646, "loss": 4.5213, "step": 314150 }, { "epoch": 0.5863321351344416, "grad_norm": 0.9296178221702576, "learning_rate": 0.00019679385612984924, "loss": 4.578, "step": 314200 }, { "epoch": 0.5864254406938202, "grad_norm": 1.0487147569656372, "learning_rate": 0.0001967928008123779, "loss": 4.5632, "step": 314250 }, { "epoch": 0.5865187462531987, "grad_norm": 1.1871947050094604, "learning_rate": 0.0001967917453240843, "loss": 4.6728, "step": 314300 }, { "epoch": 0.5866120518125771, "grad_norm": 0.8288167715072632, "learning_rate": 0.0001967906896649703, "loss": 4.598, "step": 314350 }, { "epoch": 0.5867053573719556, "grad_norm": 1.000169277191162, "learning_rate": 0.0001967896338350378, "loss": 4.49, "step": 314400 }, { "epoch": 0.5867986629313341, "grad_norm": 1.1729748249053955, "learning_rate": 0.00019678857783428857, "loss": 4.6127, "step": 314450 }, { "epoch": 0.5868919684907126, "grad_norm": 1.2053636312484741, "learning_rate": 0.00019678752166272457, "loss": 4.5601, "step": 314500 }, { "epoch": 0.586985274050091, "grad_norm": 0.915949821472168, "learning_rate": 0.0001967864653203476, "loss": 4.4926, "step": 314550 }, { "epoch": 0.5870785796094696, "grad_norm": 1.1071586608886719, "learning_rate": 0.00019678540880715956, "loss": 4.5221, "step": 314600 }, { "epoch": 0.5871718851688481, "grad_norm": 0.8765965700149536, "learning_rate": 0.0001967843521231623, "loss": 4.4275, "step": 314650 }, { "epoch": 0.5872651907282266, "grad_norm": 1.1023027896881104, "learning_rate": 0.0001967832952683577, "loss": 4.5292, "step": 314700 }, { "epoch": 0.5873584962876051, "grad_norm": 1.0290693044662476, "learning_rate": 0.0001967822382427476, "loss": 4.6013, "step": 314750 }, { "epoch": 0.5874518018469835, "grad_norm": 1.0654975175857544, "learning_rate": 0.0001967811810463339, "loss": 4.6032, "step": 314800 }, { "epoch": 0.587545107406362, "grad_norm": 1.0474263429641724, "learning_rate": 0.00019678012367911843, "loss": 4.4041, "step": 314850 }, { "epoch": 0.5876384129657405, "grad_norm": 1.2105765342712402, "learning_rate": 0.00019677906614110307, "loss": 4.4584, "step": 314900 }, { "epoch": 0.5877317185251191, "grad_norm": 1.050726056098938, "learning_rate": 0.00019677800843228966, "loss": 4.5704, "step": 314950 }, { "epoch": 0.5878250240844975, "grad_norm": 0.7779209017753601, "learning_rate": 0.00019677695055268015, "loss": 4.6361, "step": 315000 }, { "epoch": 0.5878250240844975, "eval_loss": 4.74893856048584, "eval_runtime": 233.0687, "eval_samples_per_second": 11.19, "eval_steps_per_second": 11.19, "eval_tts_loss": 7.577803379387532, "step": 315000 }, { "epoch": 0.587918329643876, "grad_norm": 1.0922472476959229, "learning_rate": 0.0001967758925022763, "loss": 4.5707, "step": 315050 }, { "epoch": 0.5880116352032545, "grad_norm": 1.1798323392868042, "learning_rate": 0.00019677483428108005, "loss": 4.5508, "step": 315100 }, { "epoch": 0.588104940762633, "grad_norm": 1.1150181293487549, "learning_rate": 0.00019677377588909326, "loss": 4.6886, "step": 315150 }, { "epoch": 0.5881982463220115, "grad_norm": 1.1528747081756592, "learning_rate": 0.00019677271732631776, "loss": 4.5604, "step": 315200 }, { "epoch": 0.5882915518813899, "grad_norm": 1.16209876537323, "learning_rate": 0.00019677165859275544, "loss": 4.6568, "step": 315250 }, { "epoch": 0.5883848574407685, "grad_norm": 1.0389351844787598, "learning_rate": 0.00019677059968840817, "loss": 4.5282, "step": 315300 }, { "epoch": 0.588478163000147, "grad_norm": 1.0864149332046509, "learning_rate": 0.00019676954061327787, "loss": 4.539, "step": 315350 }, { "epoch": 0.5885714685595255, "grad_norm": 0.9973792433738708, "learning_rate": 0.0001967684813673663, "loss": 4.5436, "step": 315400 }, { "epoch": 0.5886647741189039, "grad_norm": 1.0296919345855713, "learning_rate": 0.0001967674219506754, "loss": 4.7594, "step": 315450 }, { "epoch": 0.5887580796782824, "grad_norm": 1.2115896940231323, "learning_rate": 0.00019676636236320703, "loss": 4.676, "step": 315500 }, { "epoch": 0.5888513852376609, "grad_norm": 1.299649715423584, "learning_rate": 0.00019676530260496304, "loss": 4.6847, "step": 315550 }, { "epoch": 0.5889446907970394, "grad_norm": 1.1729029417037964, "learning_rate": 0.00019676424267594535, "loss": 4.3708, "step": 315600 }, { "epoch": 0.589037996356418, "grad_norm": 1.0936784744262695, "learning_rate": 0.00019676318257615577, "loss": 4.6959, "step": 315650 }, { "epoch": 0.5891313019157964, "grad_norm": 0.9710637331008911, "learning_rate": 0.00019676212230559622, "loss": 4.5585, "step": 315700 }, { "epoch": 0.5892246074751749, "grad_norm": 1.055420160293579, "learning_rate": 0.00019676106186426856, "loss": 4.7101, "step": 315750 }, { "epoch": 0.5893179130345534, "grad_norm": 1.2843619585037231, "learning_rate": 0.0001967600012521746, "loss": 4.4441, "step": 315800 }, { "epoch": 0.5894112185939319, "grad_norm": 1.2306737899780273, "learning_rate": 0.0001967589404693163, "loss": 4.4583, "step": 315850 }, { "epoch": 0.5895045241533103, "grad_norm": 1.2592270374298096, "learning_rate": 0.0001967578795156955, "loss": 4.6707, "step": 315900 }, { "epoch": 0.5895978297126888, "grad_norm": 1.2055423259735107, "learning_rate": 0.00019675681839131408, "loss": 4.4714, "step": 315950 }, { "epoch": 0.5896911352720674, "grad_norm": 0.9702626466751099, "learning_rate": 0.00019675575709617388, "loss": 4.5387, "step": 316000 }, { "epoch": 0.5897844408314459, "grad_norm": 0.9379885196685791, "learning_rate": 0.00019675469563027676, "loss": 4.5659, "step": 316050 }, { "epoch": 0.5898777463908244, "grad_norm": 0.9110227227210999, "learning_rate": 0.00019675363399362467, "loss": 4.5908, "step": 316100 }, { "epoch": 0.5899710519502028, "grad_norm": 1.2341092824935913, "learning_rate": 0.00019675257218621943, "loss": 4.4516, "step": 316150 }, { "epoch": 0.5900643575095813, "grad_norm": 1.1738377809524536, "learning_rate": 0.00019675151020806292, "loss": 4.4439, "step": 316200 }, { "epoch": 0.5901576630689598, "grad_norm": 0.6892131567001343, "learning_rate": 0.00019675044805915704, "loss": 4.4533, "step": 316250 }, { "epoch": 0.5902509686283383, "grad_norm": 0.9237385988235474, "learning_rate": 0.00019674938573950362, "loss": 4.479, "step": 316300 }, { "epoch": 0.5903442741877168, "grad_norm": 1.0071806907653809, "learning_rate": 0.00019674832324910457, "loss": 4.6307, "step": 316350 }, { "epoch": 0.5904375797470953, "grad_norm": 0.7943397164344788, "learning_rate": 0.00019674726058796174, "loss": 4.3602, "step": 316400 }, { "epoch": 0.5905308853064738, "grad_norm": 0.9103261828422546, "learning_rate": 0.00019674619775607703, "loss": 4.6764, "step": 316450 }, { "epoch": 0.5906241908658523, "grad_norm": 0.9532055854797363, "learning_rate": 0.00019674513475345228, "loss": 4.5013, "step": 316500 }, { "epoch": 0.5907174964252307, "grad_norm": 0.7558811902999878, "learning_rate": 0.0001967440715800894, "loss": 4.5183, "step": 316550 }, { "epoch": 0.5908108019846092, "grad_norm": 1.2909374237060547, "learning_rate": 0.00019674300823599027, "loss": 4.6535, "step": 316600 }, { "epoch": 0.5909041075439877, "grad_norm": 1.1105871200561523, "learning_rate": 0.00019674194472115678, "loss": 4.384, "step": 316650 }, { "epoch": 0.5909974131033662, "grad_norm": 1.0830601453781128, "learning_rate": 0.00019674088103559072, "loss": 4.38, "step": 316700 }, { "epoch": 0.5910907186627448, "grad_norm": 0.8381434679031372, "learning_rate": 0.00019673981717929407, "loss": 4.4828, "step": 316750 }, { "epoch": 0.5911840242221232, "grad_norm": 0.9999323487281799, "learning_rate": 0.00019673875315226865, "loss": 4.532, "step": 316800 }, { "epoch": 0.5912773297815017, "grad_norm": 0.8143675923347473, "learning_rate": 0.00019673768895451637, "loss": 4.5682, "step": 316850 }, { "epoch": 0.5913706353408802, "grad_norm": 1.1459548473358154, "learning_rate": 0.00019673662458603907, "loss": 4.4766, "step": 316900 }, { "epoch": 0.5914639409002587, "grad_norm": 0.8771162033081055, "learning_rate": 0.00019673556004683862, "loss": 4.6515, "step": 316950 }, { "epoch": 0.5915572464596371, "grad_norm": 1.024323582649231, "learning_rate": 0.00019673449533691697, "loss": 4.692, "step": 317000 }, { "epoch": 0.5916505520190156, "grad_norm": 1.246565818786621, "learning_rate": 0.00019673343045627595, "loss": 4.6446, "step": 317050 }, { "epoch": 0.5917438575783942, "grad_norm": 0.9760048389434814, "learning_rate": 0.00019673236540491745, "loss": 4.7288, "step": 317100 }, { "epoch": 0.5918371631377727, "grad_norm": 1.1647191047668457, "learning_rate": 0.00019673130018284333, "loss": 4.4353, "step": 317150 }, { "epoch": 0.5919304686971512, "grad_norm": 1.019235372543335, "learning_rate": 0.00019673023479005548, "loss": 4.612, "step": 317200 }, { "epoch": 0.5920237742565296, "grad_norm": 0.9864053726196289, "learning_rate": 0.00019672916922655582, "loss": 4.7742, "step": 317250 }, { "epoch": 0.5921170798159081, "grad_norm": 1.1842503547668457, "learning_rate": 0.00019672810349234617, "loss": 4.5876, "step": 317300 }, { "epoch": 0.5922103853752866, "grad_norm": 1.0289043188095093, "learning_rate": 0.00019672703758742842, "loss": 4.706, "step": 317350 }, { "epoch": 0.5923036909346651, "grad_norm": 0.9660289287567139, "learning_rate": 0.00019672597151180451, "loss": 4.4939, "step": 317400 }, { "epoch": 0.5923969964940436, "grad_norm": 1.1275559663772583, "learning_rate": 0.00019672490526547628, "loss": 4.517, "step": 317450 }, { "epoch": 0.5924903020534221, "grad_norm": 1.21570885181427, "learning_rate": 0.00019672383884844556, "loss": 4.8104, "step": 317500 }, { "epoch": 0.5925836076128006, "grad_norm": 0.8104692697525024, "learning_rate": 0.0001967227722607143, "loss": 4.3482, "step": 317550 }, { "epoch": 0.5926769131721791, "grad_norm": 1.08836030960083, "learning_rate": 0.0001967217055022844, "loss": 4.5051, "step": 317600 }, { "epoch": 0.5927702187315576, "grad_norm": 1.0493236780166626, "learning_rate": 0.0001967206385731577, "loss": 4.6316, "step": 317650 }, { "epoch": 0.592863524290936, "grad_norm": 0.74932861328125, "learning_rate": 0.00019671957147333609, "loss": 4.4358, "step": 317700 }, { "epoch": 0.5929568298503145, "grad_norm": 1.2372153997421265, "learning_rate": 0.00019671850420282142, "loss": 4.5856, "step": 317750 }, { "epoch": 0.5930501354096931, "grad_norm": 1.0992337465286255, "learning_rate": 0.00019671743676161562, "loss": 4.5747, "step": 317800 }, { "epoch": 0.5931434409690716, "grad_norm": 1.0623835325241089, "learning_rate": 0.0001967163691497206, "loss": 4.4083, "step": 317850 }, { "epoch": 0.59323674652845, "grad_norm": 0.9603080749511719, "learning_rate": 0.00019671530136713817, "loss": 4.6688, "step": 317900 }, { "epoch": 0.5933300520878285, "grad_norm": 0.849819540977478, "learning_rate": 0.00019671423341387028, "loss": 4.5017, "step": 317950 }, { "epoch": 0.593423357647207, "grad_norm": 0.8867913484573364, "learning_rate": 0.00019671316528991874, "loss": 4.4862, "step": 318000 }, { "epoch": 0.593423357647207, "eval_loss": 4.749143123626709, "eval_runtime": 230.0262, "eval_samples_per_second": 11.338, "eval_steps_per_second": 11.338, "eval_tts_loss": 7.517176300315815, "step": 318000 }, { "epoch": 0.5935166632065855, "grad_norm": 0.8609456419944763, "learning_rate": 0.0001967120969952855, "loss": 4.5275, "step": 318050 }, { "epoch": 0.593609968765964, "grad_norm": 1.0815144777297974, "learning_rate": 0.0001967110285299725, "loss": 4.7102, "step": 318100 }, { "epoch": 0.5937032743253425, "grad_norm": 0.7996091842651367, "learning_rate": 0.00019670995989398146, "loss": 4.6476, "step": 318150 }, { "epoch": 0.593796579884721, "grad_norm": 1.0403865575790405, "learning_rate": 0.00019670889108731437, "loss": 4.4335, "step": 318200 }, { "epoch": 0.5938898854440995, "grad_norm": 1.0239715576171875, "learning_rate": 0.00019670782210997313, "loss": 4.6901, "step": 318250 }, { "epoch": 0.593983191003478, "grad_norm": 1.121938943862915, "learning_rate": 0.00019670675296195957, "loss": 4.4311, "step": 318300 }, { "epoch": 0.5940764965628564, "grad_norm": 1.120389699935913, "learning_rate": 0.00019670568364327565, "loss": 4.6799, "step": 318350 }, { "epoch": 0.5941698021222349, "grad_norm": 1.1791412830352783, "learning_rate": 0.00019670461415392319, "loss": 4.7471, "step": 318400 }, { "epoch": 0.5942631076816134, "grad_norm": 1.0191036462783813, "learning_rate": 0.0001967035444939041, "loss": 4.6189, "step": 318450 }, { "epoch": 0.594356413240992, "grad_norm": 1.2128996849060059, "learning_rate": 0.00019670247466322026, "loss": 4.5339, "step": 318500 }, { "epoch": 0.5944497188003705, "grad_norm": 0.9911199808120728, "learning_rate": 0.00019670140466187356, "loss": 4.576, "step": 318550 }, { "epoch": 0.5945430243597489, "grad_norm": 1.3055840730667114, "learning_rate": 0.0001967003344898659, "loss": 4.5811, "step": 318600 }, { "epoch": 0.5946363299191274, "grad_norm": 0.8344447016716003, "learning_rate": 0.0001966992641471992, "loss": 4.823, "step": 318650 }, { "epoch": 0.5947296354785059, "grad_norm": 1.0247474908828735, "learning_rate": 0.0001966981936338753, "loss": 4.6469, "step": 318700 }, { "epoch": 0.5948229410378844, "grad_norm": 0.9239888787269592, "learning_rate": 0.00019669712294989608, "loss": 4.5128, "step": 318750 }, { "epoch": 0.5949162465972628, "grad_norm": 0.8752967715263367, "learning_rate": 0.00019669605209526346, "loss": 4.7802, "step": 318800 }, { "epoch": 0.5950095521566414, "grad_norm": 1.1598962545394897, "learning_rate": 0.00019669498106997932, "loss": 4.5692, "step": 318850 }, { "epoch": 0.5951028577160199, "grad_norm": 1.3275585174560547, "learning_rate": 0.00019669390987404558, "loss": 4.6229, "step": 318900 }, { "epoch": 0.5951961632753984, "grad_norm": 1.271799921989441, "learning_rate": 0.00019669283850746409, "loss": 4.5638, "step": 318950 }, { "epoch": 0.5952894688347768, "grad_norm": 1.048512578010559, "learning_rate": 0.00019669176697023676, "loss": 4.6437, "step": 319000 }, { "epoch": 0.5953827743941553, "grad_norm": 0.7832193970680237, "learning_rate": 0.00019669069526236546, "loss": 4.3553, "step": 319050 }, { "epoch": 0.5954760799535338, "grad_norm": 0.8954023718833923, "learning_rate": 0.00019668962338385208, "loss": 4.5636, "step": 319100 }, { "epoch": 0.5955693855129123, "grad_norm": 1.0028667449951172, "learning_rate": 0.00019668855133469856, "loss": 4.5156, "step": 319150 }, { "epoch": 0.5956626910722909, "grad_norm": 1.0486000776290894, "learning_rate": 0.00019668747911490673, "loss": 4.242, "step": 319200 }, { "epoch": 0.5957559966316693, "grad_norm": 0.7862451076507568, "learning_rate": 0.00019668640672447854, "loss": 4.3009, "step": 319250 }, { "epoch": 0.5958493021910478, "grad_norm": 0.964422345161438, "learning_rate": 0.00019668533416341584, "loss": 4.5527, "step": 319300 }, { "epoch": 0.5959426077504263, "grad_norm": 0.8378289341926575, "learning_rate": 0.00019668426143172054, "loss": 4.5578, "step": 319350 }, { "epoch": 0.5960359133098048, "grad_norm": 0.9096817374229431, "learning_rate": 0.00019668318852939453, "loss": 4.6013, "step": 319400 }, { "epoch": 0.5961292188691832, "grad_norm": 1.119133472442627, "learning_rate": 0.0001966821154564397, "loss": 4.6927, "step": 319450 }, { "epoch": 0.5962225244285617, "grad_norm": 1.016858458518982, "learning_rate": 0.00019668104221285795, "loss": 4.6666, "step": 319500 }, { "epoch": 0.5963158299879403, "grad_norm": 0.9536043405532837, "learning_rate": 0.00019667996879865117, "loss": 4.735, "step": 319550 }, { "epoch": 0.5964091355473188, "grad_norm": 1.2586673498153687, "learning_rate": 0.00019667889521382127, "loss": 4.7085, "step": 319600 }, { "epoch": 0.5965024411066973, "grad_norm": 1.001456379890442, "learning_rate": 0.00019667782145837013, "loss": 4.5489, "step": 319650 }, { "epoch": 0.5965957466660757, "grad_norm": 1.457856297492981, "learning_rate": 0.00019667674753229964, "loss": 4.5942, "step": 319700 }, { "epoch": 0.5966890522254542, "grad_norm": 1.1986068487167358, "learning_rate": 0.00019667567343561172, "loss": 4.6471, "step": 319750 }, { "epoch": 0.5967823577848327, "grad_norm": 0.9569765329360962, "learning_rate": 0.00019667459916830818, "loss": 4.6439, "step": 319800 }, { "epoch": 0.5968756633442112, "grad_norm": 1.1844605207443237, "learning_rate": 0.00019667352473039107, "loss": 4.6101, "step": 319850 }, { "epoch": 0.5969689689035897, "grad_norm": 1.2512829303741455, "learning_rate": 0.00019667245012186215, "loss": 4.5151, "step": 319900 }, { "epoch": 0.5970622744629682, "grad_norm": 1.1351796388626099, "learning_rate": 0.00019667137534272336, "loss": 4.2293, "step": 319950 }, { "epoch": 0.5971555800223467, "grad_norm": 1.0945556163787842, "learning_rate": 0.00019667030039297663, "loss": 4.6776, "step": 320000 }, { "epoch": 0.5972488855817252, "grad_norm": 0.9205471277236938, "learning_rate": 0.0001966692252726238, "loss": 4.3958, "step": 320050 }, { "epoch": 0.5973421911411037, "grad_norm": 1.8201923370361328, "learning_rate": 0.00019666814998166684, "loss": 4.5612, "step": 320100 }, { "epoch": 0.5974354967004821, "grad_norm": 0.9468337297439575, "learning_rate": 0.00019666707452010757, "loss": 4.7242, "step": 320150 }, { "epoch": 0.5975288022598606, "grad_norm": 1.0803794860839844, "learning_rate": 0.00019666599888794792, "loss": 4.5611, "step": 320200 }, { "epoch": 0.5976221078192392, "grad_norm": 1.1365025043487549, "learning_rate": 0.0001966649230851898, "loss": 4.4926, "step": 320250 }, { "epoch": 0.5977154133786177, "grad_norm": 0.9078410267829895, "learning_rate": 0.00019666384711183513, "loss": 4.7069, "step": 320300 }, { "epoch": 0.5978087189379961, "grad_norm": 0.7499719858169556, "learning_rate": 0.00019666277096788573, "loss": 4.4693, "step": 320350 }, { "epoch": 0.5979020244973746, "grad_norm": 1.1963690519332886, "learning_rate": 0.00019666169465334357, "loss": 4.6755, "step": 320400 }, { "epoch": 0.5979953300567531, "grad_norm": 1.081960678100586, "learning_rate": 0.0001966606181682105, "loss": 4.718, "step": 320450 }, { "epoch": 0.5980886356161316, "grad_norm": 1.1670255661010742, "learning_rate": 0.00019665954151248848, "loss": 4.4774, "step": 320500 }, { "epoch": 0.59818194117551, "grad_norm": 0.851716935634613, "learning_rate": 0.00019665846468617937, "loss": 4.6525, "step": 320550 }, { "epoch": 0.5982752467348886, "grad_norm": 1.07505464553833, "learning_rate": 0.00019665738768928506, "loss": 4.5536, "step": 320600 }, { "epoch": 0.5983685522942671, "grad_norm": 1.0349235534667969, "learning_rate": 0.0001966563105218075, "loss": 4.3931, "step": 320650 }, { "epoch": 0.5984618578536456, "grad_norm": 1.1968237161636353, "learning_rate": 0.00019665523318374858, "loss": 4.6618, "step": 320700 }, { "epoch": 0.5985551634130241, "grad_norm": 1.3349511623382568, "learning_rate": 0.0001966541556751101, "loss": 4.5394, "step": 320750 }, { "epoch": 0.5986484689724025, "grad_norm": 1.3234177827835083, "learning_rate": 0.0001966530779958941, "loss": 4.5405, "step": 320800 }, { "epoch": 0.598741774531781, "grad_norm": 1.0499647855758667, "learning_rate": 0.0001966520001461024, "loss": 4.4868, "step": 320850 }, { "epoch": 0.5988350800911595, "grad_norm": 1.1239477396011353, "learning_rate": 0.00019665092212573694, "loss": 4.6189, "step": 320900 }, { "epoch": 0.5989283856505381, "grad_norm": 1.0901520252227783, "learning_rate": 0.0001966498439347996, "loss": 4.8085, "step": 320950 }, { "epoch": 0.5990216912099166, "grad_norm": 1.0736844539642334, "learning_rate": 0.0001966487655732923, "loss": 4.6435, "step": 321000 }, { "epoch": 0.5990216912099166, "eval_loss": 4.745899200439453, "eval_runtime": 229.1766, "eval_samples_per_second": 11.38, "eval_steps_per_second": 11.38, "eval_tts_loss": 7.537432325952486, "step": 321000 }, { "epoch": 0.599114996769295, "grad_norm": 0.7315561771392822, "learning_rate": 0.00019664768704121693, "loss": 4.6113, "step": 321050 }, { "epoch": 0.5992083023286735, "grad_norm": 1.0384379625320435, "learning_rate": 0.0001966466083385754, "loss": 4.7575, "step": 321100 }, { "epoch": 0.599301607888052, "grad_norm": 1.1658464670181274, "learning_rate": 0.00019664552946536962, "loss": 4.5812, "step": 321150 }, { "epoch": 0.5993949134474305, "grad_norm": 1.0108740329742432, "learning_rate": 0.00019664445042160146, "loss": 4.5712, "step": 321200 }, { "epoch": 0.5994882190068089, "grad_norm": 0.908758819103241, "learning_rate": 0.00019664337120727288, "loss": 4.6208, "step": 321250 }, { "epoch": 0.5995815245661875, "grad_norm": 0.9615160822868347, "learning_rate": 0.00019664229182238575, "loss": 4.5738, "step": 321300 }, { "epoch": 0.599674830125566, "grad_norm": 1.0263842344284058, "learning_rate": 0.00019664121226694196, "loss": 4.4814, "step": 321350 }, { "epoch": 0.5997681356849445, "grad_norm": 0.9682742953300476, "learning_rate": 0.00019664013254094343, "loss": 4.62, "step": 321400 }, { "epoch": 0.599861441244323, "grad_norm": 1.1773290634155273, "learning_rate": 0.0001966390526443921, "loss": 4.4712, "step": 321450 }, { "epoch": 0.5999547468037014, "grad_norm": 1.0511239767074585, "learning_rate": 0.00019663797257728983, "loss": 4.6363, "step": 321500 }, { "epoch": 0.6000480523630799, "grad_norm": 1.1144319772720337, "learning_rate": 0.00019663689233963855, "loss": 4.5899, "step": 321550 }, { "epoch": 0.6001413579224584, "grad_norm": 0.9538633227348328, "learning_rate": 0.00019663581193144015, "loss": 4.5972, "step": 321600 }, { "epoch": 0.600234663481837, "grad_norm": 1.1423015594482422, "learning_rate": 0.00019663473135269653, "loss": 4.5939, "step": 321650 }, { "epoch": 0.6003279690412154, "grad_norm": 1.0630602836608887, "learning_rate": 0.00019663365060340965, "loss": 4.6106, "step": 321700 }, { "epoch": 0.6004212746005939, "grad_norm": 1.1156061887741089, "learning_rate": 0.00019663256968358134, "loss": 4.4516, "step": 321750 }, { "epoch": 0.6005145801599724, "grad_norm": 1.0120784044265747, "learning_rate": 0.00019663148859321359, "loss": 4.8363, "step": 321800 }, { "epoch": 0.6006078857193509, "grad_norm": 0.9239031076431274, "learning_rate": 0.00019663040733230823, "loss": 4.3221, "step": 321850 }, { "epoch": 0.6007011912787293, "grad_norm": 0.9571710824966431, "learning_rate": 0.00019662932590086722, "loss": 4.6294, "step": 321900 }, { "epoch": 0.6007944968381078, "grad_norm": 0.7756795883178711, "learning_rate": 0.00019662824429889246, "loss": 4.4496, "step": 321950 }, { "epoch": 0.6008878023974864, "grad_norm": 1.1667654514312744, "learning_rate": 0.00019662716252638585, "loss": 4.6833, "step": 322000 }, { "epoch": 0.6009811079568649, "grad_norm": 1.0731756687164307, "learning_rate": 0.00019662608058334928, "loss": 4.5627, "step": 322050 }, { "epoch": 0.6010744135162434, "grad_norm": 1.1679133176803589, "learning_rate": 0.00019662499846978467, "loss": 4.5755, "step": 322100 }, { "epoch": 0.6011677190756218, "grad_norm": 0.9971390962600708, "learning_rate": 0.00019662391618569398, "loss": 4.6366, "step": 322150 }, { "epoch": 0.6012610246350003, "grad_norm": 1.1999056339263916, "learning_rate": 0.00019662283373107905, "loss": 4.4628, "step": 322200 }, { "epoch": 0.6013543301943788, "grad_norm": 0.9386271834373474, "learning_rate": 0.00019662175110594183, "loss": 4.5262, "step": 322250 }, { "epoch": 0.6014476357537573, "grad_norm": 1.5020827054977417, "learning_rate": 0.00019662066831028423, "loss": 4.5555, "step": 322300 }, { "epoch": 0.6015409413131357, "grad_norm": 1.0585623979568481, "learning_rate": 0.00019661958534410813, "loss": 4.4205, "step": 322350 }, { "epoch": 0.6016342468725143, "grad_norm": 0.9941714406013489, "learning_rate": 0.00019661850220741549, "loss": 4.6991, "step": 322400 }, { "epoch": 0.6017275524318928, "grad_norm": 1.1517648696899414, "learning_rate": 0.00019661741890020817, "loss": 4.6615, "step": 322450 }, { "epoch": 0.6018208579912713, "grad_norm": 1.0368093252182007, "learning_rate": 0.00019661633542248811, "loss": 4.5439, "step": 322500 }, { "epoch": 0.6019141635506498, "grad_norm": 0.943163275718689, "learning_rate": 0.00019661525177425723, "loss": 4.5757, "step": 322550 }, { "epoch": 0.6020074691100282, "grad_norm": 1.1650466918945312, "learning_rate": 0.00019661416795551743, "loss": 4.4431, "step": 322600 }, { "epoch": 0.6021007746694067, "grad_norm": 0.9699743986129761, "learning_rate": 0.00019661308396627062, "loss": 4.5648, "step": 322650 }, { "epoch": 0.6021940802287852, "grad_norm": 1.0106393098831177, "learning_rate": 0.0001966119998065187, "loss": 4.5206, "step": 322700 }, { "epoch": 0.6022873857881638, "grad_norm": 0.9035084247589111, "learning_rate": 0.0001966109154762636, "loss": 4.4606, "step": 322750 }, { "epoch": 0.6023806913475422, "grad_norm": 1.0321792364120483, "learning_rate": 0.00019660983097550728, "loss": 4.4261, "step": 322800 }, { "epoch": 0.6024739969069207, "grad_norm": 1.080522060394287, "learning_rate": 0.00019660874630425155, "loss": 4.595, "step": 322850 }, { "epoch": 0.6025673024662992, "grad_norm": 1.063269853591919, "learning_rate": 0.0001966076614624984, "loss": 4.5491, "step": 322900 }, { "epoch": 0.6026606080256777, "grad_norm": 0.8457520008087158, "learning_rate": 0.00019660657645024972, "loss": 4.5702, "step": 322950 }, { "epoch": 0.6027539135850561, "grad_norm": 0.9796984195709229, "learning_rate": 0.00019660549126750747, "loss": 4.4528, "step": 323000 }, { "epoch": 0.6028472191444346, "grad_norm": 0.9538129568099976, "learning_rate": 0.00019660440591427348, "loss": 4.3992, "step": 323050 }, { "epoch": 0.6029405247038132, "grad_norm": 0.8515247106552124, "learning_rate": 0.00019660332039054973, "loss": 4.5346, "step": 323100 }, { "epoch": 0.6030338302631917, "grad_norm": 1.1083250045776367, "learning_rate": 0.00019660223469633814, "loss": 4.6768, "step": 323150 }, { "epoch": 0.6031271358225702, "grad_norm": 0.8830024600028992, "learning_rate": 0.00019660114883164056, "loss": 4.4938, "step": 323200 }, { "epoch": 0.6032204413819486, "grad_norm": 1.167148232460022, "learning_rate": 0.00019660006279645899, "loss": 4.4916, "step": 323250 }, { "epoch": 0.6033137469413271, "grad_norm": 1.0859688520431519, "learning_rate": 0.00019659897659079527, "loss": 4.6358, "step": 323300 }, { "epoch": 0.6034070525007056, "grad_norm": 0.9899167418479919, "learning_rate": 0.00019659789021465139, "loss": 4.4966, "step": 323350 }, { "epoch": 0.6035003580600841, "grad_norm": 1.0375924110412598, "learning_rate": 0.0001965968036680292, "loss": 4.4735, "step": 323400 }, { "epoch": 0.6035936636194627, "grad_norm": 1.1162281036376953, "learning_rate": 0.00019659571695093065, "loss": 4.7, "step": 323450 }, { "epoch": 0.6036869691788411, "grad_norm": 0.8468580842018127, "learning_rate": 0.00019659463006335768, "loss": 4.444, "step": 323500 }, { "epoch": 0.6037802747382196, "grad_norm": 1.0031423568725586, "learning_rate": 0.00019659354300531216, "loss": 4.4444, "step": 323550 }, { "epoch": 0.6038735802975981, "grad_norm": 0.908850371837616, "learning_rate": 0.00019659245577679607, "loss": 4.5278, "step": 323600 }, { "epoch": 0.6039668858569766, "grad_norm": 0.9931618571281433, "learning_rate": 0.00019659136837781125, "loss": 4.6133, "step": 323650 }, { "epoch": 0.604060191416355, "grad_norm": 1.185490369796753, "learning_rate": 0.00019659028080835968, "loss": 4.7496, "step": 323700 }, { "epoch": 0.6041534969757335, "grad_norm": 1.058326244354248, "learning_rate": 0.00019658919306844323, "loss": 4.6327, "step": 323750 }, { "epoch": 0.6042468025351121, "grad_norm": 1.1122000217437744, "learning_rate": 0.0001965881051580639, "loss": 4.7566, "step": 323800 }, { "epoch": 0.6043401080944906, "grad_norm": 0.9507600665092468, "learning_rate": 0.00019658701707722353, "loss": 4.4221, "step": 323850 }, { "epoch": 0.604433413653869, "grad_norm": 1.1114460229873657, "learning_rate": 0.00019658592882592407, "loss": 4.7225, "step": 323900 }, { "epoch": 0.6045267192132475, "grad_norm": 0.9252007603645325, "learning_rate": 0.00019658484040416743, "loss": 4.6596, "step": 323950 }, { "epoch": 0.604620024772626, "grad_norm": 1.2695643901824951, "learning_rate": 0.00019658375181195556, "loss": 4.771, "step": 324000 }, { "epoch": 0.604620024772626, "eval_loss": 4.752098560333252, "eval_runtime": 228.5606, "eval_samples_per_second": 11.411, "eval_steps_per_second": 11.411, "eval_tts_loss": 7.5316662976555415, "step": 324000 }, { "epoch": 0.6047133303320045, "grad_norm": 1.2202486991882324, "learning_rate": 0.00019658266304929035, "loss": 4.5332, "step": 324050 }, { "epoch": 0.604806635891383, "grad_norm": 0.8959473967552185, "learning_rate": 0.00019658157411617374, "loss": 4.7924, "step": 324100 }, { "epoch": 0.6048999414507615, "grad_norm": 0.8497489094734192, "learning_rate": 0.00019658048501260765, "loss": 4.5222, "step": 324150 }, { "epoch": 0.60499324701014, "grad_norm": 1.0305777788162231, "learning_rate": 0.000196579395738594, "loss": 4.5465, "step": 324200 }, { "epoch": 0.6050865525695185, "grad_norm": 1.0758594274520874, "learning_rate": 0.00019657830629413468, "loss": 4.6718, "step": 324250 }, { "epoch": 0.605179858128897, "grad_norm": 1.2500865459442139, "learning_rate": 0.0001965772166792317, "loss": 4.6455, "step": 324300 }, { "epoch": 0.6052731636882754, "grad_norm": 1.0861148834228516, "learning_rate": 0.0001965761268938869, "loss": 4.6696, "step": 324350 }, { "epoch": 0.6053664692476539, "grad_norm": 1.0526787042617798, "learning_rate": 0.00019657503693810223, "loss": 4.682, "step": 324400 }, { "epoch": 0.6054597748070324, "grad_norm": 0.9431971311569214, "learning_rate": 0.0001965739468118796, "loss": 4.4591, "step": 324450 }, { "epoch": 0.605553080366411, "grad_norm": 1.114660382270813, "learning_rate": 0.00019657285651522095, "loss": 4.5528, "step": 324500 }, { "epoch": 0.6056463859257895, "grad_norm": 1.0945026874542236, "learning_rate": 0.00019657176604812818, "loss": 4.6842, "step": 324550 }, { "epoch": 0.6057396914851679, "grad_norm": 0.9303328990936279, "learning_rate": 0.00019657067541060328, "loss": 4.7461, "step": 324600 }, { "epoch": 0.6058329970445464, "grad_norm": 0.9786701202392578, "learning_rate": 0.00019656958460264812, "loss": 4.7805, "step": 324650 }, { "epoch": 0.6059263026039249, "grad_norm": 0.8835315704345703, "learning_rate": 0.0001965684936242646, "loss": 4.7447, "step": 324700 }, { "epoch": 0.6060196081633034, "grad_norm": 1.2973759174346924, "learning_rate": 0.00019656740247545474, "loss": 4.8283, "step": 324750 }, { "epoch": 0.6061129137226818, "grad_norm": 1.4678105115890503, "learning_rate": 0.00019656631115622035, "loss": 4.6572, "step": 324800 }, { "epoch": 0.6062062192820604, "grad_norm": 0.8270549178123474, "learning_rate": 0.00019656521966656342, "loss": 4.336, "step": 324850 }, { "epoch": 0.6062995248414389, "grad_norm": 1.040366530418396, "learning_rate": 0.0001965641280064859, "loss": 4.4321, "step": 324900 }, { "epoch": 0.6063928304008174, "grad_norm": 1.086034893989563, "learning_rate": 0.00019656303617598966, "loss": 4.5608, "step": 324950 }, { "epoch": 0.6064861359601958, "grad_norm": 1.1229069232940674, "learning_rate": 0.00019656194417507665, "loss": 4.6664, "step": 325000 }, { "epoch": 0.6065794415195743, "grad_norm": 0.9855028390884399, "learning_rate": 0.00019656085200374883, "loss": 4.7031, "step": 325050 }, { "epoch": 0.6066727470789528, "grad_norm": 0.996859073638916, "learning_rate": 0.00019655975966200806, "loss": 4.4463, "step": 325100 }, { "epoch": 0.6067660526383313, "grad_norm": 1.2329950332641602, "learning_rate": 0.00019655866714985633, "loss": 4.4839, "step": 325150 }, { "epoch": 0.6068593581977099, "grad_norm": 0.7367371916770935, "learning_rate": 0.00019655757446729552, "loss": 4.488, "step": 325200 }, { "epoch": 0.6069526637570883, "grad_norm": 1.0623233318328857, "learning_rate": 0.00019655648161432757, "loss": 4.4816, "step": 325250 }, { "epoch": 0.6070459693164668, "grad_norm": 1.2126883268356323, "learning_rate": 0.00019655538859095445, "loss": 4.7547, "step": 325300 }, { "epoch": 0.6071392748758453, "grad_norm": 0.8081970810890198, "learning_rate": 0.00019655429539717803, "loss": 4.5895, "step": 325350 }, { "epoch": 0.6072325804352238, "grad_norm": 0.9101871848106384, "learning_rate": 0.00019655320203300033, "loss": 4.5339, "step": 325400 }, { "epoch": 0.6073258859946022, "grad_norm": 0.9889973402023315, "learning_rate": 0.00019655210849842315, "loss": 4.5158, "step": 325450 }, { "epoch": 0.6074191915539807, "grad_norm": 1.2756493091583252, "learning_rate": 0.00019655101479344853, "loss": 4.6334, "step": 325500 }, { "epoch": 0.6075124971133593, "grad_norm": 1.1303919553756714, "learning_rate": 0.00019654992091807833, "loss": 4.4023, "step": 325550 }, { "epoch": 0.6076058026727378, "grad_norm": 0.8666788935661316, "learning_rate": 0.00019654882687231452, "loss": 4.5264, "step": 325600 }, { "epoch": 0.6076991082321163, "grad_norm": 0.773320734500885, "learning_rate": 0.00019654773265615904, "loss": 4.8125, "step": 325650 }, { "epoch": 0.6077924137914947, "grad_norm": 0.9333234429359436, "learning_rate": 0.00019654663826961378, "loss": 4.6164, "step": 325700 }, { "epoch": 0.6078857193508732, "grad_norm": 1.298832654953003, "learning_rate": 0.00019654554371268068, "loss": 4.6582, "step": 325750 }, { "epoch": 0.6079790249102517, "grad_norm": 0.8918960690498352, "learning_rate": 0.0001965444489853617, "loss": 4.7427, "step": 325800 }, { "epoch": 0.6080723304696302, "grad_norm": 1.246414303779602, "learning_rate": 0.00019654335408765872, "loss": 4.6355, "step": 325850 }, { "epoch": 0.6081656360290087, "grad_norm": 1.3998520374298096, "learning_rate": 0.00019654225901957374, "loss": 4.6412, "step": 325900 }, { "epoch": 0.6082589415883872, "grad_norm": 1.056199312210083, "learning_rate": 0.00019654116378110866, "loss": 4.4848, "step": 325950 }, { "epoch": 0.6083522471477657, "grad_norm": 1.1381001472473145, "learning_rate": 0.0001965400683722654, "loss": 4.6771, "step": 326000 }, { "epoch": 0.6084455527071442, "grad_norm": 0.9033389687538147, "learning_rate": 0.0001965389727930459, "loss": 4.3929, "step": 326050 }, { "epoch": 0.6085388582665227, "grad_norm": 1.209061622619629, "learning_rate": 0.00019653787704345212, "loss": 4.5338, "step": 326100 }, { "epoch": 0.6086321638259011, "grad_norm": 0.8938238620758057, "learning_rate": 0.00019653678112348597, "loss": 4.5691, "step": 326150 }, { "epoch": 0.6087254693852796, "grad_norm": 1.1634777784347534, "learning_rate": 0.0001965356850331494, "loss": 4.5017, "step": 326200 }, { "epoch": 0.6088187749446582, "grad_norm": 0.9691982865333557, "learning_rate": 0.00019653458877244432, "loss": 4.6735, "step": 326250 }, { "epoch": 0.6089120805040367, "grad_norm": 1.0083643198013306, "learning_rate": 0.00019653349234137265, "loss": 4.5339, "step": 326300 }, { "epoch": 0.6090053860634151, "grad_norm": 1.2493237257003784, "learning_rate": 0.0001965323957399364, "loss": 4.6472, "step": 326350 }, { "epoch": 0.6090986916227936, "grad_norm": 0.7249242067337036, "learning_rate": 0.00019653129896813742, "loss": 4.7061, "step": 326400 }, { "epoch": 0.6091919971821721, "grad_norm": 1.0830210447311401, "learning_rate": 0.0001965302020259777, "loss": 4.4675, "step": 326450 }, { "epoch": 0.6092853027415506, "grad_norm": 0.8388657569885254, "learning_rate": 0.00019652910491345913, "loss": 4.4492, "step": 326500 }, { "epoch": 0.609378608300929, "grad_norm": 0.9378383159637451, "learning_rate": 0.0001965280076305837, "loss": 4.6353, "step": 326550 }, { "epoch": 0.6094719138603076, "grad_norm": 0.8669517636299133, "learning_rate": 0.0001965269101773533, "loss": 4.5646, "step": 326600 }, { "epoch": 0.6095652194196861, "grad_norm": 0.952555239200592, "learning_rate": 0.00019652581255376992, "loss": 4.6136, "step": 326650 }, { "epoch": 0.6096585249790646, "grad_norm": 1.1579066514968872, "learning_rate": 0.00019652471475983544, "loss": 4.4012, "step": 326700 }, { "epoch": 0.6097518305384431, "grad_norm": 1.158402442932129, "learning_rate": 0.00019652361679555185, "loss": 4.4434, "step": 326750 }, { "epoch": 0.6098451360978215, "grad_norm": 1.1343615055084229, "learning_rate": 0.00019652251866092104, "loss": 4.5192, "step": 326800 }, { "epoch": 0.6099384416572, "grad_norm": 1.0151411294937134, "learning_rate": 0.00019652142035594497, "loss": 4.5552, "step": 326850 }, { "epoch": 0.6100317472165785, "grad_norm": 1.1065433025360107, "learning_rate": 0.00019652032188062556, "loss": 4.5854, "step": 326900 }, { "epoch": 0.6101250527759571, "grad_norm": 1.094696283340454, "learning_rate": 0.0001965192232349648, "loss": 4.7376, "step": 326950 }, { "epoch": 0.6102183583353356, "grad_norm": 1.0514317750930786, "learning_rate": 0.00019651812441896457, "loss": 4.5541, "step": 327000 }, { "epoch": 0.6102183583353356, "eval_loss": 4.746123790740967, "eval_runtime": 229.2631, "eval_samples_per_second": 11.376, "eval_steps_per_second": 11.376, "eval_tts_loss": 7.612345226578975, "step": 327000 }, { "epoch": 0.610311663894714, "grad_norm": 1.2415906190872192, "learning_rate": 0.00019651702543262685, "loss": 4.4649, "step": 327050 }, { "epoch": 0.6104049694540925, "grad_norm": 1.0724464654922485, "learning_rate": 0.00019651592627595353, "loss": 4.8505, "step": 327100 }, { "epoch": 0.610498275013471, "grad_norm": 0.9932399392127991, "learning_rate": 0.00019651482694894664, "loss": 4.4503, "step": 327150 }, { "epoch": 0.6105915805728495, "grad_norm": 0.9129202961921692, "learning_rate": 0.000196513727451608, "loss": 4.5826, "step": 327200 }, { "epoch": 0.6106848861322279, "grad_norm": 1.6551772356033325, "learning_rate": 0.00019651262778393967, "loss": 4.7348, "step": 327250 }, { "epoch": 0.6107781916916065, "grad_norm": 0.7570912837982178, "learning_rate": 0.00019651152794594346, "loss": 4.504, "step": 327300 }, { "epoch": 0.610871497250985, "grad_norm": 0.8969416618347168, "learning_rate": 0.00019651042793762146, "loss": 4.6502, "step": 327350 }, { "epoch": 0.6109648028103635, "grad_norm": 1.240637183189392, "learning_rate": 0.0001965093277589755, "loss": 4.5202, "step": 327400 }, { "epoch": 0.611058108369742, "grad_norm": 1.1303819417953491, "learning_rate": 0.0001965082274100076, "loss": 4.5728, "step": 327450 }, { "epoch": 0.6111514139291204, "grad_norm": 1.054078221321106, "learning_rate": 0.0001965071268907196, "loss": 4.5804, "step": 327500 }, { "epoch": 0.6112447194884989, "grad_norm": 0.9103609919548035, "learning_rate": 0.00019650602620111354, "loss": 4.4957, "step": 327550 }, { "epoch": 0.6113380250478774, "grad_norm": 0.5890389084815979, "learning_rate": 0.0001965049253411913, "loss": 4.5091, "step": 327600 }, { "epoch": 0.6114313306072559, "grad_norm": 1.1171053647994995, "learning_rate": 0.00019650382431095483, "loss": 4.5723, "step": 327650 }, { "epoch": 0.6115246361666344, "grad_norm": 1.0724706649780273, "learning_rate": 0.00019650272311040614, "loss": 4.7041, "step": 327700 }, { "epoch": 0.6116179417260129, "grad_norm": 0.8792927861213684, "learning_rate": 0.0001965016217395471, "loss": 4.7052, "step": 327750 }, { "epoch": 0.6117112472853914, "grad_norm": 0.9894546270370483, "learning_rate": 0.00019650052019837968, "loss": 4.7049, "step": 327800 }, { "epoch": 0.6118045528447699, "grad_norm": 1.0347297191619873, "learning_rate": 0.00019649941848690582, "loss": 4.4355, "step": 327850 }, { "epoch": 0.6118978584041483, "grad_norm": 1.1192511320114136, "learning_rate": 0.00019649831660512748, "loss": 4.5305, "step": 327900 }, { "epoch": 0.6119911639635268, "grad_norm": 1.2721318006515503, "learning_rate": 0.00019649721455304657, "loss": 4.3744, "step": 327950 }, { "epoch": 0.6120844695229053, "grad_norm": 1.1381678581237793, "learning_rate": 0.00019649611233066506, "loss": 4.7289, "step": 328000 }, { "epoch": 0.6121777750822839, "grad_norm": 1.2317228317260742, "learning_rate": 0.00019649500993798488, "loss": 4.5512, "step": 328050 }, { "epoch": 0.6122710806416624, "grad_norm": 0.9952607154846191, "learning_rate": 0.000196493907375008, "loss": 4.5351, "step": 328100 }, { "epoch": 0.6123643862010408, "grad_norm": 0.5989564061164856, "learning_rate": 0.0001964928046417364, "loss": 4.3921, "step": 328150 }, { "epoch": 0.6124576917604193, "grad_norm": 0.7932347059249878, "learning_rate": 0.0001964917017381719, "loss": 4.4653, "step": 328200 }, { "epoch": 0.6125509973197978, "grad_norm": 1.1928980350494385, "learning_rate": 0.00019649059866431656, "loss": 4.5038, "step": 328250 }, { "epoch": 0.6126443028791763, "grad_norm": 1.196129560470581, "learning_rate": 0.00019648949542017228, "loss": 4.6942, "step": 328300 }, { "epoch": 0.6127376084385547, "grad_norm": 1.213106632232666, "learning_rate": 0.00019648839200574103, "loss": 4.5907, "step": 328350 }, { "epoch": 0.6128309139979333, "grad_norm": 1.1690261363983154, "learning_rate": 0.00019648728842102474, "loss": 4.6448, "step": 328400 }, { "epoch": 0.6129242195573118, "grad_norm": 1.2848875522613525, "learning_rate": 0.00019648618466602534, "loss": 4.5353, "step": 328450 }, { "epoch": 0.6130175251166903, "grad_norm": 1.101271152496338, "learning_rate": 0.00019648508074074486, "loss": 4.8021, "step": 328500 }, { "epoch": 0.6131108306760688, "grad_norm": 1.112776756286621, "learning_rate": 0.00019648397664518514, "loss": 4.4725, "step": 328550 }, { "epoch": 0.6132041362354472, "grad_norm": 1.0443400144577026, "learning_rate": 0.0001964828723793482, "loss": 4.6165, "step": 328600 }, { "epoch": 0.6132974417948257, "grad_norm": 1.1482632160186768, "learning_rate": 0.00019648176794323597, "loss": 4.4479, "step": 328650 }, { "epoch": 0.6133907473542042, "grad_norm": 1.1257174015045166, "learning_rate": 0.0001964806633368504, "loss": 4.8037, "step": 328700 }, { "epoch": 0.6134840529135828, "grad_norm": 0.9680014252662659, "learning_rate": 0.0001964795585601934, "loss": 4.6661, "step": 328750 }, { "epoch": 0.6135773584729612, "grad_norm": 1.0088255405426025, "learning_rate": 0.00019647845361326697, "loss": 4.3305, "step": 328800 }, { "epoch": 0.6136706640323397, "grad_norm": 1.115094542503357, "learning_rate": 0.00019647734849607306, "loss": 4.6662, "step": 328850 }, { "epoch": 0.6137639695917182, "grad_norm": 1.0315027236938477, "learning_rate": 0.0001964762432086136, "loss": 4.544, "step": 328900 }, { "epoch": 0.6138572751510967, "grad_norm": 0.9989319443702698, "learning_rate": 0.00019647513775089055, "loss": 4.6008, "step": 328950 }, { "epoch": 0.6139505807104751, "grad_norm": 0.8979176878929138, "learning_rate": 0.00019647403212290585, "loss": 4.5104, "step": 329000 }, { "epoch": 0.6140438862698536, "grad_norm": 0.7627320289611816, "learning_rate": 0.00019647292632466145, "loss": 4.666, "step": 329050 }, { "epoch": 0.6141371918292322, "grad_norm": 1.0842419862747192, "learning_rate": 0.00019647182035615934, "loss": 4.5201, "step": 329100 }, { "epoch": 0.6142304973886107, "grad_norm": 1.143351674079895, "learning_rate": 0.00019647071421740142, "loss": 4.5823, "step": 329150 }, { "epoch": 0.6143238029479892, "grad_norm": 1.2588927745819092, "learning_rate": 0.00019646960790838965, "loss": 4.5743, "step": 329200 }, { "epoch": 0.6144171085073676, "grad_norm": 1.2234686613082886, "learning_rate": 0.000196468501429126, "loss": 4.6187, "step": 329250 }, { "epoch": 0.6145104140667461, "grad_norm": 0.8479459881782532, "learning_rate": 0.00019646739477961246, "loss": 4.6488, "step": 329300 }, { "epoch": 0.6146037196261246, "grad_norm": 0.9231317639350891, "learning_rate": 0.00019646628795985088, "loss": 4.6329, "step": 329350 }, { "epoch": 0.6146970251855031, "grad_norm": 0.9470799565315247, "learning_rate": 0.00019646518096984333, "loss": 4.4673, "step": 329400 }, { "epoch": 0.6147903307448817, "grad_norm": 0.9557851552963257, "learning_rate": 0.0001964640738095917, "loss": 4.5174, "step": 329450 }, { "epoch": 0.6148836363042601, "grad_norm": 1.2796331644058228, "learning_rate": 0.00019646296647909793, "loss": 4.6437, "step": 329500 }, { "epoch": 0.6149769418636386, "grad_norm": 1.0128943920135498, "learning_rate": 0.000196461858978364, "loss": 4.546, "step": 329550 }, { "epoch": 0.6150702474230171, "grad_norm": 0.8020675182342529, "learning_rate": 0.00019646075130739187, "loss": 4.6278, "step": 329600 }, { "epoch": 0.6151635529823956, "grad_norm": 1.032940149307251, "learning_rate": 0.00019645964346618348, "loss": 4.4691, "step": 329650 }, { "epoch": 0.615256858541774, "grad_norm": 0.9859622120857239, "learning_rate": 0.00019645853545474082, "loss": 4.3977, "step": 329700 }, { "epoch": 0.6153501641011525, "grad_norm": 1.0631998777389526, "learning_rate": 0.00019645742727306578, "loss": 4.5291, "step": 329750 }, { "epoch": 0.6154434696605311, "grad_norm": 0.8335729837417603, "learning_rate": 0.00019645631892116039, "loss": 4.4934, "step": 329800 }, { "epoch": 0.6155367752199096, "grad_norm": 1.0670087337493896, "learning_rate": 0.00019645521039902655, "loss": 4.77, "step": 329850 }, { "epoch": 0.615630080779288, "grad_norm": 0.9757839441299438, "learning_rate": 0.00019645410170666621, "loss": 4.6015, "step": 329900 }, { "epoch": 0.6157233863386665, "grad_norm": 0.6826449036598206, "learning_rate": 0.00019645299284408137, "loss": 4.5966, "step": 329950 }, { "epoch": 0.615816691898045, "grad_norm": 1.2171297073364258, "learning_rate": 0.000196451883811274, "loss": 4.5911, "step": 330000 }, { "epoch": 0.615816691898045, "eval_loss": 4.750162601470947, "eval_runtime": 230.0209, "eval_samples_per_second": 11.338, "eval_steps_per_second": 11.338, "eval_tts_loss": 7.530302074275042, "step": 330000 }, { "epoch": 0.6159099974574235, "grad_norm": 1.0892624855041504, "learning_rate": 0.000196450774608246, "loss": 4.3776, "step": 330050 }, { "epoch": 0.616003303016802, "grad_norm": 1.1487836837768555, "learning_rate": 0.00019644966523499937, "loss": 4.5873, "step": 330100 }, { "epoch": 0.6160966085761805, "grad_norm": 1.07069993019104, "learning_rate": 0.00019644855569153605, "loss": 4.5787, "step": 330150 }, { "epoch": 0.616189914135559, "grad_norm": 1.1551017761230469, "learning_rate": 0.00019644744597785795, "loss": 4.6652, "step": 330200 }, { "epoch": 0.6162832196949375, "grad_norm": 1.1474007368087769, "learning_rate": 0.00019644633609396713, "loss": 4.407, "step": 330250 }, { "epoch": 0.616376525254316, "grad_norm": 0.6405335068702698, "learning_rate": 0.0001964452260398655, "loss": 4.5194, "step": 330300 }, { "epoch": 0.6164698308136944, "grad_norm": 0.8228738903999329, "learning_rate": 0.000196444115815555, "loss": 4.7268, "step": 330350 }, { "epoch": 0.6165631363730729, "grad_norm": 0.929740846157074, "learning_rate": 0.0001964430054210376, "loss": 4.3615, "step": 330400 }, { "epoch": 0.6166564419324514, "grad_norm": 1.1017268896102905, "learning_rate": 0.00019644189485631527, "loss": 4.5888, "step": 330450 }, { "epoch": 0.61674974749183, "grad_norm": 0.9419089555740356, "learning_rate": 0.00019644078412138995, "loss": 4.5174, "step": 330500 }, { "epoch": 0.6168430530512085, "grad_norm": 1.1504666805267334, "learning_rate": 0.00019643967321626363, "loss": 4.4526, "step": 330550 }, { "epoch": 0.6169363586105869, "grad_norm": 1.1194062232971191, "learning_rate": 0.00019643856214093824, "loss": 4.5075, "step": 330600 }, { "epoch": 0.6170296641699654, "grad_norm": 1.017431378364563, "learning_rate": 0.00019643745089541577, "loss": 4.5547, "step": 330650 }, { "epoch": 0.6171229697293439, "grad_norm": 0.9872207045555115, "learning_rate": 0.0001964363394796982, "loss": 4.5328, "step": 330700 }, { "epoch": 0.6172162752887224, "grad_norm": 0.8828074336051941, "learning_rate": 0.0001964352278937874, "loss": 4.6521, "step": 330750 }, { "epoch": 0.6173095808481008, "grad_norm": 0.8775408864021301, "learning_rate": 0.00019643411613768543, "loss": 4.7866, "step": 330800 }, { "epoch": 0.6174028864074794, "grad_norm": 0.7545701265335083, "learning_rate": 0.0001964330042113942, "loss": 4.5833, "step": 330850 }, { "epoch": 0.6174961919668579, "grad_norm": 1.1457773447036743, "learning_rate": 0.00019643189211491572, "loss": 4.5884, "step": 330900 }, { "epoch": 0.6175894975262364, "grad_norm": 1.1608442068099976, "learning_rate": 0.00019643077984825185, "loss": 4.586, "step": 330950 }, { "epoch": 0.6176828030856149, "grad_norm": 1.0994585752487183, "learning_rate": 0.00019642966741140468, "loss": 4.3464, "step": 331000 }, { "epoch": 0.6177761086449933, "grad_norm": 1.1020636558532715, "learning_rate": 0.0001964285548043761, "loss": 4.5699, "step": 331050 }, { "epoch": 0.6178694142043718, "grad_norm": 1.0513229370117188, "learning_rate": 0.00019642744202716806, "loss": 4.4984, "step": 331100 }, { "epoch": 0.6179627197637503, "grad_norm": 1.385101079940796, "learning_rate": 0.00019642632907978258, "loss": 4.5985, "step": 331150 }, { "epoch": 0.6180560253231289, "grad_norm": 1.1317663192749023, "learning_rate": 0.0001964252159622216, "loss": 4.3054, "step": 331200 }, { "epoch": 0.6181493308825073, "grad_norm": 0.9027521014213562, "learning_rate": 0.00019642410267448705, "loss": 4.6536, "step": 331250 }, { "epoch": 0.6182426364418858, "grad_norm": 1.294648289680481, "learning_rate": 0.00019642298921658096, "loss": 4.7736, "step": 331300 }, { "epoch": 0.6183359420012643, "grad_norm": 1.1973425149917603, "learning_rate": 0.00019642187558850522, "loss": 4.5538, "step": 331350 }, { "epoch": 0.6184292475606428, "grad_norm": 1.1227333545684814, "learning_rate": 0.0001964207617902619, "loss": 4.6983, "step": 331400 }, { "epoch": 0.6185225531200212, "grad_norm": 0.9092786312103271, "learning_rate": 0.00019641964782185283, "loss": 4.4916, "step": 331450 }, { "epoch": 0.6186158586793997, "grad_norm": 1.4974592924118042, "learning_rate": 0.0001964185336832801, "loss": 4.7164, "step": 331500 }, { "epoch": 0.6187091642387783, "grad_norm": 1.073347568511963, "learning_rate": 0.0001964174193745456, "loss": 4.5721, "step": 331550 }, { "epoch": 0.6188024697981568, "grad_norm": 1.0728085041046143, "learning_rate": 0.0001964163048956513, "loss": 4.6056, "step": 331600 }, { "epoch": 0.6188957753575353, "grad_norm": 0.6657750010490417, "learning_rate": 0.0001964151902465992, "loss": 4.397, "step": 331650 }, { "epoch": 0.6189890809169137, "grad_norm": 0.7082281112670898, "learning_rate": 0.00019641407542739128, "loss": 4.4559, "step": 331700 }, { "epoch": 0.6190823864762922, "grad_norm": 1.0563206672668457, "learning_rate": 0.00019641296043802946, "loss": 4.6829, "step": 331750 }, { "epoch": 0.6191756920356707, "grad_norm": 1.0246212482452393, "learning_rate": 0.00019641184527851573, "loss": 4.5568, "step": 331800 }, { "epoch": 0.6192689975950492, "grad_norm": 1.1611067056655884, "learning_rate": 0.00019641072994885207, "loss": 4.6479, "step": 331850 }, { "epoch": 0.6193623031544278, "grad_norm": 1.0870575904846191, "learning_rate": 0.00019640961444904043, "loss": 4.7873, "step": 331900 }, { "epoch": 0.6194556087138062, "grad_norm": 1.0575101375579834, "learning_rate": 0.00019640849877908275, "loss": 4.5286, "step": 331950 }, { "epoch": 0.6195489142731847, "grad_norm": 1.4265297651290894, "learning_rate": 0.00019640738293898107, "loss": 4.6938, "step": 332000 }, { "epoch": 0.6196422198325632, "grad_norm": 0.8864887356758118, "learning_rate": 0.00019640626692873735, "loss": 4.4216, "step": 332050 }, { "epoch": 0.6197355253919417, "grad_norm": 0.8195017576217651, "learning_rate": 0.00019640515074835347, "loss": 4.6461, "step": 332100 }, { "epoch": 0.6198288309513201, "grad_norm": 1.165611743927002, "learning_rate": 0.0001964040343978315, "loss": 4.6621, "step": 332150 }, { "epoch": 0.6199221365106986, "grad_norm": 1.1882660388946533, "learning_rate": 0.00019640291787717336, "loss": 4.5354, "step": 332200 }, { "epoch": 0.6200154420700772, "grad_norm": 1.2898271083831787, "learning_rate": 0.00019640180118638104, "loss": 4.777, "step": 332250 }, { "epoch": 0.6201087476294557, "grad_norm": 1.1433839797973633, "learning_rate": 0.0001964006843254565, "loss": 4.6701, "step": 332300 }, { "epoch": 0.6202020531888341, "grad_norm": 0.9746190309524536, "learning_rate": 0.0001963995672944017, "loss": 4.6845, "step": 332350 }, { "epoch": 0.6202953587482126, "grad_norm": 1.1168901920318604, "learning_rate": 0.00019639845009321864, "loss": 4.5668, "step": 332400 }, { "epoch": 0.6203886643075911, "grad_norm": 0.7818266749382019, "learning_rate": 0.00019639733272190925, "loss": 4.6588, "step": 332450 }, { "epoch": 0.6204819698669696, "grad_norm": 0.812650203704834, "learning_rate": 0.00019639621518047556, "loss": 4.6722, "step": 332500 }, { "epoch": 0.620575275426348, "grad_norm": 0.8418398499488831, "learning_rate": 0.00019639509746891952, "loss": 4.5468, "step": 332550 }, { "epoch": 0.6206685809857266, "grad_norm": 1.0831886529922485, "learning_rate": 0.00019639397958724308, "loss": 4.6973, "step": 332600 }, { "epoch": 0.6207618865451051, "grad_norm": 1.0995817184448242, "learning_rate": 0.00019639286153544822, "loss": 4.6884, "step": 332650 }, { "epoch": 0.6208551921044836, "grad_norm": 1.1868793964385986, "learning_rate": 0.00019639174331353694, "loss": 4.5945, "step": 332700 }, { "epoch": 0.6209484976638621, "grad_norm": 1.0501141548156738, "learning_rate": 0.0001963906249215112, "loss": 4.6657, "step": 332750 }, { "epoch": 0.6210418032232405, "grad_norm": 1.1675609350204468, "learning_rate": 0.0001963895063593729, "loss": 4.5511, "step": 332800 }, { "epoch": 0.621135108782619, "grad_norm": 0.9354573488235474, "learning_rate": 0.00019638838762712414, "loss": 4.6609, "step": 332850 }, { "epoch": 0.6212284143419975, "grad_norm": 0.8872344493865967, "learning_rate": 0.00019638726872476685, "loss": 4.6845, "step": 332900 }, { "epoch": 0.621321719901376, "grad_norm": 1.1700870990753174, "learning_rate": 0.00019638614965230295, "loss": 4.5461, "step": 332950 }, { "epoch": 0.6214150254607546, "grad_norm": 1.4714617729187012, "learning_rate": 0.00019638503040973447, "loss": 4.5, "step": 333000 }, { "epoch": 0.6214150254607546, "eval_loss": 4.740175247192383, "eval_runtime": 230.3212, "eval_samples_per_second": 11.323, "eval_steps_per_second": 11.323, "eval_tts_loss": 7.545454389351403, "step": 333000 }, { "epoch": 0.621508331020133, "grad_norm": 0.8408044576644897, "learning_rate": 0.00019638391099706338, "loss": 4.4279, "step": 333050 }, { "epoch": 0.6216016365795115, "grad_norm": 1.0031147003173828, "learning_rate": 0.00019638279141429162, "loss": 4.5333, "step": 333100 }, { "epoch": 0.62169494213889, "grad_norm": 0.9783959984779358, "learning_rate": 0.00019638167166142125, "loss": 4.4112, "step": 333150 }, { "epoch": 0.6217882476982685, "grad_norm": 1.0063459873199463, "learning_rate": 0.00019638055173845415, "loss": 4.7576, "step": 333200 }, { "epoch": 0.6218815532576469, "grad_norm": 0.936040461063385, "learning_rate": 0.00019637943164539233, "loss": 4.6556, "step": 333250 }, { "epoch": 0.6219748588170254, "grad_norm": 1.168675422668457, "learning_rate": 0.00019637831138223777, "loss": 4.4459, "step": 333300 }, { "epoch": 0.622068164376404, "grad_norm": 1.1717939376831055, "learning_rate": 0.00019637719094899245, "loss": 4.781, "step": 333350 }, { "epoch": 0.6221614699357825, "grad_norm": 0.8017193078994751, "learning_rate": 0.00019637607034565834, "loss": 4.73, "step": 333400 }, { "epoch": 0.622254775495161, "grad_norm": 1.0900770425796509, "learning_rate": 0.00019637494957223746, "loss": 4.7813, "step": 333450 }, { "epoch": 0.6223480810545394, "grad_norm": 1.1926219463348389, "learning_rate": 0.00019637382862873172, "loss": 4.6637, "step": 333500 }, { "epoch": 0.6224413866139179, "grad_norm": 1.0957951545715332, "learning_rate": 0.00019637270751514315, "loss": 4.6358, "step": 333550 }, { "epoch": 0.6225346921732964, "grad_norm": 1.0752534866333008, "learning_rate": 0.00019637158623147366, "loss": 4.6348, "step": 333600 }, { "epoch": 0.6226279977326749, "grad_norm": 1.108163595199585, "learning_rate": 0.00019637046477772533, "loss": 4.761, "step": 333650 }, { "epoch": 0.6227213032920534, "grad_norm": 0.9737891554832458, "learning_rate": 0.00019636934315390007, "loss": 4.4272, "step": 333700 }, { "epoch": 0.6228146088514319, "grad_norm": 1.2830264568328857, "learning_rate": 0.00019636822135999987, "loss": 4.6189, "step": 333750 }, { "epoch": 0.6229079144108104, "grad_norm": 1.093766212463379, "learning_rate": 0.00019636709939602673, "loss": 4.6984, "step": 333800 }, { "epoch": 0.6230012199701889, "grad_norm": 1.0020575523376465, "learning_rate": 0.0001963659772619826, "loss": 4.6488, "step": 333850 }, { "epoch": 0.6230945255295673, "grad_norm": 0.9393331408500671, "learning_rate": 0.00019636485495786948, "loss": 4.49, "step": 333900 }, { "epoch": 0.6231878310889458, "grad_norm": 1.1477243900299072, "learning_rate": 0.00019636373248368936, "loss": 4.6383, "step": 333950 }, { "epoch": 0.6232811366483243, "grad_norm": 1.0864436626434326, "learning_rate": 0.00019636260983944422, "loss": 4.6965, "step": 334000 }, { "epoch": 0.6233744422077029, "grad_norm": 0.757732629776001, "learning_rate": 0.00019636148702513597, "loss": 4.7961, "step": 334050 }, { "epoch": 0.6234677477670814, "grad_norm": 1.0098869800567627, "learning_rate": 0.00019636036404076669, "loss": 4.6886, "step": 334100 }, { "epoch": 0.6235610533264598, "grad_norm": 1.1908249855041504, "learning_rate": 0.00019635924088633834, "loss": 4.4147, "step": 334150 }, { "epoch": 0.6236543588858383, "grad_norm": 0.871414065361023, "learning_rate": 0.00019635811756185288, "loss": 4.7605, "step": 334200 }, { "epoch": 0.6237476644452168, "grad_norm": 1.0281577110290527, "learning_rate": 0.00019635699406731226, "loss": 4.5573, "step": 334250 }, { "epoch": 0.6238409700045953, "grad_norm": 0.8493218421936035, "learning_rate": 0.00019635587040271851, "loss": 4.3602, "step": 334300 }, { "epoch": 0.6239342755639737, "grad_norm": 0.9594669938087463, "learning_rate": 0.00019635474656807365, "loss": 4.5932, "step": 334350 }, { "epoch": 0.6240275811233523, "grad_norm": 0.9037676453590393, "learning_rate": 0.00019635362256337956, "loss": 4.5376, "step": 334400 }, { "epoch": 0.6241208866827308, "grad_norm": 1.2145931720733643, "learning_rate": 0.0001963524983886383, "loss": 4.4981, "step": 334450 }, { "epoch": 0.6242141922421093, "grad_norm": 1.0076943635940552, "learning_rate": 0.00019635137404385185, "loss": 4.558, "step": 334500 }, { "epoch": 0.6243074978014878, "grad_norm": 1.2487367391586304, "learning_rate": 0.00019635024952902215, "loss": 4.6688, "step": 334550 }, { "epoch": 0.6244008033608662, "grad_norm": 1.0044771432876587, "learning_rate": 0.00019634912484415123, "loss": 4.6066, "step": 334600 }, { "epoch": 0.6244941089202447, "grad_norm": 1.2532018423080444, "learning_rate": 0.00019634799998924108, "loss": 4.5018, "step": 334650 }, { "epoch": 0.6245874144796232, "grad_norm": 0.9828773140907288, "learning_rate": 0.0001963468749642936, "loss": 4.5975, "step": 334700 }, { "epoch": 0.6246807200390018, "grad_norm": 1.2060518264770508, "learning_rate": 0.0001963457497693109, "loss": 4.6228, "step": 334750 }, { "epoch": 0.6247740255983802, "grad_norm": 0.9976147413253784, "learning_rate": 0.0001963446244042949, "loss": 4.5998, "step": 334800 }, { "epoch": 0.6248673311577587, "grad_norm": 1.1077107191085815, "learning_rate": 0.00019634349886924753, "loss": 4.5616, "step": 334850 }, { "epoch": 0.6249606367171372, "grad_norm": 0.9811950325965881, "learning_rate": 0.00019634237316417088, "loss": 4.5548, "step": 334900 }, { "epoch": 0.6250539422765157, "grad_norm": 1.1001842021942139, "learning_rate": 0.0001963412472890669, "loss": 4.4049, "step": 334950 }, { "epoch": 0.6251472478358941, "grad_norm": 1.3210327625274658, "learning_rate": 0.00019634012124393756, "loss": 4.7167, "step": 335000 }, { "epoch": 0.6252405533952726, "grad_norm": 1.1540125608444214, "learning_rate": 0.00019633899502878487, "loss": 4.4382, "step": 335050 }, { "epoch": 0.6253338589546512, "grad_norm": 1.3277180194854736, "learning_rate": 0.00019633786864361078, "loss": 4.608, "step": 335100 }, { "epoch": 0.6254271645140297, "grad_norm": 1.3375109434127808, "learning_rate": 0.00019633674208841733, "loss": 4.3782, "step": 335150 }, { "epoch": 0.6255204700734082, "grad_norm": 1.0912272930145264, "learning_rate": 0.00019633561536320647, "loss": 4.6823, "step": 335200 }, { "epoch": 0.6256137756327866, "grad_norm": 1.2482918500900269, "learning_rate": 0.0001963344884679802, "loss": 4.6102, "step": 335250 }, { "epoch": 0.6257070811921651, "grad_norm": 1.2576767206192017, "learning_rate": 0.0001963333614027405, "loss": 4.6146, "step": 335300 }, { "epoch": 0.6258003867515436, "grad_norm": 0.9687120318412781, "learning_rate": 0.00019633223416748937, "loss": 4.5318, "step": 335350 }, { "epoch": 0.6258936923109221, "grad_norm": 0.7464566230773926, "learning_rate": 0.0001963311067622288, "loss": 4.4003, "step": 335400 }, { "epoch": 0.6259869978703007, "grad_norm": 1.0068252086639404, "learning_rate": 0.00019632997918696079, "loss": 4.6142, "step": 335450 }, { "epoch": 0.6260803034296791, "grad_norm": 1.048789143562317, "learning_rate": 0.00019632885144168732, "loss": 4.5993, "step": 335500 }, { "epoch": 0.6261736089890576, "grad_norm": 1.1501978635787964, "learning_rate": 0.00019632772352641035, "loss": 4.4245, "step": 335550 }, { "epoch": 0.6262669145484361, "grad_norm": 1.0876208543777466, "learning_rate": 0.0001963265954411319, "loss": 4.6859, "step": 335600 }, { "epoch": 0.6263602201078146, "grad_norm": 1.1919074058532715, "learning_rate": 0.00019632546718585396, "loss": 4.6262, "step": 335650 }, { "epoch": 0.626453525667193, "grad_norm": 1.3580552339553833, "learning_rate": 0.0001963243387605785, "loss": 4.5983, "step": 335700 }, { "epoch": 0.6265468312265715, "grad_norm": 0.9644747376441956, "learning_rate": 0.00019632321016530759, "loss": 4.2747, "step": 335750 }, { "epoch": 0.6266401367859501, "grad_norm": 0.8680166602134705, "learning_rate": 0.00019632208140004314, "loss": 4.5933, "step": 335800 }, { "epoch": 0.6267334423453286, "grad_norm": 0.9055318236351013, "learning_rate": 0.00019632095246478715, "loss": 4.3666, "step": 335850 }, { "epoch": 0.626826747904707, "grad_norm": 1.0014026165008545, "learning_rate": 0.00019631982335954163, "loss": 4.5995, "step": 335900 }, { "epoch": 0.6269200534640855, "grad_norm": 1.083059549331665, "learning_rate": 0.00019631869408430857, "loss": 4.635, "step": 335950 }, { "epoch": 0.627013359023464, "grad_norm": 0.9607086777687073, "learning_rate": 0.00019631756463908996, "loss": 4.7649, "step": 336000 }, { "epoch": 0.627013359023464, "eval_loss": 4.74292516708374, "eval_runtime": 230.2863, "eval_samples_per_second": 11.325, "eval_steps_per_second": 11.325, "eval_tts_loss": 7.567721385224522, "step": 336000 }, { "epoch": 0.6271066645828425, "grad_norm": 0.7822778224945068, "learning_rate": 0.0001963164350238878, "loss": 4.5734, "step": 336050 }, { "epoch": 0.627199970142221, "grad_norm": 1.253950595855713, "learning_rate": 0.00019631530523870407, "loss": 4.5294, "step": 336100 }, { "epoch": 0.6272932757015995, "grad_norm": 0.8558539748191833, "learning_rate": 0.00019631417528354078, "loss": 4.5185, "step": 336150 }, { "epoch": 0.627386581260978, "grad_norm": 0.9739894270896912, "learning_rate": 0.00019631304515839993, "loss": 4.6097, "step": 336200 }, { "epoch": 0.6274798868203565, "grad_norm": 1.14170503616333, "learning_rate": 0.00019631191486328348, "loss": 4.6264, "step": 336250 }, { "epoch": 0.627573192379735, "grad_norm": 1.0193452835083008, "learning_rate": 0.00019631078439819345, "loss": 4.699, "step": 336300 }, { "epoch": 0.6276664979391134, "grad_norm": 1.1254428625106812, "learning_rate": 0.00019630965376313185, "loss": 4.6496, "step": 336350 }, { "epoch": 0.6277598034984919, "grad_norm": 0.9944987893104553, "learning_rate": 0.00019630852295810067, "loss": 4.5546, "step": 336400 }, { "epoch": 0.6278531090578704, "grad_norm": 1.0571918487548828, "learning_rate": 0.00019630739198310185, "loss": 4.4923, "step": 336450 }, { "epoch": 0.627946414617249, "grad_norm": 1.2111427783966064, "learning_rate": 0.00019630626083813745, "loss": 4.5467, "step": 336500 }, { "epoch": 0.6280397201766275, "grad_norm": 0.6809943318367004, "learning_rate": 0.00019630512952320946, "loss": 4.57, "step": 336550 }, { "epoch": 0.6281330257360059, "grad_norm": 1.091772198677063, "learning_rate": 0.00019630399803831984, "loss": 4.6803, "step": 336600 }, { "epoch": 0.6282263312953844, "grad_norm": 1.1936126947402954, "learning_rate": 0.00019630286638347064, "loss": 4.4784, "step": 336650 }, { "epoch": 0.6283196368547629, "grad_norm": 1.1190840005874634, "learning_rate": 0.00019630173455866378, "loss": 4.7071, "step": 336700 }, { "epoch": 0.6284129424141414, "grad_norm": 1.2292462587356567, "learning_rate": 0.00019630060256390134, "loss": 4.6232, "step": 336750 }, { "epoch": 0.6285062479735198, "grad_norm": 1.1477453708648682, "learning_rate": 0.00019629947039918525, "loss": 4.7842, "step": 336800 }, { "epoch": 0.6285995535328984, "grad_norm": 1.0196152925491333, "learning_rate": 0.00019629833806451757, "loss": 4.6162, "step": 336850 }, { "epoch": 0.6286928590922769, "grad_norm": 1.0956239700317383, "learning_rate": 0.00019629720555990025, "loss": 4.4714, "step": 336900 }, { "epoch": 0.6287861646516554, "grad_norm": 0.8698761463165283, "learning_rate": 0.00019629607288533532, "loss": 4.5083, "step": 336950 }, { "epoch": 0.6288794702110339, "grad_norm": 1.1955455541610718, "learning_rate": 0.00019629494004082476, "loss": 4.585, "step": 337000 }, { "epoch": 0.6289727757704123, "grad_norm": 1.2584973573684692, "learning_rate": 0.00019629380702637053, "loss": 4.405, "step": 337050 }, { "epoch": 0.6290660813297908, "grad_norm": 0.8850629925727844, "learning_rate": 0.00019629267384197474, "loss": 4.5174, "step": 337100 }, { "epoch": 0.6291593868891693, "grad_norm": 0.861888587474823, "learning_rate": 0.00019629154048763929, "loss": 4.4397, "step": 337150 }, { "epoch": 0.6292526924485479, "grad_norm": 0.9065359830856323, "learning_rate": 0.00019629040696336622, "loss": 4.785, "step": 337200 }, { "epoch": 0.6293459980079263, "grad_norm": 0.9942500591278076, "learning_rate": 0.0001962892732691575, "loss": 4.5742, "step": 337250 }, { "epoch": 0.6294393035673048, "grad_norm": 1.2220995426177979, "learning_rate": 0.0001962881394050152, "loss": 4.6268, "step": 337300 }, { "epoch": 0.6295326091266833, "grad_norm": 1.0713739395141602, "learning_rate": 0.00019628700537094128, "loss": 4.3832, "step": 337350 }, { "epoch": 0.6296259146860618, "grad_norm": 0.8994389772415161, "learning_rate": 0.0001962858711669377, "loss": 4.7237, "step": 337400 }, { "epoch": 0.6297192202454402, "grad_norm": 1.032889723777771, "learning_rate": 0.0001962847367930065, "loss": 4.6565, "step": 337450 }, { "epoch": 0.6298125258048187, "grad_norm": 0.8554074764251709, "learning_rate": 0.0001962836022491497, "loss": 4.6186, "step": 337500 }, { "epoch": 0.6299058313641973, "grad_norm": 0.938651978969574, "learning_rate": 0.0001962824675353693, "loss": 4.5568, "step": 337550 }, { "epoch": 0.6299991369235758, "grad_norm": 0.7361330986022949, "learning_rate": 0.00019628133265166725, "loss": 4.5691, "step": 337600 }, { "epoch": 0.6300924424829543, "grad_norm": 0.9456680417060852, "learning_rate": 0.0001962801975980456, "loss": 4.6024, "step": 337650 }, { "epoch": 0.6301857480423327, "grad_norm": 1.1112492084503174, "learning_rate": 0.00019627906237450638, "loss": 4.5557, "step": 337700 }, { "epoch": 0.6302790536017112, "grad_norm": 1.0722355842590332, "learning_rate": 0.0001962779269810515, "loss": 4.5067, "step": 337750 }, { "epoch": 0.6303723591610897, "grad_norm": 1.1839230060577393, "learning_rate": 0.00019627679141768305, "loss": 4.7353, "step": 337800 }, { "epoch": 0.6304656647204682, "grad_norm": 1.101914644241333, "learning_rate": 0.00019627565568440297, "loss": 4.7507, "step": 337850 }, { "epoch": 0.6305589702798468, "grad_norm": 0.9712167382240295, "learning_rate": 0.00019627451978121335, "loss": 4.4696, "step": 337900 }, { "epoch": 0.6306522758392252, "grad_norm": 1.0436922311782837, "learning_rate": 0.0001962733837081161, "loss": 4.4775, "step": 337950 }, { "epoch": 0.6307455813986037, "grad_norm": 1.427233338356018, "learning_rate": 0.0001962722474651133, "loss": 4.6929, "step": 338000 }, { "epoch": 0.6308388869579822, "grad_norm": 0.9068809747695923, "learning_rate": 0.0001962711110522069, "loss": 4.6332, "step": 338050 }, { "epoch": 0.6309321925173607, "grad_norm": 0.6984463930130005, "learning_rate": 0.00019626997446939895, "loss": 4.6269, "step": 338100 }, { "epoch": 0.6310254980767391, "grad_norm": 1.048166275024414, "learning_rate": 0.0001962688377166914, "loss": 4.4084, "step": 338150 }, { "epoch": 0.6311188036361176, "grad_norm": 0.901068389415741, "learning_rate": 0.0001962677007940863, "loss": 4.6114, "step": 338200 }, { "epoch": 0.6312121091954962, "grad_norm": 1.1569806337356567, "learning_rate": 0.00019626656370158566, "loss": 4.7474, "step": 338250 }, { "epoch": 0.6313054147548747, "grad_norm": 1.0751103162765503, "learning_rate": 0.0001962654264391915, "loss": 4.5038, "step": 338300 }, { "epoch": 0.6313987203142531, "grad_norm": 0.8912768959999084, "learning_rate": 0.00019626428900690573, "loss": 4.4918, "step": 338350 }, { "epoch": 0.6314920258736316, "grad_norm": 1.0260379314422607, "learning_rate": 0.00019626315140473048, "loss": 4.6335, "step": 338400 }, { "epoch": 0.6315853314330101, "grad_norm": 1.0572807788848877, "learning_rate": 0.0001962620136326677, "loss": 4.6774, "step": 338450 }, { "epoch": 0.6316786369923886, "grad_norm": 0.8457328677177429, "learning_rate": 0.00019626087569071942, "loss": 4.5486, "step": 338500 }, { "epoch": 0.631771942551767, "grad_norm": 0.9951227903366089, "learning_rate": 0.0001962597375788876, "loss": 4.4332, "step": 338550 }, { "epoch": 0.6318652481111455, "grad_norm": 0.9773802757263184, "learning_rate": 0.0001962585992971743, "loss": 4.7251, "step": 338600 }, { "epoch": 0.6319585536705241, "grad_norm": 1.052430272102356, "learning_rate": 0.0001962574608455815, "loss": 4.4294, "step": 338650 }, { "epoch": 0.6320518592299026, "grad_norm": 1.225507140159607, "learning_rate": 0.0001962563222241112, "loss": 4.7384, "step": 338700 }, { "epoch": 0.6321451647892811, "grad_norm": 1.1235377788543701, "learning_rate": 0.00019625518343276545, "loss": 4.3392, "step": 338750 }, { "epoch": 0.6322384703486595, "grad_norm": 1.0696687698364258, "learning_rate": 0.00019625404447154625, "loss": 4.6364, "step": 338800 }, { "epoch": 0.632331775908038, "grad_norm": 0.9930275678634644, "learning_rate": 0.0001962529053404556, "loss": 4.6083, "step": 338850 }, { "epoch": 0.6324250814674165, "grad_norm": 1.2023861408233643, "learning_rate": 0.0001962517660394955, "loss": 4.6714, "step": 338900 }, { "epoch": 0.632518387026795, "grad_norm": 0.8565691709518433, "learning_rate": 0.00019625062656866793, "loss": 4.7474, "step": 338950 }, { "epoch": 0.6326116925861736, "grad_norm": 1.2467950582504272, "learning_rate": 0.00019624948692797498, "loss": 4.4604, "step": 339000 }, { "epoch": 0.6326116925861736, "eval_loss": 4.749098300933838, "eval_runtime": 229.0294, "eval_samples_per_second": 11.387, "eval_steps_per_second": 11.387, "eval_tts_loss": 7.533419189001919, "step": 339000 }, { "epoch": 0.632704998145552, "grad_norm": 0.9784042835235596, "learning_rate": 0.00019624834711741863, "loss": 4.6181, "step": 339050 }, { "epoch": 0.6327983037049305, "grad_norm": 1.1489437818527222, "learning_rate": 0.00019624720713700084, "loss": 4.5812, "step": 339100 }, { "epoch": 0.632891609264309, "grad_norm": 1.233275055885315, "learning_rate": 0.0001962460669867237, "loss": 4.7015, "step": 339150 }, { "epoch": 0.6329849148236875, "grad_norm": 0.9534322023391724, "learning_rate": 0.00019624492666658915, "loss": 4.589, "step": 339200 }, { "epoch": 0.6330782203830659, "grad_norm": 1.1035577058792114, "learning_rate": 0.0001962437861765993, "loss": 4.6331, "step": 339250 }, { "epoch": 0.6331715259424444, "grad_norm": 1.225083589553833, "learning_rate": 0.00019624264551675605, "loss": 4.6044, "step": 339300 }, { "epoch": 0.633264831501823, "grad_norm": 1.2086414098739624, "learning_rate": 0.00019624150468706148, "loss": 4.3887, "step": 339350 }, { "epoch": 0.6333581370612015, "grad_norm": 1.15715491771698, "learning_rate": 0.00019624036368751758, "loss": 4.7368, "step": 339400 }, { "epoch": 0.63345144262058, "grad_norm": 1.059969425201416, "learning_rate": 0.0001962392225181264, "loss": 4.6418, "step": 339450 }, { "epoch": 0.6335447481799584, "grad_norm": 1.1142557859420776, "learning_rate": 0.00019623808117888992, "loss": 4.6561, "step": 339500 }, { "epoch": 0.6336380537393369, "grad_norm": 1.1521546840667725, "learning_rate": 0.00019623693966981014, "loss": 4.6243, "step": 339550 }, { "epoch": 0.6337313592987154, "grad_norm": 1.0312284231185913, "learning_rate": 0.0001962357979908891, "loss": 4.698, "step": 339600 }, { "epoch": 0.6338246648580939, "grad_norm": 1.1738924980163574, "learning_rate": 0.00019623465614212882, "loss": 4.3374, "step": 339650 }, { "epoch": 0.6339179704174724, "grad_norm": 1.2590388059616089, "learning_rate": 0.0001962335141235313, "loss": 4.6054, "step": 339700 }, { "epoch": 0.6340112759768509, "grad_norm": 1.1100465059280396, "learning_rate": 0.00019623237193509857, "loss": 4.822, "step": 339750 }, { "epoch": 0.6341045815362294, "grad_norm": 1.1239413022994995, "learning_rate": 0.00019623122957683261, "loss": 4.5495, "step": 339800 }, { "epoch": 0.6341978870956079, "grad_norm": 1.021667242050171, "learning_rate": 0.00019623008704873547, "loss": 4.6736, "step": 339850 }, { "epoch": 0.6342911926549863, "grad_norm": 0.9145529270172119, "learning_rate": 0.00019622894435080917, "loss": 4.4413, "step": 339900 }, { "epoch": 0.6343844982143648, "grad_norm": 1.1026692390441895, "learning_rate": 0.00019622780148305572, "loss": 4.7521, "step": 339950 }, { "epoch": 0.6344778037737433, "grad_norm": 1.013266921043396, "learning_rate": 0.00019622665844547713, "loss": 4.6753, "step": 340000 }, { "epoch": 0.6345711093331219, "grad_norm": 0.95827317237854, "learning_rate": 0.00019622551523807542, "loss": 4.4974, "step": 340050 }, { "epoch": 0.6346644148925004, "grad_norm": 1.094130277633667, "learning_rate": 0.00019622437186085257, "loss": 4.7321, "step": 340100 }, { "epoch": 0.6347577204518788, "grad_norm": 0.9615687727928162, "learning_rate": 0.00019622322831381068, "loss": 4.4676, "step": 340150 }, { "epoch": 0.6348510260112573, "grad_norm": 0.982330858707428, "learning_rate": 0.0001962220845969517, "loss": 4.4142, "step": 340200 }, { "epoch": 0.6349443315706358, "grad_norm": 1.1653809547424316, "learning_rate": 0.0001962209407102777, "loss": 4.5996, "step": 340250 }, { "epoch": 0.6350376371300143, "grad_norm": 1.0871142148971558, "learning_rate": 0.00019621979665379065, "loss": 4.6704, "step": 340300 }, { "epoch": 0.6351309426893927, "grad_norm": 0.9514403939247131, "learning_rate": 0.00019621865242749258, "loss": 4.4426, "step": 340350 }, { "epoch": 0.6352242482487713, "grad_norm": 1.1947722434997559, "learning_rate": 0.0001962175080313855, "loss": 4.6943, "step": 340400 }, { "epoch": 0.6353175538081498, "grad_norm": 0.9871910810470581, "learning_rate": 0.00019621636346547149, "loss": 4.5358, "step": 340450 }, { "epoch": 0.6354108593675283, "grad_norm": 1.0049558877944946, "learning_rate": 0.00019621521872975254, "loss": 4.4537, "step": 340500 }, { "epoch": 0.6355041649269068, "grad_norm": 0.8265572190284729, "learning_rate": 0.00019621407382423064, "loss": 4.5113, "step": 340550 }, { "epoch": 0.6355974704862852, "grad_norm": 1.2362648248672485, "learning_rate": 0.0001962129287489078, "loss": 4.5154, "step": 340600 }, { "epoch": 0.6356907760456637, "grad_norm": 1.2209618091583252, "learning_rate": 0.00019621178350378608, "loss": 4.5803, "step": 340650 }, { "epoch": 0.6357840816050422, "grad_norm": 1.2041095495224, "learning_rate": 0.00019621063808886752, "loss": 4.8815, "step": 340700 }, { "epoch": 0.6358773871644208, "grad_norm": 1.169942855834961, "learning_rate": 0.00019620949250415408, "loss": 4.6944, "step": 340750 }, { "epoch": 0.6359706927237992, "grad_norm": 0.9905920624732971, "learning_rate": 0.00019620834674964784, "loss": 4.7523, "step": 340800 }, { "epoch": 0.6360639982831777, "grad_norm": 1.0956676006317139, "learning_rate": 0.0001962072008253508, "loss": 4.5541, "step": 340850 }, { "epoch": 0.6361573038425562, "grad_norm": 1.159255027770996, "learning_rate": 0.00019620605473126497, "loss": 4.5287, "step": 340900 }, { "epoch": 0.6362506094019347, "grad_norm": 1.0262867212295532, "learning_rate": 0.00019620490846739236, "loss": 4.6522, "step": 340950 }, { "epoch": 0.6363439149613132, "grad_norm": 0.8731298446655273, "learning_rate": 0.00019620376203373501, "loss": 4.4661, "step": 341000 }, { "epoch": 0.6364372205206916, "grad_norm": 1.0459544658660889, "learning_rate": 0.000196202615430295, "loss": 4.5979, "step": 341050 }, { "epoch": 0.6365305260800702, "grad_norm": 0.9693548679351807, "learning_rate": 0.00019620146865707426, "loss": 4.5553, "step": 341100 }, { "epoch": 0.6366238316394487, "grad_norm": 1.110594630241394, "learning_rate": 0.00019620032171407487, "loss": 4.402, "step": 341150 }, { "epoch": 0.6367171371988272, "grad_norm": 0.8167380094528198, "learning_rate": 0.00019619917460129882, "loss": 4.5829, "step": 341200 }, { "epoch": 0.6368104427582056, "grad_norm": 1.1120448112487793, "learning_rate": 0.00019619802731874817, "loss": 4.4741, "step": 341250 }, { "epoch": 0.6369037483175841, "grad_norm": 1.108361840248108, "learning_rate": 0.00019619687986642493, "loss": 4.4353, "step": 341300 }, { "epoch": 0.6369970538769626, "grad_norm": 0.9473559260368347, "learning_rate": 0.00019619573224433112, "loss": 4.6361, "step": 341350 }, { "epoch": 0.6370903594363411, "grad_norm": 1.1996275186538696, "learning_rate": 0.00019619458445246877, "loss": 4.6839, "step": 341400 }, { "epoch": 0.6371836649957197, "grad_norm": 0.9558074474334717, "learning_rate": 0.0001961934364908399, "loss": 4.5789, "step": 341450 }, { "epoch": 0.6372769705550981, "grad_norm": 0.9312458038330078, "learning_rate": 0.00019619228835944654, "loss": 4.4597, "step": 341500 }, { "epoch": 0.6373702761144766, "grad_norm": 0.9136353731155396, "learning_rate": 0.00019619114005829072, "loss": 4.5402, "step": 341550 }, { "epoch": 0.6374635816738551, "grad_norm": 1.1716053485870361, "learning_rate": 0.00019618999158737448, "loss": 4.6056, "step": 341600 }, { "epoch": 0.6375568872332336, "grad_norm": 0.9839272499084473, "learning_rate": 0.0001961888429466998, "loss": 4.636, "step": 341650 }, { "epoch": 0.637650192792612, "grad_norm": 1.1562862396240234, "learning_rate": 0.00019618769413626874, "loss": 4.5179, "step": 341700 }, { "epoch": 0.6377434983519905, "grad_norm": 1.0224076509475708, "learning_rate": 0.00019618654515608333, "loss": 4.3989, "step": 341750 }, { "epoch": 0.6378368039113691, "grad_norm": 1.2475993633270264, "learning_rate": 0.0001961853960061456, "loss": 4.4837, "step": 341800 }, { "epoch": 0.6379301094707476, "grad_norm": 1.106919765472412, "learning_rate": 0.00019618424668645757, "loss": 4.5131, "step": 341850 }, { "epoch": 0.638023415030126, "grad_norm": 0.9902898073196411, "learning_rate": 0.00019618309719702127, "loss": 4.5768, "step": 341900 }, { "epoch": 0.6381167205895045, "grad_norm": 1.1790491342544556, "learning_rate": 0.00019618194753783872, "loss": 4.5059, "step": 341950 }, { "epoch": 0.638210026148883, "grad_norm": 1.2637087106704712, "learning_rate": 0.00019618079770891196, "loss": 4.4652, "step": 342000 }, { "epoch": 0.638210026148883, "eval_loss": 4.740108489990234, "eval_runtime": 231.3959, "eval_samples_per_second": 11.271, "eval_steps_per_second": 11.271, "eval_tts_loss": 7.586605188102559, "step": 342000 }, { "epoch": 0.6383033317082615, "grad_norm": 1.101304531097412, "learning_rate": 0.00019617964771024302, "loss": 4.8453, "step": 342050 }, { "epoch": 0.63839663726764, "grad_norm": 1.254040002822876, "learning_rate": 0.00019617849754183394, "loss": 4.5925, "step": 342100 }, { "epoch": 0.6384899428270185, "grad_norm": 0.9550309777259827, "learning_rate": 0.0001961773472036867, "loss": 4.6263, "step": 342150 }, { "epoch": 0.638583248386397, "grad_norm": 0.9817168712615967, "learning_rate": 0.0001961761966958034, "loss": 4.4791, "step": 342200 }, { "epoch": 0.6386765539457755, "grad_norm": 0.9158251285552979, "learning_rate": 0.00019617504601818596, "loss": 4.4927, "step": 342250 }, { "epoch": 0.638769859505154, "grad_norm": 1.2883267402648926, "learning_rate": 0.00019617389517083655, "loss": 4.6163, "step": 342300 }, { "epoch": 0.6388631650645324, "grad_norm": 0.8760630488395691, "learning_rate": 0.00019617274415375716, "loss": 4.6279, "step": 342350 }, { "epoch": 0.6389564706239109, "grad_norm": 1.0041249990463257, "learning_rate": 0.00019617159296694974, "loss": 4.4317, "step": 342400 }, { "epoch": 0.6390497761832894, "grad_norm": 1.0369528532028198, "learning_rate": 0.00019617044161041643, "loss": 4.6907, "step": 342450 }, { "epoch": 0.639143081742668, "grad_norm": 0.9990165829658508, "learning_rate": 0.0001961692900841592, "loss": 4.5248, "step": 342500 }, { "epoch": 0.6392363873020465, "grad_norm": 1.1646652221679688, "learning_rate": 0.00019616813838818008, "loss": 4.5413, "step": 342550 }, { "epoch": 0.6393296928614249, "grad_norm": 0.7702269554138184, "learning_rate": 0.00019616698652248112, "loss": 4.5439, "step": 342600 }, { "epoch": 0.6394229984208034, "grad_norm": 1.0360941886901855, "learning_rate": 0.00019616583448706435, "loss": 4.6135, "step": 342650 }, { "epoch": 0.6395163039801819, "grad_norm": 1.2609333992004395, "learning_rate": 0.0001961646822819318, "loss": 4.5714, "step": 342700 }, { "epoch": 0.6396096095395604, "grad_norm": 1.1983611583709717, "learning_rate": 0.0001961635299070855, "loss": 4.722, "step": 342750 }, { "epoch": 0.6397029150989388, "grad_norm": 1.2513660192489624, "learning_rate": 0.0001961623773625275, "loss": 4.5354, "step": 342800 }, { "epoch": 0.6397962206583174, "grad_norm": 0.9544387459754944, "learning_rate": 0.00019616122464825983, "loss": 4.597, "step": 342850 }, { "epoch": 0.6398895262176959, "grad_norm": 0.9964279532432556, "learning_rate": 0.0001961600717642845, "loss": 4.7352, "step": 342900 }, { "epoch": 0.6399828317770744, "grad_norm": 0.8601515889167786, "learning_rate": 0.0001961589187106036, "loss": 4.1705, "step": 342950 }, { "epoch": 0.6400761373364529, "grad_norm": 1.0321756601333618, "learning_rate": 0.0001961577654872191, "loss": 4.6684, "step": 343000 }, { "epoch": 0.6401694428958313, "grad_norm": 1.2796767950057983, "learning_rate": 0.00019615661209413307, "loss": 4.6704, "step": 343050 }, { "epoch": 0.6402627484552098, "grad_norm": 1.3482799530029297, "learning_rate": 0.00019615545853134754, "loss": 4.3253, "step": 343100 }, { "epoch": 0.6403560540145883, "grad_norm": 1.0156035423278809, "learning_rate": 0.00019615430479886454, "loss": 4.5649, "step": 343150 }, { "epoch": 0.6404493595739669, "grad_norm": 0.8863063454627991, "learning_rate": 0.00019615315089668611, "loss": 4.6938, "step": 343200 }, { "epoch": 0.6405426651333453, "grad_norm": 0.9971023797988892, "learning_rate": 0.0001961519968248143, "loss": 4.5935, "step": 343250 }, { "epoch": 0.6406359706927238, "grad_norm": 0.84346604347229, "learning_rate": 0.00019615084258325113, "loss": 4.6598, "step": 343300 }, { "epoch": 0.6407292762521023, "grad_norm": 0.9728134274482727, "learning_rate": 0.00019614968817199864, "loss": 4.6571, "step": 343350 }, { "epoch": 0.6408225818114808, "grad_norm": 2.26699161529541, "learning_rate": 0.00019614853359105884, "loss": 4.4916, "step": 343400 }, { "epoch": 0.6409158873708592, "grad_norm": 1.1074600219726562, "learning_rate": 0.00019614737884043383, "loss": 4.5896, "step": 343450 }, { "epoch": 0.6410091929302377, "grad_norm": 0.9999982118606567, "learning_rate": 0.0001961462239201256, "loss": 4.5574, "step": 343500 }, { "epoch": 0.6411024984896163, "grad_norm": 1.1165833473205566, "learning_rate": 0.00019614506883013622, "loss": 4.5864, "step": 343550 }, { "epoch": 0.6411958040489948, "grad_norm": 0.9048035740852356, "learning_rate": 0.0001961439135704677, "loss": 4.6054, "step": 343600 }, { "epoch": 0.6412891096083733, "grad_norm": 1.244683861732483, "learning_rate": 0.00019614275814112207, "loss": 4.695, "step": 343650 }, { "epoch": 0.6413824151677517, "grad_norm": 0.6611224412918091, "learning_rate": 0.0001961416025421014, "loss": 4.4653, "step": 343700 }, { "epoch": 0.6414757207271302, "grad_norm": 0.6442915797233582, "learning_rate": 0.00019614044677340771, "loss": 4.5202, "step": 343750 }, { "epoch": 0.6415690262865087, "grad_norm": 1.1567631959915161, "learning_rate": 0.0001961392908350431, "loss": 4.6907, "step": 343800 }, { "epoch": 0.6416623318458872, "grad_norm": 0.9971233606338501, "learning_rate": 0.00019613813472700952, "loss": 4.5955, "step": 343850 }, { "epoch": 0.6417556374052656, "grad_norm": 1.0582727193832397, "learning_rate": 0.00019613697844930906, "loss": 4.5997, "step": 343900 }, { "epoch": 0.6418489429646442, "grad_norm": 1.181840419769287, "learning_rate": 0.0001961358220019437, "loss": 4.7453, "step": 343950 }, { "epoch": 0.6419422485240227, "grad_norm": 0.9383758902549744, "learning_rate": 0.00019613466538491559, "loss": 4.5304, "step": 344000 }, { "epoch": 0.6420355540834012, "grad_norm": 1.198830485343933, "learning_rate": 0.00019613350859822669, "loss": 4.4426, "step": 344050 }, { "epoch": 0.6421288596427797, "grad_norm": 1.2523542642593384, "learning_rate": 0.00019613235164187907, "loss": 4.4969, "step": 344100 }, { "epoch": 0.6422221652021581, "grad_norm": 0.9709693789482117, "learning_rate": 0.00019613119451587473, "loss": 4.5729, "step": 344150 }, { "epoch": 0.6423154707615366, "grad_norm": 1.276157021522522, "learning_rate": 0.00019613003722021577, "loss": 4.6016, "step": 344200 }, { "epoch": 0.6424087763209151, "grad_norm": 1.1887445449829102, "learning_rate": 0.0001961288797549042, "loss": 4.8722, "step": 344250 }, { "epoch": 0.6425020818802937, "grad_norm": 1.132991909980774, "learning_rate": 0.0001961277221199421, "loss": 4.5887, "step": 344300 }, { "epoch": 0.6425953874396721, "grad_norm": 0.9599609971046448, "learning_rate": 0.00019612656431533147, "loss": 4.5974, "step": 344350 }, { "epoch": 0.6426886929990506, "grad_norm": 1.0121450424194336, "learning_rate": 0.00019612540634107438, "loss": 4.6816, "step": 344400 }, { "epoch": 0.6427819985584291, "grad_norm": 0.8555306196212769, "learning_rate": 0.00019612424819717282, "loss": 4.4254, "step": 344450 }, { "epoch": 0.6428753041178076, "grad_norm": 1.1181089878082275, "learning_rate": 0.00019612308988362892, "loss": 4.4815, "step": 344500 }, { "epoch": 0.642968609677186, "grad_norm": 1.0256164073944092, "learning_rate": 0.00019612193140044465, "loss": 4.5175, "step": 344550 }, { "epoch": 0.6430619152365645, "grad_norm": 0.7289327383041382, "learning_rate": 0.00019612077274762212, "loss": 4.3594, "step": 344600 }, { "epoch": 0.6431552207959431, "grad_norm": 1.0339686870574951, "learning_rate": 0.00019611961392516328, "loss": 4.4817, "step": 344650 }, { "epoch": 0.6432485263553216, "grad_norm": 0.9205882549285889, "learning_rate": 0.00019611845493307028, "loss": 4.6025, "step": 344700 }, { "epoch": 0.6433418319147001, "grad_norm": 1.0576496124267578, "learning_rate": 0.00019611729577134513, "loss": 4.3981, "step": 344750 }, { "epoch": 0.6434351374740785, "grad_norm": 0.7700652480125427, "learning_rate": 0.00019611613643998982, "loss": 4.5055, "step": 344800 }, { "epoch": 0.643528443033457, "grad_norm": 0.9929338097572327, "learning_rate": 0.00019611497693900647, "loss": 4.6463, "step": 344850 }, { "epoch": 0.6436217485928355, "grad_norm": 0.9840719103813171, "learning_rate": 0.0001961138172683971, "loss": 4.6944, "step": 344900 }, { "epoch": 0.643715054152214, "grad_norm": 1.1600414514541626, "learning_rate": 0.00019611265742816374, "loss": 4.5144, "step": 344950 }, { "epoch": 0.6438083597115926, "grad_norm": 1.071171760559082, "learning_rate": 0.00019611149741830845, "loss": 4.6351, "step": 345000 }, { "epoch": 0.6438083597115926, "eval_loss": 4.739168643951416, "eval_runtime": 233.3772, "eval_samples_per_second": 11.175, "eval_steps_per_second": 11.175, "eval_tts_loss": 7.525498071853986, "step": 345000 }, { "epoch": 0.643901665270971, "grad_norm": 1.0771901607513428, "learning_rate": 0.00019611033723883329, "loss": 4.7873, "step": 345050 }, { "epoch": 0.6439949708303495, "grad_norm": 0.9674336314201355, "learning_rate": 0.0001961091768897403, "loss": 4.3937, "step": 345100 }, { "epoch": 0.644088276389728, "grad_norm": 0.6305899620056152, "learning_rate": 0.00019610801637103152, "loss": 4.3422, "step": 345150 }, { "epoch": 0.6441815819491065, "grad_norm": 1.0925474166870117, "learning_rate": 0.000196106855682709, "loss": 4.536, "step": 345200 }, { "epoch": 0.6442748875084849, "grad_norm": 0.9010369777679443, "learning_rate": 0.00019610569482477478, "loss": 4.6296, "step": 345250 }, { "epoch": 0.6443681930678634, "grad_norm": 1.1146564483642578, "learning_rate": 0.00019610453379723092, "loss": 4.5091, "step": 345300 }, { "epoch": 0.644461498627242, "grad_norm": 0.842292845249176, "learning_rate": 0.00019610337260007947, "loss": 4.6014, "step": 345350 }, { "epoch": 0.6445548041866205, "grad_norm": 1.0273958444595337, "learning_rate": 0.00019610221123332247, "loss": 4.571, "step": 345400 }, { "epoch": 0.644648109745999, "grad_norm": 1.2018096446990967, "learning_rate": 0.00019610104969696203, "loss": 4.6055, "step": 345450 }, { "epoch": 0.6447414153053774, "grad_norm": 1.0792499780654907, "learning_rate": 0.0001960998879910001, "loss": 4.4501, "step": 345500 }, { "epoch": 0.6448347208647559, "grad_norm": 1.2928937673568726, "learning_rate": 0.0001960987261154388, "loss": 4.7066, "step": 345550 }, { "epoch": 0.6449280264241344, "grad_norm": 0.9761016964912415, "learning_rate": 0.0001960975640702801, "loss": 4.448, "step": 345600 }, { "epoch": 0.6450213319835129, "grad_norm": 1.1374536752700806, "learning_rate": 0.00019609640185552616, "loss": 4.7384, "step": 345650 }, { "epoch": 0.6451146375428914, "grad_norm": 1.1401821374893188, "learning_rate": 0.00019609523947117897, "loss": 4.6449, "step": 345700 }, { "epoch": 0.6452079431022699, "grad_norm": 1.1021229028701782, "learning_rate": 0.0001960940769172406, "loss": 4.6003, "step": 345750 }, { "epoch": 0.6453012486616484, "grad_norm": 1.2322521209716797, "learning_rate": 0.0001960929141937131, "loss": 4.6746, "step": 345800 }, { "epoch": 0.6453945542210269, "grad_norm": 1.06720769405365, "learning_rate": 0.00019609175130059847, "loss": 4.6942, "step": 345850 }, { "epoch": 0.6454878597804053, "grad_norm": 1.0128451585769653, "learning_rate": 0.00019609058823789882, "loss": 4.5013, "step": 345900 }, { "epoch": 0.6455811653397838, "grad_norm": 0.9844326376914978, "learning_rate": 0.00019608942500561622, "loss": 4.6259, "step": 345950 }, { "epoch": 0.6456744708991623, "grad_norm": 1.1689870357513428, "learning_rate": 0.00019608826160375265, "loss": 4.673, "step": 346000 }, { "epoch": 0.6457677764585409, "grad_norm": 1.0861157178878784, "learning_rate": 0.00019608709803231028, "loss": 4.5371, "step": 346050 }, { "epoch": 0.6458610820179194, "grad_norm": 0.9207714200019836, "learning_rate": 0.00019608593429129104, "loss": 4.6154, "step": 346100 }, { "epoch": 0.6459543875772978, "grad_norm": 1.0358107089996338, "learning_rate": 0.000196084770380697, "loss": 4.5926, "step": 346150 }, { "epoch": 0.6460476931366763, "grad_norm": 1.0891337394714355, "learning_rate": 0.0001960836063005303, "loss": 4.3183, "step": 346200 }, { "epoch": 0.6461409986960548, "grad_norm": 0.8938190937042236, "learning_rate": 0.00019608244205079293, "loss": 4.5577, "step": 346250 }, { "epoch": 0.6462343042554333, "grad_norm": 1.108068823814392, "learning_rate": 0.00019608127763148691, "loss": 4.3522, "step": 346300 }, { "epoch": 0.6463276098148117, "grad_norm": 0.9931000471115112, "learning_rate": 0.00019608011304261437, "loss": 4.716, "step": 346350 }, { "epoch": 0.6464209153741903, "grad_norm": 0.6367947459220886, "learning_rate": 0.00019607894828417737, "loss": 4.5168, "step": 346400 }, { "epoch": 0.6465142209335688, "grad_norm": 1.301096796989441, "learning_rate": 0.0001960777833561779, "loss": 4.689, "step": 346450 }, { "epoch": 0.6466075264929473, "grad_norm": 1.2469911575317383, "learning_rate": 0.00019607661825861805, "loss": 4.5518, "step": 346500 }, { "epoch": 0.6467008320523258, "grad_norm": 0.7290911078453064, "learning_rate": 0.00019607545299149987, "loss": 4.5329, "step": 346550 }, { "epoch": 0.6467941376117042, "grad_norm": 0.720203697681427, "learning_rate": 0.00019607428755482543, "loss": 4.6408, "step": 346600 }, { "epoch": 0.6468874431710827, "grad_norm": 1.0603928565979004, "learning_rate": 0.00019607312194859676, "loss": 4.6633, "step": 346650 }, { "epoch": 0.6469807487304612, "grad_norm": 1.269835352897644, "learning_rate": 0.00019607195617281597, "loss": 4.6081, "step": 346700 }, { "epoch": 0.6470740542898398, "grad_norm": 0.9368632435798645, "learning_rate": 0.00019607079022748508, "loss": 4.7006, "step": 346750 }, { "epoch": 0.6471673598492182, "grad_norm": 1.0011993646621704, "learning_rate": 0.0001960696241126061, "loss": 4.5515, "step": 346800 }, { "epoch": 0.6472606654085967, "grad_norm": 0.9085133671760559, "learning_rate": 0.0001960684578281812, "loss": 4.5498, "step": 346850 }, { "epoch": 0.6473539709679752, "grad_norm": 1.0289874076843262, "learning_rate": 0.00019606729137421232, "loss": 4.6087, "step": 346900 }, { "epoch": 0.6474472765273537, "grad_norm": 1.0617784261703491, "learning_rate": 0.0001960661247507016, "loss": 4.5837, "step": 346950 }, { "epoch": 0.6475405820867322, "grad_norm": 1.0014612674713135, "learning_rate": 0.00019606495795765108, "loss": 4.3512, "step": 347000 }, { "epoch": 0.6476338876461106, "grad_norm": 0.9258717894554138, "learning_rate": 0.0001960637909950628, "loss": 4.6539, "step": 347050 }, { "epoch": 0.6477271932054892, "grad_norm": 0.7328335642814636, "learning_rate": 0.00019606262386293884, "loss": 4.7303, "step": 347100 }, { "epoch": 0.6478204987648677, "grad_norm": 0.9870677590370178, "learning_rate": 0.00019606145656128122, "loss": 4.4358, "step": 347150 }, { "epoch": 0.6479138043242462, "grad_norm": 1.216899037361145, "learning_rate": 0.00019606028909009208, "loss": 4.6176, "step": 347200 }, { "epoch": 0.6480071098836246, "grad_norm": 1.1379703283309937, "learning_rate": 0.0001960591214493734, "loss": 4.3831, "step": 347250 }, { "epoch": 0.6481004154430031, "grad_norm": 1.1051362752914429, "learning_rate": 0.0001960579536391273, "loss": 4.6166, "step": 347300 }, { "epoch": 0.6481937210023816, "grad_norm": 1.2240265607833862, "learning_rate": 0.00019605678565935578, "loss": 4.5038, "step": 347350 }, { "epoch": 0.6482870265617601, "grad_norm": 0.9971303343772888, "learning_rate": 0.00019605561751006095, "loss": 4.7579, "step": 347400 }, { "epoch": 0.6483803321211387, "grad_norm": 0.888080894947052, "learning_rate": 0.00019605444919124486, "loss": 4.5582, "step": 347450 }, { "epoch": 0.6484736376805171, "grad_norm": 1.0499393939971924, "learning_rate": 0.00019605328070290958, "loss": 4.6159, "step": 347500 }, { "epoch": 0.6485669432398956, "grad_norm": 0.9919920563697815, "learning_rate": 0.00019605211204505712, "loss": 4.521, "step": 347550 }, { "epoch": 0.6486602487992741, "grad_norm": 1.0972288846969604, "learning_rate": 0.00019605094321768963, "loss": 4.4626, "step": 347600 }, { "epoch": 0.6487535543586526, "grad_norm": 1.0431736707687378, "learning_rate": 0.00019604977422080913, "loss": 4.4371, "step": 347650 }, { "epoch": 0.648846859918031, "grad_norm": 1.1692547798156738, "learning_rate": 0.00019604860505441764, "loss": 4.6365, "step": 347700 }, { "epoch": 0.6489401654774095, "grad_norm": 1.0646686553955078, "learning_rate": 0.00019604743571851726, "loss": 4.6142, "step": 347750 }, { "epoch": 0.6490334710367881, "grad_norm": 0.9616683721542358, "learning_rate": 0.00019604626621311008, "loss": 4.6137, "step": 347800 }, { "epoch": 0.6491267765961666, "grad_norm": 1.0787080526351929, "learning_rate": 0.00019604509653819814, "loss": 4.7056, "step": 347850 }, { "epoch": 0.649220082155545, "grad_norm": 1.1852656602859497, "learning_rate": 0.0001960439266937835, "loss": 4.5779, "step": 347900 }, { "epoch": 0.6493133877149235, "grad_norm": 0.8655521869659424, "learning_rate": 0.00019604275667986824, "loss": 4.6498, "step": 347950 }, { "epoch": 0.649406693274302, "grad_norm": 1.107316493988037, "learning_rate": 0.0001960415864964544, "loss": 4.6731, "step": 348000 }, { "epoch": 0.649406693274302, "eval_loss": 4.729848861694336, "eval_runtime": 229.1395, "eval_samples_per_second": 11.382, "eval_steps_per_second": 11.382, "eval_tts_loss": 7.577490967087036, "step": 348000 }, { "epoch": 0.6494999988336805, "grad_norm": 0.722476065158844, "learning_rate": 0.00019604041614354408, "loss": 4.731, "step": 348050 }, { "epoch": 0.649593304393059, "grad_norm": 1.176112413406372, "learning_rate": 0.0001960392456211393, "loss": 4.5264, "step": 348100 }, { "epoch": 0.6496866099524375, "grad_norm": 1.0830116271972656, "learning_rate": 0.00019603807492924214, "loss": 4.3844, "step": 348150 }, { "epoch": 0.649779915511816, "grad_norm": 0.9252073764801025, "learning_rate": 0.0001960369040678547, "loss": 4.736, "step": 348200 }, { "epoch": 0.6498732210711945, "grad_norm": 1.025813102722168, "learning_rate": 0.000196035733036979, "loss": 4.7346, "step": 348250 }, { "epoch": 0.649966526630573, "grad_norm": 0.72474604845047, "learning_rate": 0.00019603456183661715, "loss": 4.6141, "step": 348300 }, { "epoch": 0.6500598321899514, "grad_norm": 1.258725881576538, "learning_rate": 0.0001960333904667712, "loss": 4.4785, "step": 348350 }, { "epoch": 0.6501531377493299, "grad_norm": 0.8761577606201172, "learning_rate": 0.0001960322189274432, "loss": 4.6884, "step": 348400 }, { "epoch": 0.6502464433087084, "grad_norm": 1.0042715072631836, "learning_rate": 0.00019603104721863522, "loss": 4.6647, "step": 348450 }, { "epoch": 0.650339748868087, "grad_norm": 1.3143467903137207, "learning_rate": 0.00019602987534034932, "loss": 4.4965, "step": 348500 }, { "epoch": 0.6504330544274655, "grad_norm": 1.3221840858459473, "learning_rate": 0.00019602870329258763, "loss": 4.5676, "step": 348550 }, { "epoch": 0.6505263599868439, "grad_norm": 1.199800729751587, "learning_rate": 0.0001960275310753522, "loss": 4.4183, "step": 348600 }, { "epoch": 0.6506196655462224, "grad_norm": 0.9423922300338745, "learning_rate": 0.000196026358688645, "loss": 4.4449, "step": 348650 }, { "epoch": 0.6507129711056009, "grad_norm": 1.0496889352798462, "learning_rate": 0.00019602518613246822, "loss": 4.3336, "step": 348700 }, { "epoch": 0.6508062766649794, "grad_norm": 1.4044764041900635, "learning_rate": 0.0001960240134068239, "loss": 4.5279, "step": 348750 }, { "epoch": 0.6508995822243578, "grad_norm": 1.3792707920074463, "learning_rate": 0.00019602284051171406, "loss": 4.5458, "step": 348800 }, { "epoch": 0.6509928877837364, "grad_norm": 1.1532506942749023, "learning_rate": 0.00019602166744714082, "loss": 4.6879, "step": 348850 }, { "epoch": 0.6510861933431149, "grad_norm": 1.7418402433395386, "learning_rate": 0.0001960204942131062, "loss": 4.6084, "step": 348900 }, { "epoch": 0.6511794989024934, "grad_norm": 1.2199933528900146, "learning_rate": 0.00019601932080961235, "loss": 4.6386, "step": 348950 }, { "epoch": 0.6512728044618719, "grad_norm": 1.0035910606384277, "learning_rate": 0.00019601814723666126, "loss": 4.4855, "step": 349000 }, { "epoch": 0.6513661100212503, "grad_norm": 0.8540064096450806, "learning_rate": 0.00019601697349425502, "loss": 4.5604, "step": 349050 }, { "epoch": 0.6514594155806288, "grad_norm": 1.0423887968063354, "learning_rate": 0.00019601579958239575, "loss": 4.4431, "step": 349100 }, { "epoch": 0.6515527211400073, "grad_norm": 0.993720531463623, "learning_rate": 0.00019601462550108547, "loss": 4.5087, "step": 349150 }, { "epoch": 0.6516460266993858, "grad_norm": 1.0810538530349731, "learning_rate": 0.00019601345125032629, "loss": 4.4436, "step": 349200 }, { "epoch": 0.6517393322587643, "grad_norm": 0.9943142533302307, "learning_rate": 0.00019601227683012023, "loss": 4.4959, "step": 349250 }, { "epoch": 0.6518326378181428, "grad_norm": 1.1125222444534302, "learning_rate": 0.00019601110224046943, "loss": 4.6668, "step": 349300 }, { "epoch": 0.6519259433775213, "grad_norm": 0.7313642501831055, "learning_rate": 0.00019600992748137593, "loss": 4.6501, "step": 349350 }, { "epoch": 0.6520192489368998, "grad_norm": 0.9498952627182007, "learning_rate": 0.00019600875255284178, "loss": 4.6597, "step": 349400 }, { "epoch": 0.6521125544962783, "grad_norm": 1.002515435218811, "learning_rate": 0.00019600757745486907, "loss": 4.6331, "step": 349450 }, { "epoch": 0.6522058600556567, "grad_norm": 0.8719493746757507, "learning_rate": 0.0001960064021874599, "loss": 4.4631, "step": 349500 }, { "epoch": 0.6522991656150352, "grad_norm": 1.3565610647201538, "learning_rate": 0.00019600522675061632, "loss": 4.6105, "step": 349550 }, { "epoch": 0.6523924711744138, "grad_norm": 1.4063760042190552, "learning_rate": 0.0001960040511443404, "loss": 4.4435, "step": 349600 }, { "epoch": 0.6524857767337923, "grad_norm": 1.1056792736053467, "learning_rate": 0.0001960028753686342, "loss": 4.8335, "step": 349650 }, { "epoch": 0.6525790822931707, "grad_norm": 0.9742892980575562, "learning_rate": 0.00019600169942349984, "loss": 4.3601, "step": 349700 }, { "epoch": 0.6526723878525492, "grad_norm": 1.1894937753677368, "learning_rate": 0.0001960005233089394, "loss": 4.7548, "step": 349750 }, { "epoch": 0.6527656934119277, "grad_norm": 1.2229342460632324, "learning_rate": 0.0001959993470249549, "loss": 4.4735, "step": 349800 }, { "epoch": 0.6528589989713062, "grad_norm": 1.0048850774765015, "learning_rate": 0.00019599817057154844, "loss": 4.4905, "step": 349850 }, { "epoch": 0.6529523045306846, "grad_norm": 1.1758800745010376, "learning_rate": 0.00019599699394872211, "loss": 4.4454, "step": 349900 }, { "epoch": 0.6530456100900632, "grad_norm": 0.9137585163116455, "learning_rate": 0.00019599581715647798, "loss": 4.4121, "step": 349950 }, { "epoch": 0.6531389156494417, "grad_norm": 1.040393590927124, "learning_rate": 0.0001959946401948181, "loss": 4.6267, "step": 350000 }, { "epoch": 0.6532322212088202, "grad_norm": 1.0021544694900513, "learning_rate": 0.00019599346306374457, "loss": 4.6441, "step": 350050 }, { "epoch": 0.6533255267681987, "grad_norm": 0.8930035829544067, "learning_rate": 0.0001959922857632595, "loss": 4.634, "step": 350100 }, { "epoch": 0.6534188323275771, "grad_norm": 1.142181634902954, "learning_rate": 0.00019599110829336493, "loss": 4.661, "step": 350150 }, { "epoch": 0.6535121378869556, "grad_norm": 1.200129747390747, "learning_rate": 0.00019598993065406294, "loss": 4.563, "step": 350200 }, { "epoch": 0.6536054434463341, "grad_norm": 1.0493800640106201, "learning_rate": 0.00019598875284535558, "loss": 4.5014, "step": 350250 }, { "epoch": 0.6536987490057127, "grad_norm": 0.9547832012176514, "learning_rate": 0.00019598757486724499, "loss": 4.645, "step": 350300 }, { "epoch": 0.6537920545650912, "grad_norm": 1.0982476472854614, "learning_rate": 0.00019598639671973323, "loss": 4.4349, "step": 350350 }, { "epoch": 0.6538853601244696, "grad_norm": 1.0742329359054565, "learning_rate": 0.00019598521840282236, "loss": 4.5784, "step": 350400 }, { "epoch": 0.6539786656838481, "grad_norm": 1.203382134437561, "learning_rate": 0.00019598403991651447, "loss": 4.7065, "step": 350450 }, { "epoch": 0.6540719712432266, "grad_norm": 1.0984567403793335, "learning_rate": 0.00019598286126081162, "loss": 4.6502, "step": 350500 }, { "epoch": 0.6541652768026051, "grad_norm": 1.1957650184631348, "learning_rate": 0.00019598168243571592, "loss": 4.5351, "step": 350550 }, { "epoch": 0.6542585823619835, "grad_norm": 1.264735460281372, "learning_rate": 0.00019598050344122944, "loss": 4.6503, "step": 350600 }, { "epoch": 0.6543518879213621, "grad_norm": 0.8449005484580994, "learning_rate": 0.00019597932427735425, "loss": 4.377, "step": 350650 }, { "epoch": 0.6544451934807406, "grad_norm": 1.2481213808059692, "learning_rate": 0.00019597814494409245, "loss": 4.55, "step": 350700 }, { "epoch": 0.6545384990401191, "grad_norm": 1.057641625404358, "learning_rate": 0.00019597696544144616, "loss": 4.41, "step": 350750 }, { "epoch": 0.6546318045994975, "grad_norm": 0.9384470582008362, "learning_rate": 0.00019597578576941734, "loss": 4.4782, "step": 350800 }, { "epoch": 0.654725110158876, "grad_norm": 0.7581617832183838, "learning_rate": 0.00019597460592800817, "loss": 4.6777, "step": 350850 }, { "epoch": 0.6548184157182545, "grad_norm": 1.164628028869629, "learning_rate": 0.0001959734259172207, "loss": 4.9107, "step": 350900 }, { "epoch": 0.654911721277633, "grad_norm": 1.0642462968826294, "learning_rate": 0.00019597224573705705, "loss": 4.7088, "step": 350950 }, { "epoch": 0.6550050268370116, "grad_norm": 1.156158447265625, "learning_rate": 0.00019597106538751926, "loss": 4.33, "step": 351000 }, { "epoch": 0.6550050268370116, "eval_loss": 4.736559867858887, "eval_runtime": 228.9675, "eval_samples_per_second": 11.39, "eval_steps_per_second": 11.39, "eval_tts_loss": 7.536694011900951, "step": 351000 }, { "epoch": 0.65509833239639, "grad_norm": 0.8227680325508118, "learning_rate": 0.00019596988486860942, "loss": 4.671, "step": 351050 }, { "epoch": 0.6551916379557685, "grad_norm": 1.0057629346847534, "learning_rate": 0.0001959687041803296, "loss": 4.557, "step": 351100 }, { "epoch": 0.655284943515147, "grad_norm": 1.1353752613067627, "learning_rate": 0.00019596752332268194, "loss": 4.4945, "step": 351150 }, { "epoch": 0.6553782490745255, "grad_norm": 1.2325284481048584, "learning_rate": 0.00019596634229566848, "loss": 4.5777, "step": 351200 }, { "epoch": 0.6554715546339039, "grad_norm": 1.2234023809432983, "learning_rate": 0.00019596516109929128, "loss": 4.6234, "step": 351250 }, { "epoch": 0.6555648601932824, "grad_norm": 1.1709771156311035, "learning_rate": 0.00019596397973355249, "loss": 4.5459, "step": 351300 }, { "epoch": 0.655658165752661, "grad_norm": 1.100290298461914, "learning_rate": 0.00019596279819845416, "loss": 4.4377, "step": 351350 }, { "epoch": 0.6557514713120395, "grad_norm": 1.1520344018936157, "learning_rate": 0.00019596161649399835, "loss": 4.3091, "step": 351400 }, { "epoch": 0.655844776871418, "grad_norm": 1.0503275394439697, "learning_rate": 0.00019596043462018718, "loss": 4.6749, "step": 351450 }, { "epoch": 0.6559380824307964, "grad_norm": 0.8855152726173401, "learning_rate": 0.00019595925257702277, "loss": 4.6931, "step": 351500 }, { "epoch": 0.6560313879901749, "grad_norm": 0.8062465786933899, "learning_rate": 0.00019595807036450712, "loss": 4.5645, "step": 351550 }, { "epoch": 0.6561246935495534, "grad_norm": 1.2063981294631958, "learning_rate": 0.00019595688798264237, "loss": 4.6111, "step": 351600 }, { "epoch": 0.6562179991089319, "grad_norm": 1.2455005645751953, "learning_rate": 0.00019595570543143057, "loss": 4.5999, "step": 351650 }, { "epoch": 0.6563113046683104, "grad_norm": 1.0943653583526611, "learning_rate": 0.00019595452271087388, "loss": 4.5354, "step": 351700 }, { "epoch": 0.6564046102276889, "grad_norm": 1.1074237823486328, "learning_rate": 0.00019595333982097432, "loss": 4.4646, "step": 351750 }, { "epoch": 0.6564979157870674, "grad_norm": 1.2968482971191406, "learning_rate": 0.000195952156761734, "loss": 4.5801, "step": 351800 }, { "epoch": 0.6565912213464459, "grad_norm": 1.4533486366271973, "learning_rate": 0.00019595097353315502, "loss": 4.5672, "step": 351850 }, { "epoch": 0.6566845269058244, "grad_norm": 1.104537844657898, "learning_rate": 0.00019594979013523944, "loss": 4.5806, "step": 351900 }, { "epoch": 0.6567778324652028, "grad_norm": 0.7298156023025513, "learning_rate": 0.00019594860656798937, "loss": 4.5813, "step": 351950 }, { "epoch": 0.6568711380245813, "grad_norm": 1.117599606513977, "learning_rate": 0.00019594742283140688, "loss": 4.6431, "step": 352000 }, { "epoch": 0.6569644435839599, "grad_norm": 1.0720120668411255, "learning_rate": 0.00019594623892549406, "loss": 4.558, "step": 352050 }, { "epoch": 0.6570577491433384, "grad_norm": 0.9862524271011353, "learning_rate": 0.00019594505485025305, "loss": 4.5758, "step": 352100 }, { "epoch": 0.6571510547027168, "grad_norm": 1.1951582431793213, "learning_rate": 0.00019594387060568587, "loss": 4.6974, "step": 352150 }, { "epoch": 0.6572443602620953, "grad_norm": 1.1000356674194336, "learning_rate": 0.00019594268619179466, "loss": 4.5496, "step": 352200 }, { "epoch": 0.6573376658214738, "grad_norm": 1.2195320129394531, "learning_rate": 0.00019594150160858148, "loss": 4.4902, "step": 352250 }, { "epoch": 0.6574309713808523, "grad_norm": 1.0285000801086426, "learning_rate": 0.00019594031685604842, "loss": 4.5667, "step": 352300 }, { "epoch": 0.6575242769402307, "grad_norm": 1.1513861417770386, "learning_rate": 0.0001959391319341976, "loss": 4.6403, "step": 352350 }, { "epoch": 0.6576175824996093, "grad_norm": 1.0986008644104004, "learning_rate": 0.0001959379468430311, "loss": 4.6092, "step": 352400 }, { "epoch": 0.6577108880589878, "grad_norm": 1.1743237972259521, "learning_rate": 0.00019593676158255097, "loss": 4.6803, "step": 352450 }, { "epoch": 0.6578041936183663, "grad_norm": 0.9898801445960999, "learning_rate": 0.00019593557615275935, "loss": 4.3212, "step": 352500 }, { "epoch": 0.6578974991777448, "grad_norm": 1.0967254638671875, "learning_rate": 0.00019593439055365833, "loss": 4.5072, "step": 352550 }, { "epoch": 0.6579908047371232, "grad_norm": 1.0041605234146118, "learning_rate": 0.00019593320478524998, "loss": 4.4197, "step": 352600 }, { "epoch": 0.6580841102965017, "grad_norm": 1.07600736618042, "learning_rate": 0.0001959320188475364, "loss": 4.6252, "step": 352650 }, { "epoch": 0.6581774158558802, "grad_norm": 0.9869751930236816, "learning_rate": 0.00019593083274051975, "loss": 4.3965, "step": 352700 }, { "epoch": 0.6582707214152588, "grad_norm": 0.9461277723312378, "learning_rate": 0.00019592964646420198, "loss": 4.4762, "step": 352750 }, { "epoch": 0.6583640269746373, "grad_norm": 1.2213248014450073, "learning_rate": 0.0001959284600185853, "loss": 4.5955, "step": 352800 }, { "epoch": 0.6584573325340157, "grad_norm": 1.1953409910202026, "learning_rate": 0.00019592727340367173, "loss": 4.5722, "step": 352850 }, { "epoch": 0.6585506380933942, "grad_norm": 0.9113812446594238, "learning_rate": 0.00019592608661946344, "loss": 4.5533, "step": 352900 }, { "epoch": 0.6586439436527727, "grad_norm": 1.179384708404541, "learning_rate": 0.00019592489966596249, "loss": 4.6624, "step": 352950 }, { "epoch": 0.6587372492121512, "grad_norm": 1.031244158744812, "learning_rate": 0.00019592371254317095, "loss": 4.4156, "step": 353000 }, { "epoch": 0.6588305547715296, "grad_norm": 1.3826984167099, "learning_rate": 0.00019592252525109092, "loss": 4.5798, "step": 353050 }, { "epoch": 0.6589238603309082, "grad_norm": 0.905318558216095, "learning_rate": 0.00019592133778972452, "loss": 4.6347, "step": 353100 }, { "epoch": 0.6590171658902867, "grad_norm": 0.880496621131897, "learning_rate": 0.00019592015015907385, "loss": 4.4655, "step": 353150 }, { "epoch": 0.6591104714496652, "grad_norm": 0.9531444907188416, "learning_rate": 0.000195918962359141, "loss": 4.6297, "step": 353200 }, { "epoch": 0.6592037770090436, "grad_norm": 1.0310364961624146, "learning_rate": 0.00019591777438992804, "loss": 4.6778, "step": 353250 }, { "epoch": 0.6592970825684221, "grad_norm": 1.0869261026382446, "learning_rate": 0.00019591658625143712, "loss": 4.6961, "step": 353300 }, { "epoch": 0.6593903881278006, "grad_norm": 0.9768805503845215, "learning_rate": 0.00019591539794367027, "loss": 4.772, "step": 353350 }, { "epoch": 0.6594836936871791, "grad_norm": 1.2685409784317017, "learning_rate": 0.0001959142094666296, "loss": 4.4007, "step": 353400 }, { "epoch": 0.6595769992465577, "grad_norm": 0.9491780400276184, "learning_rate": 0.0001959130208203173, "loss": 4.4167, "step": 353450 }, { "epoch": 0.6596703048059361, "grad_norm": 0.995601236820221, "learning_rate": 0.00019591183200473532, "loss": 4.5628, "step": 353500 }, { "epoch": 0.6597636103653146, "grad_norm": 0.6878821849822998, "learning_rate": 0.00019591064301988586, "loss": 4.5096, "step": 353550 }, { "epoch": 0.6598569159246931, "grad_norm": 1.0458654165267944, "learning_rate": 0.000195909453865771, "loss": 4.4981, "step": 353600 }, { "epoch": 0.6599502214840716, "grad_norm": 1.0209881067276, "learning_rate": 0.00019590826454239283, "loss": 4.3819, "step": 353650 }, { "epoch": 0.66004352704345, "grad_norm": 0.883777916431427, "learning_rate": 0.00019590707504975343, "loss": 4.5237, "step": 353700 }, { "epoch": 0.6601368326028285, "grad_norm": 0.9451062679290771, "learning_rate": 0.00019590588538785495, "loss": 4.5695, "step": 353750 }, { "epoch": 0.6602301381622071, "grad_norm": 1.0650455951690674, "learning_rate": 0.00019590469555669942, "loss": 4.5875, "step": 353800 }, { "epoch": 0.6603234437215856, "grad_norm": 0.835949182510376, "learning_rate": 0.000195903505556289, "loss": 4.3907, "step": 353850 }, { "epoch": 0.660416749280964, "grad_norm": 1.0406574010849, "learning_rate": 0.00019590231538662575, "loss": 4.499, "step": 353900 }, { "epoch": 0.6605100548403425, "grad_norm": 0.9681634306907654, "learning_rate": 0.0001959011250477118, "loss": 4.4877, "step": 353950 }, { "epoch": 0.660603360399721, "grad_norm": 1.1119762659072876, "learning_rate": 0.0001958999345395492, "loss": 4.596, "step": 354000 }, { "epoch": 0.660603360399721, "eval_loss": 4.737687587738037, "eval_runtime": 229.9813, "eval_samples_per_second": 11.34, "eval_steps_per_second": 11.34, "eval_tts_loss": 7.5381670642552745, "step": 354000 }, { "epoch": 0.6606966659590995, "grad_norm": 1.0130051374435425, "learning_rate": 0.00019589874386214017, "loss": 4.4346, "step": 354050 }, { "epoch": 0.660789971518478, "grad_norm": 1.2727903127670288, "learning_rate": 0.0001958975530154867, "loss": 4.5517, "step": 354100 }, { "epoch": 0.6608832770778565, "grad_norm": 1.147301435470581, "learning_rate": 0.0001958963619995909, "loss": 4.5535, "step": 354150 }, { "epoch": 0.660976582637235, "grad_norm": 0.9156633615493774, "learning_rate": 0.0001958951708144549, "loss": 4.5594, "step": 354200 }, { "epoch": 0.6610698881966135, "grad_norm": 0.8328635692596436, "learning_rate": 0.0001958939794600808, "loss": 4.5648, "step": 354250 }, { "epoch": 0.661163193755992, "grad_norm": 0.9792226552963257, "learning_rate": 0.0001958927879364707, "loss": 4.6311, "step": 354300 }, { "epoch": 0.6612564993153704, "grad_norm": 1.0784177780151367, "learning_rate": 0.0001958915962436267, "loss": 4.4926, "step": 354350 }, { "epoch": 0.6613498048747489, "grad_norm": 0.9137364029884338, "learning_rate": 0.00019589040438155092, "loss": 4.6259, "step": 354400 }, { "epoch": 0.6614431104341274, "grad_norm": 1.2003227472305298, "learning_rate": 0.00019588921235024541, "loss": 4.6057, "step": 354450 }, { "epoch": 0.661536415993506, "grad_norm": 0.9942118525505066, "learning_rate": 0.00019588802014971238, "loss": 4.5277, "step": 354500 }, { "epoch": 0.6616297215528845, "grad_norm": 1.1411582231521606, "learning_rate": 0.0001958868277799538, "loss": 4.6279, "step": 354550 }, { "epoch": 0.6617230271122629, "grad_norm": 0.9816586971282959, "learning_rate": 0.00019588563524097188, "loss": 4.4651, "step": 354600 }, { "epoch": 0.6618163326716414, "grad_norm": 1.1638094186782837, "learning_rate": 0.00019588444253276866, "loss": 4.6132, "step": 354650 }, { "epoch": 0.6619096382310199, "grad_norm": 1.100682020187378, "learning_rate": 0.0001958832496553463, "loss": 4.4892, "step": 354700 }, { "epoch": 0.6620029437903984, "grad_norm": 1.091890573501587, "learning_rate": 0.00019588205660870685, "loss": 4.5033, "step": 354750 }, { "epoch": 0.6620962493497768, "grad_norm": 1.1308807134628296, "learning_rate": 0.00019588086339285244, "loss": 4.4093, "step": 354800 }, { "epoch": 0.6621895549091553, "grad_norm": 0.8613793849945068, "learning_rate": 0.0001958796700077852, "loss": 4.6998, "step": 354850 }, { "epoch": 0.6622828604685339, "grad_norm": 0.9791272282600403, "learning_rate": 0.00019587847645350718, "loss": 4.4469, "step": 354900 }, { "epoch": 0.6623761660279124, "grad_norm": 0.8936867713928223, "learning_rate": 0.00019587728273002056, "loss": 4.5074, "step": 354950 }, { "epoch": 0.6624694715872909, "grad_norm": 1.1822404861450195, "learning_rate": 0.00019587608883732738, "loss": 4.4717, "step": 355000 }, { "epoch": 0.6625627771466693, "grad_norm": 0.9689282178878784, "learning_rate": 0.00019587489477542976, "loss": 4.56, "step": 355050 }, { "epoch": 0.6626560827060478, "grad_norm": 0.9512212872505188, "learning_rate": 0.00019587370054432985, "loss": 4.6973, "step": 355100 }, { "epoch": 0.6627493882654263, "grad_norm": 1.2343778610229492, "learning_rate": 0.00019587250614402972, "loss": 4.5545, "step": 355150 }, { "epoch": 0.6628426938248048, "grad_norm": 1.140806794166565, "learning_rate": 0.00019587131157453147, "loss": 4.5647, "step": 355200 }, { "epoch": 0.6629359993841833, "grad_norm": 0.8960227966308594, "learning_rate": 0.00019587011683583724, "loss": 4.6904, "step": 355250 }, { "epoch": 0.6630293049435618, "grad_norm": 1.0368685722351074, "learning_rate": 0.00019586892192794913, "loss": 4.5606, "step": 355300 }, { "epoch": 0.6631226105029403, "grad_norm": 1.2226470708847046, "learning_rate": 0.00019586772685086922, "loss": 4.4205, "step": 355350 }, { "epoch": 0.6632159160623188, "grad_norm": 1.3335462808609009, "learning_rate": 0.00019586653160459964, "loss": 4.3505, "step": 355400 }, { "epoch": 0.6633092216216973, "grad_norm": 0.8618627786636353, "learning_rate": 0.00019586533618914253, "loss": 4.5772, "step": 355450 }, { "epoch": 0.6634025271810757, "grad_norm": 0.8240525722503662, "learning_rate": 0.00019586414060449996, "loss": 4.4362, "step": 355500 }, { "epoch": 0.6634958327404542, "grad_norm": 1.1974875926971436, "learning_rate": 0.00019586294485067404, "loss": 4.514, "step": 355550 }, { "epoch": 0.6635891382998328, "grad_norm": 0.8910797238349915, "learning_rate": 0.0001958617489276669, "loss": 4.4302, "step": 355600 }, { "epoch": 0.6636824438592113, "grad_norm": 0.7819907069206238, "learning_rate": 0.0001958605528354806, "loss": 4.5097, "step": 355650 }, { "epoch": 0.6637757494185897, "grad_norm": 0.8999784588813782, "learning_rate": 0.00019585935657411732, "loss": 4.3274, "step": 355700 }, { "epoch": 0.6638690549779682, "grad_norm": 0.8418229222297668, "learning_rate": 0.00019585816014357915, "loss": 4.5235, "step": 355750 }, { "epoch": 0.6639623605373467, "grad_norm": 1.0023940801620483, "learning_rate": 0.0001958569635438682, "loss": 4.4684, "step": 355800 }, { "epoch": 0.6640556660967252, "grad_norm": 1.0244237184524536, "learning_rate": 0.00019585576677498656, "loss": 4.573, "step": 355850 }, { "epoch": 0.6641489716561036, "grad_norm": 1.1590559482574463, "learning_rate": 0.00019585456983693637, "loss": 4.4707, "step": 355900 }, { "epoch": 0.6642422772154822, "grad_norm": 0.977392852306366, "learning_rate": 0.0001958533727297197, "loss": 4.68, "step": 355950 }, { "epoch": 0.6643355827748607, "grad_norm": 1.1569302082061768, "learning_rate": 0.00019585217545333873, "loss": 4.5928, "step": 356000 }, { "epoch": 0.6644288883342392, "grad_norm": 1.1341270208358765, "learning_rate": 0.0001958509780077955, "loss": 4.697, "step": 356050 }, { "epoch": 0.6645221938936177, "grad_norm": 1.0654176473617554, "learning_rate": 0.0001958497803930922, "loss": 4.5901, "step": 356100 }, { "epoch": 0.6646154994529961, "grad_norm": 0.9715544581413269, "learning_rate": 0.0001958485826092309, "loss": 4.3667, "step": 356150 }, { "epoch": 0.6647088050123746, "grad_norm": 0.9358380436897278, "learning_rate": 0.00019584738465621367, "loss": 4.6559, "step": 356200 }, { "epoch": 0.6648021105717531, "grad_norm": 1.0738136768341064, "learning_rate": 0.00019584618653404272, "loss": 4.7552, "step": 356250 }, { "epoch": 0.6648954161311317, "grad_norm": 0.8731696009635925, "learning_rate": 0.0001958449882427201, "loss": 4.589, "step": 356300 }, { "epoch": 0.6649887216905102, "grad_norm": 1.1186859607696533, "learning_rate": 0.00019584378978224794, "loss": 4.4606, "step": 356350 }, { "epoch": 0.6650820272498886, "grad_norm": 1.0473381280899048, "learning_rate": 0.00019584259115262834, "loss": 4.5423, "step": 356400 }, { "epoch": 0.6651753328092671, "grad_norm": 0.7153509259223938, "learning_rate": 0.00019584139235386344, "loss": 4.5403, "step": 356450 }, { "epoch": 0.6652686383686456, "grad_norm": 0.6838157773017883, "learning_rate": 0.00019584019338595536, "loss": 4.5531, "step": 356500 }, { "epoch": 0.6653619439280241, "grad_norm": 1.0864678621292114, "learning_rate": 0.0001958389942489062, "loss": 4.5307, "step": 356550 }, { "epoch": 0.6654552494874025, "grad_norm": 1.2366557121276855, "learning_rate": 0.00019583779494271805, "loss": 4.4318, "step": 356600 }, { "epoch": 0.6655485550467811, "grad_norm": 1.153394103050232, "learning_rate": 0.00019583659546739307, "loss": 4.5182, "step": 356650 }, { "epoch": 0.6656418606061596, "grad_norm": 1.1458213329315186, "learning_rate": 0.0001958353958229334, "loss": 4.6159, "step": 356700 }, { "epoch": 0.6657351661655381, "grad_norm": 1.208235502243042, "learning_rate": 0.00019583419600934106, "loss": 4.614, "step": 356750 }, { "epoch": 0.6658284717249165, "grad_norm": 0.8747703433036804, "learning_rate": 0.00019583299602661826, "loss": 4.7372, "step": 356800 }, { "epoch": 0.665921777284295, "grad_norm": 1.0256723165512085, "learning_rate": 0.00019583179587476707, "loss": 4.4512, "step": 356850 }, { "epoch": 0.6660150828436735, "grad_norm": 1.1350177526474, "learning_rate": 0.00019583059555378963, "loss": 4.4028, "step": 356900 }, { "epoch": 0.666108388403052, "grad_norm": 0.8919464945793152, "learning_rate": 0.00019582939506368807, "loss": 4.5091, "step": 356950 }, { "epoch": 0.6662016939624306, "grad_norm": 1.2155777215957642, "learning_rate": 0.00019582819440446447, "loss": 4.6637, "step": 357000 }, { "epoch": 0.6662016939624306, "eval_loss": 4.740085124969482, "eval_runtime": 233.202, "eval_samples_per_second": 11.183, "eval_steps_per_second": 11.183, "eval_tts_loss": 7.604406293443021, "step": 357000 }, { "epoch": 0.666294999521809, "grad_norm": 1.0060449838638306, "learning_rate": 0.00019582699357612095, "loss": 4.5867, "step": 357050 }, { "epoch": 0.6663883050811875, "grad_norm": 1.1478132009506226, "learning_rate": 0.0001958257925786597, "loss": 4.4616, "step": 357100 }, { "epoch": 0.666481610640566, "grad_norm": 1.0752184391021729, "learning_rate": 0.00019582459141208273, "loss": 4.56, "step": 357150 }, { "epoch": 0.6665749161999445, "grad_norm": 1.305156946182251, "learning_rate": 0.00019582339007639225, "loss": 4.6758, "step": 357200 }, { "epoch": 0.6666682217593229, "grad_norm": 0.9395535588264465, "learning_rate": 0.00019582218857159035, "loss": 4.537, "step": 357250 }, { "epoch": 0.6667615273187014, "grad_norm": 1.31413996219635, "learning_rate": 0.00019582098689767912, "loss": 4.5146, "step": 357300 }, { "epoch": 0.66685483287808, "grad_norm": 1.1975082159042358, "learning_rate": 0.00019581978505466073, "loss": 4.5509, "step": 357350 }, { "epoch": 0.6669481384374585, "grad_norm": 0.9953302145004272, "learning_rate": 0.00019581858304253727, "loss": 4.4775, "step": 357400 }, { "epoch": 0.667041443996837, "grad_norm": 1.0209640264511108, "learning_rate": 0.00019581738086131087, "loss": 4.4213, "step": 357450 }, { "epoch": 0.6671347495562154, "grad_norm": 1.3481035232543945, "learning_rate": 0.00019581617851098366, "loss": 4.6205, "step": 357500 }, { "epoch": 0.6672280551155939, "grad_norm": 1.155086636543274, "learning_rate": 0.00019581497599155775, "loss": 4.5012, "step": 357550 }, { "epoch": 0.6673213606749724, "grad_norm": 1.0793018341064453, "learning_rate": 0.00019581377330303527, "loss": 4.6158, "step": 357600 }, { "epoch": 0.6674146662343509, "grad_norm": 1.0425056219100952, "learning_rate": 0.00019581257044541835, "loss": 4.5017, "step": 357650 }, { "epoch": 0.6675079717937294, "grad_norm": 0.9235154986381531, "learning_rate": 0.00019581136741870906, "loss": 4.5477, "step": 357700 }, { "epoch": 0.6676012773531079, "grad_norm": 0.9167792201042175, "learning_rate": 0.0001958101642229096, "loss": 4.687, "step": 357750 }, { "epoch": 0.6676945829124864, "grad_norm": 0.7527122497558594, "learning_rate": 0.00019580896085802206, "loss": 4.3443, "step": 357800 }, { "epoch": 0.6677878884718649, "grad_norm": 0.9313538670539856, "learning_rate": 0.00019580775732404854, "loss": 4.3117, "step": 357850 }, { "epoch": 0.6678811940312434, "grad_norm": 0.930475115776062, "learning_rate": 0.0001958065536209912, "loss": 4.4561, "step": 357900 }, { "epoch": 0.6679744995906218, "grad_norm": 0.9051177501678467, "learning_rate": 0.00019580534974885215, "loss": 4.6001, "step": 357950 }, { "epoch": 0.6680678051500003, "grad_norm": 0.922713041305542, "learning_rate": 0.0001958041457076335, "loss": 4.6231, "step": 358000 }, { "epoch": 0.6681611107093789, "grad_norm": 1.1627779006958008, "learning_rate": 0.00019580294149733742, "loss": 4.4918, "step": 358050 }, { "epoch": 0.6682544162687574, "grad_norm": 1.1160223484039307, "learning_rate": 0.00019580173711796598, "loss": 4.5517, "step": 358100 }, { "epoch": 0.6683477218281358, "grad_norm": 1.067106008529663, "learning_rate": 0.00019580053256952135, "loss": 4.5297, "step": 358150 }, { "epoch": 0.6684410273875143, "grad_norm": 0.7735969424247742, "learning_rate": 0.00019579932785200566, "loss": 4.4722, "step": 358200 }, { "epoch": 0.6685343329468928, "grad_norm": 1.1097967624664307, "learning_rate": 0.00019579812296542098, "loss": 4.5994, "step": 358250 }, { "epoch": 0.6686276385062713, "grad_norm": 1.2591450214385986, "learning_rate": 0.00019579691790976944, "loss": 4.6635, "step": 358300 }, { "epoch": 0.6687209440656497, "grad_norm": 1.1494500637054443, "learning_rate": 0.00019579571268505324, "loss": 4.6788, "step": 358350 }, { "epoch": 0.6688142496250283, "grad_norm": 0.966873824596405, "learning_rate": 0.00019579450729127445, "loss": 4.4838, "step": 358400 }, { "epoch": 0.6689075551844068, "grad_norm": 1.4860520362854004, "learning_rate": 0.0001957933017284352, "loss": 4.6094, "step": 358450 }, { "epoch": 0.6690008607437853, "grad_norm": 1.0449298620224, "learning_rate": 0.00019579209599653765, "loss": 4.6897, "step": 358500 }, { "epoch": 0.6690941663031638, "grad_norm": 1.0751429796218872, "learning_rate": 0.0001957908900955839, "loss": 4.5401, "step": 358550 }, { "epoch": 0.6691874718625422, "grad_norm": 0.7464136481285095, "learning_rate": 0.00019578968402557607, "loss": 4.5338, "step": 358600 }, { "epoch": 0.6692807774219207, "grad_norm": 0.916556715965271, "learning_rate": 0.0001957884777865163, "loss": 4.6826, "step": 358650 }, { "epoch": 0.6693740829812992, "grad_norm": 0.8825215101242065, "learning_rate": 0.00019578727137840673, "loss": 4.6662, "step": 358700 }, { "epoch": 0.6694673885406778, "grad_norm": 1.216315507888794, "learning_rate": 0.0001957860648012495, "loss": 4.6074, "step": 358750 }, { "epoch": 0.6695606941000563, "grad_norm": 1.1252713203430176, "learning_rate": 0.00019578485805504667, "loss": 4.624, "step": 358800 }, { "epoch": 0.6696539996594347, "grad_norm": 0.9946227073669434, "learning_rate": 0.0001957836511398005, "loss": 4.404, "step": 358850 }, { "epoch": 0.6697473052188132, "grad_norm": 0.8797948956489563, "learning_rate": 0.00019578244405551296, "loss": 4.5698, "step": 358900 }, { "epoch": 0.6698406107781917, "grad_norm": 0.8493410348892212, "learning_rate": 0.0001957812368021863, "loss": 4.5516, "step": 358950 }, { "epoch": 0.6699339163375702, "grad_norm": 0.9304854869842529, "learning_rate": 0.00019578002937982258, "loss": 4.5322, "step": 359000 }, { "epoch": 0.6700272218969486, "grad_norm": 1.0428178310394287, "learning_rate": 0.00019577882178842398, "loss": 4.3919, "step": 359050 }, { "epoch": 0.6701205274563272, "grad_norm": 0.9189246296882629, "learning_rate": 0.00019577761402799263, "loss": 4.6786, "step": 359100 }, { "epoch": 0.6702138330157057, "grad_norm": 1.20093834400177, "learning_rate": 0.00019577640609853063, "loss": 4.7049, "step": 359150 }, { "epoch": 0.6703071385750842, "grad_norm": 1.0256885290145874, "learning_rate": 0.00019577519800004009, "loss": 4.4463, "step": 359200 }, { "epoch": 0.6704004441344626, "grad_norm": 1.009544014930725, "learning_rate": 0.00019577398973252323, "loss": 4.4192, "step": 359250 }, { "epoch": 0.6704937496938411, "grad_norm": 0.9781863689422607, "learning_rate": 0.0001957727812959821, "loss": 4.7262, "step": 359300 }, { "epoch": 0.6705870552532196, "grad_norm": 0.7879834175109863, "learning_rate": 0.00019577157269041887, "loss": 4.548, "step": 359350 }, { "epoch": 0.6706803608125981, "grad_norm": 0.9353818893432617, "learning_rate": 0.00019577036391583568, "loss": 4.64, "step": 359400 }, { "epoch": 0.6707736663719767, "grad_norm": 1.063107967376709, "learning_rate": 0.00019576915497223466, "loss": 4.7691, "step": 359450 }, { "epoch": 0.6708669719313551, "grad_norm": 1.1930428743362427, "learning_rate": 0.00019576794585961788, "loss": 4.5924, "step": 359500 }, { "epoch": 0.6709602774907336, "grad_norm": 1.2417848110198975, "learning_rate": 0.00019576673657798756, "loss": 4.4907, "step": 359550 }, { "epoch": 0.6710535830501121, "grad_norm": 0.8512386679649353, "learning_rate": 0.0001957655271273458, "loss": 4.402, "step": 359600 }, { "epoch": 0.6711468886094906, "grad_norm": 1.0181959867477417, "learning_rate": 0.00019576431750769471, "loss": 4.7324, "step": 359650 }, { "epoch": 0.671240194168869, "grad_norm": 0.8717050552368164, "learning_rate": 0.0001957631077190365, "loss": 4.4207, "step": 359700 }, { "epoch": 0.6713334997282475, "grad_norm": 1.1153290271759033, "learning_rate": 0.00019576189776137324, "loss": 4.5153, "step": 359750 }, { "epoch": 0.6714268052876261, "grad_norm": 1.100089430809021, "learning_rate": 0.00019576068763470705, "loss": 4.6634, "step": 359800 }, { "epoch": 0.6715201108470046, "grad_norm": 1.1623677015304565, "learning_rate": 0.0001957594773390401, "loss": 4.5501, "step": 359850 }, { "epoch": 0.6716134164063831, "grad_norm": 1.1294299364089966, "learning_rate": 0.00019575826687437455, "loss": 4.5091, "step": 359900 }, { "epoch": 0.6717067219657615, "grad_norm": 1.3196748495101929, "learning_rate": 0.0001957570562407125, "loss": 4.5307, "step": 359950 }, { "epoch": 0.67180002752514, "grad_norm": 1.1398862600326538, "learning_rate": 0.00019575584543805608, "loss": 4.5979, "step": 360000 }, { "epoch": 0.67180002752514, "eval_loss": 4.742733478546143, "eval_runtime": 231.3397, "eval_samples_per_second": 11.273, "eval_steps_per_second": 11.273, "eval_tts_loss": 7.567788422349097, "step": 360000 }, { "epoch": 0.6718933330845185, "grad_norm": 0.9543121457099915, "learning_rate": 0.00019575463446640744, "loss": 4.5505, "step": 360050 }, { "epoch": 0.671986638643897, "grad_norm": 1.1611236333847046, "learning_rate": 0.00019575342332576873, "loss": 4.5798, "step": 360100 }, { "epoch": 0.6720799442032754, "grad_norm": 1.2661935091018677, "learning_rate": 0.0001957522120161421, "loss": 4.5332, "step": 360150 }, { "epoch": 0.672173249762654, "grad_norm": 1.149210810661316, "learning_rate": 0.0001957510005375296, "loss": 4.413, "step": 360200 }, { "epoch": 0.6722665553220325, "grad_norm": 0.6918655633926392, "learning_rate": 0.0001957497888899335, "loss": 4.4898, "step": 360250 }, { "epoch": 0.672359860881411, "grad_norm": 0.9333517551422119, "learning_rate": 0.00019574857707335583, "loss": 4.6921, "step": 360300 }, { "epoch": 0.6724531664407895, "grad_norm": 1.155081868171692, "learning_rate": 0.00019574736508779877, "loss": 4.6113, "step": 360350 }, { "epoch": 0.6725464720001679, "grad_norm": 0.6867272257804871, "learning_rate": 0.00019574615293326447, "loss": 4.5905, "step": 360400 }, { "epoch": 0.6726397775595464, "grad_norm": 1.0847917795181274, "learning_rate": 0.00019574494060975504, "loss": 4.6095, "step": 360450 }, { "epoch": 0.6727330831189249, "grad_norm": 0.8768734335899353, "learning_rate": 0.00019574372811727265, "loss": 4.6401, "step": 360500 }, { "epoch": 0.6728263886783035, "grad_norm": 1.27104651927948, "learning_rate": 0.00019574251545581946, "loss": 4.5056, "step": 360550 }, { "epoch": 0.6729196942376819, "grad_norm": 1.0164141654968262, "learning_rate": 0.00019574130262539754, "loss": 4.6382, "step": 360600 }, { "epoch": 0.6730129997970604, "grad_norm": 1.1899137496948242, "learning_rate": 0.00019574008962600908, "loss": 4.5382, "step": 360650 }, { "epoch": 0.6731063053564389, "grad_norm": 1.0625330209732056, "learning_rate": 0.0001957388764576562, "loss": 4.6297, "step": 360700 }, { "epoch": 0.6731996109158174, "grad_norm": 0.9368055462837219, "learning_rate": 0.00019573766312034103, "loss": 4.7591, "step": 360750 }, { "epoch": 0.6732929164751958, "grad_norm": 1.0346652269363403, "learning_rate": 0.00019573644961406577, "loss": 4.5228, "step": 360800 }, { "epoch": 0.6733862220345743, "grad_norm": 1.0345101356506348, "learning_rate": 0.00019573523593883249, "loss": 4.6289, "step": 360850 }, { "epoch": 0.6734795275939529, "grad_norm": 1.3405543565750122, "learning_rate": 0.00019573402209464334, "loss": 4.6347, "step": 360900 }, { "epoch": 0.6735728331533314, "grad_norm": 0.9958884716033936, "learning_rate": 0.00019573280808150053, "loss": 4.5511, "step": 360950 }, { "epoch": 0.6736661387127099, "grad_norm": 1.1798781156539917, "learning_rate": 0.00019573159389940618, "loss": 4.4071, "step": 361000 }, { "epoch": 0.6737594442720883, "grad_norm": 0.8254154324531555, "learning_rate": 0.00019573037954836236, "loss": 4.6153, "step": 361050 }, { "epoch": 0.6738527498314668, "grad_norm": 1.03472900390625, "learning_rate": 0.0001957291650283713, "loss": 4.4614, "step": 361100 }, { "epoch": 0.6739460553908453, "grad_norm": 0.7731339931488037, "learning_rate": 0.00019572795033943508, "loss": 4.5691, "step": 361150 }, { "epoch": 0.6740393609502238, "grad_norm": 0.7978653907775879, "learning_rate": 0.00019572673548155586, "loss": 4.6731, "step": 361200 }, { "epoch": 0.6741326665096024, "grad_norm": 0.8550422191619873, "learning_rate": 0.00019572552045473581, "loss": 4.4369, "step": 361250 }, { "epoch": 0.6742259720689808, "grad_norm": 1.2098850011825562, "learning_rate": 0.0001957243052589771, "loss": 4.614, "step": 361300 }, { "epoch": 0.6743192776283593, "grad_norm": 1.0120937824249268, "learning_rate": 0.00019572308989428182, "loss": 4.5158, "step": 361350 }, { "epoch": 0.6744125831877378, "grad_norm": 1.155301809310913, "learning_rate": 0.0001957218743606521, "loss": 4.5678, "step": 361400 }, { "epoch": 0.6745058887471163, "grad_norm": 1.0626083612442017, "learning_rate": 0.0001957206586580901, "loss": 4.4325, "step": 361450 }, { "epoch": 0.6745991943064947, "grad_norm": 0.9684640765190125, "learning_rate": 0.00019571944278659804, "loss": 4.61, "step": 361500 }, { "epoch": 0.6746924998658732, "grad_norm": 0.969910204410553, "learning_rate": 0.00019571822674617796, "loss": 4.4414, "step": 361550 }, { "epoch": 0.6747858054252518, "grad_norm": 1.1967800855636597, "learning_rate": 0.00019571701053683205, "loss": 4.6256, "step": 361600 }, { "epoch": 0.6748791109846303, "grad_norm": 0.913022518157959, "learning_rate": 0.00019571579415856246, "loss": 4.617, "step": 361650 }, { "epoch": 0.6749724165440087, "grad_norm": 0.9121726155281067, "learning_rate": 0.00019571457761137137, "loss": 4.5678, "step": 361700 }, { "epoch": 0.6750657221033872, "grad_norm": 1.1886693239212036, "learning_rate": 0.00019571336089526086, "loss": 4.6723, "step": 361750 }, { "epoch": 0.6751590276627657, "grad_norm": 1.0208622217178345, "learning_rate": 0.00019571214401023314, "loss": 4.5746, "step": 361800 }, { "epoch": 0.6752523332221442, "grad_norm": 1.072257399559021, "learning_rate": 0.0001957109269562903, "loss": 4.5361, "step": 361850 }, { "epoch": 0.6753456387815227, "grad_norm": 1.0889387130737305, "learning_rate": 0.00019570970973343453, "loss": 4.747, "step": 361900 }, { "epoch": 0.6754389443409012, "grad_norm": 1.2097222805023193, "learning_rate": 0.00019570849234166793, "loss": 4.5434, "step": 361950 }, { "epoch": 0.6755322499002797, "grad_norm": 1.045190453529358, "learning_rate": 0.0001957072747809927, "loss": 4.4708, "step": 362000 }, { "epoch": 0.6756255554596582, "grad_norm": 1.025077223777771, "learning_rate": 0.000195706057051411, "loss": 4.4576, "step": 362050 }, { "epoch": 0.6757188610190367, "grad_norm": 1.1330491304397583, "learning_rate": 0.0001957048391529249, "loss": 4.6768, "step": 362100 }, { "epoch": 0.6758121665784151, "grad_norm": 1.0041680335998535, "learning_rate": 0.00019570362108553663, "loss": 4.5468, "step": 362150 }, { "epoch": 0.6759054721377936, "grad_norm": 1.4580110311508179, "learning_rate": 0.00019570240284924834, "loss": 4.3164, "step": 362200 }, { "epoch": 0.6759987776971721, "grad_norm": 0.8612103462219238, "learning_rate": 0.00019570118444406208, "loss": 4.5728, "step": 362250 }, { "epoch": 0.6760920832565507, "grad_norm": 1.135723352432251, "learning_rate": 0.0001956999658699801, "loss": 4.5523, "step": 362300 }, { "epoch": 0.6761853888159292, "grad_norm": 0.9825050830841064, "learning_rate": 0.0001956987471270045, "loss": 4.5548, "step": 362350 }, { "epoch": 0.6762786943753076, "grad_norm": 1.275405764579773, "learning_rate": 0.00019569752821513748, "loss": 4.4826, "step": 362400 }, { "epoch": 0.6763719999346861, "grad_norm": 1.0499589443206787, "learning_rate": 0.00019569630913438116, "loss": 4.635, "step": 362450 }, { "epoch": 0.6764653054940646, "grad_norm": 0.9857321381568909, "learning_rate": 0.0001956950898847377, "loss": 4.5577, "step": 362500 }, { "epoch": 0.6765586110534431, "grad_norm": 1.0807452201843262, "learning_rate": 0.0001956938704662092, "loss": 4.393, "step": 362550 }, { "epoch": 0.6766519166128215, "grad_norm": 1.0702356100082397, "learning_rate": 0.00019569265087879788, "loss": 4.721, "step": 362600 }, { "epoch": 0.6767452221722001, "grad_norm": 0.8599978685379028, "learning_rate": 0.00019569143112250587, "loss": 4.5146, "step": 362650 }, { "epoch": 0.6768385277315786, "grad_norm": 0.9826505780220032, "learning_rate": 0.00019569021119733533, "loss": 4.5429, "step": 362700 }, { "epoch": 0.6769318332909571, "grad_norm": 1.0868836641311646, "learning_rate": 0.0001956889911032884, "loss": 4.527, "step": 362750 }, { "epoch": 0.6770251388503356, "grad_norm": 1.200915813446045, "learning_rate": 0.00019568777084036723, "loss": 4.5566, "step": 362800 }, { "epoch": 0.677118444409714, "grad_norm": 1.088509440422058, "learning_rate": 0.000195686550408574, "loss": 4.4763, "step": 362850 }, { "epoch": 0.6772117499690925, "grad_norm": 1.1293811798095703, "learning_rate": 0.00019568532980791082, "loss": 4.5489, "step": 362900 }, { "epoch": 0.677305055528471, "grad_norm": 1.165267825126648, "learning_rate": 0.0001956841090383799, "loss": 4.5753, "step": 362950 }, { "epoch": 0.6773983610878496, "grad_norm": 0.7049471735954285, "learning_rate": 0.00019568288809998334, "loss": 4.6423, "step": 363000 }, { "epoch": 0.6773983610878496, "eval_loss": 4.743714809417725, "eval_runtime": 231.9151, "eval_samples_per_second": 11.245, "eval_steps_per_second": 11.245, "eval_tts_loss": 7.5692039081952505, "step": 363000 }, { "epoch": 0.677491666647228, "grad_norm": 1.0355195999145508, "learning_rate": 0.00019568166699272333, "loss": 4.6849, "step": 363050 }, { "epoch": 0.6775849722066065, "grad_norm": 1.2881120443344116, "learning_rate": 0.00019568044571660197, "loss": 4.6834, "step": 363100 }, { "epoch": 0.677678277765985, "grad_norm": 1.2323970794677734, "learning_rate": 0.00019567922427162152, "loss": 4.5766, "step": 363150 }, { "epoch": 0.6777715833253635, "grad_norm": 1.278564691543579, "learning_rate": 0.00019567800265778403, "loss": 4.6025, "step": 363200 }, { "epoch": 0.677864888884742, "grad_norm": 0.8138118386268616, "learning_rate": 0.00019567678087509174, "loss": 4.5658, "step": 363250 }, { "epoch": 0.6779581944441204, "grad_norm": 0.691251814365387, "learning_rate": 0.00019567555892354674, "loss": 4.5252, "step": 363300 }, { "epoch": 0.678051500003499, "grad_norm": 0.8609877824783325, "learning_rate": 0.00019567433680315122, "loss": 4.5487, "step": 363350 }, { "epoch": 0.6781448055628775, "grad_norm": 0.8605664372444153, "learning_rate": 0.00019567311451390733, "loss": 4.4648, "step": 363400 }, { "epoch": 0.678238111122256, "grad_norm": 1.4465529918670654, "learning_rate": 0.00019567189205581723, "loss": 4.7005, "step": 363450 }, { "epoch": 0.6783314166816344, "grad_norm": 1.1591824293136597, "learning_rate": 0.00019567066942888307, "loss": 4.4901, "step": 363500 }, { "epoch": 0.6784247222410129, "grad_norm": 1.1295784711837769, "learning_rate": 0.000195669446633107, "loss": 4.583, "step": 363550 }, { "epoch": 0.6785180278003914, "grad_norm": 0.9203224778175354, "learning_rate": 0.0001956682236684912, "loss": 4.3986, "step": 363600 }, { "epoch": 0.6786113333597699, "grad_norm": 0.9815191030502319, "learning_rate": 0.00019566700053503784, "loss": 4.5553, "step": 363650 }, { "epoch": 0.6787046389191485, "grad_norm": 0.9840654730796814, "learning_rate": 0.00019566577723274903, "loss": 4.4235, "step": 363700 }, { "epoch": 0.6787979444785269, "grad_norm": 1.0703121423721313, "learning_rate": 0.00019566455376162698, "loss": 4.5737, "step": 363750 }, { "epoch": 0.6788912500379054, "grad_norm": 0.8637380599975586, "learning_rate": 0.00019566333012167379, "loss": 4.5224, "step": 363800 }, { "epoch": 0.6789845555972839, "grad_norm": 0.9189258813858032, "learning_rate": 0.00019566210631289172, "loss": 4.4868, "step": 363850 }, { "epoch": 0.6790778611566624, "grad_norm": 0.9535520076751709, "learning_rate": 0.00019566088233528278, "loss": 4.4389, "step": 363900 }, { "epoch": 0.6791711667160408, "grad_norm": 1.0472358465194702, "learning_rate": 0.00019565965818884925, "loss": 4.6446, "step": 363950 }, { "epoch": 0.6792644722754193, "grad_norm": 1.2085245847702026, "learning_rate": 0.00019565843387359325, "loss": 4.5661, "step": 364000 }, { "epoch": 0.6793577778347979, "grad_norm": 1.175116777420044, "learning_rate": 0.000195657209389517, "loss": 4.571, "step": 364050 }, { "epoch": 0.6794510833941764, "grad_norm": 1.7191215753555298, "learning_rate": 0.00019565598473662255, "loss": 4.5139, "step": 364100 }, { "epoch": 0.6795443889535548, "grad_norm": 0.8249337077140808, "learning_rate": 0.0001956547599149121, "loss": 4.6354, "step": 364150 }, { "epoch": 0.6796376945129333, "grad_norm": 0.8554646372795105, "learning_rate": 0.00019565353492438788, "loss": 4.7298, "step": 364200 }, { "epoch": 0.6797310000723118, "grad_norm": 0.9862709045410156, "learning_rate": 0.000195652309765052, "loss": 4.5764, "step": 364250 }, { "epoch": 0.6798243056316903, "grad_norm": 1.3434244394302368, "learning_rate": 0.0001956510844369066, "loss": 4.7812, "step": 364300 }, { "epoch": 0.6799176111910687, "grad_norm": 0.7632136344909668, "learning_rate": 0.00019564985893995385, "loss": 4.6408, "step": 364350 }, { "epoch": 0.6800109167504473, "grad_norm": 1.0645220279693604, "learning_rate": 0.00019564863327419598, "loss": 4.4689, "step": 364400 }, { "epoch": 0.6801042223098258, "grad_norm": 0.6220859289169312, "learning_rate": 0.00019564740743963506, "loss": 4.5658, "step": 364450 }, { "epoch": 0.6801975278692043, "grad_norm": 1.4747847318649292, "learning_rate": 0.00019564618143627332, "loss": 4.6266, "step": 364500 }, { "epoch": 0.6802908334285828, "grad_norm": 1.1730914115905762, "learning_rate": 0.00019564495526411287, "loss": 4.6482, "step": 364550 }, { "epoch": 0.6803841389879612, "grad_norm": 0.8440443873405457, "learning_rate": 0.00019564372892315592, "loss": 4.552, "step": 364600 }, { "epoch": 0.6804774445473397, "grad_norm": 1.2068709135055542, "learning_rate": 0.00019564250241340463, "loss": 4.759, "step": 364650 }, { "epoch": 0.6805707501067182, "grad_norm": 0.9385483264923096, "learning_rate": 0.00019564127573486114, "loss": 4.4719, "step": 364700 }, { "epoch": 0.6806640556660968, "grad_norm": 0.8869559168815613, "learning_rate": 0.00019564004888752762, "loss": 4.5069, "step": 364750 }, { "epoch": 0.6807573612254753, "grad_norm": 1.1940581798553467, "learning_rate": 0.00019563882187140626, "loss": 4.7019, "step": 364800 }, { "epoch": 0.6808506667848537, "grad_norm": 1.2081252336502075, "learning_rate": 0.00019563759468649918, "loss": 4.5911, "step": 364850 }, { "epoch": 0.6809439723442322, "grad_norm": 1.456714153289795, "learning_rate": 0.0001956363673328086, "loss": 4.4947, "step": 364900 }, { "epoch": 0.6810372779036107, "grad_norm": 1.0345370769500732, "learning_rate": 0.00019563513981033664, "loss": 4.5642, "step": 364950 }, { "epoch": 0.6811305834629892, "grad_norm": 0.933967113494873, "learning_rate": 0.00019563391211908552, "loss": 4.6249, "step": 365000 }, { "epoch": 0.6812238890223676, "grad_norm": 1.1223982572555542, "learning_rate": 0.00019563268425905735, "loss": 4.5568, "step": 365050 }, { "epoch": 0.6813171945817462, "grad_norm": 0.7929447889328003, "learning_rate": 0.00019563145623025432, "loss": 4.4747, "step": 365100 }, { "epoch": 0.6814105001411247, "grad_norm": 1.0209295749664307, "learning_rate": 0.0001956302280326786, "loss": 4.7439, "step": 365150 }, { "epoch": 0.6815038057005032, "grad_norm": 0.9150970578193665, "learning_rate": 0.00019562899966633234, "loss": 4.4238, "step": 365200 }, { "epoch": 0.6815971112598816, "grad_norm": 0.9653599262237549, "learning_rate": 0.00019562777113121772, "loss": 4.5515, "step": 365250 }, { "epoch": 0.6816904168192601, "grad_norm": 1.1098837852478027, "learning_rate": 0.00019562654242733692, "loss": 4.6102, "step": 365300 }, { "epoch": 0.6817837223786386, "grad_norm": 0.9841729998588562, "learning_rate": 0.0001956253135546921, "loss": 4.5646, "step": 365350 }, { "epoch": 0.6818770279380171, "grad_norm": 1.2507600784301758, "learning_rate": 0.00019562408451328543, "loss": 4.6874, "step": 365400 }, { "epoch": 0.6819703334973956, "grad_norm": 0.8660342693328857, "learning_rate": 0.00019562285530311908, "loss": 4.5343, "step": 365450 }, { "epoch": 0.6820636390567741, "grad_norm": 1.0164939165115356, "learning_rate": 0.0001956216259241952, "loss": 4.6378, "step": 365500 }, { "epoch": 0.6821569446161526, "grad_norm": 0.9830291867256165, "learning_rate": 0.000195620396376516, "loss": 4.5258, "step": 365550 }, { "epoch": 0.6822502501755311, "grad_norm": 0.8058454394340515, "learning_rate": 0.0001956191666600836, "loss": 4.5012, "step": 365600 }, { "epoch": 0.6823435557349096, "grad_norm": 0.7354638576507568, "learning_rate": 0.0001956179367749002, "loss": 4.5234, "step": 365650 }, { "epoch": 0.682436861294288, "grad_norm": 0.9891838431358337, "learning_rate": 0.00019561670672096798, "loss": 4.5401, "step": 365700 }, { "epoch": 0.6825301668536665, "grad_norm": 0.7514564394950867, "learning_rate": 0.0001956154764982891, "loss": 4.5046, "step": 365750 }, { "epoch": 0.682623472413045, "grad_norm": 1.202129602432251, "learning_rate": 0.0001956142461068657, "loss": 4.4916, "step": 365800 }, { "epoch": 0.6827167779724236, "grad_norm": 1.1306408643722534, "learning_rate": 0.0001956130155467, "loss": 4.5489, "step": 365850 }, { "epoch": 0.6828100835318021, "grad_norm": 0.8608388900756836, "learning_rate": 0.00019561178481779413, "loss": 4.4624, "step": 365900 }, { "epoch": 0.6829033890911805, "grad_norm": 1.213686227798462, "learning_rate": 0.00019561055392015031, "loss": 4.4172, "step": 365950 }, { "epoch": 0.682996694650559, "grad_norm": 1.3149348497390747, "learning_rate": 0.00019560932285377067, "loss": 4.6674, "step": 366000 }, { "epoch": 0.682996694650559, "eval_loss": 4.745150089263916, "eval_runtime": 230.3389, "eval_samples_per_second": 11.322, "eval_steps_per_second": 11.322, "eval_tts_loss": 7.550067961383097, "step": 366000 }, { "epoch": 0.6830900002099375, "grad_norm": 1.0034360885620117, "learning_rate": 0.00019560809161865743, "loss": 4.6237, "step": 366050 }, { "epoch": 0.683183305769316, "grad_norm": 1.0511972904205322, "learning_rate": 0.00019560686021481268, "loss": 4.4821, "step": 366100 }, { "epoch": 0.6832766113286944, "grad_norm": 1.0858912467956543, "learning_rate": 0.00019560562864223867, "loss": 4.4884, "step": 366150 }, { "epoch": 0.683369916888073, "grad_norm": 0.6770386099815369, "learning_rate": 0.00019560439690093756, "loss": 4.4349, "step": 366200 }, { "epoch": 0.6834632224474515, "grad_norm": 1.1042362451553345, "learning_rate": 0.00019560316499091148, "loss": 4.6419, "step": 366250 }, { "epoch": 0.68355652800683, "grad_norm": 1.1873412132263184, "learning_rate": 0.00019560193291216267, "loss": 4.4536, "step": 366300 }, { "epoch": 0.6836498335662085, "grad_norm": 1.1348243951797485, "learning_rate": 0.00019560070066469324, "loss": 4.5514, "step": 366350 }, { "epoch": 0.6837431391255869, "grad_norm": 0.8972763419151306, "learning_rate": 0.00019559946824850542, "loss": 4.639, "step": 366400 }, { "epoch": 0.6838364446849654, "grad_norm": 1.2635902166366577, "learning_rate": 0.00019559823566360135, "loss": 4.5739, "step": 366450 }, { "epoch": 0.6839297502443439, "grad_norm": 0.8604574799537659, "learning_rate": 0.00019559700290998323, "loss": 4.5483, "step": 366500 }, { "epoch": 0.6840230558037225, "grad_norm": 1.056391716003418, "learning_rate": 0.0001955957699876532, "loss": 4.4846, "step": 366550 }, { "epoch": 0.6841163613631009, "grad_norm": 1.2378082275390625, "learning_rate": 0.00019559453689661346, "loss": 4.6516, "step": 366600 }, { "epoch": 0.6842096669224794, "grad_norm": 1.1226683855056763, "learning_rate": 0.00019559330363686617, "loss": 4.4961, "step": 366650 }, { "epoch": 0.6843029724818579, "grad_norm": 0.9702514410018921, "learning_rate": 0.00019559207020841354, "loss": 4.8676, "step": 366700 }, { "epoch": 0.6843962780412364, "grad_norm": 0.9354146122932434, "learning_rate": 0.00019559083661125772, "loss": 4.5858, "step": 366750 }, { "epoch": 0.6844895836006148, "grad_norm": 0.993274986743927, "learning_rate": 0.0001955896028454009, "loss": 4.4758, "step": 366800 }, { "epoch": 0.6845828891599933, "grad_norm": 1.0623561143875122, "learning_rate": 0.00019558836891084522, "loss": 4.3529, "step": 366850 }, { "epoch": 0.6846761947193719, "grad_norm": 0.9752203226089478, "learning_rate": 0.00019558713480759293, "loss": 4.6037, "step": 366900 }, { "epoch": 0.6847695002787504, "grad_norm": 0.9326674938201904, "learning_rate": 0.00019558590053564613, "loss": 4.6296, "step": 366950 }, { "epoch": 0.6848628058381289, "grad_norm": 0.8091044425964355, "learning_rate": 0.00019558466609500706, "loss": 4.698, "step": 367000 }, { "epoch": 0.6849561113975073, "grad_norm": 0.9038903117179871, "learning_rate": 0.00019558343148567788, "loss": 4.3031, "step": 367050 }, { "epoch": 0.6850494169568858, "grad_norm": 1.1735448837280273, "learning_rate": 0.00019558219670766074, "loss": 4.7071, "step": 367100 }, { "epoch": 0.6851427225162643, "grad_norm": 1.3830543756484985, "learning_rate": 0.00019558096176095787, "loss": 4.5986, "step": 367150 }, { "epoch": 0.6852360280756428, "grad_norm": 1.1118428707122803, "learning_rate": 0.00019557972664557136, "loss": 4.458, "step": 367200 }, { "epoch": 0.6853293336350214, "grad_norm": 1.1908377408981323, "learning_rate": 0.0001955784913615035, "loss": 4.5695, "step": 367250 }, { "epoch": 0.6854226391943998, "grad_norm": 1.0074695348739624, "learning_rate": 0.0001955772559087564, "loss": 4.7084, "step": 367300 }, { "epoch": 0.6855159447537783, "grad_norm": 1.166852355003357, "learning_rate": 0.0001955760202873323, "loss": 4.6143, "step": 367350 }, { "epoch": 0.6856092503131568, "grad_norm": 0.8768144845962524, "learning_rate": 0.00019557478449723327, "loss": 4.587, "step": 367400 }, { "epoch": 0.6857025558725353, "grad_norm": 1.2766517400741577, "learning_rate": 0.00019557354853846163, "loss": 4.6175, "step": 367450 }, { "epoch": 0.6857958614319137, "grad_norm": 1.0810915231704712, "learning_rate": 0.00019557231241101945, "loss": 4.5643, "step": 367500 }, { "epoch": 0.6858891669912922, "grad_norm": 1.072299838066101, "learning_rate": 0.000195571076114909, "loss": 4.4177, "step": 367550 }, { "epoch": 0.6859824725506708, "grad_norm": 0.7774250507354736, "learning_rate": 0.0001955698396501324, "loss": 4.4628, "step": 367600 }, { "epoch": 0.6860757781100493, "grad_norm": 0.9877749085426331, "learning_rate": 0.0001955686030166918, "loss": 4.5595, "step": 367650 }, { "epoch": 0.6861690836694277, "grad_norm": 0.8914515376091003, "learning_rate": 0.00019556736621458948, "loss": 4.5708, "step": 367700 }, { "epoch": 0.6862623892288062, "grad_norm": 0.8482488989830017, "learning_rate": 0.00019556612924382754, "loss": 4.558, "step": 367750 }, { "epoch": 0.6863556947881847, "grad_norm": 1.1105358600616455, "learning_rate": 0.00019556489210440824, "loss": 4.5492, "step": 367800 }, { "epoch": 0.6864490003475632, "grad_norm": 1.0241068601608276, "learning_rate": 0.00019556365479633372, "loss": 4.3774, "step": 367850 }, { "epoch": 0.6865423059069417, "grad_norm": 1.1066597700119019, "learning_rate": 0.00019556241731960613, "loss": 4.7176, "step": 367900 }, { "epoch": 0.6866356114663202, "grad_norm": 1.0185424089431763, "learning_rate": 0.0001955611796742277, "loss": 4.3786, "step": 367950 }, { "epoch": 0.6867289170256987, "grad_norm": 1.2061375379562378, "learning_rate": 0.00019555994186020058, "loss": 4.6226, "step": 368000 }, { "epoch": 0.6868222225850772, "grad_norm": 1.1836224794387817, "learning_rate": 0.000195558703877527, "loss": 4.5512, "step": 368050 }, { "epoch": 0.6869155281444557, "grad_norm": 0.7828928232192993, "learning_rate": 0.00019555746572620916, "loss": 4.6906, "step": 368100 }, { "epoch": 0.6870088337038341, "grad_norm": 0.9676310420036316, "learning_rate": 0.00019555622740624918, "loss": 4.4422, "step": 368150 }, { "epoch": 0.6871021392632126, "grad_norm": 1.104750156402588, "learning_rate": 0.00019555498891764926, "loss": 4.6423, "step": 368200 }, { "epoch": 0.6871954448225911, "grad_norm": 1.1759475469589233, "learning_rate": 0.00019555375026041158, "loss": 4.4813, "step": 368250 }, { "epoch": 0.6872887503819697, "grad_norm": 1.128739356994629, "learning_rate": 0.00019555251143453837, "loss": 4.5668, "step": 368300 }, { "epoch": 0.6873820559413482, "grad_norm": 1.0008203983306885, "learning_rate": 0.0001955512724400318, "loss": 4.3894, "step": 368350 }, { "epoch": 0.6874753615007266, "grad_norm": 0.8730279803276062, "learning_rate": 0.00019555003327689403, "loss": 4.4255, "step": 368400 }, { "epoch": 0.6875686670601051, "grad_norm": 1.139626145362854, "learning_rate": 0.00019554879394512726, "loss": 4.5072, "step": 368450 }, { "epoch": 0.6876619726194836, "grad_norm": 0.7365878224372864, "learning_rate": 0.00019554755444473369, "loss": 4.4943, "step": 368500 }, { "epoch": 0.6877552781788621, "grad_norm": 1.0410829782485962, "learning_rate": 0.0001955463147757155, "loss": 4.5267, "step": 368550 }, { "epoch": 0.6878485837382405, "grad_norm": 1.2380183935165405, "learning_rate": 0.00019554507493807486, "loss": 4.3674, "step": 368600 }, { "epoch": 0.6879418892976191, "grad_norm": 1.0722304582595825, "learning_rate": 0.000195543834931814, "loss": 4.4804, "step": 368650 }, { "epoch": 0.6880351948569976, "grad_norm": 1.3019156455993652, "learning_rate": 0.00019554259475693507, "loss": 4.5368, "step": 368700 }, { "epoch": 0.6881285004163761, "grad_norm": 1.0796579122543335, "learning_rate": 0.00019554135441344025, "loss": 4.5918, "step": 368750 }, { "epoch": 0.6882218059757546, "grad_norm": 1.541914463043213, "learning_rate": 0.0001955401139013318, "loss": 4.4014, "step": 368800 }, { "epoch": 0.688315111535133, "grad_norm": 1.0862489938735962, "learning_rate": 0.00019553887322061182, "loss": 4.545, "step": 368850 }, { "epoch": 0.6884084170945115, "grad_norm": 1.548496961593628, "learning_rate": 0.00019553763237128256, "loss": 4.7347, "step": 368900 }, { "epoch": 0.68850172265389, "grad_norm": 1.0292510986328125, "learning_rate": 0.00019553639135334617, "loss": 4.5278, "step": 368950 }, { "epoch": 0.6885950282132686, "grad_norm": 0.8597036004066467, "learning_rate": 0.00019553515016680487, "loss": 4.4832, "step": 369000 }, { "epoch": 0.6885950282132686, "eval_loss": 4.74437952041626, "eval_runtime": 228.019, "eval_samples_per_second": 11.438, "eval_steps_per_second": 11.438, "eval_tts_loss": 7.562755081445505, "step": 369000 }, { "epoch": 0.688688333772647, "grad_norm": 0.8960505127906799, "learning_rate": 0.00019553390881166081, "loss": 4.4335, "step": 369050 }, { "epoch": 0.6887816393320255, "grad_norm": 0.9919257760047913, "learning_rate": 0.00019553266728791627, "loss": 4.4563, "step": 369100 }, { "epoch": 0.688874944891404, "grad_norm": 0.7080592513084412, "learning_rate": 0.00019553142559557336, "loss": 4.4533, "step": 369150 }, { "epoch": 0.6889682504507825, "grad_norm": 1.239512324333191, "learning_rate": 0.00019553018373463427, "loss": 4.4225, "step": 369200 }, { "epoch": 0.689061556010161, "grad_norm": 0.8886670470237732, "learning_rate": 0.00019552894170510123, "loss": 4.592, "step": 369250 }, { "epoch": 0.6891548615695394, "grad_norm": 0.9981276988983154, "learning_rate": 0.00019552769950697643, "loss": 4.4187, "step": 369300 }, { "epoch": 0.689248167128918, "grad_norm": 1.1080340147018433, "learning_rate": 0.00019552645714026202, "loss": 4.6639, "step": 369350 }, { "epoch": 0.6893414726882965, "grad_norm": 1.150093674659729, "learning_rate": 0.00019552521460496026, "loss": 4.5057, "step": 369400 }, { "epoch": 0.689434778247675, "grad_norm": 1.16274893283844, "learning_rate": 0.00019552397190107326, "loss": 4.3424, "step": 369450 }, { "epoch": 0.6895280838070534, "grad_norm": 1.1337528228759766, "learning_rate": 0.0001955227290286033, "loss": 4.4874, "step": 369500 }, { "epoch": 0.6896213893664319, "grad_norm": 0.9183812737464905, "learning_rate": 0.0001955214859875525, "loss": 4.5758, "step": 369550 }, { "epoch": 0.6897146949258104, "grad_norm": 0.9727858901023865, "learning_rate": 0.0001955202427779231, "loss": 4.4889, "step": 369600 }, { "epoch": 0.6898080004851889, "grad_norm": 0.9852572083473206, "learning_rate": 0.00019551899939971725, "loss": 4.5826, "step": 369650 }, { "epoch": 0.6899013060445675, "grad_norm": 0.9665806889533997, "learning_rate": 0.0001955177558529372, "loss": 4.4956, "step": 369700 }, { "epoch": 0.6899946116039459, "grad_norm": 1.0975180864334106, "learning_rate": 0.00019551651213758512, "loss": 4.6726, "step": 369750 }, { "epoch": 0.6900879171633244, "grad_norm": 0.9996809959411621, "learning_rate": 0.0001955152682536632, "loss": 4.6079, "step": 369800 }, { "epoch": 0.6901812227227029, "grad_norm": 1.0467857122421265, "learning_rate": 0.00019551402420117363, "loss": 4.6961, "step": 369850 }, { "epoch": 0.6902745282820814, "grad_norm": 0.9526649117469788, "learning_rate": 0.00019551277998011864, "loss": 4.4674, "step": 369900 }, { "epoch": 0.6903678338414598, "grad_norm": 1.1415717601776123, "learning_rate": 0.00019551153559050035, "loss": 4.5515, "step": 369950 }, { "epoch": 0.6904611394008383, "grad_norm": 0.8399020433425903, "learning_rate": 0.00019551029103232105, "loss": 4.5819, "step": 370000 }, { "epoch": 0.6905544449602169, "grad_norm": 0.8485755324363708, "learning_rate": 0.00019550904630558286, "loss": 4.5073, "step": 370050 }, { "epoch": 0.6906477505195954, "grad_norm": 1.2100976705551147, "learning_rate": 0.00019550780141028802, "loss": 4.7205, "step": 370100 }, { "epoch": 0.6907410560789738, "grad_norm": 1.1322890520095825, "learning_rate": 0.00019550655634643872, "loss": 4.2542, "step": 370150 }, { "epoch": 0.6908343616383523, "grad_norm": 1.1923885345458984, "learning_rate": 0.00019550531111403715, "loss": 4.6466, "step": 370200 }, { "epoch": 0.6909276671977308, "grad_norm": 0.8890280723571777, "learning_rate": 0.00019550406571308552, "loss": 4.7235, "step": 370250 }, { "epoch": 0.6910209727571093, "grad_norm": 0.961129367351532, "learning_rate": 0.000195502820143586, "loss": 4.5495, "step": 370300 }, { "epoch": 0.6911142783164878, "grad_norm": 1.1575649976730347, "learning_rate": 0.00019550157440554084, "loss": 4.6855, "step": 370350 }, { "epoch": 0.6912075838758663, "grad_norm": 0.886288046836853, "learning_rate": 0.0001955003284989522, "loss": 4.4955, "step": 370400 }, { "epoch": 0.6913008894352448, "grad_norm": 1.3565956354141235, "learning_rate": 0.00019549908242382224, "loss": 4.6876, "step": 370450 }, { "epoch": 0.6913941949946233, "grad_norm": 0.9993792772293091, "learning_rate": 0.00019549783618015324, "loss": 4.6915, "step": 370500 }, { "epoch": 0.6914875005540018, "grad_norm": 1.1421886682510376, "learning_rate": 0.00019549658976794737, "loss": 4.5883, "step": 370550 }, { "epoch": 0.6915808061133802, "grad_norm": 0.9610868096351624, "learning_rate": 0.00019549534318720677, "loss": 4.5152, "step": 370600 }, { "epoch": 0.6916741116727587, "grad_norm": 0.917777419090271, "learning_rate": 0.00019549409643793375, "loss": 4.6137, "step": 370650 }, { "epoch": 0.6917674172321372, "grad_norm": 1.0799082517623901, "learning_rate": 0.00019549284952013042, "loss": 4.5084, "step": 370700 }, { "epoch": 0.6918607227915158, "grad_norm": 0.9805840849876404, "learning_rate": 0.00019549160243379903, "loss": 4.5066, "step": 370750 }, { "epoch": 0.6919540283508943, "grad_norm": 1.0332810878753662, "learning_rate": 0.00019549035517894177, "loss": 4.5318, "step": 370800 }, { "epoch": 0.6920473339102727, "grad_norm": 1.0908498764038086, "learning_rate": 0.0001954891077555608, "loss": 4.3942, "step": 370850 }, { "epoch": 0.6921406394696512, "grad_norm": 1.0144426822662354, "learning_rate": 0.00019548786016365837, "loss": 4.5183, "step": 370900 }, { "epoch": 0.6922339450290297, "grad_norm": 0.7185133099555969, "learning_rate": 0.0001954866124032367, "loss": 4.6889, "step": 370950 }, { "epoch": 0.6923272505884082, "grad_norm": 0.8159246444702148, "learning_rate": 0.00019548536447429793, "loss": 4.5158, "step": 371000 }, { "epoch": 0.6924205561477866, "grad_norm": 1.108191967010498, "learning_rate": 0.0001954841163768443, "loss": 4.6321, "step": 371050 }, { "epoch": 0.6925138617071651, "grad_norm": 1.095153570175171, "learning_rate": 0.00019548286811087802, "loss": 4.4778, "step": 371100 }, { "epoch": 0.6926071672665437, "grad_norm": 1.071548581123352, "learning_rate": 0.00019548161967640124, "loss": 4.3622, "step": 371150 }, { "epoch": 0.6927004728259222, "grad_norm": 1.2594348192214966, "learning_rate": 0.00019548037107341624, "loss": 4.4866, "step": 371200 }, { "epoch": 0.6927937783853007, "grad_norm": 0.8258947134017944, "learning_rate": 0.00019547912230192518, "loss": 4.3015, "step": 371250 }, { "epoch": 0.6928870839446791, "grad_norm": 1.2690938711166382, "learning_rate": 0.00019547787336193025, "loss": 4.6411, "step": 371300 }, { "epoch": 0.6929803895040576, "grad_norm": 0.8823020458221436, "learning_rate": 0.0001954766242534337, "loss": 4.5026, "step": 371350 }, { "epoch": 0.6930736950634361, "grad_norm": 1.126285195350647, "learning_rate": 0.00019547537497643767, "loss": 4.5313, "step": 371400 }, { "epoch": 0.6931670006228146, "grad_norm": 1.3908414840698242, "learning_rate": 0.00019547412553094442, "loss": 4.4273, "step": 371450 }, { "epoch": 0.6932603061821931, "grad_norm": 1.1015129089355469, "learning_rate": 0.00019547287591695613, "loss": 4.4119, "step": 371500 }, { "epoch": 0.6933536117415716, "grad_norm": 1.1324613094329834, "learning_rate": 0.00019547162613447504, "loss": 4.673, "step": 371550 }, { "epoch": 0.6934469173009501, "grad_norm": 1.167299747467041, "learning_rate": 0.00019547037618350332, "loss": 4.6662, "step": 371600 }, { "epoch": 0.6935402228603286, "grad_norm": 1.0740166902542114, "learning_rate": 0.00019546912606404316, "loss": 4.475, "step": 371650 }, { "epoch": 0.693633528419707, "grad_norm": 1.5000436305999756, "learning_rate": 0.00019546787577609683, "loss": 4.6277, "step": 371700 }, { "epoch": 0.6937268339790855, "grad_norm": 1.1670953035354614, "learning_rate": 0.00019546662531966645, "loss": 4.4253, "step": 371750 }, { "epoch": 0.693820139538464, "grad_norm": 1.2359504699707031, "learning_rate": 0.0001954653746947543, "loss": 4.606, "step": 371800 }, { "epoch": 0.6939134450978426, "grad_norm": 0.9903602004051208, "learning_rate": 0.0001954641239013626, "loss": 4.5031, "step": 371850 }, { "epoch": 0.6940067506572211, "grad_norm": 1.4943287372589111, "learning_rate": 0.00019546287293949346, "loss": 4.4695, "step": 371900 }, { "epoch": 0.6941000562165995, "grad_norm": 1.1207112073898315, "learning_rate": 0.00019546162180914916, "loss": 4.6373, "step": 371950 }, { "epoch": 0.694193361775978, "grad_norm": 1.0234216451644897, "learning_rate": 0.0001954603705103319, "loss": 4.5827, "step": 372000 }, { "epoch": 0.694193361775978, "eval_loss": 4.737115859985352, "eval_runtime": 230.2817, "eval_samples_per_second": 11.325, "eval_steps_per_second": 11.325, "eval_tts_loss": 7.549153191890237, "step": 372000 }, { "epoch": 0.6942866673353565, "grad_norm": 1.2011725902557373, "learning_rate": 0.00019545911904304388, "loss": 4.5869, "step": 372050 }, { "epoch": 0.694379972894735, "grad_norm": 0.8918785452842712, "learning_rate": 0.00019545786740728734, "loss": 4.3698, "step": 372100 }, { "epoch": 0.6944732784541134, "grad_norm": 1.012628436088562, "learning_rate": 0.00019545661560306442, "loss": 4.5073, "step": 372150 }, { "epoch": 0.694566584013492, "grad_norm": 1.0200284719467163, "learning_rate": 0.0001954553636303774, "loss": 4.6447, "step": 372200 }, { "epoch": 0.6946598895728705, "grad_norm": 0.9943399429321289, "learning_rate": 0.00019545411148922846, "loss": 4.3431, "step": 372250 }, { "epoch": 0.694753195132249, "grad_norm": 0.9219287037849426, "learning_rate": 0.00019545285917961977, "loss": 4.5561, "step": 372300 }, { "epoch": 0.6948465006916275, "grad_norm": 1.167061448097229, "learning_rate": 0.0001954516067015536, "loss": 4.7088, "step": 372350 }, { "epoch": 0.6949398062510059, "grad_norm": 1.2277780771255493, "learning_rate": 0.00019545035405503213, "loss": 4.6485, "step": 372400 }, { "epoch": 0.6950331118103844, "grad_norm": 1.0556718111038208, "learning_rate": 0.00019544910124005762, "loss": 4.5689, "step": 372450 }, { "epoch": 0.6951264173697629, "grad_norm": 0.7697402834892273, "learning_rate": 0.0001954478482566322, "loss": 4.5025, "step": 372500 }, { "epoch": 0.6952197229291415, "grad_norm": 1.5871305465698242, "learning_rate": 0.00019544659510475814, "loss": 4.3397, "step": 372550 }, { "epoch": 0.69531302848852, "grad_norm": 0.971028745174408, "learning_rate": 0.00019544534178443762, "loss": 4.6454, "step": 372600 }, { "epoch": 0.6954063340478984, "grad_norm": 1.0348927974700928, "learning_rate": 0.00019544408829567288, "loss": 4.6747, "step": 372650 }, { "epoch": 0.6954996396072769, "grad_norm": 0.8070318698883057, "learning_rate": 0.0001954428346384661, "loss": 4.5578, "step": 372700 }, { "epoch": 0.6955929451666554, "grad_norm": 1.4318119287490845, "learning_rate": 0.00019544158081281953, "loss": 4.5287, "step": 372750 }, { "epoch": 0.6956862507260339, "grad_norm": 1.0857127904891968, "learning_rate": 0.00019544032681873536, "loss": 4.61, "step": 372800 }, { "epoch": 0.6957795562854123, "grad_norm": 1.012400507926941, "learning_rate": 0.0001954390726562158, "loss": 4.5802, "step": 372850 }, { "epoch": 0.6958728618447909, "grad_norm": 0.7758302688598633, "learning_rate": 0.0001954378183252631, "loss": 4.6574, "step": 372900 }, { "epoch": 0.6959661674041694, "grad_norm": 0.9272875785827637, "learning_rate": 0.0001954365638258794, "loss": 4.6471, "step": 372950 }, { "epoch": 0.6960594729635479, "grad_norm": 0.9731526970863342, "learning_rate": 0.00019543530915806697, "loss": 4.4795, "step": 373000 }, { "epoch": 0.6961527785229263, "grad_norm": 1.3055826425552368, "learning_rate": 0.00019543405432182804, "loss": 4.6253, "step": 373050 }, { "epoch": 0.6962460840823048, "grad_norm": 1.2052303552627563, "learning_rate": 0.00019543279931716475, "loss": 4.62, "step": 373100 }, { "epoch": 0.6963393896416833, "grad_norm": 1.5236045122146606, "learning_rate": 0.0001954315441440794, "loss": 4.3725, "step": 373150 }, { "epoch": 0.6964326952010618, "grad_norm": 0.7871856093406677, "learning_rate": 0.00019543028880257416, "loss": 4.5097, "step": 373200 }, { "epoch": 0.6965260007604404, "grad_norm": 0.9522252678871155, "learning_rate": 0.00019542903329265122, "loss": 4.6569, "step": 373250 }, { "epoch": 0.6966193063198188, "grad_norm": 1.0478781461715698, "learning_rate": 0.00019542777761431286, "loss": 4.5729, "step": 373300 }, { "epoch": 0.6967126118791973, "grad_norm": 1.0283840894699097, "learning_rate": 0.00019542652176756124, "loss": 4.5673, "step": 373350 }, { "epoch": 0.6968059174385758, "grad_norm": 0.9941511154174805, "learning_rate": 0.00019542526575239862, "loss": 4.3899, "step": 373400 }, { "epoch": 0.6968992229979543, "grad_norm": 0.7745673656463623, "learning_rate": 0.00019542400956882718, "loss": 4.413, "step": 373450 }, { "epoch": 0.6969925285573327, "grad_norm": 0.7503075003623962, "learning_rate": 0.00019542275321684916, "loss": 4.4243, "step": 373500 }, { "epoch": 0.6970858341167112, "grad_norm": 0.9689463973045349, "learning_rate": 0.0001954214966964668, "loss": 4.5236, "step": 373550 }, { "epoch": 0.6971791396760898, "grad_norm": 0.9288371801376343, "learning_rate": 0.00019542024000768224, "loss": 4.5392, "step": 373600 }, { "epoch": 0.6972724452354683, "grad_norm": 0.9681942462921143, "learning_rate": 0.00019541898315049777, "loss": 4.5209, "step": 373650 }, { "epoch": 0.6973657507948468, "grad_norm": 1.1082279682159424, "learning_rate": 0.00019541772612491557, "loss": 4.4681, "step": 373700 }, { "epoch": 0.6974590563542252, "grad_norm": 1.3885143995285034, "learning_rate": 0.0001954164689309379, "loss": 4.6048, "step": 373750 }, { "epoch": 0.6975523619136037, "grad_norm": 1.106627106666565, "learning_rate": 0.00019541521156856692, "loss": 4.6785, "step": 373800 }, { "epoch": 0.6976456674729822, "grad_norm": 1.0747969150543213, "learning_rate": 0.00019541395403780491, "loss": 4.8345, "step": 373850 }, { "epoch": 0.6977389730323607, "grad_norm": 1.0435670614242554, "learning_rate": 0.00019541269633865403, "loss": 4.4634, "step": 373900 }, { "epoch": 0.6978322785917392, "grad_norm": 1.3467811346054077, "learning_rate": 0.00019541143847111658, "loss": 4.4974, "step": 373950 }, { "epoch": 0.6979255841511177, "grad_norm": 1.0148619413375854, "learning_rate": 0.00019541018043519467, "loss": 4.4536, "step": 374000 }, { "epoch": 0.6980188897104962, "grad_norm": 0.9159517288208008, "learning_rate": 0.00019540892223089064, "loss": 4.4959, "step": 374050 }, { "epoch": 0.6981121952698747, "grad_norm": 0.9616969227790833, "learning_rate": 0.0001954076638582066, "loss": 4.5215, "step": 374100 }, { "epoch": 0.6982055008292531, "grad_norm": 1.0186048746109009, "learning_rate": 0.00019540640531714483, "loss": 4.7394, "step": 374150 }, { "epoch": 0.6982988063886316, "grad_norm": 1.0677536725997925, "learning_rate": 0.00019540514660770756, "loss": 4.4439, "step": 374200 }, { "epoch": 0.6983921119480101, "grad_norm": 1.3627240657806396, "learning_rate": 0.00019540388772989697, "loss": 4.6828, "step": 374250 }, { "epoch": 0.6984854175073887, "grad_norm": 0.9445304274559021, "learning_rate": 0.0001954026286837153, "loss": 4.6344, "step": 374300 }, { "epoch": 0.6985787230667672, "grad_norm": 0.7951604723930359, "learning_rate": 0.0001954013694691648, "loss": 4.6203, "step": 374350 }, { "epoch": 0.6986720286261456, "grad_norm": 1.1664619445800781, "learning_rate": 0.00019540011008624766, "loss": 4.6081, "step": 374400 }, { "epoch": 0.6987653341855241, "grad_norm": 1.1221282482147217, "learning_rate": 0.0001953988505349661, "loss": 4.7079, "step": 374450 }, { "epoch": 0.6988586397449026, "grad_norm": 1.0007444620132446, "learning_rate": 0.00019539759081532238, "loss": 4.497, "step": 374500 }, { "epoch": 0.6989519453042811, "grad_norm": 1.0179212093353271, "learning_rate": 0.0001953963309273187, "loss": 4.5788, "step": 374550 }, { "epoch": 0.6990452508636595, "grad_norm": 1.243072271347046, "learning_rate": 0.00019539507087095725, "loss": 4.5208, "step": 374600 }, { "epoch": 0.6991385564230381, "grad_norm": 1.132432222366333, "learning_rate": 0.0001953938106462403, "loss": 4.7902, "step": 374650 }, { "epoch": 0.6992318619824166, "grad_norm": 1.2821457386016846, "learning_rate": 0.00019539255025317006, "loss": 4.5234, "step": 374700 }, { "epoch": 0.6993251675417951, "grad_norm": 1.0446197986602783, "learning_rate": 0.00019539128969174878, "loss": 4.5727, "step": 374750 }, { "epoch": 0.6994184731011736, "grad_norm": 0.5463175177574158, "learning_rate": 0.00019539002896197863, "loss": 4.4624, "step": 374800 }, { "epoch": 0.699511778660552, "grad_norm": 0.9870461225509644, "learning_rate": 0.00019538876806386187, "loss": 4.5074, "step": 374850 }, { "epoch": 0.6996050842199305, "grad_norm": 1.0726468563079834, "learning_rate": 0.0001953875069974007, "loss": 4.4284, "step": 374900 }, { "epoch": 0.699698389779309, "grad_norm": 1.2654712200164795, "learning_rate": 0.0001953862457625974, "loss": 4.574, "step": 374950 }, { "epoch": 0.6997916953386876, "grad_norm": 0.9296988844871521, "learning_rate": 0.00019538498435945414, "loss": 4.5662, "step": 375000 }, { "epoch": 0.6997916953386876, "eval_loss": 4.739675521850586, "eval_runtime": 230.6566, "eval_samples_per_second": 11.307, "eval_steps_per_second": 11.307, "eval_tts_loss": 7.568869058622699, "step": 375000 }, { "epoch": 0.699885000898066, "grad_norm": 0.8359538912773132, "learning_rate": 0.00019538372278797315, "loss": 4.3433, "step": 375050 }, { "epoch": 0.6999783064574445, "grad_norm": 1.1886621713638306, "learning_rate": 0.00019538246104815668, "loss": 4.5397, "step": 375100 }, { "epoch": 0.700071612016823, "grad_norm": 0.8133793473243713, "learning_rate": 0.00019538119914000695, "loss": 4.3441, "step": 375150 }, { "epoch": 0.7001649175762015, "grad_norm": 0.9325920939445496, "learning_rate": 0.00019537993706352622, "loss": 4.5081, "step": 375200 }, { "epoch": 0.70025822313558, "grad_norm": 0.9943052530288696, "learning_rate": 0.00019537867481871665, "loss": 4.3981, "step": 375250 }, { "epoch": 0.7003515286949584, "grad_norm": 0.8147107362747192, "learning_rate": 0.0001953774124055805, "loss": 4.6357, "step": 375300 }, { "epoch": 0.700444834254337, "grad_norm": 1.0404388904571533, "learning_rate": 0.00019537614982412, "loss": 4.7682, "step": 375350 }, { "epoch": 0.7005381398137155, "grad_norm": 1.011792540550232, "learning_rate": 0.0001953748870743374, "loss": 4.3843, "step": 375400 }, { "epoch": 0.700631445373094, "grad_norm": 1.0004193782806396, "learning_rate": 0.0001953736241562349, "loss": 4.6198, "step": 375450 }, { "epoch": 0.7007247509324724, "grad_norm": 1.302375316619873, "learning_rate": 0.00019537236106981473, "loss": 4.7347, "step": 375500 }, { "epoch": 0.7008180564918509, "grad_norm": 1.1643980741500854, "learning_rate": 0.00019537109781507913, "loss": 4.6438, "step": 375550 }, { "epoch": 0.7009113620512294, "grad_norm": 0.9268198013305664, "learning_rate": 0.0001953698343920303, "loss": 4.5663, "step": 375600 }, { "epoch": 0.7010046676106079, "grad_norm": 0.8678167462348938, "learning_rate": 0.0001953685708006705, "loss": 4.6092, "step": 375650 }, { "epoch": 0.7010979731699865, "grad_norm": 0.8486493825912476, "learning_rate": 0.00019536730704100195, "loss": 4.6546, "step": 375700 }, { "epoch": 0.7011912787293649, "grad_norm": 1.1792452335357666, "learning_rate": 0.00019536604311302691, "loss": 4.6587, "step": 375750 }, { "epoch": 0.7012845842887434, "grad_norm": 1.2924436330795288, "learning_rate": 0.00019536477901674758, "loss": 4.7177, "step": 375800 }, { "epoch": 0.7013778898481219, "grad_norm": 1.1593018770217896, "learning_rate": 0.00019536351475216618, "loss": 4.3727, "step": 375850 }, { "epoch": 0.7014711954075004, "grad_norm": 1.359183430671692, "learning_rate": 0.00019536225031928494, "loss": 4.6466, "step": 375900 }, { "epoch": 0.7015645009668788, "grad_norm": 1.0704420804977417, "learning_rate": 0.00019536098571810615, "loss": 4.5263, "step": 375950 }, { "epoch": 0.7016578065262573, "grad_norm": 0.9053759574890137, "learning_rate": 0.00019535972094863197, "loss": 4.4246, "step": 376000 }, { "epoch": 0.7017511120856359, "grad_norm": 0.872847855091095, "learning_rate": 0.00019535845601086464, "loss": 4.8251, "step": 376050 }, { "epoch": 0.7018444176450144, "grad_norm": 1.0933473110198975, "learning_rate": 0.00019535719090480646, "loss": 4.6556, "step": 376100 }, { "epoch": 0.7019377232043928, "grad_norm": 1.1355681419372559, "learning_rate": 0.00019535592563045963, "loss": 4.3005, "step": 376150 }, { "epoch": 0.7020310287637713, "grad_norm": 1.106989860534668, "learning_rate": 0.00019535466018782632, "loss": 4.4891, "step": 376200 }, { "epoch": 0.7021243343231498, "grad_norm": 1.1264011859893799, "learning_rate": 0.00019535339457690882, "loss": 4.5817, "step": 376250 }, { "epoch": 0.7022176398825283, "grad_norm": 0.9267123937606812, "learning_rate": 0.00019535212879770938, "loss": 4.457, "step": 376300 }, { "epoch": 0.7023109454419068, "grad_norm": 1.2155795097351074, "learning_rate": 0.00019535086285023018, "loss": 4.5067, "step": 376350 }, { "epoch": 0.7024042510012852, "grad_norm": 1.0612845420837402, "learning_rate": 0.0001953495967344735, "loss": 4.5759, "step": 376400 }, { "epoch": 0.7024975565606638, "grad_norm": 1.1359730958938599, "learning_rate": 0.0001953483304504416, "loss": 4.5738, "step": 376450 }, { "epoch": 0.7025908621200423, "grad_norm": 1.0387494564056396, "learning_rate": 0.0001953470639981366, "loss": 4.7473, "step": 376500 }, { "epoch": 0.7026841676794208, "grad_norm": 1.2349492311477661, "learning_rate": 0.00019534579737756085, "loss": 4.6316, "step": 376550 }, { "epoch": 0.7027774732387992, "grad_norm": 1.0554500818252563, "learning_rate": 0.00019534453058871654, "loss": 4.4325, "step": 376600 }, { "epoch": 0.7028707787981777, "grad_norm": 0.9486688375473022, "learning_rate": 0.00019534326363160592, "loss": 4.5899, "step": 376650 }, { "epoch": 0.7029640843575562, "grad_norm": 0.8850482106208801, "learning_rate": 0.00019534199650623118, "loss": 4.7084, "step": 376700 }, { "epoch": 0.7030573899169347, "grad_norm": 0.7264657020568848, "learning_rate": 0.00019534072921259464, "loss": 4.4336, "step": 376750 }, { "epoch": 0.7031506954763133, "grad_norm": 1.0185405015945435, "learning_rate": 0.00019533946175069846, "loss": 4.4325, "step": 376800 }, { "epoch": 0.7032440010356917, "grad_norm": 1.1543070077896118, "learning_rate": 0.0001953381941205449, "loss": 4.6538, "step": 376850 }, { "epoch": 0.7033373065950702, "grad_norm": 1.1681410074234009, "learning_rate": 0.0001953369263221362, "loss": 4.5526, "step": 376900 }, { "epoch": 0.7034306121544487, "grad_norm": 0.8196843862533569, "learning_rate": 0.00019533565835547461, "loss": 4.601, "step": 376950 }, { "epoch": 0.7035239177138272, "grad_norm": 1.115649700164795, "learning_rate": 0.0001953343902205624, "loss": 4.4583, "step": 377000 }, { "epoch": 0.7036172232732056, "grad_norm": 1.201120376586914, "learning_rate": 0.00019533312191740172, "loss": 4.4975, "step": 377050 }, { "epoch": 0.7037105288325841, "grad_norm": 0.9064095616340637, "learning_rate": 0.00019533185344599484, "loss": 4.578, "step": 377100 }, { "epoch": 0.7038038343919627, "grad_norm": 1.2267018556594849, "learning_rate": 0.00019533058480634405, "loss": 4.5731, "step": 377150 }, { "epoch": 0.7038971399513412, "grad_norm": 1.2399940490722656, "learning_rate": 0.00019532931599845155, "loss": 4.6491, "step": 377200 }, { "epoch": 0.7039904455107197, "grad_norm": 0.991210401058197, "learning_rate": 0.00019532804702231956, "loss": 4.5972, "step": 377250 }, { "epoch": 0.7040837510700981, "grad_norm": 0.9711277484893799, "learning_rate": 0.00019532677787795035, "loss": 4.5564, "step": 377300 }, { "epoch": 0.7041770566294766, "grad_norm": 1.2676489353179932, "learning_rate": 0.00019532550856534613, "loss": 4.4359, "step": 377350 }, { "epoch": 0.7042703621888551, "grad_norm": 1.5608159303665161, "learning_rate": 0.0001953242390845092, "loss": 4.4103, "step": 377400 }, { "epoch": 0.7043636677482336, "grad_norm": 0.9402786493301392, "learning_rate": 0.00019532296943544172, "loss": 4.4642, "step": 377450 }, { "epoch": 0.7044569733076121, "grad_norm": 0.8223136067390442, "learning_rate": 0.000195321699618146, "loss": 4.4978, "step": 377500 }, { "epoch": 0.7045502788669906, "grad_norm": 0.744204044342041, "learning_rate": 0.00019532042963262423, "loss": 4.419, "step": 377550 }, { "epoch": 0.7046435844263691, "grad_norm": 1.0643330812454224, "learning_rate": 0.00019531915947887868, "loss": 4.5853, "step": 377600 }, { "epoch": 0.7047368899857476, "grad_norm": 1.1969504356384277, "learning_rate": 0.0001953178891569116, "loss": 4.6716, "step": 377650 }, { "epoch": 0.704830195545126, "grad_norm": 0.9131072759628296, "learning_rate": 0.0001953166186667252, "loss": 4.5512, "step": 377700 }, { "epoch": 0.7049235011045045, "grad_norm": 1.1578060388565063, "learning_rate": 0.00019531534800832175, "loss": 4.5252, "step": 377750 }, { "epoch": 0.705016806663883, "grad_norm": 1.3309463262557983, "learning_rate": 0.0001953140771817035, "loss": 4.8214, "step": 377800 }, { "epoch": 0.7051101122232616, "grad_norm": 1.3318638801574707, "learning_rate": 0.00019531280618687266, "loss": 4.3906, "step": 377850 }, { "epoch": 0.7052034177826401, "grad_norm": 0.9537114500999451, "learning_rate": 0.00019531153502383149, "loss": 4.6278, "step": 377900 }, { "epoch": 0.7052967233420185, "grad_norm": 1.287827968597412, "learning_rate": 0.0001953102636925822, "loss": 4.4588, "step": 377950 }, { "epoch": 0.705390028901397, "grad_norm": 1.0049210786819458, "learning_rate": 0.0001953089921931271, "loss": 4.3171, "step": 378000 }, { "epoch": 0.705390028901397, "eval_loss": 4.741976261138916, "eval_runtime": 229.2276, "eval_samples_per_second": 11.377, "eval_steps_per_second": 11.377, "eval_tts_loss": 7.590075247879626, "step": 378000 }, { "epoch": 0.7054833344607755, "grad_norm": 0.9821569919586182, "learning_rate": 0.0001953077205254684, "loss": 4.7729, "step": 378050 }, { "epoch": 0.705576640020154, "grad_norm": 1.08197820186615, "learning_rate": 0.00019530644868960833, "loss": 4.5847, "step": 378100 }, { "epoch": 0.7056699455795324, "grad_norm": 0.9573339223861694, "learning_rate": 0.00019530517668554914, "loss": 4.653, "step": 378150 }, { "epoch": 0.705763251138911, "grad_norm": 1.0623559951782227, "learning_rate": 0.0001953039045132931, "loss": 4.5052, "step": 378200 }, { "epoch": 0.7058565566982895, "grad_norm": 0.9007604718208313, "learning_rate": 0.00019530263217284243, "loss": 4.545, "step": 378250 }, { "epoch": 0.705949862257668, "grad_norm": 1.1949926614761353, "learning_rate": 0.00019530135966419938, "loss": 4.5462, "step": 378300 }, { "epoch": 0.7060431678170465, "grad_norm": 1.0200376510620117, "learning_rate": 0.00019530008698736618, "loss": 4.5997, "step": 378350 }, { "epoch": 0.7061364733764249, "grad_norm": 1.2400939464569092, "learning_rate": 0.00019529881414234515, "loss": 4.5039, "step": 378400 }, { "epoch": 0.7062297789358034, "grad_norm": 0.8792778253555298, "learning_rate": 0.00019529754112913844, "loss": 4.7137, "step": 378450 }, { "epoch": 0.7063230844951819, "grad_norm": 1.0362474918365479, "learning_rate": 0.00019529626794774835, "loss": 4.5742, "step": 378500 }, { "epoch": 0.7064163900545605, "grad_norm": 0.8532072901725769, "learning_rate": 0.0001952949945981771, "loss": 4.6874, "step": 378550 }, { "epoch": 0.706509695613939, "grad_norm": 0.7794941663742065, "learning_rate": 0.00019529372108042695, "loss": 4.3726, "step": 378600 }, { "epoch": 0.7066030011733174, "grad_norm": 1.080212116241455, "learning_rate": 0.00019529244739450018, "loss": 4.4039, "step": 378650 }, { "epoch": 0.7066963067326959, "grad_norm": 0.8216705918312073, "learning_rate": 0.000195291173540399, "loss": 4.7282, "step": 378700 }, { "epoch": 0.7067896122920744, "grad_norm": 1.154890537261963, "learning_rate": 0.00019528989951812564, "loss": 4.4009, "step": 378750 }, { "epoch": 0.7068829178514529, "grad_norm": 1.1282939910888672, "learning_rate": 0.0001952886253276824, "loss": 4.5085, "step": 378800 }, { "epoch": 0.7069762234108313, "grad_norm": 0.9745324850082397, "learning_rate": 0.0001952873509690715, "loss": 4.6955, "step": 378850 }, { "epoch": 0.7070695289702099, "grad_norm": 0.9145590662956238, "learning_rate": 0.0001952860764422952, "loss": 4.5594, "step": 378900 }, { "epoch": 0.7071628345295884, "grad_norm": 0.9380348324775696, "learning_rate": 0.00019528480174735573, "loss": 4.4592, "step": 378950 }, { "epoch": 0.7072561400889669, "grad_norm": 0.9864219427108765, "learning_rate": 0.00019528352688425533, "loss": 4.5247, "step": 379000 }, { "epoch": 0.7073494456483453, "grad_norm": 1.1214337348937988, "learning_rate": 0.0001952822518529963, "loss": 4.5793, "step": 379050 }, { "epoch": 0.7074427512077238, "grad_norm": 0.8948968648910522, "learning_rate": 0.00019528097665358086, "loss": 4.412, "step": 379100 }, { "epoch": 0.7075360567671023, "grad_norm": 0.9934725165367126, "learning_rate": 0.00019527970128601122, "loss": 4.6638, "step": 379150 }, { "epoch": 0.7076293623264808, "grad_norm": 0.9161415100097656, "learning_rate": 0.00019527842575028971, "loss": 4.6051, "step": 379200 }, { "epoch": 0.7077226678858594, "grad_norm": 1.0251588821411133, "learning_rate": 0.00019527715004641856, "loss": 4.3529, "step": 379250 }, { "epoch": 0.7078159734452378, "grad_norm": 1.0846620798110962, "learning_rate": 0.00019527587417439996, "loss": 4.5257, "step": 379300 }, { "epoch": 0.7079092790046163, "grad_norm": 1.136697769165039, "learning_rate": 0.00019527459813423623, "loss": 4.4867, "step": 379350 }, { "epoch": 0.7080025845639948, "grad_norm": 1.052685022354126, "learning_rate": 0.0001952733219259296, "loss": 4.6143, "step": 379400 }, { "epoch": 0.7080958901233733, "grad_norm": 0.9956856966018677, "learning_rate": 0.00019527204554948231, "loss": 4.5272, "step": 379450 }, { "epoch": 0.7081891956827517, "grad_norm": 1.0275317430496216, "learning_rate": 0.00019527076900489662, "loss": 4.7828, "step": 379500 }, { "epoch": 0.7082825012421302, "grad_norm": 0.8791360855102539, "learning_rate": 0.00019526949229217483, "loss": 4.5505, "step": 379550 }, { "epoch": 0.7083758068015088, "grad_norm": 0.9856266975402832, "learning_rate": 0.0001952682154113191, "loss": 4.7794, "step": 379600 }, { "epoch": 0.7084691123608873, "grad_norm": 0.8814522624015808, "learning_rate": 0.00019526693836233175, "loss": 4.5435, "step": 379650 }, { "epoch": 0.7085624179202658, "grad_norm": 0.9303852319717407, "learning_rate": 0.000195265661145215, "loss": 4.6448, "step": 379700 }, { "epoch": 0.7086557234796442, "grad_norm": 0.9985621571540833, "learning_rate": 0.00019526438375997115, "loss": 4.6232, "step": 379750 }, { "epoch": 0.7087490290390227, "grad_norm": 1.2800402641296387, "learning_rate": 0.00019526310620660242, "loss": 4.6542, "step": 379800 }, { "epoch": 0.7088423345984012, "grad_norm": 1.2168068885803223, "learning_rate": 0.00019526182848511104, "loss": 4.6047, "step": 379850 }, { "epoch": 0.7089356401577797, "grad_norm": 0.9573338627815247, "learning_rate": 0.0001952605505954993, "loss": 4.4232, "step": 379900 }, { "epoch": 0.7090289457171582, "grad_norm": 0.991240918636322, "learning_rate": 0.0001952592725377695, "loss": 4.5541, "step": 379950 }, { "epoch": 0.7091222512765367, "grad_norm": 1.210038185119629, "learning_rate": 0.0001952579943119238, "loss": 4.6207, "step": 380000 }, { "epoch": 0.7092155568359152, "grad_norm": 1.135339379310608, "learning_rate": 0.0001952567159179645, "loss": 4.5715, "step": 380050 }, { "epoch": 0.7093088623952937, "grad_norm": 1.2163465023040771, "learning_rate": 0.00019525543735589388, "loss": 4.5323, "step": 380100 }, { "epoch": 0.7094021679546721, "grad_norm": 1.2491923570632935, "learning_rate": 0.00019525415862571416, "loss": 4.6131, "step": 380150 }, { "epoch": 0.7094954735140506, "grad_norm": 0.6898648738861084, "learning_rate": 0.0001952528797274276, "loss": 4.6363, "step": 380200 }, { "epoch": 0.7095887790734291, "grad_norm": 0.7697023153305054, "learning_rate": 0.0001952516006610365, "loss": 4.4572, "step": 380250 }, { "epoch": 0.7096820846328077, "grad_norm": 0.7443915605545044, "learning_rate": 0.00019525032142654304, "loss": 4.5029, "step": 380300 }, { "epoch": 0.7097753901921862, "grad_norm": 1.4766536951065063, "learning_rate": 0.00019524904202394955, "loss": 4.5291, "step": 380350 }, { "epoch": 0.7098686957515646, "grad_norm": 0.9718853831291199, "learning_rate": 0.00019524776245325825, "loss": 4.6251, "step": 380400 }, { "epoch": 0.7099620013109431, "grad_norm": 1.1660414934158325, "learning_rate": 0.00019524648271447144, "loss": 4.5658, "step": 380450 }, { "epoch": 0.7100553068703216, "grad_norm": 1.0906556844711304, "learning_rate": 0.00019524520280759134, "loss": 4.4738, "step": 380500 }, { "epoch": 0.7101486124297001, "grad_norm": 1.1851836442947388, "learning_rate": 0.00019524392273262018, "loss": 4.5588, "step": 380550 }, { "epoch": 0.7102419179890785, "grad_norm": 1.3385717868804932, "learning_rate": 0.0001952426424895603, "loss": 4.4783, "step": 380600 }, { "epoch": 0.7103352235484571, "grad_norm": 1.1230698823928833, "learning_rate": 0.00019524136207841385, "loss": 4.7116, "step": 380650 }, { "epoch": 0.7104285291078356, "grad_norm": 1.1039788722991943, "learning_rate": 0.0001952400814991832, "loss": 4.5662, "step": 380700 }, { "epoch": 0.7105218346672141, "grad_norm": 0.9813905358314514, "learning_rate": 0.00019523880075187058, "loss": 4.612, "step": 380750 }, { "epoch": 0.7106151402265926, "grad_norm": 1.1137398481369019, "learning_rate": 0.0001952375198364782, "loss": 4.5109, "step": 380800 }, { "epoch": 0.710708445785971, "grad_norm": 1.0999574661254883, "learning_rate": 0.00019523623875300836, "loss": 4.4076, "step": 380850 }, { "epoch": 0.7108017513453495, "grad_norm": 1.1628655195236206, "learning_rate": 0.0001952349575014633, "loss": 4.6405, "step": 380900 }, { "epoch": 0.710895056904728, "grad_norm": 1.07998526096344, "learning_rate": 0.00019523367608184535, "loss": 4.588, "step": 380950 }, { "epoch": 0.7109883624641066, "grad_norm": 1.0283795595169067, "learning_rate": 0.00019523239449415668, "loss": 4.6904, "step": 381000 }, { "epoch": 0.7109883624641066, "eval_loss": 4.738658905029297, "eval_runtime": 230.5123, "eval_samples_per_second": 11.314, "eval_steps_per_second": 11.314, "eval_tts_loss": 7.550678487040068, "step": 381000 }, { "epoch": 0.711081668023485, "grad_norm": 1.0346516370773315, "learning_rate": 0.0001952311127383996, "loss": 4.3708, "step": 381050 }, { "epoch": 0.7111749735828635, "grad_norm": 1.0621981620788574, "learning_rate": 0.00019522983081457634, "loss": 4.642, "step": 381100 }, { "epoch": 0.711268279142242, "grad_norm": 1.1409293413162231, "learning_rate": 0.00019522854872268923, "loss": 4.4803, "step": 381150 }, { "epoch": 0.7113615847016205, "grad_norm": 0.9242346286773682, "learning_rate": 0.00019522726646274046, "loss": 4.4169, "step": 381200 }, { "epoch": 0.711454890260999, "grad_norm": 1.2605868577957153, "learning_rate": 0.00019522598403473232, "loss": 4.7021, "step": 381250 }, { "epoch": 0.7115481958203774, "grad_norm": 1.2490514516830444, "learning_rate": 0.00019522470143866708, "loss": 4.6978, "step": 381300 }, { "epoch": 0.711641501379756, "grad_norm": 1.114847183227539, "learning_rate": 0.000195223418674547, "loss": 4.5004, "step": 381350 }, { "epoch": 0.7117348069391345, "grad_norm": 1.267319679260254, "learning_rate": 0.00019522213574237434, "loss": 4.6773, "step": 381400 }, { "epoch": 0.711828112498513, "grad_norm": 1.009610652923584, "learning_rate": 0.00019522085264215135, "loss": 4.4301, "step": 381450 }, { "epoch": 0.7119214180578914, "grad_norm": 0.7319119572639465, "learning_rate": 0.00019521956937388032, "loss": 4.5613, "step": 381500 }, { "epoch": 0.7120147236172699, "grad_norm": 1.0810520648956299, "learning_rate": 0.00019521828593756352, "loss": 4.5926, "step": 381550 }, { "epoch": 0.7121080291766484, "grad_norm": 1.1929056644439697, "learning_rate": 0.00019521700233320318, "loss": 4.623, "step": 381600 }, { "epoch": 0.7122013347360269, "grad_norm": 1.0149575471878052, "learning_rate": 0.0001952157185608016, "loss": 4.4202, "step": 381650 }, { "epoch": 0.7122946402954053, "grad_norm": 0.918735146522522, "learning_rate": 0.000195214434620361, "loss": 4.6246, "step": 381700 }, { "epoch": 0.7123879458547839, "grad_norm": 1.206817388534546, "learning_rate": 0.0001952131505118837, "loss": 4.5127, "step": 381750 }, { "epoch": 0.7124812514141624, "grad_norm": 1.0245983600616455, "learning_rate": 0.00019521186623537196, "loss": 4.6317, "step": 381800 }, { "epoch": 0.7125745569735409, "grad_norm": 1.0655715465545654, "learning_rate": 0.000195210581790828, "loss": 4.3029, "step": 381850 }, { "epoch": 0.7126678625329194, "grad_norm": 1.1398463249206543, "learning_rate": 0.00019520929717825413, "loss": 4.5782, "step": 381900 }, { "epoch": 0.7127611680922978, "grad_norm": 1.129570484161377, "learning_rate": 0.00019520801239765258, "loss": 4.6278, "step": 381950 }, { "epoch": 0.7128544736516763, "grad_norm": 1.1474553346633911, "learning_rate": 0.00019520672744902567, "loss": 4.4637, "step": 382000 }, { "epoch": 0.7129477792110548, "grad_norm": 1.088990330696106, "learning_rate": 0.00019520544233237561, "loss": 4.4453, "step": 382050 }, { "epoch": 0.7130410847704334, "grad_norm": 0.9756652116775513, "learning_rate": 0.00019520415704770474, "loss": 4.4904, "step": 382100 }, { "epoch": 0.7131343903298119, "grad_norm": 1.2050561904907227, "learning_rate": 0.00019520287159501524, "loss": 4.6689, "step": 382150 }, { "epoch": 0.7132276958891903, "grad_norm": 1.0057941675186157, "learning_rate": 0.00019520158597430943, "loss": 4.5429, "step": 382200 }, { "epoch": 0.7133210014485688, "grad_norm": 0.9556266665458679, "learning_rate": 0.0001952003001855896, "loss": 4.4695, "step": 382250 }, { "epoch": 0.7134143070079473, "grad_norm": 0.8169947862625122, "learning_rate": 0.00019519901422885796, "loss": 4.6116, "step": 382300 }, { "epoch": 0.7135076125673258, "grad_norm": 1.0497891902923584, "learning_rate": 0.00019519772810411683, "loss": 4.5701, "step": 382350 }, { "epoch": 0.7136009181267042, "grad_norm": 0.8632709383964539, "learning_rate": 0.00019519644181136843, "loss": 4.47, "step": 382400 }, { "epoch": 0.7136942236860828, "grad_norm": 1.1212148666381836, "learning_rate": 0.0001951951553506151, "loss": 4.4381, "step": 382450 }, { "epoch": 0.7137875292454613, "grad_norm": 1.2033298015594482, "learning_rate": 0.00019519386872185907, "loss": 4.9272, "step": 382500 }, { "epoch": 0.7138808348048398, "grad_norm": 1.0442355871200562, "learning_rate": 0.0001951925819251026, "loss": 4.5303, "step": 382550 }, { "epoch": 0.7139741403642182, "grad_norm": 1.5087108612060547, "learning_rate": 0.00019519129496034797, "loss": 4.59, "step": 382600 }, { "epoch": 0.7140674459235967, "grad_norm": 1.2250038385391235, "learning_rate": 0.00019519000782759744, "loss": 4.5591, "step": 382650 }, { "epoch": 0.7141607514829752, "grad_norm": 1.4251493215560913, "learning_rate": 0.0001951887205268533, "loss": 4.4625, "step": 382700 }, { "epoch": 0.7142540570423537, "grad_norm": 1.1542210578918457, "learning_rate": 0.00019518743305811784, "loss": 4.4542, "step": 382750 }, { "epoch": 0.7143473626017323, "grad_norm": 1.1355654001235962, "learning_rate": 0.0001951861454213933, "loss": 4.3306, "step": 382800 }, { "epoch": 0.7144406681611107, "grad_norm": 0.8172572255134583, "learning_rate": 0.00019518485761668198, "loss": 4.6956, "step": 382850 }, { "epoch": 0.7145339737204892, "grad_norm": 1.0452672243118286, "learning_rate": 0.0001951835696439861, "loss": 4.577, "step": 382900 }, { "epoch": 0.7146272792798677, "grad_norm": 1.2720973491668701, "learning_rate": 0.00019518228150330798, "loss": 4.5383, "step": 382950 }, { "epoch": 0.7147205848392462, "grad_norm": 0.8681235313415527, "learning_rate": 0.00019518099319464985, "loss": 4.5858, "step": 383000 }, { "epoch": 0.7148138903986246, "grad_norm": 1.3290435075759888, "learning_rate": 0.00019517970471801408, "loss": 4.8388, "step": 383050 }, { "epoch": 0.7149071959580031, "grad_norm": 0.9616467356681824, "learning_rate": 0.00019517841607340284, "loss": 4.4234, "step": 383100 }, { "epoch": 0.7150005015173817, "grad_norm": 1.1521810293197632, "learning_rate": 0.00019517712726081844, "loss": 4.5619, "step": 383150 }, { "epoch": 0.7150938070767602, "grad_norm": 0.8024904131889343, "learning_rate": 0.00019517583828026314, "loss": 4.7557, "step": 383200 }, { "epoch": 0.7151871126361387, "grad_norm": 1.0209003686904907, "learning_rate": 0.00019517454913173927, "loss": 4.6532, "step": 383250 }, { "epoch": 0.7152804181955171, "grad_norm": 1.0985634326934814, "learning_rate": 0.00019517325981524905, "loss": 4.4483, "step": 383300 }, { "epoch": 0.7153737237548956, "grad_norm": 1.0770183801651, "learning_rate": 0.00019517197033079478, "loss": 4.7606, "step": 383350 }, { "epoch": 0.7154670293142741, "grad_norm": 0.8601162433624268, "learning_rate": 0.00019517068067837872, "loss": 4.439, "step": 383400 }, { "epoch": 0.7155603348736526, "grad_norm": 1.282110571861267, "learning_rate": 0.00019516939085800316, "loss": 4.4998, "step": 383450 }, { "epoch": 0.7156536404330311, "grad_norm": 1.1574286222457886, "learning_rate": 0.00019516810086967038, "loss": 4.5013, "step": 383500 }, { "epoch": 0.7157469459924096, "grad_norm": 1.2425411939620972, "learning_rate": 0.00019516681071338264, "loss": 4.6996, "step": 383550 }, { "epoch": 0.7158402515517881, "grad_norm": 1.076066255569458, "learning_rate": 0.00019516552038914222, "loss": 4.6142, "step": 383600 }, { "epoch": 0.7159335571111666, "grad_norm": 1.0551059246063232, "learning_rate": 0.0001951642298969514, "loss": 4.7165, "step": 383650 }, { "epoch": 0.716026862670545, "grad_norm": 1.227857232093811, "learning_rate": 0.00019516293923681246, "loss": 4.5505, "step": 383700 }, { "epoch": 0.7161201682299235, "grad_norm": 1.0538171529769897, "learning_rate": 0.00019516164840872768, "loss": 4.5371, "step": 383750 }, { "epoch": 0.716213473789302, "grad_norm": 0.9511804580688477, "learning_rate": 0.00019516035741269934, "loss": 4.5966, "step": 383800 }, { "epoch": 0.7163067793486806, "grad_norm": 0.7090573310852051, "learning_rate": 0.0001951590662487297, "loss": 4.6471, "step": 383850 }, { "epoch": 0.7164000849080591, "grad_norm": 1.0298354625701904, "learning_rate": 0.00019515777491682106, "loss": 4.5427, "step": 383900 }, { "epoch": 0.7164933904674375, "grad_norm": 1.163970947265625, "learning_rate": 0.00019515648341697568, "loss": 4.4544, "step": 383950 }, { "epoch": 0.716586696026816, "grad_norm": 1.0745140314102173, "learning_rate": 0.00019515519174919586, "loss": 4.3617, "step": 384000 }, { "epoch": 0.716586696026816, "eval_loss": 4.7351579666137695, "eval_runtime": 229.8886, "eval_samples_per_second": 11.345, "eval_steps_per_second": 11.345, "eval_tts_loss": 7.557877517531003, "step": 384000 }, { "epoch": 0.7166800015861945, "grad_norm": 0.9457404613494873, "learning_rate": 0.00019515389991348387, "loss": 4.3421, "step": 384050 }, { "epoch": 0.716773307145573, "grad_norm": 1.0447351932525635, "learning_rate": 0.000195152607909842, "loss": 4.5032, "step": 384100 }, { "epoch": 0.7168666127049514, "grad_norm": 1.032181978225708, "learning_rate": 0.0001951513157382725, "loss": 4.7111, "step": 384150 }, { "epoch": 0.71695991826433, "grad_norm": 0.8992804884910583, "learning_rate": 0.00019515002339877765, "loss": 4.5748, "step": 384200 }, { "epoch": 0.7170532238237085, "grad_norm": 1.2443760633468628, "learning_rate": 0.0001951487308913598, "loss": 4.5732, "step": 384250 }, { "epoch": 0.717146529383087, "grad_norm": 1.1091217994689941, "learning_rate": 0.00019514743821602118, "loss": 4.7897, "step": 384300 }, { "epoch": 0.7172398349424655, "grad_norm": 1.1647604703903198, "learning_rate": 0.00019514614537276403, "loss": 4.3145, "step": 384350 }, { "epoch": 0.7173331405018439, "grad_norm": 0.9686366319656372, "learning_rate": 0.0001951448523615907, "loss": 4.5574, "step": 384400 }, { "epoch": 0.7174264460612224, "grad_norm": 1.0548909902572632, "learning_rate": 0.00019514355918250344, "loss": 4.7254, "step": 384450 }, { "epoch": 0.7175197516206009, "grad_norm": 1.2261728048324585, "learning_rate": 0.00019514226583550458, "loss": 4.524, "step": 384500 }, { "epoch": 0.7176130571799795, "grad_norm": 1.1724244356155396, "learning_rate": 0.0001951409723205963, "loss": 4.6381, "step": 384550 }, { "epoch": 0.717706362739358, "grad_norm": 1.0948948860168457, "learning_rate": 0.00019513967863778097, "loss": 4.6415, "step": 384600 }, { "epoch": 0.7177996682987364, "grad_norm": 1.1220853328704834, "learning_rate": 0.00019513838478706086, "loss": 4.4522, "step": 384650 }, { "epoch": 0.7178929738581149, "grad_norm": 0.9094553589820862, "learning_rate": 0.00019513709076843822, "loss": 4.7201, "step": 384700 }, { "epoch": 0.7179862794174934, "grad_norm": 1.1259245872497559, "learning_rate": 0.00019513579658191537, "loss": 4.4064, "step": 384750 }, { "epoch": 0.7180795849768719, "grad_norm": 1.1270395517349243, "learning_rate": 0.00019513450222749456, "loss": 4.5983, "step": 384800 }, { "epoch": 0.7181728905362503, "grad_norm": 1.206202507019043, "learning_rate": 0.00019513320770517812, "loss": 4.5298, "step": 384850 }, { "epoch": 0.7182661960956289, "grad_norm": 1.294735312461853, "learning_rate": 0.0001951319130149683, "loss": 4.6011, "step": 384900 }, { "epoch": 0.7183595016550074, "grad_norm": 1.2166738510131836, "learning_rate": 0.00019513061815686737, "loss": 4.7282, "step": 384950 }, { "epoch": 0.7184528072143859, "grad_norm": 1.1213105916976929, "learning_rate": 0.00019512932313087765, "loss": 4.4658, "step": 385000 }, { "epoch": 0.7185461127737643, "grad_norm": 1.0961776971817017, "learning_rate": 0.00019512802793700143, "loss": 4.4198, "step": 385050 }, { "epoch": 0.7186394183331428, "grad_norm": 1.1053520441055298, "learning_rate": 0.00019512673257524098, "loss": 4.6368, "step": 385100 }, { "epoch": 0.7187327238925213, "grad_norm": 1.132785677909851, "learning_rate": 0.00019512543704559856, "loss": 4.6449, "step": 385150 }, { "epoch": 0.7188260294518998, "grad_norm": 0.7992883920669556, "learning_rate": 0.0001951241413480765, "loss": 4.5777, "step": 385200 }, { "epoch": 0.7189193350112784, "grad_norm": 1.2569609880447388, "learning_rate": 0.00019512284548267705, "loss": 4.7781, "step": 385250 }, { "epoch": 0.7190126405706568, "grad_norm": 0.7372807264328003, "learning_rate": 0.00019512154944940254, "loss": 4.6389, "step": 385300 }, { "epoch": 0.7191059461300353, "grad_norm": 0.9772287011146545, "learning_rate": 0.00019512025324825517, "loss": 4.5089, "step": 385350 }, { "epoch": 0.7191992516894138, "grad_norm": 1.157492995262146, "learning_rate": 0.00019511895687923737, "loss": 4.6116, "step": 385400 }, { "epoch": 0.7192925572487923, "grad_norm": 1.1132258176803589, "learning_rate": 0.0001951176603423513, "loss": 4.305, "step": 385450 }, { "epoch": 0.7193858628081707, "grad_norm": 1.1786819696426392, "learning_rate": 0.00019511636363759931, "loss": 4.6685, "step": 385500 }, { "epoch": 0.7194791683675492, "grad_norm": 1.2146203517913818, "learning_rate": 0.00019511506676498367, "loss": 4.3969, "step": 385550 }, { "epoch": 0.7195724739269278, "grad_norm": 1.0019917488098145, "learning_rate": 0.00019511376972450667, "loss": 4.7664, "step": 385600 }, { "epoch": 0.7196657794863063, "grad_norm": 1.107676386833191, "learning_rate": 0.00019511247251617064, "loss": 4.5236, "step": 385650 }, { "epoch": 0.7197590850456848, "grad_norm": 1.2492763996124268, "learning_rate": 0.0001951111751399778, "loss": 4.787, "step": 385700 }, { "epoch": 0.7198523906050632, "grad_norm": 0.9558393955230713, "learning_rate": 0.00019510987759593047, "loss": 4.4466, "step": 385750 }, { "epoch": 0.7199456961644417, "grad_norm": 1.5168671607971191, "learning_rate": 0.00019510857988403095, "loss": 4.6665, "step": 385800 }, { "epoch": 0.7200390017238202, "grad_norm": 1.0870764255523682, "learning_rate": 0.0001951072820042815, "loss": 4.4197, "step": 385850 }, { "epoch": 0.7201323072831987, "grad_norm": 1.1981571912765503, "learning_rate": 0.00019510598395668444, "loss": 4.5403, "step": 385900 }, { "epoch": 0.7202256128425772, "grad_norm": 1.0989803075790405, "learning_rate": 0.00019510468574124206, "loss": 4.6166, "step": 385950 }, { "epoch": 0.7203189184019557, "grad_norm": 1.1044586896896362, "learning_rate": 0.00019510338735795668, "loss": 4.6017, "step": 386000 }, { "epoch": 0.7204122239613342, "grad_norm": 1.1001883745193481, "learning_rate": 0.0001951020888068305, "loss": 4.6469, "step": 386050 }, { "epoch": 0.7205055295207127, "grad_norm": 1.145124077796936, "learning_rate": 0.0001951007900878659, "loss": 4.4491, "step": 386100 }, { "epoch": 0.7205988350800911, "grad_norm": 1.18637216091156, "learning_rate": 0.00019509949120106513, "loss": 4.4879, "step": 386150 }, { "epoch": 0.7206921406394696, "grad_norm": 0.5531867146492004, "learning_rate": 0.0001950981921464305, "loss": 4.6821, "step": 386200 }, { "epoch": 0.7207854461988481, "grad_norm": 1.187780737876892, "learning_rate": 0.00019509689292396428, "loss": 4.6521, "step": 386250 }, { "epoch": 0.7208787517582267, "grad_norm": 0.8865212798118591, "learning_rate": 0.00019509559353366878, "loss": 4.5438, "step": 386300 }, { "epoch": 0.7209720573176052, "grad_norm": 0.9871026277542114, "learning_rate": 0.00019509429397554628, "loss": 4.5224, "step": 386350 }, { "epoch": 0.7210653628769836, "grad_norm": 1.2513176202774048, "learning_rate": 0.00019509299424959908, "loss": 4.4853, "step": 386400 }, { "epoch": 0.7211586684363621, "grad_norm": 1.108982801437378, "learning_rate": 0.0001950916943558295, "loss": 4.7211, "step": 386450 }, { "epoch": 0.7212519739957406, "grad_norm": 1.206200361251831, "learning_rate": 0.00019509039429423979, "loss": 4.6938, "step": 386500 }, { "epoch": 0.7213452795551191, "grad_norm": 0.9535238742828369, "learning_rate": 0.00019508909406483227, "loss": 4.5804, "step": 386550 }, { "epoch": 0.7214385851144975, "grad_norm": 1.1816784143447876, "learning_rate": 0.00019508779366760926, "loss": 4.5578, "step": 386600 }, { "epoch": 0.7215318906738761, "grad_norm": 1.0024590492248535, "learning_rate": 0.00019508649310257298, "loss": 4.613, "step": 386650 }, { "epoch": 0.7216251962332546, "grad_norm": 1.145411729812622, "learning_rate": 0.0001950851923697258, "loss": 4.7277, "step": 386700 }, { "epoch": 0.7217185017926331, "grad_norm": 0.9904378056526184, "learning_rate": 0.00019508389146906998, "loss": 4.4863, "step": 386750 }, { "epoch": 0.7218118073520116, "grad_norm": 0.7934876680374146, "learning_rate": 0.0001950825904006078, "loss": 4.6157, "step": 386800 }, { "epoch": 0.72190511291139, "grad_norm": 0.9897130727767944, "learning_rate": 0.0001950812891643416, "loss": 4.5527, "step": 386850 }, { "epoch": 0.7219984184707685, "grad_norm": 1.23386549949646, "learning_rate": 0.0001950799877602737, "loss": 4.579, "step": 386900 }, { "epoch": 0.722091724030147, "grad_norm": 1.0667505264282227, "learning_rate": 0.00019507868618840629, "loss": 4.708, "step": 386950 }, { "epoch": 0.7221850295895256, "grad_norm": 1.2649896144866943, "learning_rate": 0.00019507738444874174, "loss": 4.4828, "step": 387000 }, { "epoch": 0.7221850295895256, "eval_loss": 4.725461483001709, "eval_runtime": 231.3684, "eval_samples_per_second": 11.272, "eval_steps_per_second": 11.272, "eval_tts_loss": 7.58398832848147, "step": 387000 }, { "epoch": 0.722278335148904, "grad_norm": 0.9351997971534729, "learning_rate": 0.00019507608254128232, "loss": 4.5677, "step": 387050 }, { "epoch": 0.7223716407082825, "grad_norm": 1.1455060243606567, "learning_rate": 0.00019507478046603037, "loss": 4.6401, "step": 387100 }, { "epoch": 0.722464946267661, "grad_norm": 0.8970882296562195, "learning_rate": 0.00019507347822298815, "loss": 4.4225, "step": 387150 }, { "epoch": 0.7225582518270395, "grad_norm": 1.546656847000122, "learning_rate": 0.00019507217581215798, "loss": 4.5408, "step": 387200 }, { "epoch": 0.722651557386418, "grad_norm": 1.1023125648498535, "learning_rate": 0.00019507087323354213, "loss": 4.6823, "step": 387250 }, { "epoch": 0.7227448629457964, "grad_norm": 0.9081626534461975, "learning_rate": 0.00019506957048714293, "loss": 4.3673, "step": 387300 }, { "epoch": 0.7228381685051749, "grad_norm": 1.059166669845581, "learning_rate": 0.00019506826757296266, "loss": 4.6348, "step": 387350 }, { "epoch": 0.7229314740645535, "grad_norm": 1.273818016052246, "learning_rate": 0.00019506696449100364, "loss": 4.6823, "step": 387400 }, { "epoch": 0.723024779623932, "grad_norm": 1.3179587125778198, "learning_rate": 0.00019506566124126814, "loss": 4.7256, "step": 387450 }, { "epoch": 0.7231180851833104, "grad_norm": 1.1236895322799683, "learning_rate": 0.0001950643578237585, "loss": 4.7807, "step": 387500 }, { "epoch": 0.7232113907426889, "grad_norm": 1.0101490020751953, "learning_rate": 0.00019506305423847698, "loss": 4.6532, "step": 387550 }, { "epoch": 0.7233046963020674, "grad_norm": 1.0887134075164795, "learning_rate": 0.00019506175048542587, "loss": 4.5312, "step": 387600 }, { "epoch": 0.7233980018614459, "grad_norm": 0.9442769885063171, "learning_rate": 0.00019506044656460752, "loss": 4.4949, "step": 387650 }, { "epoch": 0.7234913074208243, "grad_norm": 1.0077662467956543, "learning_rate": 0.00019505914247602423, "loss": 4.5552, "step": 387700 }, { "epoch": 0.7235846129802029, "grad_norm": 1.100584626197815, "learning_rate": 0.00019505783821967824, "loss": 4.5031, "step": 387750 }, { "epoch": 0.7236779185395814, "grad_norm": 0.8957000374794006, "learning_rate": 0.0001950565337955719, "loss": 4.5858, "step": 387800 }, { "epoch": 0.7237712240989599, "grad_norm": 1.2696412801742554, "learning_rate": 0.00019505522920370752, "loss": 4.4523, "step": 387850 }, { "epoch": 0.7238645296583384, "grad_norm": 1.0976306200027466, "learning_rate": 0.00019505392444408738, "loss": 4.574, "step": 387900 }, { "epoch": 0.7239578352177168, "grad_norm": 1.000716209411621, "learning_rate": 0.0001950526195167138, "loss": 4.4646, "step": 387950 }, { "epoch": 0.7240511407770953, "grad_norm": 0.972341775894165, "learning_rate": 0.00019505131442158907, "loss": 4.5893, "step": 388000 }, { "epoch": 0.7241444463364738, "grad_norm": 1.0794826745986938, "learning_rate": 0.00019505000915871548, "loss": 4.7191, "step": 388050 }, { "epoch": 0.7242377518958524, "grad_norm": 0.9431530833244324, "learning_rate": 0.00019504870372809535, "loss": 4.718, "step": 388100 }, { "epoch": 0.7243310574552309, "grad_norm": 1.200202465057373, "learning_rate": 0.00019504739812973098, "loss": 4.7026, "step": 388150 }, { "epoch": 0.7244243630146093, "grad_norm": 1.0569508075714111, "learning_rate": 0.00019504609236362469, "loss": 4.4948, "step": 388200 }, { "epoch": 0.7245176685739878, "grad_norm": 0.9381723999977112, "learning_rate": 0.00019504478642977876, "loss": 4.5383, "step": 388250 }, { "epoch": 0.7246109741333663, "grad_norm": 1.137317419052124, "learning_rate": 0.00019504348032819548, "loss": 4.6089, "step": 388300 }, { "epoch": 0.7247042796927448, "grad_norm": 1.215718388557434, "learning_rate": 0.0001950421740588772, "loss": 4.7085, "step": 388350 }, { "epoch": 0.7247975852521232, "grad_norm": 1.2167165279388428, "learning_rate": 0.00019504086762182622, "loss": 4.6836, "step": 388400 }, { "epoch": 0.7248908908115018, "grad_norm": 1.219132423400879, "learning_rate": 0.00019503956101704485, "loss": 4.5349, "step": 388450 }, { "epoch": 0.7249841963708803, "grad_norm": 0.8871475458145142, "learning_rate": 0.00019503825424453534, "loss": 4.5926, "step": 388500 }, { "epoch": 0.7250775019302588, "grad_norm": 1.1598362922668457, "learning_rate": 0.00019503694730430007, "loss": 4.6355, "step": 388550 }, { "epoch": 0.7251708074896372, "grad_norm": 0.7973438501358032, "learning_rate": 0.00019503564019634126, "loss": 4.4744, "step": 388600 }, { "epoch": 0.7252641130490157, "grad_norm": 1.1789216995239258, "learning_rate": 0.0001950343329206613, "loss": 4.6298, "step": 388650 }, { "epoch": 0.7253574186083942, "grad_norm": 1.0197705030441284, "learning_rate": 0.00019503302547726244, "loss": 4.4355, "step": 388700 }, { "epoch": 0.7254507241677727, "grad_norm": 1.1121525764465332, "learning_rate": 0.00019503171786614706, "loss": 4.2835, "step": 388750 }, { "epoch": 0.7255440297271513, "grad_norm": 0.8549951910972595, "learning_rate": 0.00019503041008731738, "loss": 4.3967, "step": 388800 }, { "epoch": 0.7256373352865297, "grad_norm": 0.8149085640907288, "learning_rate": 0.00019502910214077576, "loss": 4.4414, "step": 388850 }, { "epoch": 0.7257306408459082, "grad_norm": 1.0984678268432617, "learning_rate": 0.00019502779402652448, "loss": 4.3974, "step": 388900 }, { "epoch": 0.7258239464052867, "grad_norm": 1.0284241437911987, "learning_rate": 0.00019502648574456588, "loss": 4.5263, "step": 388950 }, { "epoch": 0.7259172519646652, "grad_norm": 1.1385748386383057, "learning_rate": 0.00019502517729490227, "loss": 4.6607, "step": 389000 }, { "epoch": 0.7260105575240436, "grad_norm": 1.0035887956619263, "learning_rate": 0.00019502386867753592, "loss": 4.7158, "step": 389050 }, { "epoch": 0.7261038630834221, "grad_norm": 0.9715141654014587, "learning_rate": 0.00019502255989246916, "loss": 4.4282, "step": 389100 }, { "epoch": 0.7261971686428007, "grad_norm": 1.1229904890060425, "learning_rate": 0.0001950212509397043, "loss": 4.3938, "step": 389150 }, { "epoch": 0.7262904742021792, "grad_norm": 0.7794562578201294, "learning_rate": 0.00019501994181924366, "loss": 4.4609, "step": 389200 }, { "epoch": 0.7263837797615577, "grad_norm": 1.1683100461959839, "learning_rate": 0.00019501863253108953, "loss": 4.5351, "step": 389250 }, { "epoch": 0.7264770853209361, "grad_norm": 1.011674404144287, "learning_rate": 0.00019501732307524426, "loss": 4.5928, "step": 389300 }, { "epoch": 0.7265703908803146, "grad_norm": 0.9872075319290161, "learning_rate": 0.0001950160134517101, "loss": 4.7019, "step": 389350 }, { "epoch": 0.7266636964396931, "grad_norm": 1.1409260034561157, "learning_rate": 0.00019501470366048941, "loss": 4.4976, "step": 389400 }, { "epoch": 0.7267570019990716, "grad_norm": 0.9765341877937317, "learning_rate": 0.0001950133937015845, "loss": 4.5095, "step": 389450 }, { "epoch": 0.7268503075584501, "grad_norm": 1.293187141418457, "learning_rate": 0.00019501208357499764, "loss": 4.7893, "step": 389500 }, { "epoch": 0.7269436131178286, "grad_norm": 0.8450595736503601, "learning_rate": 0.0001950107732807312, "loss": 4.5728, "step": 389550 }, { "epoch": 0.7270369186772071, "grad_norm": 0.9158079028129578, "learning_rate": 0.00019500946281878742, "loss": 4.4127, "step": 389600 }, { "epoch": 0.7271302242365856, "grad_norm": 1.1306376457214355, "learning_rate": 0.0001950081521891687, "loss": 4.5249, "step": 389650 }, { "epoch": 0.727223529795964, "grad_norm": 1.2033452987670898, "learning_rate": 0.0001950068413918773, "loss": 4.6137, "step": 389700 }, { "epoch": 0.7273168353553425, "grad_norm": 1.193956732749939, "learning_rate": 0.00019500553042691552, "loss": 4.715, "step": 389750 }, { "epoch": 0.727410140914721, "grad_norm": 1.084000825881958, "learning_rate": 0.00019500421929428573, "loss": 4.5281, "step": 389800 }, { "epoch": 0.7275034464740996, "grad_norm": 1.1593761444091797, "learning_rate": 0.00019500290799399016, "loss": 4.6083, "step": 389850 }, { "epoch": 0.7275967520334781, "grad_norm": 1.0730018615722656, "learning_rate": 0.00019500159652603122, "loss": 4.5715, "step": 389900 }, { "epoch": 0.7276900575928565, "grad_norm": 0.9677862524986267, "learning_rate": 0.00019500028489041114, "loss": 4.5463, "step": 389950 }, { "epoch": 0.727783363152235, "grad_norm": 1.1932060718536377, "learning_rate": 0.00019499897308713232, "loss": 4.5934, "step": 390000 }, { "epoch": 0.727783363152235, "eval_loss": 4.735870361328125, "eval_runtime": 230.6162, "eval_samples_per_second": 11.309, "eval_steps_per_second": 11.309, "eval_tts_loss": 7.53584766990728, "step": 390000 }, { "epoch": 0.7278766687116135, "grad_norm": 0.6686156988143921, "learning_rate": 0.000194997661116197, "loss": 4.5504, "step": 390050 }, { "epoch": 0.727969974270992, "grad_norm": 1.496148705482483, "learning_rate": 0.00019499634897760751, "loss": 4.4838, "step": 390100 }, { "epoch": 0.7280632798303704, "grad_norm": 0.8240087032318115, "learning_rate": 0.0001949950366713662, "loss": 4.5827, "step": 390150 }, { "epoch": 0.728156585389749, "grad_norm": 1.0409563779830933, "learning_rate": 0.00019499372419747534, "loss": 4.6779, "step": 390200 }, { "epoch": 0.7282498909491275, "grad_norm": 0.863301157951355, "learning_rate": 0.0001949924115559373, "loss": 4.7497, "step": 390250 }, { "epoch": 0.728343196508506, "grad_norm": 1.2422351837158203, "learning_rate": 0.00019499109874675436, "loss": 4.697, "step": 390300 }, { "epoch": 0.7284365020678845, "grad_norm": 1.0232807397842407, "learning_rate": 0.00019498978576992885, "loss": 4.6335, "step": 390350 }, { "epoch": 0.7285298076272629, "grad_norm": 0.8831896185874939, "learning_rate": 0.00019498847262546307, "loss": 4.4275, "step": 390400 }, { "epoch": 0.7286231131866414, "grad_norm": 1.2371958494186401, "learning_rate": 0.0001949871593133594, "loss": 4.4267, "step": 390450 }, { "epoch": 0.7287164187460199, "grad_norm": 0.9164360165596008, "learning_rate": 0.00019498584583362003, "loss": 4.498, "step": 390500 }, { "epoch": 0.7288097243053985, "grad_norm": 1.0729354619979858, "learning_rate": 0.0001949845321862474, "loss": 4.8734, "step": 390550 }, { "epoch": 0.728903029864777, "grad_norm": 0.5111291408538818, "learning_rate": 0.00019498321837124375, "loss": 4.4942, "step": 390600 }, { "epoch": 0.7289963354241554, "grad_norm": 0.9883329272270203, "learning_rate": 0.00019498190438861147, "loss": 4.4397, "step": 390650 }, { "epoch": 0.7290896409835339, "grad_norm": 1.101575255393982, "learning_rate": 0.0001949805902383528, "loss": 4.6804, "step": 390700 }, { "epoch": 0.7291829465429124, "grad_norm": 0.9370002746582031, "learning_rate": 0.00019497927592047014, "loss": 4.8269, "step": 390750 }, { "epoch": 0.7292762521022909, "grad_norm": 1.1477497816085815, "learning_rate": 0.00019497796143496575, "loss": 4.6717, "step": 390800 }, { "epoch": 0.7293695576616693, "grad_norm": 1.1353437900543213, "learning_rate": 0.000194976646781842, "loss": 4.7019, "step": 390850 }, { "epoch": 0.7294628632210479, "grad_norm": 0.694190263748169, "learning_rate": 0.00019497533196110114, "loss": 4.5756, "step": 390900 }, { "epoch": 0.7295561687804264, "grad_norm": 1.288985013961792, "learning_rate": 0.00019497401697274555, "loss": 4.6886, "step": 390950 }, { "epoch": 0.7296494743398049, "grad_norm": 0.5960861444473267, "learning_rate": 0.0001949727018167775, "loss": 4.4684, "step": 391000 }, { "epoch": 0.7297427798991833, "grad_norm": 1.026583194732666, "learning_rate": 0.0001949713864931994, "loss": 4.5915, "step": 391050 }, { "epoch": 0.7298360854585618, "grad_norm": 0.8665286302566528, "learning_rate": 0.00019497007100201347, "loss": 4.5365, "step": 391100 }, { "epoch": 0.7299293910179403, "grad_norm": 1.13361656665802, "learning_rate": 0.0001949687553432221, "loss": 4.5007, "step": 391150 }, { "epoch": 0.7300226965773188, "grad_norm": 1.010580062866211, "learning_rate": 0.0001949674395168276, "loss": 4.7294, "step": 391200 }, { "epoch": 0.7301160021366974, "grad_norm": 0.966787576675415, "learning_rate": 0.00019496612352283224, "loss": 4.5658, "step": 391250 }, { "epoch": 0.7302093076960758, "grad_norm": 1.1436415910720825, "learning_rate": 0.00019496480736123838, "loss": 4.7356, "step": 391300 }, { "epoch": 0.7303026132554543, "grad_norm": 1.1993733644485474, "learning_rate": 0.00019496349103204837, "loss": 4.7252, "step": 391350 }, { "epoch": 0.7303959188148328, "grad_norm": 0.901866614818573, "learning_rate": 0.00019496217453526452, "loss": 4.5208, "step": 391400 }, { "epoch": 0.7304892243742113, "grad_norm": 1.084151029586792, "learning_rate": 0.00019496085787088912, "loss": 4.5361, "step": 391450 }, { "epoch": 0.7305825299335897, "grad_norm": 0.8758310675621033, "learning_rate": 0.0001949595410389245, "loss": 4.807, "step": 391500 }, { "epoch": 0.7306758354929682, "grad_norm": 1.0187368392944336, "learning_rate": 0.000194958224039373, "loss": 4.4729, "step": 391550 }, { "epoch": 0.7307691410523468, "grad_norm": 1.1745333671569824, "learning_rate": 0.00019495690687223695, "loss": 4.3888, "step": 391600 }, { "epoch": 0.7308624466117253, "grad_norm": 0.8803514838218689, "learning_rate": 0.0001949555895375187, "loss": 4.5403, "step": 391650 }, { "epoch": 0.7309557521711038, "grad_norm": 0.7681301236152649, "learning_rate": 0.00019495427203522053, "loss": 4.5776, "step": 391700 }, { "epoch": 0.7310490577304822, "grad_norm": 0.8530518412590027, "learning_rate": 0.00019495295436534473, "loss": 4.4472, "step": 391750 }, { "epoch": 0.7311423632898607, "grad_norm": 1.137397050857544, "learning_rate": 0.00019495163652789372, "loss": 4.6771, "step": 391800 }, { "epoch": 0.7312356688492392, "grad_norm": 0.6580849289894104, "learning_rate": 0.00019495031852286978, "loss": 4.477, "step": 391850 }, { "epoch": 0.7313289744086177, "grad_norm": 1.0122594833374023, "learning_rate": 0.0001949490003502752, "loss": 4.6751, "step": 391900 }, { "epoch": 0.7314222799679962, "grad_norm": 1.0346349477767944, "learning_rate": 0.00019494768201011237, "loss": 4.465, "step": 391950 }, { "epoch": 0.7315155855273747, "grad_norm": 0.912458598613739, "learning_rate": 0.0001949463635023836, "loss": 4.5299, "step": 392000 }, { "epoch": 0.7316088910867532, "grad_norm": 1.1435613632202148, "learning_rate": 0.00019494504482709117, "loss": 4.6511, "step": 392050 }, { "epoch": 0.7317021966461317, "grad_norm": 1.081440806388855, "learning_rate": 0.00019494372598423745, "loss": 4.5773, "step": 392100 }, { "epoch": 0.7317955022055102, "grad_norm": 0.9449694156646729, "learning_rate": 0.0001949424069738248, "loss": 4.4573, "step": 392150 }, { "epoch": 0.7318888077648886, "grad_norm": 1.1139907836914062, "learning_rate": 0.00019494108779585545, "loss": 4.4003, "step": 392200 }, { "epoch": 0.7319821133242671, "grad_norm": 1.0709891319274902, "learning_rate": 0.0001949397684503318, "loss": 4.7699, "step": 392250 }, { "epoch": 0.7320754188836457, "grad_norm": 1.0212353467941284, "learning_rate": 0.00019493844893725618, "loss": 4.6232, "step": 392300 }, { "epoch": 0.7321687244430242, "grad_norm": 1.0591708421707153, "learning_rate": 0.00019493712925663088, "loss": 4.6856, "step": 392350 }, { "epoch": 0.7322620300024026, "grad_norm": 1.2571759223937988, "learning_rate": 0.00019493580940845828, "loss": 4.6591, "step": 392400 }, { "epoch": 0.7323553355617811, "grad_norm": 1.0696747303009033, "learning_rate": 0.00019493448939274068, "loss": 4.5348, "step": 392450 }, { "epoch": 0.7324486411211596, "grad_norm": 1.1913007497787476, "learning_rate": 0.00019493316920948038, "loss": 4.5864, "step": 392500 }, { "epoch": 0.7325419466805381, "grad_norm": 1.1159700155258179, "learning_rate": 0.00019493184885867977, "loss": 4.5561, "step": 392550 }, { "epoch": 0.7326352522399165, "grad_norm": 1.200270652770996, "learning_rate": 0.00019493052834034113, "loss": 4.4971, "step": 392600 }, { "epoch": 0.732728557799295, "grad_norm": 0.9367539286613464, "learning_rate": 0.00019492920765446683, "loss": 4.497, "step": 392650 }, { "epoch": 0.7328218633586736, "grad_norm": 1.0923216342926025, "learning_rate": 0.00019492788680105917, "loss": 4.69, "step": 392700 }, { "epoch": 0.7329151689180521, "grad_norm": 1.0680104494094849, "learning_rate": 0.0001949265657801205, "loss": 4.5658, "step": 392750 }, { "epoch": 0.7330084744774306, "grad_norm": 1.0249441862106323, "learning_rate": 0.00019492524459165315, "loss": 4.6498, "step": 392800 }, { "epoch": 0.733101780036809, "grad_norm": 1.0899511575698853, "learning_rate": 0.0001949239232356594, "loss": 4.6632, "step": 392850 }, { "epoch": 0.7331950855961875, "grad_norm": 0.9146956205368042, "learning_rate": 0.00019492260171214166, "loss": 4.7565, "step": 392900 }, { "epoch": 0.733288391155566, "grad_norm": 1.1608656644821167, "learning_rate": 0.00019492128002110225, "loss": 4.5894, "step": 392950 }, { "epoch": 0.7333816967149445, "grad_norm": 0.9399400353431702, "learning_rate": 0.00019491995816254347, "loss": 4.4536, "step": 393000 }, { "epoch": 0.7333816967149445, "eval_loss": 4.722550868988037, "eval_runtime": 230.5993, "eval_samples_per_second": 11.31, "eval_steps_per_second": 11.31, "eval_tts_loss": 7.581275009621574, "step": 393000 }, { "epoch": 0.733475002274323, "grad_norm": 0.8586224317550659, "learning_rate": 0.00019491863613646765, "loss": 4.4582, "step": 393050 }, { "epoch": 0.7335683078337015, "grad_norm": 1.012885332107544, "learning_rate": 0.00019491731394287717, "loss": 4.6452, "step": 393100 }, { "epoch": 0.73366161339308, "grad_norm": 1.2605918645858765, "learning_rate": 0.0001949159915817743, "loss": 4.3936, "step": 393150 }, { "epoch": 0.7337549189524585, "grad_norm": 0.7984983325004578, "learning_rate": 0.00019491466905316142, "loss": 4.6295, "step": 393200 }, { "epoch": 0.733848224511837, "grad_norm": 0.8521361947059631, "learning_rate": 0.00019491334635704085, "loss": 4.6162, "step": 393250 }, { "epoch": 0.7339415300712154, "grad_norm": 1.0116236209869385, "learning_rate": 0.00019491202349341494, "loss": 4.517, "step": 393300 }, { "epoch": 0.7340348356305939, "grad_norm": 1.4281280040740967, "learning_rate": 0.00019491070046228595, "loss": 4.5894, "step": 393350 }, { "epoch": 0.7341281411899725, "grad_norm": 1.2629574537277222, "learning_rate": 0.00019490937726365632, "loss": 4.4543, "step": 393400 }, { "epoch": 0.734221446749351, "grad_norm": 1.2775226831436157, "learning_rate": 0.00019490805389752835, "loss": 4.5062, "step": 393450 }, { "epoch": 0.7343147523087294, "grad_norm": 1.2514355182647705, "learning_rate": 0.00019490673036390435, "loss": 4.7704, "step": 393500 }, { "epoch": 0.7344080578681079, "grad_norm": 1.2405158281326294, "learning_rate": 0.00019490540666278666, "loss": 4.4944, "step": 393550 }, { "epoch": 0.7345013634274864, "grad_norm": 0.9776833057403564, "learning_rate": 0.00019490408279417766, "loss": 4.7371, "step": 393600 }, { "epoch": 0.7345946689868649, "grad_norm": 1.1885216236114502, "learning_rate": 0.0001949027587580796, "loss": 4.511, "step": 393650 }, { "epoch": 0.7346879745462433, "grad_norm": 1.2549670934677124, "learning_rate": 0.0001949014345544949, "loss": 4.5568, "step": 393700 }, { "epoch": 0.7347812801056219, "grad_norm": 1.0409431457519531, "learning_rate": 0.00019490011018342585, "loss": 4.5221, "step": 393750 }, { "epoch": 0.7348745856650004, "grad_norm": 0.9980065822601318, "learning_rate": 0.00019489878564487483, "loss": 4.6895, "step": 393800 }, { "epoch": 0.7349678912243789, "grad_norm": 1.0543571710586548, "learning_rate": 0.00019489746093884414, "loss": 4.7759, "step": 393850 }, { "epoch": 0.7350611967837574, "grad_norm": 1.1471644639968872, "learning_rate": 0.00019489613606533612, "loss": 4.5118, "step": 393900 }, { "epoch": 0.7351545023431358, "grad_norm": 0.9260897040367126, "learning_rate": 0.00019489481102435315, "loss": 4.512, "step": 393950 }, { "epoch": 0.7352478079025143, "grad_norm": 0.8788979649543762, "learning_rate": 0.0001948934858158975, "loss": 4.5552, "step": 394000 }, { "epoch": 0.7353411134618928, "grad_norm": 1.2206834554672241, "learning_rate": 0.00019489216043997156, "loss": 4.4762, "step": 394050 }, { "epoch": 0.7354344190212714, "grad_norm": 0.9536528587341309, "learning_rate": 0.00019489083489657765, "loss": 4.6025, "step": 394100 }, { "epoch": 0.7355277245806499, "grad_norm": 1.0591988563537598, "learning_rate": 0.0001948895091857181, "loss": 4.6237, "step": 394150 }, { "epoch": 0.7356210301400283, "grad_norm": 0.9356663227081299, "learning_rate": 0.00019488818330739527, "loss": 4.3627, "step": 394200 }, { "epoch": 0.7357143356994068, "grad_norm": 1.0228075981140137, "learning_rate": 0.00019488685726161151, "loss": 4.5646, "step": 394250 }, { "epoch": 0.7358076412587853, "grad_norm": 0.7471048831939697, "learning_rate": 0.00019488553104836913, "loss": 4.6334, "step": 394300 }, { "epoch": 0.7359009468181638, "grad_norm": 0.9984940886497498, "learning_rate": 0.00019488420466767047, "loss": 4.4695, "step": 394350 }, { "epoch": 0.7359942523775422, "grad_norm": 1.1658512353897095, "learning_rate": 0.00019488287811951792, "loss": 4.4986, "step": 394400 }, { "epoch": 0.7360875579369208, "grad_norm": 0.700202226638794, "learning_rate": 0.00019488155140391376, "loss": 4.4986, "step": 394450 }, { "epoch": 0.7361808634962993, "grad_norm": 1.2018382549285889, "learning_rate": 0.00019488022452086035, "loss": 4.473, "step": 394500 }, { "epoch": 0.7362741690556778, "grad_norm": 1.2141975164413452, "learning_rate": 0.00019487889747036005, "loss": 4.5214, "step": 394550 }, { "epoch": 0.7363674746150563, "grad_norm": 0.9552044868469238, "learning_rate": 0.00019487757025241518, "loss": 4.6345, "step": 394600 }, { "epoch": 0.7364607801744347, "grad_norm": 0.8278501629829407, "learning_rate": 0.0001948762428670281, "loss": 4.6117, "step": 394650 }, { "epoch": 0.7365540857338132, "grad_norm": 1.293843150138855, "learning_rate": 0.00019487491531420113, "loss": 4.5266, "step": 394700 }, { "epoch": 0.7366473912931917, "grad_norm": 1.0281263589859009, "learning_rate": 0.00019487358759393664, "loss": 4.4981, "step": 394750 }, { "epoch": 0.7367406968525703, "grad_norm": 1.3931729793548584, "learning_rate": 0.000194872259706237, "loss": 4.7601, "step": 394800 }, { "epoch": 0.7368340024119487, "grad_norm": 1.1016684770584106, "learning_rate": 0.00019487093165110447, "loss": 4.5277, "step": 394850 }, { "epoch": 0.7369273079713272, "grad_norm": 0.9391690492630005, "learning_rate": 0.00019486960342854145, "loss": 4.613, "step": 394900 }, { "epoch": 0.7370206135307057, "grad_norm": 1.0593945980072021, "learning_rate": 0.00019486827503855026, "loss": 4.53, "step": 394950 }, { "epoch": 0.7371139190900842, "grad_norm": 0.8746975660324097, "learning_rate": 0.0001948669464811333, "loss": 4.5007, "step": 395000 }, { "epoch": 0.7372072246494626, "grad_norm": 0.904080867767334, "learning_rate": 0.00019486561775629283, "loss": 4.5248, "step": 395050 }, { "epoch": 0.7373005302088411, "grad_norm": 0.9073638916015625, "learning_rate": 0.00019486428886403124, "loss": 4.2864, "step": 395100 }, { "epoch": 0.7373938357682197, "grad_norm": 1.2093853950500488, "learning_rate": 0.0001948629598043509, "loss": 4.6007, "step": 395150 }, { "epoch": 0.7374871413275982, "grad_norm": 0.8730159997940063, "learning_rate": 0.00019486163057725407, "loss": 4.5062, "step": 395200 }, { "epoch": 0.7375804468869767, "grad_norm": 0.858184278011322, "learning_rate": 0.0001948603011827432, "loss": 4.5427, "step": 395250 }, { "epoch": 0.7376737524463551, "grad_norm": 1.0963269472122192, "learning_rate": 0.0001948589716208206, "loss": 4.5644, "step": 395300 }, { "epoch": 0.7377670580057336, "grad_norm": 0.8483244776725769, "learning_rate": 0.00019485764189148855, "loss": 4.6353, "step": 395350 }, { "epoch": 0.7378603635651121, "grad_norm": 0.8771704435348511, "learning_rate": 0.0001948563119947495, "loss": 4.4817, "step": 395400 }, { "epoch": 0.7379536691244906, "grad_norm": 1.1800220012664795, "learning_rate": 0.00019485498193060576, "loss": 4.4817, "step": 395450 }, { "epoch": 0.7380469746838692, "grad_norm": 1.2803168296813965, "learning_rate": 0.00019485365169905964, "loss": 4.6603, "step": 395500 }, { "epoch": 0.7381402802432476, "grad_norm": 1.1339360475540161, "learning_rate": 0.00019485232130011352, "loss": 4.5782, "step": 395550 }, { "epoch": 0.7382335858026261, "grad_norm": 0.944157063961029, "learning_rate": 0.00019485099073376976, "loss": 4.4737, "step": 395600 }, { "epoch": 0.7383268913620046, "grad_norm": 0.9531108140945435, "learning_rate": 0.00019484966000003065, "loss": 4.571, "step": 395650 }, { "epoch": 0.738420196921383, "grad_norm": 1.2628448009490967, "learning_rate": 0.00019484832909889862, "loss": 4.6907, "step": 395700 }, { "epoch": 0.7385135024807615, "grad_norm": 0.8827887773513794, "learning_rate": 0.00019484699803037596, "loss": 4.5015, "step": 395750 }, { "epoch": 0.73860680804014, "grad_norm": 1.2830276489257812, "learning_rate": 0.00019484566679446504, "loss": 4.5319, "step": 395800 }, { "epoch": 0.7387001135995186, "grad_norm": 1.3190330266952515, "learning_rate": 0.0001948443353911682, "loss": 4.7166, "step": 395850 }, { "epoch": 0.7387934191588971, "grad_norm": 1.0245732069015503, "learning_rate": 0.0001948430038204878, "loss": 4.6743, "step": 395900 }, { "epoch": 0.7388867247182755, "grad_norm": 1.0815318822860718, "learning_rate": 0.0001948416720824262, "loss": 4.4301, "step": 395950 }, { "epoch": 0.738980030277654, "grad_norm": 1.1338343620300293, "learning_rate": 0.00019484034017698573, "loss": 4.5976, "step": 396000 }, { "epoch": 0.738980030277654, "eval_loss": 4.7241644859313965, "eval_runtime": 231.6116, "eval_samples_per_second": 11.26, "eval_steps_per_second": 11.26, "eval_tts_loss": 7.610429070300017, "step": 396000 }, { "epoch": 0.7390733358370325, "grad_norm": 1.0788607597351074, "learning_rate": 0.00019483900810416874, "loss": 4.5512, "step": 396050 }, { "epoch": 0.739166641396411, "grad_norm": 1.434680700302124, "learning_rate": 0.00019483767586397758, "loss": 4.4577, "step": 396100 }, { "epoch": 0.7392599469557894, "grad_norm": 1.0275232791900635, "learning_rate": 0.0001948363434564146, "loss": 4.6689, "step": 396150 }, { "epoch": 0.739353252515168, "grad_norm": 1.1067668199539185, "learning_rate": 0.0001948350108814822, "loss": 4.6546, "step": 396200 }, { "epoch": 0.7394465580745465, "grad_norm": 1.112076759338379, "learning_rate": 0.00019483367813918266, "loss": 4.646, "step": 396250 }, { "epoch": 0.739539863633925, "grad_norm": 1.1287716627120972, "learning_rate": 0.00019483234522951837, "loss": 4.5277, "step": 396300 }, { "epoch": 0.7396331691933035, "grad_norm": 0.8934891223907471, "learning_rate": 0.0001948310121524917, "loss": 4.748, "step": 396350 }, { "epoch": 0.7397264747526819, "grad_norm": 1.2289316654205322, "learning_rate": 0.000194829678908105, "loss": 4.5759, "step": 396400 }, { "epoch": 0.7398197803120604, "grad_norm": 0.8652569651603699, "learning_rate": 0.00019482834549636055, "loss": 4.6176, "step": 396450 }, { "epoch": 0.7399130858714389, "grad_norm": 1.4067641496658325, "learning_rate": 0.00019482701191726078, "loss": 4.5168, "step": 396500 }, { "epoch": 0.7400063914308175, "grad_norm": 1.1128288507461548, "learning_rate": 0.00019482567817080798, "loss": 4.381, "step": 396550 }, { "epoch": 0.740099696990196, "grad_norm": 0.8590818643569946, "learning_rate": 0.0001948243442570046, "loss": 4.4131, "step": 396600 }, { "epoch": 0.7401930025495744, "grad_norm": 1.0324068069458008, "learning_rate": 0.00019482301017585292, "loss": 4.6837, "step": 396650 }, { "epoch": 0.7402863081089529, "grad_norm": 0.7649784684181213, "learning_rate": 0.0001948216759273553, "loss": 4.5271, "step": 396700 }, { "epoch": 0.7403796136683314, "grad_norm": 1.2677695751190186, "learning_rate": 0.00019482034151151412, "loss": 4.5032, "step": 396750 }, { "epoch": 0.7404729192277099, "grad_norm": 1.2309435606002808, "learning_rate": 0.0001948190069283317, "loss": 4.597, "step": 396800 }, { "epoch": 0.7405662247870883, "grad_norm": 1.3792515993118286, "learning_rate": 0.00019481767217781042, "loss": 4.5447, "step": 396850 }, { "epoch": 0.7406595303464669, "grad_norm": 1.1079990863800049, "learning_rate": 0.00019481633725995266, "loss": 4.5244, "step": 396900 }, { "epoch": 0.7407528359058454, "grad_norm": 1.1744472980499268, "learning_rate": 0.0001948150021747607, "loss": 4.7786, "step": 396950 }, { "epoch": 0.7408461414652239, "grad_norm": 1.1463979482650757, "learning_rate": 0.00019481366692223698, "loss": 4.5897, "step": 397000 }, { "epoch": 0.7409394470246023, "grad_norm": 1.1247175931930542, "learning_rate": 0.00019481233150238383, "loss": 4.5714, "step": 397050 }, { "epoch": 0.7410327525839808, "grad_norm": 0.952838122844696, "learning_rate": 0.00019481099591520356, "loss": 4.4864, "step": 397100 }, { "epoch": 0.7411260581433593, "grad_norm": 1.1566669940948486, "learning_rate": 0.00019480966016069862, "loss": 4.3836, "step": 397150 }, { "epoch": 0.7412193637027378, "grad_norm": 1.0998328924179077, "learning_rate": 0.00019480832423887128, "loss": 4.4351, "step": 397200 }, { "epoch": 0.7413126692621164, "grad_norm": 0.9825552701950073, "learning_rate": 0.00019480698814972392, "loss": 4.7547, "step": 397250 }, { "epoch": 0.7414059748214948, "grad_norm": 0.9544752836227417, "learning_rate": 0.00019480565189325893, "loss": 4.3914, "step": 397300 }, { "epoch": 0.7414992803808733, "grad_norm": 1.7373327016830444, "learning_rate": 0.0001948043154694786, "loss": 4.5653, "step": 397350 }, { "epoch": 0.7415925859402518, "grad_norm": 0.8258759379386902, "learning_rate": 0.00019480297887838538, "loss": 4.7044, "step": 397400 }, { "epoch": 0.7416858914996303, "grad_norm": 0.9235637187957764, "learning_rate": 0.00019480164211998157, "loss": 4.8239, "step": 397450 }, { "epoch": 0.7417791970590087, "grad_norm": 1.1709929704666138, "learning_rate": 0.00019480030519426956, "loss": 4.797, "step": 397500 }, { "epoch": 0.7418725026183872, "grad_norm": 0.8909640908241272, "learning_rate": 0.00019479896810125166, "loss": 4.5543, "step": 397550 }, { "epoch": 0.7419658081777658, "grad_norm": 0.995737612247467, "learning_rate": 0.00019479763084093027, "loss": 4.5107, "step": 397600 }, { "epoch": 0.7420591137371443, "grad_norm": 0.8248655796051025, "learning_rate": 0.00019479629341330777, "loss": 4.3587, "step": 397650 }, { "epoch": 0.7421524192965228, "grad_norm": 0.9279022812843323, "learning_rate": 0.00019479495581838644, "loss": 4.5868, "step": 397700 }, { "epoch": 0.7422457248559012, "grad_norm": 1.4893839359283447, "learning_rate": 0.00019479361805616873, "loss": 4.4135, "step": 397750 }, { "epoch": 0.7423390304152797, "grad_norm": 0.8427762389183044, "learning_rate": 0.00019479228012665698, "loss": 4.452, "step": 397800 }, { "epoch": 0.7424323359746582, "grad_norm": 1.1434813737869263, "learning_rate": 0.00019479094202985348, "loss": 4.5496, "step": 397850 }, { "epoch": 0.7425256415340367, "grad_norm": 0.9122126698493958, "learning_rate": 0.0001947896037657607, "loss": 4.5626, "step": 397900 }, { "epoch": 0.7426189470934151, "grad_norm": 1.0076179504394531, "learning_rate": 0.00019478826533438092, "loss": 4.4451, "step": 397950 }, { "epoch": 0.7427122526527937, "grad_norm": 1.1911746263504028, "learning_rate": 0.00019478692673571656, "loss": 4.6376, "step": 398000 }, { "epoch": 0.7428055582121722, "grad_norm": 0.8936728835105896, "learning_rate": 0.0001947855879697699, "loss": 4.5737, "step": 398050 }, { "epoch": 0.7428988637715507, "grad_norm": 1.344441294670105, "learning_rate": 0.0001947842490365434, "loss": 4.6513, "step": 398100 }, { "epoch": 0.7429921693309292, "grad_norm": 1.1638767719268799, "learning_rate": 0.00019478290993603937, "loss": 4.5577, "step": 398150 }, { "epoch": 0.7430854748903076, "grad_norm": 1.0375152826309204, "learning_rate": 0.00019478157066826017, "loss": 4.432, "step": 398200 }, { "epoch": 0.7431787804496861, "grad_norm": 0.875194251537323, "learning_rate": 0.00019478023123320817, "loss": 4.5255, "step": 398250 }, { "epoch": 0.7432720860090646, "grad_norm": 1.1665624380111694, "learning_rate": 0.00019477889163088574, "loss": 4.595, "step": 398300 }, { "epoch": 0.7433653915684432, "grad_norm": 1.2813527584075928, "learning_rate": 0.00019477755186129528, "loss": 4.5058, "step": 398350 }, { "epoch": 0.7434586971278216, "grad_norm": 1.1091448068618774, "learning_rate": 0.00019477621192443906, "loss": 4.2865, "step": 398400 }, { "epoch": 0.7435520026872001, "grad_norm": 1.17946457862854, "learning_rate": 0.00019477487182031954, "loss": 4.3896, "step": 398450 }, { "epoch": 0.7436453082465786, "grad_norm": 1.1359678506851196, "learning_rate": 0.00019477353154893906, "loss": 4.3753, "step": 398500 }, { "epoch": 0.7437386138059571, "grad_norm": 0.7580427527427673, "learning_rate": 0.00019477219111029993, "loss": 4.6792, "step": 398550 }, { "epoch": 0.7438319193653355, "grad_norm": 1.2027475833892822, "learning_rate": 0.00019477085050440457, "loss": 4.4967, "step": 398600 }, { "epoch": 0.743925224924714, "grad_norm": 1.2368563413619995, "learning_rate": 0.00019476950973125535, "loss": 4.6587, "step": 398650 }, { "epoch": 0.7440185304840926, "grad_norm": 1.3931561708450317, "learning_rate": 0.0001947681687908546, "loss": 4.6635, "step": 398700 }, { "epoch": 0.7441118360434711, "grad_norm": 1.1592406034469604, "learning_rate": 0.00019476682768320474, "loss": 4.4236, "step": 398750 }, { "epoch": 0.7442051416028496, "grad_norm": 1.1834975481033325, "learning_rate": 0.00019476548640830806, "loss": 4.6033, "step": 398800 }, { "epoch": 0.744298447162228, "grad_norm": 1.156468391418457, "learning_rate": 0.00019476414496616702, "loss": 4.7581, "step": 398850 }, { "epoch": 0.7443917527216065, "grad_norm": 1.147061824798584, "learning_rate": 0.00019476280335678392, "loss": 4.5531, "step": 398900 }, { "epoch": 0.744485058280985, "grad_norm": 1.1417890787124634, "learning_rate": 0.00019476146158016114, "loss": 4.767, "step": 398950 }, { "epoch": 0.7445783638403635, "grad_norm": 0.9738073348999023, "learning_rate": 0.00019476011963630104, "loss": 4.6031, "step": 399000 }, { "epoch": 0.7445783638403635, "eval_loss": 4.733906269073486, "eval_runtime": 231.4593, "eval_samples_per_second": 11.268, "eval_steps_per_second": 11.268, "eval_tts_loss": 7.55335535614557, "step": 399000 }, { "epoch": 0.744671669399742, "grad_norm": 0.9165759682655334, "learning_rate": 0.00019475877752520604, "loss": 4.6827, "step": 399050 }, { "epoch": 0.7447649749591205, "grad_norm": 1.19134521484375, "learning_rate": 0.00019475743524687843, "loss": 4.5243, "step": 399100 }, { "epoch": 0.744858280518499, "grad_norm": 1.1840969324111938, "learning_rate": 0.00019475609280132067, "loss": 4.5984, "step": 399150 }, { "epoch": 0.7449515860778775, "grad_norm": 1.0910753011703491, "learning_rate": 0.00019475475018853502, "loss": 4.4959, "step": 399200 }, { "epoch": 0.745044891637256, "grad_norm": 0.8953019380569458, "learning_rate": 0.00019475340740852395, "loss": 4.6832, "step": 399250 }, { "epoch": 0.7451381971966344, "grad_norm": 0.8007261157035828, "learning_rate": 0.00019475206446128976, "loss": 4.6535, "step": 399300 }, { "epoch": 0.7452315027560129, "grad_norm": 0.9712862968444824, "learning_rate": 0.00019475072134683488, "loss": 4.6529, "step": 399350 }, { "epoch": 0.7453248083153915, "grad_norm": 1.3298040628433228, "learning_rate": 0.0001947493780651616, "loss": 4.5217, "step": 399400 }, { "epoch": 0.74541811387477, "grad_norm": 1.211425542831421, "learning_rate": 0.00019474803461627237, "loss": 4.452, "step": 399450 }, { "epoch": 0.7455114194341484, "grad_norm": 1.0443273782730103, "learning_rate": 0.00019474669100016953, "loss": 4.5222, "step": 399500 }, { "epoch": 0.7456047249935269, "grad_norm": 1.1909388303756714, "learning_rate": 0.00019474534721685547, "loss": 4.84, "step": 399550 }, { "epoch": 0.7456980305529054, "grad_norm": 1.1435961723327637, "learning_rate": 0.00019474400326633252, "loss": 4.589, "step": 399600 }, { "epoch": 0.7457913361122839, "grad_norm": 0.9750648140907288, "learning_rate": 0.00019474265914860308, "loss": 4.4519, "step": 399650 }, { "epoch": 0.7458846416716624, "grad_norm": 1.2055881023406982, "learning_rate": 0.00019474131486366952, "loss": 4.5154, "step": 399700 }, { "epoch": 0.7459779472310409, "grad_norm": 0.9073157906532288, "learning_rate": 0.0001947399704115342, "loss": 4.5807, "step": 399750 }, { "epoch": 0.7460712527904194, "grad_norm": 1.0163545608520508, "learning_rate": 0.00019473862579219952, "loss": 4.6481, "step": 399800 }, { "epoch": 0.7461645583497979, "grad_norm": 1.2088340520858765, "learning_rate": 0.00019473728100566782, "loss": 4.5471, "step": 399850 }, { "epoch": 0.7462578639091764, "grad_norm": 0.8323642015457153, "learning_rate": 0.0001947359360519415, "loss": 4.5136, "step": 399900 }, { "epoch": 0.7463511694685548, "grad_norm": 0.7911824584007263, "learning_rate": 0.0001947345909310229, "loss": 4.6413, "step": 399950 }, { "epoch": 0.7464444750279333, "grad_norm": 1.021842122077942, "learning_rate": 0.00019473324564291445, "loss": 4.4617, "step": 400000 }, { "epoch": 0.7465377805873118, "grad_norm": 1.359620213508606, "learning_rate": 0.00019473190018761847, "loss": 4.7336, "step": 400050 }, { "epoch": 0.7466310861466904, "grad_norm": 1.1204358339309692, "learning_rate": 0.00019473055456513736, "loss": 4.5005, "step": 400100 }, { "epoch": 0.7467243917060689, "grad_norm": 1.1570340394973755, "learning_rate": 0.00019472920877547348, "loss": 4.6138, "step": 400150 }, { "epoch": 0.7468176972654473, "grad_norm": 1.142628788948059, "learning_rate": 0.00019472786281862921, "loss": 4.3871, "step": 400200 }, { "epoch": 0.7469110028248258, "grad_norm": 1.0545388460159302, "learning_rate": 0.00019472651669460693, "loss": 4.594, "step": 400250 }, { "epoch": 0.7470043083842043, "grad_norm": 1.013261079788208, "learning_rate": 0.00019472517040340902, "loss": 4.7622, "step": 400300 }, { "epoch": 0.7470976139435828, "grad_norm": 1.195189118385315, "learning_rate": 0.00019472382394503787, "loss": 4.6786, "step": 400350 }, { "epoch": 0.7471909195029612, "grad_norm": 1.234176516532898, "learning_rate": 0.00019472247731949581, "loss": 4.5959, "step": 400400 }, { "epoch": 0.7472842250623398, "grad_norm": 0.8004019260406494, "learning_rate": 0.00019472113052678528, "loss": 4.6037, "step": 400450 }, { "epoch": 0.7473775306217183, "grad_norm": 0.8816819190979004, "learning_rate": 0.0001947197835669086, "loss": 4.5606, "step": 400500 }, { "epoch": 0.7474708361810968, "grad_norm": 0.9571092128753662, "learning_rate": 0.0001947184364398682, "loss": 4.4627, "step": 400550 }, { "epoch": 0.7475641417404753, "grad_norm": 0.8714603781700134, "learning_rate": 0.00019471708914566637, "loss": 4.621, "step": 400600 }, { "epoch": 0.7476574472998537, "grad_norm": 1.0243371725082397, "learning_rate": 0.0001947157416843056, "loss": 4.4409, "step": 400650 }, { "epoch": 0.7477507528592322, "grad_norm": 0.9247996807098389, "learning_rate": 0.00019471439405578816, "loss": 4.5011, "step": 400700 }, { "epoch": 0.7478440584186107, "grad_norm": 1.2378648519515991, "learning_rate": 0.0001947130462601165, "loss": 4.6031, "step": 400750 }, { "epoch": 0.7479373639779893, "grad_norm": 1.0939531326293945, "learning_rate": 0.00019471169829729295, "loss": 4.8021, "step": 400800 }, { "epoch": 0.7480306695373677, "grad_norm": 1.0560319423675537, "learning_rate": 0.00019471035016731998, "loss": 4.4943, "step": 400850 }, { "epoch": 0.7481239750967462, "grad_norm": 0.8283106684684753, "learning_rate": 0.00019470900187019987, "loss": 4.4933, "step": 400900 }, { "epoch": 0.7482172806561247, "grad_norm": 0.7455849051475525, "learning_rate": 0.00019470765340593503, "loss": 4.613, "step": 400950 }, { "epoch": 0.7483105862155032, "grad_norm": 1.328385829925537, "learning_rate": 0.00019470630477452786, "loss": 4.4819, "step": 401000 }, { "epoch": 0.7484038917748816, "grad_norm": 1.1238863468170166, "learning_rate": 0.00019470495597598072, "loss": 4.5988, "step": 401050 }, { "epoch": 0.7484971973342601, "grad_norm": 1.2375730276107788, "learning_rate": 0.00019470360701029601, "loss": 4.6046, "step": 401100 }, { "epoch": 0.7485905028936387, "grad_norm": 0.9606476426124573, "learning_rate": 0.00019470225787747607, "loss": 4.6423, "step": 401150 }, { "epoch": 0.7486838084530172, "grad_norm": 1.2989537715911865, "learning_rate": 0.00019470090857752332, "loss": 4.5824, "step": 401200 }, { "epoch": 0.7487771140123957, "grad_norm": 1.0816088914871216, "learning_rate": 0.00019469955911044016, "loss": 4.5863, "step": 401250 }, { "epoch": 0.7488704195717741, "grad_norm": 0.8996260762214661, "learning_rate": 0.0001946982094762289, "loss": 4.399, "step": 401300 }, { "epoch": 0.7489637251311526, "grad_norm": 0.9724602103233337, "learning_rate": 0.00019469685967489196, "loss": 4.6596, "step": 401350 }, { "epoch": 0.7490570306905311, "grad_norm": 1.0567646026611328, "learning_rate": 0.00019469550970643176, "loss": 4.5988, "step": 401400 }, { "epoch": 0.7491503362499096, "grad_norm": 1.2678072452545166, "learning_rate": 0.00019469415957085065, "loss": 4.4416, "step": 401450 }, { "epoch": 0.7492436418092882, "grad_norm": 0.6493592262268066, "learning_rate": 0.000194692809268151, "loss": 4.5179, "step": 401500 }, { "epoch": 0.7493369473686666, "grad_norm": 1.1661994457244873, "learning_rate": 0.0001946914587983352, "loss": 4.4027, "step": 401550 }, { "epoch": 0.7494302529280451, "grad_norm": 1.3520020246505737, "learning_rate": 0.0001946901081614056, "loss": 4.4424, "step": 401600 }, { "epoch": 0.7495235584874236, "grad_norm": 1.0197575092315674, "learning_rate": 0.00019468875735736467, "loss": 4.4455, "step": 401650 }, { "epoch": 0.7496168640468021, "grad_norm": 1.0030943155288696, "learning_rate": 0.0001946874063862147, "loss": 4.584, "step": 401700 }, { "epoch": 0.7497101696061805, "grad_norm": 1.0498521327972412, "learning_rate": 0.00019468605524795816, "loss": 4.6093, "step": 401750 }, { "epoch": 0.749803475165559, "grad_norm": 1.0407713651657104, "learning_rate": 0.00019468470394259736, "loss": 4.6358, "step": 401800 }, { "epoch": 0.7498967807249376, "grad_norm": 1.0960047245025635, "learning_rate": 0.00019468335247013475, "loss": 4.3322, "step": 401850 }, { "epoch": 0.7499900862843161, "grad_norm": 1.2660539150238037, "learning_rate": 0.00019468200083057268, "loss": 4.5874, "step": 401900 }, { "epoch": 0.7500833918436945, "grad_norm": 1.0143240690231323, "learning_rate": 0.00019468064902391352, "loss": 4.5931, "step": 401950 }, { "epoch": 0.750176697403073, "grad_norm": 1.3088220357894897, "learning_rate": 0.00019467929705015968, "loss": 4.4499, "step": 402000 }, { "epoch": 0.750176697403073, "eval_loss": 4.724297523498535, "eval_runtime": 230.2641, "eval_samples_per_second": 11.326, "eval_steps_per_second": 11.326, "eval_tts_loss": 7.560537891713845, "step": 402000 }, { "epoch": 0.7502700029624515, "grad_norm": 1.0270438194274902, "learning_rate": 0.00019467794490931356, "loss": 4.4617, "step": 402050 }, { "epoch": 0.75036330852183, "grad_norm": 1.3051048517227173, "learning_rate": 0.00019467659260137747, "loss": 4.4863, "step": 402100 }, { "epoch": 0.7504566140812085, "grad_norm": 0.9769117832183838, "learning_rate": 0.0001946752401263539, "loss": 4.656, "step": 402150 }, { "epoch": 0.750549919640587, "grad_norm": 1.2200247049331665, "learning_rate": 0.0001946738874842452, "loss": 4.4802, "step": 402200 }, { "epoch": 0.7506432251999655, "grad_norm": 0.6177666187286377, "learning_rate": 0.0001946725346750537, "loss": 4.658, "step": 402250 }, { "epoch": 0.750736530759344, "grad_norm": 0.9986745119094849, "learning_rate": 0.00019467118169878185, "loss": 4.4032, "step": 402300 }, { "epoch": 0.7508298363187225, "grad_norm": 1.0253387689590454, "learning_rate": 0.00019466982855543204, "loss": 4.5996, "step": 402350 }, { "epoch": 0.7509231418781009, "grad_norm": 0.9981570839881897, "learning_rate": 0.00019466847524500665, "loss": 4.5612, "step": 402400 }, { "epoch": 0.7510164474374794, "grad_norm": 1.055335283279419, "learning_rate": 0.00019466712176750806, "loss": 4.5569, "step": 402450 }, { "epoch": 0.7511097529968579, "grad_norm": 1.0270355939865112, "learning_rate": 0.0001946657681229386, "loss": 4.6409, "step": 402500 }, { "epoch": 0.7512030585562365, "grad_norm": 0.9899023175239563, "learning_rate": 0.00019466441431130078, "loss": 4.4688, "step": 402550 }, { "epoch": 0.751296364115615, "grad_norm": 1.2617480754852295, "learning_rate": 0.00019466306033259689, "loss": 4.5089, "step": 402600 }, { "epoch": 0.7513896696749934, "grad_norm": 0.9508602619171143, "learning_rate": 0.00019466170618682936, "loss": 4.8497, "step": 402650 }, { "epoch": 0.7514829752343719, "grad_norm": 1.1826248168945312, "learning_rate": 0.00019466035187400057, "loss": 4.517, "step": 402700 }, { "epoch": 0.7515762807937504, "grad_norm": 0.763058066368103, "learning_rate": 0.00019465899739411294, "loss": 4.4908, "step": 402750 }, { "epoch": 0.7516695863531289, "grad_norm": 1.045179843902588, "learning_rate": 0.00019465764274716882, "loss": 4.4248, "step": 402800 }, { "epoch": 0.7517628919125073, "grad_norm": 0.8762446641921997, "learning_rate": 0.00019465628793317063, "loss": 4.538, "step": 402850 }, { "epoch": 0.7518561974718859, "grad_norm": 0.9932625889778137, "learning_rate": 0.00019465493295212075, "loss": 4.6375, "step": 402900 }, { "epoch": 0.7519495030312644, "grad_norm": 1.0016560554504395, "learning_rate": 0.00019465357780402154, "loss": 4.3052, "step": 402950 }, { "epoch": 0.7520428085906429, "grad_norm": 1.1759898662567139, "learning_rate": 0.00019465222248887543, "loss": 4.4555, "step": 403000 }, { "epoch": 0.7521361141500214, "grad_norm": 0.9538491368293762, "learning_rate": 0.0001946508670066848, "loss": 4.4514, "step": 403050 }, { "epoch": 0.7522294197093998, "grad_norm": 1.1780041456222534, "learning_rate": 0.00019464951135745207, "loss": 4.6608, "step": 403100 }, { "epoch": 0.7523227252687783, "grad_norm": 1.1293666362762451, "learning_rate": 0.00019464815554117956, "loss": 4.425, "step": 403150 }, { "epoch": 0.7524160308281568, "grad_norm": 1.1419936418533325, "learning_rate": 0.00019464679955786976, "loss": 4.7019, "step": 403200 }, { "epoch": 0.7525093363875354, "grad_norm": 1.1940276622772217, "learning_rate": 0.00019464544340752496, "loss": 4.3547, "step": 403250 }, { "epoch": 0.7526026419469138, "grad_norm": 0.8929638862609863, "learning_rate": 0.00019464408709014762, "loss": 4.6242, "step": 403300 }, { "epoch": 0.7526959475062923, "grad_norm": 1.017694354057312, "learning_rate": 0.00019464273060574016, "loss": 4.6063, "step": 403350 }, { "epoch": 0.7527892530656708, "grad_norm": 1.4753926992416382, "learning_rate": 0.0001946413739543049, "loss": 4.4256, "step": 403400 }, { "epoch": 0.7528825586250493, "grad_norm": 0.9293118119239807, "learning_rate": 0.00019464001713584428, "loss": 4.5837, "step": 403450 }, { "epoch": 0.7529758641844277, "grad_norm": 0.9962372779846191, "learning_rate": 0.00019463866015036066, "loss": 4.6717, "step": 403500 }, { "epoch": 0.7530691697438062, "grad_norm": 1.1656463146209717, "learning_rate": 0.0001946373029978565, "loss": 4.6189, "step": 403550 }, { "epoch": 0.7531624753031847, "grad_norm": 1.0788663625717163, "learning_rate": 0.0001946359456783341, "loss": 4.3787, "step": 403600 }, { "epoch": 0.7532557808625633, "grad_norm": 1.0301463603973389, "learning_rate": 0.00019463458819179595, "loss": 4.4878, "step": 403650 }, { "epoch": 0.7533490864219418, "grad_norm": 0.9593393802642822, "learning_rate": 0.00019463323053824438, "loss": 4.4218, "step": 403700 }, { "epoch": 0.7534423919813202, "grad_norm": 0.7830232977867126, "learning_rate": 0.00019463187271768183, "loss": 4.4766, "step": 403750 }, { "epoch": 0.7535356975406987, "grad_norm": 0.786916971206665, "learning_rate": 0.00019463051473011064, "loss": 4.4523, "step": 403800 }, { "epoch": 0.7536290031000772, "grad_norm": 1.4050339460372925, "learning_rate": 0.00019462915657553326, "loss": 4.5962, "step": 403850 }, { "epoch": 0.7537223086594557, "grad_norm": 0.8983107805252075, "learning_rate": 0.00019462779825395207, "loss": 4.4027, "step": 403900 }, { "epoch": 0.7538156142188341, "grad_norm": 1.0345656871795654, "learning_rate": 0.00019462643976536948, "loss": 4.4476, "step": 403950 }, { "epoch": 0.7539089197782127, "grad_norm": 1.4160012006759644, "learning_rate": 0.00019462508110978786, "loss": 4.4757, "step": 404000 }, { "epoch": 0.7540022253375912, "grad_norm": 1.0370961427688599, "learning_rate": 0.0001946237222872096, "loss": 4.6786, "step": 404050 }, { "epoch": 0.7540955308969697, "grad_norm": 0.8926733732223511, "learning_rate": 0.00019462236329763718, "loss": 4.4315, "step": 404100 }, { "epoch": 0.7541888364563482, "grad_norm": 1.3053297996520996, "learning_rate": 0.0001946210041410729, "loss": 4.6435, "step": 404150 }, { "epoch": 0.7542821420157266, "grad_norm": 1.331666350364685, "learning_rate": 0.00019461964481751916, "loss": 4.4299, "step": 404200 }, { "epoch": 0.7543754475751051, "grad_norm": 1.0818222761154175, "learning_rate": 0.00019461828532697846, "loss": 4.6642, "step": 404250 }, { "epoch": 0.7544687531344836, "grad_norm": 1.265306830406189, "learning_rate": 0.00019461692566945308, "loss": 4.5585, "step": 404300 }, { "epoch": 0.7545620586938622, "grad_norm": 1.1073338985443115, "learning_rate": 0.0001946155658449455, "loss": 4.6652, "step": 404350 }, { "epoch": 0.7546553642532406, "grad_norm": 1.1014400720596313, "learning_rate": 0.00019461420585345813, "loss": 4.4495, "step": 404400 }, { "epoch": 0.7547486698126191, "grad_norm": 1.0301647186279297, "learning_rate": 0.0001946128456949933, "loss": 4.6083, "step": 404450 }, { "epoch": 0.7548419753719976, "grad_norm": 1.091760277748108, "learning_rate": 0.00019461148536955344, "loss": 4.4661, "step": 404500 }, { "epoch": 0.7549352809313761, "grad_norm": 1.064077615737915, "learning_rate": 0.00019461012487714097, "loss": 4.4569, "step": 404550 }, { "epoch": 0.7550285864907545, "grad_norm": 1.3203434944152832, "learning_rate": 0.00019460876421775825, "loss": 4.6581, "step": 404600 }, { "epoch": 0.755121892050133, "grad_norm": 0.9822664856910706, "learning_rate": 0.00019460740339140775, "loss": 4.5907, "step": 404650 }, { "epoch": 0.7552151976095116, "grad_norm": 0.9781966209411621, "learning_rate": 0.00019460604239809182, "loss": 4.5604, "step": 404700 }, { "epoch": 0.7553085031688901, "grad_norm": 1.252949595451355, "learning_rate": 0.00019460468123781285, "loss": 4.4489, "step": 404750 }, { "epoch": 0.7554018087282686, "grad_norm": 0.8946507573127747, "learning_rate": 0.00019460331991057327, "loss": 4.6669, "step": 404800 }, { "epoch": 0.755495114287647, "grad_norm": 1.1925349235534668, "learning_rate": 0.00019460195841637548, "loss": 4.6114, "step": 404850 }, { "epoch": 0.7555884198470255, "grad_norm": 0.9882096648216248, "learning_rate": 0.00019460059675522188, "loss": 4.6776, "step": 404900 }, { "epoch": 0.755681725406404, "grad_norm": 1.3011587858200073, "learning_rate": 0.00019459923492711487, "loss": 4.5147, "step": 404950 }, { "epoch": 0.7557750309657825, "grad_norm": 1.0943289995193481, "learning_rate": 0.00019459787293205685, "loss": 4.6341, "step": 405000 }, { "epoch": 0.7557750309657825, "eval_loss": 4.72869348526001, "eval_runtime": 232.6461, "eval_samples_per_second": 11.21, "eval_steps_per_second": 11.21, "eval_tts_loss": 7.544485853397876, "step": 405000 }, { "epoch": 0.755868336525161, "grad_norm": 1.1097865104675293, "learning_rate": 0.00019459651077005023, "loss": 4.4529, "step": 405050 }, { "epoch": 0.7559616420845395, "grad_norm": 1.0926594734191895, "learning_rate": 0.00019459514844109743, "loss": 4.4716, "step": 405100 }, { "epoch": 0.756054947643918, "grad_norm": 1.0199013948440552, "learning_rate": 0.0001945937859452008, "loss": 4.6114, "step": 405150 }, { "epoch": 0.7561482532032965, "grad_norm": 1.1694191694259644, "learning_rate": 0.00019459242328236282, "loss": 4.3158, "step": 405200 }, { "epoch": 0.756241558762675, "grad_norm": 1.148744821548462, "learning_rate": 0.00019459106045258587, "loss": 4.5701, "step": 405250 }, { "epoch": 0.7563348643220534, "grad_norm": 0.6980789303779602, "learning_rate": 0.0001945896974558723, "loss": 4.5219, "step": 405300 }, { "epoch": 0.7564281698814319, "grad_norm": 1.2195786237716675, "learning_rate": 0.00019458833429222459, "loss": 4.3908, "step": 405350 }, { "epoch": 0.7565214754408105, "grad_norm": 1.0867304801940918, "learning_rate": 0.00019458697096164508, "loss": 4.4848, "step": 405400 }, { "epoch": 0.756614781000189, "grad_norm": 1.2966946363449097, "learning_rate": 0.00019458560746413622, "loss": 4.6711, "step": 405450 }, { "epoch": 0.7567080865595674, "grad_norm": 1.1345088481903076, "learning_rate": 0.00019458424379970043, "loss": 4.941, "step": 405500 }, { "epoch": 0.7568013921189459, "grad_norm": 1.2740004062652588, "learning_rate": 0.00019458287996834009, "loss": 4.666, "step": 405550 }, { "epoch": 0.7568946976783244, "grad_norm": 0.915218710899353, "learning_rate": 0.00019458151597005757, "loss": 4.3748, "step": 405600 }, { "epoch": 0.7569880032377029, "grad_norm": 1.0948460102081299, "learning_rate": 0.00019458015180485536, "loss": 4.595, "step": 405650 }, { "epoch": 0.7570813087970814, "grad_norm": 1.0221774578094482, "learning_rate": 0.0001945787874727358, "loss": 4.3731, "step": 405700 }, { "epoch": 0.7571746143564599, "grad_norm": 1.1602940559387207, "learning_rate": 0.00019457742297370132, "loss": 4.5798, "step": 405750 }, { "epoch": 0.7572679199158384, "grad_norm": 1.1894081830978394, "learning_rate": 0.00019457605830775437, "loss": 4.5092, "step": 405800 }, { "epoch": 0.7573612254752169, "grad_norm": 1.1323769092559814, "learning_rate": 0.00019457469347489728, "loss": 4.3593, "step": 405850 }, { "epoch": 0.7574545310345954, "grad_norm": 1.1201635599136353, "learning_rate": 0.00019457332847513248, "loss": 4.6032, "step": 405900 }, { "epoch": 0.7575478365939738, "grad_norm": 0.9292293190956116, "learning_rate": 0.00019457196330846242, "loss": 4.4532, "step": 405950 }, { "epoch": 0.7576411421533523, "grad_norm": 1.1106488704681396, "learning_rate": 0.00019457059797488947, "loss": 4.5502, "step": 406000 }, { "epoch": 0.7577344477127308, "grad_norm": 1.143852710723877, "learning_rate": 0.0001945692324744161, "loss": 4.5175, "step": 406050 }, { "epoch": 0.7578277532721094, "grad_norm": 0.8539165258407593, "learning_rate": 0.0001945678668070446, "loss": 4.5666, "step": 406100 }, { "epoch": 0.7579210588314879, "grad_norm": 1.088355541229248, "learning_rate": 0.0001945665009727775, "loss": 4.4709, "step": 406150 }, { "epoch": 0.7580143643908663, "grad_norm": 1.0392314195632935, "learning_rate": 0.0001945651349716172, "loss": 4.5914, "step": 406200 }, { "epoch": 0.7581076699502448, "grad_norm": 0.8739420175552368, "learning_rate": 0.000194563768803566, "loss": 4.5308, "step": 406250 }, { "epoch": 0.7582009755096233, "grad_norm": 1.187957525253296, "learning_rate": 0.00019456240246862644, "loss": 4.6386, "step": 406300 }, { "epoch": 0.7582942810690018, "grad_norm": 0.9893680810928345, "learning_rate": 0.00019456103596680088, "loss": 4.3799, "step": 406350 }, { "epoch": 0.7583875866283802, "grad_norm": 1.0255956649780273, "learning_rate": 0.0001945596692980917, "loss": 4.7057, "step": 406400 }, { "epoch": 0.7584808921877588, "grad_norm": 1.1427487134933472, "learning_rate": 0.00019455830246250136, "loss": 4.5165, "step": 406450 }, { "epoch": 0.7585741977471373, "grad_norm": 1.0257514715194702, "learning_rate": 0.00019455693546003227, "loss": 4.6241, "step": 406500 }, { "epoch": 0.7586675033065158, "grad_norm": 1.1918649673461914, "learning_rate": 0.0001945555682906868, "loss": 4.6769, "step": 406550 }, { "epoch": 0.7587608088658943, "grad_norm": 0.8059489727020264, "learning_rate": 0.0001945542009544674, "loss": 4.5509, "step": 406600 }, { "epoch": 0.7588541144252727, "grad_norm": 1.0606913566589355, "learning_rate": 0.00019455283345137648, "loss": 4.4686, "step": 406650 }, { "epoch": 0.7589474199846512, "grad_norm": 1.049088954925537, "learning_rate": 0.00019455146578141642, "loss": 4.5742, "step": 406700 }, { "epoch": 0.7590407255440297, "grad_norm": 1.4002255201339722, "learning_rate": 0.00019455009794458968, "loss": 4.6296, "step": 406750 }, { "epoch": 0.7591340311034083, "grad_norm": 1.4870983362197876, "learning_rate": 0.00019454872994089867, "loss": 4.567, "step": 406800 }, { "epoch": 0.7592273366627867, "grad_norm": 0.9239751696586609, "learning_rate": 0.00019454736177034574, "loss": 4.4658, "step": 406850 }, { "epoch": 0.7593206422221652, "grad_norm": 1.0361567735671997, "learning_rate": 0.00019454599343293337, "loss": 4.4316, "step": 406900 }, { "epoch": 0.7594139477815437, "grad_norm": 0.99744713306427, "learning_rate": 0.000194544624928664, "loss": 4.6061, "step": 406950 }, { "epoch": 0.7595072533409222, "grad_norm": 0.8480738997459412, "learning_rate": 0.00019454325625753997, "loss": 4.2468, "step": 407000 }, { "epoch": 0.7596005589003006, "grad_norm": 1.0402275323867798, "learning_rate": 0.0001945418874195637, "loss": 4.6353, "step": 407050 }, { "epoch": 0.7596938644596791, "grad_norm": 1.025551676750183, "learning_rate": 0.00019454051841473767, "loss": 4.6685, "step": 407100 }, { "epoch": 0.7597871700190577, "grad_norm": 1.0114656686782837, "learning_rate": 0.00019453914924306425, "loss": 4.7492, "step": 407150 }, { "epoch": 0.7598804755784362, "grad_norm": 1.142307162284851, "learning_rate": 0.00019453777990454586, "loss": 4.7868, "step": 407200 }, { "epoch": 0.7599737811378147, "grad_norm": 1.160451889038086, "learning_rate": 0.00019453641039918494, "loss": 4.7449, "step": 407250 }, { "epoch": 0.7600670866971931, "grad_norm": 1.1598868370056152, "learning_rate": 0.00019453504072698387, "loss": 4.6213, "step": 407300 }, { "epoch": 0.7601603922565716, "grad_norm": 0.8196390867233276, "learning_rate": 0.0001945336708879451, "loss": 4.4612, "step": 407350 }, { "epoch": 0.7602536978159501, "grad_norm": 1.2065792083740234, "learning_rate": 0.00019453230088207105, "loss": 4.5206, "step": 407400 }, { "epoch": 0.7603470033753286, "grad_norm": 1.200914740562439, "learning_rate": 0.00019453093070936407, "loss": 4.574, "step": 407450 }, { "epoch": 0.7604403089347072, "grad_norm": 1.0380594730377197, "learning_rate": 0.0001945295603698267, "loss": 4.5967, "step": 407500 }, { "epoch": 0.7605336144940856, "grad_norm": 0.948436439037323, "learning_rate": 0.00019452818986346123, "loss": 4.4912, "step": 407550 }, { "epoch": 0.7606269200534641, "grad_norm": 1.1930091381072998, "learning_rate": 0.00019452681919027018, "loss": 4.2002, "step": 407600 }, { "epoch": 0.7607202256128426, "grad_norm": 0.9641384482383728, "learning_rate": 0.0001945254483502559, "loss": 4.5361, "step": 407650 }, { "epoch": 0.7608135311722211, "grad_norm": 0.9800957441329956, "learning_rate": 0.00019452407734342085, "loss": 4.7418, "step": 407700 }, { "epoch": 0.7609068367315995, "grad_norm": 0.9638832807540894, "learning_rate": 0.00019452270616976742, "loss": 4.6648, "step": 407750 }, { "epoch": 0.761000142290978, "grad_norm": 0.9217967987060547, "learning_rate": 0.00019452133482929807, "loss": 4.4051, "step": 407800 }, { "epoch": 0.7610934478503566, "grad_norm": 0.972507894039154, "learning_rate": 0.00019451996332201516, "loss": 4.5565, "step": 407850 }, { "epoch": 0.7611867534097351, "grad_norm": 1.100412130355835, "learning_rate": 0.0001945185916479212, "loss": 4.6013, "step": 407900 }, { "epoch": 0.7612800589691135, "grad_norm": 1.2424074411392212, "learning_rate": 0.00019451721980701851, "loss": 4.5051, "step": 407950 }, { "epoch": 0.761373364528492, "grad_norm": 0.6012529134750366, "learning_rate": 0.00019451584779930955, "loss": 4.5495, "step": 408000 }, { "epoch": 0.761373364528492, "eval_loss": 4.730047225952148, "eval_runtime": 231.8619, "eval_samples_per_second": 11.248, "eval_steps_per_second": 11.248, "eval_tts_loss": 7.581340771529858, "step": 408000 }, { "epoch": 0.7614666700878705, "grad_norm": 1.240382432937622, "learning_rate": 0.00019451447562479678, "loss": 4.44, "step": 408050 }, { "epoch": 0.761559975647249, "grad_norm": 0.8447127938270569, "learning_rate": 0.0001945131032834826, "loss": 4.5198, "step": 408100 }, { "epoch": 0.7616532812066275, "grad_norm": 1.0848270654678345, "learning_rate": 0.00019451173077536938, "loss": 4.6223, "step": 408150 }, { "epoch": 0.761746586766006, "grad_norm": 1.1647281646728516, "learning_rate": 0.0001945103581004596, "loss": 4.4722, "step": 408200 }, { "epoch": 0.7618398923253845, "grad_norm": 1.1008877754211426, "learning_rate": 0.0001945089852587557, "loss": 4.5822, "step": 408250 }, { "epoch": 0.761933197884763, "grad_norm": 1.0027493238449097, "learning_rate": 0.00019450761225026005, "loss": 4.5399, "step": 408300 }, { "epoch": 0.7620265034441415, "grad_norm": 1.3000314235687256, "learning_rate": 0.0001945062390749751, "loss": 4.5339, "step": 408350 }, { "epoch": 0.7621198090035199, "grad_norm": 1.1075400114059448, "learning_rate": 0.00019450486573290325, "loss": 4.2773, "step": 408400 }, { "epoch": 0.7622131145628984, "grad_norm": 1.1882368326187134, "learning_rate": 0.00019450349222404695, "loss": 4.5993, "step": 408450 }, { "epoch": 0.7623064201222769, "grad_norm": 0.9934725761413574, "learning_rate": 0.0001945021185484086, "loss": 4.6568, "step": 408500 }, { "epoch": 0.7623997256816555, "grad_norm": 0.9820505976676941, "learning_rate": 0.00019450074470599068, "loss": 4.4453, "step": 408550 }, { "epoch": 0.762493031241034, "grad_norm": 1.356992244720459, "learning_rate": 0.00019449937069679555, "loss": 4.5526, "step": 408600 }, { "epoch": 0.7625863368004124, "grad_norm": 0.980800986289978, "learning_rate": 0.00019449799652082562, "loss": 4.497, "step": 408650 }, { "epoch": 0.7626796423597909, "grad_norm": 1.1779906749725342, "learning_rate": 0.0001944966221780834, "loss": 4.5632, "step": 408700 }, { "epoch": 0.7627729479191694, "grad_norm": 0.7027290463447571, "learning_rate": 0.00019449524766857127, "loss": 4.5865, "step": 408750 }, { "epoch": 0.7628662534785479, "grad_norm": 0.9588069915771484, "learning_rate": 0.00019449387299229164, "loss": 4.6278, "step": 408800 }, { "epoch": 0.7629595590379263, "grad_norm": 1.077543020248413, "learning_rate": 0.00019449249814924694, "loss": 4.5214, "step": 408850 }, { "epoch": 0.7630528645973048, "grad_norm": 0.7556289434432983, "learning_rate": 0.00019449112313943962, "loss": 4.3555, "step": 408900 }, { "epoch": 0.7631461701566834, "grad_norm": 1.1389154195785522, "learning_rate": 0.0001944897479628721, "loss": 4.6045, "step": 408950 }, { "epoch": 0.7632394757160619, "grad_norm": 1.0199394226074219, "learning_rate": 0.00019448837261954678, "loss": 4.3561, "step": 409000 }, { "epoch": 0.7633327812754404, "grad_norm": 1.3064322471618652, "learning_rate": 0.00019448699710946613, "loss": 4.4714, "step": 409050 }, { "epoch": 0.7634260868348188, "grad_norm": 1.4018571376800537, "learning_rate": 0.00019448562143263256, "loss": 4.4632, "step": 409100 }, { "epoch": 0.7635193923941973, "grad_norm": 1.2147985696792603, "learning_rate": 0.00019448424558904848, "loss": 4.8225, "step": 409150 }, { "epoch": 0.7636126979535758, "grad_norm": 1.0787280797958374, "learning_rate": 0.00019448286957871632, "loss": 4.4554, "step": 409200 }, { "epoch": 0.7637060035129543, "grad_norm": 1.1411551237106323, "learning_rate": 0.0001944814934016385, "loss": 4.4555, "step": 409250 }, { "epoch": 0.7637993090723328, "grad_norm": 0.8726068735122681, "learning_rate": 0.00019448011705781754, "loss": 4.648, "step": 409300 }, { "epoch": 0.7638926146317113, "grad_norm": 0.9773291945457458, "learning_rate": 0.00019447874054725578, "loss": 4.6669, "step": 409350 }, { "epoch": 0.7639859201910898, "grad_norm": 0.9928725361824036, "learning_rate": 0.00019447736386995563, "loss": 4.7582, "step": 409400 }, { "epoch": 0.7640792257504683, "grad_norm": 1.1772263050079346, "learning_rate": 0.00019447598702591957, "loss": 4.7123, "step": 409450 }, { "epoch": 0.7641725313098467, "grad_norm": 1.0281718969345093, "learning_rate": 0.00019447461001515003, "loss": 4.5314, "step": 409500 }, { "epoch": 0.7642658368692252, "grad_norm": 1.2277331352233887, "learning_rate": 0.0001944732328376494, "loss": 4.6218, "step": 409550 }, { "epoch": 0.7643591424286037, "grad_norm": 0.8721914887428284, "learning_rate": 0.00019447185549342018, "loss": 4.5136, "step": 409600 }, { "epoch": 0.7644524479879823, "grad_norm": 0.7213169932365417, "learning_rate": 0.00019447047798246472, "loss": 4.5493, "step": 409650 }, { "epoch": 0.7645457535473608, "grad_norm": 1.10368812084198, "learning_rate": 0.00019446910030478548, "loss": 4.6382, "step": 409700 }, { "epoch": 0.7646390591067392, "grad_norm": 1.156548261642456, "learning_rate": 0.00019446772246038493, "loss": 4.8191, "step": 409750 }, { "epoch": 0.7647323646661177, "grad_norm": 1.1590478420257568, "learning_rate": 0.00019446634444926546, "loss": 4.7934, "step": 409800 }, { "epoch": 0.7648256702254962, "grad_norm": 0.8049784302711487, "learning_rate": 0.00019446496627142953, "loss": 4.5612, "step": 409850 }, { "epoch": 0.7649189757848747, "grad_norm": 1.069915771484375, "learning_rate": 0.00019446358792687954, "loss": 4.5778, "step": 409900 }, { "epoch": 0.7650122813442531, "grad_norm": 1.0776913166046143, "learning_rate": 0.0001944622094156179, "loss": 4.7124, "step": 409950 }, { "epoch": 0.7651055869036317, "grad_norm": 1.3020472526550293, "learning_rate": 0.00019446083073764716, "loss": 4.4051, "step": 410000 }, { "epoch": 0.7651988924630102, "grad_norm": 1.0878071784973145, "learning_rate": 0.00019445945189296965, "loss": 4.6188, "step": 410050 }, { "epoch": 0.7652921980223887, "grad_norm": 1.1448755264282227, "learning_rate": 0.00019445807288158778, "loss": 4.6531, "step": 410100 }, { "epoch": 0.7653855035817672, "grad_norm": 0.8183618783950806, "learning_rate": 0.0001944566937035041, "loss": 4.751, "step": 410150 }, { "epoch": 0.7654788091411456, "grad_norm": 0.969052791595459, "learning_rate": 0.00019445531435872092, "loss": 4.5996, "step": 410200 }, { "epoch": 0.7655721147005241, "grad_norm": 0.8166422247886658, "learning_rate": 0.00019445393484724077, "loss": 4.5969, "step": 410250 }, { "epoch": 0.7656654202599026, "grad_norm": 0.8773173093795776, "learning_rate": 0.000194452555169066, "loss": 4.5966, "step": 410300 }, { "epoch": 0.7657587258192812, "grad_norm": 1.2333333492279053, "learning_rate": 0.00019445117532419914, "loss": 4.5009, "step": 410350 }, { "epoch": 0.7658520313786596, "grad_norm": 1.1725988388061523, "learning_rate": 0.00019444979531264256, "loss": 4.6075, "step": 410400 }, { "epoch": 0.7659453369380381, "grad_norm": 1.0543071031570435, "learning_rate": 0.00019444841513439868, "loss": 4.6056, "step": 410450 }, { "epoch": 0.7660386424974166, "grad_norm": 1.2511577606201172, "learning_rate": 0.00019444703478947, "loss": 4.5412, "step": 410500 }, { "epoch": 0.7661319480567951, "grad_norm": 3.9236738681793213, "learning_rate": 0.0001944456542778589, "loss": 4.5994, "step": 410550 }, { "epoch": 0.7662252536161736, "grad_norm": 1.1855190992355347, "learning_rate": 0.00019444427359956788, "loss": 4.4763, "step": 410600 }, { "epoch": 0.766318559175552, "grad_norm": 0.8772841095924377, "learning_rate": 0.00019444289275459927, "loss": 4.7665, "step": 410650 }, { "epoch": 0.7664118647349306, "grad_norm": 0.7112861275672913, "learning_rate": 0.0001944415117429556, "loss": 4.5654, "step": 410700 }, { "epoch": 0.7665051702943091, "grad_norm": 1.1377826929092407, "learning_rate": 0.0001944401305646393, "loss": 4.5007, "step": 410750 }, { "epoch": 0.7665984758536876, "grad_norm": 0.8896861672401428, "learning_rate": 0.00019443874921965277, "loss": 4.5797, "step": 410800 }, { "epoch": 0.766691781413066, "grad_norm": 1.1979306936264038, "learning_rate": 0.00019443736770799849, "loss": 4.4931, "step": 410850 }, { "epoch": 0.7667850869724445, "grad_norm": 1.0301620960235596, "learning_rate": 0.00019443598602967882, "loss": 4.4661, "step": 410900 }, { "epoch": 0.766878392531823, "grad_norm": 1.393868088722229, "learning_rate": 0.00019443460418469628, "loss": 4.5384, "step": 410950 }, { "epoch": 0.7669716980912015, "grad_norm": 1.014648199081421, "learning_rate": 0.0001944332221730533, "loss": 4.4657, "step": 411000 }, { "epoch": 0.7669716980912015, "eval_loss": 4.729247093200684, "eval_runtime": 234.8612, "eval_samples_per_second": 11.104, "eval_steps_per_second": 11.104, "eval_tts_loss": 7.555574542404284, "step": 411000 }, { "epoch": 0.7670650036505801, "grad_norm": 1.115856647491455, "learning_rate": 0.00019443183999475229, "loss": 4.71, "step": 411050 }, { "epoch": 0.7671583092099585, "grad_norm": 0.8463857769966125, "learning_rate": 0.00019443045764979568, "loss": 4.4989, "step": 411100 }, { "epoch": 0.767251614769337, "grad_norm": 0.9952225089073181, "learning_rate": 0.0001944290751381859, "loss": 4.7664, "step": 411150 }, { "epoch": 0.7673449203287155, "grad_norm": 0.9273759126663208, "learning_rate": 0.00019442769245992547, "loss": 4.602, "step": 411200 }, { "epoch": 0.767438225888094, "grad_norm": 1.1283340454101562, "learning_rate": 0.00019442630961501677, "loss": 4.5756, "step": 411250 }, { "epoch": 0.7675315314474724, "grad_norm": 1.040757417678833, "learning_rate": 0.00019442492660346223, "loss": 4.5112, "step": 411300 }, { "epoch": 0.7676248370068509, "grad_norm": 1.1037769317626953, "learning_rate": 0.00019442354342526433, "loss": 4.6146, "step": 411350 }, { "epoch": 0.7677181425662295, "grad_norm": 1.0502253770828247, "learning_rate": 0.00019442216008042545, "loss": 4.6156, "step": 411400 }, { "epoch": 0.767811448125608, "grad_norm": 0.9768388271331787, "learning_rate": 0.00019442077656894812, "loss": 4.4807, "step": 411450 }, { "epoch": 0.7679047536849865, "grad_norm": 1.213956356048584, "learning_rate": 0.0001944193928908347, "loss": 4.4356, "step": 411500 }, { "epoch": 0.7679980592443649, "grad_norm": 1.2873953580856323, "learning_rate": 0.00019441800904608769, "loss": 4.5072, "step": 411550 }, { "epoch": 0.7680913648037434, "grad_norm": 0.7170881628990173, "learning_rate": 0.0001944166250347095, "loss": 4.4458, "step": 411600 }, { "epoch": 0.7681846703631219, "grad_norm": 0.972938060760498, "learning_rate": 0.00019441524085670258, "loss": 4.5461, "step": 411650 }, { "epoch": 0.7682779759225004, "grad_norm": 1.1991459131240845, "learning_rate": 0.00019441385651206936, "loss": 4.4705, "step": 411700 }, { "epoch": 0.7683712814818789, "grad_norm": 1.1323275566101074, "learning_rate": 0.0001944124720008123, "loss": 4.6942, "step": 411750 }, { "epoch": 0.7684645870412574, "grad_norm": 1.1658388376235962, "learning_rate": 0.00019441108732293384, "loss": 4.7014, "step": 411800 }, { "epoch": 0.7685578926006359, "grad_norm": 0.9996901750564575, "learning_rate": 0.00019440970247843644, "loss": 4.5796, "step": 411850 }, { "epoch": 0.7686511981600144, "grad_norm": 0.9950489401817322, "learning_rate": 0.0001944083174673225, "loss": 4.6371, "step": 411900 }, { "epoch": 0.7687445037193928, "grad_norm": 0.9656080007553101, "learning_rate": 0.00019440693228959453, "loss": 4.6424, "step": 411950 }, { "epoch": 0.7688378092787713, "grad_norm": 1.0694754123687744, "learning_rate": 0.0001944055469452549, "loss": 4.4645, "step": 412000 }, { "epoch": 0.7689311148381498, "grad_norm": 1.3352670669555664, "learning_rate": 0.0001944041614343061, "loss": 4.5704, "step": 412050 }, { "epoch": 0.7690244203975284, "grad_norm": 1.2701009511947632, "learning_rate": 0.00019440277575675057, "loss": 4.7115, "step": 412100 }, { "epoch": 0.7691177259569069, "grad_norm": 1.1434177160263062, "learning_rate": 0.00019440138991259076, "loss": 4.6429, "step": 412150 }, { "epoch": 0.7692110315162853, "grad_norm": 0.9898224472999573, "learning_rate": 0.0001944000039018291, "loss": 4.4657, "step": 412200 }, { "epoch": 0.7693043370756638, "grad_norm": 1.0095897912979126, "learning_rate": 0.00019439861772446802, "loss": 4.3326, "step": 412250 }, { "epoch": 0.7693976426350423, "grad_norm": 1.093815565109253, "learning_rate": 0.00019439723138051004, "loss": 4.5972, "step": 412300 }, { "epoch": 0.7694909481944208, "grad_norm": 1.092368483543396, "learning_rate": 0.0001943958448699575, "loss": 4.5492, "step": 412350 }, { "epoch": 0.7695842537537992, "grad_norm": 1.0995482206344604, "learning_rate": 0.00019439445819281292, "loss": 4.5641, "step": 412400 }, { "epoch": 0.7696775593131778, "grad_norm": 1.3554096221923828, "learning_rate": 0.00019439307134907872, "loss": 4.415, "step": 412450 }, { "epoch": 0.7697708648725563, "grad_norm": 1.2155762910842896, "learning_rate": 0.0001943916843387574, "loss": 4.4498, "step": 412500 }, { "epoch": 0.7698641704319348, "grad_norm": 1.114269733428955, "learning_rate": 0.00019439029716185135, "loss": 4.5365, "step": 412550 }, { "epoch": 0.7699574759913133, "grad_norm": 1.1325422525405884, "learning_rate": 0.00019438890981836302, "loss": 4.5969, "step": 412600 }, { "epoch": 0.7700507815506917, "grad_norm": 1.136763095855713, "learning_rate": 0.0001943875223082949, "loss": 4.6564, "step": 412650 }, { "epoch": 0.7701440871100702, "grad_norm": 1.1705725193023682, "learning_rate": 0.0001943861346316494, "loss": 4.3569, "step": 412700 }, { "epoch": 0.7702373926694487, "grad_norm": 1.3166948556900024, "learning_rate": 0.00019438474678842895, "loss": 4.2721, "step": 412750 }, { "epoch": 0.7703306982288273, "grad_norm": 1.262305736541748, "learning_rate": 0.00019438335877863604, "loss": 4.6105, "step": 412800 }, { "epoch": 0.7704240037882057, "grad_norm": 1.311882734298706, "learning_rate": 0.00019438197060227312, "loss": 4.5957, "step": 412850 }, { "epoch": 0.7705173093475842, "grad_norm": 1.2204020023345947, "learning_rate": 0.00019438058225934265, "loss": 4.5673, "step": 412900 }, { "epoch": 0.7706106149069627, "grad_norm": 1.0517381429672241, "learning_rate": 0.000194379193749847, "loss": 4.6357, "step": 412950 }, { "epoch": 0.7707039204663412, "grad_norm": 1.0536929368972778, "learning_rate": 0.00019437780507378872, "loss": 4.4254, "step": 413000 }, { "epoch": 0.7707972260257197, "grad_norm": 1.018850326538086, "learning_rate": 0.00019437641623117023, "loss": 4.5578, "step": 413050 }, { "epoch": 0.7708905315850981, "grad_norm": 1.2499680519104004, "learning_rate": 0.00019437502722199397, "loss": 4.6343, "step": 413100 }, { "epoch": 0.7709838371444767, "grad_norm": 1.065850853919983, "learning_rate": 0.00019437363804626235, "loss": 4.5045, "step": 413150 }, { "epoch": 0.7710771427038552, "grad_norm": 1.150632619857788, "learning_rate": 0.00019437224870397793, "loss": 4.613, "step": 413200 }, { "epoch": 0.7711704482632337, "grad_norm": 0.9182499647140503, "learning_rate": 0.00019437085919514306, "loss": 4.5223, "step": 413250 }, { "epoch": 0.7712637538226121, "grad_norm": 1.01868736743927, "learning_rate": 0.00019436946951976024, "loss": 4.6569, "step": 413300 }, { "epoch": 0.7713570593819906, "grad_norm": 1.1553199291229248, "learning_rate": 0.0001943680796778319, "loss": 4.7115, "step": 413350 }, { "epoch": 0.7714503649413691, "grad_norm": 1.235852599143982, "learning_rate": 0.00019436668966936048, "loss": 4.5456, "step": 413400 }, { "epoch": 0.7715436705007476, "grad_norm": 1.1578147411346436, "learning_rate": 0.00019436529949434848, "loss": 4.5546, "step": 413450 }, { "epoch": 0.7716369760601262, "grad_norm": 1.095115303993225, "learning_rate": 0.00019436390915279837, "loss": 4.6062, "step": 413500 }, { "epoch": 0.7717302816195046, "grad_norm": 0.9559109807014465, "learning_rate": 0.0001943625186447125, "loss": 4.5222, "step": 413550 }, { "epoch": 0.7718235871788831, "grad_norm": 0.964905321598053, "learning_rate": 0.00019436112797009342, "loss": 4.7261, "step": 413600 }, { "epoch": 0.7719168927382616, "grad_norm": 1.1534135341644287, "learning_rate": 0.00019435973712894356, "loss": 4.575, "step": 413650 }, { "epoch": 0.7720101982976401, "grad_norm": 0.8084390163421631, "learning_rate": 0.00019435834612126535, "loss": 4.6461, "step": 413700 }, { "epoch": 0.7721035038570185, "grad_norm": 0.9302859306335449, "learning_rate": 0.00019435695494706128, "loss": 4.5341, "step": 413750 }, { "epoch": 0.772196809416397, "grad_norm": 0.7920522689819336, "learning_rate": 0.00019435556360633375, "loss": 4.3171, "step": 413800 }, { "epoch": 0.7722901149757756, "grad_norm": 1.0716264247894287, "learning_rate": 0.0001943541720990853, "loss": 4.7698, "step": 413850 }, { "epoch": 0.7723834205351541, "grad_norm": 0.8790009021759033, "learning_rate": 0.0001943527804253183, "loss": 4.5874, "step": 413900 }, { "epoch": 0.7724767260945326, "grad_norm": 1.0554299354553223, "learning_rate": 0.00019435138858503527, "loss": 4.4623, "step": 413950 }, { "epoch": 0.772570031653911, "grad_norm": 0.9873991012573242, "learning_rate": 0.00019434999657823863, "loss": 4.4922, "step": 414000 }, { "epoch": 0.772570031653911, "eval_loss": 4.728475093841553, "eval_runtime": 234.8475, "eval_samples_per_second": 11.105, "eval_steps_per_second": 11.105, "eval_tts_loss": 7.568875565229304, "step": 414000 }, { "epoch": 0.7726633372132895, "grad_norm": 1.185647964477539, "learning_rate": 0.00019434860440493084, "loss": 4.5529, "step": 414050 }, { "epoch": 0.772756642772668, "grad_norm": 1.039781093597412, "learning_rate": 0.00019434721206511437, "loss": 4.3436, "step": 414100 }, { "epoch": 0.7728499483320465, "grad_norm": 1.0758379697799683, "learning_rate": 0.00019434581955879166, "loss": 4.652, "step": 414150 }, { "epoch": 0.7729432538914249, "grad_norm": 1.0461156368255615, "learning_rate": 0.0001943444268859652, "loss": 4.6459, "step": 414200 }, { "epoch": 0.7730365594508035, "grad_norm": 1.0226956605911255, "learning_rate": 0.00019434303404663745, "loss": 4.3713, "step": 414250 }, { "epoch": 0.773129865010182, "grad_norm": 0.9251855611801147, "learning_rate": 0.0001943416410408108, "loss": 4.6817, "step": 414300 }, { "epoch": 0.7732231705695605, "grad_norm": 0.9334062337875366, "learning_rate": 0.00019434024786848777, "loss": 4.4471, "step": 414350 }, { "epoch": 0.773316476128939, "grad_norm": 1.0316689014434814, "learning_rate": 0.0001943388545296708, "loss": 4.5474, "step": 414400 }, { "epoch": 0.7734097816883174, "grad_norm": 1.1120091676712036, "learning_rate": 0.00019433746102436237, "loss": 4.699, "step": 414450 }, { "epoch": 0.7735030872476959, "grad_norm": 0.9110586047172546, "learning_rate": 0.00019433606735256488, "loss": 4.5102, "step": 414500 }, { "epoch": 0.7735963928070744, "grad_norm": 1.002780795097351, "learning_rate": 0.00019433467351428086, "loss": 4.6266, "step": 414550 }, { "epoch": 0.773689698366453, "grad_norm": 1.3112727403640747, "learning_rate": 0.00019433327950951274, "loss": 4.6851, "step": 414600 }, { "epoch": 0.7737830039258314, "grad_norm": 1.1229636669158936, "learning_rate": 0.00019433188533826297, "loss": 4.6857, "step": 414650 }, { "epoch": 0.7738763094852099, "grad_norm": 0.9074994325637817, "learning_rate": 0.00019433049100053403, "loss": 4.656, "step": 414700 }, { "epoch": 0.7739696150445884, "grad_norm": 0.9077115654945374, "learning_rate": 0.00019432909649632835, "loss": 4.3477, "step": 414750 }, { "epoch": 0.7740629206039669, "grad_norm": 0.8067662119865417, "learning_rate": 0.00019432770182564843, "loss": 4.5333, "step": 414800 }, { "epoch": 0.7741562261633453, "grad_norm": 1.110912799835205, "learning_rate": 0.00019432630698849673, "loss": 4.6428, "step": 414850 }, { "epoch": 0.7742495317227238, "grad_norm": 1.0513705015182495, "learning_rate": 0.00019432491198487567, "loss": 4.5769, "step": 414900 }, { "epoch": 0.7743428372821024, "grad_norm": 1.2311010360717773, "learning_rate": 0.00019432351681478773, "loss": 4.7393, "step": 414950 }, { "epoch": 0.7744361428414809, "grad_norm": 1.2063549757003784, "learning_rate": 0.0001943221214782354, "loss": 4.3179, "step": 415000 }, { "epoch": 0.7745294484008594, "grad_norm": 1.0355026721954346, "learning_rate": 0.0001943207259752211, "loss": 4.4265, "step": 415050 }, { "epoch": 0.7746227539602378, "grad_norm": 0.9668757915496826, "learning_rate": 0.00019431933030574733, "loss": 4.5367, "step": 415100 }, { "epoch": 0.7747160595196163, "grad_norm": 1.165673017501831, "learning_rate": 0.00019431793446981655, "loss": 4.5728, "step": 415150 }, { "epoch": 0.7748093650789948, "grad_norm": 0.811471700668335, "learning_rate": 0.00019431653846743122, "loss": 4.5724, "step": 415200 }, { "epoch": 0.7749026706383733, "grad_norm": 1.1002155542373657, "learning_rate": 0.00019431514229859375, "loss": 4.4974, "step": 415250 }, { "epoch": 0.7749959761977518, "grad_norm": 0.9775769710540771, "learning_rate": 0.00019431374596330666, "loss": 4.135, "step": 415300 }, { "epoch": 0.7750892817571303, "grad_norm": 0.9746343493461609, "learning_rate": 0.00019431234946157245, "loss": 4.503, "step": 415350 }, { "epoch": 0.7751825873165088, "grad_norm": 1.0700674057006836, "learning_rate": 0.00019431095279339347, "loss": 4.6995, "step": 415400 }, { "epoch": 0.7752758928758873, "grad_norm": 1.1561682224273682, "learning_rate": 0.0001943095559587723, "loss": 4.4893, "step": 415450 }, { "epoch": 0.7753691984352657, "grad_norm": 0.9776704907417297, "learning_rate": 0.00019430815895771136, "loss": 4.368, "step": 415500 }, { "epoch": 0.7754625039946442, "grad_norm": 1.591834306716919, "learning_rate": 0.0001943067617902131, "loss": 4.7457, "step": 415550 }, { "epoch": 0.7755558095540227, "grad_norm": 1.0882806777954102, "learning_rate": 0.00019430536445628, "loss": 4.618, "step": 415600 }, { "epoch": 0.7756491151134013, "grad_norm": 0.9813977479934692, "learning_rate": 0.00019430396695591454, "loss": 4.6303, "step": 415650 }, { "epoch": 0.7757424206727798, "grad_norm": 1.1258888244628906, "learning_rate": 0.0001943025692891191, "loss": 4.4641, "step": 415700 }, { "epoch": 0.7758357262321582, "grad_norm": 1.2065447568893433, "learning_rate": 0.0001943011714558963, "loss": 4.573, "step": 415750 }, { "epoch": 0.7759290317915367, "grad_norm": 0.7935813665390015, "learning_rate": 0.0001942997734562485, "loss": 4.7439, "step": 415800 }, { "epoch": 0.7760223373509152, "grad_norm": 0.7973695993423462, "learning_rate": 0.00019429837529017817, "loss": 4.5464, "step": 415850 }, { "epoch": 0.7761156429102937, "grad_norm": 1.237538456916809, "learning_rate": 0.00019429697695768784, "loss": 4.5519, "step": 415900 }, { "epoch": 0.7762089484696721, "grad_norm": 1.0046755075454712, "learning_rate": 0.00019429557845877994, "loss": 4.584, "step": 415950 }, { "epoch": 0.7763022540290507, "grad_norm": 1.007865071296692, "learning_rate": 0.0001942941797934569, "loss": 4.5836, "step": 416000 }, { "epoch": 0.7763955595884292, "grad_norm": 1.0284510850906372, "learning_rate": 0.00019429278096172125, "loss": 4.5441, "step": 416050 }, { "epoch": 0.7764888651478077, "grad_norm": 1.0060741901397705, "learning_rate": 0.0001942913819635754, "loss": 4.4357, "step": 416100 }, { "epoch": 0.7765821707071862, "grad_norm": 1.2664929628372192, "learning_rate": 0.00019428998279902188, "loss": 4.3837, "step": 416150 }, { "epoch": 0.7766754762665646, "grad_norm": 1.3339954614639282, "learning_rate": 0.00019428858346806312, "loss": 4.5083, "step": 416200 }, { "epoch": 0.7767687818259431, "grad_norm": 1.8308089971542358, "learning_rate": 0.00019428718397070162, "loss": 4.6836, "step": 416250 }, { "epoch": 0.7768620873853216, "grad_norm": 1.1787583827972412, "learning_rate": 0.0001942857843069398, "loss": 4.4676, "step": 416300 }, { "epoch": 0.7769553929447002, "grad_norm": 1.283090591430664, "learning_rate": 0.0001942843844767802, "loss": 4.5517, "step": 416350 }, { "epoch": 0.7770486985040786, "grad_norm": 0.8927239179611206, "learning_rate": 0.00019428298448022523, "loss": 4.7133, "step": 416400 }, { "epoch": 0.7771420040634571, "grad_norm": 1.2088508605957031, "learning_rate": 0.00019428158431727738, "loss": 4.5483, "step": 416450 }, { "epoch": 0.7772353096228356, "grad_norm": 0.8650713562965393, "learning_rate": 0.00019428018398793913, "loss": 4.6464, "step": 416500 }, { "epoch": 0.7773286151822141, "grad_norm": 0.8104282021522522, "learning_rate": 0.00019427878349221293, "loss": 4.2484, "step": 416550 }, { "epoch": 0.7774219207415926, "grad_norm": 1.0021339654922485, "learning_rate": 0.00019427738283010128, "loss": 4.5643, "step": 416600 }, { "epoch": 0.777515226300971, "grad_norm": 1.0103039741516113, "learning_rate": 0.00019427598200160662, "loss": 4.4514, "step": 416650 }, { "epoch": 0.7776085318603496, "grad_norm": 1.0550456047058105, "learning_rate": 0.00019427458100673147, "loss": 4.4182, "step": 416700 }, { "epoch": 0.7777018374197281, "grad_norm": 0.9796375036239624, "learning_rate": 0.00019427317984547823, "loss": 4.7182, "step": 416750 }, { "epoch": 0.7777951429791066, "grad_norm": 0.9060117602348328, "learning_rate": 0.00019427177851784943, "loss": 4.7177, "step": 416800 }, { "epoch": 0.777888448538485, "grad_norm": 1.0749459266662598, "learning_rate": 0.00019427037702384755, "loss": 4.6499, "step": 416850 }, { "epoch": 0.7779817540978635, "grad_norm": 1.1595515012741089, "learning_rate": 0.00019426897536347504, "loss": 4.6817, "step": 416900 }, { "epoch": 0.778075059657242, "grad_norm": 1.2265338897705078, "learning_rate": 0.00019426757353673436, "loss": 4.6216, "step": 416950 }, { "epoch": 0.7781683652166205, "grad_norm": 0.9206924438476562, "learning_rate": 0.000194266171543628, "loss": 4.4135, "step": 417000 }, { "epoch": 0.7781683652166205, "eval_loss": 4.7275614738464355, "eval_runtime": 233.8432, "eval_samples_per_second": 11.153, "eval_steps_per_second": 11.153, "eval_tts_loss": 7.560000432109563, "step": 417000 }, { "epoch": 0.7782616707759991, "grad_norm": 1.1508795022964478, "learning_rate": 0.00019426476938415844, "loss": 4.4384, "step": 417050 }, { "epoch": 0.7783549763353775, "grad_norm": 1.052635669708252, "learning_rate": 0.00019426336705832817, "loss": 4.5719, "step": 417100 }, { "epoch": 0.778448281894756, "grad_norm": 1.231458306312561, "learning_rate": 0.0001942619645661396, "loss": 4.7393, "step": 417150 }, { "epoch": 0.7785415874541345, "grad_norm": 0.6703341603279114, "learning_rate": 0.00019426056190759527, "loss": 4.3872, "step": 417200 }, { "epoch": 0.778634893013513, "grad_norm": 1.0658252239227295, "learning_rate": 0.00019425915908269766, "loss": 4.4999, "step": 417250 }, { "epoch": 0.7787281985728914, "grad_norm": 1.0017874240875244, "learning_rate": 0.0001942577560914492, "loss": 4.766, "step": 417300 }, { "epoch": 0.7788215041322699, "grad_norm": 1.4141888618469238, "learning_rate": 0.00019425635293385236, "loss": 4.4379, "step": 417350 }, { "epoch": 0.7789148096916485, "grad_norm": 1.4482828378677368, "learning_rate": 0.00019425494960990965, "loss": 4.5795, "step": 417400 }, { "epoch": 0.779008115251027, "grad_norm": 1.072346568107605, "learning_rate": 0.00019425354611962353, "loss": 4.4124, "step": 417450 }, { "epoch": 0.7791014208104055, "grad_norm": 1.0275367498397827, "learning_rate": 0.00019425214246299651, "loss": 4.6065, "step": 417500 }, { "epoch": 0.7791947263697839, "grad_norm": 1.1461256742477417, "learning_rate": 0.00019425073864003106, "loss": 4.5195, "step": 417550 }, { "epoch": 0.7792880319291624, "grad_norm": 1.0955123901367188, "learning_rate": 0.00019424933465072962, "loss": 4.6588, "step": 417600 }, { "epoch": 0.7793813374885409, "grad_norm": 1.1902095079421997, "learning_rate": 0.0001942479304950947, "loss": 4.5148, "step": 417650 }, { "epoch": 0.7794746430479194, "grad_norm": 1.45046865940094, "learning_rate": 0.00019424652617312872, "loss": 4.676, "step": 417700 }, { "epoch": 0.779567948607298, "grad_norm": 1.0053950548171997, "learning_rate": 0.00019424512168483425, "loss": 4.5951, "step": 417750 }, { "epoch": 0.7796612541666764, "grad_norm": 1.1858372688293457, "learning_rate": 0.00019424371703021372, "loss": 4.4666, "step": 417800 }, { "epoch": 0.7797545597260549, "grad_norm": 1.2483307123184204, "learning_rate": 0.00019424231220926957, "loss": 4.4132, "step": 417850 }, { "epoch": 0.7798478652854334, "grad_norm": 1.159818410873413, "learning_rate": 0.0001942409072220044, "loss": 4.296, "step": 417900 }, { "epoch": 0.7799411708448118, "grad_norm": 0.9254298806190491, "learning_rate": 0.00019423950206842055, "loss": 4.5937, "step": 417950 }, { "epoch": 0.7800344764041903, "grad_norm": 0.9275083541870117, "learning_rate": 0.00019423809674852058, "loss": 4.5276, "step": 418000 }, { "epoch": 0.7801277819635688, "grad_norm": 0.9060862064361572, "learning_rate": 0.00019423669126230695, "loss": 4.6078, "step": 418050 }, { "epoch": 0.7802210875229474, "grad_norm": 0.9902940988540649, "learning_rate": 0.00019423528560978212, "loss": 4.6819, "step": 418100 }, { "epoch": 0.7803143930823259, "grad_norm": 0.9466397166252136, "learning_rate": 0.0001942338797909486, "loss": 4.578, "step": 418150 }, { "epoch": 0.7804076986417043, "grad_norm": 1.061043381690979, "learning_rate": 0.0001942324738058089, "loss": 4.6206, "step": 418200 }, { "epoch": 0.7805010042010828, "grad_norm": 0.9925414323806763, "learning_rate": 0.00019423106765436543, "loss": 4.4128, "step": 418250 }, { "epoch": 0.7805943097604613, "grad_norm": 1.138055443763733, "learning_rate": 0.0001942296613366207, "loss": 4.5836, "step": 418300 }, { "epoch": 0.7806876153198398, "grad_norm": 0.9431651830673218, "learning_rate": 0.0001942282548525772, "loss": 4.488, "step": 418350 }, { "epoch": 0.7807809208792182, "grad_norm": 1.1069378852844238, "learning_rate": 0.00019422684820223743, "loss": 4.3909, "step": 418400 }, { "epoch": 0.7808742264385968, "grad_norm": 1.473007321357727, "learning_rate": 0.00019422544138560387, "loss": 4.7102, "step": 418450 }, { "epoch": 0.7809675319979753, "grad_norm": 1.063603162765503, "learning_rate": 0.00019422403440267892, "loss": 4.4908, "step": 418500 }, { "epoch": 0.7810608375573538, "grad_norm": 0.9246348142623901, "learning_rate": 0.0001942226272534652, "loss": 4.5294, "step": 418550 }, { "epoch": 0.7811541431167323, "grad_norm": 1.083772897720337, "learning_rate": 0.00019422121993796507, "loss": 4.6143, "step": 418600 }, { "epoch": 0.7812474486761107, "grad_norm": 1.3056730031967163, "learning_rate": 0.0001942198124561811, "loss": 4.6124, "step": 418650 }, { "epoch": 0.7813407542354892, "grad_norm": 0.9770277142524719, "learning_rate": 0.00019421840480811573, "loss": 4.3734, "step": 418700 }, { "epoch": 0.7814340597948677, "grad_norm": 1.2718496322631836, "learning_rate": 0.00019421699699377146, "loss": 4.6823, "step": 418750 }, { "epoch": 0.7815273653542463, "grad_norm": 0.8844640254974365, "learning_rate": 0.00019421558901315072, "loss": 4.5507, "step": 418800 }, { "epoch": 0.7816206709136247, "grad_norm": 0.9227458834648132, "learning_rate": 0.0001942141808662561, "loss": 4.3661, "step": 418850 }, { "epoch": 0.7817139764730032, "grad_norm": 1.2249149084091187, "learning_rate": 0.00019421277255309, "loss": 4.4293, "step": 418900 }, { "epoch": 0.7818072820323817, "grad_norm": 1.1062428951263428, "learning_rate": 0.00019421136407365497, "loss": 4.6696, "step": 418950 }, { "epoch": 0.7819005875917602, "grad_norm": 0.9474209547042847, "learning_rate": 0.00019420995542795342, "loss": 4.5205, "step": 419000 }, { "epoch": 0.7819938931511387, "grad_norm": 1.0487412214279175, "learning_rate": 0.00019420854661598787, "loss": 4.5216, "step": 419050 }, { "epoch": 0.7820871987105171, "grad_norm": 0.9993675947189331, "learning_rate": 0.00019420713763776084, "loss": 4.6082, "step": 419100 }, { "epoch": 0.7821805042698957, "grad_norm": 0.8808112740516663, "learning_rate": 0.0001942057284932748, "loss": 4.505, "step": 419150 }, { "epoch": 0.7822738098292742, "grad_norm": 1.1803745031356812, "learning_rate": 0.0001942043191825322, "loss": 4.5235, "step": 419200 }, { "epoch": 0.7823671153886527, "grad_norm": 0.852111279964447, "learning_rate": 0.00019420290970553552, "loss": 4.6029, "step": 419250 }, { "epoch": 0.7824604209480311, "grad_norm": 1.1742595434188843, "learning_rate": 0.00019420150006228733, "loss": 4.6342, "step": 419300 }, { "epoch": 0.7825537265074096, "grad_norm": 0.799828290939331, "learning_rate": 0.00019420009025279005, "loss": 4.6569, "step": 419350 }, { "epoch": 0.7826470320667881, "grad_norm": 1.0880378484725952, "learning_rate": 0.00019419868027704617, "loss": 4.4393, "step": 419400 }, { "epoch": 0.7827403376261666, "grad_norm": 1.1590566635131836, "learning_rate": 0.00019419727013505824, "loss": 4.6774, "step": 419450 }, { "epoch": 0.782833643185545, "grad_norm": 0.9538164138793945, "learning_rate": 0.00019419585982682867, "loss": 4.5779, "step": 419500 }, { "epoch": 0.7829269487449236, "grad_norm": 1.166024923324585, "learning_rate": 0.00019419444935235997, "loss": 4.6071, "step": 419550 }, { "epoch": 0.7830202543043021, "grad_norm": 1.2705340385437012, "learning_rate": 0.00019419303871165465, "loss": 4.5342, "step": 419600 }, { "epoch": 0.7831135598636806, "grad_norm": 1.2403308153152466, "learning_rate": 0.0001941916279047152, "loss": 4.4299, "step": 419650 }, { "epoch": 0.7832068654230591, "grad_norm": 0.6722421050071716, "learning_rate": 0.0001941902169315441, "loss": 4.7047, "step": 419700 }, { "epoch": 0.7833001709824375, "grad_norm": 1.151554822921753, "learning_rate": 0.00019418880579214384, "loss": 4.6488, "step": 419750 }, { "epoch": 0.783393476541816, "grad_norm": 0.9347752928733826, "learning_rate": 0.00019418739448651693, "loss": 4.3575, "step": 419800 }, { "epoch": 0.7834867821011945, "grad_norm": 1.110811471939087, "learning_rate": 0.00019418598301466583, "loss": 4.5186, "step": 419850 }, { "epoch": 0.7835800876605731, "grad_norm": 1.1932270526885986, "learning_rate": 0.00019418457137659304, "loss": 4.6109, "step": 419900 }, { "epoch": 0.7836733932199516, "grad_norm": 1.063226342201233, "learning_rate": 0.00019418315957230102, "loss": 4.4918, "step": 419950 }, { "epoch": 0.78376669877933, "grad_norm": 1.0015634298324585, "learning_rate": 0.00019418174760179232, "loss": 4.5527, "step": 420000 }, { "epoch": 0.78376669877933, "eval_loss": 4.721020221710205, "eval_runtime": 233.7819, "eval_samples_per_second": 11.156, "eval_steps_per_second": 11.156, "eval_tts_loss": 7.572293719996704, "step": 420000 }, { "epoch": 0.7838600043387085, "grad_norm": 1.0639441013336182, "learning_rate": 0.00019418033546506942, "loss": 4.7157, "step": 420050 }, { "epoch": 0.783953309898087, "grad_norm": 1.0074838399887085, "learning_rate": 0.0001941789231621348, "loss": 4.7882, "step": 420100 }, { "epoch": 0.7840466154574655, "grad_norm": 0.9637343883514404, "learning_rate": 0.00019417751069299093, "loss": 4.5151, "step": 420150 }, { "epoch": 0.7841399210168439, "grad_norm": 0.9664762020111084, "learning_rate": 0.00019417609805764033, "loss": 4.7028, "step": 420200 }, { "epoch": 0.7842332265762225, "grad_norm": 1.0026178359985352, "learning_rate": 0.00019417468525608552, "loss": 4.5529, "step": 420250 }, { "epoch": 0.784326532135601, "grad_norm": 0.8280647993087769, "learning_rate": 0.00019417327228832891, "loss": 4.6032, "step": 420300 }, { "epoch": 0.7844198376949795, "grad_norm": 1.0198395252227783, "learning_rate": 0.0001941718591543731, "loss": 4.5357, "step": 420350 }, { "epoch": 0.784513143254358, "grad_norm": 1.0759913921356201, "learning_rate": 0.00019417044585422048, "loss": 4.6143, "step": 420400 }, { "epoch": 0.7846064488137364, "grad_norm": 1.1430230140686035, "learning_rate": 0.0001941690323878736, "loss": 4.6088, "step": 420450 }, { "epoch": 0.7846997543731149, "grad_norm": 1.0650012493133545, "learning_rate": 0.00019416761875533495, "loss": 4.6748, "step": 420500 }, { "epoch": 0.7847930599324934, "grad_norm": 0.8512048721313477, "learning_rate": 0.00019416620495660704, "loss": 4.5298, "step": 420550 }, { "epoch": 0.784886365491872, "grad_norm": 1.2103396654129028, "learning_rate": 0.00019416479099169235, "loss": 4.6461, "step": 420600 }, { "epoch": 0.7849796710512504, "grad_norm": 0.8439053893089294, "learning_rate": 0.00019416337686059337, "loss": 4.4296, "step": 420650 }, { "epoch": 0.7850729766106289, "grad_norm": 1.0167698860168457, "learning_rate": 0.00019416196256331257, "loss": 4.6308, "step": 420700 }, { "epoch": 0.7851662821700074, "grad_norm": 0.9046244025230408, "learning_rate": 0.0001941605480998525, "loss": 4.4991, "step": 420750 }, { "epoch": 0.7852595877293859, "grad_norm": 0.9066252708435059, "learning_rate": 0.0001941591334702156, "loss": 4.6585, "step": 420800 }, { "epoch": 0.7853528932887643, "grad_norm": 1.1947590112686157, "learning_rate": 0.00019415771867440444, "loss": 4.6536, "step": 420850 }, { "epoch": 0.7854461988481428, "grad_norm": 0.8964353799819946, "learning_rate": 0.00019415630371242144, "loss": 4.5808, "step": 420900 }, { "epoch": 0.7855395044075214, "grad_norm": 1.4293516874313354, "learning_rate": 0.00019415488858426917, "loss": 4.4272, "step": 420950 }, { "epoch": 0.7856328099668999, "grad_norm": 1.0083626508712769, "learning_rate": 0.00019415347328995006, "loss": 4.4853, "step": 421000 }, { "epoch": 0.7857261155262784, "grad_norm": 1.1188188791275024, "learning_rate": 0.00019415205782946665, "loss": 4.3596, "step": 421050 }, { "epoch": 0.7858194210856568, "grad_norm": 1.024791955947876, "learning_rate": 0.0001941506422028214, "loss": 4.5746, "step": 421100 }, { "epoch": 0.7859127266450353, "grad_norm": 1.1521062850952148, "learning_rate": 0.00019414922641001686, "loss": 4.5167, "step": 421150 }, { "epoch": 0.7860060322044138, "grad_norm": 1.025851845741272, "learning_rate": 0.00019414781045105548, "loss": 4.4505, "step": 421200 }, { "epoch": 0.7860993377637923, "grad_norm": 0.8077706694602966, "learning_rate": 0.0001941463943259398, "loss": 4.4218, "step": 421250 }, { "epoch": 0.7861926433231708, "grad_norm": 1.1391069889068604, "learning_rate": 0.00019414497803467228, "loss": 4.7806, "step": 421300 }, { "epoch": 0.7862859488825493, "grad_norm": 0.9531813859939575, "learning_rate": 0.00019414356157725546, "loss": 4.4662, "step": 421350 }, { "epoch": 0.7863792544419278, "grad_norm": 0.9496504664421082, "learning_rate": 0.0001941421449536918, "loss": 4.4881, "step": 421400 }, { "epoch": 0.7864725600013063, "grad_norm": 1.1776880025863647, "learning_rate": 0.00019414072816398384, "loss": 4.5501, "step": 421450 }, { "epoch": 0.7865658655606848, "grad_norm": 1.0474661588668823, "learning_rate": 0.00019413931120813404, "loss": 4.6452, "step": 421500 }, { "epoch": 0.7866591711200632, "grad_norm": 1.3503599166870117, "learning_rate": 0.00019413789408614492, "loss": 4.7286, "step": 421550 }, { "epoch": 0.7867524766794417, "grad_norm": 1.51902174949646, "learning_rate": 0.00019413647679801898, "loss": 4.4683, "step": 421600 }, { "epoch": 0.7868457822388203, "grad_norm": 1.1486494541168213, "learning_rate": 0.00019413505934375876, "loss": 4.4698, "step": 421650 }, { "epoch": 0.7869390877981988, "grad_norm": 1.0835721492767334, "learning_rate": 0.00019413364172336667, "loss": 4.4485, "step": 421700 }, { "epoch": 0.7870323933575772, "grad_norm": 1.2232539653778076, "learning_rate": 0.0001941322239368453, "loss": 4.7299, "step": 421750 }, { "epoch": 0.7871256989169557, "grad_norm": 1.496379017829895, "learning_rate": 0.0001941308059841971, "loss": 4.5478, "step": 421800 }, { "epoch": 0.7872190044763342, "grad_norm": 0.9647490382194519, "learning_rate": 0.0001941293878654246, "loss": 4.6043, "step": 421850 }, { "epoch": 0.7873123100357127, "grad_norm": 1.1284325122833252, "learning_rate": 0.0001941279695805303, "loss": 4.5567, "step": 421900 }, { "epoch": 0.7874056155950911, "grad_norm": 1.2496758699417114, "learning_rate": 0.00019412655112951668, "loss": 4.4419, "step": 421950 }, { "epoch": 0.7874989211544697, "grad_norm": 1.2127939462661743, "learning_rate": 0.00019412513251238625, "loss": 4.5831, "step": 422000 }, { "epoch": 0.7875922267138482, "grad_norm": 1.3008806705474854, "learning_rate": 0.00019412371372914155, "loss": 4.582, "step": 422050 }, { "epoch": 0.7876855322732267, "grad_norm": 1.2855007648468018, "learning_rate": 0.00019412229477978505, "loss": 4.5988, "step": 422100 }, { "epoch": 0.7877788378326052, "grad_norm": 1.103704571723938, "learning_rate": 0.00019412087566431926, "loss": 4.7151, "step": 422150 }, { "epoch": 0.7878721433919836, "grad_norm": 0.8745959997177124, "learning_rate": 0.0001941194563827467, "loss": 4.6391, "step": 422200 }, { "epoch": 0.7879654489513621, "grad_norm": 2.252020835876465, "learning_rate": 0.00019411803693506982, "loss": 4.7215, "step": 422250 }, { "epoch": 0.7880587545107406, "grad_norm": 1.0912048816680908, "learning_rate": 0.00019411661732129118, "loss": 4.7557, "step": 422300 }, { "epoch": 0.7881520600701192, "grad_norm": 1.004123568534851, "learning_rate": 0.00019411519754141326, "loss": 4.4193, "step": 422350 }, { "epoch": 0.7882453656294977, "grad_norm": 1.1392632722854614, "learning_rate": 0.0001941137775954386, "loss": 4.379, "step": 422400 }, { "epoch": 0.7883386711888761, "grad_norm": 1.1152008771896362, "learning_rate": 0.00019411235748336968, "loss": 4.6683, "step": 422450 }, { "epoch": 0.7884319767482546, "grad_norm": 1.1675337553024292, "learning_rate": 0.000194110937205209, "loss": 4.6349, "step": 422500 }, { "epoch": 0.7885252823076331, "grad_norm": 1.2631796598434448, "learning_rate": 0.0001941095167609591, "loss": 4.4409, "step": 422550 }, { "epoch": 0.7886185878670116, "grad_norm": 1.056110143661499, "learning_rate": 0.00019410809615062245, "loss": 4.5492, "step": 422600 }, { "epoch": 0.78871189342639, "grad_norm": 1.232704520225525, "learning_rate": 0.00019410667537420157, "loss": 4.7121, "step": 422650 }, { "epoch": 0.7888051989857686, "grad_norm": 0.8401553630828857, "learning_rate": 0.00019410525443169896, "loss": 4.3471, "step": 422700 }, { "epoch": 0.7888985045451471, "grad_norm": 1.2476882934570312, "learning_rate": 0.00019410383332311713, "loss": 4.6391, "step": 422750 }, { "epoch": 0.7889918101045256, "grad_norm": 0.9736795425415039, "learning_rate": 0.0001941024120484586, "loss": 4.6042, "step": 422800 }, { "epoch": 0.789085115663904, "grad_norm": 1.1006672382354736, "learning_rate": 0.00019410099060772584, "loss": 4.6395, "step": 422850 }, { "epoch": 0.7891784212232825, "grad_norm": 1.1191246509552002, "learning_rate": 0.00019409956900092143, "loss": 4.5684, "step": 422900 }, { "epoch": 0.789271726782661, "grad_norm": 1.0905956029891968, "learning_rate": 0.0001940981472280478, "loss": 4.5317, "step": 422950 }, { "epoch": 0.7893650323420395, "grad_norm": 0.9905209541320801, "learning_rate": 0.00019409672528910755, "loss": 4.6819, "step": 423000 }, { "epoch": 0.7893650323420395, "eval_loss": 4.721234321594238, "eval_runtime": 234.0422, "eval_samples_per_second": 11.143, "eval_steps_per_second": 11.143, "eval_tts_loss": 7.554684614714098, "step": 423000 }, { "epoch": 0.7894583379014181, "grad_norm": 1.1053640842437744, "learning_rate": 0.00019409530318410313, "loss": 4.7404, "step": 423050 }, { "epoch": 0.7895516434607965, "grad_norm": 1.2071901559829712, "learning_rate": 0.000194093880913037, "loss": 4.4602, "step": 423100 }, { "epoch": 0.789644949020175, "grad_norm": 0.8920719027519226, "learning_rate": 0.00019409245847591176, "loss": 4.4808, "step": 423150 }, { "epoch": 0.7897382545795535, "grad_norm": 1.0834848880767822, "learning_rate": 0.00019409103587272989, "loss": 4.3641, "step": 423200 }, { "epoch": 0.789831560138932, "grad_norm": 1.171592116355896, "learning_rate": 0.00019408961310349388, "loss": 4.7109, "step": 423250 }, { "epoch": 0.7899248656983104, "grad_norm": 1.1035345792770386, "learning_rate": 0.0001940881901682063, "loss": 4.5078, "step": 423300 }, { "epoch": 0.7900181712576889, "grad_norm": 1.0811997652053833, "learning_rate": 0.0001940867670668696, "loss": 4.5882, "step": 423350 }, { "epoch": 0.7901114768170675, "grad_norm": 1.1453145742416382, "learning_rate": 0.00019408534379948629, "loss": 4.6713, "step": 423400 }, { "epoch": 0.790204782376446, "grad_norm": 0.9412307739257812, "learning_rate": 0.00019408392036605892, "loss": 4.5075, "step": 423450 }, { "epoch": 0.7902980879358245, "grad_norm": 1.229821801185608, "learning_rate": 0.00019408249676658998, "loss": 4.507, "step": 423500 }, { "epoch": 0.7903913934952029, "grad_norm": 1.3128808736801147, "learning_rate": 0.000194081073001082, "loss": 4.5877, "step": 423550 }, { "epoch": 0.7904846990545814, "grad_norm": 1.1726655960083008, "learning_rate": 0.00019407964906953745, "loss": 4.5936, "step": 423600 }, { "epoch": 0.7905780046139599, "grad_norm": 1.0003212690353394, "learning_rate": 0.0001940782249719589, "loss": 4.4855, "step": 423650 }, { "epoch": 0.7906713101733384, "grad_norm": 1.2849147319793701, "learning_rate": 0.00019407680070834883, "loss": 4.7337, "step": 423700 }, { "epoch": 0.790764615732717, "grad_norm": 1.2079015970230103, "learning_rate": 0.00019407537627870976, "loss": 4.4842, "step": 423750 }, { "epoch": 0.7908579212920954, "grad_norm": 1.1246628761291504, "learning_rate": 0.0001940739516830442, "loss": 4.4204, "step": 423800 }, { "epoch": 0.7909512268514739, "grad_norm": 0.8938810229301453, "learning_rate": 0.00019407252692135466, "loss": 4.5199, "step": 423850 }, { "epoch": 0.7910445324108524, "grad_norm": 1.0785510540008545, "learning_rate": 0.00019407110199364368, "loss": 4.4074, "step": 423900 }, { "epoch": 0.7911378379702309, "grad_norm": 0.9102846384048462, "learning_rate": 0.00019406967689991378, "loss": 4.42, "step": 423950 }, { "epoch": 0.7912311435296093, "grad_norm": 0.9787639379501343, "learning_rate": 0.00019406825164016743, "loss": 4.5467, "step": 424000 }, { "epoch": 0.7913244490889878, "grad_norm": 1.0713123083114624, "learning_rate": 0.00019406682621440715, "loss": 4.7614, "step": 424050 }, { "epoch": 0.7914177546483664, "grad_norm": 0.9125092625617981, "learning_rate": 0.00019406540062263548, "loss": 4.6056, "step": 424100 }, { "epoch": 0.7915110602077449, "grad_norm": 1.1230067014694214, "learning_rate": 0.00019406397486485493, "loss": 4.499, "step": 424150 }, { "epoch": 0.7916043657671233, "grad_norm": 0.9560022354125977, "learning_rate": 0.00019406254894106803, "loss": 4.3798, "step": 424200 }, { "epoch": 0.7916976713265018, "grad_norm": 0.8355482816696167, "learning_rate": 0.00019406112285127726, "loss": 4.3809, "step": 424250 }, { "epoch": 0.7917909768858803, "grad_norm": 1.1256998777389526, "learning_rate": 0.00019405969659548519, "loss": 4.7232, "step": 424300 }, { "epoch": 0.7918842824452588, "grad_norm": 1.2029259204864502, "learning_rate": 0.00019405827017369428, "loss": 4.7615, "step": 424350 }, { "epoch": 0.7919775880046372, "grad_norm": 1.0264108180999756, "learning_rate": 0.00019405684358590706, "loss": 4.4148, "step": 424400 }, { "epoch": 0.7920708935640158, "grad_norm": 0.945449709892273, "learning_rate": 0.00019405541683212608, "loss": 4.569, "step": 424450 }, { "epoch": 0.7921641991233943, "grad_norm": 1.1476784944534302, "learning_rate": 0.00019405398991235385, "loss": 4.6362, "step": 424500 }, { "epoch": 0.7922575046827728, "grad_norm": 1.0508506298065186, "learning_rate": 0.00019405256282659287, "loss": 4.7232, "step": 424550 }, { "epoch": 0.7923508102421513, "grad_norm": 0.9733095765113831, "learning_rate": 0.00019405113557484564, "loss": 4.5333, "step": 424600 }, { "epoch": 0.7924441158015297, "grad_norm": 1.1830148696899414, "learning_rate": 0.00019404970815711475, "loss": 4.5946, "step": 424650 }, { "epoch": 0.7925374213609082, "grad_norm": 0.968312680721283, "learning_rate": 0.00019404828057340264, "loss": 4.3865, "step": 424700 }, { "epoch": 0.7926307269202867, "grad_norm": 1.0268644094467163, "learning_rate": 0.00019404685282371187, "loss": 4.6542, "step": 424750 }, { "epoch": 0.7927240324796653, "grad_norm": 0.8817425966262817, "learning_rate": 0.00019404542490804495, "loss": 4.4903, "step": 424800 }, { "epoch": 0.7928173380390438, "grad_norm": 1.0740628242492676, "learning_rate": 0.0001940439968264044, "loss": 4.7789, "step": 424850 }, { "epoch": 0.7929106435984222, "grad_norm": 1.2251063585281372, "learning_rate": 0.00019404256857879278, "loss": 4.6793, "step": 424900 }, { "epoch": 0.7930039491578007, "grad_norm": 1.330265998840332, "learning_rate": 0.00019404114016521255, "loss": 4.6689, "step": 424950 }, { "epoch": 0.7930972547171792, "grad_norm": 1.1080074310302734, "learning_rate": 0.00019403971158566625, "loss": 4.5088, "step": 425000 }, { "epoch": 0.7931905602765577, "grad_norm": 1.2231508493423462, "learning_rate": 0.00019403828284015638, "loss": 4.38, "step": 425050 }, { "epoch": 0.7932838658359361, "grad_norm": 1.0425812005996704, "learning_rate": 0.00019403685392868552, "loss": 4.5114, "step": 425100 }, { "epoch": 0.7933771713953146, "grad_norm": 0.8140989542007446, "learning_rate": 0.00019403542485125615, "loss": 4.6362, "step": 425150 }, { "epoch": 0.7934704769546932, "grad_norm": 1.0934182405471802, "learning_rate": 0.0001940339956078708, "loss": 4.3847, "step": 425200 }, { "epoch": 0.7935637825140717, "grad_norm": 1.2771217823028564, "learning_rate": 0.000194032566198532, "loss": 4.5295, "step": 425250 }, { "epoch": 0.7936570880734501, "grad_norm": 1.1933695077896118, "learning_rate": 0.0001940311366232423, "loss": 4.687, "step": 425300 }, { "epoch": 0.7937503936328286, "grad_norm": 1.0973583459854126, "learning_rate": 0.00019402970688200413, "loss": 4.6445, "step": 425350 }, { "epoch": 0.7938436991922071, "grad_norm": 1.0848637819290161, "learning_rate": 0.0001940282769748201, "loss": 4.4759, "step": 425400 }, { "epoch": 0.7939370047515856, "grad_norm": 0.9990158081054688, "learning_rate": 0.00019402684690169268, "loss": 4.5484, "step": 425450 }, { "epoch": 0.794030310310964, "grad_norm": 0.8308629989624023, "learning_rate": 0.00019402541666262445, "loss": 4.3825, "step": 425500 }, { "epoch": 0.7941236158703426, "grad_norm": 1.1615198850631714, "learning_rate": 0.0001940239862576179, "loss": 4.6229, "step": 425550 }, { "epoch": 0.7942169214297211, "grad_norm": 1.0980902910232544, "learning_rate": 0.00019402255568667557, "loss": 4.5118, "step": 425600 }, { "epoch": 0.7943102269890996, "grad_norm": 1.2873594760894775, "learning_rate": 0.00019402112494979994, "loss": 4.404, "step": 425650 }, { "epoch": 0.7944035325484781, "grad_norm": 0.7118760347366333, "learning_rate": 0.00019401969404699357, "loss": 4.3621, "step": 425700 }, { "epoch": 0.7944968381078565, "grad_norm": 0.9714242219924927, "learning_rate": 0.00019401826297825903, "loss": 4.5299, "step": 425750 }, { "epoch": 0.794590143667235, "grad_norm": 1.098937749862671, "learning_rate": 0.00019401683174359875, "loss": 4.6598, "step": 425800 }, { "epoch": 0.7946834492266135, "grad_norm": 0.9259284734725952, "learning_rate": 0.00019401540034301532, "loss": 4.2878, "step": 425850 }, { "epoch": 0.7947767547859921, "grad_norm": 0.9571802020072937, "learning_rate": 0.00019401396877651123, "loss": 4.5047, "step": 425900 }, { "epoch": 0.7948700603453706, "grad_norm": 1.0199216604232788, "learning_rate": 0.00019401253704408905, "loss": 4.6847, "step": 425950 }, { "epoch": 0.794963365904749, "grad_norm": 1.0073426961898804, "learning_rate": 0.0001940111051457513, "loss": 4.6102, "step": 426000 }, { "epoch": 0.794963365904749, "eval_loss": 4.718998432159424, "eval_runtime": 231.5463, "eval_samples_per_second": 11.263, "eval_steps_per_second": 11.263, "eval_tts_loss": 7.552527759125616, "step": 426000 }, { "epoch": 0.7950566714641275, "grad_norm": 0.927761971950531, "learning_rate": 0.00019400967308150045, "loss": 4.6581, "step": 426050 }, { "epoch": 0.795149977023506, "grad_norm": 1.1682859659194946, "learning_rate": 0.0001940082408513391, "loss": 4.5859, "step": 426100 }, { "epoch": 0.7952432825828845, "grad_norm": 0.9072175621986389, "learning_rate": 0.00019400680845526976, "loss": 4.4219, "step": 426150 }, { "epoch": 0.7953365881422629, "grad_norm": 0.6768325567245483, "learning_rate": 0.00019400537589329492, "loss": 4.5796, "step": 426200 }, { "epoch": 0.7954298937016415, "grad_norm": 1.156038522720337, "learning_rate": 0.00019400394316541712, "loss": 4.7026, "step": 426250 }, { "epoch": 0.79552319926102, "grad_norm": 1.071454405784607, "learning_rate": 0.00019400251027163894, "loss": 4.6118, "step": 426300 }, { "epoch": 0.7956165048203985, "grad_norm": 1.1358624696731567, "learning_rate": 0.00019400107721196283, "loss": 4.511, "step": 426350 }, { "epoch": 0.795709810379777, "grad_norm": 1.1195452213287354, "learning_rate": 0.00019399964398639138, "loss": 4.4004, "step": 426400 }, { "epoch": 0.7958031159391554, "grad_norm": 1.0327094793319702, "learning_rate": 0.0001939982105949271, "loss": 4.6984, "step": 426450 }, { "epoch": 0.7958964214985339, "grad_norm": 1.180101990699768, "learning_rate": 0.00019399677703757251, "loss": 4.701, "step": 426500 }, { "epoch": 0.7959897270579124, "grad_norm": 0.8951889276504517, "learning_rate": 0.00019399534331433015, "loss": 4.5, "step": 426550 }, { "epoch": 0.796083032617291, "grad_norm": 1.1613125801086426, "learning_rate": 0.00019399390942520254, "loss": 4.6163, "step": 426600 }, { "epoch": 0.7961763381766694, "grad_norm": 3.089205503463745, "learning_rate": 0.00019399247537019222, "loss": 4.5153, "step": 426650 }, { "epoch": 0.7962696437360479, "grad_norm": 1.1623356342315674, "learning_rate": 0.00019399104114930173, "loss": 4.5372, "step": 426700 }, { "epoch": 0.7963629492954264, "grad_norm": 0.8393343687057495, "learning_rate": 0.00019398960676253357, "loss": 4.3724, "step": 426750 }, { "epoch": 0.7964562548548049, "grad_norm": 1.1424648761749268, "learning_rate": 0.0001939881722098903, "loss": 4.6238, "step": 426800 }, { "epoch": 0.7965495604141833, "grad_norm": 1.0593382120132446, "learning_rate": 0.00019398673749137447, "loss": 4.5704, "step": 426850 }, { "epoch": 0.7966428659735618, "grad_norm": 0.9022440910339355, "learning_rate": 0.00019398530260698857, "loss": 4.3505, "step": 426900 }, { "epoch": 0.7967361715329404, "grad_norm": 1.0963038206100464, "learning_rate": 0.00019398386755673514, "loss": 4.5951, "step": 426950 }, { "epoch": 0.7968294770923189, "grad_norm": 1.1834969520568848, "learning_rate": 0.00019398243234061674, "loss": 4.453, "step": 427000 }, { "epoch": 0.7969227826516974, "grad_norm": 0.8198856711387634, "learning_rate": 0.00019398099695863588, "loss": 4.5635, "step": 427050 }, { "epoch": 0.7970160882110758, "grad_norm": 0.9053031802177429, "learning_rate": 0.00019397956141079504, "loss": 4.6795, "step": 427100 }, { "epoch": 0.7971093937704543, "grad_norm": 1.189697265625, "learning_rate": 0.00019397812569709691, "loss": 4.7513, "step": 427150 }, { "epoch": 0.7972026993298328, "grad_norm": 0.9244701862335205, "learning_rate": 0.00019397668981754385, "loss": 4.4181, "step": 427200 }, { "epoch": 0.7972960048892113, "grad_norm": 0.9748154878616333, "learning_rate": 0.0001939752537721385, "loss": 4.6368, "step": 427250 }, { "epoch": 0.7973893104485898, "grad_norm": 1.1173096895217896, "learning_rate": 0.00019397381756088334, "loss": 4.6749, "step": 427300 }, { "epoch": 0.7974826160079683, "grad_norm": 0.9567530751228333, "learning_rate": 0.00019397238118378093, "loss": 4.4394, "step": 427350 }, { "epoch": 0.7975759215673468, "grad_norm": 0.9598793387413025, "learning_rate": 0.0001939709446408338, "loss": 4.5758, "step": 427400 }, { "epoch": 0.7976692271267253, "grad_norm": 1.2359800338745117, "learning_rate": 0.00019396950793204451, "loss": 4.4107, "step": 427450 }, { "epoch": 0.7977625326861038, "grad_norm": 1.8980746269226074, "learning_rate": 0.00019396807105741558, "loss": 4.5808, "step": 427500 }, { "epoch": 0.7978558382454822, "grad_norm": 0.9553906917572021, "learning_rate": 0.0001939666340169495, "loss": 4.4524, "step": 427550 }, { "epoch": 0.7979491438048607, "grad_norm": 1.077438235282898, "learning_rate": 0.00019396519681064885, "loss": 4.516, "step": 427600 }, { "epoch": 0.7980424493642393, "grad_norm": 0.6697947382926941, "learning_rate": 0.00019396375943851618, "loss": 4.5095, "step": 427650 }, { "epoch": 0.7981357549236178, "grad_norm": 1.3011325597763062, "learning_rate": 0.000193962321900554, "loss": 4.5906, "step": 427700 }, { "epoch": 0.7982290604829962, "grad_norm": 1.0033787488937378, "learning_rate": 0.00019396088419676487, "loss": 4.5734, "step": 427750 }, { "epoch": 0.7983223660423747, "grad_norm": 1.2484372854232788, "learning_rate": 0.0001939594463271513, "loss": 4.724, "step": 427800 }, { "epoch": 0.7984156716017532, "grad_norm": 1.0030156373977661, "learning_rate": 0.00019395800829171585, "loss": 4.5814, "step": 427850 }, { "epoch": 0.7985089771611317, "grad_norm": 0.8476771116256714, "learning_rate": 0.00019395657009046104, "loss": 4.7281, "step": 427900 }, { "epoch": 0.7986022827205101, "grad_norm": 1.0120536088943481, "learning_rate": 0.0001939551317233894, "loss": 4.5716, "step": 427950 }, { "epoch": 0.7986955882798887, "grad_norm": 1.2245855331420898, "learning_rate": 0.00019395369319050352, "loss": 4.6486, "step": 428000 }, { "epoch": 0.7987888938392672, "grad_norm": 1.0530380010604858, "learning_rate": 0.0001939522544918059, "loss": 4.8216, "step": 428050 }, { "epoch": 0.7988821993986457, "grad_norm": 1.4846291542053223, "learning_rate": 0.00019395081562729904, "loss": 4.6106, "step": 428100 }, { "epoch": 0.7989755049580242, "grad_norm": 1.0235326290130615, "learning_rate": 0.00019394937659698554, "loss": 4.5297, "step": 428150 }, { "epoch": 0.7990688105174026, "grad_norm": 1.066236138343811, "learning_rate": 0.00019394793740086793, "loss": 4.663, "step": 428200 }, { "epoch": 0.7991621160767811, "grad_norm": 0.7847051620483398, "learning_rate": 0.00019394649803894875, "loss": 4.5661, "step": 428250 }, { "epoch": 0.7992554216361596, "grad_norm": 0.9233770370483398, "learning_rate": 0.00019394505851123053, "loss": 4.5384, "step": 428300 }, { "epoch": 0.7993487271955382, "grad_norm": 1.0995171070098877, "learning_rate": 0.0001939436188177158, "loss": 4.6251, "step": 428350 }, { "epoch": 0.7994420327549167, "grad_norm": 1.2371164560317993, "learning_rate": 0.0001939421789584071, "loss": 4.5591, "step": 428400 }, { "epoch": 0.7995353383142951, "grad_norm": 1.12126886844635, "learning_rate": 0.000193940738933307, "loss": 4.5319, "step": 428450 }, { "epoch": 0.7996286438736736, "grad_norm": 1.0614880323410034, "learning_rate": 0.000193939298742418, "loss": 4.6363, "step": 428500 }, { "epoch": 0.7997219494330521, "grad_norm": 0.83480304479599, "learning_rate": 0.0001939378583857427, "loss": 4.6423, "step": 428550 }, { "epoch": 0.7998152549924306, "grad_norm": 1.0235836505889893, "learning_rate": 0.0001939364178632836, "loss": 4.5759, "step": 428600 }, { "epoch": 0.799908560551809, "grad_norm": 1.0326169729232788, "learning_rate": 0.00019393497717504326, "loss": 4.4497, "step": 428650 }, { "epoch": 0.8000018661111876, "grad_norm": 0.876973032951355, "learning_rate": 0.00019393353632102421, "loss": 4.3632, "step": 428700 }, { "epoch": 0.8000951716705661, "grad_norm": 1.080389380455017, "learning_rate": 0.000193932095301229, "loss": 4.6233, "step": 428750 }, { "epoch": 0.8001884772299446, "grad_norm": 0.9304258227348328, "learning_rate": 0.00019393065411566016, "loss": 4.5492, "step": 428800 }, { "epoch": 0.800281782789323, "grad_norm": 1.1061395406723022, "learning_rate": 0.00019392921276432023, "loss": 4.5125, "step": 428850 }, { "epoch": 0.8003750883487015, "grad_norm": 1.3603273630142212, "learning_rate": 0.00019392777124721178, "loss": 4.3903, "step": 428900 }, { "epoch": 0.80046839390808, "grad_norm": 1.0902111530303955, "learning_rate": 0.00019392632956433733, "loss": 4.5412, "step": 428950 }, { "epoch": 0.8005616994674585, "grad_norm": 0.8139966726303101, "learning_rate": 0.00019392488771569945, "loss": 4.4993, "step": 429000 }, { "epoch": 0.8005616994674585, "eval_loss": 4.7199482917785645, "eval_runtime": 232.6367, "eval_samples_per_second": 11.211, "eval_steps_per_second": 11.211, "eval_tts_loss": 7.580851806817645, "step": 429000 }, { "epoch": 0.8006550050268371, "grad_norm": 1.0927164554595947, "learning_rate": 0.0001939234457013007, "loss": 4.5409, "step": 429050 }, { "epoch": 0.8007483105862155, "grad_norm": 0.8889446258544922, "learning_rate": 0.00019392200352114355, "loss": 4.5445, "step": 429100 }, { "epoch": 0.800841616145594, "grad_norm": 1.0192301273345947, "learning_rate": 0.0001939205611752306, "loss": 4.633, "step": 429150 }, { "epoch": 0.8009349217049725, "grad_norm": 0.770257294178009, "learning_rate": 0.00019391911866356435, "loss": 4.4848, "step": 429200 }, { "epoch": 0.801028227264351, "grad_norm": 1.1421622037887573, "learning_rate": 0.00019391767598614745, "loss": 4.544, "step": 429250 }, { "epoch": 0.8011215328237294, "grad_norm": 0.9731292128562927, "learning_rate": 0.00019391623314298233, "loss": 4.406, "step": 429300 }, { "epoch": 0.8012148383831079, "grad_norm": 0.8407917022705078, "learning_rate": 0.00019391479013407162, "loss": 4.7693, "step": 429350 }, { "epoch": 0.8013081439424865, "grad_norm": 0.8591404557228088, "learning_rate": 0.00019391334695941778, "loss": 4.5128, "step": 429400 }, { "epoch": 0.801401449501865, "grad_norm": 0.9232378602027893, "learning_rate": 0.00019391190361902342, "loss": 4.5328, "step": 429450 }, { "epoch": 0.8014947550612435, "grad_norm": 1.314203143119812, "learning_rate": 0.0001939104601128911, "loss": 4.5138, "step": 429500 }, { "epoch": 0.8015880606206219, "grad_norm": 0.9116572141647339, "learning_rate": 0.00019390901644102335, "loss": 4.7407, "step": 429550 }, { "epoch": 0.8016813661800004, "grad_norm": 1.0177239179611206, "learning_rate": 0.00019390757260342268, "loss": 4.5381, "step": 429600 }, { "epoch": 0.8017746717393789, "grad_norm": 0.8840582966804504, "learning_rate": 0.00019390612860009169, "loss": 4.4708, "step": 429650 }, { "epoch": 0.8018679772987574, "grad_norm": 1.3547614812850952, "learning_rate": 0.00019390468443103288, "loss": 4.4395, "step": 429700 }, { "epoch": 0.801961282858136, "grad_norm": 1.4510341882705688, "learning_rate": 0.00019390324009624886, "loss": 4.5966, "step": 429750 }, { "epoch": 0.8020545884175144, "grad_norm": 0.9955864548683167, "learning_rate": 0.0001939017955957421, "loss": 4.4213, "step": 429800 }, { "epoch": 0.8021478939768929, "grad_norm": 1.103618860244751, "learning_rate": 0.00019390035092951523, "loss": 4.6235, "step": 429850 }, { "epoch": 0.8022411995362714, "grad_norm": 1.115838885307312, "learning_rate": 0.00019389890609757075, "loss": 4.643, "step": 429900 }, { "epoch": 0.8023345050956499, "grad_norm": 0.9047982692718506, "learning_rate": 0.00019389746109991121, "loss": 4.4769, "step": 429950 }, { "epoch": 0.8024278106550283, "grad_norm": 1.1972705125808716, "learning_rate": 0.00019389601593653919, "loss": 4.3868, "step": 430000 }, { "epoch": 0.8025211162144068, "grad_norm": 0.8004233837127686, "learning_rate": 0.00019389457060745724, "loss": 4.7359, "step": 430050 }, { "epoch": 0.8026144217737854, "grad_norm": 1.0934377908706665, "learning_rate": 0.00019389312511266786, "loss": 4.5359, "step": 430100 }, { "epoch": 0.8027077273331639, "grad_norm": 1.0269830226898193, "learning_rate": 0.00019389167945217366, "loss": 4.4079, "step": 430150 }, { "epoch": 0.8028010328925423, "grad_norm": 1.061034917831421, "learning_rate": 0.00019389023362597713, "loss": 4.5938, "step": 430200 }, { "epoch": 0.8028943384519208, "grad_norm": 1.0427353382110596, "learning_rate": 0.00019388878763408092, "loss": 4.5804, "step": 430250 }, { "epoch": 0.8029876440112993, "grad_norm": 1.1512060165405273, "learning_rate": 0.00019388734147648745, "loss": 4.432, "step": 430300 }, { "epoch": 0.8030809495706778, "grad_norm": 0.8705995678901672, "learning_rate": 0.0001938858951531994, "loss": 4.5563, "step": 430350 }, { "epoch": 0.8031742551300562, "grad_norm": 0.9586858153343201, "learning_rate": 0.00019388444866421924, "loss": 4.4358, "step": 430400 }, { "epoch": 0.8032675606894347, "grad_norm": 1.1046531200408936, "learning_rate": 0.00019388300200954954, "loss": 4.5264, "step": 430450 }, { "epoch": 0.8033608662488133, "grad_norm": 1.192919373512268, "learning_rate": 0.00019388155518919287, "loss": 4.6478, "step": 430500 }, { "epoch": 0.8034541718081918, "grad_norm": 1.0320477485656738, "learning_rate": 0.0001938801082031518, "loss": 4.6415, "step": 430550 }, { "epoch": 0.8035474773675703, "grad_norm": 0.8191885948181152, "learning_rate": 0.0001938786610514288, "loss": 4.418, "step": 430600 }, { "epoch": 0.8036407829269487, "grad_norm": 1.1384339332580566, "learning_rate": 0.00019387721373402653, "loss": 4.6896, "step": 430650 }, { "epoch": 0.8037340884863272, "grad_norm": 0.8888775706291199, "learning_rate": 0.00019387576625094747, "loss": 4.5304, "step": 430700 }, { "epoch": 0.8038273940457057, "grad_norm": 1.0733976364135742, "learning_rate": 0.0001938743186021942, "loss": 4.5005, "step": 430750 }, { "epoch": 0.8039206996050842, "grad_norm": 1.3514388799667358, "learning_rate": 0.00019387287078776927, "loss": 4.7052, "step": 430800 }, { "epoch": 0.8040140051644628, "grad_norm": 1.0057579278945923, "learning_rate": 0.00019387142280767527, "loss": 4.5062, "step": 430850 }, { "epoch": 0.8041073107238412, "grad_norm": 1.267685055732727, "learning_rate": 0.00019386997466191473, "loss": 4.4404, "step": 430900 }, { "epoch": 0.8042006162832197, "grad_norm": 1.0608516931533813, "learning_rate": 0.00019386852635049017, "loss": 4.5119, "step": 430950 }, { "epoch": 0.8042939218425982, "grad_norm": 1.2537400722503662, "learning_rate": 0.0001938670778734042, "loss": 4.7115, "step": 431000 }, { "epoch": 0.8043872274019767, "grad_norm": 1.269854187965393, "learning_rate": 0.00019386562923065935, "loss": 4.7092, "step": 431050 }, { "epoch": 0.8044805329613551, "grad_norm": 1.3239473104476929, "learning_rate": 0.00019386418042225816, "loss": 4.5927, "step": 431100 }, { "epoch": 0.8045738385207336, "grad_norm": 1.200350046157837, "learning_rate": 0.00019386273144820324, "loss": 4.8531, "step": 431150 }, { "epoch": 0.8046671440801122, "grad_norm": 0.8199599981307983, "learning_rate": 0.0001938612823084971, "loss": 4.5809, "step": 431200 }, { "epoch": 0.8047604496394907, "grad_norm": 1.2164291143417358, "learning_rate": 0.0001938598330031423, "loss": 4.5588, "step": 431250 }, { "epoch": 0.8048537551988691, "grad_norm": 0.743262767791748, "learning_rate": 0.00019385838353214142, "loss": 4.711, "step": 431300 }, { "epoch": 0.8049470607582476, "grad_norm": 1.2544654607772827, "learning_rate": 0.000193856933895497, "loss": 4.4243, "step": 431350 }, { "epoch": 0.8050403663176261, "grad_norm": 1.6423524618148804, "learning_rate": 0.0001938554840932116, "loss": 4.401, "step": 431400 }, { "epoch": 0.8051336718770046, "grad_norm": 0.9570430517196655, "learning_rate": 0.00019385403412528783, "loss": 4.5196, "step": 431450 }, { "epoch": 0.805226977436383, "grad_norm": 0.5382121205329895, "learning_rate": 0.00019385258399172813, "loss": 4.4504, "step": 431500 }, { "epoch": 0.8053202829957616, "grad_norm": 0.977709174156189, "learning_rate": 0.0001938511336925352, "loss": 4.4906, "step": 431550 }, { "epoch": 0.8054135885551401, "grad_norm": 1.1496464014053345, "learning_rate": 0.0001938496832277115, "loss": 4.644, "step": 431600 }, { "epoch": 0.8055068941145186, "grad_norm": 0.735632598400116, "learning_rate": 0.00019384823259725964, "loss": 4.4586, "step": 431650 }, { "epoch": 0.8056001996738971, "grad_norm": 1.154396653175354, "learning_rate": 0.00019384678180118213, "loss": 4.4867, "step": 431700 }, { "epoch": 0.8056935052332755, "grad_norm": 0.8699199557304382, "learning_rate": 0.0001938453308394816, "loss": 4.5057, "step": 431750 }, { "epoch": 0.805786810792654, "grad_norm": 1.3764482736587524, "learning_rate": 0.0001938438797121606, "loss": 4.4259, "step": 431800 }, { "epoch": 0.8058801163520325, "grad_norm": 1.190111756324768, "learning_rate": 0.0001938424284192216, "loss": 4.6185, "step": 431850 }, { "epoch": 0.8059734219114111, "grad_norm": 0.7815505862236023, "learning_rate": 0.00019384097696066723, "loss": 4.3942, "step": 431900 }, { "epoch": 0.8060667274707896, "grad_norm": 1.2037734985351562, "learning_rate": 0.00019383952533650006, "loss": 4.6106, "step": 431950 }, { "epoch": 0.806160033030168, "grad_norm": 0.9880151748657227, "learning_rate": 0.00019383807354672266, "loss": 4.6179, "step": 432000 }, { "epoch": 0.806160033030168, "eval_loss": 4.717226505279541, "eval_runtime": 232.5206, "eval_samples_per_second": 11.216, "eval_steps_per_second": 11.216, "eval_tts_loss": 7.522934058765406, "step": 432000 }, { "epoch": 0.8062533385895465, "grad_norm": 1.1217437982559204, "learning_rate": 0.00019383662159133756, "loss": 4.5934, "step": 432050 }, { "epoch": 0.806346644148925, "grad_norm": 1.386207103729248, "learning_rate": 0.00019383516947034735, "loss": 4.7539, "step": 432100 }, { "epoch": 0.8064399497083035, "grad_norm": 1.0124891996383667, "learning_rate": 0.00019383371718375455, "loss": 4.3381, "step": 432150 }, { "epoch": 0.8065332552676819, "grad_norm": 0.9883818626403809, "learning_rate": 0.00019383226473156174, "loss": 4.3125, "step": 432200 }, { "epoch": 0.8066265608270605, "grad_norm": 0.9008289575576782, "learning_rate": 0.00019383081211377153, "loss": 4.3435, "step": 432250 }, { "epoch": 0.806719866386439, "grad_norm": 1.0534547567367554, "learning_rate": 0.00019382935933038643, "loss": 4.6482, "step": 432300 }, { "epoch": 0.8068131719458175, "grad_norm": 0.7578873634338379, "learning_rate": 0.000193827906381409, "loss": 4.4737, "step": 432350 }, { "epoch": 0.806906477505196, "grad_norm": 1.3317744731903076, "learning_rate": 0.00019382645326684186, "loss": 4.7364, "step": 432400 }, { "epoch": 0.8069997830645744, "grad_norm": 1.2212891578674316, "learning_rate": 0.0001938249999866875, "loss": 4.4826, "step": 432450 }, { "epoch": 0.8070930886239529, "grad_norm": 1.2318320274353027, "learning_rate": 0.00019382354654094857, "loss": 4.7266, "step": 432500 }, { "epoch": 0.8071863941833314, "grad_norm": 1.1068717241287231, "learning_rate": 0.00019382209292962755, "loss": 4.4529, "step": 432550 }, { "epoch": 0.80727969974271, "grad_norm": 0.8774948716163635, "learning_rate": 0.00019382063915272707, "loss": 4.5036, "step": 432600 }, { "epoch": 0.8073730053020884, "grad_norm": 1.1480398178100586, "learning_rate": 0.00019381918521024967, "loss": 4.3559, "step": 432650 }, { "epoch": 0.8074663108614669, "grad_norm": 1.0650641918182373, "learning_rate": 0.0001938177311021979, "loss": 4.6795, "step": 432700 }, { "epoch": 0.8075596164208454, "grad_norm": 1.1158818006515503, "learning_rate": 0.00019381627682857436, "loss": 4.5262, "step": 432750 }, { "epoch": 0.8076529219802239, "grad_norm": 0.9662179350852966, "learning_rate": 0.00019381482238938158, "loss": 4.5011, "step": 432800 }, { "epoch": 0.8077462275396023, "grad_norm": 1.0558754205703735, "learning_rate": 0.00019381336778462216, "loss": 4.4317, "step": 432850 }, { "epoch": 0.8078395330989808, "grad_norm": 1.1265623569488525, "learning_rate": 0.00019381191301429864, "loss": 4.5669, "step": 432900 }, { "epoch": 0.8079328386583594, "grad_norm": 1.1973166465759277, "learning_rate": 0.0001938104580784136, "loss": 4.4845, "step": 432950 }, { "epoch": 0.8080261442177379, "grad_norm": 1.4364298582077026, "learning_rate": 0.0001938090029769696, "loss": 4.4275, "step": 433000 }, { "epoch": 0.8081194497771164, "grad_norm": 1.2793713808059692, "learning_rate": 0.00019380754770996925, "loss": 4.6811, "step": 433050 }, { "epoch": 0.8082127553364948, "grad_norm": 0.9621243476867676, "learning_rate": 0.00019380609227741507, "loss": 4.4556, "step": 433100 }, { "epoch": 0.8083060608958733, "grad_norm": 1.181762933731079, "learning_rate": 0.00019380463667930965, "loss": 4.4645, "step": 433150 }, { "epoch": 0.8083993664552518, "grad_norm": 1.160107135772705, "learning_rate": 0.00019380318091565552, "loss": 4.4719, "step": 433200 }, { "epoch": 0.8084926720146303, "grad_norm": 1.064186692237854, "learning_rate": 0.0001938017249864553, "loss": 4.5754, "step": 433250 }, { "epoch": 0.8085859775740089, "grad_norm": 0.985802173614502, "learning_rate": 0.00019380026889171154, "loss": 4.7013, "step": 433300 }, { "epoch": 0.8086792831333873, "grad_norm": 1.314144253730774, "learning_rate": 0.00019379881263142682, "loss": 4.4962, "step": 433350 }, { "epoch": 0.8087725886927658, "grad_norm": 1.0957282781600952, "learning_rate": 0.0001937973562056037, "loss": 4.5735, "step": 433400 }, { "epoch": 0.8088658942521443, "grad_norm": 1.0866283178329468, "learning_rate": 0.00019379589961424475, "loss": 4.5805, "step": 433450 }, { "epoch": 0.8089591998115228, "grad_norm": 1.0170081853866577, "learning_rate": 0.0001937944428573525, "loss": 4.6924, "step": 433500 }, { "epoch": 0.8090525053709012, "grad_norm": 1.194145917892456, "learning_rate": 0.0001937929859349296, "loss": 4.4469, "step": 433550 }, { "epoch": 0.8091458109302797, "grad_norm": 0.969913125038147, "learning_rate": 0.00019379152884697857, "loss": 4.4026, "step": 433600 }, { "epoch": 0.8092391164896583, "grad_norm": 1.0304129123687744, "learning_rate": 0.00019379007159350201, "loss": 4.4583, "step": 433650 }, { "epoch": 0.8093324220490368, "grad_norm": 0.905687153339386, "learning_rate": 0.00019378861417450248, "loss": 4.5727, "step": 433700 }, { "epoch": 0.8094257276084152, "grad_norm": 1.0900845527648926, "learning_rate": 0.00019378715658998252, "loss": 4.5235, "step": 433750 }, { "epoch": 0.8095190331677937, "grad_norm": 1.0319151878356934, "learning_rate": 0.00019378569883994476, "loss": 4.563, "step": 433800 }, { "epoch": 0.8096123387271722, "grad_norm": 0.8259133100509644, "learning_rate": 0.00019378424092439172, "loss": 4.5616, "step": 433850 }, { "epoch": 0.8097056442865507, "grad_norm": 0.8913447260856628, "learning_rate": 0.00019378278284332603, "loss": 4.6049, "step": 433900 }, { "epoch": 0.8097989498459292, "grad_norm": 0.9409626722335815, "learning_rate": 0.00019378132459675022, "loss": 4.5054, "step": 433950 }, { "epoch": 0.8098922554053077, "grad_norm": 1.0639225244522095, "learning_rate": 0.00019377986618466687, "loss": 4.5129, "step": 434000 }, { "epoch": 0.8099855609646862, "grad_norm": 1.1352790594100952, "learning_rate": 0.00019377840760707852, "loss": 4.5635, "step": 434050 }, { "epoch": 0.8100788665240647, "grad_norm": 0.8985616564750671, "learning_rate": 0.00019377694886398784, "loss": 4.4284, "step": 434100 }, { "epoch": 0.8101721720834432, "grad_norm": 1.2385417222976685, "learning_rate": 0.0001937754899553973, "loss": 4.5906, "step": 434150 }, { "epoch": 0.8102654776428216, "grad_norm": 1.389296293258667, "learning_rate": 0.00019377403088130955, "loss": 4.5582, "step": 434200 }, { "epoch": 0.8103587832022001, "grad_norm": 0.6670157313346863, "learning_rate": 0.0001937725716417271, "loss": 4.426, "step": 434250 }, { "epoch": 0.8104520887615786, "grad_norm": 0.9546065330505371, "learning_rate": 0.0001937711122366526, "loss": 4.632, "step": 434300 }, { "epoch": 0.8105453943209572, "grad_norm": 1.1238577365875244, "learning_rate": 0.00019376965266608854, "loss": 4.6131, "step": 434350 }, { "epoch": 0.8106386998803357, "grad_norm": 1.2199645042419434, "learning_rate": 0.0001937681929300376, "loss": 4.6498, "step": 434400 }, { "epoch": 0.8107320054397141, "grad_norm": 1.0700517892837524, "learning_rate": 0.00019376673302850227, "loss": 4.7002, "step": 434450 }, { "epoch": 0.8108253109990926, "grad_norm": 1.0681084394454956, "learning_rate": 0.00019376527296148513, "loss": 4.4733, "step": 434500 }, { "epoch": 0.8109186165584711, "grad_norm": 0.9579760432243347, "learning_rate": 0.00019376381272898883, "loss": 4.376, "step": 434550 }, { "epoch": 0.8110119221178496, "grad_norm": 1.084481954574585, "learning_rate": 0.00019376235233101586, "loss": 4.5135, "step": 434600 }, { "epoch": 0.811105227677228, "grad_norm": 1.28203547000885, "learning_rate": 0.00019376089176756888, "loss": 4.5879, "step": 434650 }, { "epoch": 0.8111985332366066, "grad_norm": 1.1159179210662842, "learning_rate": 0.0001937594310386504, "loss": 4.5683, "step": 434700 }, { "epoch": 0.8112918387959851, "grad_norm": 1.1986109018325806, "learning_rate": 0.00019375797014426301, "loss": 4.5916, "step": 434750 }, { "epoch": 0.8113851443553636, "grad_norm": 1.1072922945022583, "learning_rate": 0.0001937565090844093, "loss": 4.4843, "step": 434800 }, { "epoch": 0.811478449914742, "grad_norm": 0.9535547494888306, "learning_rate": 0.00019375504785909184, "loss": 4.3743, "step": 434850 }, { "epoch": 0.8115717554741205, "grad_norm": 1.0558990240097046, "learning_rate": 0.00019375358646831324, "loss": 4.5874, "step": 434900 }, { "epoch": 0.811665061033499, "grad_norm": 1.1166068315505981, "learning_rate": 0.00019375212491207607, "loss": 4.4974, "step": 434950 }, { "epoch": 0.8117583665928775, "grad_norm": 0.9038137197494507, "learning_rate": 0.00019375066319038287, "loss": 4.6755, "step": 435000 }, { "epoch": 0.8117583665928775, "eval_loss": 4.713137149810791, "eval_runtime": 233.558, "eval_samples_per_second": 11.166, "eval_steps_per_second": 11.166, "eval_tts_loss": 7.611853272671257, "step": 435000 }, { "epoch": 0.8118516721522561, "grad_norm": 1.1564393043518066, "learning_rate": 0.00019374920130323624, "loss": 4.5115, "step": 435050 }, { "epoch": 0.8119449777116345, "grad_norm": 1.019436001777649, "learning_rate": 0.00019374773925063879, "loss": 4.5726, "step": 435100 }, { "epoch": 0.812038283271013, "grad_norm": 1.456741213798523, "learning_rate": 0.00019374627703259306, "loss": 4.5041, "step": 435150 }, { "epoch": 0.8121315888303915, "grad_norm": 1.1122186183929443, "learning_rate": 0.00019374481464910163, "loss": 4.4887, "step": 435200 }, { "epoch": 0.81222489438977, "grad_norm": 0.973937451839447, "learning_rate": 0.00019374335210016713, "loss": 4.6768, "step": 435250 }, { "epoch": 0.8123181999491484, "grad_norm": 1.2644262313842773, "learning_rate": 0.00019374188938579208, "loss": 4.625, "step": 435300 }, { "epoch": 0.8124115055085269, "grad_norm": 0.8474568724632263, "learning_rate": 0.0001937404265059791, "loss": 4.4576, "step": 435350 }, { "epoch": 0.8125048110679055, "grad_norm": 1.687286615371704, "learning_rate": 0.00019373896346073076, "loss": 4.6964, "step": 435400 }, { "epoch": 0.812598116627284, "grad_norm": 1.0351998805999756, "learning_rate": 0.00019373750025004967, "loss": 4.5602, "step": 435450 }, { "epoch": 0.8126914221866625, "grad_norm": 1.0308910608291626, "learning_rate": 0.00019373603687393835, "loss": 4.6834, "step": 435500 }, { "epoch": 0.8127847277460409, "grad_norm": 1.1440367698669434, "learning_rate": 0.00019373457333239944, "loss": 4.4099, "step": 435550 }, { "epoch": 0.8128780333054194, "grad_norm": 0.7390778064727783, "learning_rate": 0.0001937331096254355, "loss": 4.6605, "step": 435600 }, { "epoch": 0.8129713388647979, "grad_norm": 0.8651524782180786, "learning_rate": 0.0001937316457530491, "loss": 4.692, "step": 435650 }, { "epoch": 0.8130646444241764, "grad_norm": 0.9918853640556335, "learning_rate": 0.00019373018171524285, "loss": 4.5824, "step": 435700 }, { "epoch": 0.8131579499835548, "grad_norm": 1.238487720489502, "learning_rate": 0.0001937287175120193, "loss": 4.5105, "step": 435750 }, { "epoch": 0.8132512555429334, "grad_norm": 1.0910192728042603, "learning_rate": 0.0001937272531433811, "loss": 4.4074, "step": 435800 }, { "epoch": 0.8133445611023119, "grad_norm": 0.9015228748321533, "learning_rate": 0.00019372578860933076, "loss": 4.7231, "step": 435850 }, { "epoch": 0.8134378666616904, "grad_norm": 0.8462371826171875, "learning_rate": 0.0001937243239098709, "loss": 4.4896, "step": 435900 }, { "epoch": 0.8135311722210689, "grad_norm": 0.8238480687141418, "learning_rate": 0.00019372285904500406, "loss": 4.5205, "step": 435950 }, { "epoch": 0.8136244777804473, "grad_norm": 0.7250654101371765, "learning_rate": 0.00019372139401473292, "loss": 4.5316, "step": 436000 }, { "epoch": 0.8137177833398258, "grad_norm": 1.0538784265518188, "learning_rate": 0.00019371992881906001, "loss": 4.5203, "step": 436050 }, { "epoch": 0.8138110888992043, "grad_norm": 0.9435088038444519, "learning_rate": 0.00019371846345798787, "loss": 4.4478, "step": 436100 }, { "epoch": 0.8139043944585829, "grad_norm": 1.0073535442352295, "learning_rate": 0.00019371699793151916, "loss": 4.685, "step": 436150 }, { "epoch": 0.8139977000179613, "grad_norm": 1.3032444715499878, "learning_rate": 0.00019371553223965644, "loss": 4.6852, "step": 436200 }, { "epoch": 0.8140910055773398, "grad_norm": 1.096046805381775, "learning_rate": 0.0001937140663824023, "loss": 4.4044, "step": 436250 }, { "epoch": 0.8141843111367183, "grad_norm": 0.7758470773696899, "learning_rate": 0.00019371260035975932, "loss": 4.3741, "step": 436300 }, { "epoch": 0.8142776166960968, "grad_norm": 0.9498625993728638, "learning_rate": 0.0001937111341717301, "loss": 4.7392, "step": 436350 }, { "epoch": 0.8143709222554752, "grad_norm": 1.3436883687973022, "learning_rate": 0.0001937096678183172, "loss": 4.4038, "step": 436400 }, { "epoch": 0.8144642278148537, "grad_norm": 0.9300359487533569, "learning_rate": 0.00019370820129952323, "loss": 4.5385, "step": 436450 }, { "epoch": 0.8145575333742323, "grad_norm": 1.2873274087905884, "learning_rate": 0.00019370673461535076, "loss": 4.6742, "step": 436500 }, { "epoch": 0.8146508389336108, "grad_norm": 1.1295846700668335, "learning_rate": 0.0001937052677658024, "loss": 4.5639, "step": 436550 }, { "epoch": 0.8147441444929893, "grad_norm": 1.1773862838745117, "learning_rate": 0.00019370380075088074, "loss": 4.6209, "step": 436600 }, { "epoch": 0.8148374500523677, "grad_norm": 1.0682029724121094, "learning_rate": 0.00019370233357058836, "loss": 4.5695, "step": 436650 }, { "epoch": 0.8149307556117462, "grad_norm": 0.7991012334823608, "learning_rate": 0.00019370086622492782, "loss": 4.475, "step": 436700 }, { "epoch": 0.8150240611711247, "grad_norm": 1.3142451047897339, "learning_rate": 0.0001936993987139018, "loss": 4.5237, "step": 436750 }, { "epoch": 0.8151173667305032, "grad_norm": 1.2649680376052856, "learning_rate": 0.00019369793103751276, "loss": 4.6097, "step": 436800 }, { "epoch": 0.8152106722898818, "grad_norm": 0.9599008560180664, "learning_rate": 0.0001936964631957634, "loss": 4.583, "step": 436850 }, { "epoch": 0.8153039778492602, "grad_norm": 1.1062766313552856, "learning_rate": 0.00019369499518865623, "loss": 4.479, "step": 436900 }, { "epoch": 0.8153972834086387, "grad_norm": 0.7987418174743652, "learning_rate": 0.0001936935270161939, "loss": 4.4334, "step": 436950 }, { "epoch": 0.8154905889680172, "grad_norm": 0.9704287648200989, "learning_rate": 0.000193692058678379, "loss": 4.4151, "step": 437000 }, { "epoch": 0.8155838945273957, "grad_norm": 0.8681563138961792, "learning_rate": 0.0001936905901752141, "loss": 4.6867, "step": 437050 }, { "epoch": 0.8156772000867741, "grad_norm": 0.8741142153739929, "learning_rate": 0.00019368912150670176, "loss": 4.4993, "step": 437100 }, { "epoch": 0.8157705056461526, "grad_norm": 0.6997228264808655, "learning_rate": 0.00019368765267284465, "loss": 4.5395, "step": 437150 }, { "epoch": 0.8158638112055312, "grad_norm": 1.2820758819580078, "learning_rate": 0.0001936861836736453, "loss": 4.2663, "step": 437200 }, { "epoch": 0.8159571167649097, "grad_norm": 1.075132131576538, "learning_rate": 0.0001936847145091063, "loss": 4.6971, "step": 437250 }, { "epoch": 0.8160504223242881, "grad_norm": 1.4322841167449951, "learning_rate": 0.0001936832451792303, "loss": 4.5894, "step": 437300 }, { "epoch": 0.8161437278836666, "grad_norm": 1.2873334884643555, "learning_rate": 0.00019368177568401983, "loss": 4.549, "step": 437350 }, { "epoch": 0.8162370334430451, "grad_norm": 1.1981624364852905, "learning_rate": 0.00019368030602347748, "loss": 4.7408, "step": 437400 }, { "epoch": 0.8163303390024236, "grad_norm": 1.266121506690979, "learning_rate": 0.00019367883619760595, "loss": 4.4947, "step": 437450 }, { "epoch": 0.816423644561802, "grad_norm": 1.034895896911621, "learning_rate": 0.0001936773662064077, "loss": 4.5056, "step": 437500 }, { "epoch": 0.8165169501211806, "grad_norm": 0.8305200338363647, "learning_rate": 0.00019367589604988542, "loss": 4.3609, "step": 437550 }, { "epoch": 0.8166102556805591, "grad_norm": 1.0634491443634033, "learning_rate": 0.00019367442572804162, "loss": 4.5135, "step": 437600 }, { "epoch": 0.8167035612399376, "grad_norm": 1.3011289834976196, "learning_rate": 0.00019367295524087895, "loss": 4.4633, "step": 437650 }, { "epoch": 0.8167968667993161, "grad_norm": 1.2609500885009766, "learning_rate": 0.0001936714845884, "loss": 4.6228, "step": 437700 }, { "epoch": 0.8168901723586945, "grad_norm": 0.9911806583404541, "learning_rate": 0.00019367001377060738, "loss": 4.5142, "step": 437750 }, { "epoch": 0.816983477918073, "grad_norm": 1.2613277435302734, "learning_rate": 0.0001936685427875037, "loss": 4.6284, "step": 437800 }, { "epoch": 0.8170767834774515, "grad_norm": 1.2594963312149048, "learning_rate": 0.00019366707163909146, "loss": 4.4815, "step": 437850 }, { "epoch": 0.8171700890368301, "grad_norm": 0.746685802936554, "learning_rate": 0.00019366560032537335, "loss": 4.4356, "step": 437900 }, { "epoch": 0.8172633945962086, "grad_norm": 1.3757548332214355, "learning_rate": 0.00019366412884635193, "loss": 4.59, "step": 437950 }, { "epoch": 0.817356700155587, "grad_norm": 1.0559455156326294, "learning_rate": 0.0001936626572020298, "loss": 4.545, "step": 438000 }, { "epoch": 0.817356700155587, "eval_loss": 4.7206220626831055, "eval_runtime": 230.9923, "eval_samples_per_second": 11.29, "eval_steps_per_second": 11.29, "eval_tts_loss": 7.6207517155023305, "step": 438000 }, { "epoch": 0.8174500057149655, "grad_norm": 0.9019753336906433, "learning_rate": 0.00019366118539240957, "loss": 4.6121, "step": 438050 }, { "epoch": 0.817543311274344, "grad_norm": 0.8623765110969543, "learning_rate": 0.00019365971341749383, "loss": 4.4661, "step": 438100 }, { "epoch": 0.8176366168337225, "grad_norm": 1.2834802865982056, "learning_rate": 0.00019365824127728516, "loss": 4.5664, "step": 438150 }, { "epoch": 0.8177299223931009, "grad_norm": 1.1096104383468628, "learning_rate": 0.00019365676897178618, "loss": 4.4834, "step": 438200 }, { "epoch": 0.8178232279524795, "grad_norm": 1.050219178199768, "learning_rate": 0.00019365529650099947, "loss": 4.3719, "step": 438250 }, { "epoch": 0.817916533511858, "grad_norm": 1.2993484735488892, "learning_rate": 0.00019365382386492765, "loss": 4.4403, "step": 438300 }, { "epoch": 0.8180098390712365, "grad_norm": 1.1241735219955444, "learning_rate": 0.0001936523510635733, "loss": 4.475, "step": 438350 }, { "epoch": 0.818103144630615, "grad_norm": 1.0164114236831665, "learning_rate": 0.00019365087809693906, "loss": 4.5896, "step": 438400 }, { "epoch": 0.8181964501899934, "grad_norm": 1.322221279144287, "learning_rate": 0.00019364940496502747, "loss": 4.6524, "step": 438450 }, { "epoch": 0.8182897557493719, "grad_norm": 1.0273584127426147, "learning_rate": 0.00019364793166784118, "loss": 4.5349, "step": 438500 }, { "epoch": 0.8183830613087504, "grad_norm": 1.0621941089630127, "learning_rate": 0.00019364645820538277, "loss": 4.4498, "step": 438550 }, { "epoch": 0.818476366868129, "grad_norm": 1.1163345575332642, "learning_rate": 0.00019364498457765482, "loss": 4.4727, "step": 438600 }, { "epoch": 0.8185696724275074, "grad_norm": 1.2609418630599976, "learning_rate": 0.00019364351078465997, "loss": 4.6392, "step": 438650 }, { "epoch": 0.8186629779868859, "grad_norm": 1.1600068807601929, "learning_rate": 0.0001936420368264008, "loss": 4.347, "step": 438700 }, { "epoch": 0.8187562835462644, "grad_norm": 0.9375897645950317, "learning_rate": 0.0001936405627028799, "loss": 4.6244, "step": 438750 }, { "epoch": 0.8188495891056429, "grad_norm": 1.4730111360549927, "learning_rate": 0.00019363908841409986, "loss": 4.6538, "step": 438800 }, { "epoch": 0.8189428946650213, "grad_norm": 0.9291381239891052, "learning_rate": 0.00019363761396006332, "loss": 4.5404, "step": 438850 }, { "epoch": 0.8190362002243998, "grad_norm": 1.0139800310134888, "learning_rate": 0.0001936361393407729, "loss": 4.6218, "step": 438900 }, { "epoch": 0.8191295057837784, "grad_norm": 1.3425135612487793, "learning_rate": 0.00019363466455623113, "loss": 4.5433, "step": 438950 }, { "epoch": 0.8192228113431569, "grad_norm": 0.9264724850654602, "learning_rate": 0.00019363318960644068, "loss": 4.6103, "step": 439000 }, { "epoch": 0.8193161169025354, "grad_norm": 1.0425794124603271, "learning_rate": 0.0001936317144914041, "loss": 4.595, "step": 439050 }, { "epoch": 0.8194094224619138, "grad_norm": 1.0097206830978394, "learning_rate": 0.00019363023921112405, "loss": 4.4658, "step": 439100 }, { "epoch": 0.8195027280212923, "grad_norm": 1.0204837322235107, "learning_rate": 0.00019362876376560306, "loss": 4.4472, "step": 439150 }, { "epoch": 0.8195960335806708, "grad_norm": 1.0780389308929443, "learning_rate": 0.0001936272881548438, "loss": 4.6381, "step": 439200 }, { "epoch": 0.8196893391400493, "grad_norm": 1.4371711015701294, "learning_rate": 0.00019362581237884885, "loss": 4.5083, "step": 439250 }, { "epoch": 0.8197826446994279, "grad_norm": 1.2701109647750854, "learning_rate": 0.0001936243364376208, "loss": 4.6079, "step": 439300 }, { "epoch": 0.8198759502588063, "grad_norm": 1.156437635421753, "learning_rate": 0.00019362286033116233, "loss": 4.6392, "step": 439350 }, { "epoch": 0.8199692558181848, "grad_norm": 1.3134959936141968, "learning_rate": 0.00019362138405947592, "loss": 4.5783, "step": 439400 }, { "epoch": 0.8200625613775633, "grad_norm": 0.8655474781990051, "learning_rate": 0.00019361990762256425, "loss": 4.5355, "step": 439450 }, { "epoch": 0.8201558669369418, "grad_norm": 1.014864206314087, "learning_rate": 0.00019361843102042996, "loss": 4.3339, "step": 439500 }, { "epoch": 0.8202491724963202, "grad_norm": 0.9473825097084045, "learning_rate": 0.00019361695425307556, "loss": 4.6462, "step": 439550 }, { "epoch": 0.8203424780556987, "grad_norm": 1.0384535789489746, "learning_rate": 0.00019361547732050374, "loss": 4.6507, "step": 439600 }, { "epoch": 0.8204357836150773, "grad_norm": 0.7609256505966187, "learning_rate": 0.00019361400022271705, "loss": 4.3934, "step": 439650 }, { "epoch": 0.8205290891744558, "grad_norm": 0.8633242845535278, "learning_rate": 0.00019361252295971813, "loss": 4.441, "step": 439700 }, { "epoch": 0.8206223947338342, "grad_norm": 0.8905895352363586, "learning_rate": 0.0001936110455315096, "loss": 4.562, "step": 439750 }, { "epoch": 0.8207157002932127, "grad_norm": 0.65688157081604, "learning_rate": 0.000193609567938094, "loss": 4.6056, "step": 439800 }, { "epoch": 0.8208090058525912, "grad_norm": 0.895016610622406, "learning_rate": 0.00019360809017947403, "loss": 4.4881, "step": 439850 }, { "epoch": 0.8209023114119697, "grad_norm": 1.1564828157424927, "learning_rate": 0.00019360661225565224, "loss": 4.8174, "step": 439900 }, { "epoch": 0.8209956169713482, "grad_norm": 1.0620310306549072, "learning_rate": 0.00019360513416663124, "loss": 4.648, "step": 439950 }, { "epoch": 0.8210889225307267, "grad_norm": 1.1277257204055786, "learning_rate": 0.00019360365591241364, "loss": 4.4791, "step": 440000 }, { "epoch": 0.8211822280901052, "grad_norm": 1.2287598848342896, "learning_rate": 0.0001936021774930021, "loss": 4.5892, "step": 440050 }, { "epoch": 0.8212755336494837, "grad_norm": 0.9048253297805786, "learning_rate": 0.00019360069890839913, "loss": 4.4444, "step": 440100 }, { "epoch": 0.8213688392088622, "grad_norm": 0.9036170244216919, "learning_rate": 0.00019359922015860744, "loss": 4.7466, "step": 440150 }, { "epoch": 0.8214621447682406, "grad_norm": 1.0846253633499146, "learning_rate": 0.00019359774124362957, "loss": 4.5556, "step": 440200 }, { "epoch": 0.8215554503276191, "grad_norm": 1.0259082317352295, "learning_rate": 0.00019359626216346816, "loss": 4.447, "step": 440250 }, { "epoch": 0.8216487558869976, "grad_norm": 1.0791857242584229, "learning_rate": 0.00019359478291812581, "loss": 4.604, "step": 440300 }, { "epoch": 0.8217420614463762, "grad_norm": 1.085229754447937, "learning_rate": 0.00019359330350760516, "loss": 4.6088, "step": 440350 }, { "epoch": 0.8218353670057547, "grad_norm": 0.9532181620597839, "learning_rate": 0.00019359182393190877, "loss": 4.5741, "step": 440400 }, { "epoch": 0.8219286725651331, "grad_norm": 1.000110149383545, "learning_rate": 0.0001935903441910393, "loss": 4.5447, "step": 440450 }, { "epoch": 0.8220219781245116, "grad_norm": 1.0213303565979004, "learning_rate": 0.00019358886428499935, "loss": 4.7264, "step": 440500 }, { "epoch": 0.8221152836838901, "grad_norm": 1.0686661005020142, "learning_rate": 0.0001935873842137915, "loss": 4.5832, "step": 440550 }, { "epoch": 0.8222085892432686, "grad_norm": 0.8314534425735474, "learning_rate": 0.0001935859039774184, "loss": 4.5014, "step": 440600 }, { "epoch": 0.822301894802647, "grad_norm": 1.1675671339035034, "learning_rate": 0.00019358442357588263, "loss": 4.6003, "step": 440650 }, { "epoch": 0.8223952003620256, "grad_norm": 1.1414916515350342, "learning_rate": 0.00019358294300918683, "loss": 4.4115, "step": 440700 }, { "epoch": 0.8224885059214041, "grad_norm": 1.1694287061691284, "learning_rate": 0.0001935814622773336, "loss": 4.4839, "step": 440750 }, { "epoch": 0.8225818114807826, "grad_norm": 0.8491055369377136, "learning_rate": 0.00019357998138032554, "loss": 4.3969, "step": 440800 }, { "epoch": 0.822675117040161, "grad_norm": 0.9321352243423462, "learning_rate": 0.00019357850031816528, "loss": 4.6168, "step": 440850 }, { "epoch": 0.8227684225995395, "grad_norm": 1.036413550376892, "learning_rate": 0.00019357701909085545, "loss": 4.4682, "step": 440900 }, { "epoch": 0.822861728158918, "grad_norm": 1.0816128253936768, "learning_rate": 0.00019357553769839864, "loss": 4.7283, "step": 440950 }, { "epoch": 0.8229550337182965, "grad_norm": 1.1739145517349243, "learning_rate": 0.00019357405614079748, "loss": 4.47, "step": 441000 }, { "epoch": 0.8229550337182965, "eval_loss": 4.712830066680908, "eval_runtime": 232.585, "eval_samples_per_second": 11.213, "eval_steps_per_second": 11.213, "eval_tts_loss": 7.592319561862758, "step": 441000 }, { "epoch": 0.8230483392776751, "grad_norm": 1.0819491147994995, "learning_rate": 0.00019357257441805455, "loss": 4.574, "step": 441050 }, { "epoch": 0.8231416448370535, "grad_norm": 1.0389968156814575, "learning_rate": 0.00019357109253017254, "loss": 4.5989, "step": 441100 }, { "epoch": 0.823234950396432, "grad_norm": 1.2745685577392578, "learning_rate": 0.00019356961047715395, "loss": 4.4777, "step": 441150 }, { "epoch": 0.8233282559558105, "grad_norm": 1.6084719896316528, "learning_rate": 0.00019356812825900154, "loss": 4.3616, "step": 441200 }, { "epoch": 0.823421561515189, "grad_norm": 1.1243865489959717, "learning_rate": 0.0001935666458757178, "loss": 4.6195, "step": 441250 }, { "epoch": 0.8235148670745674, "grad_norm": 1.3631200790405273, "learning_rate": 0.00019356516332730536, "loss": 4.4103, "step": 441300 }, { "epoch": 0.8236081726339459, "grad_norm": 1.0318950414657593, "learning_rate": 0.00019356368061376693, "loss": 4.4224, "step": 441350 }, { "epoch": 0.8237014781933244, "grad_norm": 1.1379915475845337, "learning_rate": 0.00019356219773510502, "loss": 4.8016, "step": 441400 }, { "epoch": 0.823794783752703, "grad_norm": 0.9971288442611694, "learning_rate": 0.00019356071469132233, "loss": 4.4852, "step": 441450 }, { "epoch": 0.8238880893120815, "grad_norm": 0.6444038152694702, "learning_rate": 0.00019355923148242141, "loss": 4.5769, "step": 441500 }, { "epoch": 0.8239813948714599, "grad_norm": 0.9256520867347717, "learning_rate": 0.00019355774810840494, "loss": 4.5454, "step": 441550 }, { "epoch": 0.8240747004308384, "grad_norm": 1.0571428537368774, "learning_rate": 0.00019355626456927548, "loss": 4.5266, "step": 441600 }, { "epoch": 0.8241680059902169, "grad_norm": 1.3807666301727295, "learning_rate": 0.00019355478086503572, "loss": 4.7166, "step": 441650 }, { "epoch": 0.8242613115495954, "grad_norm": 0.9384042024612427, "learning_rate": 0.00019355329699568817, "loss": 4.5248, "step": 441700 }, { "epoch": 0.8243546171089738, "grad_norm": 0.9789107441902161, "learning_rate": 0.00019355181296123558, "loss": 4.4164, "step": 441750 }, { "epoch": 0.8244479226683524, "grad_norm": 0.6336019039154053, "learning_rate": 0.00019355032876168044, "loss": 4.5301, "step": 441800 }, { "epoch": 0.8245412282277309, "grad_norm": 0.892485499382019, "learning_rate": 0.00019354884439702548, "loss": 4.7402, "step": 441850 }, { "epoch": 0.8246345337871094, "grad_norm": 1.0237869024276733, "learning_rate": 0.00019354735986727324, "loss": 4.5871, "step": 441900 }, { "epoch": 0.8247278393464879, "grad_norm": 1.1821402311325073, "learning_rate": 0.00019354587517242635, "loss": 4.619, "step": 441950 }, { "epoch": 0.8248211449058663, "grad_norm": 1.0676908493041992, "learning_rate": 0.00019354439031248747, "loss": 4.331, "step": 442000 }, { "epoch": 0.8249144504652448, "grad_norm": 0.8039354681968689, "learning_rate": 0.00019354290528745922, "loss": 4.8554, "step": 442050 }, { "epoch": 0.8250077560246233, "grad_norm": 1.2127957344055176, "learning_rate": 0.00019354142009734418, "loss": 4.601, "step": 442100 }, { "epoch": 0.8251010615840019, "grad_norm": 1.4200031757354736, "learning_rate": 0.00019353993474214502, "loss": 4.71, "step": 442150 }, { "epoch": 0.8251943671433803, "grad_norm": 1.0917919874191284, "learning_rate": 0.0001935384492218643, "loss": 4.524, "step": 442200 }, { "epoch": 0.8252876727027588, "grad_norm": 1.008728265762329, "learning_rate": 0.0001935369635365047, "loss": 4.4806, "step": 442250 }, { "epoch": 0.8253809782621373, "grad_norm": 1.1972861289978027, "learning_rate": 0.0001935354776860688, "loss": 4.4702, "step": 442300 }, { "epoch": 0.8254742838215158, "grad_norm": 0.9816621541976929, "learning_rate": 0.00019353399167055925, "loss": 4.6638, "step": 442350 }, { "epoch": 0.8255675893808943, "grad_norm": 1.1558983325958252, "learning_rate": 0.00019353250548997865, "loss": 4.6179, "step": 442400 }, { "epoch": 0.8256608949402727, "grad_norm": 0.9303699731826782, "learning_rate": 0.00019353101914432966, "loss": 4.618, "step": 442450 }, { "epoch": 0.8257542004996513, "grad_norm": 1.1025595664978027, "learning_rate": 0.00019352953263361485, "loss": 4.7275, "step": 442500 }, { "epoch": 0.8258475060590298, "grad_norm": 1.0343772172927856, "learning_rate": 0.0001935280459578369, "loss": 4.5344, "step": 442550 }, { "epoch": 0.8259408116184083, "grad_norm": 0.9679977893829346, "learning_rate": 0.0001935265591169984, "loss": 4.5776, "step": 442600 }, { "epoch": 0.8260341171777867, "grad_norm": 1.3095729351043701, "learning_rate": 0.00019352507211110195, "loss": 4.5481, "step": 442650 }, { "epoch": 0.8261274227371652, "grad_norm": 0.7980509400367737, "learning_rate": 0.00019352358494015024, "loss": 4.3728, "step": 442700 }, { "epoch": 0.8262207282965437, "grad_norm": 1.1941235065460205, "learning_rate": 0.00019352209760414583, "loss": 4.4981, "step": 442750 }, { "epoch": 0.8263140338559222, "grad_norm": 1.2921700477600098, "learning_rate": 0.0001935206101030914, "loss": 4.5507, "step": 442800 }, { "epoch": 0.8264073394153008, "grad_norm": 1.0769251585006714, "learning_rate": 0.00019351912243698952, "loss": 4.5446, "step": 442850 }, { "epoch": 0.8265006449746792, "grad_norm": 0.9345519542694092, "learning_rate": 0.00019351763460584288, "loss": 4.583, "step": 442900 }, { "epoch": 0.8265939505340577, "grad_norm": 1.0591378211975098, "learning_rate": 0.00019351614660965405, "loss": 4.5456, "step": 442950 }, { "epoch": 0.8266872560934362, "grad_norm": 1.0100284814834595, "learning_rate": 0.00019351465844842566, "loss": 4.4957, "step": 443000 }, { "epoch": 0.8267805616528147, "grad_norm": 1.1169507503509521, "learning_rate": 0.00019351317012216037, "loss": 4.4462, "step": 443050 }, { "epoch": 0.8268738672121931, "grad_norm": 1.3374896049499512, "learning_rate": 0.00019351168163086077, "loss": 4.6489, "step": 443100 }, { "epoch": 0.8269671727715716, "grad_norm": 0.8487604856491089, "learning_rate": 0.00019351019297452954, "loss": 4.5124, "step": 443150 }, { "epoch": 0.8270604783309502, "grad_norm": 0.9288005232810974, "learning_rate": 0.00019350870415316926, "loss": 4.8345, "step": 443200 }, { "epoch": 0.8271537838903287, "grad_norm": 1.0807114839553833, "learning_rate": 0.00019350721516678257, "loss": 4.6849, "step": 443250 }, { "epoch": 0.8272470894497072, "grad_norm": 1.2470729351043701, "learning_rate": 0.0001935057260153721, "loss": 4.6842, "step": 443300 }, { "epoch": 0.8273403950090856, "grad_norm": 1.235500454902649, "learning_rate": 0.00019350423669894047, "loss": 4.5028, "step": 443350 }, { "epoch": 0.8274337005684641, "grad_norm": 0.8499687314033508, "learning_rate": 0.00019350274721749032, "loss": 4.4574, "step": 443400 }, { "epoch": 0.8275270061278426, "grad_norm": 1.2059576511383057, "learning_rate": 0.00019350125757102427, "loss": 4.5983, "step": 443450 }, { "epoch": 0.827620311687221, "grad_norm": 0.9806116819381714, "learning_rate": 0.00019349976775954493, "loss": 4.594, "step": 443500 }, { "epoch": 0.8277136172465996, "grad_norm": 0.8075485229492188, "learning_rate": 0.00019349827778305498, "loss": 4.536, "step": 443550 }, { "epoch": 0.8278069228059781, "grad_norm": 0.8367143869400024, "learning_rate": 0.00019349678764155704, "loss": 4.5917, "step": 443600 }, { "epoch": 0.8279002283653566, "grad_norm": 0.9745396375656128, "learning_rate": 0.00019349529733505368, "loss": 4.3854, "step": 443650 }, { "epoch": 0.8279935339247351, "grad_norm": 1.4076462984085083, "learning_rate": 0.00019349380686354759, "loss": 4.539, "step": 443700 }, { "epoch": 0.8280868394841135, "grad_norm": 1.1576460599899292, "learning_rate": 0.00019349231622704135, "loss": 4.5147, "step": 443750 }, { "epoch": 0.828180145043492, "grad_norm": 0.9356547594070435, "learning_rate": 0.00019349082542553767, "loss": 4.8188, "step": 443800 }, { "epoch": 0.8282734506028705, "grad_norm": 0.9233013987541199, "learning_rate": 0.0001934893344590391, "loss": 4.6146, "step": 443850 }, { "epoch": 0.8283667561622491, "grad_norm": 1.172934651374817, "learning_rate": 0.00019348784332754834, "loss": 4.5364, "step": 443900 }, { "epoch": 0.8284600617216276, "grad_norm": 1.0102208852767944, "learning_rate": 0.00019348635203106796, "loss": 4.5929, "step": 443950 }, { "epoch": 0.828553367281006, "grad_norm": 1.102264165878296, "learning_rate": 0.0001934848605696006, "loss": 4.6786, "step": 444000 }, { "epoch": 0.828553367281006, "eval_loss": 4.721865177154541, "eval_runtime": 232.6378, "eval_samples_per_second": 11.211, "eval_steps_per_second": 11.211, "eval_tts_loss": 7.58550966499899, "step": 444000 }, { "epoch": 0.8286466728403845, "grad_norm": 0.9398770332336426, "learning_rate": 0.00019348336894314894, "loss": 4.5993, "step": 444050 }, { "epoch": 0.828739978399763, "grad_norm": 0.9913604259490967, "learning_rate": 0.0001934818771517156, "loss": 4.244, "step": 444100 }, { "epoch": 0.8288332839591415, "grad_norm": 1.2712340354919434, "learning_rate": 0.00019348038519530317, "loss": 4.6613, "step": 444150 }, { "epoch": 0.8289265895185199, "grad_norm": 0.976055920124054, "learning_rate": 0.0001934788930739143, "loss": 4.5837, "step": 444200 }, { "epoch": 0.8290198950778985, "grad_norm": 1.0469322204589844, "learning_rate": 0.00019347740078755164, "loss": 4.2978, "step": 444250 }, { "epoch": 0.829113200637277, "grad_norm": 0.8743473887443542, "learning_rate": 0.00019347590833621782, "loss": 4.439, "step": 444300 }, { "epoch": 0.8292065061966555, "grad_norm": 1.1809208393096924, "learning_rate": 0.00019347441571991548, "loss": 4.388, "step": 444350 }, { "epoch": 0.829299811756034, "grad_norm": 1.614380121231079, "learning_rate": 0.0001934729229386472, "loss": 4.3623, "step": 444400 }, { "epoch": 0.8293931173154124, "grad_norm": 0.9954262971878052, "learning_rate": 0.00019347142999241572, "loss": 4.6601, "step": 444450 }, { "epoch": 0.8294864228747909, "grad_norm": 0.9364942312240601, "learning_rate": 0.00019346993688122358, "loss": 4.693, "step": 444500 }, { "epoch": 0.8295797284341694, "grad_norm": 1.0278654098510742, "learning_rate": 0.00019346844360507344, "loss": 4.6674, "step": 444550 }, { "epoch": 0.829673033993548, "grad_norm": 1.1544060707092285, "learning_rate": 0.00019346695016396795, "loss": 4.4789, "step": 444600 }, { "epoch": 0.8297663395529264, "grad_norm": 1.6491734981536865, "learning_rate": 0.00019346545655790976, "loss": 4.7639, "step": 444650 }, { "epoch": 0.8298596451123049, "grad_norm": 0.6557138562202454, "learning_rate": 0.00019346396278690145, "loss": 4.3905, "step": 444700 }, { "epoch": 0.8299529506716834, "grad_norm": 1.094346523284912, "learning_rate": 0.0001934624688509457, "loss": 4.4474, "step": 444750 }, { "epoch": 0.8300462562310619, "grad_norm": 1.0265722274780273, "learning_rate": 0.00019346097475004515, "loss": 4.5761, "step": 444800 }, { "epoch": 0.8301395617904404, "grad_norm": 1.1581714153289795, "learning_rate": 0.00019345948048420244, "loss": 4.4582, "step": 444850 }, { "epoch": 0.8302328673498188, "grad_norm": 1.1250214576721191, "learning_rate": 0.00019345798605342016, "loss": 4.5792, "step": 444900 }, { "epoch": 0.8303261729091974, "grad_norm": 1.219117283821106, "learning_rate": 0.00019345649145770098, "loss": 4.642, "step": 444950 }, { "epoch": 0.8304194784685759, "grad_norm": 0.9941939115524292, "learning_rate": 0.00019345499669704754, "loss": 4.4172, "step": 445000 }, { "epoch": 0.8305127840279544, "grad_norm": 1.004757046699524, "learning_rate": 0.0001934535017714625, "loss": 4.4208, "step": 445050 }, { "epoch": 0.8306060895873328, "grad_norm": 1.1099998950958252, "learning_rate": 0.00019345200668094846, "loss": 4.6247, "step": 445100 }, { "epoch": 0.8306993951467113, "grad_norm": 1.1278831958770752, "learning_rate": 0.00019345051142550808, "loss": 4.6416, "step": 445150 }, { "epoch": 0.8307927007060898, "grad_norm": 1.3656575679779053, "learning_rate": 0.00019344901600514396, "loss": 4.7044, "step": 445200 }, { "epoch": 0.8308860062654683, "grad_norm": 1.220974326133728, "learning_rate": 0.00019344752041985878, "loss": 4.3872, "step": 445250 }, { "epoch": 0.8309793118248469, "grad_norm": 1.3309504985809326, "learning_rate": 0.00019344602466965518, "loss": 4.5627, "step": 445300 }, { "epoch": 0.8310726173842253, "grad_norm": 1.1383757591247559, "learning_rate": 0.0001934445287545358, "loss": 4.5751, "step": 445350 }, { "epoch": 0.8311659229436038, "grad_norm": 1.1477620601654053, "learning_rate": 0.00019344303267450322, "loss": 4.5198, "step": 445400 }, { "epoch": 0.8312592285029823, "grad_norm": 0.9861552119255066, "learning_rate": 0.00019344153642956018, "loss": 4.247, "step": 445450 }, { "epoch": 0.8313525340623608, "grad_norm": 0.9693005084991455, "learning_rate": 0.00019344004001970924, "loss": 4.6788, "step": 445500 }, { "epoch": 0.8314458396217392, "grad_norm": 1.1747108697891235, "learning_rate": 0.0001934385434449531, "loss": 4.5564, "step": 445550 }, { "epoch": 0.8315391451811177, "grad_norm": 0.9544690251350403, "learning_rate": 0.00019343704670529433, "loss": 4.4388, "step": 445600 }, { "epoch": 0.8316324507404963, "grad_norm": 1.0575664043426514, "learning_rate": 0.00019343554980073567, "loss": 4.4988, "step": 445650 }, { "epoch": 0.8317257562998748, "grad_norm": 1.320414662361145, "learning_rate": 0.00019343405273127966, "loss": 4.6764, "step": 445700 }, { "epoch": 0.8318190618592533, "grad_norm": 1.0772422552108765, "learning_rate": 0.000193432555496929, "loss": 4.5405, "step": 445750 }, { "epoch": 0.8319123674186317, "grad_norm": 0.8643651604652405, "learning_rate": 0.00019343105809768631, "loss": 4.3712, "step": 445800 }, { "epoch": 0.8320056729780102, "grad_norm": 1.0002784729003906, "learning_rate": 0.00019342956053355429, "loss": 4.6529, "step": 445850 }, { "epoch": 0.8320989785373887, "grad_norm": 1.0368218421936035, "learning_rate": 0.00019342806280453546, "loss": 4.5132, "step": 445900 }, { "epoch": 0.8321922840967672, "grad_norm": 1.2993515729904175, "learning_rate": 0.0001934265649106326, "loss": 4.7953, "step": 445950 }, { "epoch": 0.8322855896561457, "grad_norm": 0.9636172652244568, "learning_rate": 0.00019342506685184825, "loss": 4.433, "step": 446000 }, { "epoch": 0.8323788952155242, "grad_norm": 1.0712823867797852, "learning_rate": 0.00019342356862818511, "loss": 4.4565, "step": 446050 }, { "epoch": 0.8324722007749027, "grad_norm": 1.0290355682373047, "learning_rate": 0.00019342207023964584, "loss": 4.5186, "step": 446100 }, { "epoch": 0.8325655063342812, "grad_norm": 0.8416425585746765, "learning_rate": 0.000193420571686233, "loss": 4.6492, "step": 446150 }, { "epoch": 0.8326588118936596, "grad_norm": 1.026302695274353, "learning_rate": 0.00019341907296794932, "loss": 4.4113, "step": 446200 }, { "epoch": 0.8327521174530381, "grad_norm": 1.206878423690796, "learning_rate": 0.00019341757408479742, "loss": 4.4899, "step": 446250 }, { "epoch": 0.8328454230124166, "grad_norm": 0.9927154779434204, "learning_rate": 0.00019341607503677992, "loss": 4.6825, "step": 446300 }, { "epoch": 0.8329387285717952, "grad_norm": 1.098044753074646, "learning_rate": 0.0001934145758238995, "loss": 4.584, "step": 446350 }, { "epoch": 0.8330320341311737, "grad_norm": 1.096203088760376, "learning_rate": 0.00019341307644615878, "loss": 4.5713, "step": 446400 }, { "epoch": 0.8331253396905521, "grad_norm": 1.0846002101898193, "learning_rate": 0.00019341157690356043, "loss": 4.5434, "step": 446450 }, { "epoch": 0.8332186452499306, "grad_norm": 1.1647289991378784, "learning_rate": 0.00019341007719610707, "loss": 4.4887, "step": 446500 }, { "epoch": 0.8333119508093091, "grad_norm": 1.1023021936416626, "learning_rate": 0.00019340857732380134, "loss": 4.4969, "step": 446550 }, { "epoch": 0.8334052563686876, "grad_norm": 0.9889746308326721, "learning_rate": 0.00019340707728664596, "loss": 4.468, "step": 446600 }, { "epoch": 0.833498561928066, "grad_norm": 1.0913546085357666, "learning_rate": 0.00019340557708464348, "loss": 4.5224, "step": 446650 }, { "epoch": 0.8335918674874445, "grad_norm": 1.1410549879074097, "learning_rate": 0.0001934040767177966, "loss": 4.5776, "step": 446700 }, { "epoch": 0.8336851730468231, "grad_norm": 0.9060096740722656, "learning_rate": 0.00019340257618610798, "loss": 4.5885, "step": 446750 }, { "epoch": 0.8337784786062016, "grad_norm": 1.1176276206970215, "learning_rate": 0.00019340107548958022, "loss": 4.4966, "step": 446800 }, { "epoch": 0.83387178416558, "grad_norm": 1.2471281290054321, "learning_rate": 0.00019339957462821602, "loss": 4.4346, "step": 446850 }, { "epoch": 0.8339650897249585, "grad_norm": 1.1243484020233154, "learning_rate": 0.00019339807360201802, "loss": 4.7867, "step": 446900 }, { "epoch": 0.834058395284337, "grad_norm": 1.058711051940918, "learning_rate": 0.00019339657241098878, "loss": 4.8493, "step": 446950 }, { "epoch": 0.8341517008437155, "grad_norm": 1.22663152217865, "learning_rate": 0.0001933950710551311, "loss": 4.4856, "step": 447000 }, { "epoch": 0.8341517008437155, "eval_loss": 4.712674140930176, "eval_runtime": 232.2896, "eval_samples_per_second": 11.227, "eval_steps_per_second": 11.227, "eval_tts_loss": 7.630600184322181, "step": 447000 }, { "epoch": 0.834245006403094, "grad_norm": 1.2073787450790405, "learning_rate": 0.00019339356953444753, "loss": 4.6652, "step": 447050 }, { "epoch": 0.8343383119624725, "grad_norm": 1.0807472467422485, "learning_rate": 0.0001933920678489407, "loss": 4.6302, "step": 447100 }, { "epoch": 0.834431617521851, "grad_norm": 0.8505199551582336, "learning_rate": 0.00019339056599861336, "loss": 4.6837, "step": 447150 }, { "epoch": 0.8345249230812295, "grad_norm": 1.016205906867981, "learning_rate": 0.00019338906398346807, "loss": 4.6417, "step": 447200 }, { "epoch": 0.834618228640608, "grad_norm": 1.0800621509552002, "learning_rate": 0.00019338756180350755, "loss": 4.5582, "step": 447250 }, { "epoch": 0.8347115341999864, "grad_norm": 0.9196884036064148, "learning_rate": 0.0001933860594587344, "loss": 4.4846, "step": 447300 }, { "epoch": 0.8348048397593649, "grad_norm": 1.1872416734695435, "learning_rate": 0.00019338455694915127, "loss": 4.5343, "step": 447350 }, { "epoch": 0.8348981453187434, "grad_norm": 1.0017876625061035, "learning_rate": 0.00019338305427476083, "loss": 4.3577, "step": 447400 }, { "epoch": 0.834991450878122, "grad_norm": 1.4030224084854126, "learning_rate": 0.00019338155143556573, "loss": 4.6123, "step": 447450 }, { "epoch": 0.8350847564375005, "grad_norm": 1.8109701871871948, "learning_rate": 0.00019338004843156865, "loss": 4.5135, "step": 447500 }, { "epoch": 0.8351780619968789, "grad_norm": 1.2797409296035767, "learning_rate": 0.00019337854526277218, "loss": 4.4158, "step": 447550 }, { "epoch": 0.8352713675562574, "grad_norm": 0.9986948370933533, "learning_rate": 0.00019337704192917904, "loss": 4.6389, "step": 447600 }, { "epoch": 0.8353646731156359, "grad_norm": 1.0423827171325684, "learning_rate": 0.00019337553843079184, "loss": 4.4192, "step": 447650 }, { "epoch": 0.8354579786750144, "grad_norm": 0.9322047233581543, "learning_rate": 0.00019337403476761325, "loss": 4.5607, "step": 447700 }, { "epoch": 0.8355512842343928, "grad_norm": 1.076505422592163, "learning_rate": 0.0001933725309396459, "loss": 4.8305, "step": 447750 }, { "epoch": 0.8356445897937714, "grad_norm": 1.1497913599014282, "learning_rate": 0.00019337102694689246, "loss": 4.4388, "step": 447800 }, { "epoch": 0.8357378953531499, "grad_norm": 1.237391710281372, "learning_rate": 0.00019336952278935563, "loss": 4.6497, "step": 447850 }, { "epoch": 0.8358312009125284, "grad_norm": 1.1664336919784546, "learning_rate": 0.00019336801846703798, "loss": 4.4486, "step": 447900 }, { "epoch": 0.8359245064719069, "grad_norm": 1.0214166641235352, "learning_rate": 0.0001933665139799422, "loss": 4.6264, "step": 447950 }, { "epoch": 0.8360178120312853, "grad_norm": 1.5139281749725342, "learning_rate": 0.00019336500932807098, "loss": 4.7044, "step": 448000 }, { "epoch": 0.8361111175906638, "grad_norm": 1.286423683166504, "learning_rate": 0.00019336350451142694, "loss": 4.4449, "step": 448050 }, { "epoch": 0.8362044231500423, "grad_norm": 0.8456434011459351, "learning_rate": 0.00019336199953001273, "loss": 4.3987, "step": 448100 }, { "epoch": 0.8362977287094209, "grad_norm": 0.9992003440856934, "learning_rate": 0.00019336049438383103, "loss": 4.6912, "step": 448150 }, { "epoch": 0.8363910342687993, "grad_norm": 1.1905733346939087, "learning_rate": 0.00019335898907288447, "loss": 4.6875, "step": 448200 }, { "epoch": 0.8364843398281778, "grad_norm": 0.9549011588096619, "learning_rate": 0.00019335748359717574, "loss": 4.6345, "step": 448250 }, { "epoch": 0.8365776453875563, "grad_norm": 1.0515804290771484, "learning_rate": 0.00019335597795670746, "loss": 4.4891, "step": 448300 }, { "epoch": 0.8366709509469348, "grad_norm": 1.0489537715911865, "learning_rate": 0.0001933544721514823, "loss": 4.7241, "step": 448350 }, { "epoch": 0.8367642565063133, "grad_norm": 1.0642502307891846, "learning_rate": 0.00019335296618150296, "loss": 4.6523, "step": 448400 }, { "epoch": 0.8368575620656917, "grad_norm": 1.2421493530273438, "learning_rate": 0.00019335146004677203, "loss": 4.5251, "step": 448450 }, { "epoch": 0.8369508676250703, "grad_norm": 1.2859892845153809, "learning_rate": 0.0001933499537472922, "loss": 4.5942, "step": 448500 }, { "epoch": 0.8370441731844488, "grad_norm": 1.2976332902908325, "learning_rate": 0.00019334844728306614, "loss": 4.4986, "step": 448550 }, { "epoch": 0.8371374787438273, "grad_norm": 0.7612985372543335, "learning_rate": 0.00019334694065409647, "loss": 4.601, "step": 448600 }, { "epoch": 0.8372307843032057, "grad_norm": 1.1025867462158203, "learning_rate": 0.00019334543386038592, "loss": 4.4959, "step": 448650 }, { "epoch": 0.8373240898625842, "grad_norm": 1.1251939535140991, "learning_rate": 0.00019334392690193708, "loss": 4.6645, "step": 448700 }, { "epoch": 0.8374173954219627, "grad_norm": 1.149271845817566, "learning_rate": 0.0001933424197787526, "loss": 4.5686, "step": 448750 }, { "epoch": 0.8375107009813412, "grad_norm": 1.1676266193389893, "learning_rate": 0.00019334091249083524, "loss": 4.5399, "step": 448800 }, { "epoch": 0.8376040065407198, "grad_norm": 0.7529256343841553, "learning_rate": 0.00019333940503818756, "loss": 4.4607, "step": 448850 }, { "epoch": 0.8376973121000982, "grad_norm": 1.1759973764419556, "learning_rate": 0.00019333789742081222, "loss": 4.3374, "step": 448900 }, { "epoch": 0.8377906176594767, "grad_norm": 1.1348012685775757, "learning_rate": 0.00019333638963871198, "loss": 4.5555, "step": 448950 }, { "epoch": 0.8378839232188552, "grad_norm": 1.3209272623062134, "learning_rate": 0.00019333488169188937, "loss": 4.5982, "step": 449000 }, { "epoch": 0.8379772287782337, "grad_norm": 1.3228648900985718, "learning_rate": 0.00019333337358034715, "loss": 4.4296, "step": 449050 }, { "epoch": 0.8380705343376121, "grad_norm": 1.0187352895736694, "learning_rate": 0.00019333186530408797, "loss": 4.7661, "step": 449100 }, { "epoch": 0.8381638398969906, "grad_norm": 1.0453397035598755, "learning_rate": 0.00019333035686311444, "loss": 4.5145, "step": 449150 }, { "epoch": 0.8382571454563692, "grad_norm": 1.3582311868667603, "learning_rate": 0.00019332884825742924, "loss": 4.6269, "step": 449200 }, { "epoch": 0.8383504510157477, "grad_norm": 1.093062400817871, "learning_rate": 0.00019332733948703506, "loss": 4.619, "step": 449250 }, { "epoch": 0.8384437565751262, "grad_norm": 0.9017040133476257, "learning_rate": 0.00019332583055193454, "loss": 4.6322, "step": 449300 }, { "epoch": 0.8385370621345046, "grad_norm": 1.103871464729309, "learning_rate": 0.00019332432145213038, "loss": 4.5635, "step": 449350 }, { "epoch": 0.8386303676938831, "grad_norm": 1.134921908378601, "learning_rate": 0.0001933228121876252, "loss": 4.4465, "step": 449400 }, { "epoch": 0.8387236732532616, "grad_norm": 1.078826665878296, "learning_rate": 0.00019332130275842167, "loss": 4.466, "step": 449450 }, { "epoch": 0.8388169788126401, "grad_norm": 1.0170928239822388, "learning_rate": 0.00019331979316452243, "loss": 4.4511, "step": 449500 }, { "epoch": 0.8389102843720186, "grad_norm": 1.1686103343963623, "learning_rate": 0.0001933182834059302, "loss": 4.4879, "step": 449550 }, { "epoch": 0.8390035899313971, "grad_norm": 1.040514349937439, "learning_rate": 0.00019331677348264766, "loss": 4.5254, "step": 449600 }, { "epoch": 0.8390968954907756, "grad_norm": 0.905274510383606, "learning_rate": 0.0001933152633946774, "loss": 4.6875, "step": 449650 }, { "epoch": 0.8391902010501541, "grad_norm": 0.9785981774330139, "learning_rate": 0.0001933137531420221, "loss": 4.6597, "step": 449700 }, { "epoch": 0.8392835066095325, "grad_norm": 1.2745859622955322, "learning_rate": 0.00019331224272468446, "loss": 4.4091, "step": 449750 }, { "epoch": 0.839376812168911, "grad_norm": 1.7750924825668335, "learning_rate": 0.00019331073214266716, "loss": 4.564, "step": 449800 }, { "epoch": 0.8394701177282895, "grad_norm": 1.0283859968185425, "learning_rate": 0.0001933092213959728, "loss": 4.5776, "step": 449850 }, { "epoch": 0.8395634232876681, "grad_norm": 1.2207344770431519, "learning_rate": 0.0001933077104846041, "loss": 4.6865, "step": 449900 }, { "epoch": 0.8396567288470466, "grad_norm": 1.1184113025665283, "learning_rate": 0.00019330619940856369, "loss": 4.6266, "step": 449950 }, { "epoch": 0.839750034406425, "grad_norm": 1.1503829956054688, "learning_rate": 0.00019330468816785426, "loss": 4.5217, "step": 450000 }, { "epoch": 0.839750034406425, "eval_loss": 4.710883617401123, "eval_runtime": 232.9307, "eval_samples_per_second": 11.196, "eval_steps_per_second": 11.196, "eval_tts_loss": 7.568327237897627, "step": 450000 }, { "epoch": 0.8398433399658035, "grad_norm": 1.1530593633651733, "learning_rate": 0.0001933031767624785, "loss": 4.6247, "step": 450050 }, { "epoch": 0.839936645525182, "grad_norm": 0.7560440301895142, "learning_rate": 0.00019330166519243905, "loss": 4.5636, "step": 450100 }, { "epoch": 0.8400299510845605, "grad_norm": 1.126522183418274, "learning_rate": 0.00019330015345773855, "loss": 4.6178, "step": 450150 }, { "epoch": 0.8401232566439389, "grad_norm": 1.2579351663589478, "learning_rate": 0.0001932986415583797, "loss": 4.8406, "step": 450200 }, { "epoch": 0.8402165622033175, "grad_norm": 1.1700242757797241, "learning_rate": 0.00019329712949436517, "loss": 4.5789, "step": 450250 }, { "epoch": 0.840309867762696, "grad_norm": 1.1610465049743652, "learning_rate": 0.00019329561726569763, "loss": 4.7092, "step": 450300 }, { "epoch": 0.8404031733220745, "grad_norm": 0.921472430229187, "learning_rate": 0.0001932941048723797, "loss": 4.4653, "step": 450350 }, { "epoch": 0.840496478881453, "grad_norm": 1.2933341264724731, "learning_rate": 0.00019329259231441414, "loss": 4.5035, "step": 450400 }, { "epoch": 0.8405897844408314, "grad_norm": 0.9860426187515259, "learning_rate": 0.00019329107959180358, "loss": 4.5394, "step": 450450 }, { "epoch": 0.8406830900002099, "grad_norm": 1.5691264867782593, "learning_rate": 0.00019328956670455064, "loss": 4.4438, "step": 450500 }, { "epoch": 0.8407763955595884, "grad_norm": 1.0524669885635376, "learning_rate": 0.00019328805365265806, "loss": 4.5818, "step": 450550 }, { "epoch": 0.840869701118967, "grad_norm": 1.1248055696487427, "learning_rate": 0.00019328654043612844, "loss": 4.6727, "step": 450600 }, { "epoch": 0.8409630066783454, "grad_norm": 0.8522668480873108, "learning_rate": 0.00019328502705496451, "loss": 4.4992, "step": 450650 }, { "epoch": 0.8410563122377239, "grad_norm": 0.8684114217758179, "learning_rate": 0.00019328351350916893, "loss": 4.5486, "step": 450700 }, { "epoch": 0.8411496177971024, "grad_norm": 1.2275203466415405, "learning_rate": 0.00019328199979874437, "loss": 4.5316, "step": 450750 }, { "epoch": 0.8412429233564809, "grad_norm": 1.0412169694900513, "learning_rate": 0.00019328048592369345, "loss": 4.2662, "step": 450800 }, { "epoch": 0.8413362289158594, "grad_norm": 0.8822068572044373, "learning_rate": 0.00019327897188401894, "loss": 4.6678, "step": 450850 }, { "epoch": 0.8414295344752378, "grad_norm": 1.0500564575195312, "learning_rate": 0.00019327745767972344, "loss": 4.6266, "step": 450900 }, { "epoch": 0.8415228400346164, "grad_norm": 1.1681212186813354, "learning_rate": 0.00019327594331080961, "loss": 4.5943, "step": 450950 }, { "epoch": 0.8416161455939949, "grad_norm": 1.0938808917999268, "learning_rate": 0.00019327442877728018, "loss": 4.5703, "step": 451000 }, { "epoch": 0.8417094511533734, "grad_norm": 0.9540387988090515, "learning_rate": 0.0001932729140791378, "loss": 4.5122, "step": 451050 }, { "epoch": 0.8418027567127518, "grad_norm": 0.9375895857810974, "learning_rate": 0.00019327139921638514, "loss": 4.5586, "step": 451100 }, { "epoch": 0.8418960622721303, "grad_norm": 1.1446915864944458, "learning_rate": 0.00019326988418902485, "loss": 4.5502, "step": 451150 }, { "epoch": 0.8419893678315088, "grad_norm": 1.0916074514389038, "learning_rate": 0.00019326836899705965, "loss": 4.654, "step": 451200 }, { "epoch": 0.8420826733908873, "grad_norm": 1.0063576698303223, "learning_rate": 0.00019326685364049219, "loss": 4.4077, "step": 451250 }, { "epoch": 0.8421759789502659, "grad_norm": 1.0716850757598877, "learning_rate": 0.0001932653381193251, "loss": 4.6463, "step": 451300 }, { "epoch": 0.8422692845096443, "grad_norm": 1.2872897386550903, "learning_rate": 0.00019326382243356114, "loss": 4.4904, "step": 451350 }, { "epoch": 0.8423625900690228, "grad_norm": 0.8096011877059937, "learning_rate": 0.00019326230658320292, "loss": 4.6754, "step": 451400 }, { "epoch": 0.8424558956284013, "grad_norm": 1.3478927612304688, "learning_rate": 0.00019326079056825317, "loss": 4.5584, "step": 451450 }, { "epoch": 0.8425492011877798, "grad_norm": 1.3702021837234497, "learning_rate": 0.0001932592743887145, "loss": 4.5173, "step": 451500 }, { "epoch": 0.8426425067471582, "grad_norm": 0.9578404426574707, "learning_rate": 0.00019325775804458965, "loss": 4.7119, "step": 451550 }, { "epoch": 0.8427358123065367, "grad_norm": 1.2032324075698853, "learning_rate": 0.00019325624153588127, "loss": 4.5057, "step": 451600 }, { "epoch": 0.8428291178659153, "grad_norm": 1.288559079170227, "learning_rate": 0.000193254724862592, "loss": 4.5697, "step": 451650 }, { "epoch": 0.8429224234252938, "grad_norm": 0.963632345199585, "learning_rate": 0.00019325320802472456, "loss": 4.4117, "step": 451700 }, { "epoch": 0.8430157289846723, "grad_norm": 1.1363961696624756, "learning_rate": 0.00019325169102228166, "loss": 4.5567, "step": 451750 }, { "epoch": 0.8431090345440507, "grad_norm": 1.131595492362976, "learning_rate": 0.00019325017385526586, "loss": 4.3638, "step": 451800 }, { "epoch": 0.8432023401034292, "grad_norm": 0.9933236241340637, "learning_rate": 0.00019324865652367996, "loss": 4.4048, "step": 451850 }, { "epoch": 0.8432956456628077, "grad_norm": 0.9534032344818115, "learning_rate": 0.00019324713902752657, "loss": 4.5009, "step": 451900 }, { "epoch": 0.8433889512221862, "grad_norm": 1.0373858213424683, "learning_rate": 0.00019324562136680839, "loss": 4.7383, "step": 451950 }, { "epoch": 0.8434822567815646, "grad_norm": 0.8199633955955505, "learning_rate": 0.00019324410354152808, "loss": 4.5314, "step": 452000 }, { "epoch": 0.8435755623409432, "grad_norm": 1.0837599039077759, "learning_rate": 0.00019324258555168837, "loss": 4.5894, "step": 452050 }, { "epoch": 0.8436688679003217, "grad_norm": 1.2364952564239502, "learning_rate": 0.00019324106739729188, "loss": 4.6582, "step": 452100 }, { "epoch": 0.8437621734597002, "grad_norm": 0.7789540886878967, "learning_rate": 0.00019323954907834133, "loss": 4.5806, "step": 452150 }, { "epoch": 0.8438554790190786, "grad_norm": 1.2450273036956787, "learning_rate": 0.00019323803059483936, "loss": 4.6015, "step": 452200 }, { "epoch": 0.8439487845784571, "grad_norm": 1.2874577045440674, "learning_rate": 0.00019323651194678868, "loss": 4.5458, "step": 452250 }, { "epoch": 0.8440420901378356, "grad_norm": 1.3656573295593262, "learning_rate": 0.00019323499313419196, "loss": 4.6391, "step": 452300 }, { "epoch": 0.8441353956972141, "grad_norm": 1.064369797706604, "learning_rate": 0.00019323347415705188, "loss": 4.4416, "step": 452350 }, { "epoch": 0.8442287012565927, "grad_norm": 1.1734198331832886, "learning_rate": 0.00019323195501537115, "loss": 4.362, "step": 452400 }, { "epoch": 0.8443220068159711, "grad_norm": 1.1021287441253662, "learning_rate": 0.00019323043570915237, "loss": 4.678, "step": 452450 }, { "epoch": 0.8444153123753496, "grad_norm": 1.1119120121002197, "learning_rate": 0.00019322891623839834, "loss": 4.7595, "step": 452500 }, { "epoch": 0.8445086179347281, "grad_norm": 1.0412399768829346, "learning_rate": 0.00019322739660311165, "loss": 4.6088, "step": 452550 }, { "epoch": 0.8446019234941066, "grad_norm": 0.7403721213340759, "learning_rate": 0.00019322587680329497, "loss": 4.5932, "step": 452600 }, { "epoch": 0.844695229053485, "grad_norm": 1.14961576461792, "learning_rate": 0.00019322435683895108, "loss": 4.5157, "step": 452650 }, { "epoch": 0.8447885346128635, "grad_norm": 1.1009234189987183, "learning_rate": 0.00019322283671008257, "loss": 4.5013, "step": 452700 }, { "epoch": 0.8448818401722421, "grad_norm": 0.8400731086730957, "learning_rate": 0.00019322131641669218, "loss": 4.5795, "step": 452750 }, { "epoch": 0.8449751457316206, "grad_norm": 1.0518465042114258, "learning_rate": 0.00019321979595878254, "loss": 4.6606, "step": 452800 }, { "epoch": 0.8450684512909991, "grad_norm": 1.2242203950881958, "learning_rate": 0.00019321827533635638, "loss": 4.4912, "step": 452850 }, { "epoch": 0.8451617568503775, "grad_norm": 1.163169026374817, "learning_rate": 0.00019321675454941634, "loss": 4.5729, "step": 452900 }, { "epoch": 0.845255062409756, "grad_norm": 1.1667548418045044, "learning_rate": 0.00019321523359796516, "loss": 4.5619, "step": 452950 }, { "epoch": 0.8453483679691345, "grad_norm": 1.4125514030456543, "learning_rate": 0.00019321371248200552, "loss": 4.5351, "step": 453000 }, { "epoch": 0.8453483679691345, "eval_loss": 4.710817813873291, "eval_runtime": 233.9958, "eval_samples_per_second": 11.145, "eval_steps_per_second": 11.145, "eval_tts_loss": 7.61211348108753, "step": 453000 } ], "logging_steps": 50, "max_steps": 3751118, "num_input_tokens_seen": 0, "num_train_epochs": 7, "save_steps": 3000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.206774371991683e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }