| { | |
| "best_metric": 0.2958612975391499, | |
| "best_model_checkpoint": "./save/jonatasgrosman/wav2vec2-large-xlsr-53-chinese-zh-cn/checkpoint-35728", | |
| "epoch": 30.0, | |
| "global_step": 36960, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.993235930735931e-05, | |
| "loss": 3.1964, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.9864718614718616e-05, | |
| "loss": 2.1923, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.979707792207792e-05, | |
| "loss": 1.9162, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.972943722943723e-05, | |
| "loss": 1.8832, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.966179653679654e-05, | |
| "loss": 1.8216, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.959415584415585e-05, | |
| "loss": 1.7282, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.952651515151515e-05, | |
| "loss": 1.6915, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.945887445887446e-05, | |
| "loss": 1.7018, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.939123376623377e-05, | |
| "loss": 1.5646, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.932359307359308e-05, | |
| "loss": 1.5312, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.925595238095238e-05, | |
| "loss": 1.5983, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.918831168831169e-05, | |
| "loss": 1.5388, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.9120670995671e-05, | |
| "loss": 1.491, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.905303030303031e-05, | |
| "loss": 1.4054, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.898538961038962e-05, | |
| "loss": 1.4775, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.8917748917748915e-05, | |
| "loss": 1.4369, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.8850108225108225e-05, | |
| "loss": 1.4013, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.8782467532467536e-05, | |
| "loss": 1.3465, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.8714826839826846e-05, | |
| "loss": 1.4206, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.864718614718615e-05, | |
| "loss": 1.2723, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.857954545454545e-05, | |
| "loss": 1.2339, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.8511904761904764e-05, | |
| "loss": 1.2496, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.8444264069264074e-05, | |
| "loss": 1.2337, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.8376623376623384e-05, | |
| "loss": 1.2815, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_cer": 0.40496900379643425, | |
| "eval_loss": 1.455051064491272, | |
| "eval_mer": 0.42002237136465326, | |
| "eval_runtime": 216.1491, | |
| "eval_samples_per_second": 5.214, | |
| "eval_steps_per_second": 0.652, | |
| "step": 1232 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.830898268398268e-05, | |
| "loss": 1.1819, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 4.824134199134199e-05, | |
| "loss": 1.1662, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 4.81737012987013e-05, | |
| "loss": 1.1465, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 4.810606060606061e-05, | |
| "loss": 1.011, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 4.8038419913419916e-05, | |
| "loss": 1.0133, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 4.797077922077922e-05, | |
| "loss": 1.072, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 4.790313852813853e-05, | |
| "loss": 1.1135, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 4.783549783549784e-05, | |
| "loss": 0.9754, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 4.7767857142857144e-05, | |
| "loss": 1.052, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 4.770021645021645e-05, | |
| "loss": 1.0488, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 4.763257575757576e-05, | |
| "loss": 1.0345, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 4.756493506493507e-05, | |
| "loss": 1.0513, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 4.749729437229438e-05, | |
| "loss": 0.9665, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 4.742965367965368e-05, | |
| "loss": 1.0715, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 4.7362012987012986e-05, | |
| "loss": 1.0641, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 4.72943722943723e-05, | |
| "loss": 0.9736, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 4.722673160173161e-05, | |
| "loss": 0.9678, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 4.715909090909091e-05, | |
| "loss": 1.0556, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 4.7091450216450214e-05, | |
| "loss": 0.9854, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 4.7023809523809525e-05, | |
| "loss": 0.9546, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 4.6956168831168835e-05, | |
| "loss": 1.0219, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 4.6888528138528146e-05, | |
| "loss": 0.9491, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 4.682088744588744e-05, | |
| "loss": 1.0235, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 4.675324675324675e-05, | |
| "loss": 1.0252, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 4.668560606060606e-05, | |
| "loss": 0.9278, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_cer": 0.3652265846508722, | |
| "eval_loss": 1.3382420539855957, | |
| "eval_mer": 0.3908596995845318, | |
| "eval_runtime": 218.6583, | |
| "eval_samples_per_second": 5.154, | |
| "eval_steps_per_second": 0.645, | |
| "step": 2464 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 4.6617965367965374e-05, | |
| "loss": 0.8101, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 4.655032467532468e-05, | |
| "loss": 0.7881, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 4.648268398268398e-05, | |
| "loss": 0.8176, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 4.641504329004329e-05, | |
| "loss": 0.8024, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 4.63474025974026e-05, | |
| "loss": 0.7591, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 4.627976190476191e-05, | |
| "loss": 0.86, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 4.621212121212121e-05, | |
| "loss": 0.7364, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 4.614448051948052e-05, | |
| "loss": 0.8359, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 4.607683982683983e-05, | |
| "loss": 0.8329, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 4.600919913419914e-05, | |
| "loss": 0.8302, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 4.5941558441558444e-05, | |
| "loss": 0.7797, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 4.587391774891775e-05, | |
| "loss": 0.8389, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 4.580627705627706e-05, | |
| "loss": 0.8095, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 4.573863636363637e-05, | |
| "loss": 0.7784, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 4.567099567099568e-05, | |
| "loss": 0.7606, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 4.5603354978354976e-05, | |
| "loss": 0.8153, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 4.5535714285714286e-05, | |
| "loss": 0.7842, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.5468073593073596e-05, | |
| "loss": 0.8033, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.540043290043291e-05, | |
| "loss": 0.7306, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.533279220779221e-05, | |
| "loss": 0.7315, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 4.5265151515151514e-05, | |
| "loss": 0.7986, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 4.5197510822510825e-05, | |
| "loss": 0.8575, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 4.5129870129870135e-05, | |
| "loss": 0.8591, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 4.506222943722944e-05, | |
| "loss": 0.7015, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_cer": 0.3540295064635494, | |
| "eval_loss": 1.530730128288269, | |
| "eval_mer": 0.39429530201342283, | |
| "eval_runtime": 217.1668, | |
| "eval_samples_per_second": 5.19, | |
| "eval_steps_per_second": 0.649, | |
| "step": 3696 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 4.499458874458874e-05, | |
| "loss": 0.8337, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 3.04, | |
| "learning_rate": 4.492694805194805e-05, | |
| "loss": 0.6677, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 3.08, | |
| "learning_rate": 4.485930735930736e-05, | |
| "loss": 0.6976, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 3.12, | |
| "learning_rate": 4.4791666666666673e-05, | |
| "loss": 0.6131, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 3.17, | |
| "learning_rate": 4.472402597402598e-05, | |
| "loss": 0.6671, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 3.21, | |
| "learning_rate": 4.465638528138528e-05, | |
| "loss": 0.6474, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 3.25, | |
| "learning_rate": 4.458874458874459e-05, | |
| "loss": 0.6593, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 4.45211038961039e-05, | |
| "loss": 0.6617, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "learning_rate": 4.4453463203463205e-05, | |
| "loss": 0.607, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 3.37, | |
| "learning_rate": 4.438582251082251e-05, | |
| "loss": 0.6391, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 3.41, | |
| "learning_rate": 4.431818181818182e-05, | |
| "loss": 0.5903, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 3.45, | |
| "learning_rate": 4.425054112554113e-05, | |
| "loss": 0.6669, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 4.418290043290044e-05, | |
| "loss": 0.6639, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "learning_rate": 4.411525974025974e-05, | |
| "loss": 0.5948, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 3.57, | |
| "learning_rate": 4.404761904761905e-05, | |
| "loss": 0.6522, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 3.61, | |
| "learning_rate": 4.397997835497836e-05, | |
| "loss": 0.6048, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 3.65, | |
| "learning_rate": 4.391233766233767e-05, | |
| "loss": 0.5718, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 3.69, | |
| "learning_rate": 4.384469696969697e-05, | |
| "loss": 0.6911, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 3.73, | |
| "learning_rate": 4.3777056277056275e-05, | |
| "loss": 0.6172, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 3.77, | |
| "learning_rate": 4.3709415584415586e-05, | |
| "loss": 0.6573, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 3.81, | |
| "learning_rate": 4.3641774891774896e-05, | |
| "loss": 0.671, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 3.86, | |
| "learning_rate": 4.357413419913421e-05, | |
| "loss": 0.6021, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "learning_rate": 4.3506493506493503e-05, | |
| "loss": 0.5882, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 3.94, | |
| "learning_rate": 4.3438852813852814e-05, | |
| "loss": 0.6395, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 3.98, | |
| "learning_rate": 4.3371212121212124e-05, | |
| "loss": 0.6266, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_cer": 0.3161612763708011, | |
| "eval_loss": 1.3133615255355835, | |
| "eval_mer": 0.35802173218280603, | |
| "eval_runtime": 216.4936, | |
| "eval_samples_per_second": 5.206, | |
| "eval_steps_per_second": 0.651, | |
| "step": 4928 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "learning_rate": 4.3303571428571435e-05, | |
| "loss": 0.6378, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 4.06, | |
| "learning_rate": 4.323593073593074e-05, | |
| "loss": 0.5454, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 4.1, | |
| "learning_rate": 4.316829004329004e-05, | |
| "loss": 0.4825, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 4.14, | |
| "learning_rate": 4.310064935064935e-05, | |
| "loss": 0.5549, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 4.18, | |
| "learning_rate": 4.303300865800866e-05, | |
| "loss": 0.515, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 4.22, | |
| "learning_rate": 4.2965367965367966e-05, | |
| "loss": 0.5566, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 4.26, | |
| "learning_rate": 4.289772727272727e-05, | |
| "loss": 0.5484, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 4.3, | |
| "learning_rate": 4.283008658008658e-05, | |
| "loss": 0.5033, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 4.34, | |
| "learning_rate": 4.276244588744589e-05, | |
| "loss": 0.4821, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 4.38, | |
| "learning_rate": 4.26948051948052e-05, | |
| "loss": 0.4676, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 4.42, | |
| "learning_rate": 4.2627164502164505e-05, | |
| "loss": 0.5176, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 4.46, | |
| "learning_rate": 4.255952380952381e-05, | |
| "loss": 0.5493, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 4.5, | |
| "learning_rate": 4.249188311688312e-05, | |
| "loss": 0.5177, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 4.55, | |
| "learning_rate": 4.242424242424243e-05, | |
| "loss": 0.4891, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 4.59, | |
| "learning_rate": 4.235660173160173e-05, | |
| "loss": 0.5674, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 4.63, | |
| "learning_rate": 4.228896103896104e-05, | |
| "loss": 0.4763, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 4.67, | |
| "learning_rate": 4.222132034632035e-05, | |
| "loss": 0.5431, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 4.71, | |
| "learning_rate": 4.215367965367966e-05, | |
| "loss": 0.5039, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 4.75, | |
| "learning_rate": 4.208603896103897e-05, | |
| "loss": 0.5064, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 4.79, | |
| "learning_rate": 4.2018398268398265e-05, | |
| "loss": 0.5018, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 4.83, | |
| "learning_rate": 4.1950757575757575e-05, | |
| "loss": 0.5066, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 4.87, | |
| "learning_rate": 4.1883116883116886e-05, | |
| "loss": 0.5357, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 4.91, | |
| "learning_rate": 4.1815476190476196e-05, | |
| "loss": 0.5285, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 4.95, | |
| "learning_rate": 4.17478354978355e-05, | |
| "loss": 0.5814, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 4.99, | |
| "learning_rate": 4.16801948051948e-05, | |
| "loss": 0.5216, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "eval_cer": 0.30030275361622377, | |
| "eval_loss": 1.3027092218399048, | |
| "eval_mer": 0.34603707254713967, | |
| "eval_runtime": 217.8086, | |
| "eval_samples_per_second": 5.174, | |
| "eval_steps_per_second": 0.647, | |
| "step": 6160 | |
| }, | |
| { | |
| "epoch": 5.03, | |
| "learning_rate": 4.1612554112554114e-05, | |
| "loss": 0.4249, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 5.07, | |
| "learning_rate": 4.1544913419913424e-05, | |
| "loss": 0.3902, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 5.11, | |
| "learning_rate": 4.1477272727272734e-05, | |
| "loss": 0.399, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 5.15, | |
| "learning_rate": 4.140963203463203e-05, | |
| "loss": 0.4107, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 5.19, | |
| "learning_rate": 4.134199134199134e-05, | |
| "loss": 0.4767, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 5.24, | |
| "learning_rate": 4.127435064935065e-05, | |
| "loss": 0.4182, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 5.28, | |
| "learning_rate": 4.120670995670996e-05, | |
| "loss": 0.4235, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 5.32, | |
| "learning_rate": 4.1139069264069266e-05, | |
| "loss": 0.4028, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 5.36, | |
| "learning_rate": 4.107142857142857e-05, | |
| "loss": 0.4497, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 5.4, | |
| "learning_rate": 4.100378787878788e-05, | |
| "loss": 0.4645, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 5.44, | |
| "learning_rate": 4.093614718614719e-05, | |
| "loss": 0.3768, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 5.48, | |
| "learning_rate": 4.08685064935065e-05, | |
| "loss": 0.5213, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 5.52, | |
| "learning_rate": 4.08008658008658e-05, | |
| "loss": 0.4512, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 5.56, | |
| "learning_rate": 4.073322510822511e-05, | |
| "loss": 0.4031, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 5.6, | |
| "learning_rate": 4.066558441558442e-05, | |
| "loss": 0.4361, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 5.64, | |
| "learning_rate": 4.059794372294373e-05, | |
| "loss": 0.4232, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 5.68, | |
| "learning_rate": 4.053030303030303e-05, | |
| "loss": 0.4205, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 5.72, | |
| "learning_rate": 4.0462662337662336e-05, | |
| "loss": 0.4145, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 5.76, | |
| "learning_rate": 4.039502164502165e-05, | |
| "loss": 0.5018, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 5.8, | |
| "learning_rate": 4.032738095238096e-05, | |
| "loss": 0.4073, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 5.84, | |
| "learning_rate": 4.025974025974026e-05, | |
| "loss": 0.4543, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 5.88, | |
| "learning_rate": 4.0192099567099564e-05, | |
| "loss": 0.3707, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 5.93, | |
| "learning_rate": 4.0124458874458875e-05, | |
| "loss": 0.4212, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 5.97, | |
| "learning_rate": 4.0056818181818185e-05, | |
| "loss": 0.4556, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "eval_cer": 0.2909318083521553, | |
| "eval_loss": 1.3074078559875488, | |
| "eval_mer": 0.3376478108021732, | |
| "eval_runtime": 217.3379, | |
| "eval_samples_per_second": 5.185, | |
| "eval_steps_per_second": 0.649, | |
| "step": 7392 | |
| }, | |
| { | |
| "epoch": 6.01, | |
| "learning_rate": 3.9989177489177496e-05, | |
| "loss": 0.4065, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 6.05, | |
| "learning_rate": 3.99215367965368e-05, | |
| "loss": 0.3537, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 6.09, | |
| "learning_rate": 3.98538961038961e-05, | |
| "loss": 0.3525, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 6.13, | |
| "learning_rate": 3.978625541125541e-05, | |
| "loss": 0.4335, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 6.17, | |
| "learning_rate": 3.9718614718614724e-05, | |
| "loss": 0.3785, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 6.21, | |
| "learning_rate": 3.965097402597403e-05, | |
| "loss": 0.3757, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 6.25, | |
| "learning_rate": 3.958333333333333e-05, | |
| "loss": 0.3563, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 6.29, | |
| "learning_rate": 3.951569264069264e-05, | |
| "loss": 0.3371, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 6.33, | |
| "learning_rate": 3.944805194805195e-05, | |
| "loss": 0.3632, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 6.37, | |
| "learning_rate": 3.938041125541126e-05, | |
| "loss": 0.365, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 6.41, | |
| "learning_rate": 3.931277056277056e-05, | |
| "loss": 0.355, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 6.45, | |
| "learning_rate": 3.924512987012987e-05, | |
| "loss": 0.3097, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 6.49, | |
| "learning_rate": 3.917748917748918e-05, | |
| "loss": 0.3188, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 6.53, | |
| "learning_rate": 3.910984848484849e-05, | |
| "loss": 0.3512, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 6.57, | |
| "learning_rate": 3.9042207792207794e-05, | |
| "loss": 0.3474, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 6.62, | |
| "learning_rate": 3.89745670995671e-05, | |
| "loss": 0.3851, | |
| "step": 8150 | |
| }, | |
| { | |
| "epoch": 6.66, | |
| "learning_rate": 3.890692640692641e-05, | |
| "loss": 0.3295, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 6.7, | |
| "learning_rate": 3.883928571428572e-05, | |
| "loss": 0.3445, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 6.74, | |
| "learning_rate": 3.877164502164503e-05, | |
| "loss": 0.3924, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 6.78, | |
| "learning_rate": 3.8704004329004326e-05, | |
| "loss": 0.3905, | |
| "step": 8350 | |
| }, | |
| { | |
| "epoch": 6.82, | |
| "learning_rate": 3.8636363636363636e-05, | |
| "loss": 0.3809, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 6.86, | |
| "learning_rate": 3.8568722943722946e-05, | |
| "loss": 0.3692, | |
| "step": 8450 | |
| }, | |
| { | |
| "epoch": 6.9, | |
| "learning_rate": 3.850108225108226e-05, | |
| "loss": 0.3243, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 6.94, | |
| "learning_rate": 3.843344155844156e-05, | |
| "loss": 0.3845, | |
| "step": 8550 | |
| }, | |
| { | |
| "epoch": 6.98, | |
| "learning_rate": 3.8365800865800864e-05, | |
| "loss": 0.3212, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "eval_cer": 0.284684511509443, | |
| "eval_loss": 1.3443255424499512, | |
| "eval_mer": 0.33237456056248005, | |
| "eval_runtime": 216.9522, | |
| "eval_samples_per_second": 5.195, | |
| "eval_steps_per_second": 0.65, | |
| "step": 8624 | |
| }, | |
| { | |
| "epoch": 7.02, | |
| "learning_rate": 3.8298160173160175e-05, | |
| "loss": 0.3799, | |
| "step": 8650 | |
| }, | |
| { | |
| "epoch": 7.06, | |
| "learning_rate": 3.8230519480519485e-05, | |
| "loss": 0.31, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 7.1, | |
| "learning_rate": 3.816287878787879e-05, | |
| "loss": 0.2646, | |
| "step": 8750 | |
| }, | |
| { | |
| "epoch": 7.14, | |
| "learning_rate": 3.809523809523809e-05, | |
| "loss": 0.2783, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 7.18, | |
| "learning_rate": 3.80275974025974e-05, | |
| "loss": 0.364, | |
| "step": 8850 | |
| }, | |
| { | |
| "epoch": 7.22, | |
| "learning_rate": 3.795995670995671e-05, | |
| "loss": 0.2917, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 7.26, | |
| "learning_rate": 3.7892316017316023e-05, | |
| "loss": 0.3235, | |
| "step": 8950 | |
| }, | |
| { | |
| "epoch": 7.31, | |
| "learning_rate": 3.782467532467533e-05, | |
| "loss": 0.2982, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 7.35, | |
| "learning_rate": 3.775703463203463e-05, | |
| "loss": 0.323, | |
| "step": 9050 | |
| }, | |
| { | |
| "epoch": 7.39, | |
| "learning_rate": 3.768939393939394e-05, | |
| "loss": 0.2758, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 7.43, | |
| "learning_rate": 3.762175324675325e-05, | |
| "loss": 0.3276, | |
| "step": 9150 | |
| }, | |
| { | |
| "epoch": 7.47, | |
| "learning_rate": 3.7554112554112555e-05, | |
| "loss": 0.3555, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 7.51, | |
| "learning_rate": 3.748647186147186e-05, | |
| "loss": 0.2968, | |
| "step": 9250 | |
| }, | |
| { | |
| "epoch": 7.55, | |
| "learning_rate": 3.741883116883117e-05, | |
| "loss": 0.3656, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 7.59, | |
| "learning_rate": 3.735119047619048e-05, | |
| "loss": 0.3429, | |
| "step": 9350 | |
| }, | |
| { | |
| "epoch": 7.63, | |
| "learning_rate": 3.728354978354979e-05, | |
| "loss": 0.3144, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 7.67, | |
| "learning_rate": 3.721590909090909e-05, | |
| "loss": 0.3158, | |
| "step": 9450 | |
| }, | |
| { | |
| "epoch": 7.71, | |
| "learning_rate": 3.71482683982684e-05, | |
| "loss": 0.3139, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 7.75, | |
| "learning_rate": 3.708062770562771e-05, | |
| "loss": 0.2774, | |
| "step": 9550 | |
| }, | |
| { | |
| "epoch": 7.79, | |
| "learning_rate": 3.701298701298702e-05, | |
| "loss": 0.2942, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 7.83, | |
| "learning_rate": 3.694534632034632e-05, | |
| "loss": 0.3149, | |
| "step": 9650 | |
| }, | |
| { | |
| "epoch": 7.87, | |
| "learning_rate": 3.6877705627705625e-05, | |
| "loss": 0.2894, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 7.91, | |
| "learning_rate": 3.6810064935064936e-05, | |
| "loss": 0.286, | |
| "step": 9750 | |
| }, | |
| { | |
| "epoch": 7.95, | |
| "learning_rate": 3.6742424242424246e-05, | |
| "loss": 0.278, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "learning_rate": 3.6674783549783557e-05, | |
| "loss": 0.2932, | |
| "step": 9850 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "eval_cer": 0.28132058244029023, | |
| "eval_loss": 1.3463859558105469, | |
| "eval_mer": 0.33013742409715563, | |
| "eval_runtime": 217.5907, | |
| "eval_samples_per_second": 5.179, | |
| "eval_steps_per_second": 0.648, | |
| "step": 9856 | |
| }, | |
| { | |
| "epoch": 8.04, | |
| "learning_rate": 3.6607142857142853e-05, | |
| "loss": 0.2679, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 8.08, | |
| "learning_rate": 3.6539502164502164e-05, | |
| "loss": 0.2465, | |
| "step": 9950 | |
| }, | |
| { | |
| "epoch": 8.12, | |
| "learning_rate": 3.6471861471861474e-05, | |
| "loss": 0.2562, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 8.16, | |
| "learning_rate": 3.6404220779220785e-05, | |
| "loss": 0.2457, | |
| "step": 10050 | |
| }, | |
| { | |
| "epoch": 8.2, | |
| "learning_rate": 3.633658008658009e-05, | |
| "loss": 0.2664, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 8.24, | |
| "learning_rate": 3.626893939393939e-05, | |
| "loss": 0.2926, | |
| "step": 10150 | |
| }, | |
| { | |
| "epoch": 8.28, | |
| "learning_rate": 3.62012987012987e-05, | |
| "loss": 0.2467, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 8.32, | |
| "learning_rate": 3.613365800865801e-05, | |
| "loss": 0.2339, | |
| "step": 10250 | |
| }, | |
| { | |
| "epoch": 8.36, | |
| "learning_rate": 3.606601731601732e-05, | |
| "loss": 0.2644, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 8.4, | |
| "learning_rate": 3.599837662337662e-05, | |
| "loss": 0.2132, | |
| "step": 10350 | |
| }, | |
| { | |
| "epoch": 8.44, | |
| "learning_rate": 3.593073593073593e-05, | |
| "loss": 0.2529, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 8.48, | |
| "learning_rate": 3.586309523809524e-05, | |
| "loss": 0.228, | |
| "step": 10450 | |
| }, | |
| { | |
| "epoch": 8.52, | |
| "learning_rate": 3.579545454545455e-05, | |
| "loss": 0.2362, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 8.56, | |
| "learning_rate": 3.5727813852813855e-05, | |
| "loss": 0.2834, | |
| "step": 10550 | |
| }, | |
| { | |
| "epoch": 8.6, | |
| "learning_rate": 3.566017316017316e-05, | |
| "loss": 0.2586, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 8.64, | |
| "learning_rate": 3.559253246753247e-05, | |
| "loss": 0.2667, | |
| "step": 10650 | |
| }, | |
| { | |
| "epoch": 8.69, | |
| "learning_rate": 3.552489177489178e-05, | |
| "loss": 0.264, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 8.73, | |
| "learning_rate": 3.545725108225108e-05, | |
| "loss": 0.2493, | |
| "step": 10750 | |
| }, | |
| { | |
| "epoch": 8.77, | |
| "learning_rate": 3.5389610389610387e-05, | |
| "loss": 0.2459, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 8.81, | |
| "learning_rate": 3.53219696969697e-05, | |
| "loss": 0.2584, | |
| "step": 10850 | |
| }, | |
| { | |
| "epoch": 8.85, | |
| "learning_rate": 3.525432900432901e-05, | |
| "loss": 0.2871, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 8.89, | |
| "learning_rate": 3.518668831168832e-05, | |
| "loss": 0.2366, | |
| "step": 10950 | |
| }, | |
| { | |
| "epoch": 8.93, | |
| "learning_rate": 3.511904761904762e-05, | |
| "loss": 0.283, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 8.97, | |
| "learning_rate": 3.5051406926406925e-05, | |
| "loss": 0.2787, | |
| "step": 11050 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "eval_cer": 0.2752655101158153, | |
| "eval_loss": 1.368951678276062, | |
| "eval_mer": 0.3227069351230425, | |
| "eval_runtime": 216.8467, | |
| "eval_samples_per_second": 5.197, | |
| "eval_steps_per_second": 0.65, | |
| "step": 11088 | |
| }, | |
| { | |
| "epoch": 9.01, | |
| "learning_rate": 3.4983766233766235e-05, | |
| "loss": 0.2791, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 9.05, | |
| "learning_rate": 3.4916125541125546e-05, | |
| "loss": 0.2467, | |
| "step": 11150 | |
| }, | |
| { | |
| "epoch": 9.09, | |
| "learning_rate": 3.484848484848485e-05, | |
| "loss": 0.2426, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 9.13, | |
| "learning_rate": 3.478084415584415e-05, | |
| "loss": 0.2166, | |
| "step": 11250 | |
| }, | |
| { | |
| "epoch": 9.17, | |
| "learning_rate": 3.4713203463203464e-05, | |
| "loss": 0.2115, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 9.21, | |
| "learning_rate": 3.4645562770562774e-05, | |
| "loss": 0.2068, | |
| "step": 11350 | |
| }, | |
| { | |
| "epoch": 9.25, | |
| "learning_rate": 3.4577922077922084e-05, | |
| "loss": 0.2428, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 9.29, | |
| "learning_rate": 3.451028138528138e-05, | |
| "loss": 0.2099, | |
| "step": 11450 | |
| }, | |
| { | |
| "epoch": 9.33, | |
| "learning_rate": 3.444264069264069e-05, | |
| "loss": 0.1878, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 9.38, | |
| "learning_rate": 3.4375e-05, | |
| "loss": 0.2063, | |
| "step": 11550 | |
| }, | |
| { | |
| "epoch": 9.42, | |
| "learning_rate": 3.430735930735931e-05, | |
| "loss": 0.2298, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 9.46, | |
| "learning_rate": 3.4239718614718616e-05, | |
| "loss": 0.2199, | |
| "step": 11650 | |
| }, | |
| { | |
| "epoch": 9.5, | |
| "learning_rate": 3.417207792207792e-05, | |
| "loss": 0.2313, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 9.54, | |
| "learning_rate": 3.410443722943723e-05, | |
| "loss": 0.1911, | |
| "step": 11750 | |
| }, | |
| { | |
| "epoch": 9.58, | |
| "learning_rate": 3.403679653679654e-05, | |
| "loss": 0.2218, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 9.62, | |
| "learning_rate": 3.396915584415585e-05, | |
| "loss": 0.2401, | |
| "step": 11850 | |
| }, | |
| { | |
| "epoch": 9.66, | |
| "learning_rate": 3.390151515151515e-05, | |
| "loss": 0.1973, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 9.7, | |
| "learning_rate": 3.383387445887446e-05, | |
| "loss": 0.2304, | |
| "step": 11950 | |
| }, | |
| { | |
| "epoch": 9.74, | |
| "learning_rate": 3.376623376623377e-05, | |
| "loss": 0.2131, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 9.78, | |
| "learning_rate": 3.369859307359308e-05, | |
| "loss": 0.2249, | |
| "step": 12050 | |
| }, | |
| { | |
| "epoch": 9.82, | |
| "learning_rate": 3.363095238095238e-05, | |
| "loss": 0.2094, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 9.86, | |
| "learning_rate": 3.3563311688311686e-05, | |
| "loss": 0.2466, | |
| "step": 12150 | |
| }, | |
| { | |
| "epoch": 9.9, | |
| "learning_rate": 3.3495670995671e-05, | |
| "loss": 0.2148, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 9.94, | |
| "learning_rate": 3.342803030303031e-05, | |
| "loss": 0.1822, | |
| "step": 12250 | |
| }, | |
| { | |
| "epoch": 9.98, | |
| "learning_rate": 3.336038961038961e-05, | |
| "loss": 0.2482, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "eval_cer": 0.27065212167812003, | |
| "eval_loss": 1.430137276649475, | |
| "eval_mer": 0.3196708213486737, | |
| "eval_runtime": 217.321, | |
| "eval_samples_per_second": 5.186, | |
| "eval_steps_per_second": 0.649, | |
| "step": 12320 | |
| }, | |
| { | |
| "epoch": 10.02, | |
| "learning_rate": 3.3292748917748914e-05, | |
| "loss": 0.1891, | |
| "step": 12350 | |
| }, | |
| { | |
| "epoch": 10.06, | |
| "learning_rate": 3.3225108225108225e-05, | |
| "loss": 0.1518, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 10.11, | |
| "learning_rate": 3.3157467532467535e-05, | |
| "loss": 0.2368, | |
| "step": 12450 | |
| }, | |
| { | |
| "epoch": 10.15, | |
| "learning_rate": 3.3089826839826846e-05, | |
| "loss": 0.226, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 10.19, | |
| "learning_rate": 3.302218614718615e-05, | |
| "loss": 0.2228, | |
| "step": 12550 | |
| }, | |
| { | |
| "epoch": 10.23, | |
| "learning_rate": 3.295454545454545e-05, | |
| "loss": 0.1958, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 10.27, | |
| "learning_rate": 3.288690476190476e-05, | |
| "loss": 0.1967, | |
| "step": 12650 | |
| }, | |
| { | |
| "epoch": 10.31, | |
| "learning_rate": 3.2819264069264074e-05, | |
| "loss": 0.1856, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 10.35, | |
| "learning_rate": 3.275162337662338e-05, | |
| "loss": 0.1832, | |
| "step": 12750 | |
| }, | |
| { | |
| "epoch": 10.39, | |
| "learning_rate": 3.268398268398268e-05, | |
| "loss": 0.1739, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 10.43, | |
| "learning_rate": 3.261634199134199e-05, | |
| "loss": 0.1737, | |
| "step": 12850 | |
| }, | |
| { | |
| "epoch": 10.47, | |
| "learning_rate": 3.25487012987013e-05, | |
| "loss": 0.1694, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 10.51, | |
| "learning_rate": 3.248106060606061e-05, | |
| "loss": 0.1867, | |
| "step": 12950 | |
| }, | |
| { | |
| "epoch": 10.55, | |
| "learning_rate": 3.241341991341991e-05, | |
| "loss": 0.1907, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 10.59, | |
| "learning_rate": 3.234577922077922e-05, | |
| "loss": 0.1684, | |
| "step": 13050 | |
| }, | |
| { | |
| "epoch": 10.63, | |
| "learning_rate": 3.227813852813853e-05, | |
| "loss": 0.1795, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 10.67, | |
| "learning_rate": 3.221049783549784e-05, | |
| "loss": 0.1838, | |
| "step": 13150 | |
| }, | |
| { | |
| "epoch": 10.71, | |
| "learning_rate": 3.2142857142857144e-05, | |
| "loss": 0.198, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 10.75, | |
| "learning_rate": 3.207521645021645e-05, | |
| "loss": 0.19, | |
| "step": 13250 | |
| }, | |
| { | |
| "epoch": 10.8, | |
| "learning_rate": 3.200757575757576e-05, | |
| "loss": 0.1725, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 10.84, | |
| "learning_rate": 3.193993506493507e-05, | |
| "loss": 0.2693, | |
| "step": 13350 | |
| }, | |
| { | |
| "epoch": 10.88, | |
| "learning_rate": 3.187229437229438e-05, | |
| "loss": 0.2033, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 10.92, | |
| "learning_rate": 3.1804653679653676e-05, | |
| "loss": 0.2122, | |
| "step": 13450 | |
| }, | |
| { | |
| "epoch": 10.96, | |
| "learning_rate": 3.1737012987012986e-05, | |
| "loss": 0.1898, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "learning_rate": 3.1669372294372296e-05, | |
| "loss": 0.2176, | |
| "step": 13550 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "eval_cer": 0.26897015714354366, | |
| "eval_loss": 1.4202213287353516, | |
| "eval_mer": 0.3180728667305848, | |
| "eval_runtime": 215.6805, | |
| "eval_samples_per_second": 5.225, | |
| "eval_steps_per_second": 0.654, | |
| "step": 13552 | |
| }, | |
| { | |
| "epoch": 11.04, | |
| "learning_rate": 3.160173160173161e-05, | |
| "loss": 0.1901, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 11.08, | |
| "learning_rate": 3.153409090909091e-05, | |
| "loss": 0.1398, | |
| "step": 13650 | |
| }, | |
| { | |
| "epoch": 11.12, | |
| "learning_rate": 3.1466450216450214e-05, | |
| "loss": 0.1427, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 11.16, | |
| "learning_rate": 3.1398809523809525e-05, | |
| "loss": 0.1461, | |
| "step": 13750 | |
| }, | |
| { | |
| "epoch": 11.2, | |
| "learning_rate": 3.1331168831168835e-05, | |
| "loss": 0.1296, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 11.24, | |
| "learning_rate": 3.1263528138528145e-05, | |
| "loss": 0.1695, | |
| "step": 13850 | |
| }, | |
| { | |
| "epoch": 11.28, | |
| "learning_rate": 3.119588744588744e-05, | |
| "loss": 0.1431, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 11.32, | |
| "learning_rate": 3.112824675324675e-05, | |
| "loss": 0.1482, | |
| "step": 13950 | |
| }, | |
| { | |
| "epoch": 11.36, | |
| "learning_rate": 3.106060606060606e-05, | |
| "loss": 0.155, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 11.4, | |
| "learning_rate": 3.0992965367965373e-05, | |
| "loss": 0.1611, | |
| "step": 14050 | |
| }, | |
| { | |
| "epoch": 11.44, | |
| "learning_rate": 3.092532467532468e-05, | |
| "loss": 0.1815, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 11.49, | |
| "learning_rate": 3.085768398268398e-05, | |
| "loss": 0.173, | |
| "step": 14150 | |
| }, | |
| { | |
| "epoch": 11.53, | |
| "learning_rate": 3.079004329004329e-05, | |
| "loss": 0.1693, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 11.57, | |
| "learning_rate": 3.07224025974026e-05, | |
| "loss": 0.1859, | |
| "step": 14250 | |
| }, | |
| { | |
| "epoch": 11.61, | |
| "learning_rate": 3.0654761904761905e-05, | |
| "loss": 0.1763, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 11.65, | |
| "learning_rate": 3.058712121212121e-05, | |
| "loss": 0.1747, | |
| "step": 14350 | |
| }, | |
| { | |
| "epoch": 11.69, | |
| "learning_rate": 3.051948051948052e-05, | |
| "loss": 0.176, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 11.73, | |
| "learning_rate": 3.045183982683983e-05, | |
| "loss": 0.1514, | |
| "step": 14450 | |
| }, | |
| { | |
| "epoch": 11.77, | |
| "learning_rate": 3.0384199134199137e-05, | |
| "loss": 0.2071, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 11.81, | |
| "learning_rate": 3.031655844155844e-05, | |
| "loss": 0.1684, | |
| "step": 14550 | |
| }, | |
| { | |
| "epoch": 11.85, | |
| "learning_rate": 3.0248917748917747e-05, | |
| "loss": 0.1552, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 11.89, | |
| "learning_rate": 3.0181277056277058e-05, | |
| "loss": 0.2061, | |
| "step": 14650 | |
| }, | |
| { | |
| "epoch": 11.93, | |
| "learning_rate": 3.0113636363636365e-05, | |
| "loss": 0.1661, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 11.97, | |
| "learning_rate": 3.0045995670995675e-05, | |
| "loss": 0.1789, | |
| "step": 14750 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "eval_cer": 0.266951799702052, | |
| "eval_loss": 1.5919291973114014, | |
| "eval_mer": 0.3156759348034516, | |
| "eval_runtime": 215.4469, | |
| "eval_samples_per_second": 5.231, | |
| "eval_steps_per_second": 0.654, | |
| "step": 14784 | |
| }, | |
| { | |
| "epoch": 12.01, | |
| "learning_rate": 2.997835497835498e-05, | |
| "loss": 0.175, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 12.05, | |
| "learning_rate": 2.9910714285714286e-05, | |
| "loss": 0.1368, | |
| "step": 14850 | |
| }, | |
| { | |
| "epoch": 12.09, | |
| "learning_rate": 2.9843073593073596e-05, | |
| "loss": 0.121, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 12.13, | |
| "learning_rate": 2.9775432900432903e-05, | |
| "loss": 0.1425, | |
| "step": 14950 | |
| }, | |
| { | |
| "epoch": 12.18, | |
| "learning_rate": 2.9707792207792207e-05, | |
| "loss": 0.1534, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 12.22, | |
| "learning_rate": 2.9640151515151514e-05, | |
| "loss": 0.1691, | |
| "step": 15050 | |
| }, | |
| { | |
| "epoch": 12.26, | |
| "learning_rate": 2.9572510822510824e-05, | |
| "loss": 0.1276, | |
| "step": 15100 | |
| }, | |
| { | |
| "epoch": 12.3, | |
| "learning_rate": 2.950487012987013e-05, | |
| "loss": 0.1304, | |
| "step": 15150 | |
| }, | |
| { | |
| "epoch": 12.34, | |
| "learning_rate": 2.943722943722944e-05, | |
| "loss": 0.1564, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 12.38, | |
| "learning_rate": 2.9369588744588745e-05, | |
| "loss": 0.1219, | |
| "step": 15250 | |
| }, | |
| { | |
| "epoch": 12.42, | |
| "learning_rate": 2.9301948051948052e-05, | |
| "loss": 0.1604, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 12.46, | |
| "learning_rate": 2.9234307359307363e-05, | |
| "loss": 0.145, | |
| "step": 15350 | |
| }, | |
| { | |
| "epoch": 12.5, | |
| "learning_rate": 2.916666666666667e-05, | |
| "loss": 0.136, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 12.54, | |
| "learning_rate": 2.9099025974025973e-05, | |
| "loss": 0.1394, | |
| "step": 15450 | |
| }, | |
| { | |
| "epoch": 12.58, | |
| "learning_rate": 2.903138528138528e-05, | |
| "loss": 0.1355, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 12.62, | |
| "learning_rate": 2.896374458874459e-05, | |
| "loss": 0.1293, | |
| "step": 15550 | |
| }, | |
| { | |
| "epoch": 12.66, | |
| "learning_rate": 2.8896103896103898e-05, | |
| "loss": 0.1454, | |
| "step": 15600 | |
| }, | |
| { | |
| "epoch": 12.7, | |
| "learning_rate": 2.8828463203463208e-05, | |
| "loss": 0.1378, | |
| "step": 15650 | |
| }, | |
| { | |
| "epoch": 12.74, | |
| "learning_rate": 2.8760822510822512e-05, | |
| "loss": 0.1405, | |
| "step": 15700 | |
| }, | |
| { | |
| "epoch": 12.78, | |
| "learning_rate": 2.869318181818182e-05, | |
| "loss": 0.1397, | |
| "step": 15750 | |
| }, | |
| { | |
| "epoch": 12.82, | |
| "learning_rate": 2.862554112554113e-05, | |
| "loss": 0.1307, | |
| "step": 15800 | |
| }, | |
| { | |
| "epoch": 12.87, | |
| "learning_rate": 2.8557900432900436e-05, | |
| "loss": 0.1331, | |
| "step": 15850 | |
| }, | |
| { | |
| "epoch": 12.91, | |
| "learning_rate": 2.849025974025974e-05, | |
| "loss": 0.1569, | |
| "step": 15900 | |
| }, | |
| { | |
| "epoch": 12.95, | |
| "learning_rate": 2.8422619047619047e-05, | |
| "loss": 0.1444, | |
| "step": 15950 | |
| }, | |
| { | |
| "epoch": 12.99, | |
| "learning_rate": 2.8354978354978357e-05, | |
| "loss": 0.1557, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "eval_cer": 0.26709596809073, | |
| "eval_loss": 1.5367108583450317, | |
| "eval_mer": 0.317913071268776, | |
| "eval_runtime": 216.3585, | |
| "eval_samples_per_second": 5.209, | |
| "eval_steps_per_second": 0.652, | |
| "step": 16016 | |
| }, | |
| { | |
| "epoch": 13.03, | |
| "learning_rate": 2.8287337662337664e-05, | |
| "loss": 0.1467, | |
| "step": 16050 | |
| }, | |
| { | |
| "epoch": 13.07, | |
| "learning_rate": 2.8219696969696975e-05, | |
| "loss": 0.1197, | |
| "step": 16100 | |
| }, | |
| { | |
| "epoch": 13.11, | |
| "learning_rate": 2.815205627705628e-05, | |
| "loss": 0.1162, | |
| "step": 16150 | |
| }, | |
| { | |
| "epoch": 13.15, | |
| "learning_rate": 2.8084415584415585e-05, | |
| "loss": 0.093, | |
| "step": 16200 | |
| }, | |
| { | |
| "epoch": 13.19, | |
| "learning_rate": 2.8016774891774892e-05, | |
| "loss": 0.1373, | |
| "step": 16250 | |
| }, | |
| { | |
| "epoch": 13.23, | |
| "learning_rate": 2.7949134199134203e-05, | |
| "loss": 0.121, | |
| "step": 16300 | |
| }, | |
| { | |
| "epoch": 13.27, | |
| "learning_rate": 2.7881493506493507e-05, | |
| "loss": 0.1317, | |
| "step": 16350 | |
| }, | |
| { | |
| "epoch": 13.31, | |
| "learning_rate": 2.7813852813852814e-05, | |
| "loss": 0.1352, | |
| "step": 16400 | |
| }, | |
| { | |
| "epoch": 13.35, | |
| "learning_rate": 2.7746212121212124e-05, | |
| "loss": 0.1073, | |
| "step": 16450 | |
| }, | |
| { | |
| "epoch": 13.39, | |
| "learning_rate": 2.767857142857143e-05, | |
| "loss": 0.1212, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 13.43, | |
| "learning_rate": 2.7610930735930735e-05, | |
| "loss": 0.1542, | |
| "step": 16550 | |
| }, | |
| { | |
| "epoch": 13.47, | |
| "learning_rate": 2.754329004329004e-05, | |
| "loss": 0.1445, | |
| "step": 16600 | |
| }, | |
| { | |
| "epoch": 13.51, | |
| "learning_rate": 2.7475649350649352e-05, | |
| "loss": 0.106, | |
| "step": 16650 | |
| }, | |
| { | |
| "epoch": 13.56, | |
| "learning_rate": 2.740800865800866e-05, | |
| "loss": 0.1023, | |
| "step": 16700 | |
| }, | |
| { | |
| "epoch": 13.6, | |
| "learning_rate": 2.734036796536797e-05, | |
| "loss": 0.131, | |
| "step": 16750 | |
| }, | |
| { | |
| "epoch": 13.64, | |
| "learning_rate": 2.7272727272727273e-05, | |
| "loss": 0.1182, | |
| "step": 16800 | |
| }, | |
| { | |
| "epoch": 13.68, | |
| "learning_rate": 2.720508658008658e-05, | |
| "loss": 0.1086, | |
| "step": 16850 | |
| }, | |
| { | |
| "epoch": 13.72, | |
| "learning_rate": 2.713744588744589e-05, | |
| "loss": 0.1214, | |
| "step": 16900 | |
| }, | |
| { | |
| "epoch": 13.76, | |
| "learning_rate": 2.7069805194805198e-05, | |
| "loss": 0.1353, | |
| "step": 16950 | |
| }, | |
| { | |
| "epoch": 13.8, | |
| "learning_rate": 2.70021645021645e-05, | |
| "loss": 0.1273, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 13.84, | |
| "learning_rate": 2.6934523809523808e-05, | |
| "loss": 0.1184, | |
| "step": 17050 | |
| }, | |
| { | |
| "epoch": 13.88, | |
| "learning_rate": 2.686688311688312e-05, | |
| "loss": 0.1343, | |
| "step": 17100 | |
| }, | |
| { | |
| "epoch": 13.92, | |
| "learning_rate": 2.6799242424242426e-05, | |
| "loss": 0.1314, | |
| "step": 17150 | |
| }, | |
| { | |
| "epoch": 13.96, | |
| "learning_rate": 2.6731601731601736e-05, | |
| "loss": 0.1328, | |
| "step": 17200 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "eval_cer": 0.2673843048680859, | |
| "eval_loss": 1.5152249336242676, | |
| "eval_mer": 0.3162352189197827, | |
| "eval_runtime": 217.4486, | |
| "eval_samples_per_second": 5.183, | |
| "eval_steps_per_second": 0.648, | |
| "step": 17248 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "learning_rate": 2.666396103896104e-05, | |
| "loss": 0.1186, | |
| "step": 17250 | |
| }, | |
| { | |
| "epoch": 14.04, | |
| "learning_rate": 2.6596320346320347e-05, | |
| "loss": 0.0985, | |
| "step": 17300 | |
| }, | |
| { | |
| "epoch": 14.08, | |
| "learning_rate": 2.6528679653679657e-05, | |
| "loss": 0.11, | |
| "step": 17350 | |
| }, | |
| { | |
| "epoch": 14.12, | |
| "learning_rate": 2.6461038961038964e-05, | |
| "loss": 0.1185, | |
| "step": 17400 | |
| }, | |
| { | |
| "epoch": 14.16, | |
| "learning_rate": 2.6393398268398268e-05, | |
| "loss": 0.1099, | |
| "step": 17450 | |
| }, | |
| { | |
| "epoch": 14.2, | |
| "learning_rate": 2.6325757575757575e-05, | |
| "loss": 0.106, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 14.25, | |
| "learning_rate": 2.6258116883116885e-05, | |
| "loss": 0.0785, | |
| "step": 17550 | |
| }, | |
| { | |
| "epoch": 14.29, | |
| "learning_rate": 2.6190476190476192e-05, | |
| "loss": 0.1029, | |
| "step": 17600 | |
| }, | |
| { | |
| "epoch": 14.33, | |
| "learning_rate": 2.6122835497835503e-05, | |
| "loss": 0.1203, | |
| "step": 17650 | |
| }, | |
| { | |
| "epoch": 14.37, | |
| "learning_rate": 2.6055194805194806e-05, | |
| "loss": 0.1184, | |
| "step": 17700 | |
| }, | |
| { | |
| "epoch": 14.41, | |
| "learning_rate": 2.5987554112554113e-05, | |
| "loss": 0.1351, | |
| "step": 17750 | |
| }, | |
| { | |
| "epoch": 14.45, | |
| "learning_rate": 2.591991341991342e-05, | |
| "loss": 0.1235, | |
| "step": 17800 | |
| }, | |
| { | |
| "epoch": 14.49, | |
| "learning_rate": 2.585227272727273e-05, | |
| "loss": 0.1144, | |
| "step": 17850 | |
| }, | |
| { | |
| "epoch": 14.53, | |
| "learning_rate": 2.5784632034632034e-05, | |
| "loss": 0.147, | |
| "step": 17900 | |
| }, | |
| { | |
| "epoch": 14.57, | |
| "learning_rate": 2.571699134199134e-05, | |
| "loss": 0.1162, | |
| "step": 17950 | |
| }, | |
| { | |
| "epoch": 14.61, | |
| "learning_rate": 2.5649350649350652e-05, | |
| "loss": 0.1331, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 14.65, | |
| "learning_rate": 2.558170995670996e-05, | |
| "loss": 0.0867, | |
| "step": 18050 | |
| }, | |
| { | |
| "epoch": 14.69, | |
| "learning_rate": 2.551406926406927e-05, | |
| "loss": 0.1207, | |
| "step": 18100 | |
| }, | |
| { | |
| "epoch": 14.73, | |
| "learning_rate": 2.544642857142857e-05, | |
| "loss": 0.1016, | |
| "step": 18150 | |
| }, | |
| { | |
| "epoch": 14.77, | |
| "learning_rate": 2.537878787878788e-05, | |
| "loss": 0.1218, | |
| "step": 18200 | |
| }, | |
| { | |
| "epoch": 14.81, | |
| "learning_rate": 2.5311147186147187e-05, | |
| "loss": 0.1225, | |
| "step": 18250 | |
| }, | |
| { | |
| "epoch": 14.85, | |
| "learning_rate": 2.5243506493506497e-05, | |
| "loss": 0.1284, | |
| "step": 18300 | |
| }, | |
| { | |
| "epoch": 14.89, | |
| "learning_rate": 2.51758658008658e-05, | |
| "loss": 0.1163, | |
| "step": 18350 | |
| }, | |
| { | |
| "epoch": 14.94, | |
| "learning_rate": 2.5108225108225108e-05, | |
| "loss": 0.1368, | |
| "step": 18400 | |
| }, | |
| { | |
| "epoch": 14.98, | |
| "learning_rate": 2.504058441558442e-05, | |
| "loss": 0.1094, | |
| "step": 18450 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "eval_cer": 0.261040895766255, | |
| "eval_loss": 1.6068286895751953, | |
| "eval_mer": 0.30864493448386066, | |
| "eval_runtime": 217.2374, | |
| "eval_samples_per_second": 5.188, | |
| "eval_steps_per_second": 0.649, | |
| "step": 18480 | |
| }, | |
| { | |
| "epoch": 15.02, | |
| "learning_rate": 2.4972943722943722e-05, | |
| "loss": 0.122, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 15.06, | |
| "learning_rate": 2.4905303030303032e-05, | |
| "loss": 0.0876, | |
| "step": 18550 | |
| }, | |
| { | |
| "epoch": 15.1, | |
| "learning_rate": 2.483766233766234e-05, | |
| "loss": 0.0845, | |
| "step": 18600 | |
| }, | |
| { | |
| "epoch": 15.14, | |
| "learning_rate": 2.4770021645021646e-05, | |
| "loss": 0.079, | |
| "step": 18650 | |
| }, | |
| { | |
| "epoch": 15.18, | |
| "learning_rate": 2.4702380952380953e-05, | |
| "loss": 0.108, | |
| "step": 18700 | |
| }, | |
| { | |
| "epoch": 15.22, | |
| "learning_rate": 2.463474025974026e-05, | |
| "loss": 0.098, | |
| "step": 18750 | |
| }, | |
| { | |
| "epoch": 15.26, | |
| "learning_rate": 2.4567099567099567e-05, | |
| "loss": 0.0936, | |
| "step": 18800 | |
| }, | |
| { | |
| "epoch": 15.3, | |
| "learning_rate": 2.4499458874458875e-05, | |
| "loss": 0.0766, | |
| "step": 18850 | |
| }, | |
| { | |
| "epoch": 15.34, | |
| "learning_rate": 2.4431818181818185e-05, | |
| "loss": 0.0993, | |
| "step": 18900 | |
| }, | |
| { | |
| "epoch": 15.38, | |
| "learning_rate": 2.436417748917749e-05, | |
| "loss": 0.0851, | |
| "step": 18950 | |
| }, | |
| { | |
| "epoch": 15.42, | |
| "learning_rate": 2.42965367965368e-05, | |
| "loss": 0.1116, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 15.46, | |
| "learning_rate": 2.4228896103896103e-05, | |
| "loss": 0.0895, | |
| "step": 19050 | |
| }, | |
| { | |
| "epoch": 15.5, | |
| "learning_rate": 2.4161255411255413e-05, | |
| "loss": 0.0996, | |
| "step": 19100 | |
| }, | |
| { | |
| "epoch": 15.54, | |
| "learning_rate": 2.409361471861472e-05, | |
| "loss": 0.0965, | |
| "step": 19150 | |
| }, | |
| { | |
| "epoch": 15.58, | |
| "learning_rate": 2.4025974025974027e-05, | |
| "loss": 0.0833, | |
| "step": 19200 | |
| }, | |
| { | |
| "epoch": 15.62, | |
| "learning_rate": 2.3958333333333334e-05, | |
| "loss": 0.0886, | |
| "step": 19250 | |
| }, | |
| { | |
| "epoch": 15.67, | |
| "learning_rate": 2.389069264069264e-05, | |
| "loss": 0.1058, | |
| "step": 19300 | |
| }, | |
| { | |
| "epoch": 15.71, | |
| "learning_rate": 2.382305194805195e-05, | |
| "loss": 0.1254, | |
| "step": 19350 | |
| }, | |
| { | |
| "epoch": 15.75, | |
| "learning_rate": 2.3755411255411255e-05, | |
| "loss": 0.1088, | |
| "step": 19400 | |
| }, | |
| { | |
| "epoch": 15.79, | |
| "learning_rate": 2.3687770562770566e-05, | |
| "loss": 0.1121, | |
| "step": 19450 | |
| }, | |
| { | |
| "epoch": 15.83, | |
| "learning_rate": 2.362012987012987e-05, | |
| "loss": 0.1078, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 15.87, | |
| "learning_rate": 2.355248917748918e-05, | |
| "loss": 0.0855, | |
| "step": 19550 | |
| }, | |
| { | |
| "epoch": 15.91, | |
| "learning_rate": 2.3484848484848487e-05, | |
| "loss": 0.1041, | |
| "step": 19600 | |
| }, | |
| { | |
| "epoch": 15.95, | |
| "learning_rate": 2.3417207792207794e-05, | |
| "loss": 0.0947, | |
| "step": 19650 | |
| }, | |
| { | |
| "epoch": 15.99, | |
| "learning_rate": 2.33495670995671e-05, | |
| "loss": 0.204, | |
| "step": 19700 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "eval_cer": 0.2614253448027296, | |
| "eval_loss": 1.6343796253204346, | |
| "eval_mer": 0.30752636625119845, | |
| "eval_runtime": 215.5234, | |
| "eval_samples_per_second": 5.229, | |
| "eval_steps_per_second": 0.654, | |
| "step": 19712 | |
| }, | |
| { | |
| "epoch": 16.03, | |
| "learning_rate": 2.3281926406926408e-05, | |
| "loss": 0.0996, | |
| "step": 19750 | |
| }, | |
| { | |
| "epoch": 16.07, | |
| "learning_rate": 2.3214285714285715e-05, | |
| "loss": 0.0752, | |
| "step": 19800 | |
| }, | |
| { | |
| "epoch": 16.11, | |
| "learning_rate": 2.314664502164502e-05, | |
| "loss": 0.0983, | |
| "step": 19850 | |
| }, | |
| { | |
| "epoch": 16.15, | |
| "learning_rate": 2.3079004329004332e-05, | |
| "loss": 0.0698, | |
| "step": 19900 | |
| }, | |
| { | |
| "epoch": 16.19, | |
| "learning_rate": 2.3011363636363636e-05, | |
| "loss": 0.0724, | |
| "step": 19950 | |
| }, | |
| { | |
| "epoch": 16.23, | |
| "learning_rate": 2.2943722943722946e-05, | |
| "loss": 0.088, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 16.27, | |
| "learning_rate": 2.287608225108225e-05, | |
| "loss": 0.1087, | |
| "step": 20050 | |
| }, | |
| { | |
| "epoch": 16.31, | |
| "learning_rate": 2.280844155844156e-05, | |
| "loss": 0.084, | |
| "step": 20100 | |
| }, | |
| { | |
| "epoch": 16.36, | |
| "learning_rate": 2.2740800865800867e-05, | |
| "loss": 0.0737, | |
| "step": 20150 | |
| }, | |
| { | |
| "epoch": 16.4, | |
| "learning_rate": 2.2673160173160174e-05, | |
| "loss": 0.1058, | |
| "step": 20200 | |
| }, | |
| { | |
| "epoch": 16.44, | |
| "learning_rate": 2.260551948051948e-05, | |
| "loss": 0.0756, | |
| "step": 20250 | |
| }, | |
| { | |
| "epoch": 16.48, | |
| "learning_rate": 2.2537878787878788e-05, | |
| "loss": 0.0806, | |
| "step": 20300 | |
| }, | |
| { | |
| "epoch": 16.52, | |
| "learning_rate": 2.2470238095238095e-05, | |
| "loss": 0.0861, | |
| "step": 20350 | |
| }, | |
| { | |
| "epoch": 16.56, | |
| "learning_rate": 2.2402597402597402e-05, | |
| "loss": 0.0787, | |
| "step": 20400 | |
| }, | |
| { | |
| "epoch": 16.6, | |
| "learning_rate": 2.2334956709956713e-05, | |
| "loss": 0.079, | |
| "step": 20450 | |
| }, | |
| { | |
| "epoch": 16.64, | |
| "learning_rate": 2.2267316017316016e-05, | |
| "loss": 0.0798, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 16.68, | |
| "learning_rate": 2.2199675324675327e-05, | |
| "loss": 0.0894, | |
| "step": 20550 | |
| }, | |
| { | |
| "epoch": 16.72, | |
| "learning_rate": 2.2132034632034634e-05, | |
| "loss": 0.0994, | |
| "step": 20600 | |
| }, | |
| { | |
| "epoch": 16.76, | |
| "learning_rate": 2.206439393939394e-05, | |
| "loss": 0.0971, | |
| "step": 20650 | |
| }, | |
| { | |
| "epoch": 16.8, | |
| "learning_rate": 2.1996753246753248e-05, | |
| "loss": 0.0809, | |
| "step": 20700 | |
| }, | |
| { | |
| "epoch": 16.84, | |
| "learning_rate": 2.1929112554112555e-05, | |
| "loss": 0.0848, | |
| "step": 20750 | |
| }, | |
| { | |
| "epoch": 16.88, | |
| "learning_rate": 2.1861471861471862e-05, | |
| "loss": 0.0883, | |
| "step": 20800 | |
| }, | |
| { | |
| "epoch": 16.92, | |
| "learning_rate": 2.179383116883117e-05, | |
| "loss": 0.1039, | |
| "step": 20850 | |
| }, | |
| { | |
| "epoch": 16.96, | |
| "learning_rate": 2.172619047619048e-05, | |
| "loss": 0.1116, | |
| "step": 20900 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "eval_cer": 0.2613772886731703, | |
| "eval_loss": 1.701643466949463, | |
| "eval_mer": 0.30824544582933844, | |
| "eval_runtime": 216.1518, | |
| "eval_samples_per_second": 5.214, | |
| "eval_steps_per_second": 0.652, | |
| "step": 20944 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "learning_rate": 2.1658549783549783e-05, | |
| "loss": 0.0909, | |
| "step": 20950 | |
| }, | |
| { | |
| "epoch": 17.05, | |
| "learning_rate": 2.1590909090909093e-05, | |
| "loss": 0.0984, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 17.09, | |
| "learning_rate": 2.1523268398268397e-05, | |
| "loss": 0.0839, | |
| "step": 21050 | |
| }, | |
| { | |
| "epoch": 17.13, | |
| "learning_rate": 2.1455627705627707e-05, | |
| "loss": 0.0687, | |
| "step": 21100 | |
| }, | |
| { | |
| "epoch": 17.17, | |
| "learning_rate": 2.1387987012987014e-05, | |
| "loss": 0.0681, | |
| "step": 21150 | |
| }, | |
| { | |
| "epoch": 17.21, | |
| "learning_rate": 2.132034632034632e-05, | |
| "loss": 0.0737, | |
| "step": 21200 | |
| }, | |
| { | |
| "epoch": 17.25, | |
| "learning_rate": 2.125270562770563e-05, | |
| "loss": 0.0852, | |
| "step": 21250 | |
| }, | |
| { | |
| "epoch": 17.29, | |
| "learning_rate": 2.1185064935064935e-05, | |
| "loss": 0.0715, | |
| "step": 21300 | |
| }, | |
| { | |
| "epoch": 17.33, | |
| "learning_rate": 2.1117424242424242e-05, | |
| "loss": 0.0832, | |
| "step": 21350 | |
| }, | |
| { | |
| "epoch": 17.37, | |
| "learning_rate": 2.104978354978355e-05, | |
| "loss": 0.0956, | |
| "step": 21400 | |
| }, | |
| { | |
| "epoch": 17.41, | |
| "learning_rate": 2.098214285714286e-05, | |
| "loss": 0.075, | |
| "step": 21450 | |
| }, | |
| { | |
| "epoch": 17.45, | |
| "learning_rate": 2.0914502164502164e-05, | |
| "loss": 0.0763, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 17.49, | |
| "learning_rate": 2.0846861471861474e-05, | |
| "loss": 0.0774, | |
| "step": 21550 | |
| }, | |
| { | |
| "epoch": 17.53, | |
| "learning_rate": 2.077922077922078e-05, | |
| "loss": 0.0778, | |
| "step": 21600 | |
| }, | |
| { | |
| "epoch": 17.57, | |
| "learning_rate": 2.0711580086580088e-05, | |
| "loss": 0.0864, | |
| "step": 21650 | |
| }, | |
| { | |
| "epoch": 17.61, | |
| "learning_rate": 2.0643939393939395e-05, | |
| "loss": 0.082, | |
| "step": 21700 | |
| }, | |
| { | |
| "epoch": 17.65, | |
| "learning_rate": 2.0576298701298702e-05, | |
| "loss": 0.0764, | |
| "step": 21750 | |
| }, | |
| { | |
| "epoch": 17.69, | |
| "learning_rate": 2.050865800865801e-05, | |
| "loss": 0.0813, | |
| "step": 21800 | |
| }, | |
| { | |
| "epoch": 17.74, | |
| "learning_rate": 2.0441017316017316e-05, | |
| "loss": 0.0775, | |
| "step": 21850 | |
| }, | |
| { | |
| "epoch": 17.78, | |
| "learning_rate": 2.0373376623376626e-05, | |
| "loss": 0.0764, | |
| "step": 21900 | |
| }, | |
| { | |
| "epoch": 17.82, | |
| "learning_rate": 2.030573593073593e-05, | |
| "loss": 0.0711, | |
| "step": 21950 | |
| }, | |
| { | |
| "epoch": 17.86, | |
| "learning_rate": 2.023809523809524e-05, | |
| "loss": 0.0886, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 17.9, | |
| "learning_rate": 2.0170454545454544e-05, | |
| "loss": 0.0798, | |
| "step": 22050 | |
| }, | |
| { | |
| "epoch": 17.94, | |
| "learning_rate": 2.0102813852813855e-05, | |
| "loss": 0.086, | |
| "step": 22100 | |
| }, | |
| { | |
| "epoch": 17.98, | |
| "learning_rate": 2.003517316017316e-05, | |
| "loss": 0.0874, | |
| "step": 22150 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "eval_cer": 0.26253063578259406, | |
| "eval_loss": 1.7160123586654663, | |
| "eval_mer": 0.3069670821348674, | |
| "eval_runtime": 215.4465, | |
| "eval_samples_per_second": 5.231, | |
| "eval_steps_per_second": 0.654, | |
| "step": 22176 | |
| }, | |
| { | |
| "epoch": 18.02, | |
| "learning_rate": 1.996753246753247e-05, | |
| "loss": 0.0931, | |
| "step": 22200 | |
| }, | |
| { | |
| "epoch": 18.06, | |
| "learning_rate": 1.9899891774891776e-05, | |
| "loss": 0.069, | |
| "step": 22250 | |
| }, | |
| { | |
| "epoch": 18.1, | |
| "learning_rate": 1.9832251082251083e-05, | |
| "loss": 0.0586, | |
| "step": 22300 | |
| }, | |
| { | |
| "epoch": 18.14, | |
| "learning_rate": 1.976461038961039e-05, | |
| "loss": 0.0624, | |
| "step": 22350 | |
| }, | |
| { | |
| "epoch": 18.18, | |
| "learning_rate": 1.9696969696969697e-05, | |
| "loss": 0.0672, | |
| "step": 22400 | |
| }, | |
| { | |
| "epoch": 18.22, | |
| "learning_rate": 1.9629329004329007e-05, | |
| "loss": 0.0896, | |
| "step": 22450 | |
| }, | |
| { | |
| "epoch": 18.26, | |
| "learning_rate": 1.956168831168831e-05, | |
| "loss": 0.0876, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 18.3, | |
| "learning_rate": 1.949404761904762e-05, | |
| "loss": 0.0629, | |
| "step": 22550 | |
| }, | |
| { | |
| "epoch": 18.34, | |
| "learning_rate": 1.9426406926406928e-05, | |
| "loss": 0.0743, | |
| "step": 22600 | |
| }, | |
| { | |
| "epoch": 18.38, | |
| "learning_rate": 1.9358766233766235e-05, | |
| "loss": 0.0825, | |
| "step": 22650 | |
| }, | |
| { | |
| "epoch": 18.43, | |
| "learning_rate": 1.9291125541125542e-05, | |
| "loss": 0.0786, | |
| "step": 22700 | |
| }, | |
| { | |
| "epoch": 18.47, | |
| "learning_rate": 1.922348484848485e-05, | |
| "loss": 0.0519, | |
| "step": 22750 | |
| }, | |
| { | |
| "epoch": 18.51, | |
| "learning_rate": 1.9155844155844156e-05, | |
| "loss": 0.0624, | |
| "step": 22800 | |
| }, | |
| { | |
| "epoch": 18.55, | |
| "learning_rate": 1.9088203463203463e-05, | |
| "loss": 0.0635, | |
| "step": 22850 | |
| }, | |
| { | |
| "epoch": 18.59, | |
| "learning_rate": 1.9020562770562774e-05, | |
| "loss": 0.0611, | |
| "step": 22900 | |
| }, | |
| { | |
| "epoch": 18.63, | |
| "learning_rate": 1.8952922077922077e-05, | |
| "loss": 0.0815, | |
| "step": 22950 | |
| }, | |
| { | |
| "epoch": 18.67, | |
| "learning_rate": 1.8885281385281388e-05, | |
| "loss": 0.0776, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 18.71, | |
| "learning_rate": 1.881764069264069e-05, | |
| "loss": 0.0869, | |
| "step": 23050 | |
| }, | |
| { | |
| "epoch": 18.75, | |
| "learning_rate": 1.8750000000000002e-05, | |
| "loss": 0.06, | |
| "step": 23100 | |
| }, | |
| { | |
| "epoch": 18.79, | |
| "learning_rate": 1.868235930735931e-05, | |
| "loss": 0.0667, | |
| "step": 23150 | |
| }, | |
| { | |
| "epoch": 18.83, | |
| "learning_rate": 1.8614718614718616e-05, | |
| "loss": 0.0608, | |
| "step": 23200 | |
| }, | |
| { | |
| "epoch": 18.87, | |
| "learning_rate": 1.8547077922077923e-05, | |
| "loss": 0.0744, | |
| "step": 23250 | |
| }, | |
| { | |
| "epoch": 18.91, | |
| "learning_rate": 1.847943722943723e-05, | |
| "loss": 0.0843, | |
| "step": 23300 | |
| }, | |
| { | |
| "epoch": 18.95, | |
| "learning_rate": 1.8411796536796537e-05, | |
| "loss": 0.0877, | |
| "step": 23350 | |
| }, | |
| { | |
| "epoch": 18.99, | |
| "learning_rate": 1.8344155844155844e-05, | |
| "loss": 0.0495, | |
| "step": 23400 | |
| }, | |
| { | |
| "epoch": 19.0, | |
| "eval_cer": 0.2569561247537123, | |
| "eval_loss": 1.73832106590271, | |
| "eval_mer": 0.30377117289868966, | |
| "eval_runtime": 216.9476, | |
| "eval_samples_per_second": 5.195, | |
| "eval_steps_per_second": 0.65, | |
| "step": 23408 | |
| }, | |
| { | |
| "epoch": 19.03, | |
| "learning_rate": 1.8276515151515154e-05, | |
| "loss": 0.0714, | |
| "step": 23450 | |
| }, | |
| { | |
| "epoch": 19.07, | |
| "learning_rate": 1.8208874458874458e-05, | |
| "loss": 0.0916, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 19.12, | |
| "learning_rate": 1.814123376623377e-05, | |
| "loss": 0.0573, | |
| "step": 23550 | |
| }, | |
| { | |
| "epoch": 19.16, | |
| "learning_rate": 1.8073593073593072e-05, | |
| "loss": 0.0617, | |
| "step": 23600 | |
| }, | |
| { | |
| "epoch": 19.2, | |
| "learning_rate": 1.8005952380952382e-05, | |
| "loss": 0.0823, | |
| "step": 23650 | |
| }, | |
| { | |
| "epoch": 19.24, | |
| "learning_rate": 1.793831168831169e-05, | |
| "loss": 0.0498, | |
| "step": 23700 | |
| }, | |
| { | |
| "epoch": 19.28, | |
| "learning_rate": 1.7870670995670996e-05, | |
| "loss": 0.064, | |
| "step": 23750 | |
| }, | |
| { | |
| "epoch": 19.32, | |
| "learning_rate": 1.7803030303030303e-05, | |
| "loss": 0.0582, | |
| "step": 23800 | |
| }, | |
| { | |
| "epoch": 19.36, | |
| "learning_rate": 1.773538961038961e-05, | |
| "loss": 0.0841, | |
| "step": 23850 | |
| }, | |
| { | |
| "epoch": 19.4, | |
| "learning_rate": 1.766774891774892e-05, | |
| "loss": 0.0671, | |
| "step": 23900 | |
| }, | |
| { | |
| "epoch": 19.44, | |
| "learning_rate": 1.7600108225108224e-05, | |
| "loss": 0.0567, | |
| "step": 23950 | |
| }, | |
| { | |
| "epoch": 19.48, | |
| "learning_rate": 1.7532467532467535e-05, | |
| "loss": 0.0602, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 19.52, | |
| "learning_rate": 1.746482683982684e-05, | |
| "loss": 0.0696, | |
| "step": 24050 | |
| }, | |
| { | |
| "epoch": 19.56, | |
| "learning_rate": 1.739718614718615e-05, | |
| "loss": 0.0496, | |
| "step": 24100 | |
| }, | |
| { | |
| "epoch": 19.6, | |
| "learning_rate": 1.7329545454545456e-05, | |
| "loss": 0.0666, | |
| "step": 24150 | |
| }, | |
| { | |
| "epoch": 19.64, | |
| "learning_rate": 1.7261904761904763e-05, | |
| "loss": 0.0864, | |
| "step": 24200 | |
| }, | |
| { | |
| "epoch": 19.68, | |
| "learning_rate": 1.719426406926407e-05, | |
| "loss": 0.0633, | |
| "step": 24250 | |
| }, | |
| { | |
| "epoch": 19.72, | |
| "learning_rate": 1.7126623376623377e-05, | |
| "loss": 0.0577, | |
| "step": 24300 | |
| }, | |
| { | |
| "epoch": 19.76, | |
| "learning_rate": 1.7058982683982684e-05, | |
| "loss": 0.0527, | |
| "step": 24350 | |
| }, | |
| { | |
| "epoch": 19.81, | |
| "learning_rate": 1.699134199134199e-05, | |
| "loss": 0.0573, | |
| "step": 24400 | |
| }, | |
| { | |
| "epoch": 19.85, | |
| "learning_rate": 1.69237012987013e-05, | |
| "loss": 0.0662, | |
| "step": 24450 | |
| }, | |
| { | |
| "epoch": 19.89, | |
| "learning_rate": 1.6856060606060605e-05, | |
| "loss": 0.0595, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 19.93, | |
| "learning_rate": 1.6788419913419916e-05, | |
| "loss": 0.0667, | |
| "step": 24550 | |
| }, | |
| { | |
| "epoch": 19.97, | |
| "learning_rate": 1.672077922077922e-05, | |
| "loss": 0.0514, | |
| "step": 24600 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "eval_cer": 0.2549377673122207, | |
| "eval_loss": 1.7997801303863525, | |
| "eval_mer": 0.3000958772770853, | |
| "eval_runtime": 216.9181, | |
| "eval_samples_per_second": 5.196, | |
| "eval_steps_per_second": 0.65, | |
| "step": 24640 | |
| }, | |
| { | |
| "epoch": 20.01, | |
| "learning_rate": 1.665313852813853e-05, | |
| "loss": 0.064, | |
| "step": 24650 | |
| }, | |
| { | |
| "epoch": 20.05, | |
| "learning_rate": 1.6585497835497837e-05, | |
| "loss": 0.0531, | |
| "step": 24700 | |
| }, | |
| { | |
| "epoch": 20.09, | |
| "learning_rate": 1.6517857142857144e-05, | |
| "loss": 0.0659, | |
| "step": 24750 | |
| }, | |
| { | |
| "epoch": 20.13, | |
| "learning_rate": 1.645021645021645e-05, | |
| "loss": 0.0699, | |
| "step": 24800 | |
| }, | |
| { | |
| "epoch": 20.17, | |
| "learning_rate": 1.6382575757575758e-05, | |
| "loss": 0.073, | |
| "step": 24850 | |
| }, | |
| { | |
| "epoch": 20.21, | |
| "learning_rate": 1.6314935064935065e-05, | |
| "loss": 0.0483, | |
| "step": 24900 | |
| }, | |
| { | |
| "epoch": 20.25, | |
| "learning_rate": 1.624729437229437e-05, | |
| "loss": 0.0535, | |
| "step": 24950 | |
| }, | |
| { | |
| "epoch": 20.29, | |
| "learning_rate": 1.6179653679653682e-05, | |
| "loss": 0.0534, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 20.33, | |
| "learning_rate": 1.6112012987012986e-05, | |
| "loss": 0.0549, | |
| "step": 25050 | |
| }, | |
| { | |
| "epoch": 20.37, | |
| "learning_rate": 1.6044372294372296e-05, | |
| "loss": 0.0675, | |
| "step": 25100 | |
| }, | |
| { | |
| "epoch": 20.41, | |
| "learning_rate": 1.5976731601731603e-05, | |
| "loss": 0.0607, | |
| "step": 25150 | |
| }, | |
| { | |
| "epoch": 20.45, | |
| "learning_rate": 1.590909090909091e-05, | |
| "loss": 0.0469, | |
| "step": 25200 | |
| }, | |
| { | |
| "epoch": 20.5, | |
| "learning_rate": 1.5841450216450217e-05, | |
| "loss": 0.0484, | |
| "step": 25250 | |
| }, | |
| { | |
| "epoch": 20.54, | |
| "learning_rate": 1.5773809523809524e-05, | |
| "loss": 0.0506, | |
| "step": 25300 | |
| }, | |
| { | |
| "epoch": 20.58, | |
| "learning_rate": 1.570616883116883e-05, | |
| "loss": 0.0574, | |
| "step": 25350 | |
| }, | |
| { | |
| "epoch": 20.62, | |
| "learning_rate": 1.5638528138528138e-05, | |
| "loss": 0.0706, | |
| "step": 25400 | |
| }, | |
| { | |
| "epoch": 20.66, | |
| "learning_rate": 1.557088744588745e-05, | |
| "loss": 0.0419, | |
| "step": 25450 | |
| }, | |
| { | |
| "epoch": 20.7, | |
| "learning_rate": 1.5503246753246752e-05, | |
| "loss": 0.0583, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 20.74, | |
| "learning_rate": 1.5435606060606063e-05, | |
| "loss": 0.0766, | |
| "step": 25550 | |
| }, | |
| { | |
| "epoch": 20.78, | |
| "learning_rate": 1.5367965367965366e-05, | |
| "loss": 0.057, | |
| "step": 25600 | |
| }, | |
| { | |
| "epoch": 20.82, | |
| "learning_rate": 1.5300324675324677e-05, | |
| "loss": 0.0686, | |
| "step": 25650 | |
| }, | |
| { | |
| "epoch": 20.86, | |
| "learning_rate": 1.5232683982683984e-05, | |
| "loss": 0.0466, | |
| "step": 25700 | |
| }, | |
| { | |
| "epoch": 20.9, | |
| "learning_rate": 1.516504329004329e-05, | |
| "loss": 0.0614, | |
| "step": 25750 | |
| }, | |
| { | |
| "epoch": 20.94, | |
| "learning_rate": 1.50974025974026e-05, | |
| "loss": 0.0537, | |
| "step": 25800 | |
| }, | |
| { | |
| "epoch": 20.98, | |
| "learning_rate": 1.5029761904761905e-05, | |
| "loss": 0.0414, | |
| "step": 25850 | |
| }, | |
| { | |
| "epoch": 21.0, | |
| "eval_cer": 0.25315969051852566, | |
| "eval_loss": 1.8042283058166504, | |
| "eval_mer": 0.2984979226589965, | |
| "eval_runtime": 216.3821, | |
| "eval_samples_per_second": 5.208, | |
| "eval_steps_per_second": 0.652, | |
| "step": 25872 | |
| }, | |
| { | |
| "epoch": 21.02, | |
| "learning_rate": 1.4962121212121214e-05, | |
| "loss": 0.0628, | |
| "step": 25900 | |
| }, | |
| { | |
| "epoch": 21.06, | |
| "learning_rate": 1.4894480519480519e-05, | |
| "loss": 0.0408, | |
| "step": 25950 | |
| }, | |
| { | |
| "epoch": 21.1, | |
| "learning_rate": 1.4826839826839828e-05, | |
| "loss": 0.0497, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 21.14, | |
| "learning_rate": 1.4759199134199133e-05, | |
| "loss": 0.044, | |
| "step": 26050 | |
| }, | |
| { | |
| "epoch": 21.19, | |
| "learning_rate": 1.4691558441558442e-05, | |
| "loss": 0.0361, | |
| "step": 26100 | |
| }, | |
| { | |
| "epoch": 21.23, | |
| "learning_rate": 1.462391774891775e-05, | |
| "loss": 0.0498, | |
| "step": 26150 | |
| }, | |
| { | |
| "epoch": 21.27, | |
| "learning_rate": 1.4556277056277057e-05, | |
| "loss": 0.0485, | |
| "step": 26200 | |
| }, | |
| { | |
| "epoch": 21.31, | |
| "learning_rate": 1.4488636363636366e-05, | |
| "loss": 0.0438, | |
| "step": 26250 | |
| }, | |
| { | |
| "epoch": 21.35, | |
| "learning_rate": 1.4420995670995671e-05, | |
| "loss": 0.0749, | |
| "step": 26300 | |
| }, | |
| { | |
| "epoch": 21.39, | |
| "learning_rate": 1.435335497835498e-05, | |
| "loss": 0.0467, | |
| "step": 26350 | |
| }, | |
| { | |
| "epoch": 21.43, | |
| "learning_rate": 1.4285714285714285e-05, | |
| "loss": 0.0613, | |
| "step": 26400 | |
| }, | |
| { | |
| "epoch": 21.47, | |
| "learning_rate": 1.4218073593073594e-05, | |
| "loss": 0.0746, | |
| "step": 26450 | |
| }, | |
| { | |
| "epoch": 21.51, | |
| "learning_rate": 1.41504329004329e-05, | |
| "loss": 0.0516, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 21.55, | |
| "learning_rate": 1.4082792207792208e-05, | |
| "loss": 0.0478, | |
| "step": 26550 | |
| }, | |
| { | |
| "epoch": 21.59, | |
| "learning_rate": 1.4015151515151515e-05, | |
| "loss": 0.0412, | |
| "step": 26600 | |
| }, | |
| { | |
| "epoch": 21.63, | |
| "learning_rate": 1.3947510822510824e-05, | |
| "loss": 0.0495, | |
| "step": 26650 | |
| }, | |
| { | |
| "epoch": 21.67, | |
| "learning_rate": 1.3879870129870131e-05, | |
| "loss": 0.072, | |
| "step": 26700 | |
| }, | |
| { | |
| "epoch": 21.71, | |
| "learning_rate": 1.3812229437229438e-05, | |
| "loss": 0.054, | |
| "step": 26750 | |
| }, | |
| { | |
| "epoch": 21.75, | |
| "learning_rate": 1.3744588744588747e-05, | |
| "loss": 0.0664, | |
| "step": 26800 | |
| }, | |
| { | |
| "epoch": 21.79, | |
| "learning_rate": 1.3676948051948052e-05, | |
| "loss": 0.057, | |
| "step": 26850 | |
| }, | |
| { | |
| "epoch": 21.83, | |
| "learning_rate": 1.360930735930736e-05, | |
| "loss": 0.0503, | |
| "step": 26900 | |
| }, | |
| { | |
| "epoch": 21.88, | |
| "learning_rate": 1.3541666666666666e-05, | |
| "loss": 0.0741, | |
| "step": 26950 | |
| }, | |
| { | |
| "epoch": 21.92, | |
| "learning_rate": 1.3474025974025975e-05, | |
| "loss": 0.06, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 21.96, | |
| "learning_rate": 1.340638528138528e-05, | |
| "loss": 0.047, | |
| "step": 27050 | |
| }, | |
| { | |
| "epoch": 22.0, | |
| "learning_rate": 1.3338744588744589e-05, | |
| "loss": 0.0397, | |
| "step": 27100 | |
| }, | |
| { | |
| "epoch": 22.0, | |
| "eval_cer": 0.256908068624153, | |
| "eval_loss": 1.8340286016464233, | |
| "eval_mer": 0.3004154682007031, | |
| "eval_runtime": 216.2013, | |
| "eval_samples_per_second": 5.213, | |
| "eval_steps_per_second": 0.652, | |
| "step": 27104 | |
| }, | |
| { | |
| "epoch": 22.04, | |
| "learning_rate": 1.3271103896103898e-05, | |
| "loss": 0.055, | |
| "step": 27150 | |
| }, | |
| { | |
| "epoch": 22.08, | |
| "learning_rate": 1.3203463203463205e-05, | |
| "loss": 0.0488, | |
| "step": 27200 | |
| }, | |
| { | |
| "epoch": 22.12, | |
| "learning_rate": 1.3135822510822513e-05, | |
| "loss": 0.0452, | |
| "step": 27250 | |
| }, | |
| { | |
| "epoch": 22.16, | |
| "learning_rate": 1.3068181818181819e-05, | |
| "loss": 0.0625, | |
| "step": 27300 | |
| }, | |
| { | |
| "epoch": 22.2, | |
| "learning_rate": 1.3000541125541127e-05, | |
| "loss": 0.0555, | |
| "step": 27350 | |
| }, | |
| { | |
| "epoch": 22.24, | |
| "learning_rate": 1.2932900432900433e-05, | |
| "loss": 0.0482, | |
| "step": 27400 | |
| }, | |
| { | |
| "epoch": 22.28, | |
| "learning_rate": 1.2865259740259741e-05, | |
| "loss": 0.0447, | |
| "step": 27450 | |
| }, | |
| { | |
| "epoch": 22.32, | |
| "learning_rate": 1.2797619047619047e-05, | |
| "loss": 0.0539, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 22.36, | |
| "learning_rate": 1.2729978354978355e-05, | |
| "loss": 0.0593, | |
| "step": 27550 | |
| }, | |
| { | |
| "epoch": 22.4, | |
| "learning_rate": 1.2662337662337662e-05, | |
| "loss": 0.0418, | |
| "step": 27600 | |
| }, | |
| { | |
| "epoch": 22.44, | |
| "learning_rate": 1.259469696969697e-05, | |
| "loss": 0.0507, | |
| "step": 27650 | |
| }, | |
| { | |
| "epoch": 22.48, | |
| "learning_rate": 1.2527056277056278e-05, | |
| "loss": 0.0327, | |
| "step": 27700 | |
| }, | |
| { | |
| "epoch": 22.52, | |
| "learning_rate": 1.2459415584415585e-05, | |
| "loss": 0.0572, | |
| "step": 27750 | |
| }, | |
| { | |
| "epoch": 22.56, | |
| "learning_rate": 1.2391774891774892e-05, | |
| "loss": 0.07, | |
| "step": 27800 | |
| }, | |
| { | |
| "epoch": 22.61, | |
| "learning_rate": 1.23241341991342e-05, | |
| "loss": 0.0471, | |
| "step": 27850 | |
| }, | |
| { | |
| "epoch": 22.65, | |
| "learning_rate": 1.2256493506493508e-05, | |
| "loss": 0.059, | |
| "step": 27900 | |
| }, | |
| { | |
| "epoch": 22.69, | |
| "learning_rate": 1.2188852813852815e-05, | |
| "loss": 0.0594, | |
| "step": 27950 | |
| }, | |
| { | |
| "epoch": 22.73, | |
| "learning_rate": 1.2121212121212122e-05, | |
| "loss": 0.0368, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 22.77, | |
| "learning_rate": 1.2053571428571429e-05, | |
| "loss": 0.0663, | |
| "step": 28050 | |
| }, | |
| { | |
| "epoch": 22.81, | |
| "learning_rate": 1.1985930735930736e-05, | |
| "loss": 0.0508, | |
| "step": 28100 | |
| }, | |
| { | |
| "epoch": 22.85, | |
| "learning_rate": 1.1918290043290043e-05, | |
| "loss": 0.0621, | |
| "step": 28150 | |
| }, | |
| { | |
| "epoch": 22.89, | |
| "learning_rate": 1.1850649350649352e-05, | |
| "loss": 0.0455, | |
| "step": 28200 | |
| }, | |
| { | |
| "epoch": 22.93, | |
| "learning_rate": 1.1783008658008659e-05, | |
| "loss": 0.0606, | |
| "step": 28250 | |
| }, | |
| { | |
| "epoch": 22.97, | |
| "learning_rate": 1.1715367965367966e-05, | |
| "loss": 0.0303, | |
| "step": 28300 | |
| }, | |
| { | |
| "epoch": 23.0, | |
| "eval_cer": 0.25364025181411887, | |
| "eval_loss": 1.8699883222579956, | |
| "eval_mer": 0.29841802492809205, | |
| "eval_runtime": 216.7323, | |
| "eval_samples_per_second": 5.2, | |
| "eval_steps_per_second": 0.651, | |
| "step": 28336 | |
| }, | |
| { | |
| "epoch": 23.01, | |
| "learning_rate": 1.1647727272727273e-05, | |
| "loss": 0.0703, | |
| "step": 28350 | |
| }, | |
| { | |
| "epoch": 23.05, | |
| "learning_rate": 1.1580086580086581e-05, | |
| "loss": 0.0535, | |
| "step": 28400 | |
| }, | |
| { | |
| "epoch": 23.09, | |
| "learning_rate": 1.1512445887445889e-05, | |
| "loss": 0.0511, | |
| "step": 28450 | |
| }, | |
| { | |
| "epoch": 23.13, | |
| "learning_rate": 1.1444805194805196e-05, | |
| "loss": 0.0423, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 23.17, | |
| "learning_rate": 1.1377164502164503e-05, | |
| "loss": 0.0577, | |
| "step": 28550 | |
| }, | |
| { | |
| "epoch": 23.21, | |
| "learning_rate": 1.130952380952381e-05, | |
| "loss": 0.0388, | |
| "step": 28600 | |
| }, | |
| { | |
| "epoch": 23.25, | |
| "learning_rate": 1.1241883116883117e-05, | |
| "loss": 0.0426, | |
| "step": 28650 | |
| }, | |
| { | |
| "epoch": 23.3, | |
| "learning_rate": 1.1174242424242425e-05, | |
| "loss": 0.0597, | |
| "step": 28700 | |
| }, | |
| { | |
| "epoch": 23.34, | |
| "learning_rate": 1.1106601731601732e-05, | |
| "loss": 0.0409, | |
| "step": 28750 | |
| }, | |
| { | |
| "epoch": 23.38, | |
| "learning_rate": 1.103896103896104e-05, | |
| "loss": 0.0542, | |
| "step": 28800 | |
| }, | |
| { | |
| "epoch": 23.42, | |
| "learning_rate": 1.0971320346320346e-05, | |
| "loss": 0.0396, | |
| "step": 28850 | |
| }, | |
| { | |
| "epoch": 23.46, | |
| "learning_rate": 1.0903679653679655e-05, | |
| "loss": 0.0713, | |
| "step": 28900 | |
| }, | |
| { | |
| "epoch": 23.5, | |
| "learning_rate": 1.0836038961038962e-05, | |
| "loss": 0.0322, | |
| "step": 28950 | |
| }, | |
| { | |
| "epoch": 23.54, | |
| "learning_rate": 1.0768398268398269e-05, | |
| "loss": 0.0392, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 23.58, | |
| "learning_rate": 1.0700757575757576e-05, | |
| "loss": 0.0495, | |
| "step": 29050 | |
| }, | |
| { | |
| "epoch": 23.62, | |
| "learning_rate": 1.0633116883116883e-05, | |
| "loss": 0.0474, | |
| "step": 29100 | |
| }, | |
| { | |
| "epoch": 23.66, | |
| "learning_rate": 1.056547619047619e-05, | |
| "loss": 0.0647, | |
| "step": 29150 | |
| }, | |
| { | |
| "epoch": 23.7, | |
| "learning_rate": 1.0497835497835499e-05, | |
| "loss": 0.0423, | |
| "step": 29200 | |
| }, | |
| { | |
| "epoch": 23.74, | |
| "learning_rate": 1.0430194805194806e-05, | |
| "loss": 0.0563, | |
| "step": 29250 | |
| }, | |
| { | |
| "epoch": 23.78, | |
| "learning_rate": 1.0362554112554113e-05, | |
| "loss": 0.0581, | |
| "step": 29300 | |
| }, | |
| { | |
| "epoch": 23.82, | |
| "learning_rate": 1.029491341991342e-05, | |
| "loss": 0.0321, | |
| "step": 29350 | |
| }, | |
| { | |
| "epoch": 23.86, | |
| "learning_rate": 1.0227272727272729e-05, | |
| "loss": 0.0433, | |
| "step": 29400 | |
| }, | |
| { | |
| "epoch": 23.9, | |
| "learning_rate": 1.0159632034632036e-05, | |
| "loss": 0.0589, | |
| "step": 29450 | |
| }, | |
| { | |
| "epoch": 23.94, | |
| "learning_rate": 1.0091991341991343e-05, | |
| "loss": 0.0347, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 23.99, | |
| "learning_rate": 1.002435064935065e-05, | |
| "loss": 0.0552, | |
| "step": 29550 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "eval_cer": 0.253207746648085, | |
| "eval_loss": 1.9628108739852905, | |
| "eval_mer": 0.29729945669542984, | |
| "eval_runtime": 215.5186, | |
| "eval_samples_per_second": 5.229, | |
| "eval_steps_per_second": 0.654, | |
| "step": 29568 | |
| }, | |
| { | |
| "epoch": 24.03, | |
| "learning_rate": 9.956709956709957e-06, | |
| "loss": 0.0639, | |
| "step": 29600 | |
| }, | |
| { | |
| "epoch": 24.07, | |
| "learning_rate": 9.889069264069264e-06, | |
| "loss": 0.049, | |
| "step": 29650 | |
| }, | |
| { | |
| "epoch": 24.11, | |
| "learning_rate": 9.821428571428573e-06, | |
| "loss": 0.0438, | |
| "step": 29700 | |
| }, | |
| { | |
| "epoch": 24.15, | |
| "learning_rate": 9.75378787878788e-06, | |
| "loss": 0.0417, | |
| "step": 29750 | |
| }, | |
| { | |
| "epoch": 24.19, | |
| "learning_rate": 9.686147186147187e-06, | |
| "loss": 0.0402, | |
| "step": 29800 | |
| }, | |
| { | |
| "epoch": 24.23, | |
| "learning_rate": 9.618506493506494e-06, | |
| "loss": 0.0517, | |
| "step": 29850 | |
| }, | |
| { | |
| "epoch": 24.27, | |
| "learning_rate": 9.550865800865802e-06, | |
| "loss": 0.0439, | |
| "step": 29900 | |
| }, | |
| { | |
| "epoch": 24.31, | |
| "learning_rate": 9.48322510822511e-06, | |
| "loss": 0.0426, | |
| "step": 29950 | |
| }, | |
| { | |
| "epoch": 24.35, | |
| "learning_rate": 9.415584415584416e-06, | |
| "loss": 0.0567, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 24.39, | |
| "learning_rate": 9.347943722943723e-06, | |
| "loss": 0.0529, | |
| "step": 30050 | |
| }, | |
| { | |
| "epoch": 24.43, | |
| "learning_rate": 9.28030303030303e-06, | |
| "loss": 0.0479, | |
| "step": 30100 | |
| }, | |
| { | |
| "epoch": 24.47, | |
| "learning_rate": 9.212662337662337e-06, | |
| "loss": 0.0483, | |
| "step": 30150 | |
| }, | |
| { | |
| "epoch": 24.51, | |
| "learning_rate": 9.145021645021646e-06, | |
| "loss": 0.0385, | |
| "step": 30200 | |
| }, | |
| { | |
| "epoch": 24.55, | |
| "learning_rate": 9.077380952380953e-06, | |
| "loss": 0.048, | |
| "step": 30250 | |
| }, | |
| { | |
| "epoch": 24.59, | |
| "learning_rate": 9.00974025974026e-06, | |
| "loss": 0.0458, | |
| "step": 30300 | |
| }, | |
| { | |
| "epoch": 24.63, | |
| "learning_rate": 8.942099567099567e-06, | |
| "loss": 0.0519, | |
| "step": 30350 | |
| }, | |
| { | |
| "epoch": 24.68, | |
| "learning_rate": 8.874458874458876e-06, | |
| "loss": 0.0443, | |
| "step": 30400 | |
| }, | |
| { | |
| "epoch": 24.72, | |
| "learning_rate": 8.806818181818183e-06, | |
| "loss": 0.0636, | |
| "step": 30450 | |
| }, | |
| { | |
| "epoch": 24.76, | |
| "learning_rate": 8.73917748917749e-06, | |
| "loss": 0.0398, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 24.8, | |
| "learning_rate": 8.671536796536797e-06, | |
| "loss": 0.0587, | |
| "step": 30550 | |
| }, | |
| { | |
| "epoch": 24.84, | |
| "learning_rate": 8.603896103896104e-06, | |
| "loss": 0.05, | |
| "step": 30600 | |
| }, | |
| { | |
| "epoch": 24.88, | |
| "learning_rate": 8.536255411255411e-06, | |
| "loss": 0.0364, | |
| "step": 30650 | |
| }, | |
| { | |
| "epoch": 24.92, | |
| "learning_rate": 8.468614718614718e-06, | |
| "loss": 0.0454, | |
| "step": 30700 | |
| }, | |
| { | |
| "epoch": 24.96, | |
| "learning_rate": 8.400974025974027e-06, | |
| "loss": 0.0432, | |
| "step": 30750 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "learning_rate": 8.333333333333334e-06, | |
| "loss": 0.0401, | |
| "step": 30800 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "eval_cer": 0.2528713537411697, | |
| "eval_loss": 1.9063920974731445, | |
| "eval_mer": 0.29825822946628316, | |
| "eval_runtime": 216.7371, | |
| "eval_samples_per_second": 5.2, | |
| "eval_steps_per_second": 0.651, | |
| "step": 30800 | |
| }, | |
| { | |
| "epoch": 25.04, | |
| "learning_rate": 8.26569264069264e-06, | |
| "loss": 0.0462, | |
| "step": 30850 | |
| }, | |
| { | |
| "epoch": 25.08, | |
| "learning_rate": 8.19805194805195e-06, | |
| "loss": 0.0334, | |
| "step": 30900 | |
| }, | |
| { | |
| "epoch": 25.12, | |
| "learning_rate": 8.130411255411256e-06, | |
| "loss": 0.0404, | |
| "step": 30950 | |
| }, | |
| { | |
| "epoch": 25.16, | |
| "learning_rate": 8.062770562770564e-06, | |
| "loss": 0.0527, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 25.2, | |
| "learning_rate": 7.99512987012987e-06, | |
| "loss": 0.0458, | |
| "step": 31050 | |
| }, | |
| { | |
| "epoch": 25.24, | |
| "learning_rate": 7.927489177489178e-06, | |
| "loss": 0.0502, | |
| "step": 31100 | |
| }, | |
| { | |
| "epoch": 25.28, | |
| "learning_rate": 7.859848484848485e-06, | |
| "loss": 0.0313, | |
| "step": 31150 | |
| }, | |
| { | |
| "epoch": 25.32, | |
| "learning_rate": 7.792207792207792e-06, | |
| "loss": 0.0302, | |
| "step": 31200 | |
| }, | |
| { | |
| "epoch": 25.37, | |
| "learning_rate": 7.7245670995671e-06, | |
| "loss": 0.0432, | |
| "step": 31250 | |
| }, | |
| { | |
| "epoch": 25.41, | |
| "learning_rate": 7.656926406926407e-06, | |
| "loss": 0.0735, | |
| "step": 31300 | |
| }, | |
| { | |
| "epoch": 25.45, | |
| "learning_rate": 7.589285714285714e-06, | |
| "loss": 0.0408, | |
| "step": 31350 | |
| }, | |
| { | |
| "epoch": 25.49, | |
| "learning_rate": 7.521645021645023e-06, | |
| "loss": 0.0249, | |
| "step": 31400 | |
| }, | |
| { | |
| "epoch": 25.53, | |
| "learning_rate": 7.45400432900433e-06, | |
| "loss": 0.0527, | |
| "step": 31450 | |
| }, | |
| { | |
| "epoch": 25.57, | |
| "learning_rate": 7.386363636363637e-06, | |
| "loss": 0.0442, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 25.61, | |
| "learning_rate": 7.318722943722944e-06, | |
| "loss": 0.0394, | |
| "step": 31550 | |
| }, | |
| { | |
| "epoch": 25.65, | |
| "learning_rate": 7.251082251082251e-06, | |
| "loss": 0.0363, | |
| "step": 31600 | |
| }, | |
| { | |
| "epoch": 25.69, | |
| "learning_rate": 7.183441558441559e-06, | |
| "loss": 0.0359, | |
| "step": 31650 | |
| }, | |
| { | |
| "epoch": 25.73, | |
| "learning_rate": 7.115800865800866e-06, | |
| "loss": 0.0367, | |
| "step": 31700 | |
| }, | |
| { | |
| "epoch": 25.77, | |
| "learning_rate": 7.048160173160173e-06, | |
| "loss": 0.0451, | |
| "step": 31750 | |
| }, | |
| { | |
| "epoch": 25.81, | |
| "learning_rate": 6.98051948051948e-06, | |
| "loss": 0.0373, | |
| "step": 31800 | |
| }, | |
| { | |
| "epoch": 25.85, | |
| "learning_rate": 6.912878787878788e-06, | |
| "loss": 0.0452, | |
| "step": 31850 | |
| }, | |
| { | |
| "epoch": 25.89, | |
| "learning_rate": 6.845238095238096e-06, | |
| "loss": 0.046, | |
| "step": 31900 | |
| }, | |
| { | |
| "epoch": 25.93, | |
| "learning_rate": 6.777597402597404e-06, | |
| "loss": 0.0466, | |
| "step": 31950 | |
| }, | |
| { | |
| "epoch": 25.97, | |
| "learning_rate": 6.709956709956711e-06, | |
| "loss": 0.0417, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 26.0, | |
| "eval_cer": 0.25364025181411887, | |
| "eval_loss": 1.9402902126312256, | |
| "eval_mer": 0.2979386385426654, | |
| "eval_runtime": 217.1614, | |
| "eval_samples_per_second": 5.19, | |
| "eval_steps_per_second": 0.649, | |
| "step": 32032 | |
| }, | |
| { | |
| "epoch": 26.01, | |
| "learning_rate": 6.642316017316018e-06, | |
| "loss": 0.0395, | |
| "step": 32050 | |
| }, | |
| { | |
| "epoch": 26.06, | |
| "learning_rate": 6.574675324675325e-06, | |
| "loss": 0.046, | |
| "step": 32100 | |
| }, | |
| { | |
| "epoch": 26.1, | |
| "learning_rate": 6.507034632034633e-06, | |
| "loss": 0.0412, | |
| "step": 32150 | |
| }, | |
| { | |
| "epoch": 26.14, | |
| "learning_rate": 6.43939393939394e-06, | |
| "loss": 0.0322, | |
| "step": 32200 | |
| }, | |
| { | |
| "epoch": 26.18, | |
| "learning_rate": 6.371753246753247e-06, | |
| "loss": 0.0337, | |
| "step": 32250 | |
| }, | |
| { | |
| "epoch": 26.22, | |
| "learning_rate": 6.304112554112554e-06, | |
| "loss": 0.0366, | |
| "step": 32300 | |
| }, | |
| { | |
| "epoch": 26.26, | |
| "learning_rate": 6.2364718614718615e-06, | |
| "loss": 0.035, | |
| "step": 32350 | |
| }, | |
| { | |
| "epoch": 26.3, | |
| "learning_rate": 6.168831168831169e-06, | |
| "loss": 0.048, | |
| "step": 32400 | |
| }, | |
| { | |
| "epoch": 26.34, | |
| "learning_rate": 6.101190476190476e-06, | |
| "loss": 0.0188, | |
| "step": 32450 | |
| }, | |
| { | |
| "epoch": 26.38, | |
| "learning_rate": 6.0335497835497834e-06, | |
| "loss": 0.0469, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 26.42, | |
| "learning_rate": 5.965909090909091e-06, | |
| "loss": 0.03, | |
| "step": 32550 | |
| }, | |
| { | |
| "epoch": 26.46, | |
| "learning_rate": 5.898268398268398e-06, | |
| "loss": 0.0509, | |
| "step": 32600 | |
| }, | |
| { | |
| "epoch": 26.5, | |
| "learning_rate": 5.830627705627706e-06, | |
| "loss": 0.053, | |
| "step": 32650 | |
| }, | |
| { | |
| "epoch": 26.54, | |
| "learning_rate": 5.762987012987013e-06, | |
| "loss": 0.0413, | |
| "step": 32700 | |
| }, | |
| { | |
| "epoch": 26.58, | |
| "learning_rate": 5.69534632034632e-06, | |
| "loss": 0.0343, | |
| "step": 32750 | |
| }, | |
| { | |
| "epoch": 26.62, | |
| "learning_rate": 5.627705627705628e-06, | |
| "loss": 0.046, | |
| "step": 32800 | |
| }, | |
| { | |
| "epoch": 26.66, | |
| "learning_rate": 5.560064935064935e-06, | |
| "loss": 0.0529, | |
| "step": 32850 | |
| }, | |
| { | |
| "epoch": 26.7, | |
| "learning_rate": 5.492424242424243e-06, | |
| "loss": 0.0425, | |
| "step": 32900 | |
| }, | |
| { | |
| "epoch": 26.75, | |
| "learning_rate": 5.42478354978355e-06, | |
| "loss": 0.0322, | |
| "step": 32950 | |
| }, | |
| { | |
| "epoch": 26.79, | |
| "learning_rate": 5.357142857142857e-06, | |
| "loss": 0.044, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 26.83, | |
| "learning_rate": 5.289502164502165e-06, | |
| "loss": 0.0329, | |
| "step": 33050 | |
| }, | |
| { | |
| "epoch": 26.87, | |
| "learning_rate": 5.221861471861472e-06, | |
| "loss": 0.0385, | |
| "step": 33100 | |
| }, | |
| { | |
| "epoch": 26.91, | |
| "learning_rate": 5.15422077922078e-06, | |
| "loss": 0.0436, | |
| "step": 33150 | |
| }, | |
| { | |
| "epoch": 26.95, | |
| "learning_rate": 5.086580086580087e-06, | |
| "loss": 0.0524, | |
| "step": 33200 | |
| }, | |
| { | |
| "epoch": 26.99, | |
| "learning_rate": 5.018939393939394e-06, | |
| "loss": 0.0301, | |
| "step": 33250 | |
| }, | |
| { | |
| "epoch": 27.0, | |
| "eval_cer": 0.2514296698543899, | |
| "eval_loss": 1.915743350982666, | |
| "eval_mer": 0.29658037711728985, | |
| "eval_runtime": 215.5246, | |
| "eval_samples_per_second": 5.229, | |
| "eval_steps_per_second": 0.654, | |
| "step": 33264 | |
| }, | |
| { | |
| "epoch": 27.03, | |
| "learning_rate": 4.951298701298702e-06, | |
| "loss": 0.0274, | |
| "step": 33300 | |
| }, | |
| { | |
| "epoch": 27.07, | |
| "learning_rate": 4.883658008658009e-06, | |
| "loss": 0.0548, | |
| "step": 33350 | |
| }, | |
| { | |
| "epoch": 27.11, | |
| "learning_rate": 4.816017316017317e-06, | |
| "loss": 0.0354, | |
| "step": 33400 | |
| }, | |
| { | |
| "epoch": 27.15, | |
| "learning_rate": 4.748376623376624e-06, | |
| "loss": 0.0333, | |
| "step": 33450 | |
| }, | |
| { | |
| "epoch": 27.19, | |
| "learning_rate": 4.680735930735931e-06, | |
| "loss": 0.0361, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 27.23, | |
| "learning_rate": 4.6130952380952385e-06, | |
| "loss": 0.0292, | |
| "step": 33550 | |
| }, | |
| { | |
| "epoch": 27.27, | |
| "learning_rate": 4.5454545454545455e-06, | |
| "loss": 0.0312, | |
| "step": 33600 | |
| }, | |
| { | |
| "epoch": 27.31, | |
| "learning_rate": 4.477813852813853e-06, | |
| "loss": 0.0365, | |
| "step": 33650 | |
| }, | |
| { | |
| "epoch": 27.35, | |
| "learning_rate": 4.41017316017316e-06, | |
| "loss": 0.0383, | |
| "step": 33700 | |
| }, | |
| { | |
| "epoch": 27.39, | |
| "learning_rate": 4.3425324675324674e-06, | |
| "loss": 0.0331, | |
| "step": 33750 | |
| }, | |
| { | |
| "epoch": 27.44, | |
| "learning_rate": 4.274891774891775e-06, | |
| "loss": 0.024, | |
| "step": 33800 | |
| }, | |
| { | |
| "epoch": 27.48, | |
| "learning_rate": 4.207251082251082e-06, | |
| "loss": 0.0417, | |
| "step": 33850 | |
| }, | |
| { | |
| "epoch": 27.52, | |
| "learning_rate": 4.13961038961039e-06, | |
| "loss": 0.0447, | |
| "step": 33900 | |
| }, | |
| { | |
| "epoch": 27.56, | |
| "learning_rate": 4.071969696969697e-06, | |
| "loss": 0.0436, | |
| "step": 33950 | |
| }, | |
| { | |
| "epoch": 27.6, | |
| "learning_rate": 4.004329004329004e-06, | |
| "loss": 0.0597, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 27.64, | |
| "learning_rate": 3.936688311688312e-06, | |
| "loss": 0.0297, | |
| "step": 34050 | |
| }, | |
| { | |
| "epoch": 27.68, | |
| "learning_rate": 3.869047619047619e-06, | |
| "loss": 0.0383, | |
| "step": 34100 | |
| }, | |
| { | |
| "epoch": 27.72, | |
| "learning_rate": 3.8014069264069266e-06, | |
| "loss": 0.0443, | |
| "step": 34150 | |
| }, | |
| { | |
| "epoch": 27.76, | |
| "learning_rate": 3.733766233766234e-06, | |
| "loss": 0.0339, | |
| "step": 34200 | |
| }, | |
| { | |
| "epoch": 27.8, | |
| "learning_rate": 3.666125541125541e-06, | |
| "loss": 0.0338, | |
| "step": 34250 | |
| }, | |
| { | |
| "epoch": 27.84, | |
| "learning_rate": 3.598484848484849e-06, | |
| "loss": 0.0271, | |
| "step": 34300 | |
| }, | |
| { | |
| "epoch": 27.88, | |
| "learning_rate": 3.5308441558441563e-06, | |
| "loss": 0.0296, | |
| "step": 34350 | |
| }, | |
| { | |
| "epoch": 27.92, | |
| "learning_rate": 3.4632034632034634e-06, | |
| "loss": 0.0314, | |
| "step": 34400 | |
| }, | |
| { | |
| "epoch": 27.96, | |
| "learning_rate": 3.3955627705627704e-06, | |
| "loss": 0.0267, | |
| "step": 34450 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "eval_cer": 0.2515738382430679, | |
| "eval_loss": 1.982408881187439, | |
| "eval_mer": 0.2975391498881432, | |
| "eval_runtime": 217.1186, | |
| "eval_samples_per_second": 5.191, | |
| "eval_steps_per_second": 0.649, | |
| "step": 34496 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "learning_rate": 3.327922077922078e-06, | |
| "loss": 0.0364, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 28.04, | |
| "learning_rate": 3.2602813852813857e-06, | |
| "loss": 0.0339, | |
| "step": 34550 | |
| }, | |
| { | |
| "epoch": 28.08, | |
| "learning_rate": 3.192640692640693e-06, | |
| "loss": 0.0314, | |
| "step": 34600 | |
| }, | |
| { | |
| "epoch": 28.12, | |
| "learning_rate": 3.125e-06, | |
| "loss": 0.0226, | |
| "step": 34650 | |
| }, | |
| { | |
| "epoch": 28.17, | |
| "learning_rate": 3.057359307359307e-06, | |
| "loss": 0.0422, | |
| "step": 34700 | |
| }, | |
| { | |
| "epoch": 28.21, | |
| "learning_rate": 2.989718614718615e-06, | |
| "loss": 0.0367, | |
| "step": 34750 | |
| }, | |
| { | |
| "epoch": 28.25, | |
| "learning_rate": 2.922077922077922e-06, | |
| "loss": 0.0462, | |
| "step": 34800 | |
| }, | |
| { | |
| "epoch": 28.29, | |
| "learning_rate": 2.85443722943723e-06, | |
| "loss": 0.0377, | |
| "step": 34850 | |
| }, | |
| { | |
| "epoch": 28.33, | |
| "learning_rate": 2.786796536796537e-06, | |
| "loss": 0.0295, | |
| "step": 34900 | |
| }, | |
| { | |
| "epoch": 28.37, | |
| "learning_rate": 2.719155844155844e-06, | |
| "loss": 0.0313, | |
| "step": 34950 | |
| }, | |
| { | |
| "epoch": 28.41, | |
| "learning_rate": 2.651515151515152e-06, | |
| "loss": 0.0341, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 28.45, | |
| "learning_rate": 2.583874458874459e-06, | |
| "loss": 0.0451, | |
| "step": 35050 | |
| }, | |
| { | |
| "epoch": 28.49, | |
| "learning_rate": 2.5162337662337663e-06, | |
| "loss": 0.0432, | |
| "step": 35100 | |
| }, | |
| { | |
| "epoch": 28.53, | |
| "learning_rate": 2.4485930735930737e-06, | |
| "loss": 0.0531, | |
| "step": 35150 | |
| }, | |
| { | |
| "epoch": 28.57, | |
| "learning_rate": 2.3809523809523808e-06, | |
| "loss": 0.0283, | |
| "step": 35200 | |
| }, | |
| { | |
| "epoch": 28.61, | |
| "learning_rate": 2.3133116883116886e-06, | |
| "loss": 0.0256, | |
| "step": 35250 | |
| }, | |
| { | |
| "epoch": 28.65, | |
| "learning_rate": 2.2456709956709957e-06, | |
| "loss": 0.0289, | |
| "step": 35300 | |
| }, | |
| { | |
| "epoch": 28.69, | |
| "learning_rate": 2.178030303030303e-06, | |
| "loss": 0.0277, | |
| "step": 35350 | |
| }, | |
| { | |
| "epoch": 28.73, | |
| "learning_rate": 2.1103896103896105e-06, | |
| "loss": 0.0239, | |
| "step": 35400 | |
| }, | |
| { | |
| "epoch": 28.77, | |
| "learning_rate": 2.0427489177489176e-06, | |
| "loss": 0.0483, | |
| "step": 35450 | |
| }, | |
| { | |
| "epoch": 28.81, | |
| "learning_rate": 1.9751082251082254e-06, | |
| "loss": 0.0449, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 28.86, | |
| "learning_rate": 1.9074675324675324e-06, | |
| "loss": 0.0276, | |
| "step": 35550 | |
| }, | |
| { | |
| "epoch": 28.9, | |
| "learning_rate": 1.83982683982684e-06, | |
| "loss": 0.0276, | |
| "step": 35600 | |
| }, | |
| { | |
| "epoch": 28.94, | |
| "learning_rate": 1.7721861471861473e-06, | |
| "loss": 0.0466, | |
| "step": 35650 | |
| }, | |
| { | |
| "epoch": 28.98, | |
| "learning_rate": 1.7045454545454546e-06, | |
| "loss": 0.0448, | |
| "step": 35700 | |
| }, | |
| { | |
| "epoch": 29.0, | |
| "eval_cer": 0.25046854726320344, | |
| "eval_loss": 1.9711192846298218, | |
| "eval_mer": 0.2958612975391499, | |
| "eval_runtime": 217.3139, | |
| "eval_samples_per_second": 5.186, | |
| "eval_steps_per_second": 0.649, | |
| "step": 35728 | |
| }, | |
| { | |
| "epoch": 29.02, | |
| "learning_rate": 1.636904761904762e-06, | |
| "loss": 0.0394, | |
| "step": 35750 | |
| }, | |
| { | |
| "epoch": 29.06, | |
| "learning_rate": 1.5692640692640692e-06, | |
| "loss": 0.0343, | |
| "step": 35800 | |
| }, | |
| { | |
| "epoch": 29.1, | |
| "learning_rate": 1.5016233766233767e-06, | |
| "loss": 0.0333, | |
| "step": 35850 | |
| }, | |
| { | |
| "epoch": 29.14, | |
| "learning_rate": 1.4339826839826841e-06, | |
| "loss": 0.0462, | |
| "step": 35900 | |
| }, | |
| { | |
| "epoch": 29.18, | |
| "learning_rate": 1.3663419913419914e-06, | |
| "loss": 0.0377, | |
| "step": 35950 | |
| }, | |
| { | |
| "epoch": 29.22, | |
| "learning_rate": 1.2987012987012988e-06, | |
| "loss": 0.0233, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 29.26, | |
| "learning_rate": 1.231060606060606e-06, | |
| "loss": 0.031, | |
| "step": 36050 | |
| }, | |
| { | |
| "epoch": 29.3, | |
| "learning_rate": 1.1634199134199135e-06, | |
| "loss": 0.0302, | |
| "step": 36100 | |
| }, | |
| { | |
| "epoch": 29.34, | |
| "learning_rate": 1.095779220779221e-06, | |
| "loss": 0.0397, | |
| "step": 36150 | |
| }, | |
| { | |
| "epoch": 29.38, | |
| "learning_rate": 1.0281385281385282e-06, | |
| "loss": 0.0266, | |
| "step": 36200 | |
| }, | |
| { | |
| "epoch": 29.42, | |
| "learning_rate": 9.604978354978354e-07, | |
| "loss": 0.0242, | |
| "step": 36250 | |
| }, | |
| { | |
| "epoch": 29.46, | |
| "learning_rate": 8.928571428571428e-07, | |
| "loss": 0.0339, | |
| "step": 36300 | |
| }, | |
| { | |
| "epoch": 29.5, | |
| "learning_rate": 8.252164502164503e-07, | |
| "loss": 0.0364, | |
| "step": 36350 | |
| }, | |
| { | |
| "epoch": 29.55, | |
| "learning_rate": 7.575757575757576e-07, | |
| "loss": 0.0485, | |
| "step": 36400 | |
| }, | |
| { | |
| "epoch": 29.59, | |
| "learning_rate": 6.89935064935065e-07, | |
| "loss": 0.0299, | |
| "step": 36450 | |
| }, | |
| { | |
| "epoch": 29.63, | |
| "learning_rate": 6.222943722943723e-07, | |
| "loss": 0.0276, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 29.67, | |
| "learning_rate": 5.546536796536797e-07, | |
| "loss": 0.04, | |
| "step": 36550 | |
| }, | |
| { | |
| "epoch": 29.71, | |
| "learning_rate": 4.87012987012987e-07, | |
| "loss": 0.0426, | |
| "step": 36600 | |
| }, | |
| { | |
| "epoch": 29.75, | |
| "learning_rate": 4.193722943722944e-07, | |
| "loss": 0.0266, | |
| "step": 36650 | |
| }, | |
| { | |
| "epoch": 29.79, | |
| "learning_rate": 3.5173160173160176e-07, | |
| "loss": 0.0222, | |
| "step": 36700 | |
| }, | |
| { | |
| "epoch": 29.83, | |
| "learning_rate": 2.840909090909091e-07, | |
| "loss": 0.0347, | |
| "step": 36750 | |
| }, | |
| { | |
| "epoch": 29.87, | |
| "learning_rate": 2.1645021645021646e-07, | |
| "loss": 0.041, | |
| "step": 36800 | |
| }, | |
| { | |
| "epoch": 29.91, | |
| "learning_rate": 1.488095238095238e-07, | |
| "loss": 0.034, | |
| "step": 36850 | |
| }, | |
| { | |
| "epoch": 29.95, | |
| "learning_rate": 8.116883116883118e-08, | |
| "loss": 0.043, | |
| "step": 36900 | |
| }, | |
| { | |
| "epoch": 29.99, | |
| "learning_rate": 1.3528138528138529e-08, | |
| "loss": 0.0336, | |
| "step": 36950 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "eval_cer": 0.2511893892065933, | |
| "eval_loss": 1.9807147979736328, | |
| "eval_mer": 0.29674017257909874, | |
| "eval_runtime": 216.3415, | |
| "eval_samples_per_second": 5.209, | |
| "eval_steps_per_second": 0.652, | |
| "step": 36960 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "step": 36960, | |
| "total_flos": 3.3776644953273782e+19, | |
| "train_loss": 0.25857088244661236, | |
| "train_runtime": 51050.909, | |
| "train_samples_per_second": 5.792, | |
| "train_steps_per_second": 0.724 | |
| } | |
| ], | |
| "max_steps": 36960, | |
| "num_train_epochs": 30, | |
| "total_flos": 3.3776644953273782e+19, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |