| { | |
| "best_metric": 24.71689240755159, | |
| "best_model_checkpoint": "output_dir_run_2/checkpoint-42000", | |
| "epoch": 1.834423353006859, | |
| "global_step": 46000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.4e-07, | |
| "loss": 3.8256, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 9.200000000000001e-07, | |
| "loss": 3.5128, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.42e-06, | |
| "loss": 2.8953, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.9200000000000003e-06, | |
| "loss": 2.3779, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 2.42e-06, | |
| "loss": 2.065, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2.92e-06, | |
| "loss": 1.8663, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.4200000000000007e-06, | |
| "loss": 1.6813, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.920000000000001e-06, | |
| "loss": 1.6136, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.42e-06, | |
| "loss": 1.5066, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.92e-06, | |
| "loss": 1.3971, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 5.420000000000001e-06, | |
| "loss": 1.3437, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 5.92e-06, | |
| "loss": 1.2227, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 6.42e-06, | |
| "loss": 1.2204, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 6.92e-06, | |
| "loss": 1.1546, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 7.420000000000001e-06, | |
| "loss": 1.1521, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 7.92e-06, | |
| "loss": 1.0848, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 8.42e-06, | |
| "loss": 1.0626, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 8.920000000000001e-06, | |
| "loss": 1.0112, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 9.42e-06, | |
| "loss": 0.988, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 9.920000000000002e-06, | |
| "loss": 0.9656, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 9.99957957957958e-06, | |
| "loss": 0.9418, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 9.99907907907908e-06, | |
| "loss": 0.9225, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 9.99857857857858e-06, | |
| "loss": 0.9479, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 9.998078078078079e-06, | |
| "loss": 0.8927, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 9.997577577577578e-06, | |
| "loss": 0.8688, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 9.997077077077078e-06, | |
| "loss": 0.8846, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 9.996576576576577e-06, | |
| "loss": 0.8514, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 9.996076076076076e-06, | |
| "loss": 0.8254, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 9.995575575575577e-06, | |
| "loss": 0.8167, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 9.995075075075076e-06, | |
| "loss": 0.7695, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 9.994574574574576e-06, | |
| "loss": 0.8109, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 9.994074074074075e-06, | |
| "loss": 0.7718, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 9.993573573573574e-06, | |
| "loss": 0.7937, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 9.993073073073074e-06, | |
| "loss": 0.7566, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 9.992572572572573e-06, | |
| "loss": 0.7627, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 9.992072072072074e-06, | |
| "loss": 0.7564, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 9.991571571571572e-06, | |
| "loss": 0.766, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 9.991071071071073e-06, | |
| "loss": 0.7255, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 9.990570570570572e-06, | |
| "loss": 0.7385, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 9.990070070070071e-06, | |
| "loss": 0.7455, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "eval_loss": 0.9031199216842651, | |
| "eval_runtime": 2867.5101, | |
| "eval_samples_per_second": 5.654, | |
| "eval_steps_per_second": 0.142, | |
| "eval_wer": 58.282005226601704, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 9.98956956956957e-06, | |
| "loss": 0.7141, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 9.98906906906907e-06, | |
| "loss": 0.7036, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 9.988568568568569e-06, | |
| "loss": 0.7157, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 9.988068068068068e-06, | |
| "loss": 0.7362, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 9.98756756756757e-06, | |
| "loss": 0.7134, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.987067067067067e-06, | |
| "loss": 0.6704, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.986566566566568e-06, | |
| "loss": 0.6907, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.986066066066067e-06, | |
| "loss": 0.6851, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.985565565565567e-06, | |
| "loss": 0.7214, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.985065065065066e-06, | |
| "loss": 0.6876, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.984564564564565e-06, | |
| "loss": 0.6627, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.984064064064064e-06, | |
| "loss": 0.6739, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.983563563563564e-06, | |
| "loss": 0.6629, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.983063063063065e-06, | |
| "loss": 0.6493, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.982562562562564e-06, | |
| "loss": 0.6512, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 9.982062062062063e-06, | |
| "loss": 0.6481, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 9.981561561561563e-06, | |
| "loss": 0.6459, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 9.981061061061062e-06, | |
| "loss": 0.643, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 9.980560560560561e-06, | |
| "loss": 0.6344, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 9.98006006006006e-06, | |
| "loss": 0.6312, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 9.979559559559561e-06, | |
| "loss": 0.6587, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 9.979059059059059e-06, | |
| "loss": 0.6156, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 9.97855855855856e-06, | |
| "loss": 0.64, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 9.97805805805806e-06, | |
| "loss": 0.5947, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 9.977557557557559e-06, | |
| "loss": 0.6496, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 9.977057057057058e-06, | |
| "loss": 0.6434, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 9.976556556556557e-06, | |
| "loss": 0.5917, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 9.976056056056057e-06, | |
| "loss": 0.6057, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 9.975555555555556e-06, | |
| "loss": 0.6132, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 9.975055055055057e-06, | |
| "loss": 0.5929, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 9.974554554554554e-06, | |
| "loss": 0.6166, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 9.974054054054055e-06, | |
| "loss": 0.5866, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 9.973553553553555e-06, | |
| "loss": 0.6123, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 9.973053053053054e-06, | |
| "loss": 0.594, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 9.972552552552553e-06, | |
| "loss": 0.5949, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 9.972052052052053e-06, | |
| "loss": 0.5862, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 9.971551551551552e-06, | |
| "loss": 0.609, | |
| "step": 1925 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 9.971051051051051e-06, | |
| "loss": 0.5688, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 9.970550550550552e-06, | |
| "loss": 0.592, | |
| "step": 1975 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 9.970050050050052e-06, | |
| "loss": 0.5906, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "eval_loss": 0.7455159425735474, | |
| "eval_runtime": 3054.18, | |
| "eval_samples_per_second": 5.308, | |
| "eval_steps_per_second": 0.133, | |
| "eval_wer": 50.6001341340261, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 9.96954954954955e-06, | |
| "loss": 0.5755, | |
| "step": 2025 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 9.96904904904905e-06, | |
| "loss": 0.5761, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 9.96854854854855e-06, | |
| "loss": 0.5832, | |
| "step": 2075 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 9.968048048048049e-06, | |
| "loss": 0.5661, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 9.967547547547548e-06, | |
| "loss": 0.5906, | |
| "step": 2125 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 9.967047047047049e-06, | |
| "loss": 0.5769, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 9.966546546546547e-06, | |
| "loss": 0.5491, | |
| "step": 2175 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 9.966046046046048e-06, | |
| "loss": 0.5794, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 9.965545545545547e-06, | |
| "loss": 0.5319, | |
| "step": 2225 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 9.965045045045044e-06, | |
| "loss": 0.5656, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 9.964544544544545e-06, | |
| "loss": 0.5714, | |
| "step": 2275 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 9.964044044044045e-06, | |
| "loss": 0.5667, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 9.963543543543544e-06, | |
| "loss": 0.5628, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 9.963043043043043e-06, | |
| "loss": 0.5816, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 9.962542542542544e-06, | |
| "loss": 0.5713, | |
| "step": 2375 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 9.962042042042042e-06, | |
| "loss": 0.5513, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 9.961541541541543e-06, | |
| "loss": 0.5416, | |
| "step": 2425 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 9.961041041041042e-06, | |
| "loss": 0.5273, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 9.960540540540542e-06, | |
| "loss": 0.5414, | |
| "step": 2475 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 9.960040040040041e-06, | |
| "loss": 0.5454, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 9.95953953953954e-06, | |
| "loss": 0.5335, | |
| "step": 2525 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 9.95903903903904e-06, | |
| "loss": 0.538, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 9.958538538538539e-06, | |
| "loss": 0.5275, | |
| "step": 2575 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 9.95803803803804e-06, | |
| "loss": 0.541, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 9.957537537537539e-06, | |
| "loss": 0.5653, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 9.957037037037038e-06, | |
| "loss": 0.5407, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 9.956536536536538e-06, | |
| "loss": 0.5508, | |
| "step": 2675 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 9.956036036036037e-06, | |
| "loss": 0.507, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 9.955535535535536e-06, | |
| "loss": 0.5289, | |
| "step": 2725 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 9.955035035035036e-06, | |
| "loss": 0.543, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 9.954534534534537e-06, | |
| "loss": 0.5489, | |
| "step": 2775 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 9.954034034034034e-06, | |
| "loss": 0.5433, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 9.953533533533535e-06, | |
| "loss": 0.5372, | |
| "step": 2825 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 9.953033033033034e-06, | |
| "loss": 0.5397, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 9.952532532532532e-06, | |
| "loss": 0.5429, | |
| "step": 2875 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 9.952032032032033e-06, | |
| "loss": 0.5178, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 9.951531531531532e-06, | |
| "loss": 0.5426, | |
| "step": 2925 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 9.951031031031032e-06, | |
| "loss": 0.5301, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 9.950530530530531e-06, | |
| "loss": 0.5113, | |
| "step": 2975 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 9.950030030030032e-06, | |
| "loss": 0.5015, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "eval_loss": 0.6700246334075928, | |
| "eval_runtime": 2829.2547, | |
| "eval_samples_per_second": 5.73, | |
| "eval_steps_per_second": 0.144, | |
| "eval_wer": 45.80060283223225, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 9.94952952952953e-06, | |
| "loss": 0.5075, | |
| "step": 3025 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 9.94902902902903e-06, | |
| "loss": 0.548, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 9.94852852852853e-06, | |
| "loss": 0.5221, | |
| "step": 3075 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 9.948028028028029e-06, | |
| "loss": 0.5108, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 9.947527527527528e-06, | |
| "loss": 0.4792, | |
| "step": 3125 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 9.947027027027028e-06, | |
| "loss": 0.4993, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 9.946526526526527e-06, | |
| "loss": 0.5159, | |
| "step": 3175 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 9.946026026026026e-06, | |
| "loss": 0.4948, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 9.945525525525527e-06, | |
| "loss": 0.527, | |
| "step": 3225 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 9.945025025025027e-06, | |
| "loss": 0.4831, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 9.944524524524526e-06, | |
| "loss": 0.5106, | |
| "step": 3275 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 9.944024024024025e-06, | |
| "loss": 0.5323, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 9.943523523523524e-06, | |
| "loss": 0.502, | |
| "step": 3325 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 9.943023023023024e-06, | |
| "loss": 0.495, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 9.942522522522523e-06, | |
| "loss": 0.519, | |
| "step": 3375 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 9.942022022022022e-06, | |
| "loss": 0.5351, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 9.941521521521522e-06, | |
| "loss": 0.4833, | |
| "step": 3425 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 9.941021021021023e-06, | |
| "loss": 0.5227, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 9.940520520520522e-06, | |
| "loss": 0.4969, | |
| "step": 3475 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 9.94002002002002e-06, | |
| "loss": 0.5103, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 9.93951951951952e-06, | |
| "loss": 0.5109, | |
| "step": 3525 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 9.93901901901902e-06, | |
| "loss": 0.5228, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 9.938518518518519e-06, | |
| "loss": 0.4983, | |
| "step": 3575 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 9.938018018018018e-06, | |
| "loss": 0.4873, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 9.93751751751752e-06, | |
| "loss": 0.4955, | |
| "step": 3625 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 9.937017017017017e-06, | |
| "loss": 0.5064, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 9.936516516516518e-06, | |
| "loss": 0.506, | |
| "step": 3675 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 9.936016016016017e-06, | |
| "loss": 0.483, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 9.935515515515517e-06, | |
| "loss": 0.4931, | |
| "step": 3725 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 9.935015015015016e-06, | |
| "loss": 0.4854, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 9.934514514514515e-06, | |
| "loss": 0.4697, | |
| "step": 3775 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 9.934014014014014e-06, | |
| "loss": 0.5078, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 9.933513513513514e-06, | |
| "loss": 0.4759, | |
| "step": 3825 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 9.933013013013015e-06, | |
| "loss": 0.5202, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 9.932512512512514e-06, | |
| "loss": 0.4633, | |
| "step": 3875 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 9.932012012012013e-06, | |
| "loss": 0.4947, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 9.931511511511513e-06, | |
| "loss": 0.4952, | |
| "step": 3925 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 9.931011011011012e-06, | |
| "loss": 0.4768, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 9.930510510510511e-06, | |
| "loss": 0.485, | |
| "step": 3975 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 9.93001001001001e-06, | |
| "loss": 0.4855, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "eval_loss": 0.6402216553688049, | |
| "eval_runtime": 2763.3301, | |
| "eval_samples_per_second": 5.867, | |
| "eval_steps_per_second": 0.147, | |
| "eval_wer": 41.90377810840188, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 9.92950950950951e-06, | |
| "loss": 0.4653, | |
| "step": 4025 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 9.929009009009009e-06, | |
| "loss": 0.4994, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 9.92850850850851e-06, | |
| "loss": 0.4641, | |
| "step": 4075 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 9.92800800800801e-06, | |
| "loss": 0.4851, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 9.927507507507507e-06, | |
| "loss": 0.489, | |
| "step": 4125 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 9.927007007007008e-06, | |
| "loss": 0.4881, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 9.926506506506507e-06, | |
| "loss": 0.4975, | |
| "step": 4175 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 9.926006006006007e-06, | |
| "loss": 0.4785, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 9.925505505505506e-06, | |
| "loss": 0.4689, | |
| "step": 4225 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 9.925005005005007e-06, | |
| "loss": 0.4688, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 9.924504504504505e-06, | |
| "loss": 0.4842, | |
| "step": 4275 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 9.924004004004006e-06, | |
| "loss": 0.462, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 9.923503503503505e-06, | |
| "loss": 0.4674, | |
| "step": 4325 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 9.923003003003004e-06, | |
| "loss": 0.4632, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 9.922502502502503e-06, | |
| "loss": 0.4618, | |
| "step": 4375 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 9.922002002002003e-06, | |
| "loss": 0.4655, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 9.921501501501502e-06, | |
| "loss": 0.4557, | |
| "step": 4425 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 9.921001001001001e-06, | |
| "loss": 0.4549, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 9.920500500500502e-06, | |
| "loss": 0.4757, | |
| "step": 4475 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 9.920000000000002e-06, | |
| "loss": 0.4501, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 9.919499499499501e-06, | |
| "loss": 0.4789, | |
| "step": 4525 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 9.918998998999e-06, | |
| "loss": 0.4768, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 9.9184984984985e-06, | |
| "loss": 0.4799, | |
| "step": 4575 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 9.917997997997999e-06, | |
| "loss": 0.4648, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 9.917497497497498e-06, | |
| "loss": 0.4671, | |
| "step": 4625 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 9.916996996996997e-06, | |
| "loss": 0.4672, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 9.916496496496497e-06, | |
| "loss": 0.4664, | |
| "step": 4675 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 9.915995995995998e-06, | |
| "loss": 0.4769, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 9.915495495495497e-06, | |
| "loss": 0.4828, | |
| "step": 4725 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 9.914994994994995e-06, | |
| "loss": 0.4648, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 9.914494494494496e-06, | |
| "loss": 0.4501, | |
| "step": 4775 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 9.913993993993995e-06, | |
| "loss": 0.4638, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 9.913493493493494e-06, | |
| "loss": 0.463, | |
| "step": 4825 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 9.912992992992993e-06, | |
| "loss": 0.4759, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 9.912492492492494e-06, | |
| "loss": 0.4967, | |
| "step": 4875 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.911991991991992e-06, | |
| "loss": 0.4483, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.911491491491493e-06, | |
| "loss": 0.4514, | |
| "step": 4925 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.910990990990992e-06, | |
| "loss": 0.482, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.910490490490492e-06, | |
| "loss": 0.4439, | |
| "step": 4975 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.909989989989991e-06, | |
| "loss": 0.4379, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "eval_loss": 0.6058462858200073, | |
| "eval_runtime": 2619.0085, | |
| "eval_samples_per_second": 6.191, | |
| "eval_steps_per_second": 0.155, | |
| "eval_wer": 38.43171113389505, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.90948948948949e-06, | |
| "loss": 0.4567, | |
| "step": 5025 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.90898898898899e-06, | |
| "loss": 0.4506, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.908488488488489e-06, | |
| "loss": 0.4717, | |
| "step": 5075 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.90798798798799e-06, | |
| "loss": 0.4775, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.907487487487487e-06, | |
| "loss": 0.4344, | |
| "step": 5125 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 9.906986986986987e-06, | |
| "loss": 0.4462, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 9.906486486486488e-06, | |
| "loss": 0.4492, | |
| "step": 5175 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 9.905985985985987e-06, | |
| "loss": 0.4532, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 9.905485485485486e-06, | |
| "loss": 0.4595, | |
| "step": 5225 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 9.904984984984986e-06, | |
| "loss": 0.4369, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 9.904484484484485e-06, | |
| "loss": 0.4668, | |
| "step": 5275 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 9.903983983983984e-06, | |
| "loss": 0.4507, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 9.903483483483485e-06, | |
| "loss": 0.4438, | |
| "step": 5325 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 9.902982982982984e-06, | |
| "loss": 0.4679, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 9.902482482482482e-06, | |
| "loss": 0.468, | |
| "step": 5375 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 9.901981981981983e-06, | |
| "loss": 0.4502, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 9.901481481481482e-06, | |
| "loss": 0.4319, | |
| "step": 5425 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 9.900980980980982e-06, | |
| "loss": 0.4799, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 9.900480480480481e-06, | |
| "loss": 0.4553, | |
| "step": 5475 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 9.899979979979982e-06, | |
| "loss": 0.4309, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 9.89947947947948e-06, | |
| "loss": 0.4385, | |
| "step": 5525 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 9.89897897897898e-06, | |
| "loss": 0.431, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 9.89847847847848e-06, | |
| "loss": 0.4478, | |
| "step": 5575 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 9.897977977977979e-06, | |
| "loss": 0.4379, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 9.897477477477478e-06, | |
| "loss": 0.4334, | |
| "step": 5625 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 9.896976976976978e-06, | |
| "loss": 0.4542, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 9.896476476476477e-06, | |
| "loss": 0.4454, | |
| "step": 5675 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 9.895975975975976e-06, | |
| "loss": 0.4519, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 9.895475475475477e-06, | |
| "loss": 0.4625, | |
| "step": 5725 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 9.894974974974975e-06, | |
| "loss": 0.4361, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 9.894474474474474e-06, | |
| "loss": 0.4305, | |
| "step": 5775 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 9.893973973973975e-06, | |
| "loss": 0.4558, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 9.893473473473475e-06, | |
| "loss": 0.4554, | |
| "step": 5825 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 9.892972972972974e-06, | |
| "loss": 0.4562, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 9.892472472472473e-06, | |
| "loss": 0.4378, | |
| "step": 5875 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 9.891971971971972e-06, | |
| "loss": 0.464, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 9.891471471471472e-06, | |
| "loss": 0.4279, | |
| "step": 5925 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 9.890970970970973e-06, | |
| "loss": 0.4414, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 9.890470470470472e-06, | |
| "loss": 0.4537, | |
| "step": 5975 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 9.88996996996997e-06, | |
| "loss": 0.4399, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "eval_loss": 0.5941766500473022, | |
| "eval_runtime": 2644.5549, | |
| "eval_samples_per_second": 6.131, | |
| "eval_steps_per_second": 0.154, | |
| "eval_wer": 37.588362716907824, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 9.88946946946947e-06, | |
| "loss": 0.4419, | |
| "step": 6025 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 9.88896896896897e-06, | |
| "loss": 0.4219, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 9.88846846846847e-06, | |
| "loss": 0.4271, | |
| "step": 6075 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 9.887967967967968e-06, | |
| "loss": 0.4565, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 9.88746746746747e-06, | |
| "loss": 0.4588, | |
| "step": 6125 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 9.886966966966967e-06, | |
| "loss": 0.4384, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 9.886466466466468e-06, | |
| "loss": 0.4341, | |
| "step": 6175 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 9.885965965965967e-06, | |
| "loss": 0.4302, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 9.885465465465467e-06, | |
| "loss": 0.4307, | |
| "step": 6225 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 9.884964964964966e-06, | |
| "loss": 0.4634, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 9.884464464464465e-06, | |
| "loss": 0.4374, | |
| "step": 6275 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 9.883963963963965e-06, | |
| "loss": 0.4352, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 9.883463463463464e-06, | |
| "loss": 0.4376, | |
| "step": 6325 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 9.882962962962965e-06, | |
| "loss": 0.4284, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 9.882462462462462e-06, | |
| "loss": 0.4143, | |
| "step": 6375 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.881961961961962e-06, | |
| "loss": 0.4286, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.881461461461463e-06, | |
| "loss": 0.4358, | |
| "step": 6425 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.880960960960962e-06, | |
| "loss": 0.4306, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.880460460460461e-06, | |
| "loss": 0.4357, | |
| "step": 6475 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.87995995995996e-06, | |
| "loss": 0.4344, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.87945945945946e-06, | |
| "loss": 0.4508, | |
| "step": 6525 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.87895895895896e-06, | |
| "loss": 0.441, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.87845845845846e-06, | |
| "loss": 0.419, | |
| "step": 6575 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.87795795795796e-06, | |
| "loss": 0.4221, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.877457457457457e-06, | |
| "loss": 0.4288, | |
| "step": 6625 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 9.876956956956958e-06, | |
| "loss": 0.4288, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 9.876456456456457e-06, | |
| "loss": 0.4314, | |
| "step": 6675 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 9.875955955955957e-06, | |
| "loss": 0.4138, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 9.875455455455456e-06, | |
| "loss": 0.4291, | |
| "step": 6725 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 9.874954954954957e-06, | |
| "loss": 0.4446, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 9.874454454454455e-06, | |
| "loss": 0.4153, | |
| "step": 6775 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 9.873953953953956e-06, | |
| "loss": 0.4325, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 9.873453453453455e-06, | |
| "loss": 0.4158, | |
| "step": 6825 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 9.872952952952952e-06, | |
| "loss": 0.4248, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 9.872452452452453e-06, | |
| "loss": 0.4221, | |
| "step": 6875 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 9.871951951951953e-06, | |
| "loss": 0.4206, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 9.871451451451452e-06, | |
| "loss": 0.4173, | |
| "step": 6925 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 9.870950950950951e-06, | |
| "loss": 0.4285, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 9.870450450450452e-06, | |
| "loss": 0.4102, | |
| "step": 6975 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 9.86994994994995e-06, | |
| "loss": 0.4335, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "eval_loss": 0.5607422590255737, | |
| "eval_runtime": 2701.8903, | |
| "eval_samples_per_second": 6.001, | |
| "eval_steps_per_second": 0.15, | |
| "eval_wer": 36.83443698398872, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 9.86944944944945e-06, | |
| "loss": 0.4267, | |
| "step": 7025 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 9.86894894894895e-06, | |
| "loss": 0.439, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 9.86844844844845e-06, | |
| "loss": 0.4139, | |
| "step": 7075 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 9.867947947947949e-06, | |
| "loss": 0.4248, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 9.867447447447448e-06, | |
| "loss": 0.4279, | |
| "step": 7125 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 9.866946946946947e-06, | |
| "loss": 0.4172, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 9.866446446446447e-06, | |
| "loss": 0.4382, | |
| "step": 7175 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 9.865945945945948e-06, | |
| "loss": 0.4367, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 9.865445445445447e-06, | |
| "loss": 0.4247, | |
| "step": 7225 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 9.864944944944945e-06, | |
| "loss": 0.4223, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 9.864444444444446e-06, | |
| "loss": 0.4207, | |
| "step": 7275 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 9.863943943943945e-06, | |
| "loss": 0.4166, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 9.863443443443444e-06, | |
| "loss": 0.4155, | |
| "step": 7325 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 9.862942942942944e-06, | |
| "loss": 0.4131, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 9.862442442442443e-06, | |
| "loss": 0.4, | |
| "step": 7375 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 9.861941941941942e-06, | |
| "loss": 0.4051, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 9.861441441441443e-06, | |
| "loss": 0.3976, | |
| "step": 7425 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 9.860940940940942e-06, | |
| "loss": 0.4396, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 9.86044044044044e-06, | |
| "loss": 0.3999, | |
| "step": 7475 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 9.859939939939941e-06, | |
| "loss": 0.4167, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 9.85943943943944e-06, | |
| "loss": 0.4324, | |
| "step": 7525 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 9.85893893893894e-06, | |
| "loss": 0.4185, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 9.858438438438439e-06, | |
| "loss": 0.4205, | |
| "step": 7575 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 9.85793793793794e-06, | |
| "loss": 0.4011, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 9.857437437437437e-06, | |
| "loss": 0.4032, | |
| "step": 7625 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 9.856936936936937e-06, | |
| "loss": 0.4223, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 9.856436436436438e-06, | |
| "loss": 0.4024, | |
| "step": 7675 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 9.855935935935937e-06, | |
| "loss": 0.4071, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 9.855435435435436e-06, | |
| "loss": 0.4024, | |
| "step": 7725 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 9.854934934934936e-06, | |
| "loss": 0.3904, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 9.854434434434435e-06, | |
| "loss": 0.4041, | |
| "step": 7775 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 9.853933933933934e-06, | |
| "loss": 0.4201, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 9.853433433433435e-06, | |
| "loss": 0.393, | |
| "step": 7825 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 9.852932932932935e-06, | |
| "loss": 0.4405, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 9.852432432432432e-06, | |
| "loss": 0.402, | |
| "step": 7875 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 9.851931931931933e-06, | |
| "loss": 0.4064, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 9.851431431431432e-06, | |
| "loss": 0.3908, | |
| "step": 7925 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 9.850930930930932e-06, | |
| "loss": 0.4232, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 9.850430430430431e-06, | |
| "loss": 0.4311, | |
| "step": 7975 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 9.84992992992993e-06, | |
| "loss": 0.4266, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "eval_loss": 0.5630306601524353, | |
| "eval_runtime": 2643.0014, | |
| "eval_samples_per_second": 6.134, | |
| "eval_steps_per_second": 0.154, | |
| "eval_wer": 34.73531656401045, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 9.84942942942943e-06, | |
| "loss": 0.4117, | |
| "step": 8025 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 9.84894894894895e-06, | |
| "loss": 0.3983, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 9.848448448448448e-06, | |
| "loss": 0.4025, | |
| "step": 8075 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 9.847947947947949e-06, | |
| "loss": 0.4119, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 9.847447447447448e-06, | |
| "loss": 0.4135, | |
| "step": 8125 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 9.846946946946948e-06, | |
| "loss": 0.4048, | |
| "step": 8150 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 9.846446446446447e-06, | |
| "loss": 0.4371, | |
| "step": 8175 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 9.845945945945948e-06, | |
| "loss": 0.3954, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 9.845445445445446e-06, | |
| "loss": 0.3933, | |
| "step": 8225 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 9.844944944944945e-06, | |
| "loss": 0.3956, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 9.844444444444446e-06, | |
| "loss": 0.4109, | |
| "step": 8275 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 9.843943943943945e-06, | |
| "loss": 0.4087, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 9.843443443443444e-06, | |
| "loss": 0.3904, | |
| "step": 8325 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 9.842942942942944e-06, | |
| "loss": 0.4309, | |
| "step": 8350 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 9.842442442442443e-06, | |
| "loss": 0.411, | |
| "step": 8375 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 9.841941941941942e-06, | |
| "loss": 0.384, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 9.841441441441443e-06, | |
| "loss": 0.3993, | |
| "step": 8425 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 9.840940940940943e-06, | |
| "loss": 0.3931, | |
| "step": 8450 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 9.84044044044044e-06, | |
| "loss": 0.4044, | |
| "step": 8475 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 9.839939939939941e-06, | |
| "loss": 0.3985, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 9.83943943943944e-06, | |
| "loss": 0.4317, | |
| "step": 8525 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 9.83893893893894e-06, | |
| "loss": 0.3932, | |
| "step": 8550 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 9.838438438438439e-06, | |
| "loss": 0.3925, | |
| "step": 8575 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 9.837937937937938e-06, | |
| "loss": 0.3799, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 9.837437437437438e-06, | |
| "loss": 0.3975, | |
| "step": 8625 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 9.836936936936937e-06, | |
| "loss": 0.4047, | |
| "step": 8650 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 9.836436436436438e-06, | |
| "loss": 0.4092, | |
| "step": 8675 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 9.835935935935936e-06, | |
| "loss": 0.3945, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 9.835435435435437e-06, | |
| "loss": 0.4068, | |
| "step": 8725 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 9.834934934934936e-06, | |
| "loss": 0.3877, | |
| "step": 8750 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 9.834434434434435e-06, | |
| "loss": 0.3949, | |
| "step": 8775 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 9.833933933933934e-06, | |
| "loss": 0.3841, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 9.833433433433435e-06, | |
| "loss": 0.4074, | |
| "step": 8825 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 9.832932932932933e-06, | |
| "loss": 0.4171, | |
| "step": 8850 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 9.832452452452454e-06, | |
| "loss": 0.3953, | |
| "step": 8875 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 9.831951951951953e-06, | |
| "loss": 0.3761, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 9.831451451451452e-06, | |
| "loss": 0.4115, | |
| "step": 8925 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 9.830950950950952e-06, | |
| "loss": 0.3979, | |
| "step": 8950 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 9.830450450450451e-06, | |
| "loss": 0.405, | |
| "step": 8975 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 9.82994994994995e-06, | |
| "loss": 0.4057, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "eval_loss": 0.5405702590942383, | |
| "eval_runtime": 2698.8083, | |
| "eval_samples_per_second": 6.007, | |
| "eval_steps_per_second": 0.15, | |
| "eval_wer": 34.710648237370975, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 9.829449449449451e-06, | |
| "loss": 0.3833, | |
| "step": 9025 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 9.82894894894895e-06, | |
| "loss": 0.4058, | |
| "step": 9050 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 9.828448448448448e-06, | |
| "loss": 0.3861, | |
| "step": 9075 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 9.82794794794795e-06, | |
| "loss": 0.4031, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 9.827447447447449e-06, | |
| "loss": 0.3894, | |
| "step": 9125 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 9.826946946946948e-06, | |
| "loss": 0.3892, | |
| "step": 9150 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 9.826446446446447e-06, | |
| "loss": 0.3991, | |
| "step": 9175 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 9.825945945945946e-06, | |
| "loss": 0.3803, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 9.825445445445446e-06, | |
| "loss": 0.3927, | |
| "step": 9225 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 9.824944944944945e-06, | |
| "loss": 0.4084, | |
| "step": 9250 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 9.824444444444446e-06, | |
| "loss": 0.3916, | |
| "step": 9275 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 9.823943943943944e-06, | |
| "loss": 0.3674, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 9.823443443443445e-06, | |
| "loss": 0.3683, | |
| "step": 9325 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 9.822942942942944e-06, | |
| "loss": 0.4016, | |
| "step": 9350 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 9.822442442442443e-06, | |
| "loss": 0.395, | |
| "step": 9375 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 9.821941941941942e-06, | |
| "loss": 0.3906, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 9.821441441441443e-06, | |
| "loss": 0.3843, | |
| "step": 9425 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 9.820940940940941e-06, | |
| "loss": 0.3732, | |
| "step": 9450 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 9.82044044044044e-06, | |
| "loss": 0.3927, | |
| "step": 9475 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 9.819939939939941e-06, | |
| "loss": 0.4151, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 9.81943943943944e-06, | |
| "loss": 0.3933, | |
| "step": 9525 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 9.81893893893894e-06, | |
| "loss": 0.3856, | |
| "step": 9550 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 9.81843843843844e-06, | |
| "loss": 0.3818, | |
| "step": 9575 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 9.817937937937939e-06, | |
| "loss": 0.3932, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 9.817437437437438e-06, | |
| "loss": 0.4085, | |
| "step": 9625 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 9.816936936936939e-06, | |
| "loss": 0.4023, | |
| "step": 9650 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.816436436436436e-06, | |
| "loss": 0.3918, | |
| "step": 9675 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.815935935935936e-06, | |
| "loss": 0.3965, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.815435435435437e-06, | |
| "loss": 0.3934, | |
| "step": 9725 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.814934934934936e-06, | |
| "loss": 0.3807, | |
| "step": 9750 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.814434434434435e-06, | |
| "loss": 0.3809, | |
| "step": 9775 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.813933933933935e-06, | |
| "loss": 0.3804, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.813433433433434e-06, | |
| "loss": 0.4112, | |
| "step": 9825 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.812932932932933e-06, | |
| "loss": 0.3853, | |
| "step": 9850 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.812432432432433e-06, | |
| "loss": 0.3967, | |
| "step": 9875 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.811931931931934e-06, | |
| "loss": 0.3934, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 9.811431431431431e-06, | |
| "loss": 0.4013, | |
| "step": 9925 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 9.810930930930932e-06, | |
| "loss": 0.4004, | |
| "step": 9950 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 9.810430430430431e-06, | |
| "loss": 0.3825, | |
| "step": 9975 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 9.80992992992993e-06, | |
| "loss": 0.3798, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "eval_loss": 0.5242385864257812, | |
| "eval_runtime": 2613.1802, | |
| "eval_samples_per_second": 6.204, | |
| "eval_steps_per_second": 0.155, | |
| "eval_wer": 32.9584261607604, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 9.80942942942943e-06, | |
| "loss": 0.379, | |
| "step": 10025 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 9.808928928928931e-06, | |
| "loss": 0.397, | |
| "step": 10050 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 9.808428428428429e-06, | |
| "loss": 0.3818, | |
| "step": 10075 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 9.807927927927928e-06, | |
| "loss": 0.373, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 9.807427427427429e-06, | |
| "loss": 0.3592, | |
| "step": 10125 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 9.806926926926928e-06, | |
| "loss": 0.3824, | |
| "step": 10150 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 9.806426426426427e-06, | |
| "loss": 0.3907, | |
| "step": 10175 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 9.805925925925927e-06, | |
| "loss": 0.3787, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 9.805425425425426e-06, | |
| "loss": 0.3571, | |
| "step": 10225 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 9.804924924924925e-06, | |
| "loss": 0.3785, | |
| "step": 10250 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 9.804424424424426e-06, | |
| "loss": 0.41, | |
| "step": 10275 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 9.803923923923924e-06, | |
| "loss": 0.3703, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 9.803423423423423e-06, | |
| "loss": 0.38, | |
| "step": 10325 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 9.802922922922924e-06, | |
| "loss": 0.3787, | |
| "step": 10350 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 9.802422422422424e-06, | |
| "loss": 0.3948, | |
| "step": 10375 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 9.801921921921923e-06, | |
| "loss": 0.3775, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 9.801421421421422e-06, | |
| "loss": 0.3823, | |
| "step": 10425 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 9.800920920920921e-06, | |
| "loss": 0.3868, | |
| "step": 10450 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 9.80042042042042e-06, | |
| "loss": 0.378, | |
| "step": 10475 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 9.79991991991992e-06, | |
| "loss": 0.3911, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 9.799419419419421e-06, | |
| "loss": 0.3845, | |
| "step": 10525 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 9.798918918918919e-06, | |
| "loss": 0.3899, | |
| "step": 10550 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 9.79841841841842e-06, | |
| "loss": 0.3919, | |
| "step": 10575 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 9.797917917917919e-06, | |
| "loss": 0.4163, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 9.797417417417418e-06, | |
| "loss": 0.382, | |
| "step": 10625 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 9.796916916916918e-06, | |
| "loss": 0.3796, | |
| "step": 10650 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 9.796416416416419e-06, | |
| "loss": 0.3751, | |
| "step": 10675 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 9.795915915915916e-06, | |
| "loss": 0.3768, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 9.795415415415415e-06, | |
| "loss": 0.3878, | |
| "step": 10725 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 9.794914914914916e-06, | |
| "loss": 0.3821, | |
| "step": 10750 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 9.794414414414416e-06, | |
| "loss": 0.3879, | |
| "step": 10775 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 9.793913913913915e-06, | |
| "loss": 0.3685, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 9.793413413413414e-06, | |
| "loss": 0.3673, | |
| "step": 10825 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 9.792912912912914e-06, | |
| "loss": 0.3955, | |
| "step": 10850 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 9.792412412412413e-06, | |
| "loss": 0.3616, | |
| "step": 10875 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 9.791911911911914e-06, | |
| "loss": 0.3692, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 9.791411411411411e-06, | |
| "loss": 0.3766, | |
| "step": 10925 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 9.79091091091091e-06, | |
| "loss": 0.3873, | |
| "step": 10950 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 9.790410410410412e-06, | |
| "loss": 0.3792, | |
| "step": 10975 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 9.789909909909911e-06, | |
| "loss": 0.3781, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "eval_loss": 0.5272489190101624, | |
| "eval_runtime": 2804.5214, | |
| "eval_samples_per_second": 5.781, | |
| "eval_steps_per_second": 0.145, | |
| "eval_wer": 34.24349180163582, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 9.78940940940941e-06, | |
| "loss": 0.3926, | |
| "step": 11025 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 9.78890890890891e-06, | |
| "loss": 0.3849, | |
| "step": 11050 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 9.788408408408409e-06, | |
| "loss": 0.3726, | |
| "step": 11075 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 9.787907907907908e-06, | |
| "loss": 0.3798, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 9.787407407407408e-06, | |
| "loss": 0.3877, | |
| "step": 11125 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 9.786906906906909e-06, | |
| "loss": 0.3873, | |
| "step": 11150 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 9.786426426426428e-06, | |
| "loss": 0.381, | |
| "step": 11175 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 9.785925925925927e-06, | |
| "loss": 0.3769, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 9.785425425425426e-06, | |
| "loss": 0.3903, | |
| "step": 11225 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 9.784924924924926e-06, | |
| "loss": 0.3778, | |
| "step": 11250 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 9.784424424424427e-06, | |
| "loss": 0.3966, | |
| "step": 11275 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 9.783923923923924e-06, | |
| "loss": 0.3851, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 9.783423423423423e-06, | |
| "loss": 0.3737, | |
| "step": 11325 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 9.782922922922924e-06, | |
| "loss": 0.3602, | |
| "step": 11350 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 9.782422422422424e-06, | |
| "loss": 0.3722, | |
| "step": 11375 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 9.781921921921923e-06, | |
| "loss": 0.3874, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 9.781421421421422e-06, | |
| "loss": 0.3895, | |
| "step": 11425 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 9.780920920920922e-06, | |
| "loss": 0.375, | |
| "step": 11450 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 9.780420420420421e-06, | |
| "loss": 0.3858, | |
| "step": 11475 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 9.77991991991992e-06, | |
| "loss": 0.3877, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 9.77941941941942e-06, | |
| "loss": 0.3932, | |
| "step": 11525 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 9.778918918918919e-06, | |
| "loss": 0.3792, | |
| "step": 11550 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 9.77841841841842e-06, | |
| "loss": 0.3709, | |
| "step": 11575 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 9.777917917917919e-06, | |
| "loss": 0.3815, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 9.777417417417418e-06, | |
| "loss": 0.3834, | |
| "step": 11625 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 9.776916916916918e-06, | |
| "loss": 0.3598, | |
| "step": 11650 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 9.776416416416417e-06, | |
| "loss": 0.3648, | |
| "step": 11675 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 9.775915915915916e-06, | |
| "loss": 0.3582, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 9.775415415415416e-06, | |
| "loss": 0.3753, | |
| "step": 11725 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 9.774914914914917e-06, | |
| "loss": 0.3689, | |
| "step": 11750 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 9.774414414414414e-06, | |
| "loss": 0.3774, | |
| "step": 11775 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 9.773913913913915e-06, | |
| "loss": 0.3825, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 9.773413413413414e-06, | |
| "loss": 0.36, | |
| "step": 11825 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 9.772912912912914e-06, | |
| "loss": 0.3672, | |
| "step": 11850 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 9.772412412412413e-06, | |
| "loss": 0.3848, | |
| "step": 11875 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 9.771911911911914e-06, | |
| "loss": 0.357, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 9.771411411411412e-06, | |
| "loss": 0.3658, | |
| "step": 11925 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 9.770910910910911e-06, | |
| "loss": 0.3733, | |
| "step": 11950 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 9.770410410410412e-06, | |
| "loss": 0.3731, | |
| "step": 11975 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 9.76990990990991e-06, | |
| "loss": 0.3674, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "eval_loss": 0.5061578750610352, | |
| "eval_runtime": 2740.8866, | |
| "eval_samples_per_second": 5.915, | |
| "eval_steps_per_second": 0.148, | |
| "eval_wer": 33.014700780906715, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 9.76940940940941e-06, | |
| "loss": 0.3981, | |
| "step": 12025 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 9.76890890890891e-06, | |
| "loss": 0.3759, | |
| "step": 12050 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 9.76840840840841e-06, | |
| "loss": 0.3647, | |
| "step": 12075 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 9.767907907907908e-06, | |
| "loss": 0.3756, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 9.767407407407408e-06, | |
| "loss": 0.3444, | |
| "step": 12125 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 9.766906906906907e-06, | |
| "loss": 0.3536, | |
| "step": 12150 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 9.766406406406406e-06, | |
| "loss": 0.3584, | |
| "step": 12175 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 9.765905905905907e-06, | |
| "loss": 0.3879, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 9.765405405405407e-06, | |
| "loss": 0.3403, | |
| "step": 12225 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 9.764904904904906e-06, | |
| "loss": 0.3646, | |
| "step": 12250 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 9.764404404404405e-06, | |
| "loss": 0.3697, | |
| "step": 12275 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 9.763903903903905e-06, | |
| "loss": 0.3607, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 9.763403403403404e-06, | |
| "loss": 0.3588, | |
| "step": 12325 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 9.762902902902903e-06, | |
| "loss": 0.3668, | |
| "step": 12350 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 9.762402402402404e-06, | |
| "loss": 0.3933, | |
| "step": 12375 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 9.761901901901902e-06, | |
| "loss": 0.3744, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 9.761401401401403e-06, | |
| "loss": 0.3905, | |
| "step": 12425 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 9.760900900900902e-06, | |
| "loss": 0.3641, | |
| "step": 12450 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 9.760400400400401e-06, | |
| "loss": 0.3647, | |
| "step": 12475 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 9.7598998998999e-06, | |
| "loss": 0.35, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 9.7593993993994e-06, | |
| "loss": 0.3545, | |
| "step": 12525 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 9.7588988988989e-06, | |
| "loss": 0.3645, | |
| "step": 12550 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 9.758398398398398e-06, | |
| "loss": 0.3625, | |
| "step": 12575 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 9.7578978978979e-06, | |
| "loss": 0.3801, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 9.757397397397397e-06, | |
| "loss": 0.3562, | |
| "step": 12625 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 9.756896896896898e-06, | |
| "loss": 0.3594, | |
| "step": 12650 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 9.756396396396397e-06, | |
| "loss": 0.3644, | |
| "step": 12675 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 9.755895895895897e-06, | |
| "loss": 0.356, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 9.755395395395396e-06, | |
| "loss": 0.3676, | |
| "step": 12725 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 9.754894894894895e-06, | |
| "loss": 0.3722, | |
| "step": 12750 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 9.754394394394395e-06, | |
| "loss": 0.396, | |
| "step": 12775 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 9.753893893893894e-06, | |
| "loss": 0.3613, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 9.753393393393395e-06, | |
| "loss": 0.3626, | |
| "step": 12825 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 9.752892892892894e-06, | |
| "loss": 0.3642, | |
| "step": 12850 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 9.752392392392393e-06, | |
| "loss": 0.3674, | |
| "step": 12875 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 9.751891891891893e-06, | |
| "loss": 0.3526, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.751391391391392e-06, | |
| "loss": 0.3767, | |
| "step": 12925 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.750890890890891e-06, | |
| "loss": 0.3616, | |
| "step": 12950 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.75039039039039e-06, | |
| "loss": 0.3475, | |
| "step": 12975 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.749889889889892e-06, | |
| "loss": 0.3507, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "eval_loss": 0.49944427609443665, | |
| "eval_runtime": 2897.6163, | |
| "eval_samples_per_second": 5.595, | |
| "eval_steps_per_second": 0.14, | |
| "eval_wer": 33.72160251616932, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.74938938938939e-06, | |
| "loss": 0.3743, | |
| "step": 13025 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.74888888888889e-06, | |
| "loss": 0.3597, | |
| "step": 13050 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.74838838838839e-06, | |
| "loss": 0.3862, | |
| "step": 13075 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.747887887887889e-06, | |
| "loss": 0.3606, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.747387387387388e-06, | |
| "loss": 0.4099, | |
| "step": 13125 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.746886886886887e-06, | |
| "loss": 0.3583, | |
| "step": 13150 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.746386386386387e-06, | |
| "loss": 0.3673, | |
| "step": 13175 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.745885885885886e-06, | |
| "loss": 0.3597, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.745385385385387e-06, | |
| "loss": 0.3715, | |
| "step": 13225 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.744884884884885e-06, | |
| "loss": 0.3681, | |
| "step": 13250 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.744384384384386e-06, | |
| "loss": 0.3793, | |
| "step": 13275 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.743883883883885e-06, | |
| "loss": 0.3574, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.743383383383384e-06, | |
| "loss": 0.3602, | |
| "step": 13325 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.742882882882883e-06, | |
| "loss": 0.3663, | |
| "step": 13350 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.742382382382383e-06, | |
| "loss": 0.3947, | |
| "step": 13375 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.741881881881882e-06, | |
| "loss": 0.3709, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.741381381381381e-06, | |
| "loss": 0.3625, | |
| "step": 13425 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.740880880880882e-06, | |
| "loss": 0.3574, | |
| "step": 13450 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.740380380380382e-06, | |
| "loss": 0.3642, | |
| "step": 13475 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.739879879879881e-06, | |
| "loss": 0.3639, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.73937937937938e-06, | |
| "loss": 0.3669, | |
| "step": 13525 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.73887887887888e-06, | |
| "loss": 0.3762, | |
| "step": 13550 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.738378378378379e-06, | |
| "loss": 0.3564, | |
| "step": 13575 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.737877877877878e-06, | |
| "loss": 0.3602, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.737377377377379e-06, | |
| "loss": 0.3464, | |
| "step": 13625 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.736876876876877e-06, | |
| "loss": 0.3618, | |
| "step": 13650 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 9.736376376376378e-06, | |
| "loss": 0.3761, | |
| "step": 13675 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 9.735875875875877e-06, | |
| "loss": 0.3604, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 9.735375375375376e-06, | |
| "loss": 0.3714, | |
| "step": 13725 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 9.734874874874876e-06, | |
| "loss": 0.3699, | |
| "step": 13750 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 9.734374374374375e-06, | |
| "loss": 0.3728, | |
| "step": 13775 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 9.733873873873874e-06, | |
| "loss": 0.3878, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 9.733373373373374e-06, | |
| "loss": 0.3461, | |
| "step": 13825 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 9.732872872872875e-06, | |
| "loss": 0.3415, | |
| "step": 13850 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 9.732372372372372e-06, | |
| "loss": 0.3548, | |
| "step": 13875 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 9.731871871871873e-06, | |
| "loss": 0.3729, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.731371371371372e-06, | |
| "loss": 0.3705, | |
| "step": 13925 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.730870870870872e-06, | |
| "loss": 0.3605, | |
| "step": 13950 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.730370370370371e-06, | |
| "loss": 0.3631, | |
| "step": 13975 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.72986986986987e-06, | |
| "loss": 0.375, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "eval_loss": 0.5072089433670044, | |
| "eval_runtime": 2512.5863, | |
| "eval_samples_per_second": 6.453, | |
| "eval_steps_per_second": 0.162, | |
| "eval_wer": 30.171676135706633, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.72936936936937e-06, | |
| "loss": 0.3642, | |
| "step": 14025 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.728868868868869e-06, | |
| "loss": 0.348, | |
| "step": 14050 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.72836836836837e-06, | |
| "loss": 0.3413, | |
| "step": 14075 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.72786786786787e-06, | |
| "loss": 0.3528, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.727367367367368e-06, | |
| "loss": 0.3501, | |
| "step": 14125 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.726866866866868e-06, | |
| "loss": 0.3318, | |
| "step": 14150 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.726366366366367e-06, | |
| "loss": 0.3655, | |
| "step": 14175 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.725865865865866e-06, | |
| "loss": 0.3496, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.725365365365366e-06, | |
| "loss": 0.3379, | |
| "step": 14225 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.724884884884885e-06, | |
| "loss": 0.3893, | |
| "step": 14250 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.724384384384386e-06, | |
| "loss": 0.3703, | |
| "step": 14275 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.723883883883885e-06, | |
| "loss": 0.3531, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.723383383383384e-06, | |
| "loss": 0.3604, | |
| "step": 14325 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.722882882882884e-06, | |
| "loss": 0.3527, | |
| "step": 14350 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.722382382382383e-06, | |
| "loss": 0.3556, | |
| "step": 14375 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.721881881881882e-06, | |
| "loss": 0.3664, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 9.721381381381382e-06, | |
| "loss": 0.3641, | |
| "step": 14425 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 9.720880880880883e-06, | |
| "loss": 0.3553, | |
| "step": 14450 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 9.72038038038038e-06, | |
| "loss": 0.3605, | |
| "step": 14475 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 9.719879879879881e-06, | |
| "loss": 0.3412, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 9.71937937937938e-06, | |
| "loss": 0.3588, | |
| "step": 14525 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 9.71887887887888e-06, | |
| "loss": 0.3334, | |
| "step": 14550 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 9.718378378378379e-06, | |
| "loss": 0.3624, | |
| "step": 14575 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 9.717877877877878e-06, | |
| "loss": 0.3609, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 9.717377377377378e-06, | |
| "loss": 0.3339, | |
| "step": 14625 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 9.716876876876877e-06, | |
| "loss": 0.353, | |
| "step": 14650 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 9.716376376376378e-06, | |
| "loss": 0.3517, | |
| "step": 14675 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 9.715875875875877e-06, | |
| "loss": 0.3626, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 9.715375375375377e-06, | |
| "loss": 0.3268, | |
| "step": 14725 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 9.714874874874876e-06, | |
| "loss": 0.3648, | |
| "step": 14750 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 9.714374374374375e-06, | |
| "loss": 0.3579, | |
| "step": 14775 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 9.713873873873874e-06, | |
| "loss": 0.3403, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 9.713373373373374e-06, | |
| "loss": 0.3597, | |
| "step": 14825 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 9.712872872872873e-06, | |
| "loss": 0.3323, | |
| "step": 14850 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 9.712372372372372e-06, | |
| "loss": 0.3412, | |
| "step": 14875 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 9.711871871871873e-06, | |
| "loss": 0.3574, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 9.711371371371373e-06, | |
| "loss": 0.348, | |
| "step": 14925 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 9.710870870870872e-06, | |
| "loss": 0.3588, | |
| "step": 14950 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 9.710370370370371e-06, | |
| "loss": 0.3445, | |
| "step": 14975 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 9.70986986986987e-06, | |
| "loss": 0.3545, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "eval_loss": 0.5023101568222046, | |
| "eval_runtime": 2571.7112, | |
| "eval_samples_per_second": 6.304, | |
| "eval_steps_per_second": 0.158, | |
| "eval_wer": 30.215616592533205, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 9.70936936936937e-06, | |
| "loss": 0.3526, | |
| "step": 15025 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 9.708868868868869e-06, | |
| "loss": 0.361, | |
| "step": 15050 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 9.70836836836837e-06, | |
| "loss": 0.3731, | |
| "step": 15075 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 9.707867867867868e-06, | |
| "loss": 0.3499, | |
| "step": 15100 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 9.707367367367369e-06, | |
| "loss": 0.3608, | |
| "step": 15125 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 9.706866866866868e-06, | |
| "loss": 0.3465, | |
| "step": 15150 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 9.706366366366367e-06, | |
| "loss": 0.362, | |
| "step": 15175 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 9.705865865865867e-06, | |
| "loss": 0.348, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 9.705365365365366e-06, | |
| "loss": 0.3532, | |
| "step": 15225 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 9.704864864864865e-06, | |
| "loss": 0.3352, | |
| "step": 15250 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 9.704364364364364e-06, | |
| "loss": 0.3424, | |
| "step": 15275 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 9.703863863863865e-06, | |
| "loss": 0.3569, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 9.703363363363365e-06, | |
| "loss": 0.329, | |
| "step": 15325 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 9.702862862862864e-06, | |
| "loss": 0.3551, | |
| "step": 15350 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 9.702362362362363e-06, | |
| "loss": 0.3592, | |
| "step": 15375 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 9.701861861861863e-06, | |
| "loss": 0.359, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 9.701361361361362e-06, | |
| "loss": 0.3498, | |
| "step": 15425 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 9.700860860860861e-06, | |
| "loss": 0.352, | |
| "step": 15450 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 9.70036036036036e-06, | |
| "loss": 0.3451, | |
| "step": 15475 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 9.69985985985986e-06, | |
| "loss": 0.3582, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 9.69935935935936e-06, | |
| "loss": 0.3495, | |
| "step": 15525 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 9.69885885885886e-06, | |
| "loss": 0.3603, | |
| "step": 15550 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 9.69835835835836e-06, | |
| "loss": 0.3398, | |
| "step": 15575 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 9.697857857857859e-06, | |
| "loss": 0.3532, | |
| "step": 15600 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 9.697357357357358e-06, | |
| "loss": 0.3848, | |
| "step": 15625 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 9.696856856856857e-06, | |
| "loss": 0.3341, | |
| "step": 15650 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 9.696356356356357e-06, | |
| "loss": 0.37, | |
| "step": 15675 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 9.695855855855858e-06, | |
| "loss": 0.3635, | |
| "step": 15700 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 9.695355355355355e-06, | |
| "loss": 0.3614, | |
| "step": 15725 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 9.694854854854856e-06, | |
| "loss": 0.3651, | |
| "step": 15750 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 9.694354354354355e-06, | |
| "loss": 0.3733, | |
| "step": 15775 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 9.693853853853855e-06, | |
| "loss": 0.3455, | |
| "step": 15800 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 9.693353353353354e-06, | |
| "loss": 0.3551, | |
| "step": 15825 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 9.692852852852853e-06, | |
| "loss": 0.3517, | |
| "step": 15850 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 9.692352352352353e-06, | |
| "loss": 0.3595, | |
| "step": 15875 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 9.691851851851852e-06, | |
| "loss": 0.3694, | |
| "step": 15900 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 9.691351351351353e-06, | |
| "loss": 0.3433, | |
| "step": 15925 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 9.69085085085085e-06, | |
| "loss": 0.3443, | |
| "step": 15950 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 9.690350350350352e-06, | |
| "loss": 0.3754, | |
| "step": 15975 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 9.689849849849851e-06, | |
| "loss": 0.369, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "eval_loss": 0.4858362674713135, | |
| "eval_runtime": 2697.5515, | |
| "eval_samples_per_second": 6.01, | |
| "eval_steps_per_second": 0.151, | |
| "eval_wer": 31.24320657410905, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 9.68934934934935e-06, | |
| "loss": 0.3536, | |
| "step": 16025 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 9.68884884884885e-06, | |
| "loss": 0.339, | |
| "step": 16050 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 9.688348348348349e-06, | |
| "loss": 0.3319, | |
| "step": 16075 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 9.687847847847848e-06, | |
| "loss": 0.3742, | |
| "step": 16100 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 9.687347347347347e-06, | |
| "loss": 0.3299, | |
| "step": 16125 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 9.686846846846848e-06, | |
| "loss": 0.3526, | |
| "step": 16150 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 9.686346346346348e-06, | |
| "loss": 0.3479, | |
| "step": 16175 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 9.685845845845847e-06, | |
| "loss": 0.3631, | |
| "step": 16200 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 9.685345345345346e-06, | |
| "loss": 0.3545, | |
| "step": 16225 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 9.684844844844846e-06, | |
| "loss": 0.3651, | |
| "step": 16250 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 9.684344344344345e-06, | |
| "loss": 0.3438, | |
| "step": 16275 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 9.683843843843844e-06, | |
| "loss": 0.3502, | |
| "step": 16300 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 9.683343343343345e-06, | |
| "loss": 0.357, | |
| "step": 16325 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 9.682842842842843e-06, | |
| "loss": 0.349, | |
| "step": 16350 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 9.682342342342344e-06, | |
| "loss": 0.3369, | |
| "step": 16375 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 9.681841841841843e-06, | |
| "loss": 0.361, | |
| "step": 16400 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 9.681341341341342e-06, | |
| "loss": 0.3435, | |
| "step": 16425 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 9.680840840840842e-06, | |
| "loss": 0.3185, | |
| "step": 16450 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 9.680340340340341e-06, | |
| "loss": 0.3341, | |
| "step": 16475 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 9.67983983983984e-06, | |
| "loss": 0.3349, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 9.67933933933934e-06, | |
| "loss": 0.3453, | |
| "step": 16525 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 9.67883883883884e-06, | |
| "loss": 0.3519, | |
| "step": 16550 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 9.678338338338338e-06, | |
| "loss": 0.3529, | |
| "step": 16575 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 9.677857857857859e-06, | |
| "loss": 0.3486, | |
| "step": 16600 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 9.677357357357358e-06, | |
| "loss": 0.3424, | |
| "step": 16625 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 9.676856856856858e-06, | |
| "loss": 0.3455, | |
| "step": 16650 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 9.676356356356357e-06, | |
| "loss": 0.3431, | |
| "step": 16675 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.675855855855856e-06, | |
| "loss": 0.3379, | |
| "step": 16700 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.675355355355355e-06, | |
| "loss": 0.3539, | |
| "step": 16725 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.674854854854856e-06, | |
| "loss": 0.327, | |
| "step": 16750 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.674354354354356e-06, | |
| "loss": 0.3275, | |
| "step": 16775 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.673853853853855e-06, | |
| "loss": 0.3396, | |
| "step": 16800 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.673353353353354e-06, | |
| "loss": 0.3476, | |
| "step": 16825 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.672852852852854e-06, | |
| "loss": 0.3675, | |
| "step": 16850 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.672352352352353e-06, | |
| "loss": 0.3571, | |
| "step": 16875 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.671851851851852e-06, | |
| "loss": 0.3463, | |
| "step": 16900 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.671351351351353e-06, | |
| "loss": 0.3381, | |
| "step": 16925 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.67085085085085e-06, | |
| "loss": 0.3582, | |
| "step": 16950 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.670350350350352e-06, | |
| "loss": 0.3633, | |
| "step": 16975 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.669849849849851e-06, | |
| "loss": 0.3585, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "eval_loss": 0.4826071560382843, | |
| "eval_runtime": 2556.6287, | |
| "eval_samples_per_second": 6.342, | |
| "eval_steps_per_second": 0.159, | |
| "eval_wer": 29.16952536597775, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.66934934934935e-06, | |
| "loss": 0.3504, | |
| "step": 17025 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.66884884884885e-06, | |
| "loss": 0.3523, | |
| "step": 17050 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.668348348348349e-06, | |
| "loss": 0.3572, | |
| "step": 17075 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.667847847847848e-06, | |
| "loss": 0.364, | |
| "step": 17100 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.667347347347348e-06, | |
| "loss": 0.3251, | |
| "step": 17125 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.666846846846849e-06, | |
| "loss": 0.3318, | |
| "step": 17150 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.666346346346346e-06, | |
| "loss": 0.3456, | |
| "step": 17175 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.665845845845847e-06, | |
| "loss": 0.345, | |
| "step": 17200 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.665345345345346e-06, | |
| "loss": 0.3224, | |
| "step": 17225 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.664844844844846e-06, | |
| "loss": 0.3397, | |
| "step": 17250 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.664344344344345e-06, | |
| "loss": 0.3484, | |
| "step": 17275 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.663843843843844e-06, | |
| "loss": 0.3267, | |
| "step": 17300 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.663343343343344e-06, | |
| "loss": 0.3642, | |
| "step": 17325 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.662842842842843e-06, | |
| "loss": 0.3447, | |
| "step": 17350 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.662342342342344e-06, | |
| "loss": 0.3429, | |
| "step": 17375 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.661841841841843e-06, | |
| "loss": 0.3258, | |
| "step": 17400 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.661341341341343e-06, | |
| "loss": 0.3305, | |
| "step": 17425 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.660840840840842e-06, | |
| "loss": 0.3376, | |
| "step": 17450 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.660340340340341e-06, | |
| "loss": 0.3417, | |
| "step": 17475 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.65983983983984e-06, | |
| "loss": 0.3326, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.65933933933934e-06, | |
| "loss": 0.334, | |
| "step": 17525 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.65883883883884e-06, | |
| "loss": 0.3461, | |
| "step": 17550 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.658338338338338e-06, | |
| "loss": 0.3512, | |
| "step": 17575 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.65783783783784e-06, | |
| "loss": 0.3609, | |
| "step": 17600 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.657337337337339e-06, | |
| "loss": 0.3502, | |
| "step": 17625 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.656836836836838e-06, | |
| "loss": 0.3508, | |
| "step": 17650 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.656336336336337e-06, | |
| "loss": 0.3436, | |
| "step": 17675 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 9.655835835835836e-06, | |
| "loss": 0.3525, | |
| "step": 17700 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 9.655335335335336e-06, | |
| "loss": 0.339, | |
| "step": 17725 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 9.654834834834835e-06, | |
| "loss": 0.3507, | |
| "step": 17750 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 9.654334334334336e-06, | |
| "loss": 0.3277, | |
| "step": 17775 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 9.653833833833834e-06, | |
| "loss": 0.3486, | |
| "step": 17800 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 9.653353353353354e-06, | |
| "loss": 0.3529, | |
| "step": 17825 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 9.652852852852854e-06, | |
| "loss": 0.3228, | |
| "step": 17850 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 9.652352352352353e-06, | |
| "loss": 0.335, | |
| "step": 17875 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 9.651851851851852e-06, | |
| "loss": 0.3641, | |
| "step": 17900 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 9.651351351351352e-06, | |
| "loss": 0.3248, | |
| "step": 17925 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 9.650850850850851e-06, | |
| "loss": 0.3308, | |
| "step": 17950 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 9.650350350350352e-06, | |
| "loss": 0.333, | |
| "step": 17975 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 9.649849849849851e-06, | |
| "loss": 0.3302, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "eval_loss": 0.4842660427093506, | |
| "eval_runtime": 2644.843, | |
| "eval_samples_per_second": 6.13, | |
| "eval_steps_per_second": 0.154, | |
| "eval_wer": 29.56498947741692, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 9.64934934934935e-06, | |
| "loss": 0.3624, | |
| "step": 18025 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 9.64884884884885e-06, | |
| "loss": 0.3645, | |
| "step": 18050 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 9.648348348348349e-06, | |
| "loss": 0.3563, | |
| "step": 18075 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 9.647847847847848e-06, | |
| "loss": 0.3354, | |
| "step": 18100 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 9.647347347347348e-06, | |
| "loss": 0.3352, | |
| "step": 18125 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 9.646846846846849e-06, | |
| "loss": 0.3426, | |
| "step": 18150 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 9.646346346346346e-06, | |
| "loss": 0.3534, | |
| "step": 18175 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 9.645845845845847e-06, | |
| "loss": 0.3316, | |
| "step": 18200 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 9.645345345345347e-06, | |
| "loss": 0.3387, | |
| "step": 18225 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 9.644844844844846e-06, | |
| "loss": 0.3489, | |
| "step": 18250 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 9.644344344344345e-06, | |
| "loss": 0.3339, | |
| "step": 18275 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 9.643843843843845e-06, | |
| "loss": 0.3416, | |
| "step": 18300 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 9.643343343343344e-06, | |
| "loss": 0.3259, | |
| "step": 18325 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 9.642842842842843e-06, | |
| "loss": 0.3476, | |
| "step": 18350 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 9.642342342342344e-06, | |
| "loss": 0.3499, | |
| "step": 18375 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 9.641841841841842e-06, | |
| "loss": 0.3438, | |
| "step": 18400 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 9.641341341341343e-06, | |
| "loss": 0.3316, | |
| "step": 18425 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 9.640840840840842e-06, | |
| "loss": 0.3397, | |
| "step": 18450 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 9.640340340340341e-06, | |
| "loss": 0.3281, | |
| "step": 18475 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 9.63983983983984e-06, | |
| "loss": 0.3316, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 9.63933933933934e-06, | |
| "loss": 0.3487, | |
| "step": 18525 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 9.63883883883884e-06, | |
| "loss": 0.3273, | |
| "step": 18550 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 9.638338338338338e-06, | |
| "loss": 0.3522, | |
| "step": 18575 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 9.63783783783784e-06, | |
| "loss": 0.3261, | |
| "step": 18600 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 9.637337337337339e-06, | |
| "loss": 0.3551, | |
| "step": 18625 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 9.636836836836838e-06, | |
| "loss": 0.343, | |
| "step": 18650 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 9.636336336336337e-06, | |
| "loss": 0.3302, | |
| "step": 18675 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 9.635835835835837e-06, | |
| "loss": 0.3287, | |
| "step": 18700 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 9.635335335335336e-06, | |
| "loss": 0.3317, | |
| "step": 18725 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 9.634834834834835e-06, | |
| "loss": 0.3346, | |
| "step": 18750 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 9.634334334334336e-06, | |
| "loss": 0.3531, | |
| "step": 18775 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 9.633833833833834e-06, | |
| "loss": 0.35, | |
| "step": 18800 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 9.633333333333335e-06, | |
| "loss": 0.3407, | |
| "step": 18825 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 9.632832832832834e-06, | |
| "loss": 0.3193, | |
| "step": 18850 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 9.632332332332333e-06, | |
| "loss": 0.351, | |
| "step": 18875 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 9.631831831831833e-06, | |
| "loss": 0.3554, | |
| "step": 18900 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 9.631331331331332e-06, | |
| "loss": 0.3389, | |
| "step": 18925 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 9.630830830830831e-06, | |
| "loss": 0.3196, | |
| "step": 18950 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 9.63033033033033e-06, | |
| "loss": 0.3271, | |
| "step": 18975 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 9.629829829829832e-06, | |
| "loss": 0.3252, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "eval_loss": 0.4709344506263733, | |
| "eval_runtime": 2536.9152, | |
| "eval_samples_per_second": 6.391, | |
| "eval_steps_per_second": 0.16, | |
| "eval_wer": 28.698514504205182, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 9.62932932932933e-06, | |
| "loss": 0.3142, | |
| "step": 19025 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 9.62882882882883e-06, | |
| "loss": 0.3346, | |
| "step": 19050 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 9.62832832832833e-06, | |
| "loss": 0.3582, | |
| "step": 19075 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 9.627827827827829e-06, | |
| "loss": 0.327, | |
| "step": 19100 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 9.627327327327328e-06, | |
| "loss": 0.3347, | |
| "step": 19125 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 9.626826826826827e-06, | |
| "loss": 0.3509, | |
| "step": 19150 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 9.626326326326327e-06, | |
| "loss": 0.3349, | |
| "step": 19175 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 9.625825825825826e-06, | |
| "loss": 0.335, | |
| "step": 19200 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 9.625325325325327e-06, | |
| "loss": 0.3371, | |
| "step": 19225 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 9.624824824824826e-06, | |
| "loss": 0.3475, | |
| "step": 19250 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 9.624324324324326e-06, | |
| "loss": 0.3428, | |
| "step": 19275 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 9.623823823823825e-06, | |
| "loss": 0.3375, | |
| "step": 19300 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 9.623323323323324e-06, | |
| "loss": 0.335, | |
| "step": 19325 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 9.622822822822823e-06, | |
| "loss": 0.3488, | |
| "step": 19350 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 9.622322322322323e-06, | |
| "loss": 0.3255, | |
| "step": 19375 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 9.621821821821822e-06, | |
| "loss": 0.3419, | |
| "step": 19400 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 9.621321321321321e-06, | |
| "loss": 0.3257, | |
| "step": 19425 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 9.620820820820822e-06, | |
| "loss": 0.337, | |
| "step": 19450 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 9.620320320320322e-06, | |
| "loss": 0.3159, | |
| "step": 19475 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 9.619819819819821e-06, | |
| "loss": 0.336, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 9.61931931931932e-06, | |
| "loss": 0.3218, | |
| "step": 19525 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 9.61881881881882e-06, | |
| "loss": 0.3447, | |
| "step": 19550 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 9.618318318318319e-06, | |
| "loss": 0.3395, | |
| "step": 19575 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 9.617817817817818e-06, | |
| "loss": 0.3237, | |
| "step": 19600 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 9.617317317317319e-06, | |
| "loss": 0.331, | |
| "step": 19625 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 9.616816816816817e-06, | |
| "loss": 0.3152, | |
| "step": 19650 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 9.616316316316318e-06, | |
| "loss": 0.3391, | |
| "step": 19675 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 9.615815815815817e-06, | |
| "loss": 0.3293, | |
| "step": 19700 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 9.615315315315316e-06, | |
| "loss": 0.3322, | |
| "step": 19725 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 9.614814814814816e-06, | |
| "loss": 0.3294, | |
| "step": 19750 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 9.614314314314315e-06, | |
| "loss": 0.334, | |
| "step": 19775 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 9.613813813813814e-06, | |
| "loss": 0.3418, | |
| "step": 19800 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 9.613313313313314e-06, | |
| "loss": 0.3456, | |
| "step": 19825 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 9.612812812812815e-06, | |
| "loss": 0.3513, | |
| "step": 19850 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 9.612312312312314e-06, | |
| "loss": 0.3395, | |
| "step": 19875 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 9.611811811811813e-06, | |
| "loss": 0.3478, | |
| "step": 19900 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 9.611311311311312e-06, | |
| "loss": 0.3278, | |
| "step": 19925 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.610810810810812e-06, | |
| "loss": 0.3012, | |
| "step": 19950 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.610310310310311e-06, | |
| "loss": 0.3453, | |
| "step": 19975 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.60980980980981e-06, | |
| "loss": 0.3453, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "eval_loss": 0.471629798412323, | |
| "eval_runtime": 2478.9552, | |
| "eval_samples_per_second": 6.54, | |
| "eval_steps_per_second": 0.164, | |
| "eval_wer": 27.84051926827576, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.60930930930931e-06, | |
| "loss": 0.3442, | |
| "step": 20025 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.608808808808809e-06, | |
| "loss": 0.3431, | |
| "step": 20050 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.60830830830831e-06, | |
| "loss": 0.3146, | |
| "step": 20075 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.60780780780781e-06, | |
| "loss": 0.3474, | |
| "step": 20100 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.607307307307307e-06, | |
| "loss": 0.3335, | |
| "step": 20125 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.606806806806808e-06, | |
| "loss": 0.3374, | |
| "step": 20150 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.606306306306307e-06, | |
| "loss": 0.3235, | |
| "step": 20175 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.605805805805806e-06, | |
| "loss": 0.339, | |
| "step": 20200 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.605305305305306e-06, | |
| "loss": 0.3468, | |
| "step": 20225 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.604804804804807e-06, | |
| "loss": 0.3587, | |
| "step": 20250 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.604304304304304e-06, | |
| "loss": 0.3113, | |
| "step": 20275 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.603803803803805e-06, | |
| "loss": 0.3224, | |
| "step": 20300 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.603303303303305e-06, | |
| "loss": 0.3298, | |
| "step": 20325 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.602802802802804e-06, | |
| "loss": 0.3489, | |
| "step": 20350 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.602302302302303e-06, | |
| "loss": 0.2943, | |
| "step": 20375 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.601801801801802e-06, | |
| "loss": 0.3283, | |
| "step": 20400 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.601301301301302e-06, | |
| "loss": 0.3167, | |
| "step": 20425 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.600800800800801e-06, | |
| "loss": 0.3427, | |
| "step": 20450 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.600300300300302e-06, | |
| "loss": 0.3538, | |
| "step": 20475 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.599799799799801e-06, | |
| "loss": 0.3353, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.5992992992993e-06, | |
| "loss": 0.3427, | |
| "step": 20525 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.5987987987988e-06, | |
| "loss": 0.3349, | |
| "step": 20550 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.5982982982983e-06, | |
| "loss": 0.334, | |
| "step": 20575 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.597797797797799e-06, | |
| "loss": 0.3281, | |
| "step": 20600 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.597297297297298e-06, | |
| "loss": 0.3231, | |
| "step": 20625 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.596796796796797e-06, | |
| "loss": 0.336, | |
| "step": 20650 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.596296296296296e-06, | |
| "loss": 0.3208, | |
| "step": 20675 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 9.595795795795797e-06, | |
| "loss": 0.3261, | |
| "step": 20700 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 9.595295295295297e-06, | |
| "loss": 0.3354, | |
| "step": 20725 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 9.594794794794794e-06, | |
| "loss": 0.3311, | |
| "step": 20750 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 9.594294294294295e-06, | |
| "loss": 0.3372, | |
| "step": 20775 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 9.593793793793795e-06, | |
| "loss": 0.3263, | |
| "step": 20800 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 9.593293293293294e-06, | |
| "loss": 0.3187, | |
| "step": 20825 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 9.592792792792793e-06, | |
| "loss": 0.317, | |
| "step": 20850 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 9.592292292292294e-06, | |
| "loss": 0.3428, | |
| "step": 20875 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 9.591791791791792e-06, | |
| "loss": 0.3049, | |
| "step": 20900 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 9.591291291291293e-06, | |
| "loss": 0.3125, | |
| "step": 20925 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 9.590790790790792e-06, | |
| "loss": 0.3206, | |
| "step": 20950 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 9.590290290290291e-06, | |
| "loss": 0.3515, | |
| "step": 20975 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 9.58978978978979e-06, | |
| "loss": 0.3412, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "eval_loss": 0.4714227020740509, | |
| "eval_runtime": 2637.8451, | |
| "eval_samples_per_second": 6.146, | |
| "eval_steps_per_second": 0.154, | |
| "eval_wer": 29.17954687367504, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 9.58928928928929e-06, | |
| "loss": 0.3176, | |
| "step": 21025 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 9.58878878878879e-06, | |
| "loss": 0.3355, | |
| "step": 21050 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 9.588288288288289e-06, | |
| "loss": 0.3306, | |
| "step": 21075 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 9.58778778778779e-06, | |
| "loss": 0.3532, | |
| "step": 21100 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 9.587287287287287e-06, | |
| "loss": 0.3311, | |
| "step": 21125 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 9.586786786786788e-06, | |
| "loss": 0.3377, | |
| "step": 21150 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 9.586286286286287e-06, | |
| "loss": 0.3171, | |
| "step": 21175 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 9.585785785785787e-06, | |
| "loss": 0.3454, | |
| "step": 21200 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 9.585285285285286e-06, | |
| "loss": 0.311, | |
| "step": 21225 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 9.584784784784785e-06, | |
| "loss": 0.3303, | |
| "step": 21250 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 9.584284284284285e-06, | |
| "loss": 0.3204, | |
| "step": 21275 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 9.583783783783784e-06, | |
| "loss": 0.3104, | |
| "step": 21300 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 9.583283283283285e-06, | |
| "loss": 0.3244, | |
| "step": 21325 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 9.582782782782784e-06, | |
| "loss": 0.3192, | |
| "step": 21350 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 9.582282282282282e-06, | |
| "loss": 0.3238, | |
| "step": 21375 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 9.581781781781783e-06, | |
| "loss": 0.3298, | |
| "step": 21400 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 9.581281281281282e-06, | |
| "loss": 0.3223, | |
| "step": 21425 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 9.580780780780781e-06, | |
| "loss": 0.3576, | |
| "step": 21450 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 9.58028028028028e-06, | |
| "loss": 0.305, | |
| "step": 21475 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 9.579779779779782e-06, | |
| "loss": 0.3287, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 9.57927927927928e-06, | |
| "loss": 0.3372, | |
| "step": 21525 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 9.57877877877878e-06, | |
| "loss": 0.3142, | |
| "step": 21550 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 9.57827827827828e-06, | |
| "loss": 0.3222, | |
| "step": 21575 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 9.577777777777779e-06, | |
| "loss": 0.3413, | |
| "step": 21600 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 9.577277277277278e-06, | |
| "loss": 0.3194, | |
| "step": 21625 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 9.576776776776777e-06, | |
| "loss": 0.3226, | |
| "step": 21650 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 9.576276276276277e-06, | |
| "loss": 0.3162, | |
| "step": 21675 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 9.575775775775776e-06, | |
| "loss": 0.3097, | |
| "step": 21700 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 9.575275275275277e-06, | |
| "loss": 0.3366, | |
| "step": 21725 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 9.574774774774775e-06, | |
| "loss": 0.322, | |
| "step": 21750 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 9.574274274274276e-06, | |
| "loss": 0.3407, | |
| "step": 21775 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 9.573773773773775e-06, | |
| "loss": 0.325, | |
| "step": 21800 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 9.573273273273274e-06, | |
| "loss": 0.319, | |
| "step": 21825 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 9.572772772772774e-06, | |
| "loss": 0.318, | |
| "step": 21850 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 9.572272272272273e-06, | |
| "loss": 0.3029, | |
| "step": 21875 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 9.571771771771772e-06, | |
| "loss": 0.3398, | |
| "step": 21900 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 9.571271271271271e-06, | |
| "loss": 0.345, | |
| "step": 21925 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 9.570770770770772e-06, | |
| "loss": 0.3366, | |
| "step": 21950 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 9.570270270270272e-06, | |
| "loss": 0.3192, | |
| "step": 21975 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 9.56976976976977e-06, | |
| "loss": 0.3221, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "eval_loss": 0.4648185670375824, | |
| "eval_runtime": 2509.1979, | |
| "eval_samples_per_second": 6.461, | |
| "eval_steps_per_second": 0.162, | |
| "eval_wer": 27.938421689626196, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 9.56926926926927e-06, | |
| "loss": 0.3333, | |
| "step": 22025 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 9.56876876876877e-06, | |
| "loss": 0.3309, | |
| "step": 22050 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 9.568268268268269e-06, | |
| "loss": 0.3252, | |
| "step": 22075 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 9.567767767767768e-06, | |
| "loss": 0.339, | |
| "step": 22100 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 9.56726726726727e-06, | |
| "loss": 0.3185, | |
| "step": 22125 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 9.566766766766767e-06, | |
| "loss": 0.3312, | |
| "step": 22150 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 9.566266266266268e-06, | |
| "loss": 0.3279, | |
| "step": 22175 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 9.565765765765767e-06, | |
| "loss": 0.3111, | |
| "step": 22200 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 9.565265265265266e-06, | |
| "loss": 0.3294, | |
| "step": 22225 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 9.564764764764766e-06, | |
| "loss": 0.3359, | |
| "step": 22250 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 9.564264264264265e-06, | |
| "loss": 0.3324, | |
| "step": 22275 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 9.563763763763764e-06, | |
| "loss": 0.3331, | |
| "step": 22300 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 9.563263263263264e-06, | |
| "loss": 0.3331, | |
| "step": 22325 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 9.562762762762765e-06, | |
| "loss": 0.307, | |
| "step": 22350 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 9.562262262262262e-06, | |
| "loss": 0.3384, | |
| "step": 22375 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 9.561761761761763e-06, | |
| "loss": 0.3291, | |
| "step": 22400 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 9.561261261261262e-06, | |
| "loss": 0.3411, | |
| "step": 22425 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 9.560760760760762e-06, | |
| "loss": 0.3141, | |
| "step": 22450 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 9.560260260260261e-06, | |
| "loss": 0.3289, | |
| "step": 22475 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 9.55975975975976e-06, | |
| "loss": 0.3289, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 9.55925925925926e-06, | |
| "loss": 0.2942, | |
| "step": 22525 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 9.558758758758759e-06, | |
| "loss": 0.3307, | |
| "step": 22550 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 9.55825825825826e-06, | |
| "loss": 0.3565, | |
| "step": 22575 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 9.55775775775776e-06, | |
| "loss": 0.3065, | |
| "step": 22600 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 9.557257257257257e-06, | |
| "loss": 0.3128, | |
| "step": 22625 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 9.556756756756758e-06, | |
| "loss": 0.3181, | |
| "step": 22650 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 9.556256256256257e-06, | |
| "loss": 0.321, | |
| "step": 22675 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 9.555755755755756e-06, | |
| "loss": 0.3305, | |
| "step": 22700 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 9.555255255255256e-06, | |
| "loss": 0.3454, | |
| "step": 22725 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 9.554754754754757e-06, | |
| "loss": 0.3236, | |
| "step": 22750 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 9.554254254254254e-06, | |
| "loss": 0.3345, | |
| "step": 22775 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 9.553753753753755e-06, | |
| "loss": 0.3221, | |
| "step": 22800 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 9.553253253253255e-06, | |
| "loss": 0.3333, | |
| "step": 22825 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 9.552752752752752e-06, | |
| "loss": 0.3176, | |
| "step": 22850 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 9.552252252252253e-06, | |
| "loss": 0.3286, | |
| "step": 22875 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 9.551751751751753e-06, | |
| "loss": 0.3165, | |
| "step": 22900 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 9.551251251251252e-06, | |
| "loss": 0.3312, | |
| "step": 22925 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 9.550750750750751e-06, | |
| "loss": 0.3334, | |
| "step": 22950 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 9.550250250250252e-06, | |
| "loss": 0.3092, | |
| "step": 22975 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 9.54974974974975e-06, | |
| "loss": 0.322, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "eval_loss": 0.46254613995552063, | |
| "eval_runtime": 2464.8795, | |
| "eval_samples_per_second": 6.578, | |
| "eval_steps_per_second": 0.165, | |
| "eval_wer": 27.390322307105247, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 9.54924924924925e-06, | |
| "loss": 0.3245, | |
| "step": 23025 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 9.54874874874875e-06, | |
| "loss": 0.3263, | |
| "step": 23050 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 9.54824824824825e-06, | |
| "loss": 0.3246, | |
| "step": 23075 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 9.547747747747749e-06, | |
| "loss": 0.3407, | |
| "step": 23100 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 9.547247247247248e-06, | |
| "loss": 0.3116, | |
| "step": 23125 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 9.546746746746747e-06, | |
| "loss": 0.3278, | |
| "step": 23150 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 9.546246246246246e-06, | |
| "loss": 0.3369, | |
| "step": 23175 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 9.545745745745747e-06, | |
| "loss": 0.3133, | |
| "step": 23200 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 9.545245245245247e-06, | |
| "loss": 0.3112, | |
| "step": 23225 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 9.544744744744744e-06, | |
| "loss": 0.3316, | |
| "step": 23250 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 9.544244244244245e-06, | |
| "loss": 0.3266, | |
| "step": 23275 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 9.543743743743745e-06, | |
| "loss": 0.3237, | |
| "step": 23300 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 9.543243243243244e-06, | |
| "loss": 0.3073, | |
| "step": 23325 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 9.542742742742743e-06, | |
| "loss": 0.3326, | |
| "step": 23350 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 9.542242242242243e-06, | |
| "loss": 0.3063, | |
| "step": 23375 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 9.541741741741742e-06, | |
| "loss": 0.3145, | |
| "step": 23400 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 9.541241241241243e-06, | |
| "loss": 0.3329, | |
| "step": 23425 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 9.540740740740742e-06, | |
| "loss": 0.3285, | |
| "step": 23450 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 9.54024024024024e-06, | |
| "loss": 0.3312, | |
| "step": 23475 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 9.53973973973974e-06, | |
| "loss": 0.3203, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 9.53923923923924e-06, | |
| "loss": 0.327, | |
| "step": 23525 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 9.53873873873874e-06, | |
| "loss": 0.3136, | |
| "step": 23550 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 9.538238238238239e-06, | |
| "loss": 0.314, | |
| "step": 23575 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 9.53773773773774e-06, | |
| "loss": 0.3294, | |
| "step": 23600 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 9.537237237237237e-06, | |
| "loss": 0.3331, | |
| "step": 23625 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 9.536736736736738e-06, | |
| "loss": 0.3112, | |
| "step": 23650 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 9.536236236236237e-06, | |
| "loss": 0.3047, | |
| "step": 23675 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 9.535735735735737e-06, | |
| "loss": 0.3153, | |
| "step": 23700 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 9.535235235235236e-06, | |
| "loss": 0.3195, | |
| "step": 23725 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 9.534734734734735e-06, | |
| "loss": 0.3187, | |
| "step": 23750 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 9.534234234234235e-06, | |
| "loss": 0.3351, | |
| "step": 23775 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 9.533733733733734e-06, | |
| "loss": 0.3172, | |
| "step": 23800 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 9.533233233233235e-06, | |
| "loss": 0.3206, | |
| "step": 23825 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 9.532752752752752e-06, | |
| "loss": 0.3261, | |
| "step": 23850 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 9.532252252252253e-06, | |
| "loss": 0.3204, | |
| "step": 23875 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 9.531751751751753e-06, | |
| "loss": 0.3154, | |
| "step": 23900 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 9.531251251251252e-06, | |
| "loss": 0.3159, | |
| "step": 23925 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 9.530750750750751e-06, | |
| "loss": 0.326, | |
| "step": 23950 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 9.53025025025025e-06, | |
| "loss": 0.3014, | |
| "step": 23975 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 9.52974974974975e-06, | |
| "loss": 0.312, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "eval_loss": 0.46919143199920654, | |
| "eval_runtime": 2551.3752, | |
| "eval_samples_per_second": 6.355, | |
| "eval_steps_per_second": 0.159, | |
| "eval_wer": 28.381680683929357, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 9.529249249249251e-06, | |
| "loss": 0.317, | |
| "step": 24025 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 9.52874874874875e-06, | |
| "loss": 0.3128, | |
| "step": 24050 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 9.528248248248248e-06, | |
| "loss": 0.324, | |
| "step": 24075 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 9.527747747747749e-06, | |
| "loss": 0.3367, | |
| "step": 24100 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 9.527247247247248e-06, | |
| "loss": 0.3159, | |
| "step": 24125 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 9.526746746746747e-06, | |
| "loss": 0.3288, | |
| "step": 24150 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 9.526246246246247e-06, | |
| "loss": 0.3379, | |
| "step": 24175 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 9.525745745745748e-06, | |
| "loss": 0.3359, | |
| "step": 24200 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 9.525245245245245e-06, | |
| "loss": 0.3146, | |
| "step": 24225 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 9.524744744744745e-06, | |
| "loss": 0.3054, | |
| "step": 24250 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 9.524244244244246e-06, | |
| "loss": 0.3329, | |
| "step": 24275 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 9.523743743743745e-06, | |
| "loss": 0.3157, | |
| "step": 24300 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 9.523243243243244e-06, | |
| "loss": 0.3252, | |
| "step": 24325 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 9.522742742742743e-06, | |
| "loss": 0.3199, | |
| "step": 24350 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 9.522242242242243e-06, | |
| "loss": 0.2962, | |
| "step": 24375 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 9.521741741741742e-06, | |
| "loss": 0.3197, | |
| "step": 24400 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 9.521241241241243e-06, | |
| "loss": 0.3105, | |
| "step": 24425 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 9.520740740740742e-06, | |
| "loss": 0.3158, | |
| "step": 24450 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 9.52024024024024e-06, | |
| "loss": 0.3084, | |
| "step": 24475 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 9.519739739739741e-06, | |
| "loss": 0.3258, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 9.51923923923924e-06, | |
| "loss": 0.3087, | |
| "step": 24525 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 9.51873873873874e-06, | |
| "loss": 0.323, | |
| "step": 24550 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 9.518238238238239e-06, | |
| "loss": 0.3153, | |
| "step": 24575 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 9.517737737737738e-06, | |
| "loss": 0.3256, | |
| "step": 24600 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 9.517237237237237e-06, | |
| "loss": 0.3461, | |
| "step": 24625 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 9.516736736736738e-06, | |
| "loss": 0.319, | |
| "step": 24650 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 9.516236236236238e-06, | |
| "loss": 0.3091, | |
| "step": 24675 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 9.515735735735735e-06, | |
| "loss": 0.3109, | |
| "step": 24700 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 9.515235235235236e-06, | |
| "loss": 0.3062, | |
| "step": 24725 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 9.514734734734736e-06, | |
| "loss": 0.3162, | |
| "step": 24750 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 9.514234234234235e-06, | |
| "loss": 0.3184, | |
| "step": 24775 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 9.513733733733734e-06, | |
| "loss": 0.3168, | |
| "step": 24800 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 9.513233233233235e-06, | |
| "loss": 0.315, | |
| "step": 24825 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 9.512732732732733e-06, | |
| "loss": 0.3023, | |
| "step": 24850 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 9.512232232232232e-06, | |
| "loss": 0.3242, | |
| "step": 24875 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 9.511731731731733e-06, | |
| "loss": 0.3337, | |
| "step": 24900 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 9.511231231231232e-06, | |
| "loss": 0.3149, | |
| "step": 24925 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 9.510730730730732e-06, | |
| "loss": 0.3085, | |
| "step": 24950 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 9.510230230230231e-06, | |
| "loss": 0.3263, | |
| "step": 24975 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 9.50972972972973e-06, | |
| "loss": 0.3089, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 0.457874059677124, | |
| "eval_runtime": 2679.5488, | |
| "eval_samples_per_second": 6.051, | |
| "eval_steps_per_second": 0.152, | |
| "eval_wer": 27.82818510495602, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 9.50922922922923e-06, | |
| "loss": 0.3246, | |
| "step": 25025 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 9.50872872872873e-06, | |
| "loss": 0.3371, | |
| "step": 25050 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 9.50822822822823e-06, | |
| "loss": 0.2992, | |
| "step": 25075 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 9.507727727727727e-06, | |
| "loss": 0.2763, | |
| "step": 25100 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 9.507227227227228e-06, | |
| "loss": 0.277, | |
| "step": 25125 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 9.506726726726728e-06, | |
| "loss": 0.2702, | |
| "step": 25150 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 9.506226226226227e-06, | |
| "loss": 0.2656, | |
| "step": 25175 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 9.505725725725726e-06, | |
| "loss": 0.2767, | |
| "step": 25200 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 9.505225225225226e-06, | |
| "loss": 0.2745, | |
| "step": 25225 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 9.504724724724725e-06, | |
| "loss": 0.2637, | |
| "step": 25250 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 9.504224224224226e-06, | |
| "loss": 0.2704, | |
| "step": 25275 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 9.503723723723725e-06, | |
| "loss": 0.2554, | |
| "step": 25300 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 9.503223223223223e-06, | |
| "loss": 0.271, | |
| "step": 25325 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 9.502722722722724e-06, | |
| "loss": 0.2757, | |
| "step": 25350 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 9.502222222222223e-06, | |
| "loss": 0.2606, | |
| "step": 25375 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 9.501721721721722e-06, | |
| "loss": 0.2664, | |
| "step": 25400 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 9.501221221221222e-06, | |
| "loss": 0.2702, | |
| "step": 25425 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 9.500720720720723e-06, | |
| "loss": 0.2633, | |
| "step": 25450 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 9.50022022022022e-06, | |
| "loss": 0.2816, | |
| "step": 25475 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 9.49971971971972e-06, | |
| "loss": 0.2848, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 9.49921921921922e-06, | |
| "loss": 0.2655, | |
| "step": 25525 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 9.49871871871872e-06, | |
| "loss": 0.2652, | |
| "step": 25550 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 9.49821821821822e-06, | |
| "loss": 0.2704, | |
| "step": 25575 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 9.497717717717718e-06, | |
| "loss": 0.2605, | |
| "step": 25600 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 9.497217217217218e-06, | |
| "loss": 0.2674, | |
| "step": 25625 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 9.496716716716717e-06, | |
| "loss": 0.2677, | |
| "step": 25650 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 9.496216216216218e-06, | |
| "loss": 0.2761, | |
| "step": 25675 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 9.495715715715716e-06, | |
| "loss": 0.2626, | |
| "step": 25700 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 9.495215215215215e-06, | |
| "loss": 0.2637, | |
| "step": 25725 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 9.494714714714716e-06, | |
| "loss": 0.2704, | |
| "step": 25750 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 9.494214214214215e-06, | |
| "loss": 0.2756, | |
| "step": 25775 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 9.493713713713715e-06, | |
| "loss": 0.2807, | |
| "step": 25800 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 9.493213213213214e-06, | |
| "loss": 0.2798, | |
| "step": 25825 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 9.492712712712713e-06, | |
| "loss": 0.2728, | |
| "step": 25850 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 9.492212212212212e-06, | |
| "loss": 0.2664, | |
| "step": 25875 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 9.491731731731733e-06, | |
| "loss": 0.2811, | |
| "step": 25900 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 9.491231231231231e-06, | |
| "loss": 0.2582, | |
| "step": 25925 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 9.490730730730732e-06, | |
| "loss": 0.2739, | |
| "step": 25950 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 9.490230230230231e-06, | |
| "loss": 0.264, | |
| "step": 25975 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 9.48972972972973e-06, | |
| "loss": 0.2625, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "eval_loss": 0.45843926072120667, | |
| "eval_runtime": 2831.694, | |
| "eval_samples_per_second": 5.726, | |
| "eval_steps_per_second": 0.143, | |
| "eval_wer": 27.744929502547777, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 9.48922922922923e-06, | |
| "loss": 0.2841, | |
| "step": 26025 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 9.48872872872873e-06, | |
| "loss": 0.2881, | |
| "step": 26050 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 9.488228228228228e-06, | |
| "loss": 0.2915, | |
| "step": 26075 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 9.487727727727728e-06, | |
| "loss": 0.2687, | |
| "step": 26100 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 9.487227227227229e-06, | |
| "loss": 0.2678, | |
| "step": 26125 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 9.486726726726728e-06, | |
| "loss": 0.2724, | |
| "step": 26150 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 9.486226226226227e-06, | |
| "loss": 0.2792, | |
| "step": 26175 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 9.485725725725727e-06, | |
| "loss": 0.2657, | |
| "step": 26200 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 9.485225225225226e-06, | |
| "loss": 0.2722, | |
| "step": 26225 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 9.484724724724725e-06, | |
| "loss": 0.2704, | |
| "step": 26250 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 9.484224224224226e-06, | |
| "loss": 0.2595, | |
| "step": 26275 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 9.483723723723724e-06, | |
| "loss": 0.2699, | |
| "step": 26300 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 9.483223223223223e-06, | |
| "loss": 0.2857, | |
| "step": 26325 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 9.482722722722724e-06, | |
| "loss": 0.2664, | |
| "step": 26350 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 9.482222222222223e-06, | |
| "loss": 0.2606, | |
| "step": 26375 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 9.481721721721723e-06, | |
| "loss": 0.2607, | |
| "step": 26400 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 9.481221221221222e-06, | |
| "loss": 0.2487, | |
| "step": 26425 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 9.480720720720721e-06, | |
| "loss": 0.2603, | |
| "step": 26450 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 9.48022022022022e-06, | |
| "loss": 0.2819, | |
| "step": 26475 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 9.479719719719721e-06, | |
| "loss": 0.27, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 9.47921921921922e-06, | |
| "loss": 0.2577, | |
| "step": 26525 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 9.478718718718718e-06, | |
| "loss": 0.2743, | |
| "step": 26550 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 9.47821821821822e-06, | |
| "loss": 0.2806, | |
| "step": 26575 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 9.477717717717719e-06, | |
| "loss": 0.2736, | |
| "step": 26600 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 9.477217217217218e-06, | |
| "loss": 0.2699, | |
| "step": 26625 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 9.476716716716717e-06, | |
| "loss": 0.2743, | |
| "step": 26650 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 9.476216216216218e-06, | |
| "loss": 0.2706, | |
| "step": 26675 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 9.475715715715716e-06, | |
| "loss": 0.2658, | |
| "step": 26700 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 9.475215215215215e-06, | |
| "loss": 0.263, | |
| "step": 26725 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 9.474714714714716e-06, | |
| "loss": 0.2653, | |
| "step": 26750 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 9.474214214214215e-06, | |
| "loss": 0.2622, | |
| "step": 26775 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 9.473713713713715e-06, | |
| "loss": 0.2743, | |
| "step": 26800 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 9.473213213213214e-06, | |
| "loss": 0.2531, | |
| "step": 26825 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 9.472712712712713e-06, | |
| "loss": 0.2686, | |
| "step": 26850 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 9.472212212212213e-06, | |
| "loss": 0.2847, | |
| "step": 26875 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 9.471711711711714e-06, | |
| "loss": 0.2556, | |
| "step": 26900 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 9.471211211211211e-06, | |
| "loss": 0.2642, | |
| "step": 26925 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 9.47071071071071e-06, | |
| "loss": 0.2719, | |
| "step": 26950 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 9.470210210210212e-06, | |
| "loss": 0.2632, | |
| "step": 26975 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 9.46970970970971e-06, | |
| "loss": 0.2758, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "eval_loss": 0.45501989126205444, | |
| "eval_runtime": 2547.6663, | |
| "eval_samples_per_second": 6.364, | |
| "eval_steps_per_second": 0.159, | |
| "eval_wer": 26.85301531748907, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 9.46920920920921e-06, | |
| "loss": 0.2766, | |
| "step": 27025 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 9.46870870870871e-06, | |
| "loss": 0.2724, | |
| "step": 27050 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 9.468208208208209e-06, | |
| "loss": 0.265, | |
| "step": 27075 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 9.467707707707708e-06, | |
| "loss": 0.2627, | |
| "step": 27100 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 9.467207207207207e-06, | |
| "loss": 0.2648, | |
| "step": 27125 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 9.466706706706708e-06, | |
| "loss": 0.2676, | |
| "step": 27150 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 9.466206206206206e-06, | |
| "loss": 0.2634, | |
| "step": 27175 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 9.465705705705707e-06, | |
| "loss": 0.2631, | |
| "step": 27200 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 9.465205205205206e-06, | |
| "loss": 0.2729, | |
| "step": 27225 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 9.464704704704705e-06, | |
| "loss": 0.2772, | |
| "step": 27250 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 9.464204204204205e-06, | |
| "loss": 0.2605, | |
| "step": 27275 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 9.463703703703706e-06, | |
| "loss": 0.2753, | |
| "step": 27300 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 9.463203203203203e-06, | |
| "loss": 0.2719, | |
| "step": 27325 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 9.462702702702703e-06, | |
| "loss": 0.2802, | |
| "step": 27350 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 9.462202202202204e-06, | |
| "loss": 0.2579, | |
| "step": 27375 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 9.461701701701701e-06, | |
| "loss": 0.2781, | |
| "step": 27400 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 9.461201201201202e-06, | |
| "loss": 0.292, | |
| "step": 27425 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 9.460700700700702e-06, | |
| "loss": 0.25, | |
| "step": 27450 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 9.460200200200201e-06, | |
| "loss": 0.2518, | |
| "step": 27475 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 9.4596996996997e-06, | |
| "loss": 0.264, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 9.459199199199201e-06, | |
| "loss": 0.2603, | |
| "step": 27525 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 9.458698698698699e-06, | |
| "loss": 0.2601, | |
| "step": 27550 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 9.458198198198198e-06, | |
| "loss": 0.275, | |
| "step": 27575 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 9.457697697697699e-06, | |
| "loss": 0.2681, | |
| "step": 27600 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 9.457197197197198e-06, | |
| "loss": 0.2595, | |
| "step": 27625 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 9.456696696696698e-06, | |
| "loss": 0.2797, | |
| "step": 27650 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 9.456196196196197e-06, | |
| "loss": 0.2564, | |
| "step": 27675 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 9.455695695695696e-06, | |
| "loss": 0.2565, | |
| "step": 27700 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 9.455195195195196e-06, | |
| "loss": 0.2833, | |
| "step": 27725 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 9.454694694694695e-06, | |
| "loss": 0.2648, | |
| "step": 27750 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 9.454194194194196e-06, | |
| "loss": 0.2607, | |
| "step": 27775 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 9.453693693693693e-06, | |
| "loss": 0.2697, | |
| "step": 27800 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 9.453193193193194e-06, | |
| "loss": 0.2778, | |
| "step": 27825 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 9.452692692692694e-06, | |
| "loss": 0.2616, | |
| "step": 27850 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 9.452192192192193e-06, | |
| "loss": 0.27, | |
| "step": 27875 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 9.451691691691692e-06, | |
| "loss": 0.2726, | |
| "step": 27900 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 9.451191191191193e-06, | |
| "loss": 0.2444, | |
| "step": 27925 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 9.450690690690691e-06, | |
| "loss": 0.258, | |
| "step": 27950 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 9.45019019019019e-06, | |
| "loss": 0.2758, | |
| "step": 27975 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 9.449689689689691e-06, | |
| "loss": 0.265, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "eval_loss": 0.45008718967437744, | |
| "eval_runtime": 2652.1712, | |
| "eval_samples_per_second": 6.113, | |
| "eval_steps_per_second": 0.153, | |
| "eval_wer": 28.59213234557242, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 9.44920920920921e-06, | |
| "loss": 0.2703, | |
| "step": 28025 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 9.44870870870871e-06, | |
| "loss": 0.2573, | |
| "step": 28050 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 9.448208208208209e-06, | |
| "loss": 0.2824, | |
| "step": 28075 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 9.447707707707708e-06, | |
| "loss": 0.2697, | |
| "step": 28100 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 9.44720720720721e-06, | |
| "loss": 0.2586, | |
| "step": 28125 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 9.446706706706707e-06, | |
| "loss": 0.2882, | |
| "step": 28150 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 9.446206206206206e-06, | |
| "loss": 0.2649, | |
| "step": 28175 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 9.445705705705707e-06, | |
| "loss": 0.2542, | |
| "step": 28200 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 9.445205205205206e-06, | |
| "loss": 0.263, | |
| "step": 28225 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 9.444704704704706e-06, | |
| "loss": 0.2729, | |
| "step": 28250 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 9.444204204204205e-06, | |
| "loss": 0.2861, | |
| "step": 28275 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 9.443703703703704e-06, | |
| "loss": 0.2692, | |
| "step": 28300 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 9.443203203203204e-06, | |
| "loss": 0.2826, | |
| "step": 28325 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 9.442702702702703e-06, | |
| "loss": 0.2795, | |
| "step": 28350 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 9.442202202202204e-06, | |
| "loss": 0.2655, | |
| "step": 28375 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 9.441701701701701e-06, | |
| "loss": 0.2605, | |
| "step": 28400 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 9.441201201201202e-06, | |
| "loss": 0.2626, | |
| "step": 28425 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 9.440700700700702e-06, | |
| "loss": 0.2867, | |
| "step": 28450 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 9.440200200200201e-06, | |
| "loss": 0.272, | |
| "step": 28475 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 9.4396996996997e-06, | |
| "loss": 0.257, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 9.4391991991992e-06, | |
| "loss": 0.2783, | |
| "step": 28525 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 9.438698698698699e-06, | |
| "loss": 0.2602, | |
| "step": 28550 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 9.438198198198198e-06, | |
| "loss": 0.2644, | |
| "step": 28575 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 9.4376976976977e-06, | |
| "loss": 0.2554, | |
| "step": 28600 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 9.437197197197197e-06, | |
| "loss": 0.2743, | |
| "step": 28625 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 9.436696696696698e-06, | |
| "loss": 0.2724, | |
| "step": 28650 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 9.436196196196197e-06, | |
| "loss": 0.2607, | |
| "step": 28675 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 9.435695695695696e-06, | |
| "loss": 0.2626, | |
| "step": 28700 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 9.435195195195196e-06, | |
| "loss": 0.2616, | |
| "step": 28725 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 9.434694694694697e-06, | |
| "loss": 0.2602, | |
| "step": 28750 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 9.434194194194194e-06, | |
| "loss": 0.2639, | |
| "step": 28775 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 9.433693693693694e-06, | |
| "loss": 0.2564, | |
| "step": 28800 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 9.433193193193195e-06, | |
| "loss": 0.2544, | |
| "step": 28825 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 9.432692692692694e-06, | |
| "loss": 0.2874, | |
| "step": 28850 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 9.432192192192193e-06, | |
| "loss": 0.2663, | |
| "step": 28875 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 9.431691691691692e-06, | |
| "loss": 0.2658, | |
| "step": 28900 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 9.431191191191192e-06, | |
| "loss": 0.286, | |
| "step": 28925 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 9.430690690690691e-06, | |
| "loss": 0.2451, | |
| "step": 28950 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 9.43019019019019e-06, | |
| "loss": 0.2811, | |
| "step": 28975 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 9.429689689689691e-06, | |
| "loss": 0.258, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "eval_loss": 0.44391068816185, | |
| "eval_runtime": 2504.9501, | |
| "eval_samples_per_second": 6.472, | |
| "eval_steps_per_second": 0.162, | |
| "eval_wer": 26.30877036100554, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 9.429189189189189e-06, | |
| "loss": 0.2654, | |
| "step": 29025 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 9.42868868868869e-06, | |
| "loss": 0.2587, | |
| "step": 29050 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 9.42818818818819e-06, | |
| "loss": 0.2571, | |
| "step": 29075 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 9.427687687687689e-06, | |
| "loss": 0.2734, | |
| "step": 29100 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 9.427187187187188e-06, | |
| "loss": 0.2898, | |
| "step": 29125 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 9.426686686686687e-06, | |
| "loss": 0.2514, | |
| "step": 29150 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 9.426186186186186e-06, | |
| "loss": 0.2804, | |
| "step": 29175 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 9.425685685685686e-06, | |
| "loss": 0.2812, | |
| "step": 29200 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 9.425185185185187e-06, | |
| "loss": 0.265, | |
| "step": 29225 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 9.424684684684684e-06, | |
| "loss": 0.2546, | |
| "step": 29250 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 9.424184184184185e-06, | |
| "loss": 0.2818, | |
| "step": 29275 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 9.423683683683685e-06, | |
| "loss": 0.2448, | |
| "step": 29300 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 9.423183183183184e-06, | |
| "loss": 0.2791, | |
| "step": 29325 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 9.422682682682683e-06, | |
| "loss": 0.2583, | |
| "step": 29350 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 9.422182182182184e-06, | |
| "loss": 0.2737, | |
| "step": 29375 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 9.421701701701702e-06, | |
| "loss": 0.266, | |
| "step": 29400 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 9.421201201201203e-06, | |
| "loss": 0.265, | |
| "step": 29425 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 9.420700700700702e-06, | |
| "loss": 0.2771, | |
| "step": 29450 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 9.420200200200201e-06, | |
| "loss": 0.2569, | |
| "step": 29475 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 9.4196996996997e-06, | |
| "loss": 0.2621, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 9.4191991991992e-06, | |
| "loss": 0.2703, | |
| "step": 29525 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 9.418698698698699e-06, | |
| "loss": 0.2786, | |
| "step": 29550 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 9.418198198198198e-06, | |
| "loss": 0.2628, | |
| "step": 29575 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 9.4176976976977e-06, | |
| "loss": 0.295, | |
| "step": 29600 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 9.417197197197197e-06, | |
| "loss": 0.2829, | |
| "step": 29625 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 9.416696696696698e-06, | |
| "loss": 0.2721, | |
| "step": 29650 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 9.416196196196197e-06, | |
| "loss": 0.2721, | |
| "step": 29675 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 9.415695695695697e-06, | |
| "loss": 0.2683, | |
| "step": 29700 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 9.415195195195196e-06, | |
| "loss": 0.2655, | |
| "step": 29725 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 9.414694694694695e-06, | |
| "loss": 0.2567, | |
| "step": 29750 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 9.414194194194194e-06, | |
| "loss": 0.2542, | |
| "step": 29775 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 9.413693693693694e-06, | |
| "loss": 0.2722, | |
| "step": 29800 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 9.413193193193195e-06, | |
| "loss": 0.2707, | |
| "step": 29825 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 9.412692692692692e-06, | |
| "loss": 0.2718, | |
| "step": 29850 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 9.412192192192193e-06, | |
| "loss": 0.2634, | |
| "step": 29875 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 9.411691691691693e-06, | |
| "loss": 0.2643, | |
| "step": 29900 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 9.411191191191192e-06, | |
| "loss": 0.2618, | |
| "step": 29925 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 9.410690690690691e-06, | |
| "loss": 0.2684, | |
| "step": 29950 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 9.41019019019019e-06, | |
| "loss": 0.282, | |
| "step": 29975 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 9.40968968968969e-06, | |
| "loss": 0.2837, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "eval_loss": 0.4546422064304352, | |
| "eval_runtime": 2695.5326, | |
| "eval_samples_per_second": 6.015, | |
| "eval_steps_per_second": 0.151, | |
| "eval_wer": 27.167536482142445, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 9.40918918918919e-06, | |
| "loss": 0.2669, | |
| "step": 30025 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 9.40868868868869e-06, | |
| "loss": 0.2881, | |
| "step": 30050 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 9.40818818818819e-06, | |
| "loss": 0.2759, | |
| "step": 30075 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 9.407687687687689e-06, | |
| "loss": 0.2568, | |
| "step": 30100 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 9.407187187187188e-06, | |
| "loss": 0.2779, | |
| "step": 30125 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 9.406686686686687e-06, | |
| "loss": 0.2668, | |
| "step": 30150 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 9.406186186186187e-06, | |
| "loss": 0.272, | |
| "step": 30175 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 9.405685685685686e-06, | |
| "loss": 0.2638, | |
| "step": 30200 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 9.405185185185187e-06, | |
| "loss": 0.27, | |
| "step": 30225 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 9.404684684684685e-06, | |
| "loss": 0.2757, | |
| "step": 30250 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 9.404184184184186e-06, | |
| "loss": 0.2499, | |
| "step": 30275 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 9.403683683683685e-06, | |
| "loss": 0.2886, | |
| "step": 30300 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 9.403183183183184e-06, | |
| "loss": 0.2756, | |
| "step": 30325 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 9.402682682682683e-06, | |
| "loss": 0.261, | |
| "step": 30350 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 9.402182182182183e-06, | |
| "loss": 0.2748, | |
| "step": 30375 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 9.401681681681682e-06, | |
| "loss": 0.261, | |
| "step": 30400 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 9.401181181181181e-06, | |
| "loss": 0.2839, | |
| "step": 30425 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 9.400680680680682e-06, | |
| "loss": 0.2735, | |
| "step": 30450 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 9.40018018018018e-06, | |
| "loss": 0.2659, | |
| "step": 30475 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 9.3996996996997e-06, | |
| "loss": 0.2621, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 9.3991991991992e-06, | |
| "loss": 0.2744, | |
| "step": 30525 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 9.3986986986987e-06, | |
| "loss": 0.2632, | |
| "step": 30550 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 9.398198198198199e-06, | |
| "loss": 0.2589, | |
| "step": 30575 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 9.397697697697698e-06, | |
| "loss": 0.2528, | |
| "step": 30600 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 9.397197197197197e-06, | |
| "loss": 0.265, | |
| "step": 30625 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 9.396696696696698e-06, | |
| "loss": 0.2569, | |
| "step": 30650 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 9.396196196196197e-06, | |
| "loss": 0.2646, | |
| "step": 30675 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 9.395695695695697e-06, | |
| "loss": 0.2737, | |
| "step": 30700 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 9.395195195195196e-06, | |
| "loss": 0.2555, | |
| "step": 30725 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 9.394694694694695e-06, | |
| "loss": 0.2671, | |
| "step": 30750 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 9.394194194194195e-06, | |
| "loss": 0.2693, | |
| "step": 30775 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 9.393693693693694e-06, | |
| "loss": 0.2795, | |
| "step": 30800 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 9.393193193193193e-06, | |
| "loss": 0.2641, | |
| "step": 30825 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 9.392692692692693e-06, | |
| "loss": 0.2789, | |
| "step": 30850 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 9.392192192192194e-06, | |
| "loss": 0.2617, | |
| "step": 30875 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 9.391691691691693e-06, | |
| "loss": 0.2536, | |
| "step": 30900 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 9.391191191191192e-06, | |
| "loss": 0.2636, | |
| "step": 30925 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 9.390690690690691e-06, | |
| "loss": 0.2713, | |
| "step": 30950 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 9.39019019019019e-06, | |
| "loss": 0.2741, | |
| "step": 30975 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 9.38968968968969e-06, | |
| "loss": 0.2804, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "eval_loss": 0.44737279415130615, | |
| "eval_runtime": 2512.7508, | |
| "eval_samples_per_second": 6.452, | |
| "eval_steps_per_second": 0.162, | |
| "eval_wer": 26.34577285096476, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 9.38918918918919e-06, | |
| "loss": 0.277, | |
| "step": 31025 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 9.38868868868869e-06, | |
| "loss": 0.2563, | |
| "step": 31050 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 9.388188188188188e-06, | |
| "loss": 0.2455, | |
| "step": 31075 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 9.387687687687689e-06, | |
| "loss": 0.2606, | |
| "step": 31100 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 9.387187187187188e-06, | |
| "loss": 0.2602, | |
| "step": 31125 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 9.386686686686688e-06, | |
| "loss": 0.2576, | |
| "step": 31150 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 9.386186186186187e-06, | |
| "loss": 0.27, | |
| "step": 31175 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 9.385685685685686e-06, | |
| "loss": 0.2806, | |
| "step": 31200 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 9.385185185185185e-06, | |
| "loss": 0.2673, | |
| "step": 31225 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 9.384684684684685e-06, | |
| "loss": 0.2766, | |
| "step": 31250 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 9.384184184184186e-06, | |
| "loss": 0.2575, | |
| "step": 31275 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 9.383683683683685e-06, | |
| "loss": 0.2526, | |
| "step": 31300 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 9.383183183183184e-06, | |
| "loss": 0.2678, | |
| "step": 31325 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 9.382682682682684e-06, | |
| "loss": 0.2711, | |
| "step": 31350 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 9.382182182182183e-06, | |
| "loss": 0.2889, | |
| "step": 31375 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 9.381681681681682e-06, | |
| "loss": 0.2709, | |
| "step": 31400 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 9.381181181181182e-06, | |
| "loss": 0.2736, | |
| "step": 31425 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 9.38068068068068e-06, | |
| "loss": 0.2699, | |
| "step": 31450 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 9.38018018018018e-06, | |
| "loss": 0.2801, | |
| "step": 31475 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 9.379679679679681e-06, | |
| "loss": 0.2774, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 9.37917917917918e-06, | |
| "loss": 0.2658, | |
| "step": 31525 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 9.37867867867868e-06, | |
| "loss": 0.2663, | |
| "step": 31550 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 9.378178178178179e-06, | |
| "loss": 0.2673, | |
| "step": 31575 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 9.377677677677678e-06, | |
| "loss": 0.2722, | |
| "step": 31600 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 9.377177177177178e-06, | |
| "loss": 0.2818, | |
| "step": 31625 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 9.376676676676677e-06, | |
| "loss": 0.2689, | |
| "step": 31650 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 9.376176176176178e-06, | |
| "loss": 0.2755, | |
| "step": 31675 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 9.375675675675675e-06, | |
| "loss": 0.2563, | |
| "step": 31700 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 9.375175175175176e-06, | |
| "loss": 0.2691, | |
| "step": 31725 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 9.374674674674676e-06, | |
| "loss": 0.2775, | |
| "step": 31750 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 9.374174174174175e-06, | |
| "loss": 0.2625, | |
| "step": 31775 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 9.373673673673674e-06, | |
| "loss": 0.2526, | |
| "step": 31800 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 9.373173173173174e-06, | |
| "loss": 0.2624, | |
| "step": 31825 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 9.372672672672673e-06, | |
| "loss": 0.2823, | |
| "step": 31850 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 9.372172172172172e-06, | |
| "loss": 0.2637, | |
| "step": 31875 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 9.371671671671673e-06, | |
| "loss": 0.2747, | |
| "step": 31900 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 9.371171171171173e-06, | |
| "loss": 0.2795, | |
| "step": 31925 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 9.370670670670672e-06, | |
| "loss": 0.2643, | |
| "step": 31950 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 9.370170170170171e-06, | |
| "loss": 0.2799, | |
| "step": 31975 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 9.36966966966967e-06, | |
| "loss": 0.2612, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "eval_loss": 0.4532637298107147, | |
| "eval_runtime": 2555.128, | |
| "eval_samples_per_second": 6.345, | |
| "eval_steps_per_second": 0.159, | |
| "eval_wer": 25.984998573862367, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 9.36916916916917e-06, | |
| "loss": 0.2583, | |
| "step": 32025 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 9.368668668668669e-06, | |
| "loss": 0.257, | |
| "step": 32050 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 9.368168168168168e-06, | |
| "loss": 0.2726, | |
| "step": 32075 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 9.367667667667668e-06, | |
| "loss": 0.2675, | |
| "step": 32100 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 9.367167167167169e-06, | |
| "loss": 0.2606, | |
| "step": 32125 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 9.366666666666668e-06, | |
| "loss": 0.2672, | |
| "step": 32150 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 9.366166166166167e-06, | |
| "loss": 0.2622, | |
| "step": 32175 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 9.365665665665666e-06, | |
| "loss": 0.2592, | |
| "step": 32200 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 9.365165165165166e-06, | |
| "loss": 0.2618, | |
| "step": 32225 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 9.364664664664665e-06, | |
| "loss": 0.2669, | |
| "step": 32250 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 9.364164164164164e-06, | |
| "loss": 0.2725, | |
| "step": 32275 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 9.363663663663665e-06, | |
| "loss": 0.2695, | |
| "step": 32300 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 9.363163163163163e-06, | |
| "loss": 0.2727, | |
| "step": 32325 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 9.362662662662664e-06, | |
| "loss": 0.28, | |
| "step": 32350 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 9.362162162162163e-06, | |
| "loss": 0.2826, | |
| "step": 32375 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 9.361661661661663e-06, | |
| "loss": 0.2728, | |
| "step": 32400 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 9.361161161161162e-06, | |
| "loss": 0.2653, | |
| "step": 32425 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 9.360660660660661e-06, | |
| "loss": 0.2655, | |
| "step": 32450 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 9.36016016016016e-06, | |
| "loss": 0.2787, | |
| "step": 32475 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 9.35965965965966e-06, | |
| "loss": 0.2677, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 9.35915915915916e-06, | |
| "loss": 0.2587, | |
| "step": 32525 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 9.358658658658658e-06, | |
| "loss": 0.2685, | |
| "step": 32550 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 9.35815815815816e-06, | |
| "loss": 0.2583, | |
| "step": 32575 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 9.357657657657659e-06, | |
| "loss": 0.2673, | |
| "step": 32600 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 9.357157157157158e-06, | |
| "loss": 0.2828, | |
| "step": 32625 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 9.356656656656657e-06, | |
| "loss": 0.2753, | |
| "step": 32650 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 9.356156156156157e-06, | |
| "loss": 0.2847, | |
| "step": 32675 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 9.355655655655656e-06, | |
| "loss": 0.2671, | |
| "step": 32700 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 9.355155155155155e-06, | |
| "loss": 0.2655, | |
| "step": 32725 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 9.354654654654656e-06, | |
| "loss": 0.2602, | |
| "step": 32750 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 9.354154154154155e-06, | |
| "loss": 0.2707, | |
| "step": 32775 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 9.353653653653655e-06, | |
| "loss": 0.2577, | |
| "step": 32800 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 9.353153153153154e-06, | |
| "loss": 0.2786, | |
| "step": 32825 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 9.352652652652653e-06, | |
| "loss": 0.2529, | |
| "step": 32850 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 9.352152152152153e-06, | |
| "loss": 0.2544, | |
| "step": 32875 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 9.351651651651652e-06, | |
| "loss": 0.2748, | |
| "step": 32900 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 9.351151151151153e-06, | |
| "loss": 0.2696, | |
| "step": 32925 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 9.35065065065065e-06, | |
| "loss": 0.2646, | |
| "step": 32950 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 9.350150150150151e-06, | |
| "loss": 0.2598, | |
| "step": 32975 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 9.34964964964965e-06, | |
| "loss": 0.2551, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "eval_loss": 0.4496789276599884, | |
| "eval_runtime": 2555.1297, | |
| "eval_samples_per_second": 6.345, | |
| "eval_steps_per_second": 0.159, | |
| "eval_wer": 26.183886957393177, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 9.34914914914915e-06, | |
| "loss": 0.2877, | |
| "step": 33025 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 9.34864864864865e-06, | |
| "loss": 0.2536, | |
| "step": 33050 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 9.348148148148149e-06, | |
| "loss": 0.2592, | |
| "step": 33075 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 9.347647647647648e-06, | |
| "loss": 0.2691, | |
| "step": 33100 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 9.347147147147147e-06, | |
| "loss": 0.269, | |
| "step": 33125 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 9.346646646646648e-06, | |
| "loss": 0.2585, | |
| "step": 33150 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 9.346146146146146e-06, | |
| "loss": 0.2547, | |
| "step": 33175 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 9.345645645645647e-06, | |
| "loss": 0.278, | |
| "step": 33200 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 9.345145145145146e-06, | |
| "loss": 0.283, | |
| "step": 33225 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 9.344644644644645e-06, | |
| "loss": 0.2708, | |
| "step": 33250 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 9.344144144144145e-06, | |
| "loss": 0.2549, | |
| "step": 33275 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 9.343643643643644e-06, | |
| "loss": 0.2601, | |
| "step": 33300 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 9.343143143143143e-06, | |
| "loss": 0.2585, | |
| "step": 33325 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 9.342642642642643e-06, | |
| "loss": 0.2794, | |
| "step": 33350 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 9.342142142142144e-06, | |
| "loss": 0.2721, | |
| "step": 33375 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 9.341641641641643e-06, | |
| "loss": 0.278, | |
| "step": 33400 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 9.341141141141142e-06, | |
| "loss": 0.2702, | |
| "step": 33425 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 9.340640640640642e-06, | |
| "loss": 0.2577, | |
| "step": 33450 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 9.34014014014014e-06, | |
| "loss": 0.2706, | |
| "step": 33475 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 9.33963963963964e-06, | |
| "loss": 0.25, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 9.33913913913914e-06, | |
| "loss": 0.2774, | |
| "step": 33525 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 9.33863863863864e-06, | |
| "loss": 0.2676, | |
| "step": 33550 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 9.338138138138138e-06, | |
| "loss": 0.2776, | |
| "step": 33575 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 9.337637637637639e-06, | |
| "loss": 0.2542, | |
| "step": 33600 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 9.337137137137138e-06, | |
| "loss": 0.273, | |
| "step": 33625 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 9.336636636636638e-06, | |
| "loss": 0.2498, | |
| "step": 33650 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 9.336136136136137e-06, | |
| "loss": 0.2624, | |
| "step": 33675 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 9.335635635635636e-06, | |
| "loss": 0.255, | |
| "step": 33700 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 9.335135135135135e-06, | |
| "loss": 0.2639, | |
| "step": 33725 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 9.334634634634635e-06, | |
| "loss": 0.281, | |
| "step": 33750 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 9.334134134134136e-06, | |
| "loss": 0.2733, | |
| "step": 33775 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 9.333633633633633e-06, | |
| "loss": 0.2613, | |
| "step": 33800 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 9.333133133133134e-06, | |
| "loss": 0.273, | |
| "step": 33825 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 9.332632632632634e-06, | |
| "loss": 0.264, | |
| "step": 33850 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 9.332132132132133e-06, | |
| "loss": 0.2665, | |
| "step": 33875 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 9.331631631631632e-06, | |
| "loss": 0.264, | |
| "step": 33900 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 9.331131131131132e-06, | |
| "loss": 0.2686, | |
| "step": 33925 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 9.330630630630631e-06, | |
| "loss": 0.2709, | |
| "step": 33950 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 9.33013013013013e-06, | |
| "loss": 0.2731, | |
| "step": 33975 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 9.329629629629631e-06, | |
| "loss": 0.28, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "eval_loss": 0.44749680161476135, | |
| "eval_runtime": 2532.2254, | |
| "eval_samples_per_second": 6.403, | |
| "eval_steps_per_second": 0.16, | |
| "eval_wer": 26.161531286376142, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 9.32912912912913e-06, | |
| "loss": 0.2572, | |
| "step": 34025 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 9.32862862862863e-06, | |
| "loss": 0.2828, | |
| "step": 34050 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 9.328128128128129e-06, | |
| "loss": 0.268, | |
| "step": 34075 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 9.327627627627628e-06, | |
| "loss": 0.2602, | |
| "step": 34100 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 9.327127127127128e-06, | |
| "loss": 0.2651, | |
| "step": 34125 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 9.326626626626627e-06, | |
| "loss": 0.2796, | |
| "step": 34150 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 9.326126126126128e-06, | |
| "loss": 0.2649, | |
| "step": 34175 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 9.325625625625626e-06, | |
| "loss": 0.2729, | |
| "step": 34200 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 9.325125125125127e-06, | |
| "loss": 0.2492, | |
| "step": 34225 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 9.324624624624626e-06, | |
| "loss": 0.2389, | |
| "step": 34250 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 9.324124124124125e-06, | |
| "loss": 0.2584, | |
| "step": 34275 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 9.323623623623624e-06, | |
| "loss": 0.2635, | |
| "step": 34300 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 9.323123123123124e-06, | |
| "loss": 0.2556, | |
| "step": 34325 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 9.322622622622623e-06, | |
| "loss": 0.2661, | |
| "step": 34350 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 9.322122122122122e-06, | |
| "loss": 0.2698, | |
| "step": 34375 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 9.321621621621623e-06, | |
| "loss": 0.2622, | |
| "step": 34400 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 9.321121121121121e-06, | |
| "loss": 0.2606, | |
| "step": 34425 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 9.320620620620622e-06, | |
| "loss": 0.2561, | |
| "step": 34450 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 9.320120120120121e-06, | |
| "loss": 0.2681, | |
| "step": 34475 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 9.31961961961962e-06, | |
| "loss": 0.2686, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 9.31911911911912e-06, | |
| "loss": 0.2691, | |
| "step": 34525 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 9.318618618618619e-06, | |
| "loss": 0.2481, | |
| "step": 34550 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 9.318118118118118e-06, | |
| "loss": 0.2579, | |
| "step": 34575 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 9.317617617617618e-06, | |
| "loss": 0.2442, | |
| "step": 34600 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 9.317117117117119e-06, | |
| "loss": 0.2605, | |
| "step": 34625 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 9.316616616616618e-06, | |
| "loss": 0.2717, | |
| "step": 34650 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 9.316116116116117e-06, | |
| "loss": 0.2749, | |
| "step": 34675 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 9.315615615615617e-06, | |
| "loss": 0.2525, | |
| "step": 34700 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 9.315115115115116e-06, | |
| "loss": 0.2586, | |
| "step": 34725 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 9.314614614614615e-06, | |
| "loss": 0.2683, | |
| "step": 34750 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 9.314114114114114e-06, | |
| "loss": 0.2538, | |
| "step": 34775 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 9.313613613613614e-06, | |
| "loss": 0.2354, | |
| "step": 34800 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 9.313113113113113e-06, | |
| "loss": 0.2591, | |
| "step": 34825 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 9.312612612612614e-06, | |
| "loss": 0.28, | |
| "step": 34850 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 9.312112112112113e-06, | |
| "loss": 0.2789, | |
| "step": 34875 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 9.311611611611613e-06, | |
| "loss": 0.2741, | |
| "step": 34900 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 9.311111111111112e-06, | |
| "loss": 0.2803, | |
| "step": 34925 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 9.310610610610611e-06, | |
| "loss": 0.2699, | |
| "step": 34950 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 9.31011011011011e-06, | |
| "loss": 0.2618, | |
| "step": 34975 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 9.30960960960961e-06, | |
| "loss": 0.2775, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "eval_loss": 0.44763508439064026, | |
| "eval_runtime": 2480.936, | |
| "eval_samples_per_second": 6.535, | |
| "eval_steps_per_second": 0.164, | |
| "eval_wer": 25.96958086971269, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 9.30910910910911e-06, | |
| "loss": 0.2801, | |
| "step": 35025 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 9.308608608608608e-06, | |
| "loss": 0.2666, | |
| "step": 35050 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 9.30810810810811e-06, | |
| "loss": 0.2817, | |
| "step": 35075 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 9.307607607607609e-06, | |
| "loss": 0.2618, | |
| "step": 35100 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 9.307107107107108e-06, | |
| "loss": 0.2637, | |
| "step": 35125 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 9.306606606606607e-06, | |
| "loss": 0.2679, | |
| "step": 35150 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 9.306106106106107e-06, | |
| "loss": 0.2729, | |
| "step": 35175 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 9.305605605605606e-06, | |
| "loss": 0.2659, | |
| "step": 35200 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 9.305105105105105e-06, | |
| "loss": 0.2642, | |
| "step": 35225 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 9.304604604604606e-06, | |
| "loss": 0.2607, | |
| "step": 35250 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 9.304104104104105e-06, | |
| "loss": 0.2831, | |
| "step": 35275 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 9.303603603603605e-06, | |
| "loss": 0.2681, | |
| "step": 35300 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 9.303103103103104e-06, | |
| "loss": 0.2611, | |
| "step": 35325 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 9.302602602602603e-06, | |
| "loss": 0.2525, | |
| "step": 35350 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 9.302102102102103e-06, | |
| "loss": 0.2629, | |
| "step": 35375 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 9.301601601601602e-06, | |
| "loss": 0.2576, | |
| "step": 35400 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 9.301101101101101e-06, | |
| "loss": 0.2576, | |
| "step": 35425 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 9.3006006006006e-06, | |
| "loss": 0.2558, | |
| "step": 35450 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 9.300100100100102e-06, | |
| "loss": 0.2591, | |
| "step": 35475 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 9.299599599599601e-06, | |
| "loss": 0.2646, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 9.2990990990991e-06, | |
| "loss": 0.2515, | |
| "step": 35525 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 9.2985985985986e-06, | |
| "loss": 0.274, | |
| "step": 35550 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 9.298098098098099e-06, | |
| "loss": 0.2696, | |
| "step": 35575 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 9.297597597597598e-06, | |
| "loss": 0.2637, | |
| "step": 35600 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 9.297097097097097e-06, | |
| "loss": 0.2644, | |
| "step": 35625 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 9.296596596596598e-06, | |
| "loss": 0.262, | |
| "step": 35650 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 9.296096096096096e-06, | |
| "loss": 0.2709, | |
| "step": 35675 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 9.295595595595597e-06, | |
| "loss": 0.263, | |
| "step": 35700 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 9.295095095095096e-06, | |
| "loss": 0.2569, | |
| "step": 35725 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 9.294594594594596e-06, | |
| "loss": 0.2525, | |
| "step": 35750 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 9.294094094094095e-06, | |
| "loss": 0.2584, | |
| "step": 35775 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 9.293593593593594e-06, | |
| "loss": 0.2665, | |
| "step": 35800 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 9.293093093093093e-06, | |
| "loss": 0.2693, | |
| "step": 35825 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 9.292592592592593e-06, | |
| "loss": 0.257, | |
| "step": 35850 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 9.292092092092094e-06, | |
| "loss": 0.2616, | |
| "step": 35875 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 9.291591591591593e-06, | |
| "loss": 0.2711, | |
| "step": 35900 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 9.291091091091092e-06, | |
| "loss": 0.2573, | |
| "step": 35925 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 9.290590590590592e-06, | |
| "loss": 0.2708, | |
| "step": 35950 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 9.290090090090091e-06, | |
| "loss": 0.2423, | |
| "step": 35975 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 9.28958958958959e-06, | |
| "loss": 0.2609, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "eval_loss": 0.4396921396255493, | |
| "eval_runtime": 2445.1333, | |
| "eval_samples_per_second": 6.631, | |
| "eval_steps_per_second": 0.166, | |
| "eval_wer": 25.296598083579376, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 9.28908908908909e-06, | |
| "loss": 0.2572, | |
| "step": 36025 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 9.288588588588589e-06, | |
| "loss": 0.2676, | |
| "step": 36050 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 9.288088088088088e-06, | |
| "loss": 0.2508, | |
| "step": 36075 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 9.287587587587589e-06, | |
| "loss": 0.2714, | |
| "step": 36100 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 9.287087087087088e-06, | |
| "loss": 0.26, | |
| "step": 36125 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 9.286586586586588e-06, | |
| "loss": 0.2648, | |
| "step": 36150 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 9.286086086086087e-06, | |
| "loss": 0.2748, | |
| "step": 36175 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 9.285585585585586e-06, | |
| "loss": 0.2602, | |
| "step": 36200 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 9.285085085085086e-06, | |
| "loss": 0.2658, | |
| "step": 36225 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 9.284584584584585e-06, | |
| "loss": 0.2623, | |
| "step": 36250 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 9.284084084084086e-06, | |
| "loss": 0.2634, | |
| "step": 36275 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 9.283583583583583e-06, | |
| "loss": 0.2832, | |
| "step": 36300 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 9.283083083083084e-06, | |
| "loss": 0.2656, | |
| "step": 36325 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 9.282582582582584e-06, | |
| "loss": 0.2565, | |
| "step": 36350 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 9.282082082082083e-06, | |
| "loss": 0.2542, | |
| "step": 36375 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 9.281581581581582e-06, | |
| "loss": 0.2594, | |
| "step": 36400 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 9.281081081081082e-06, | |
| "loss": 0.2654, | |
| "step": 36425 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 9.280580580580581e-06, | |
| "loss": 0.292, | |
| "step": 36450 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 9.28008008008008e-06, | |
| "loss": 0.282, | |
| "step": 36475 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 9.279579579579581e-06, | |
| "loss": 0.2795, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 9.27907907907908e-06, | |
| "loss": 0.263, | |
| "step": 36525 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 9.27857857857858e-06, | |
| "loss": 0.272, | |
| "step": 36550 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 9.278098098098099e-06, | |
| "loss": 0.2498, | |
| "step": 36575 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 9.277597597597598e-06, | |
| "loss": 0.2626, | |
| "step": 36600 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 9.277097097097098e-06, | |
| "loss": 0.2671, | |
| "step": 36625 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 9.276596596596597e-06, | |
| "loss": 0.2622, | |
| "step": 36650 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 9.276096096096096e-06, | |
| "loss": 0.2507, | |
| "step": 36675 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 9.275595595595597e-06, | |
| "loss": 0.2606, | |
| "step": 36700 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 9.275095095095096e-06, | |
| "loss": 0.2716, | |
| "step": 36725 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 9.274594594594596e-06, | |
| "loss": 0.2699, | |
| "step": 36750 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 9.274094094094095e-06, | |
| "loss": 0.2819, | |
| "step": 36775 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 9.273593593593594e-06, | |
| "loss": 0.2681, | |
| "step": 36800 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 9.273093093093094e-06, | |
| "loss": 0.2603, | |
| "step": 36825 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 9.272592592592593e-06, | |
| "loss": 0.2556, | |
| "step": 36850 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 9.272092092092094e-06, | |
| "loss": 0.2798, | |
| "step": 36875 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 9.271591591591592e-06, | |
| "loss": 0.2623, | |
| "step": 36900 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 9.271091091091092e-06, | |
| "loss": 0.2502, | |
| "step": 36925 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 9.270590590590592e-06, | |
| "loss": 0.262, | |
| "step": 36950 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 9.270090090090091e-06, | |
| "loss": 0.2504, | |
| "step": 36975 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 9.26958958958959e-06, | |
| "loss": 0.2681, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "eval_loss": 0.4397401511669159, | |
| "eval_runtime": 2578.9596, | |
| "eval_samples_per_second": 6.287, | |
| "eval_steps_per_second": 0.157, | |
| "eval_wer": 26.684962342257613, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 9.26908908908909e-06, | |
| "loss": 0.2618, | |
| "step": 37025 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 9.268588588588589e-06, | |
| "loss": 0.2717, | |
| "step": 37050 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 9.268088088088088e-06, | |
| "loss": 0.2628, | |
| "step": 37075 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 9.26758758758759e-06, | |
| "loss": 0.262, | |
| "step": 37100 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 9.267087087087087e-06, | |
| "loss": 0.2531, | |
| "step": 37125 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 9.266586586586588e-06, | |
| "loss": 0.2407, | |
| "step": 37150 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 9.266086086086087e-06, | |
| "loss": 0.2717, | |
| "step": 37175 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 9.265585585585586e-06, | |
| "loss": 0.2696, | |
| "step": 37200 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 9.265085085085086e-06, | |
| "loss": 0.2718, | |
| "step": 37225 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 9.264584584584585e-06, | |
| "loss": 0.2523, | |
| "step": 37250 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 9.264084084084084e-06, | |
| "loss": 0.2637, | |
| "step": 37275 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 9.263583583583584e-06, | |
| "loss": 0.2818, | |
| "step": 37300 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 9.263083083083085e-06, | |
| "loss": 0.2812, | |
| "step": 37325 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 9.262582582582584e-06, | |
| "loss": 0.2623, | |
| "step": 37350 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 9.262082082082083e-06, | |
| "loss": 0.2615, | |
| "step": 37375 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 9.261581581581583e-06, | |
| "loss": 0.2783, | |
| "step": 37400 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 9.261081081081082e-06, | |
| "loss": 0.2577, | |
| "step": 37425 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 9.260580580580581e-06, | |
| "loss": 0.2718, | |
| "step": 37450 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 9.26008008008008e-06, | |
| "loss": 0.2569, | |
| "step": 37475 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 9.259579579579581e-06, | |
| "loss": 0.2604, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 9.259079079079079e-06, | |
| "loss": 0.265, | |
| "step": 37525 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 9.25857857857858e-06, | |
| "loss": 0.2688, | |
| "step": 37550 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 9.25807807807808e-06, | |
| "loss": 0.2567, | |
| "step": 37575 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 9.257577577577579e-06, | |
| "loss": 0.2747, | |
| "step": 37600 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 9.257077077077078e-06, | |
| "loss": 0.27, | |
| "step": 37625 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 9.256576576576577e-06, | |
| "loss": 0.2699, | |
| "step": 37650 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 9.256076076076076e-06, | |
| "loss": 0.2645, | |
| "step": 37675 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 9.255575575575576e-06, | |
| "loss": 0.2723, | |
| "step": 37700 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 9.255075075075077e-06, | |
| "loss": 0.2693, | |
| "step": 37725 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 9.254574574574574e-06, | |
| "loss": 0.2509, | |
| "step": 37750 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 9.254074074074075e-06, | |
| "loss": 0.2698, | |
| "step": 37775 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 9.253573573573575e-06, | |
| "loss": 0.2359, | |
| "step": 37800 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 9.253073073073074e-06, | |
| "loss": 0.2707, | |
| "step": 37825 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 9.252572572572573e-06, | |
| "loss": 0.2534, | |
| "step": 37850 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 9.252072072072073e-06, | |
| "loss": 0.2554, | |
| "step": 37875 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 9.251571571571572e-06, | |
| "loss": 0.2753, | |
| "step": 37900 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 9.251071071071071e-06, | |
| "loss": 0.2551, | |
| "step": 37925 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 9.250570570570572e-06, | |
| "loss": 0.2576, | |
| "step": 37950 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 9.250070070070071e-06, | |
| "loss": 0.2579, | |
| "step": 37975 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 9.24956956956957e-06, | |
| "loss": 0.2661, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "eval_loss": 0.4426652491092682, | |
| "eval_runtime": 2522.4342, | |
| "eval_samples_per_second": 6.428, | |
| "eval_steps_per_second": 0.161, | |
| "eval_wer": 25.423023257606708, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 9.24906906906907e-06, | |
| "loss": 0.2819, | |
| "step": 38025 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 9.24856856856857e-06, | |
| "loss": 0.2773, | |
| "step": 38050 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 9.248068068068069e-06, | |
| "loss": 0.2604, | |
| "step": 38075 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 9.247567567567568e-06, | |
| "loss": 0.2645, | |
| "step": 38100 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 9.247067067067069e-06, | |
| "loss": 0.2504, | |
| "step": 38125 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 9.246566566566567e-06, | |
| "loss": 0.2731, | |
| "step": 38150 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 9.246066066066068e-06, | |
| "loss": 0.2543, | |
| "step": 38175 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 9.245565565565567e-06, | |
| "loss": 0.2552, | |
| "step": 38200 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 9.245065065065066e-06, | |
| "loss": 0.2789, | |
| "step": 38225 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.244564564564565e-06, | |
| "loss": 0.2668, | |
| "step": 38250 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.244064064064065e-06, | |
| "loss": 0.2635, | |
| "step": 38275 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.243563563563564e-06, | |
| "loss": 0.248, | |
| "step": 38300 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.243063063063063e-06, | |
| "loss": 0.2675, | |
| "step": 38325 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.242562562562564e-06, | |
| "loss": 0.265, | |
| "step": 38350 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.242062062062062e-06, | |
| "loss": 0.2509, | |
| "step": 38375 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.241561561561563e-06, | |
| "loss": 0.2535, | |
| "step": 38400 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.241061061061062e-06, | |
| "loss": 0.2716, | |
| "step": 38425 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.240560560560561e-06, | |
| "loss": 0.2578, | |
| "step": 38450 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.24006006006006e-06, | |
| "loss": 0.2743, | |
| "step": 38475 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 9.23955955955956e-06, | |
| "loss": 0.2701, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 9.23905905905906e-06, | |
| "loss": 0.2716, | |
| "step": 38525 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 9.238558558558559e-06, | |
| "loss": 0.2696, | |
| "step": 38550 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 9.23805805805806e-06, | |
| "loss": 0.2701, | |
| "step": 38575 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 9.237577577577577e-06, | |
| "loss": 0.2573, | |
| "step": 38600 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 9.237077077077078e-06, | |
| "loss": 0.2641, | |
| "step": 38625 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 9.236576576576577e-06, | |
| "loss": 0.277, | |
| "step": 38650 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 9.236076076076077e-06, | |
| "loss": 0.2562, | |
| "step": 38675 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 9.235575575575576e-06, | |
| "loss": 0.2584, | |
| "step": 38700 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 9.235075075075077e-06, | |
| "loss": 0.2571, | |
| "step": 38725 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.234574574574575e-06, | |
| "loss": 0.2652, | |
| "step": 38750 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.234074074074076e-06, | |
| "loss": 0.26, | |
| "step": 38775 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.233573573573575e-06, | |
| "loss": 0.2475, | |
| "step": 38800 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.233073073073074e-06, | |
| "loss": 0.2728, | |
| "step": 38825 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.232572572572573e-06, | |
| "loss": 0.2678, | |
| "step": 38850 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.232072072072073e-06, | |
| "loss": 0.2597, | |
| "step": 38875 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.231571571571572e-06, | |
| "loss": 0.2532, | |
| "step": 38900 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.231071071071071e-06, | |
| "loss": 0.2552, | |
| "step": 38925 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.230570570570572e-06, | |
| "loss": 0.2528, | |
| "step": 38950 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.23007007007007e-06, | |
| "loss": 0.2356, | |
| "step": 38975 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.229569569569571e-06, | |
| "loss": 0.2512, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "eval_loss": 0.4447844326496124, | |
| "eval_runtime": 2535.3087, | |
| "eval_samples_per_second": 6.395, | |
| "eval_steps_per_second": 0.16, | |
| "eval_wer": 26.18311607218569, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.22906906906907e-06, | |
| "loss": 0.28, | |
| "step": 39025 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.22856856856857e-06, | |
| "loss": 0.2598, | |
| "step": 39050 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.228068068068069e-06, | |
| "loss": 0.2805, | |
| "step": 39075 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.227567567567568e-06, | |
| "loss": 0.2677, | |
| "step": 39100 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.227067067067067e-06, | |
| "loss": 0.2749, | |
| "step": 39125 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.226566566566567e-06, | |
| "loss": 0.2652, | |
| "step": 39150 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.226066066066068e-06, | |
| "loss": 0.2609, | |
| "step": 39175 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.225565565565567e-06, | |
| "loss": 0.2656, | |
| "step": 39200 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.225065065065065e-06, | |
| "loss": 0.2648, | |
| "step": 39225 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.224564564564566e-06, | |
| "loss": 0.2567, | |
| "step": 39250 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.224064064064065e-06, | |
| "loss": 0.2433, | |
| "step": 39275 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.223563563563564e-06, | |
| "loss": 0.276, | |
| "step": 39300 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.223063063063064e-06, | |
| "loss": 0.2651, | |
| "step": 39325 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.222562562562564e-06, | |
| "loss": 0.2591, | |
| "step": 39350 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.222062062062062e-06, | |
| "loss": 0.2616, | |
| "step": 39375 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.221561561561563e-06, | |
| "loss": 0.2563, | |
| "step": 39400 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.221061061061062e-06, | |
| "loss": 0.2595, | |
| "step": 39425 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.22056056056056e-06, | |
| "loss": 0.2646, | |
| "step": 39450 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.220060060060061e-06, | |
| "loss": 0.2676, | |
| "step": 39475 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.21955955955956e-06, | |
| "loss": 0.2585, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.21905905905906e-06, | |
| "loss": 0.2599, | |
| "step": 39525 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.218558558558559e-06, | |
| "loss": 0.2477, | |
| "step": 39550 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.21805805805806e-06, | |
| "loss": 0.2376, | |
| "step": 39575 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.217557557557557e-06, | |
| "loss": 0.2491, | |
| "step": 39600 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.217057057057058e-06, | |
| "loss": 0.2533, | |
| "step": 39625 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.216556556556558e-06, | |
| "loss": 0.2646, | |
| "step": 39650 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.216056056056057e-06, | |
| "loss": 0.2824, | |
| "step": 39675 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.215555555555556e-06, | |
| "loss": 0.2854, | |
| "step": 39700 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.215055055055056e-06, | |
| "loss": 0.2553, | |
| "step": 39725 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.214554554554555e-06, | |
| "loss": 0.2484, | |
| "step": 39750 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.214054054054054e-06, | |
| "loss": 0.2362, | |
| "step": 39775 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.213553553553555e-06, | |
| "loss": 0.2634, | |
| "step": 39800 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.213053053053055e-06, | |
| "loss": 0.2554, | |
| "step": 39825 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.212552552552552e-06, | |
| "loss": 0.2502, | |
| "step": 39850 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.212052052052053e-06, | |
| "loss": 0.2546, | |
| "step": 39875 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.211551551551552e-06, | |
| "loss": 0.2723, | |
| "step": 39900 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.211051051051052e-06, | |
| "loss": 0.2674, | |
| "step": 39925 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.210550550550551e-06, | |
| "loss": 0.2688, | |
| "step": 39950 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.21005005005005e-06, | |
| "loss": 0.2512, | |
| "step": 39975 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.20954954954955e-06, | |
| "loss": 0.2541, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "eval_loss": 0.4458891749382019, | |
| "eval_runtime": 2617.6449, | |
| "eval_samples_per_second": 6.194, | |
| "eval_steps_per_second": 0.155, | |
| "eval_wer": 26.314937442665414, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.20904904904905e-06, | |
| "loss": 0.2547, | |
| "step": 40025 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.20854854854855e-06, | |
| "loss": 0.2557, | |
| "step": 40050 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.208048048048048e-06, | |
| "loss": 0.2526, | |
| "step": 40075 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.207547547547549e-06, | |
| "loss": 0.2709, | |
| "step": 40100 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.207047047047048e-06, | |
| "loss": 0.2643, | |
| "step": 40125 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.206546546546547e-06, | |
| "loss": 0.2716, | |
| "step": 40150 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.206046046046046e-06, | |
| "loss": 0.2725, | |
| "step": 40175 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.205545545545547e-06, | |
| "loss": 0.2532, | |
| "step": 40200 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.205045045045045e-06, | |
| "loss": 0.2595, | |
| "step": 40225 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.204544544544546e-06, | |
| "loss": 0.2619, | |
| "step": 40250 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.204044044044045e-06, | |
| "loss": 0.2627, | |
| "step": 40275 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.203543543543545e-06, | |
| "loss": 0.2581, | |
| "step": 40300 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.203043043043044e-06, | |
| "loss": 0.2634, | |
| "step": 40325 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.202542542542543e-06, | |
| "loss": 0.2392, | |
| "step": 40350 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.202042042042042e-06, | |
| "loss": 0.2694, | |
| "step": 40375 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.201541541541542e-06, | |
| "loss": 0.2596, | |
| "step": 40400 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.201041041041043e-06, | |
| "loss": 0.2844, | |
| "step": 40425 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.200540540540542e-06, | |
| "loss": 0.2679, | |
| "step": 40450 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.20004004004004e-06, | |
| "loss": 0.2592, | |
| "step": 40475 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.19953953953954e-06, | |
| "loss": 0.2523, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.19903903903904e-06, | |
| "loss": 0.27, | |
| "step": 40525 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.19853853853854e-06, | |
| "loss": 0.272, | |
| "step": 40550 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.198038038038039e-06, | |
| "loss": 0.2621, | |
| "step": 40575 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.197537537537538e-06, | |
| "loss": 0.262, | |
| "step": 40600 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.197057057057059e-06, | |
| "loss": 0.2642, | |
| "step": 40625 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.196556556556558e-06, | |
| "loss": 0.278, | |
| "step": 40650 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.196056056056056e-06, | |
| "loss": 0.2711, | |
| "step": 40675 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.195555555555557e-06, | |
| "loss": 0.2728, | |
| "step": 40700 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.195055055055056e-06, | |
| "loss": 0.2844, | |
| "step": 40725 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.194554554554555e-06, | |
| "loss": 0.2657, | |
| "step": 40750 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.194054054054054e-06, | |
| "loss": 0.2559, | |
| "step": 40775 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.193553553553555e-06, | |
| "loss": 0.2671, | |
| "step": 40800 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.193053053053053e-06, | |
| "loss": 0.2559, | |
| "step": 40825 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.192552552552554e-06, | |
| "loss": 0.2546, | |
| "step": 40850 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.192052052052053e-06, | |
| "loss": 0.2607, | |
| "step": 40875 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.191551551551553e-06, | |
| "loss": 0.2718, | |
| "step": 40900 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.191051051051052e-06, | |
| "loss": 0.2672, | |
| "step": 40925 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.190550550550551e-06, | |
| "loss": 0.2682, | |
| "step": 40950 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.19005005005005e-06, | |
| "loss": 0.2646, | |
| "step": 40975 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.18954954954955e-06, | |
| "loss": 0.2535, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "eval_loss": 0.44114458560943604, | |
| "eval_runtime": 2473.7687, | |
| "eval_samples_per_second": 6.554, | |
| "eval_steps_per_second": 0.164, | |
| "eval_wer": 25.31355755814402, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.18904904904905e-06, | |
| "loss": 0.2634, | |
| "step": 41025 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.18854854854855e-06, | |
| "loss": 0.2616, | |
| "step": 41050 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.188048048048048e-06, | |
| "loss": 0.266, | |
| "step": 41075 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.187547547547549e-06, | |
| "loss": 0.264, | |
| "step": 41100 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.187047047047048e-06, | |
| "loss": 0.2604, | |
| "step": 41125 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.186546546546547e-06, | |
| "loss": 0.2552, | |
| "step": 41150 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.186046046046047e-06, | |
| "loss": 0.2539, | |
| "step": 41175 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.185545545545546e-06, | |
| "loss": 0.2854, | |
| "step": 41200 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.185045045045045e-06, | |
| "loss": 0.2629, | |
| "step": 41225 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.184544544544546e-06, | |
| "loss": 0.2613, | |
| "step": 41250 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 9.184044044044045e-06, | |
| "loss": 0.2561, | |
| "step": 41275 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 9.183543543543543e-06, | |
| "loss": 0.2617, | |
| "step": 41300 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 9.183043043043044e-06, | |
| "loss": 0.2573, | |
| "step": 41325 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 9.182542542542543e-06, | |
| "loss": 0.2622, | |
| "step": 41350 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 9.182042042042043e-06, | |
| "loss": 0.2472, | |
| "step": 41375 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 9.181541541541542e-06, | |
| "loss": 0.262, | |
| "step": 41400 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 9.181041041041043e-06, | |
| "loss": 0.2558, | |
| "step": 41425 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 9.18054054054054e-06, | |
| "loss": 0.2572, | |
| "step": 41450 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 9.180040040040042e-06, | |
| "loss": 0.2618, | |
| "step": 41475 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 9.17953953953954e-06, | |
| "loss": 0.2532, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 9.17903903903904e-06, | |
| "loss": 0.2522, | |
| "step": 41525 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 9.17853853853854e-06, | |
| "loss": 0.275, | |
| "step": 41550 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 9.178038038038039e-06, | |
| "loss": 0.2754, | |
| "step": 41575 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 9.177537537537538e-06, | |
| "loss": 0.2599, | |
| "step": 41600 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 9.177037037037037e-06, | |
| "loss": 0.2509, | |
| "step": 41625 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 9.176536536536538e-06, | |
| "loss": 0.2554, | |
| "step": 41650 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 9.176036036036036e-06, | |
| "loss": 0.2426, | |
| "step": 41675 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 9.175535535535535e-06, | |
| "loss": 0.2675, | |
| "step": 41700 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 9.175035035035036e-06, | |
| "loss": 0.2559, | |
| "step": 41725 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 9.174534534534536e-06, | |
| "loss": 0.2716, | |
| "step": 41750 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 9.174034034034035e-06, | |
| "loss": 0.2686, | |
| "step": 41775 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 9.173533533533534e-06, | |
| "loss": 0.259, | |
| "step": 41800 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 9.173033033033033e-06, | |
| "loss": 0.2747, | |
| "step": 41825 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 9.172532532532533e-06, | |
| "loss": 0.2507, | |
| "step": 41850 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 9.172032032032034e-06, | |
| "loss": 0.2573, | |
| "step": 41875 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 9.171531531531533e-06, | |
| "loss": 0.2542, | |
| "step": 41900 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 9.17103103103103e-06, | |
| "loss": 0.2584, | |
| "step": 41925 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 9.170530530530532e-06, | |
| "loss": 0.2459, | |
| "step": 41950 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 9.170030030030031e-06, | |
| "loss": 0.2533, | |
| "step": 41975 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 9.16952952952953e-06, | |
| "loss": 0.2518, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "eval_loss": 0.42200759053230286, | |
| "eval_runtime": 2416.12, | |
| "eval_samples_per_second": 6.71, | |
| "eval_steps_per_second": 0.168, | |
| "eval_wer": 24.71689240755159, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 9.16902902902903e-06, | |
| "loss": 0.2759, | |
| "step": 42025 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 9.16852852852853e-06, | |
| "loss": 0.258, | |
| "step": 42050 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 9.168028028028028e-06, | |
| "loss": 0.2424, | |
| "step": 42075 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 9.167527527527527e-06, | |
| "loss": 0.258, | |
| "step": 42100 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 9.167027027027028e-06, | |
| "loss": 0.274, | |
| "step": 42125 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 9.166526526526528e-06, | |
| "loss": 0.2479, | |
| "step": 42150 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 9.166026026026027e-06, | |
| "loss": 0.2697, | |
| "step": 42175 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 9.165525525525526e-06, | |
| "loss": 0.2621, | |
| "step": 42200 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 9.165025025025026e-06, | |
| "loss": 0.2645, | |
| "step": 42225 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 9.164524524524525e-06, | |
| "loss": 0.2624, | |
| "step": 42250 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 9.164024024024026e-06, | |
| "loss": 0.2592, | |
| "step": 42275 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 9.163523523523523e-06, | |
| "loss": 0.255, | |
| "step": 42300 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 9.163023023023023e-06, | |
| "loss": 0.262, | |
| "step": 42325 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 9.162522522522524e-06, | |
| "loss": 0.2687, | |
| "step": 42350 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 9.162022022022023e-06, | |
| "loss": 0.2566, | |
| "step": 42375 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 9.161521521521522e-06, | |
| "loss": 0.2563, | |
| "step": 42400 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 9.161021021021022e-06, | |
| "loss": 0.2559, | |
| "step": 42425 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 9.160520520520521e-06, | |
| "loss": 0.2641, | |
| "step": 42450 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 9.16002002002002e-06, | |
| "loss": 0.2729, | |
| "step": 42475 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 9.159519519519521e-06, | |
| "loss": 0.2531, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 9.15901901901902e-06, | |
| "loss": 0.2819, | |
| "step": 42525 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 9.158518518518518e-06, | |
| "loss": 0.2594, | |
| "step": 42550 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 9.158018018018019e-06, | |
| "loss": 0.2649, | |
| "step": 42575 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 9.157517517517518e-06, | |
| "loss": 0.2593, | |
| "step": 42600 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 9.157017017017018e-06, | |
| "loss": 0.2636, | |
| "step": 42625 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 9.156536536536539e-06, | |
| "loss": 0.2698, | |
| "step": 42650 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 9.156036036036036e-06, | |
| "loss": 0.254, | |
| "step": 42675 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 9.155535535535535e-06, | |
| "loss": 0.2356, | |
| "step": 42700 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 9.155035035035036e-06, | |
| "loss": 0.2506, | |
| "step": 42725 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 9.154534534534536e-06, | |
| "loss": 0.2669, | |
| "step": 42750 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 9.154034034034035e-06, | |
| "loss": 0.2576, | |
| "step": 42775 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 9.153533533533534e-06, | |
| "loss": 0.2536, | |
| "step": 42800 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 9.153033033033034e-06, | |
| "loss": 0.2686, | |
| "step": 42825 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 9.152532532532533e-06, | |
| "loss": 0.2511, | |
| "step": 42850 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 9.152032032032034e-06, | |
| "loss": 0.2667, | |
| "step": 42875 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 9.151531531531531e-06, | |
| "loss": 0.2593, | |
| "step": 42900 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 9.15103103103103e-06, | |
| "loss": 0.2656, | |
| "step": 42925 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 9.150530530530532e-06, | |
| "loss": 0.2574, | |
| "step": 42950 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 9.150030030030031e-06, | |
| "loss": 0.2557, | |
| "step": 42975 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 9.14952952952953e-06, | |
| "loss": 0.255, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "eval_loss": 0.43135425448417664, | |
| "eval_runtime": 2587.1846, | |
| "eval_samples_per_second": 6.267, | |
| "eval_steps_per_second": 0.157, | |
| "eval_wer": 25.259595593620155, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 9.14902902902903e-06, | |
| "loss": 0.2639, | |
| "step": 43025 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 9.148528528528529e-06, | |
| "loss": 0.256, | |
| "step": 43050 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 9.148028028028028e-06, | |
| "loss": 0.2588, | |
| "step": 43075 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 9.14752752752753e-06, | |
| "loss": 0.2768, | |
| "step": 43100 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 9.147027027027029e-06, | |
| "loss": 0.2557, | |
| "step": 43125 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 9.146526526526526e-06, | |
| "loss": 0.2506, | |
| "step": 43150 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 9.146026026026027e-06, | |
| "loss": 0.2495, | |
| "step": 43175 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 9.145525525525526e-06, | |
| "loss": 0.2652, | |
| "step": 43200 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 9.145025025025026e-06, | |
| "loss": 0.2525, | |
| "step": 43225 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 9.144524524524525e-06, | |
| "loss": 0.2694, | |
| "step": 43250 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 9.144024024024026e-06, | |
| "loss": 0.2489, | |
| "step": 43275 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 9.143523523523524e-06, | |
| "loss": 0.2531, | |
| "step": 43300 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 9.143023023023023e-06, | |
| "loss": 0.2549, | |
| "step": 43325 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 9.142522522522524e-06, | |
| "loss": 0.2526, | |
| "step": 43350 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 9.142022022022023e-06, | |
| "loss": 0.2604, | |
| "step": 43375 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 9.141521521521523e-06, | |
| "loss": 0.2683, | |
| "step": 43400 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 9.141021021021022e-06, | |
| "loss": 0.2659, | |
| "step": 43425 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 9.140520520520521e-06, | |
| "loss": 0.2569, | |
| "step": 43450 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 9.14002002002002e-06, | |
| "loss": 0.2556, | |
| "step": 43475 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 9.139519519519521e-06, | |
| "loss": 0.2752, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 9.139019019019019e-06, | |
| "loss": 0.2562, | |
| "step": 43525 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 9.138518518518518e-06, | |
| "loss": 0.2848, | |
| "step": 43550 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 9.13801801801802e-06, | |
| "loss": 0.2598, | |
| "step": 43575 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 9.137517517517519e-06, | |
| "loss": 0.2663, | |
| "step": 43600 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 9.137017017017018e-06, | |
| "loss": 0.2582, | |
| "step": 43625 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 9.136516516516517e-06, | |
| "loss": 0.2646, | |
| "step": 43650 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 9.136016016016016e-06, | |
| "loss": 0.2623, | |
| "step": 43675 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 9.135515515515516e-06, | |
| "loss": 0.2569, | |
| "step": 43700 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 9.135015015015017e-06, | |
| "loss": 0.2504, | |
| "step": 43725 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 9.134514514514516e-06, | |
| "loss": 0.2726, | |
| "step": 43750 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 9.134014014014014e-06, | |
| "loss": 0.2593, | |
| "step": 43775 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 9.133513513513515e-06, | |
| "loss": 0.2478, | |
| "step": 43800 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 9.133013013013014e-06, | |
| "loss": 0.2582, | |
| "step": 43825 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 9.132512512512513e-06, | |
| "loss": 0.2646, | |
| "step": 43850 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 9.132012012012013e-06, | |
| "loss": 0.2822, | |
| "step": 43875 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 9.131511511511514e-06, | |
| "loss": 0.2379, | |
| "step": 43900 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 9.131011011011011e-06, | |
| "loss": 0.2728, | |
| "step": 43925 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 9.13051051051051e-06, | |
| "loss": 0.274, | |
| "step": 43950 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 9.130010010010011e-06, | |
| "loss": 0.2838, | |
| "step": 43975 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 9.129509509509509e-06, | |
| "loss": 0.2639, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "eval_loss": 0.43265146017074585, | |
| "eval_runtime": 2614.8693, | |
| "eval_samples_per_second": 6.2, | |
| "eval_steps_per_second": 0.155, | |
| "eval_wer": 25.521696564164632, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 9.12900900900901e-06, | |
| "loss": 0.2588, | |
| "step": 44025 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 9.12850850850851e-06, | |
| "loss": 0.2764, | |
| "step": 44050 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 9.128008008008009e-06, | |
| "loss": 0.263, | |
| "step": 44075 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 9.127507507507508e-06, | |
| "loss": 0.2644, | |
| "step": 44100 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 9.127007007007009e-06, | |
| "loss": 0.2476, | |
| "step": 44125 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 9.126506506506507e-06, | |
| "loss": 0.2604, | |
| "step": 44150 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 9.126006006006006e-06, | |
| "loss": 0.2747, | |
| "step": 44175 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 9.125505505505507e-06, | |
| "loss": 0.265, | |
| "step": 44200 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 9.125005005005006e-06, | |
| "loss": 0.2568, | |
| "step": 44225 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 9.124504504504505e-06, | |
| "loss": 0.2485, | |
| "step": 44250 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 9.124004004004005e-06, | |
| "loss": 0.2742, | |
| "step": 44275 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 9.123503503503504e-06, | |
| "loss": 0.2634, | |
| "step": 44300 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 9.123003003003003e-06, | |
| "loss": 0.2805, | |
| "step": 44325 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 9.122502502502504e-06, | |
| "loss": 0.2543, | |
| "step": 44350 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 9.122002002002004e-06, | |
| "loss": 0.2753, | |
| "step": 44375 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 9.121501501501501e-06, | |
| "loss": 0.2607, | |
| "step": 44400 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 9.121001001001002e-06, | |
| "loss": 0.242, | |
| "step": 44425 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 9.120500500500501e-06, | |
| "loss": 0.2633, | |
| "step": 44450 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 9.12e-06, | |
| "loss": 0.2686, | |
| "step": 44475 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 9.1194994994995e-06, | |
| "loss": 0.2608, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 9.118998998999e-06, | |
| "loss": 0.2779, | |
| "step": 44525 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 9.118498498498499e-06, | |
| "loss": 0.2648, | |
| "step": 44550 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 9.117997997997998e-06, | |
| "loss": 0.2701, | |
| "step": 44575 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 9.117497497497499e-06, | |
| "loss": 0.245, | |
| "step": 44600 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 9.116996996996997e-06, | |
| "loss": 0.2607, | |
| "step": 44625 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 9.116516516516517e-06, | |
| "loss": 0.2622, | |
| "step": 44650 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 9.116016016016017e-06, | |
| "loss": 0.2466, | |
| "step": 44675 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 9.115515515515516e-06, | |
| "loss": 0.2627, | |
| "step": 44700 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 9.115015015015017e-06, | |
| "loss": 0.26, | |
| "step": 44725 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 9.114514514514515e-06, | |
| "loss": 0.2837, | |
| "step": 44750 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 9.114014014014014e-06, | |
| "loss": 0.2417, | |
| "step": 44775 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 9.113513513513515e-06, | |
| "loss": 0.2663, | |
| "step": 44800 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 9.113013013013014e-06, | |
| "loss": 0.2606, | |
| "step": 44825 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 9.112512512512513e-06, | |
| "loss": 0.2416, | |
| "step": 44850 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 9.112012012012013e-06, | |
| "loss": 0.2537, | |
| "step": 44875 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 9.111511511511512e-06, | |
| "loss": 0.2587, | |
| "step": 44900 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 9.111011011011011e-06, | |
| "loss": 0.2658, | |
| "step": 44925 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 9.11051051051051e-06, | |
| "loss": 0.2653, | |
| "step": 44950 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 9.110010010010012e-06, | |
| "loss": 0.2643, | |
| "step": 44975 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 9.10950950950951e-06, | |
| "loss": 0.2638, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "eval_loss": 0.432108074426651, | |
| "eval_runtime": 2486.7031, | |
| "eval_samples_per_second": 6.52, | |
| "eval_steps_per_second": 0.163, | |
| "eval_wer": 24.733851882116234, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 9.10900900900901e-06, | |
| "loss": 0.268, | |
| "step": 45025 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 9.10850850850851e-06, | |
| "loss": 0.2566, | |
| "step": 45050 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 9.108008008008009e-06, | |
| "loss": 0.2603, | |
| "step": 45075 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 9.107507507507508e-06, | |
| "loss": 0.2463, | |
| "step": 45100 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 9.107007007007007e-06, | |
| "loss": 0.2597, | |
| "step": 45125 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 9.106506506506507e-06, | |
| "loss": 0.2531, | |
| "step": 45150 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 9.106006006006006e-06, | |
| "loss": 0.2675, | |
| "step": 45175 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 9.105505505505507e-06, | |
| "loss": 0.2445, | |
| "step": 45200 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 9.105005005005005e-06, | |
| "loss": 0.2576, | |
| "step": 45225 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 9.104504504504506e-06, | |
| "loss": 0.2473, | |
| "step": 45250 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 9.104004004004005e-06, | |
| "loss": 0.2551, | |
| "step": 45275 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 9.103503503503504e-06, | |
| "loss": 0.2658, | |
| "step": 45300 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 9.103003003003003e-06, | |
| "loss": 0.2603, | |
| "step": 45325 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 9.102502502502504e-06, | |
| "loss": 0.2618, | |
| "step": 45350 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 9.102002002002002e-06, | |
| "loss": 0.2405, | |
| "step": 45375 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 9.101501501501501e-06, | |
| "loss": 0.263, | |
| "step": 45400 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 9.101001001001002e-06, | |
| "loss": 0.2673, | |
| "step": 45425 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 9.100500500500502e-06, | |
| "loss": 0.2638, | |
| "step": 45450 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 9.100000000000001e-06, | |
| "loss": 0.2583, | |
| "step": 45475 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 9.0994994994995e-06, | |
| "loss": 0.2695, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 9.098998998999e-06, | |
| "loss": 0.2544, | |
| "step": 45525 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 9.098498498498499e-06, | |
| "loss": 0.2634, | |
| "step": 45550 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 9.097997997997998e-06, | |
| "loss": 0.2532, | |
| "step": 45575 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 9.097497497497499e-06, | |
| "loss": 0.2559, | |
| "step": 45600 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 9.096996996996997e-06, | |
| "loss": 0.2459, | |
| "step": 45625 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 9.096496496496498e-06, | |
| "loss": 0.2781, | |
| "step": 45650 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 9.095995995995997e-06, | |
| "loss": 0.2712, | |
| "step": 45675 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 9.095495495495496e-06, | |
| "loss": 0.2597, | |
| "step": 45700 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 9.094994994994996e-06, | |
| "loss": 0.275, | |
| "step": 45725 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 9.094494494494495e-06, | |
| "loss": 0.2458, | |
| "step": 45750 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 9.093993993993994e-06, | |
| "loss": 0.2752, | |
| "step": 45775 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 9.093493493493494e-06, | |
| "loss": 0.261, | |
| "step": 45800 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 9.092992992992995e-06, | |
| "loss": 0.2532, | |
| "step": 45825 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 9.092492492492492e-06, | |
| "loss": 0.2847, | |
| "step": 45850 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 9.091991991991993e-06, | |
| "loss": 0.2607, | |
| "step": 45875 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 9.091491491491492e-06, | |
| "loss": 0.2678, | |
| "step": 45900 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 9.090990990990992e-06, | |
| "loss": 0.2488, | |
| "step": 45925 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 9.090490490490491e-06, | |
| "loss": 0.2628, | |
| "step": 45950 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 9.089989989989992e-06, | |
| "loss": 0.2625, | |
| "step": 45975 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 9.08948948948949e-06, | |
| "loss": 0.2647, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "eval_loss": 0.44105586409568787, | |
| "eval_runtime": 2496.4644, | |
| "eval_samples_per_second": 6.494, | |
| "eval_steps_per_second": 0.163, | |
| "eval_wer": 24.86027705614357, | |
| "step": 46000 | |
| } | |
| ], | |
| "max_steps": 500000, | |
| "num_train_epochs": 20, | |
| "total_flos": 5.309913924895223e+20, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |