| { |
| "best_metric": 56.294326241134755, |
| "best_model_checkpoint": "./whisper-small-hi/checkpoint-2000", |
| "epoch": 33.333333333333336, |
| "global_step": 4000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.21, |
| "learning_rate": 4.2000000000000006e-07, |
| "loss": 7.0837, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 9.200000000000001e-07, |
| "loss": 6.0185, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 1.42e-06, |
| "loss": 4.3032, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.9200000000000003e-06, |
| "loss": 2.4463, |
| "step": 100 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 2.42e-06, |
| "loss": 1.3751, |
| "step": 125 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 2.92e-06, |
| "loss": 1.0035, |
| "step": 150 |
| }, |
| { |
| "epoch": 1.46, |
| "learning_rate": 3.4200000000000007e-06, |
| "loss": 0.7718, |
| "step": 175 |
| }, |
| { |
| "epoch": 1.67, |
| "learning_rate": 3.920000000000001e-06, |
| "loss": 0.6373, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 4.42e-06, |
| "loss": 0.5507, |
| "step": 225 |
| }, |
| { |
| "epoch": 2.08, |
| "learning_rate": 4.92e-06, |
| "loss": 0.4714, |
| "step": 250 |
| }, |
| { |
| "epoch": 2.29, |
| "learning_rate": 5.420000000000001e-06, |
| "loss": 0.3697, |
| "step": 275 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 5.92e-06, |
| "loss": 0.3467, |
| "step": 300 |
| }, |
| { |
| "epoch": 2.71, |
| "learning_rate": 6.42e-06, |
| "loss": 0.2925, |
| "step": 325 |
| }, |
| { |
| "epoch": 2.92, |
| "learning_rate": 6.92e-06, |
| "loss": 0.3204, |
| "step": 350 |
| }, |
| { |
| "epoch": 3.12, |
| "learning_rate": 7.420000000000001e-06, |
| "loss": 0.2179, |
| "step": 375 |
| }, |
| { |
| "epoch": 3.33, |
| "learning_rate": 7.92e-06, |
| "loss": 0.1938, |
| "step": 400 |
| }, |
| { |
| "epoch": 3.54, |
| "learning_rate": 8.42e-06, |
| "loss": 0.1639, |
| "step": 425 |
| }, |
| { |
| "epoch": 3.75, |
| "learning_rate": 8.920000000000001e-06, |
| "loss": 0.1491, |
| "step": 450 |
| }, |
| { |
| "epoch": 3.96, |
| "learning_rate": 9.42e-06, |
| "loss": 0.1562, |
| "step": 475 |
| }, |
| { |
| "epoch": 4.17, |
| "learning_rate": 9.920000000000002e-06, |
| "loss": 0.1057, |
| "step": 500 |
| }, |
| { |
| "epoch": 4.38, |
| "learning_rate": 9.940000000000001e-06, |
| "loss": 0.1096, |
| "step": 525 |
| }, |
| { |
| "epoch": 4.58, |
| "learning_rate": 9.86857142857143e-06, |
| "loss": 0.1015, |
| "step": 550 |
| }, |
| { |
| "epoch": 4.79, |
| "learning_rate": 9.797142857142858e-06, |
| "loss": 0.1179, |
| "step": 575 |
| }, |
| { |
| "epoch": 5.0, |
| "learning_rate": 9.725714285714287e-06, |
| "loss": 0.0859, |
| "step": 600 |
| }, |
| { |
| "epoch": 5.21, |
| "learning_rate": 9.654285714285716e-06, |
| "loss": 0.0669, |
| "step": 625 |
| }, |
| { |
| "epoch": 5.42, |
| "learning_rate": 9.582857142857143e-06, |
| "loss": 0.055, |
| "step": 650 |
| }, |
| { |
| "epoch": 5.62, |
| "learning_rate": 9.511428571428572e-06, |
| "loss": 0.0683, |
| "step": 675 |
| }, |
| { |
| "epoch": 5.83, |
| "learning_rate": 9.440000000000001e-06, |
| "loss": 0.0582, |
| "step": 700 |
| }, |
| { |
| "epoch": 6.04, |
| "learning_rate": 9.368571428571428e-06, |
| "loss": 0.0528, |
| "step": 725 |
| }, |
| { |
| "epoch": 6.25, |
| "learning_rate": 9.297142857142857e-06, |
| "loss": 0.0415, |
| "step": 750 |
| }, |
| { |
| "epoch": 6.46, |
| "learning_rate": 9.225714285714286e-06, |
| "loss": 0.0343, |
| "step": 775 |
| }, |
| { |
| "epoch": 6.67, |
| "learning_rate": 9.154285714285715e-06, |
| "loss": 0.038, |
| "step": 800 |
| }, |
| { |
| "epoch": 6.88, |
| "learning_rate": 9.082857142857143e-06, |
| "loss": 0.0509, |
| "step": 825 |
| }, |
| { |
| "epoch": 7.08, |
| "learning_rate": 9.011428571428572e-06, |
| "loss": 0.0277, |
| "step": 850 |
| }, |
| { |
| "epoch": 7.29, |
| "learning_rate": 8.94e-06, |
| "loss": 0.0283, |
| "step": 875 |
| }, |
| { |
| "epoch": 7.5, |
| "learning_rate": 8.86857142857143e-06, |
| "loss": 0.0322, |
| "step": 900 |
| }, |
| { |
| "epoch": 7.71, |
| "learning_rate": 8.797142857142857e-06, |
| "loss": 0.0171, |
| "step": 925 |
| }, |
| { |
| "epoch": 7.92, |
| "learning_rate": 8.725714285714286e-06, |
| "loss": 0.0311, |
| "step": 950 |
| }, |
| { |
| "epoch": 8.12, |
| "learning_rate": 8.654285714285715e-06, |
| "loss": 0.0368, |
| "step": 975 |
| }, |
| { |
| "epoch": 8.33, |
| "learning_rate": 8.582857142857144e-06, |
| "loss": 0.0214, |
| "step": 1000 |
| }, |
| { |
| "epoch": 8.33, |
| "eval_loss": 0.5996837019920349, |
| "eval_runtime": 178.3082, |
| "eval_samples_per_second": 3.572, |
| "eval_steps_per_second": 0.449, |
| "eval_wer": 97.25177304964538, |
| "step": 1000 |
| }, |
| { |
| "epoch": 8.54, |
| "learning_rate": 8.511428571428571e-06, |
| "loss": 0.0092, |
| "step": 1025 |
| }, |
| { |
| "epoch": 8.75, |
| "learning_rate": 8.44e-06, |
| "loss": 0.0162, |
| "step": 1050 |
| }, |
| { |
| "epoch": 8.96, |
| "learning_rate": 8.36857142857143e-06, |
| "loss": 0.0166, |
| "step": 1075 |
| }, |
| { |
| "epoch": 9.17, |
| "learning_rate": 8.297142857142859e-06, |
| "loss": 0.0087, |
| "step": 1100 |
| }, |
| { |
| "epoch": 9.38, |
| "learning_rate": 8.225714285714288e-06, |
| "loss": 0.0091, |
| "step": 1125 |
| }, |
| { |
| "epoch": 9.58, |
| "learning_rate": 8.154285714285715e-06, |
| "loss": 0.005, |
| "step": 1150 |
| }, |
| { |
| "epoch": 9.79, |
| "learning_rate": 8.082857142857144e-06, |
| "loss": 0.0134, |
| "step": 1175 |
| }, |
| { |
| "epoch": 10.0, |
| "learning_rate": 8.011428571428573e-06, |
| "loss": 0.006, |
| "step": 1200 |
| }, |
| { |
| "epoch": 10.21, |
| "learning_rate": 7.94e-06, |
| "loss": 0.0038, |
| "step": 1225 |
| }, |
| { |
| "epoch": 10.42, |
| "learning_rate": 7.86857142857143e-06, |
| "loss": 0.0051, |
| "step": 1250 |
| }, |
| { |
| "epoch": 10.62, |
| "learning_rate": 7.797142857142858e-06, |
| "loss": 0.009, |
| "step": 1275 |
| }, |
| { |
| "epoch": 10.83, |
| "learning_rate": 7.725714285714286e-06, |
| "loss": 0.0062, |
| "step": 1300 |
| }, |
| { |
| "epoch": 11.04, |
| "learning_rate": 7.654285714285715e-06, |
| "loss": 0.0071, |
| "step": 1325 |
| }, |
| { |
| "epoch": 11.25, |
| "learning_rate": 7.5828571428571444e-06, |
| "loss": 0.0065, |
| "step": 1350 |
| }, |
| { |
| "epoch": 11.46, |
| "learning_rate": 7.511428571428572e-06, |
| "loss": 0.0066, |
| "step": 1375 |
| }, |
| { |
| "epoch": 11.67, |
| "learning_rate": 7.440000000000001e-06, |
| "loss": 0.0053, |
| "step": 1400 |
| }, |
| { |
| "epoch": 11.88, |
| "learning_rate": 7.36857142857143e-06, |
| "loss": 0.0019, |
| "step": 1425 |
| }, |
| { |
| "epoch": 12.08, |
| "learning_rate": 7.297142857142858e-06, |
| "loss": 0.0078, |
| "step": 1450 |
| }, |
| { |
| "epoch": 12.29, |
| "learning_rate": 7.225714285714286e-06, |
| "loss": 0.0119, |
| "step": 1475 |
| }, |
| { |
| "epoch": 12.5, |
| "learning_rate": 7.154285714285715e-06, |
| "loss": 0.004, |
| "step": 1500 |
| }, |
| { |
| "epoch": 12.71, |
| "learning_rate": 7.082857142857143e-06, |
| "loss": 0.002, |
| "step": 1525 |
| }, |
| { |
| "epoch": 12.92, |
| "learning_rate": 7.011428571428572e-06, |
| "loss": 0.0084, |
| "step": 1550 |
| }, |
| { |
| "epoch": 13.12, |
| "learning_rate": 6.9400000000000005e-06, |
| "loss": 0.0008, |
| "step": 1575 |
| }, |
| { |
| "epoch": 13.33, |
| "learning_rate": 6.868571428571429e-06, |
| "loss": 0.004, |
| "step": 1600 |
| }, |
| { |
| "epoch": 13.54, |
| "learning_rate": 6.797142857142858e-06, |
| "loss": 0.0045, |
| "step": 1625 |
| }, |
| { |
| "epoch": 13.75, |
| "learning_rate": 6.725714285714287e-06, |
| "loss": 0.0027, |
| "step": 1650 |
| }, |
| { |
| "epoch": 13.96, |
| "learning_rate": 6.654285714285716e-06, |
| "loss": 0.0018, |
| "step": 1675 |
| }, |
| { |
| "epoch": 14.17, |
| "learning_rate": 6.582857142857143e-06, |
| "loss": 0.0014, |
| "step": 1700 |
| }, |
| { |
| "epoch": 14.38, |
| "learning_rate": 6.511428571428572e-06, |
| "loss": 0.001, |
| "step": 1725 |
| }, |
| { |
| "epoch": 14.58, |
| "learning_rate": 6.440000000000001e-06, |
| "loss": 0.002, |
| "step": 1750 |
| }, |
| { |
| "epoch": 14.79, |
| "learning_rate": 6.368571428571429e-06, |
| "loss": 0.0085, |
| "step": 1775 |
| }, |
| { |
| "epoch": 15.0, |
| "learning_rate": 6.297142857142857e-06, |
| "loss": 0.0011, |
| "step": 1800 |
| }, |
| { |
| "epoch": 15.21, |
| "learning_rate": 6.225714285714286e-06, |
| "loss": 0.0009, |
| "step": 1825 |
| }, |
| { |
| "epoch": 15.42, |
| "learning_rate": 6.1542857142857145e-06, |
| "loss": 0.0017, |
| "step": 1850 |
| }, |
| { |
| "epoch": 15.62, |
| "learning_rate": 6.0828571428571435e-06, |
| "loss": 0.0004, |
| "step": 1875 |
| }, |
| { |
| "epoch": 15.83, |
| "learning_rate": 6.011428571428572e-06, |
| "loss": 0.0005, |
| "step": 1900 |
| }, |
| { |
| "epoch": 16.04, |
| "learning_rate": 5.94e-06, |
| "loss": 0.0019, |
| "step": 1925 |
| }, |
| { |
| "epoch": 16.25, |
| "learning_rate": 5.868571428571429e-06, |
| "loss": 0.0002, |
| "step": 1950 |
| }, |
| { |
| "epoch": 16.46, |
| "learning_rate": 5.797142857142858e-06, |
| "loss": 0.0006, |
| "step": 1975 |
| }, |
| { |
| "epoch": 16.67, |
| "learning_rate": 5.725714285714287e-06, |
| "loss": 0.0002, |
| "step": 2000 |
| }, |
| { |
| "epoch": 16.67, |
| "eval_loss": 0.4707730710506439, |
| "eval_runtime": 179.3219, |
| "eval_samples_per_second": 3.552, |
| "eval_steps_per_second": 0.446, |
| "eval_wer": 56.294326241134755, |
| "step": 2000 |
| }, |
| { |
| "epoch": 16.88, |
| "learning_rate": 5.654285714285714e-06, |
| "loss": 0.0002, |
| "step": 2025 |
| }, |
| { |
| "epoch": 17.08, |
| "learning_rate": 5.582857142857143e-06, |
| "loss": 0.0016, |
| "step": 2050 |
| }, |
| { |
| "epoch": 17.29, |
| "learning_rate": 5.511428571428572e-06, |
| "loss": 0.0017, |
| "step": 2075 |
| }, |
| { |
| "epoch": 17.5, |
| "learning_rate": 5.4400000000000004e-06, |
| "loss": 0.0002, |
| "step": 2100 |
| }, |
| { |
| "epoch": 17.71, |
| "learning_rate": 5.368571428571429e-06, |
| "loss": 0.0002, |
| "step": 2125 |
| }, |
| { |
| "epoch": 17.92, |
| "learning_rate": 5.297142857142858e-06, |
| "loss": 0.0001, |
| "step": 2150 |
| }, |
| { |
| "epoch": 18.12, |
| "learning_rate": 5.225714285714286e-06, |
| "loss": 0.0001, |
| "step": 2175 |
| }, |
| { |
| "epoch": 18.33, |
| "learning_rate": 5.154285714285715e-06, |
| "loss": 0.0003, |
| "step": 2200 |
| }, |
| { |
| "epoch": 18.54, |
| "learning_rate": 5.082857142857144e-06, |
| "loss": 0.0001, |
| "step": 2225 |
| }, |
| { |
| "epoch": 18.75, |
| "learning_rate": 5.011428571428571e-06, |
| "loss": 0.0001, |
| "step": 2250 |
| }, |
| { |
| "epoch": 18.96, |
| "learning_rate": 4.94e-06, |
| "loss": 0.0001, |
| "step": 2275 |
| }, |
| { |
| "epoch": 19.17, |
| "learning_rate": 4.868571428571429e-06, |
| "loss": 0.0001, |
| "step": 2300 |
| }, |
| { |
| "epoch": 19.38, |
| "learning_rate": 4.797142857142857e-06, |
| "loss": 0.0001, |
| "step": 2325 |
| }, |
| { |
| "epoch": 19.58, |
| "learning_rate": 4.725714285714286e-06, |
| "loss": 0.0001, |
| "step": 2350 |
| }, |
| { |
| "epoch": 19.79, |
| "learning_rate": 4.6542857142857145e-06, |
| "loss": 0.0001, |
| "step": 2375 |
| }, |
| { |
| "epoch": 20.0, |
| "learning_rate": 4.5828571428571435e-06, |
| "loss": 0.0007, |
| "step": 2400 |
| }, |
| { |
| "epoch": 20.21, |
| "learning_rate": 4.511428571428572e-06, |
| "loss": 0.0001, |
| "step": 2425 |
| }, |
| { |
| "epoch": 20.42, |
| "learning_rate": 4.440000000000001e-06, |
| "loss": 0.0007, |
| "step": 2450 |
| }, |
| { |
| "epoch": 20.62, |
| "learning_rate": 4.368571428571429e-06, |
| "loss": 0.0001, |
| "step": 2475 |
| }, |
| { |
| "epoch": 20.83, |
| "learning_rate": 4.297142857142858e-06, |
| "loss": 0.0001, |
| "step": 2500 |
| }, |
| { |
| "epoch": 21.04, |
| "learning_rate": 4.225714285714286e-06, |
| "loss": 0.0001, |
| "step": 2525 |
| }, |
| { |
| "epoch": 21.25, |
| "learning_rate": 4.154285714285714e-06, |
| "loss": 0.0001, |
| "step": 2550 |
| }, |
| { |
| "epoch": 21.46, |
| "learning_rate": 4.082857142857143e-06, |
| "loss": 0.0001, |
| "step": 2575 |
| }, |
| { |
| "epoch": 21.67, |
| "learning_rate": 4.011428571428571e-06, |
| "loss": 0.0022, |
| "step": 2600 |
| }, |
| { |
| "epoch": 21.88, |
| "learning_rate": 3.94e-06, |
| "loss": 0.0001, |
| "step": 2625 |
| }, |
| { |
| "epoch": 22.08, |
| "learning_rate": 3.8685714285714286e-06, |
| "loss": 0.0001, |
| "step": 2650 |
| }, |
| { |
| "epoch": 22.29, |
| "learning_rate": 3.7971428571428576e-06, |
| "loss": 0.0001, |
| "step": 2675 |
| }, |
| { |
| "epoch": 22.5, |
| "learning_rate": 3.7257142857142857e-06, |
| "loss": 0.0023, |
| "step": 2700 |
| }, |
| { |
| "epoch": 22.71, |
| "learning_rate": 3.6542857142857148e-06, |
| "loss": 0.0001, |
| "step": 2725 |
| }, |
| { |
| "epoch": 22.92, |
| "learning_rate": 3.582857142857143e-06, |
| "loss": 0.0001, |
| "step": 2750 |
| }, |
| { |
| "epoch": 23.12, |
| "learning_rate": 3.511428571428572e-06, |
| "loss": 0.0001, |
| "step": 2775 |
| }, |
| { |
| "epoch": 23.33, |
| "learning_rate": 3.44e-06, |
| "loss": 0.0001, |
| "step": 2800 |
| }, |
| { |
| "epoch": 23.54, |
| "learning_rate": 3.3685714285714287e-06, |
| "loss": 0.0001, |
| "step": 2825 |
| }, |
| { |
| "epoch": 23.75, |
| "learning_rate": 3.2971428571428577e-06, |
| "loss": 0.0001, |
| "step": 2850 |
| }, |
| { |
| "epoch": 23.96, |
| "learning_rate": 3.225714285714286e-06, |
| "loss": 0.0007, |
| "step": 2875 |
| }, |
| { |
| "epoch": 24.17, |
| "learning_rate": 3.154285714285715e-06, |
| "loss": 0.0001, |
| "step": 2900 |
| }, |
| { |
| "epoch": 24.38, |
| "learning_rate": 3.082857142857143e-06, |
| "loss": 0.0001, |
| "step": 2925 |
| }, |
| { |
| "epoch": 24.58, |
| "learning_rate": 3.0114285714285716e-06, |
| "loss": 0.0001, |
| "step": 2950 |
| }, |
| { |
| "epoch": 24.79, |
| "learning_rate": 2.9400000000000002e-06, |
| "loss": 0.0001, |
| "step": 2975 |
| }, |
| { |
| "epoch": 25.0, |
| "learning_rate": 2.868571428571429e-06, |
| "loss": 0.0011, |
| "step": 3000 |
| }, |
| { |
| "epoch": 25.0, |
| "eval_loss": 0.5027238726615906, |
| "eval_runtime": 180.7143, |
| "eval_samples_per_second": 3.525, |
| "eval_steps_per_second": 0.443, |
| "eval_wer": 57.801418439716315, |
| "step": 3000 |
| }, |
| { |
| "epoch": 25.21, |
| "learning_rate": 2.797142857142857e-06, |
| "loss": 0.0001, |
| "step": 3025 |
| }, |
| { |
| "epoch": 25.42, |
| "learning_rate": 2.725714285714286e-06, |
| "loss": 0.0001, |
| "step": 3050 |
| }, |
| { |
| "epoch": 25.62, |
| "learning_rate": 2.654285714285714e-06, |
| "loss": 0.0001, |
| "step": 3075 |
| }, |
| { |
| "epoch": 25.83, |
| "learning_rate": 2.582857142857143e-06, |
| "loss": 0.0001, |
| "step": 3100 |
| }, |
| { |
| "epoch": 26.04, |
| "learning_rate": 2.5114285714285718e-06, |
| "loss": 0.0001, |
| "step": 3125 |
| }, |
| { |
| "epoch": 26.25, |
| "learning_rate": 2.4400000000000004e-06, |
| "loss": 0.0001, |
| "step": 3150 |
| }, |
| { |
| "epoch": 26.46, |
| "learning_rate": 2.3685714285714285e-06, |
| "loss": 0.0001, |
| "step": 3175 |
| }, |
| { |
| "epoch": 26.67, |
| "learning_rate": 2.297142857142857e-06, |
| "loss": 0.0001, |
| "step": 3200 |
| }, |
| { |
| "epoch": 26.88, |
| "learning_rate": 2.2257142857142857e-06, |
| "loss": 0.0001, |
| "step": 3225 |
| }, |
| { |
| "epoch": 27.08, |
| "learning_rate": 2.1542857142857147e-06, |
| "loss": 0.0001, |
| "step": 3250 |
| }, |
| { |
| "epoch": 27.29, |
| "learning_rate": 2.0828571428571433e-06, |
| "loss": 0.0001, |
| "step": 3275 |
| }, |
| { |
| "epoch": 27.5, |
| "learning_rate": 2.0114285714285715e-06, |
| "loss": 0.0001, |
| "step": 3300 |
| }, |
| { |
| "epoch": 27.71, |
| "learning_rate": 1.94e-06, |
| "loss": 0.0001, |
| "step": 3325 |
| }, |
| { |
| "epoch": 27.92, |
| "learning_rate": 1.8685714285714289e-06, |
| "loss": 0.0001, |
| "step": 3350 |
| }, |
| { |
| "epoch": 28.12, |
| "learning_rate": 1.7971428571428572e-06, |
| "loss": 0.0001, |
| "step": 3375 |
| }, |
| { |
| "epoch": 28.33, |
| "learning_rate": 1.7257142857142858e-06, |
| "loss": 0.0001, |
| "step": 3400 |
| }, |
| { |
| "epoch": 28.54, |
| "learning_rate": 1.6542857142857144e-06, |
| "loss": 0.0001, |
| "step": 3425 |
| }, |
| { |
| "epoch": 28.75, |
| "learning_rate": 1.582857142857143e-06, |
| "loss": 0.0001, |
| "step": 3450 |
| }, |
| { |
| "epoch": 28.96, |
| "learning_rate": 1.5114285714285714e-06, |
| "loss": 0.0, |
| "step": 3475 |
| }, |
| { |
| "epoch": 29.17, |
| "learning_rate": 1.44e-06, |
| "loss": 0.0001, |
| "step": 3500 |
| }, |
| { |
| "epoch": 29.38, |
| "learning_rate": 1.3685714285714286e-06, |
| "loss": 0.0001, |
| "step": 3525 |
| }, |
| { |
| "epoch": 29.58, |
| "learning_rate": 1.2971428571428574e-06, |
| "loss": 0.0, |
| "step": 3550 |
| }, |
| { |
| "epoch": 29.79, |
| "learning_rate": 1.2257142857142857e-06, |
| "loss": 0.0001, |
| "step": 3575 |
| }, |
| { |
| "epoch": 30.0, |
| "learning_rate": 1.1542857142857143e-06, |
| "loss": 0.0001, |
| "step": 3600 |
| }, |
| { |
| "epoch": 30.21, |
| "learning_rate": 1.082857142857143e-06, |
| "loss": 0.0001, |
| "step": 3625 |
| }, |
| { |
| "epoch": 30.42, |
| "learning_rate": 1.0114285714285715e-06, |
| "loss": 0.0, |
| "step": 3650 |
| }, |
| { |
| "epoch": 30.62, |
| "learning_rate": 9.400000000000001e-07, |
| "loss": 0.0001, |
| "step": 3675 |
| }, |
| { |
| "epoch": 30.83, |
| "learning_rate": 8.685714285714286e-07, |
| "loss": 0.0001, |
| "step": 3700 |
| }, |
| { |
| "epoch": 31.04, |
| "learning_rate": 7.971428571428572e-07, |
| "loss": 0.0001, |
| "step": 3725 |
| }, |
| { |
| "epoch": 31.25, |
| "learning_rate": 7.257142857142857e-07, |
| "loss": 0.0, |
| "step": 3750 |
| }, |
| { |
| "epoch": 31.46, |
| "learning_rate": 6.542857142857144e-07, |
| "loss": 0.0001, |
| "step": 3775 |
| }, |
| { |
| "epoch": 31.67, |
| "learning_rate": 5.82857142857143e-07, |
| "loss": 0.0001, |
| "step": 3800 |
| }, |
| { |
| "epoch": 31.88, |
| "learning_rate": 5.114285714285714e-07, |
| "loss": 0.0001, |
| "step": 3825 |
| }, |
| { |
| "epoch": 32.08, |
| "learning_rate": 4.4e-07, |
| "loss": 0.0, |
| "step": 3850 |
| }, |
| { |
| "epoch": 32.29, |
| "learning_rate": 3.685714285714286e-07, |
| "loss": 0.0001, |
| "step": 3875 |
| }, |
| { |
| "epoch": 32.5, |
| "learning_rate": 2.9714285714285715e-07, |
| "loss": 0.0001, |
| "step": 3900 |
| }, |
| { |
| "epoch": 32.71, |
| "learning_rate": 2.2571428571428574e-07, |
| "loss": 0.0, |
| "step": 3925 |
| }, |
| { |
| "epoch": 32.92, |
| "learning_rate": 1.542857142857143e-07, |
| "loss": 0.0001, |
| "step": 3950 |
| }, |
| { |
| "epoch": 33.12, |
| "learning_rate": 8.285714285714285e-08, |
| "loss": 0.0001, |
| "step": 3975 |
| }, |
| { |
| "epoch": 33.33, |
| "learning_rate": 1.142857142857143e-08, |
| "loss": 0.0001, |
| "step": 4000 |
| }, |
| { |
| "epoch": 33.33, |
| "eval_loss": 0.5104972720146179, |
| "eval_runtime": 176.8369, |
| "eval_samples_per_second": 3.602, |
| "eval_steps_per_second": 0.452, |
| "eval_wer": 57.71276595744681, |
| "step": 4000 |
| } |
| ], |
| "max_steps": 4000, |
| "num_train_epochs": 34, |
| "total_flos": 1.834566248374272e+19, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|