| { |
| "best_metric": 33.282732447817835, |
| "best_model_checkpoint": "whisper-lv-small-augmented/checkpoint-4000", |
| "epoch": 83.33333333333333, |
| "global_step": 5000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.83, |
| "learning_rate": 4.6e-06, |
| "loss": 2.9758, |
| "step": 50 |
| }, |
| { |
| "epoch": 1.67, |
| "learning_rate": 9.600000000000001e-06, |
| "loss": 1.2165, |
| "step": 100 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 1.4599999999999999e-05, |
| "loss": 0.7096, |
| "step": 150 |
| }, |
| { |
| "epoch": 3.33, |
| "learning_rate": 1.9600000000000002e-05, |
| "loss": 0.4438, |
| "step": 200 |
| }, |
| { |
| "epoch": 4.17, |
| "learning_rate": 2.46e-05, |
| "loss": 0.3072, |
| "step": 250 |
| }, |
| { |
| "epoch": 5.0, |
| "learning_rate": 2.96e-05, |
| "loss": 0.1973, |
| "step": 300 |
| }, |
| { |
| "epoch": 5.83, |
| "learning_rate": 3.46e-05, |
| "loss": 0.1346, |
| "step": 350 |
| }, |
| { |
| "epoch": 6.67, |
| "learning_rate": 3.960000000000001e-05, |
| "loss": 0.107, |
| "step": 400 |
| }, |
| { |
| "epoch": 7.5, |
| "learning_rate": 4.46e-05, |
| "loss": 0.0858, |
| "step": 450 |
| }, |
| { |
| "epoch": 8.33, |
| "learning_rate": 4.96e-05, |
| "loss": 0.0758, |
| "step": 500 |
| }, |
| { |
| "epoch": 9.17, |
| "learning_rate": 4.948888888888889e-05, |
| "loss": 0.0645, |
| "step": 550 |
| }, |
| { |
| "epoch": 10.0, |
| "learning_rate": 4.8933333333333335e-05, |
| "loss": 0.0608, |
| "step": 600 |
| }, |
| { |
| "epoch": 10.83, |
| "learning_rate": 4.837777777777778e-05, |
| "loss": 0.0443, |
| "step": 650 |
| }, |
| { |
| "epoch": 11.67, |
| "learning_rate": 4.782222222222222e-05, |
| "loss": 0.0415, |
| "step": 700 |
| }, |
| { |
| "epoch": 12.5, |
| "learning_rate": 4.726666666666667e-05, |
| "loss": 0.0378, |
| "step": 750 |
| }, |
| { |
| "epoch": 13.33, |
| "learning_rate": 4.671111111111111e-05, |
| "loss": 0.0312, |
| "step": 800 |
| }, |
| { |
| "epoch": 14.17, |
| "learning_rate": 4.615555555555556e-05, |
| "loss": 0.0278, |
| "step": 850 |
| }, |
| { |
| "epoch": 15.0, |
| "learning_rate": 4.5600000000000004e-05, |
| "loss": 0.0266, |
| "step": 900 |
| }, |
| { |
| "epoch": 15.83, |
| "learning_rate": 4.504444444444445e-05, |
| "loss": 0.0241, |
| "step": 950 |
| }, |
| { |
| "epoch": 16.67, |
| "learning_rate": 4.448888888888889e-05, |
| "loss": 0.0219, |
| "step": 1000 |
| }, |
| { |
| "epoch": 16.67, |
| "eval_loss": 0.6128966808319092, |
| "eval_runtime": 463.1108, |
| "eval_samples_per_second": 4.789, |
| "eval_steps_per_second": 0.151, |
| "eval_wer": 41.15749525616698, |
| "step": 1000 |
| }, |
| { |
| "epoch": 17.5, |
| "learning_rate": 4.3933333333333335e-05, |
| "loss": 0.0199, |
| "step": 1050 |
| }, |
| { |
| "epoch": 18.33, |
| "learning_rate": 4.337777777777778e-05, |
| "loss": 0.0187, |
| "step": 1100 |
| }, |
| { |
| "epoch": 19.17, |
| "learning_rate": 4.282222222222222e-05, |
| "loss": 0.015, |
| "step": 1150 |
| }, |
| { |
| "epoch": 20.0, |
| "learning_rate": 4.226666666666667e-05, |
| "loss": 0.0139, |
| "step": 1200 |
| }, |
| { |
| "epoch": 20.83, |
| "learning_rate": 4.171111111111111e-05, |
| "loss": 0.0123, |
| "step": 1250 |
| }, |
| { |
| "epoch": 21.67, |
| "learning_rate": 4.115555555555556e-05, |
| "loss": 0.0114, |
| "step": 1300 |
| }, |
| { |
| "epoch": 22.5, |
| "learning_rate": 4.0600000000000004e-05, |
| "loss": 0.0103, |
| "step": 1350 |
| }, |
| { |
| "epoch": 23.33, |
| "learning_rate": 4.004444444444445e-05, |
| "loss": 0.0101, |
| "step": 1400 |
| }, |
| { |
| "epoch": 24.17, |
| "learning_rate": 3.948888888888889e-05, |
| "loss": 0.0097, |
| "step": 1450 |
| }, |
| { |
| "epoch": 25.0, |
| "learning_rate": 3.8933333333333336e-05, |
| "loss": 0.0092, |
| "step": 1500 |
| }, |
| { |
| "epoch": 25.83, |
| "learning_rate": 3.837777777777778e-05, |
| "loss": 0.0087, |
| "step": 1550 |
| }, |
| { |
| "epoch": 26.67, |
| "learning_rate": 3.782222222222222e-05, |
| "loss": 0.0073, |
| "step": 1600 |
| }, |
| { |
| "epoch": 27.5, |
| "learning_rate": 3.726666666666667e-05, |
| "loss": 0.0065, |
| "step": 1650 |
| }, |
| { |
| "epoch": 28.33, |
| "learning_rate": 3.671111111111111e-05, |
| "loss": 0.0076, |
| "step": 1700 |
| }, |
| { |
| "epoch": 29.17, |
| "learning_rate": 3.615555555555556e-05, |
| "loss": 0.0069, |
| "step": 1750 |
| }, |
| { |
| "epoch": 30.0, |
| "learning_rate": 3.56e-05, |
| "loss": 0.0057, |
| "step": 1800 |
| }, |
| { |
| "epoch": 30.83, |
| "learning_rate": 3.504444444444445e-05, |
| "loss": 0.0055, |
| "step": 1850 |
| }, |
| { |
| "epoch": 31.67, |
| "learning_rate": 3.448888888888889e-05, |
| "loss": 0.0052, |
| "step": 1900 |
| }, |
| { |
| "epoch": 32.5, |
| "learning_rate": 3.3933333333333336e-05, |
| "loss": 0.0039, |
| "step": 1950 |
| }, |
| { |
| "epoch": 33.33, |
| "learning_rate": 3.337777777777778e-05, |
| "loss": 0.0029, |
| "step": 2000 |
| }, |
| { |
| "epoch": 33.33, |
| "eval_loss": 0.5974836349487305, |
| "eval_runtime": 428.1165, |
| "eval_samples_per_second": 5.181, |
| "eval_steps_per_second": 0.164, |
| "eval_wer": 36.14800759013283, |
| "step": 2000 |
| }, |
| { |
| "epoch": 34.17, |
| "learning_rate": 3.2822222222222223e-05, |
| "loss": 0.0035, |
| "step": 2050 |
| }, |
| { |
| "epoch": 35.0, |
| "learning_rate": 3.226666666666667e-05, |
| "loss": 0.0029, |
| "step": 2100 |
| }, |
| { |
| "epoch": 35.83, |
| "learning_rate": 3.171111111111111e-05, |
| "loss": 0.003, |
| "step": 2150 |
| }, |
| { |
| "epoch": 36.67, |
| "learning_rate": 3.1155555555555555e-05, |
| "loss": 0.0037, |
| "step": 2200 |
| }, |
| { |
| "epoch": 37.5, |
| "learning_rate": 3.06e-05, |
| "loss": 0.0035, |
| "step": 2250 |
| }, |
| { |
| "epoch": 38.33, |
| "learning_rate": 3.004444444444445e-05, |
| "loss": 0.0029, |
| "step": 2300 |
| }, |
| { |
| "epoch": 39.17, |
| "learning_rate": 2.948888888888889e-05, |
| "loss": 0.003, |
| "step": 2350 |
| }, |
| { |
| "epoch": 40.0, |
| "learning_rate": 2.8933333333333333e-05, |
| "loss": 0.0041, |
| "step": 2400 |
| }, |
| { |
| "epoch": 40.83, |
| "learning_rate": 2.837777777777778e-05, |
| "loss": 0.0034, |
| "step": 2450 |
| }, |
| { |
| "epoch": 41.67, |
| "learning_rate": 2.782222222222222e-05, |
| "loss": 0.0026, |
| "step": 2500 |
| }, |
| { |
| "epoch": 42.5, |
| "learning_rate": 2.7266666666666668e-05, |
| "loss": 0.0031, |
| "step": 2550 |
| }, |
| { |
| "epoch": 43.33, |
| "learning_rate": 2.6711111111111115e-05, |
| "loss": 0.002, |
| "step": 2600 |
| }, |
| { |
| "epoch": 44.17, |
| "learning_rate": 2.6155555555555555e-05, |
| "loss": 0.0009, |
| "step": 2650 |
| }, |
| { |
| "epoch": 45.0, |
| "learning_rate": 2.5600000000000002e-05, |
| "loss": 0.002, |
| "step": 2700 |
| }, |
| { |
| "epoch": 45.83, |
| "learning_rate": 2.504444444444445e-05, |
| "loss": 0.0016, |
| "step": 2750 |
| }, |
| { |
| "epoch": 46.67, |
| "learning_rate": 2.448888888888889e-05, |
| "loss": 0.0009, |
| "step": 2800 |
| }, |
| { |
| "epoch": 47.5, |
| "learning_rate": 2.3933333333333337e-05, |
| "loss": 0.001, |
| "step": 2850 |
| }, |
| { |
| "epoch": 48.33, |
| "learning_rate": 2.337777777777778e-05, |
| "loss": 0.0012, |
| "step": 2900 |
| }, |
| { |
| "epoch": 49.17, |
| "learning_rate": 2.282222222222222e-05, |
| "loss": 0.0012, |
| "step": 2950 |
| }, |
| { |
| "epoch": 50.0, |
| "learning_rate": 2.2266666666666668e-05, |
| "loss": 0.0003, |
| "step": 3000 |
| }, |
| { |
| "epoch": 50.0, |
| "eval_loss": 0.5625514388084412, |
| "eval_runtime": 427.1642, |
| "eval_samples_per_second": 5.192, |
| "eval_steps_per_second": 0.164, |
| "eval_wer": 33.68121442125237, |
| "step": 3000 |
| }, |
| { |
| "epoch": 50.83, |
| "learning_rate": 2.1711111111111112e-05, |
| "loss": 0.0011, |
| "step": 3050 |
| }, |
| { |
| "epoch": 51.67, |
| "learning_rate": 2.1155555555555556e-05, |
| "loss": 0.0012, |
| "step": 3100 |
| }, |
| { |
| "epoch": 52.5, |
| "learning_rate": 2.06e-05, |
| "loss": 0.0007, |
| "step": 3150 |
| }, |
| { |
| "epoch": 53.33, |
| "learning_rate": 2.0044444444444446e-05, |
| "loss": 0.0009, |
| "step": 3200 |
| }, |
| { |
| "epoch": 54.17, |
| "learning_rate": 1.948888888888889e-05, |
| "loss": 0.0005, |
| "step": 3250 |
| }, |
| { |
| "epoch": 55.0, |
| "learning_rate": 1.8933333333333334e-05, |
| "loss": 0.0003, |
| "step": 3300 |
| }, |
| { |
| "epoch": 55.83, |
| "learning_rate": 1.837777777777778e-05, |
| "loss": 0.0003, |
| "step": 3350 |
| }, |
| { |
| "epoch": 56.67, |
| "learning_rate": 1.7822222222222225e-05, |
| "loss": 0.0004, |
| "step": 3400 |
| }, |
| { |
| "epoch": 57.5, |
| "learning_rate": 1.726666666666667e-05, |
| "loss": 0.0006, |
| "step": 3450 |
| }, |
| { |
| "epoch": 58.33, |
| "learning_rate": 1.6711111111111112e-05, |
| "loss": 0.0014, |
| "step": 3500 |
| }, |
| { |
| "epoch": 59.17, |
| "learning_rate": 1.6155555555555556e-05, |
| "loss": 0.0006, |
| "step": 3550 |
| }, |
| { |
| "epoch": 60.0, |
| "learning_rate": 1.56e-05, |
| "loss": 0.0004, |
| "step": 3600 |
| }, |
| { |
| "epoch": 60.83, |
| "learning_rate": 1.5044444444444445e-05, |
| "loss": 0.0004, |
| "step": 3650 |
| }, |
| { |
| "epoch": 61.67, |
| "learning_rate": 1.448888888888889e-05, |
| "loss": 0.0004, |
| "step": 3700 |
| }, |
| { |
| "epoch": 62.5, |
| "learning_rate": 1.3933333333333334e-05, |
| "loss": 0.0002, |
| "step": 3750 |
| }, |
| { |
| "epoch": 63.33, |
| "learning_rate": 1.3377777777777778e-05, |
| "loss": 0.0004, |
| "step": 3800 |
| }, |
| { |
| "epoch": 64.17, |
| "learning_rate": 1.2822222222222222e-05, |
| "loss": 0.0002, |
| "step": 3850 |
| }, |
| { |
| "epoch": 65.0, |
| "learning_rate": 1.2266666666666667e-05, |
| "loss": 0.0002, |
| "step": 3900 |
| }, |
| { |
| "epoch": 65.83, |
| "learning_rate": 1.1711111111111111e-05, |
| "loss": 0.0004, |
| "step": 3950 |
| }, |
| { |
| "epoch": 66.67, |
| "learning_rate": 1.1155555555555556e-05, |
| "loss": 0.0004, |
| "step": 4000 |
| }, |
| { |
| "epoch": 66.67, |
| "eval_loss": 0.560806930065155, |
| "eval_runtime": 426.638, |
| "eval_samples_per_second": 5.199, |
| "eval_steps_per_second": 0.164, |
| "eval_wer": 33.282732447817835, |
| "step": 4000 |
| }, |
| { |
| "epoch": 67.5, |
| "learning_rate": 1.06e-05, |
| "loss": 0.0002, |
| "step": 4050 |
| }, |
| { |
| "epoch": 68.33, |
| "learning_rate": 1.0044444444444446e-05, |
| "loss": 0.0002, |
| "step": 4100 |
| }, |
| { |
| "epoch": 69.17, |
| "learning_rate": 9.48888888888889e-06, |
| "loss": 0.0003, |
| "step": 4150 |
| }, |
| { |
| "epoch": 70.0, |
| "learning_rate": 8.933333333333333e-06, |
| "loss": 0.0003, |
| "step": 4200 |
| }, |
| { |
| "epoch": 70.83, |
| "learning_rate": 8.377777777777779e-06, |
| "loss": 0.0002, |
| "step": 4250 |
| }, |
| { |
| "epoch": 71.67, |
| "learning_rate": 7.822222222222222e-06, |
| "loss": 0.0004, |
| "step": 4300 |
| }, |
| { |
| "epoch": 72.5, |
| "learning_rate": 7.266666666666668e-06, |
| "loss": 0.0002, |
| "step": 4350 |
| }, |
| { |
| "epoch": 73.33, |
| "learning_rate": 6.711111111111111e-06, |
| "loss": 0.0001, |
| "step": 4400 |
| }, |
| { |
| "epoch": 74.17, |
| "learning_rate": 6.155555555555556e-06, |
| "loss": 0.0001, |
| "step": 4450 |
| }, |
| { |
| "epoch": 75.0, |
| "learning_rate": 5.600000000000001e-06, |
| "loss": 0.0001, |
| "step": 4500 |
| }, |
| { |
| "epoch": 75.83, |
| "learning_rate": 5.044444444444444e-06, |
| "loss": 0.0001, |
| "step": 4550 |
| }, |
| { |
| "epoch": 76.67, |
| "learning_rate": 4.488888888888889e-06, |
| "loss": 0.0001, |
| "step": 4600 |
| }, |
| { |
| "epoch": 77.5, |
| "learning_rate": 3.9333333333333335e-06, |
| "loss": 0.0001, |
| "step": 4650 |
| }, |
| { |
| "epoch": 78.33, |
| "learning_rate": 3.3777777777777777e-06, |
| "loss": 0.0003, |
| "step": 4700 |
| }, |
| { |
| "epoch": 79.17, |
| "learning_rate": 2.8222222222222223e-06, |
| "loss": 0.0002, |
| "step": 4750 |
| }, |
| { |
| "epoch": 80.0, |
| "learning_rate": 2.266666666666667e-06, |
| "loss": 0.0001, |
| "step": 4800 |
| }, |
| { |
| "epoch": 80.83, |
| "learning_rate": 1.7111111111111112e-06, |
| "loss": 0.0001, |
| "step": 4850 |
| }, |
| { |
| "epoch": 81.67, |
| "learning_rate": 1.1555555555555556e-06, |
| "loss": 0.0002, |
| "step": 4900 |
| }, |
| { |
| "epoch": 82.5, |
| "learning_rate": 6.000000000000001e-07, |
| "loss": 0.0001, |
| "step": 4950 |
| }, |
| { |
| "epoch": 83.33, |
| "learning_rate": 4.444444444444445e-08, |
| "loss": 0.0001, |
| "step": 5000 |
| }, |
| { |
| "epoch": 83.33, |
| "eval_loss": 0.5715191960334778, |
| "eval_runtime": 429.9918, |
| "eval_samples_per_second": 5.158, |
| "eval_steps_per_second": 0.163, |
| "eval_wer": 33.804554079696395, |
| "step": 5000 |
| }, |
| { |
| "epoch": 83.33, |
| "step": 5000, |
| "total_flos": 9.148503492624384e+19, |
| "train_loss": 0.06885672171264887, |
| "train_runtime": 14920.221, |
| "train_samples_per_second": 21.447, |
| "train_steps_per_second": 0.335 |
| } |
| ], |
| "max_steps": 5000, |
| "num_train_epochs": 84, |
| "total_flos": 9.148503492624384e+19, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|