| { |
| "best_metric": 49.03147699757869, |
| "best_model_checkpoint": "./checkpoint-400", |
| "epoch": 9.51764705882353, |
| "global_step": 400, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.24, |
| "learning_rate": 2.666666666666667e-06, |
| "loss": 4.3134, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 6e-06, |
| "loss": 2.9329, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 9.333333333333334e-06, |
| "loss": 2.0584, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 9.829787234042554e-06, |
| "loss": 1.566, |
| "step": 40 |
| }, |
| { |
| "epoch": 1.19, |
| "learning_rate": 9.617021276595745e-06, |
| "loss": 1.3777, |
| "step": 50 |
| }, |
| { |
| "epoch": 1.42, |
| "learning_rate": 9.404255319148937e-06, |
| "loss": 1.1469, |
| "step": 60 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 9.191489361702128e-06, |
| "loss": 1.0638, |
| "step": 70 |
| }, |
| { |
| "epoch": 1.89, |
| "learning_rate": 8.97872340425532e-06, |
| "loss": 0.9974, |
| "step": 80 |
| }, |
| { |
| "epoch": 2.14, |
| "learning_rate": 8.765957446808512e-06, |
| "loss": 0.9615, |
| "step": 90 |
| }, |
| { |
| "epoch": 2.38, |
| "learning_rate": 8.553191489361703e-06, |
| "loss": 0.8262, |
| "step": 100 |
| }, |
| { |
| "epoch": 2.38, |
| "eval_cer": 149.33157314260887, |
| "eval_loss": 0.8188337683677673, |
| "eval_runtime": 459.9395, |
| "eval_samples_per_second": 1.113, |
| "eval_steps_per_second": 0.139, |
| "eval_wer": 146.85230024213075, |
| "step": 100 |
| }, |
| { |
| "epoch": 2.61, |
| "learning_rate": 8.340425531914894e-06, |
| "loss": 0.7986, |
| "step": 110 |
| }, |
| { |
| "epoch": 2.85, |
| "learning_rate": 8.127659574468085e-06, |
| "loss": 0.7707, |
| "step": 120 |
| }, |
| { |
| "epoch": 3.09, |
| "learning_rate": 7.914893617021278e-06, |
| "loss": 0.7623, |
| "step": 130 |
| }, |
| { |
| "epoch": 3.33, |
| "learning_rate": 7.702127659574469e-06, |
| "loss": 0.6451, |
| "step": 140 |
| }, |
| { |
| "epoch": 3.56, |
| "learning_rate": 7.48936170212766e-06, |
| "loss": 0.6205, |
| "step": 150 |
| }, |
| { |
| "epoch": 3.8, |
| "learning_rate": 7.2765957446808524e-06, |
| "loss": 0.6097, |
| "step": 160 |
| }, |
| { |
| "epoch": 4.05, |
| "learning_rate": 7.0638297872340434e-06, |
| "loss": 0.6217, |
| "step": 170 |
| }, |
| { |
| "epoch": 4.28, |
| "learning_rate": 6.8510638297872344e-06, |
| "loss": 0.4584, |
| "step": 180 |
| }, |
| { |
| "epoch": 4.52, |
| "learning_rate": 6.6382978723404254e-06, |
| "loss": 0.4929, |
| "step": 190 |
| }, |
| { |
| "epoch": 4.75, |
| "learning_rate": 6.425531914893618e-06, |
| "loss": 0.4843, |
| "step": 200 |
| }, |
| { |
| "epoch": 4.75, |
| "eval_cer": 22.099862972494236, |
| "eval_loss": 0.6699215173721313, |
| "eval_runtime": 1062.3991, |
| "eval_samples_per_second": 0.482, |
| "eval_steps_per_second": 0.06, |
| "eval_wer": 52.66343825665859, |
| "step": 200 |
| }, |
| { |
| "epoch": 4.99, |
| "learning_rate": 6.212765957446809e-06, |
| "loss": 0.4678, |
| "step": 210 |
| }, |
| { |
| "epoch": 5.24, |
| "learning_rate": 6e-06, |
| "loss": 0.4014, |
| "step": 220 |
| }, |
| { |
| "epoch": 5.47, |
| "learning_rate": 5.787234042553191e-06, |
| "loss": 0.3926, |
| "step": 230 |
| }, |
| { |
| "epoch": 5.71, |
| "learning_rate": 5.574468085106384e-06, |
| "loss": 0.3806, |
| "step": 240 |
| }, |
| { |
| "epoch": 5.94, |
| "learning_rate": 5.361702127659575e-06, |
| "loss": 0.3855, |
| "step": 250 |
| }, |
| { |
| "epoch": 6.19, |
| "learning_rate": 5.148936170212766e-06, |
| "loss": 0.3478, |
| "step": 260 |
| }, |
| { |
| "epoch": 6.42, |
| "learning_rate": 4.936170212765958e-06, |
| "loss": 0.3078, |
| "step": 270 |
| }, |
| { |
| "epoch": 6.66, |
| "learning_rate": 4.7234042553191496e-06, |
| "loss": 0.3147, |
| "step": 280 |
| }, |
| { |
| "epoch": 6.89, |
| "learning_rate": 4.5106382978723406e-06, |
| "loss": 0.3213, |
| "step": 290 |
| }, |
| { |
| "epoch": 7.14, |
| "learning_rate": 4.297872340425532e-06, |
| "loss": 0.287, |
| "step": 300 |
| }, |
| { |
| "epoch": 7.14, |
| "eval_cer": 20.677784833394604, |
| "eval_loss": 0.6913720369338989, |
| "eval_runtime": 1003.603, |
| "eval_samples_per_second": 0.51, |
| "eval_steps_per_second": 0.064, |
| "eval_wer": 50.13619854721549, |
| "step": 300 |
| }, |
| { |
| "epoch": 7.38, |
| "learning_rate": 4.085106382978723e-06, |
| "loss": 0.2602, |
| "step": 310 |
| }, |
| { |
| "epoch": 7.61, |
| "learning_rate": 3.872340425531915e-06, |
| "loss": 0.2612, |
| "step": 320 |
| }, |
| { |
| "epoch": 7.85, |
| "learning_rate": 3.6595744680851063e-06, |
| "loss": 0.2618, |
| "step": 330 |
| }, |
| { |
| "epoch": 8.09, |
| "learning_rate": 3.446808510638298e-06, |
| "loss": 0.2494, |
| "step": 340 |
| }, |
| { |
| "epoch": 8.33, |
| "learning_rate": 3.23404255319149e-06, |
| "loss": 0.2123, |
| "step": 350 |
| }, |
| { |
| "epoch": 8.56, |
| "learning_rate": 3.021276595744681e-06, |
| "loss": 0.2173, |
| "step": 360 |
| }, |
| { |
| "epoch": 8.8, |
| "learning_rate": 2.808510638297873e-06, |
| "loss": 0.2093, |
| "step": 370 |
| }, |
| { |
| "epoch": 9.05, |
| "learning_rate": 2.595744680851064e-06, |
| "loss": 0.2209, |
| "step": 380 |
| }, |
| { |
| "epoch": 9.28, |
| "learning_rate": 2.3829787234042557e-06, |
| "loss": 0.1742, |
| "step": 390 |
| }, |
| { |
| "epoch": 9.52, |
| "learning_rate": 2.170212765957447e-06, |
| "loss": 0.1828, |
| "step": 400 |
| }, |
| { |
| "epoch": 9.52, |
| "eval_cer": 20.460546104742487, |
| "eval_loss": 0.7369570136070251, |
| "eval_runtime": 1004.0546, |
| "eval_samples_per_second": 0.51, |
| "eval_steps_per_second": 0.064, |
| "eval_wer": 49.03147699757869, |
| "step": 400 |
| } |
| ], |
| "max_steps": 500, |
| "num_train_epochs": 12, |
| "total_flos": 7.44752342605824e+18, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|