{ "best_metric": 92.65656565656566, "best_model_checkpoint": "./whisper-bangla-dialect/checkpoint-1128", "epoch": 5.984084880636605, "eval_steps": 500, "global_step": 1128, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.26525198938992045, "grad_norm": 0.8929822444915771, "learning_rate": 4.000000000000001e-06, "loss": 1.4277, "step": 50 }, { "epoch": 0.5305039787798409, "grad_norm": 1.3129897117614746, "learning_rate": 8.000000000000001e-06, "loss": 1.4261, "step": 100 }, { "epoch": 0.7957559681697612, "grad_norm": 2.1221041679382324, "learning_rate": 1.2e-05, "loss": 1.4282, "step": 150 }, { "epoch": 0.9973474801061007, "eval_loss": 1.4541715383529663, "eval_norm_levenshtein_similarity": 90.81818181818181, "eval_runtime": 234.1728, "eval_samples_per_second": 1.431, "eval_steps_per_second": 0.359, "eval_wer": 12.320441988950275, "step": 188 }, { "epoch": 1.0610079575596818, "grad_norm": 0.727924108505249, "learning_rate": 1.6000000000000003e-05, "loss": 1.428, "step": 200 }, { "epoch": 1.3262599469496021, "grad_norm": 1.6670674085617065, "learning_rate": 2e-05, "loss": 1.429, "step": 250 }, { "epoch": 1.5915119363395225, "grad_norm": 1.1281551122665405, "learning_rate": 1.9840389289753895e-05, "loss": 1.4315, "step": 300 }, { "epoch": 1.8567639257294428, "grad_norm": 1.7024176120758057, "learning_rate": 1.936665227478063e-05, "loss": 1.4319, "step": 350 }, { "epoch": 2.0, "eval_loss": 1.4596483707427979, "eval_norm_levenshtein_similarity": 90.71717171717172, "eval_runtime": 231.2305, "eval_samples_per_second": 1.449, "eval_steps_per_second": 0.363, "eval_wer": 12.154696132596685, "step": 377 }, { "epoch": 2.1220159151193636, "grad_norm": 1.5983550548553467, "learning_rate": 1.8593911655366154e-05, "loss": 1.43, "step": 400 }, { "epoch": 2.387267904509284, "grad_norm": 1.2418725490570068, "learning_rate": 1.754683496733062e-05, "loss": 1.4283, "step": 450 }, { "epoch": 2.6525198938992043, "grad_norm": 1.0546443462371826, "learning_rate": 1.625884714144593e-05, "loss": 1.4274, "step": 500 }, { "epoch": 2.9177718832891246, "grad_norm": 0.7676965594291687, "learning_rate": 1.4771063508047636e-05, "loss": 1.4256, "step": 550 }, { "epoch": 2.9973474801061006, "eval_loss": 1.4547041654586792, "eval_norm_levenshtein_similarity": 90.95959595959596, "eval_runtime": 231.2052, "eval_samples_per_second": 1.449, "eval_steps_per_second": 0.363, "eval_wer": 12.099447513812155, "step": 565 }, { "epoch": 3.183023872679045, "grad_norm": 0.8181115984916687, "learning_rate": 1.3130977307619595e-05, "loss": 1.4236, "step": 600 }, { "epoch": 3.4482758620689653, "grad_norm": 1.0643696784973145, "learning_rate": 1.1390943604824827e-05, "loss": 1.4239, "step": 650 }, { "epoch": 3.713527851458886, "grad_norm": 1.1705901622772217, "learning_rate": 9.606508002694387e-06, "loss": 1.4229, "step": 700 }, { "epoch": 3.9787798408488064, "grad_norm": 0.8113660216331482, "learning_rate": 7.83463350799717e-06, "loss": 1.4218, "step": 750 }, { "epoch": 4.0, "eval_loss": 1.4521433115005493, "eval_norm_levenshtein_similarity": 91.84848484848484, "eval_runtime": 225.5507, "eval_samples_per_second": 1.485, "eval_steps_per_second": 0.372, "eval_wer": 10.718232044198896, "step": 754 }, { "epoch": 4.244031830238727, "grad_norm": 0.4397776126861572, "learning_rate": 6.131882150046291e-06, "loss": 1.4204, "step": 800 }, { "epoch": 4.5092838196286475, "grad_norm": 0.6042605042457581, "learning_rate": 4.552609399564763e-06, "loss": 1.4202, "step": 850 }, { "epoch": 4.774535809018568, "grad_norm": 0.3798394799232483, "learning_rate": 3.1472290256279227e-06, "loss": 1.4198, "step": 900 }, { "epoch": 4.997347480106101, "eval_loss": 1.4484983682632446, "eval_norm_levenshtein_similarity": 92.35353535353535, "eval_runtime": 229.5151, "eval_samples_per_second": 1.46, "eval_steps_per_second": 0.366, "eval_wer": 10.276243093922652, "step": 942 }, { "epoch": 5.039787798408488, "grad_norm": 0.8835919499397278, "learning_rate": 1.9606037801657674e-06, "loss": 1.4192, "step": 950 }, { "epoch": 5.305039787798409, "grad_norm": 0.6534780859947205, "learning_rate": 1.0306132828231318e-06, "loss": 1.4182, "step": 1000 }, { "epoch": 5.570291777188329, "grad_norm": 0.7457737326622009, "learning_rate": 3.8694482236061646e-07, "loss": 1.4185, "step": 1050 }, { "epoch": 5.835543766578249, "grad_norm": 0.349679559469223, "learning_rate": 5.0145674805708314e-08, "loss": 1.418, "step": 1100 }, { "epoch": 5.984084880636605, "eval_loss": 1.4477653503417969, "eval_norm_levenshtein_similarity": 92.65656565656566, "eval_runtime": 229.6795, "eval_samples_per_second": 1.459, "eval_steps_per_second": 0.366, "eval_wer": 9.779005524861878, "step": 1128 } ], "logging_steps": 50, "max_steps": 1128, "num_input_tokens_seen": 0, "num_train_epochs": 6, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.38442639917056e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }