{ "best_global_step": 2000, "best_metric": 42.951101658018516, "best_model_checkpoint": "./whisper-small-yoruba-438h/checkpoint-2000", "epoch": 0.7312614259597806, "eval_steps": 2000, "global_step": 2000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.18281535648994515, "grad_norm": 23.418228149414062, "learning_rate": 2.4950000000000003e-06, "loss": 8.61669921875, "step": 500 }, { "epoch": 0.3656307129798903, "grad_norm": 18.097488403320312, "learning_rate": 4.9950000000000005e-06, "loss": 2.76252197265625, "step": 1000 }, { "epoch": 0.5484460694698354, "grad_norm": 14.418298721313477, "learning_rate": 4.980903172380229e-06, "loss": 1.8946580810546876, "step": 1500 }, { "epoch": 0.7312614259597806, "grad_norm": 13.605561256408691, "learning_rate": 4.9237526483866514e-06, "loss": 1.6098297119140625, "step": 2000 }, { "epoch": 0.7312614259597806, "eval_loss": 0.37521448731422424, "eval_runtime": 6135.7464, "eval_samples_per_second": 1.585, "eval_steps_per_second": 0.198, "eval_wer": 42.951101658018516, "step": 2000 } ], "logging_steps": 500, "max_steps": 13675, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 2000, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 3, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.846946562048e+19, "train_batch_size": 8, "trial_name": null, "trial_params": null }