File size: 2,593 Bytes
09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d 70ab39c 09fe84d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 |
{
"best_metric": 39.43644074527058,
"best_model_checkpoint": "./whisper-lora-15k-adapters/checkpoint-237",
"epoch": 0.2775175644028103,
"eval_steps": 237,
"global_step": 237,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.02927400468384075,
"grad_norm": 0.616703450679779,
"learning_rate": 0.0005,
"loss": 0.8917,
"step": 25
},
{
"epoch": 0.0585480093676815,
"grad_norm": 0.4405761659145355,
"learning_rate": 0.001,
"loss": 0.736,
"step": 50
},
{
"epoch": 0.08782201405152225,
"grad_norm": 0.4590846598148346,
"learning_rate": 0.0009940758293838863,
"loss": 0.7102,
"step": 75
},
{
"epoch": 0.117096018735363,
"grad_norm": 0.4082449972629547,
"learning_rate": 0.0009881516587677726,
"loss": 0.6302,
"step": 100
},
{
"epoch": 0.14637002341920374,
"grad_norm": 0.3845512866973877,
"learning_rate": 0.0009822274881516586,
"loss": 0.6538,
"step": 125
},
{
"epoch": 0.1756440281030445,
"grad_norm": 0.47404810786247253,
"learning_rate": 0.000976303317535545,
"loss": 0.6261,
"step": 150
},
{
"epoch": 0.20491803278688525,
"grad_norm": 0.38350749015808105,
"learning_rate": 0.0009703791469194313,
"loss": 0.6408,
"step": 175
},
{
"epoch": 0.234192037470726,
"grad_norm": 0.4432656466960907,
"learning_rate": 0.0009644549763033176,
"loss": 0.5789,
"step": 200
},
{
"epoch": 0.26346604215456676,
"grad_norm": 0.4181855618953705,
"learning_rate": 0.0009585308056872039,
"loss": 0.5673,
"step": 225
},
{
"epoch": 0.2775175644028103,
"eval_loss": 0.6236673593521118,
"eval_runtime": 10166.9784,
"eval_samples_per_second": 0.149,
"eval_steps_per_second": 0.009,
"eval_wer": 39.43644074527058,
"step": 237
}
],
"logging_steps": 25,
"max_steps": 4270,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 237,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3.92166536380416e+18,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}
|