File size: 2,593 Bytes
09fe84d
70ab39c
 
09fe84d
 
 
 
 
 
 
 
 
70ab39c
09fe84d
70ab39c
09fe84d
 
 
 
70ab39c
09fe84d
70ab39c
09fe84d
 
 
 
70ab39c
09fe84d
70ab39c
09fe84d
 
 
 
70ab39c
09fe84d
70ab39c
09fe84d
 
 
 
70ab39c
09fe84d
70ab39c
09fe84d
 
 
 
70ab39c
09fe84d
70ab39c
09fe84d
 
 
 
70ab39c
09fe84d
70ab39c
09fe84d
 
 
 
70ab39c
09fe84d
70ab39c
09fe84d
 
 
 
70ab39c
09fe84d
70ab39c
09fe84d
 
 
 
70ab39c
 
 
 
 
09fe84d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
{
  "best_metric": 39.43644074527058,
  "best_model_checkpoint": "./whisper-lora-15k-adapters/checkpoint-237",
  "epoch": 0.2775175644028103,
  "eval_steps": 237,
  "global_step": 237,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.02927400468384075,
      "grad_norm": 0.616703450679779,
      "learning_rate": 0.0005,
      "loss": 0.8917,
      "step": 25
    },
    {
      "epoch": 0.0585480093676815,
      "grad_norm": 0.4405761659145355,
      "learning_rate": 0.001,
      "loss": 0.736,
      "step": 50
    },
    {
      "epoch": 0.08782201405152225,
      "grad_norm": 0.4590846598148346,
      "learning_rate": 0.0009940758293838863,
      "loss": 0.7102,
      "step": 75
    },
    {
      "epoch": 0.117096018735363,
      "grad_norm": 0.4082449972629547,
      "learning_rate": 0.0009881516587677726,
      "loss": 0.6302,
      "step": 100
    },
    {
      "epoch": 0.14637002341920374,
      "grad_norm": 0.3845512866973877,
      "learning_rate": 0.0009822274881516586,
      "loss": 0.6538,
      "step": 125
    },
    {
      "epoch": 0.1756440281030445,
      "grad_norm": 0.47404810786247253,
      "learning_rate": 0.000976303317535545,
      "loss": 0.6261,
      "step": 150
    },
    {
      "epoch": 0.20491803278688525,
      "grad_norm": 0.38350749015808105,
      "learning_rate": 0.0009703791469194313,
      "loss": 0.6408,
      "step": 175
    },
    {
      "epoch": 0.234192037470726,
      "grad_norm": 0.4432656466960907,
      "learning_rate": 0.0009644549763033176,
      "loss": 0.5789,
      "step": 200
    },
    {
      "epoch": 0.26346604215456676,
      "grad_norm": 0.4181855618953705,
      "learning_rate": 0.0009585308056872039,
      "loss": 0.5673,
      "step": 225
    },
    {
      "epoch": 0.2775175644028103,
      "eval_loss": 0.6236673593521118,
      "eval_runtime": 10166.9784,
      "eval_samples_per_second": 0.149,
      "eval_steps_per_second": 0.009,
      "eval_wer": 39.43644074527058,
      "step": 237
    }
  ],
  "logging_steps": 25,
  "max_steps": 4270,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 5,
  "save_steps": 237,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": false
      },
      "attributes": {}
    }
  },
  "total_flos": 3.92166536380416e+18,
  "train_batch_size": 16,
  "trial_name": null,
  "trial_params": null
}