Baselhany commited on
Commit
badda50
·
verified ·
1 Parent(s): 85da1da

Training in progress, step 28000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e8697dd2e30554939f3fdca476e8f8c14ef00252787d6ed337c4f9c24c9f4e2e
3
  size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e9c5b5c507c25afe3cf4d5aec82c7836af315794f646ccd78331b1ef69b9976
3
  size 223144592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f55a81dbcdab168b5b56232e68b26448a34e74f5b0c7df45184f5e18e31d7e3
3
  size 281574266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f37dd8023aeda17192fa3cd957ae713351d08b56ea14a8b05ae51168a5029d04
3
  size 281574266
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:32de07988befbf143be2b9e32a6f8bc0b1b31db8931faad85d1d92ab132d45e6
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2940908599d69a8410ea56cae7b048228912580fd3318d429840a0c58b1c925b
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a9b4a296cdf5148ddd1594a16e140d1d657863c4c29f02fff39c8abb2ac48881
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbd563d83932cc7d0da7780f68883324db1a55e2e28771f70432de2590ff8acb
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ce753046bc02fca905121dc376e6ab4266dc1a838826882290e3881ef7934be
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a62d34f5456d530d45292bdc7a7b57a95952306ee03d49e82ed6ce0cb924518d
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 25000,
3
- "best_metric": 0.18243442381373415,
4
- "best_model_checkpoint": "./distil-whisper/checkpoint-25000",
5
- "epoch": 16.0,
6
  "eval_steps": 1000,
7
- "global_step": 27312,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2162,12 +2162,70 @@
2162
  "learning_rate": 1.0070117857675669e-07,
2163
  "loss": 0.7324,
2164
  "step": 27300
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2165
  }
2166
  ],
2167
  "logging_steps": 100,
2168
- "max_steps": 27312,
2169
  "num_input_tokens_seen": 0,
2170
- "num_train_epochs": 16,
2171
  "save_steps": 1000,
2172
  "stateful_callbacks": {
2173
  "TrainerControl": {
@@ -2176,12 +2234,12 @@
2176
  "should_evaluate": false,
2177
  "should_log": false,
2178
  "should_save": true,
2179
- "should_training_stop": true
2180
  },
2181
  "attributes": {}
2182
  }
2183
  },
2184
- "total_flos": 3.55254335963136e+19,
2185
  "train_batch_size": 8,
2186
  "trial_name": null,
2187
  "trial_params": null
 
1
  {
2
+ "best_global_step": 28000,
3
+ "best_metric": 0.18110816386678455,
4
+ "best_model_checkpoint": "./distil-whisper/checkpoint-28000",
5
+ "epoch": 16.403046280023432,
6
  "eval_steps": 1000,
7
+ "global_step": 28000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2162
  "learning_rate": 1.0070117857675669e-07,
2163
  "loss": 0.7324,
2164
  "step": 27300
2165
+ },
2166
+ {
2167
+ "epoch": 16.05155243116579,
2168
+ "grad_norm": 12.168989181518555,
2169
+ "learning_rate": 2.7443730771306743e-05,
2170
+ "loss": 0.7807,
2171
+ "step": 27400
2172
+ },
2173
+ {
2174
+ "epoch": 16.11013473930873,
2175
+ "grad_norm": 5.686388969421387,
2176
+ "learning_rate": 2.717385437469639e-05,
2177
+ "loss": 0.8098,
2178
+ "step": 27500
2179
+ },
2180
+ {
2181
+ "epoch": 16.16871704745167,
2182
+ "grad_norm": 6.920952320098877,
2183
+ "learning_rate": 2.6903977978086036e-05,
2184
+ "loss": 0.7786,
2185
+ "step": 27600
2186
+ },
2187
+ {
2188
+ "epoch": 16.22729935559461,
2189
+ "grad_norm": 8.017998695373535,
2190
+ "learning_rate": 2.6634101581475683e-05,
2191
+ "loss": 0.8257,
2192
+ "step": 27700
2193
+ },
2194
+ {
2195
+ "epoch": 16.285881663737552,
2196
+ "grad_norm": 6.6050262451171875,
2197
+ "learning_rate": 2.6364225184865333e-05,
2198
+ "loss": 0.7671,
2199
+ "step": 27800
2200
+ },
2201
+ {
2202
+ "epoch": 16.344463971880494,
2203
+ "grad_norm": 8.146703720092773,
2204
+ "learning_rate": 2.609434878825498e-05,
2205
+ "loss": 0.7733,
2206
+ "step": 27900
2207
+ },
2208
+ {
2209
+ "epoch": 16.403046280023432,
2210
+ "grad_norm": 11.652145385742188,
2211
+ "learning_rate": 2.5824472391644626e-05,
2212
+ "loss": 0.7902,
2213
+ "step": 28000
2214
+ },
2215
+ {
2216
+ "epoch": 16.403046280023432,
2217
+ "eval_loss": 0.08416531980037689,
2218
+ "eval_runtime": 148.3349,
2219
+ "eval_samples_per_second": 3.371,
2220
+ "eval_steps_per_second": 0.425,
2221
+ "eval_wer": 0.18110816386678455,
2222
+ "step": 28000
2223
  }
2224
  ],
2225
  "logging_steps": 100,
2226
+ "max_steps": 37554,
2227
  "num_input_tokens_seen": 0,
2228
+ "num_train_epochs": 22,
2229
  "save_steps": 1000,
2230
  "stateful_callbacks": {
2231
  "TrainerControl": {
 
2234
  "should_evaluate": false,
2235
  "should_log": false,
2236
  "should_save": true,
2237
+ "should_training_stop": false
2238
  },
2239
  "attributes": {}
2240
  }
2241
  },
2242
+ "total_flos": 3.642039874879488e+19,
2243
  "train_batch_size": 8,
2244
  "trial_name": null,
2245
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8c88ae5031dc0d52713a2b3c363a1fb376926954db49624064f91e4b907e5be
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e81ddeed28465bb51d4919630f0406d970c910999565b6a4160cb0f11418cd99
3
  size 5496