Baselhany commited on
Commit
d885617
·
verified ·
1 Parent(s): 6a770df

Training in progress, step 34400, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:20bdfe64bb7787dbde06ed0249c104c2a75f74f77a4da7b6d17b0be3c907bffe
3
  size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c7bac0ecd11e105ba77bc6b9c1e49b56f9b9523b4862c6989ad212585b9e24c
3
  size 223144592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:55dd9570954d73b04e0eb476a55f9ed87ca0571abe03bdd05b897829def60e9b
3
  size 281574266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19981691c62f8f347e95b064d87dda28cd05190b739901bc88cf7a374983b649
3
  size 281574266
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:19d659484e215c1796fcedcf712444963aec532ba7ef9b9f53c66f0a4322a2d0
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce15f59f8cb64bc668dfb4e258be90d962b7c8172b3c45431bfac7361d50154f
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e82de6898de222ba34f053188f784b1daff292a8caca719707d751a738a4d585
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9dbb3e5d0fc7426e887c7892235a6eb2991774261a344c85f92b5b41d28ebaf
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8627c62832e908e5ee44d6dc9ec10410d33b0d8b29cb328e2941165b3e5c98eb
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee882eac40b50fbdc8d881b62f827901f35cd88d24357d996c81a66bb3343b11
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 32400,
3
  "best_metric": 0.19510757441791923,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-32400",
5
- "epoch": 24.228754676643508,
6
  "eval_steps": 400,
7
- "global_step": 34000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -3153,6 +3153,43 @@
3153
  "eval_steps_per_second": 0.393,
3154
  "eval_wer": 0.19790745652814618,
3155
  "step": 34000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3156
  }
3157
  ],
3158
  "logging_steps": 100,
@@ -3172,7 +3209,7 @@
3172
  "attributes": {}
3173
  }
3174
  },
3175
- "total_flos": 4.420329474097152e+19,
3176
  "train_batch_size": 8,
3177
  "trial_name": null,
3178
  "trial_params": null
 
2
  "best_global_step": 32400,
3
  "best_metric": 0.19510757441791923,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-32400",
5
+ "epoch": 24.513807233208624,
6
  "eval_steps": 400,
7
+ "global_step": 34400,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
3153
  "eval_steps_per_second": 0.393,
3154
  "eval_wer": 0.19790745652814618,
3155
  "step": 34000
3156
+ },
3157
+ {
3158
+ "epoch": 24.300017815784784,
3159
+ "grad_norm": 11.129683494567871,
3160
+ "learning_rate": 2.8778018799710775e-06,
3161
+ "loss": 0.9646,
3162
+ "step": 34100
3163
+ },
3164
+ {
3165
+ "epoch": 24.371280954926064,
3166
+ "grad_norm": 11.022905349731445,
3167
+ "learning_rate": 2.588575560375994e-06,
3168
+ "loss": 0.9873,
3169
+ "step": 34200
3170
+ },
3171
+ {
3172
+ "epoch": 24.442544094067344,
3173
+ "grad_norm": 9.839102745056152,
3174
+ "learning_rate": 2.299349240780911e-06,
3175
+ "loss": 0.9995,
3176
+ "step": 34300
3177
+ },
3178
+ {
3179
+ "epoch": 24.513807233208624,
3180
+ "grad_norm": 10.704620361328125,
3181
+ "learning_rate": 2.010122921185828e-06,
3182
+ "loss": 0.9504,
3183
+ "step": 34400
3184
+ },
3185
+ {
3186
+ "epoch": 24.513807233208624,
3187
+ "eval_loss": 0.09273621439933777,
3188
+ "eval_runtime": 158.2643,
3189
+ "eval_samples_per_second": 3.159,
3190
+ "eval_steps_per_second": 0.398,
3191
+ "eval_wer": 0.19599174771588565,
3192
+ "step": 34400
3193
  }
3194
  ],
3195
  "logging_steps": 100,
 
3209
  "attributes": {}
3210
  }
3211
  },
3212
+ "total_flos": 4.472362331799552e+19,
3213
  "train_batch_size": 8,
3214
  "trial_name": null,
3215
  "trial_params": null