Baselhany commited on
Commit
47dc569
·
verified ·
1 Parent(s): 4d32390

Training in progress, step 32800, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:720ec26f34508162cd3ed48aa354f1a55bde3c940e3b941db4493d80be90061c
3
  size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5c9a0489a2cf843b764d138398df3a7a6a5d1d0e8773f219c72b4c09ccfae61
3
  size 223144592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:288024677ae015fd9f07724ebcc4a5c2dde656deb4c2988d4acbc6416d681ba4
3
  size 281574266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db6620df8d73410ffb0bf9409a6ade7370440551ac5d5b4428a111e658eecd44
3
  size 281574266
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:732b979418123e1232daa0215db26fcf01224d94c16971ed415238028e85e575
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28cbe766e4153cd167aa13892e6d8f2b2fb01f0cf38c806d1ede3210aea189b8
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:676a0c92f411f549dea3afe0b01db3d2c7fd7a67a5bdccc0755488565188383a
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ef2649ceb258c6256acc0ae4c4605bb5b2c0540ea4c5d0deec7e198b382fb7a
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:50ca82bbea02fc725f68bafbc19bec35c56c4c96823814172201a68e0f179761
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51726f789396c230f9beb438f4af63a6120e53498ea8e472d26397da95a5f066
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 32400,
3
  "best_metric": 0.19510757441791923,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-32400",
5
- "epoch": 23.0890789239266,
6
  "eval_steps": 400,
7
- "global_step": 32400,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -3005,6 +3005,43 @@
3005
  "eval_steps_per_second": 0.396,
3006
  "eval_wer": 0.19510757441791923,
3007
  "step": 32400
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3008
  }
3009
  ],
3010
  "logging_steps": 100,
@@ -3024,7 +3061,7 @@
3024
  "attributes": {}
3025
  }
3026
  },
3027
- "total_flos": 4.212315930230784e+19,
3028
  "train_batch_size": 8,
3029
  "trial_name": null,
3030
  "trial_params": null
 
2
  "best_global_step": 32400,
3
  "best_metric": 0.19510757441791923,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-32400",
5
+ "epoch": 23.374131480491716,
6
  "eval_steps": 400,
7
+ "global_step": 32800,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
3005
  "eval_steps_per_second": 0.396,
3006
  "eval_wer": 0.19510757441791923,
3007
  "step": 32400
3008
+ },
3009
+ {
3010
+ "epoch": 23.16034206306788,
3011
+ "grad_norm": 15.528765678405762,
3012
+ "learning_rate": 7.502530730296457e-06,
3013
+ "loss": 0.9855,
3014
+ "step": 32500
3015
+ },
3016
+ {
3017
+ "epoch": 23.231605202209156,
3018
+ "grad_norm": 11.74361801147461,
3019
+ "learning_rate": 7.213304410701374e-06,
3020
+ "loss": 1.0193,
3021
+ "step": 32600
3022
+ },
3023
+ {
3024
+ "epoch": 23.302868341350436,
3025
+ "grad_norm": 11.150557518005371,
3026
+ "learning_rate": 6.924078091106291e-06,
3027
+ "loss": 0.9878,
3028
+ "step": 32700
3029
+ },
3030
+ {
3031
+ "epoch": 23.374131480491716,
3032
+ "grad_norm": 11.871788024902344,
3033
+ "learning_rate": 6.634851771511208e-06,
3034
+ "loss": 1.0443,
3035
+ "step": 32800
3036
+ },
3037
+ {
3038
+ "epoch": 23.374131480491716,
3039
+ "eval_loss": 0.09271341562271118,
3040
+ "eval_runtime": 157.7057,
3041
+ "eval_samples_per_second": 3.17,
3042
+ "eval_steps_per_second": 0.399,
3043
+ "eval_wer": 0.19952844090775126,
3044
+ "step": 32800
3045
  }
3046
  ],
3047
  "logging_steps": 100,
 
3061
  "attributes": {}
3062
  }
3063
  },
3064
+ "total_flos": 4.264348787933184e+19,
3065
  "train_batch_size": 8,
3066
  "trial_name": null,
3067
  "trial_params": null