Baselhany commited on
Commit
3d21634
·
verified ·
1 Parent(s): 219dd05

Training in progress, step 32400, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ee40e71245bf0f66c619c4566f44ff371a139d5158466fdc9d52c3e8ec2e3176
3
  size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:720ec26f34508162cd3ed48aa354f1a55bde3c940e3b941db4493d80be90061c
3
  size 223144592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:23d7f9f3b44eb0e3f4ac71847a56e56f764cdaa70a0e972cf685f8a504364deb
3
  size 281574266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:288024677ae015fd9f07724ebcc4a5c2dde656deb4c2988d4acbc6416d681ba4
3
  size 281574266
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:889fba2dda612c6c0d4296e738b798d52e27bd546723613a9f1022c81f0a1b37
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:732b979418123e1232daa0215db26fcf01224d94c16971ed415238028e85e575
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:01a2425ae56cb630c35a174629553276a10429b35fc53ac726db0dfb5fc49bea
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:676a0c92f411f549dea3afe0b01db3d2c7fd7a67a5bdccc0755488565188383a
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e27c322f011be2d3e98ac457c9a136957b37708f9a786651d2be8677d0d76424
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50ca82bbea02fc725f68bafbc19bec35c56c4c96823814172201a68e0f179761
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 18400,
3
- "best_metric": 0.19658119658119658,
4
- "best_model_checkpoint": "./distil-whisper/checkpoint-18400",
5
- "epoch": 22.80456084090504,
6
  "eval_steps": 400,
7
- "global_step": 32000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2968,6 +2968,43 @@
2968
  "eval_steps_per_second": 0.396,
2969
  "eval_wer": 0.201886236368995,
2970
  "step": 32000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2971
  }
2972
  ],
2973
  "logging_steps": 100,
@@ -2987,7 +3024,7 @@
2987
  "attributes": {}
2988
  }
2989
  },
2990
- "total_flos": 4.160400959471616e+19,
2991
  "train_batch_size": 8,
2992
  "trial_name": null,
2993
  "trial_params": null
 
1
  {
2
+ "best_global_step": 32400,
3
+ "best_metric": 0.19510757441791923,
4
+ "best_model_checkpoint": "./distil-whisper/checkpoint-32400",
5
+ "epoch": 23.0890789239266,
6
  "eval_steps": 400,
7
+ "global_step": 32400,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2968
  "eval_steps_per_second": 0.396,
2969
  "eval_wer": 0.201886236368995,
2970
  "step": 32000
2971
+ },
2972
+ {
2973
+ "epoch": 22.87582398004632,
2974
+ "grad_norm": 13.594120025634766,
2975
+ "learning_rate": 8.65943600867679e-06,
2976
+ "loss": 1.0777,
2977
+ "step": 32100
2978
+ },
2979
+ {
2980
+ "epoch": 22.9470871191876,
2981
+ "grad_norm": 15.224822044372559,
2982
+ "learning_rate": 8.370209689081707e-06,
2983
+ "loss": 1.0544,
2984
+ "step": 32200
2985
+ },
2986
+ {
2987
+ "epoch": 23.01781578478532,
2988
+ "grad_norm": 11.079718589782715,
2989
+ "learning_rate": 8.080983369486623e-06,
2990
+ "loss": 1.0563,
2991
+ "step": 32300
2992
+ },
2993
+ {
2994
+ "epoch": 23.0890789239266,
2995
+ "grad_norm": 12.417219161987305,
2996
+ "learning_rate": 7.791757049891541e-06,
2997
+ "loss": 0.9784,
2998
+ "step": 32400
2999
+ },
3000
+ {
3001
+ "epoch": 23.0890789239266,
3002
+ "eval_loss": 0.0926588624715805,
3003
+ "eval_runtime": 159.1992,
3004
+ "eval_samples_per_second": 3.141,
3005
+ "eval_steps_per_second": 0.396,
3006
+ "eval_wer": 0.19510757441791923,
3007
+ "step": 32400
3008
  }
3009
  ],
3010
  "logging_steps": 100,
 
3024
  "attributes": {}
3025
  }
3026
  },
3027
+ "total_flos": 4.212315930230784e+19,
3028
  "train_batch_size": 8,
3029
  "trial_name": null,
3030
  "trial_params": null