Baselhany commited on
Commit
ab607e3
·
verified ·
1 Parent(s): 88eaa93

Training in progress, step 42000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8f4de11682a0c8e88d71359dafa5b34053b806b63c76bc17b578626c59d10d7
3
  size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc425d05484c4b99e7a21e5e88f1efa1777409c46771fd4a3fa63fe044be2eb5
3
  size 223144592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:60be441bde6c082a27eac961cb3a0e53373da53b188daae45ac42c8b9f832bbd
3
  size 281574266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea0c99f2a920cd5b3db1fe7e99fae1fcad95da2e7a5b7dfac0d182daaf15dfbe
3
  size 281574266
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8fd9a99c413ca4719ee97edd196fbd497aaa8fa62bf6aa7ecef1095789a7c684
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82e429a69160a7e218f681be79bc0a1dd4abcd753291d3cd09970598fd33c375
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8e0c332553653cd9d4981e38bf20cbc0d49f17070be19495c137d21df729fecf
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4858df2f172563cac3713edd793643c1b95146d389b9ae8616518d8c80c3945
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e2e5b2934f5267113b7be9d670e6b6f99e8588ce7390623947ef0eb6778afb9f
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0aa6b63fe8f302c8544aed592f7c3cd8ff925a68ee9acd3a9154b7f528575009
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 32400,
3
  "best_metric": 0.19510757441791923,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-32400",
5
- "epoch": 29.647781934794228,
6
  "eval_steps": 400,
7
- "global_step": 41600,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -3856,6 +3856,43 @@
3856
  "eval_steps_per_second": 0.433,
3857
  "eval_wer": 0.19820218096080164,
3858
  "step": 41600
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3859
  }
3860
  ],
3861
  "logging_steps": 100,
@@ -3875,7 +3912,7 @@
3875
  "attributes": {}
3876
  }
3877
  },
3878
- "total_flos": 5.408482222669824e+19,
3879
  "train_batch_size": 8,
3880
  "trial_name": null,
3881
  "trial_params": null
 
2
  "best_global_step": 32400,
3
  "best_metric": 0.19510757441791923,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-32400",
5
+ "epoch": 29.932834491359344,
6
  "eval_steps": 400,
7
+ "global_step": 42000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
3856
  "eval_steps_per_second": 0.433,
3857
  "eval_wer": 0.19820218096080164,
3858
  "step": 41600
3859
+ },
3860
+ {
3861
+ "epoch": 29.719045073935508,
3862
+ "grad_norm": 11.282709121704102,
3863
+ "learning_rate": 9.930271699927866e-07,
3864
+ "loss": 0.9995,
3865
+ "step": 41700
3866
+ },
3867
+ {
3868
+ "epoch": 29.790308213076784,
3869
+ "grad_norm": 10.81017780303955,
3870
+ "learning_rate": 7.525847559509498e-07,
3871
+ "loss": 0.9876,
3872
+ "step": 41800
3873
+ },
3874
+ {
3875
+ "epoch": 29.861571352218064,
3876
+ "grad_norm": 11.903881072998047,
3877
+ "learning_rate": 5.121423419091128e-07,
3878
+ "loss": 0.9524,
3879
+ "step": 41900
3880
+ },
3881
+ {
3882
+ "epoch": 29.932834491359344,
3883
+ "grad_norm": 11.278658866882324,
3884
+ "learning_rate": 2.7410435200769414e-07,
3885
+ "loss": 0.9491,
3886
+ "step": 42000
3887
+ },
3888
+ {
3889
+ "epoch": 29.932834491359344,
3890
+ "eval_loss": 0.09202782809734344,
3891
+ "eval_runtime": 145.311,
3892
+ "eval_samples_per_second": 3.441,
3893
+ "eval_steps_per_second": 0.434,
3894
+ "eval_wer": 0.19908635425876806,
3895
+ "step": 42000
3896
  }
3897
  ],
3898
  "logging_steps": 100,
 
3912
  "attributes": {}
3913
  }
3914
  },
3915
+ "total_flos": 5.460515080372224e+19,
3916
  "train_batch_size": 8,
3917
  "trial_name": null,
3918
  "trial_params": null