Baselhany commited on
Commit
134c2bf
·
verified ·
1 Parent(s): 8607c73

Training in progress, step 9200, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eb5d7da1b209571a42761f2acda70045e87f2f5393538333c69aa6be5730a682
3
  size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2634ee5c91e47c9d9438951c5a31b6505f921ccdc3c0a21f3663947156853947
3
  size 223144592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4045b8c929a0fb9e7576abf8eb3f4b952fdd4666d17e7350e182722fbc30d896
3
  size 281574266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75bb3146d72d9dd817a812afea831681c0b29bc8550b271772873f4b41dbd457
3
  size 281574266
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:573b8846751747c8c9f821d740773f26bdc18311fd829c492aee93a262a7fd26
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e464bae5ad4106f5a3627e7d672fcdac62ea5f98f2885a50ca5598f6f6382a9
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:688eaddc6ab64410ae7ae3de0b5c007c2fdc95cc8779ec9261dd523d672f7de3
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a55405e1660e5257b9afd90e5fb43c8cb59734b3562afe4dea9642684f8e6541
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:07c6d599d4044e43352a07d178cf8318fcc3fe97bb451adab81781c82561eae9
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba8b9029ae05745a1673f92384ec22df6e7e540c2703e5221e20ea21bb31a083
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 4000,
3
  "best_metric": 0.18950781019746538,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-4000",
5
- "epoch": 5.216360403082395,
6
  "eval_steps": 400,
7
- "global_step": 8800,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -822,6 +822,43 @@
822
  "eval_steps_per_second": 0.425,
823
  "eval_wer": 0.19422340111995284,
824
  "step": 8800
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
825
  }
826
  ],
827
  "logging_steps": 100,
@@ -841,7 +878,7 @@
841
  "attributes": {}
842
  }
843
  },
844
- "total_flos": 1.14458059210752e+19,
845
  "train_batch_size": 8,
846
  "trial_name": null,
847
  "trial_params": null
 
2
  "best_global_step": 4000,
3
  "best_metric": 0.18950781019746538,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-4000",
5
+ "epoch": 5.453467694131595,
6
  "eval_steps": 400,
7
+ "global_step": 9200,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
822
  "eval_steps_per_second": 0.425,
823
  "eval_wer": 0.19422340111995284,
824
  "step": 8800
825
+ },
826
+ {
827
+ "epoch": 5.275637225844695,
828
+ "grad_norm": 21.900598526000977,
829
+ "learning_rate": 3.543397968605725e-05,
830
+ "loss": 1.3348,
831
+ "step": 8900
832
+ },
833
+ {
834
+ "epoch": 5.334914048606994,
835
+ "grad_norm": 11.572765350341797,
836
+ "learning_rate": 3.4664512157586956e-05,
837
+ "loss": 1.3409,
838
+ "step": 9000
839
+ },
840
+ {
841
+ "epoch": 5.394190871369295,
842
+ "grad_norm": 9.629471778869629,
843
+ "learning_rate": 3.389504462911665e-05,
844
+ "loss": 1.3919,
845
+ "step": 9100
846
+ },
847
+ {
848
+ "epoch": 5.453467694131595,
849
+ "grad_norm": 12.661375045776367,
850
+ "learning_rate": 3.3125577100646356e-05,
851
+ "loss": 1.3186,
852
+ "step": 9200
853
+ },
854
+ {
855
+ "epoch": 5.453467694131595,
856
+ "eval_loss": 0.09043103456497192,
857
+ "eval_runtime": 145.9203,
858
+ "eval_samples_per_second": 3.427,
859
+ "eval_steps_per_second": 0.432,
860
+ "eval_wer": 0.1943707633362806,
861
+ "step": 9200
862
  }
863
  ],
864
  "logging_steps": 100,
 
878
  "attributes": {}
879
  }
880
  },
881
+ "total_flos": 1.19661344980992e+19,
882
  "train_batch_size": 8,
883
  "trial_name": null,
884
  "trial_params": null