Baselhany commited on
Commit
8811605
·
verified ·
1 Parent(s): 0c5256c

Training in progress, step 41200, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eb3428b5d7351f2f983a8817140c1c8b2e96b29566ec79f4307a4b81ec849b68
3
  size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6e2823f88f5f24074d58b21cfe18f66f131ac34667095575c4c40e4eae863e3
3
  size 223144592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:24a096e10e2fc8125e4596c05a55c5be93a424630c3a8d00b9012ce34e0eb121
3
  size 281574266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:337d3b57220580ed03ee608a9a877cb6776a0f7fb530d8a389436c6d77d26aed
3
  size 281574266
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0409b0246d35c8ef79cb2b0e43178bdf95a92462fbd64e2c9dfd213533047dc5
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d37da37f32b9da157d0f354c842374bbe28e2210ef8b4e1a2f62909ef09fdefb
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d0b922550ce28a3681322bb9777a8dd637293414e098f71686c9e28adbd7a3a6
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bbf1059b349dfce5e1aca93ba02e0a5c14f3145f54aaedb0a25d075e6e0aaa9
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ea63f0486f536d4a4ac9c36e06e08d63effd7c038d0a64ea9126e98bd07aa911
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7e2692c415195d4cb598a1230fff1568c572f90b174e77b60f1d07171445123
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 32400,
3
  "best_metric": 0.19510757441791923,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-32400",
5
- "epoch": 29.077676821663996,
6
  "eval_steps": 400,
7
- "global_step": 40800,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -3782,6 +3782,43 @@
3782
  "eval_steps_per_second": 0.434,
3783
  "eval_wer": 0.1976127320954907,
3784
  "step": 40800
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3785
  }
3786
  ],
3787
  "logging_steps": 100,
@@ -3801,7 +3838,7 @@
3801
  "attributes": {}
3802
  }
3803
  },
3804
- "total_flos": 5.304416507265024e+19,
3805
  "train_batch_size": 8,
3806
  "trial_name": null,
3807
  "trial_params": null
 
2
  "best_global_step": 32400,
3
  "best_metric": 0.19510757441791923,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-32400",
5
+ "epoch": 29.362729378229112,
6
  "eval_steps": 400,
7
+ "global_step": 41200,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
3782
  "eval_steps_per_second": 0.434,
3783
  "eval_wer": 0.1976127320954907,
3784
  "step": 40800
3785
+ },
3786
+ {
3787
+ "epoch": 29.148939960805272,
3788
+ "grad_norm": 12.47839641571045,
3789
+ "learning_rate": 2.9165664823274828e-06,
3790
+ "loss": 0.9257,
3791
+ "step": 40900
3792
+ },
3793
+ {
3794
+ "epoch": 29.220203099946552,
3795
+ "grad_norm": 9.78573989868164,
3796
+ "learning_rate": 2.676124068285646e-06,
3797
+ "loss": 0.9376,
3798
+ "step": 41000
3799
+ },
3800
+ {
3801
+ "epoch": 29.291466239087832,
3802
+ "grad_norm": 10.32386589050293,
3803
+ "learning_rate": 2.4356816542438087e-06,
3804
+ "loss": 0.9428,
3805
+ "step": 41100
3806
+ },
3807
+ {
3808
+ "epoch": 29.362729378229112,
3809
+ "grad_norm": 12.5963773727417,
3810
+ "learning_rate": 2.195239240201972e-06,
3811
+ "loss": 0.9519,
3812
+ "step": 41200
3813
+ },
3814
+ {
3815
+ "epoch": 29.362729378229112,
3816
+ "eval_loss": 0.09224098175764084,
3817
+ "eval_runtime": 145.4228,
3818
+ "eval_samples_per_second": 3.438,
3819
+ "eval_steps_per_second": 0.433,
3820
+ "eval_wer": 0.19687592101385204,
3821
+ "step": 41200
3822
  }
3823
  ],
3824
  "logging_steps": 100,
 
3838
  "attributes": {}
3839
  }
3840
  },
3841
+ "total_flos": 5.356449364967424e+19,
3842
  "train_batch_size": 8,
3843
  "trial_name": null,
3844
  "trial_params": null