Baselhany commited on
Commit
69618e3
·
verified ·
1 Parent(s): 2e9e645

Training in progress, step 35000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:396a29a5a4787dbf082bfc1745fc5d4dbaf49c674cf87663f742af78ff889ac4
3
  size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c40ac5768678ec5e02f1ef67d5a4309a7523549c081d4aa4dcec3a9e62775b9
3
  size 223144592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2a681bae809c2c7355b83c2ba73ac835d0bd5c7c190fd2e0484e22164316f803
3
  size 281574266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04e97e0d3303ce9166cbde3be58ec202701027a873651cfd877e10f3438e1bef
3
  size 281574266
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ce2c6e00c180a2aec543a6c39f38b41238eae7e450c89c57b59b23abde749eed
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9bbeae69b0c3ccad7b2da286244601c7cab8dcaae00fabe926ee3c6e16123c7
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d82f68a924bc6f264caa5b5167990f1fdb3b2c8f9254156d8c0f9ccc3c5489db
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f23d0e9ac8ef166c949e157928330a7cb0661782ade2d248d1a928b2ac29805e
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:122a12c57bcb396697297c9c6111a0a8854f31fea532275de745584ad3eee2c3
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d08ed08052d252f71cbe4b116766edbb8b5b9589f8f40bb95cb4854fc9cf848
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 28000,
3
  "best_metric": 0.18110816386678455,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-28000",
5
- "epoch": 19.917984768599883,
6
  "eval_steps": 1000,
7
- "global_step": 34000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2694,6 +2694,85 @@
2694
  "eval_steps_per_second": 0.425,
2695
  "eval_wer": 0.18715001473622164,
2696
  "step": 34000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2697
  }
2698
  ],
2699
  "logging_steps": 100,
@@ -2713,7 +2792,7 @@
2713
  "attributes": {}
2714
  }
2715
  },
2716
- "total_flos": 4.422483959611392e+19,
2717
  "train_batch_size": 8,
2718
  "trial_name": null,
2719
  "trial_params": null
 
2
  "best_global_step": 28000,
3
  "best_metric": 0.18110816386678455,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-28000",
5
+ "epoch": 20.50380785002929,
6
  "eval_steps": 1000,
7
+ "global_step": 35000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2694
  "eval_steps_per_second": 0.425,
2695
  "eval_wer": 0.18715001473622164,
2696
  "step": 34000
2697
+ },
2698
+ {
2699
+ "epoch": 19.976567076742825,
2700
+ "grad_norm": 7.8604302406311035,
2701
+ "learning_rate": 9.367409726345334e-06,
2702
+ "loss": 0.7127,
2703
+ "step": 34100
2704
+ },
2705
+ {
2706
+ "epoch": 20.035149384885763,
2707
+ "grad_norm": 6.885805130004883,
2708
+ "learning_rate": 9.097533329734982e-06,
2709
+ "loss": 0.6805,
2710
+ "step": 34200
2711
+ },
2712
+ {
2713
+ "epoch": 20.093731693028705,
2714
+ "grad_norm": 5.695769786834717,
2715
+ "learning_rate": 8.82765693312463e-06,
2716
+ "loss": 0.7139,
2717
+ "step": 34300
2718
+ },
2719
+ {
2720
+ "epoch": 20.152314001171646,
2721
+ "grad_norm": 6.634642124176025,
2722
+ "learning_rate": 8.557780536514277e-06,
2723
+ "loss": 0.6917,
2724
+ "step": 34400
2725
+ },
2726
+ {
2727
+ "epoch": 20.210896309314588,
2728
+ "grad_norm": 5.601979732513428,
2729
+ "learning_rate": 8.287904139903925e-06,
2730
+ "loss": 0.6576,
2731
+ "step": 34500
2732
+ },
2733
+ {
2734
+ "epoch": 20.269478617457526,
2735
+ "grad_norm": 4.387629508972168,
2736
+ "learning_rate": 8.018027743293572e-06,
2737
+ "loss": 0.7389,
2738
+ "step": 34600
2739
+ },
2740
+ {
2741
+ "epoch": 20.328060925600468,
2742
+ "grad_norm": 5.184898853302002,
2743
+ "learning_rate": 7.748151346683219e-06,
2744
+ "loss": 0.6803,
2745
+ "step": 34700
2746
+ },
2747
+ {
2748
+ "epoch": 20.38664323374341,
2749
+ "grad_norm": 9.722779273986816,
2750
+ "learning_rate": 7.478274950072867e-06,
2751
+ "loss": 0.7095,
2752
+ "step": 34800
2753
+ },
2754
+ {
2755
+ "epoch": 20.44522554188635,
2756
+ "grad_norm": 6.964946269989014,
2757
+ "learning_rate": 7.208398553462514e-06,
2758
+ "loss": 0.7372,
2759
+ "step": 34900
2760
+ },
2761
+ {
2762
+ "epoch": 20.50380785002929,
2763
+ "grad_norm": 9.976523399353027,
2764
+ "learning_rate": 6.938522156852162e-06,
2765
+ "loss": 0.6865,
2766
+ "step": 35000
2767
+ },
2768
+ {
2769
+ "epoch": 20.50380785002929,
2770
+ "eval_loss": 0.08354520797729492,
2771
+ "eval_runtime": 147.4678,
2772
+ "eval_samples_per_second": 3.391,
2773
+ "eval_steps_per_second": 0.427,
2774
+ "eval_wer": 0.1843501326259947,
2775
+ "step": 35000
2776
  }
2777
  ],
2778
  "logging_steps": 100,
 
2792
  "attributes": {}
2793
  }
2794
  },
2795
+ "total_flos": 4.55254984359936e+19,
2796
  "train_batch_size": 8,
2797
  "trial_name": null,
2798
  "trial_params": null