Baselhany commited on
Commit
83dea89
·
verified ·
1 Parent(s): 24828ca

Training in progress, step 34000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0cfc2369854afc94fdfa5a772fefa0c2c28012dcdc0bc0ea80f23b73f346c34a
3
  size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:396a29a5a4787dbf082bfc1745fc5d4dbaf49c674cf87663f742af78ff889ac4
3
  size 223144592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:59a0c5737e43b8c351af0642dc5e297ba4cfc310d2333fb3515ae84242ec7206
3
  size 281574266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a681bae809c2c7355b83c2ba73ac835d0bd5c7c190fd2e0484e22164316f803
3
  size 281574266
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:576380a256a4340c462ffc82306c52abbab7f38e65a7b12e9ab2634163e0fae7
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce2c6e00c180a2aec543a6c39f38b41238eae7e450c89c57b59b23abde749eed
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5680c8a2f1d27c25e08e493d045a3f651f13f689a05908b8e76432cb1f7ace3b
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d82f68a924bc6f264caa5b5167990f1fdb3b2c8f9254156d8c0f9ccc3c5489db
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3bb6ef00c1c865d5fe6826f62f53b17ea0c3488cfae31fdb4898bd3e20b4343e
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:122a12c57bcb396697297c9c6111a0a8854f31fea532275de745584ad3eee2c3
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 28000,
3
  "best_metric": 0.18110816386678455,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-28000",
5
- "epoch": 19.332161687170476,
6
  "eval_steps": 1000,
7
- "global_step": 33000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2615,6 +2615,85 @@
2615
  "eval_steps_per_second": 0.426,
2616
  "eval_wer": 0.18641320365458297,
2617
  "step": 33000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2618
  }
2619
  ],
2620
  "logging_steps": 100,
@@ -2634,7 +2713,7 @@
2634
  "attributes": {}
2635
  }
2636
  },
2637
- "total_flos": 4.292401815355392e+19,
2638
  "train_batch_size": 8,
2639
  "trial_name": null,
2640
  "trial_params": null
 
2
  "best_global_step": 28000,
3
  "best_metric": 0.18110816386678455,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-28000",
5
+ "epoch": 19.917984768599883,
6
  "eval_steps": 1000,
7
+ "global_step": 34000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2615
  "eval_steps_per_second": 0.426,
2616
  "eval_wer": 0.18641320365458297,
2617
  "step": 33000
2618
+ },
2619
+ {
2620
+ "epoch": 19.390743995313414,
2621
+ "grad_norm": 7.711204528808594,
2622
+ "learning_rate": 1.206617369244886e-05,
2623
+ "loss": 0.7326,
2624
+ "step": 33100
2625
+ },
2626
+ {
2627
+ "epoch": 19.449326303456356,
2628
+ "grad_norm": 6.137068271636963,
2629
+ "learning_rate": 1.1796297295838508e-05,
2630
+ "loss": 0.6976,
2631
+ "step": 33200
2632
+ },
2633
+ {
2634
+ "epoch": 19.507908611599298,
2635
+ "grad_norm": 6.798486709594727,
2636
+ "learning_rate": 1.1526420899228154e-05,
2637
+ "loss": 0.7118,
2638
+ "step": 33300
2639
+ },
2640
+ {
2641
+ "epoch": 19.566490919742236,
2642
+ "grad_norm": 6.276034355163574,
2643
+ "learning_rate": 1.12565445026178e-05,
2644
+ "loss": 0.7112,
2645
+ "step": 33400
2646
+ },
2647
+ {
2648
+ "epoch": 19.625073227885178,
2649
+ "grad_norm": 7.021731853485107,
2650
+ "learning_rate": 1.0986668106007449e-05,
2651
+ "loss": 0.6984,
2652
+ "step": 33500
2653
+ },
2654
+ {
2655
+ "epoch": 19.68365553602812,
2656
+ "grad_norm": 5.709362983703613,
2657
+ "learning_rate": 1.0716791709397096e-05,
2658
+ "loss": 0.7376,
2659
+ "step": 33600
2660
+ },
2661
+ {
2662
+ "epoch": 19.74223784417106,
2663
+ "grad_norm": 7.312982559204102,
2664
+ "learning_rate": 1.0446915312786744e-05,
2665
+ "loss": 0.7448,
2666
+ "step": 33700
2667
+ },
2668
+ {
2669
+ "epoch": 19.800820152314,
2670
+ "grad_norm": 5.710330486297607,
2671
+ "learning_rate": 1.0177038916176392e-05,
2672
+ "loss": 0.7524,
2673
+ "step": 33800
2674
+ },
2675
+ {
2676
+ "epoch": 19.85940246045694,
2677
+ "grad_norm": 6.241232395172119,
2678
+ "learning_rate": 9.907162519566039e-06,
2679
+ "loss": 0.6942,
2680
+ "step": 33900
2681
+ },
2682
+ {
2683
+ "epoch": 19.917984768599883,
2684
+ "grad_norm": 6.5843024253845215,
2685
+ "learning_rate": 9.637286122955687e-06,
2686
+ "loss": 0.707,
2687
+ "step": 34000
2688
+ },
2689
+ {
2690
+ "epoch": 19.917984768599883,
2691
+ "eval_loss": 0.08329460024833679,
2692
+ "eval_runtime": 148.0997,
2693
+ "eval_samples_per_second": 3.376,
2694
+ "eval_steps_per_second": 0.425,
2695
+ "eval_wer": 0.18715001473622164,
2696
+ "step": 34000
2697
  }
2698
  ],
2699
  "logging_steps": 100,
 
2713
  "attributes": {}
2714
  }
2715
  },
2716
+ "total_flos": 4.422483959611392e+19,
2717
  "train_batch_size": 8,
2718
  "trial_name": null,
2719
  "trial_params": null