Check commited on
Commit
58df76a
Β·
1 Parent(s): 4dbb8f1

"auto-commit"

Browse files
Files changed (28) hide show
  1. model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67447}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67447}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67447}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67447}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67447}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67447}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67447}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67447}/trainer_state.json +641 -5
  9. model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67447}/training_args.bin +0 -0
  10. model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67572}/config.json +0 -0
  11. model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67572}/optimizer.pt +2 -2
  12. model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67572}/preprocessor_config.json +0 -0
  13. model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67572}/pytorch_model.bin +1 -1
  14. model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67572}/rng_state.pth +1 -1
  15. model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67572}/scaler.pt +1 -1
  16. model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67572}/scheduler.pt +1 -1
  17. model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67572}/trainer_state.json +0 -0
  18. model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67572}/training_args.bin +0 -0
  19. model-bin/finetune/base/log/1629834135.9336932/events.out.tfevents.1629834135.c435e1c5ee04.920.211 +3 -0
  20. model-bin/finetune/base/log/1629834756.205052/events.out.tfevents.1629834756.c435e1c5ee04.920.213 +3 -0
  21. model-bin/finetune/base/log/1629835507.9034107/events.out.tfevents.1629835507.c435e1c5ee04.920.215 +3 -0
  22. model-bin/finetune/base/log/1629836153.2164783/events.out.tfevents.1629836153.c435e1c5ee04.920.217 +3 -0
  23. model-bin/finetune/base/log/1629836799.1113734/events.out.tfevents.1629836799.c435e1c5ee04.920.219 +3 -0
  24. model-bin/finetune/base/log/events.out.tfevents.1629834135.c435e1c5ee04.920.210 +3 -0
  25. model-bin/finetune/base/log/events.out.tfevents.1629834756.c435e1c5ee04.920.212 +3 -0
  26. model-bin/finetune/base/log/events.out.tfevents.1629835507.c435e1c5ee04.920.214 +3 -0
  27. model-bin/finetune/base/log/events.out.tfevents.1629836153.c435e1c5ee04.920.216 +3 -0
  28. model-bin/finetune/base/log/events.out.tfevents.1629836799.c435e1c5ee04.920.218 +3 -0
model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67447}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67447}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a4305b79063344e8dcfa70fc08a440bf763d7a5555ef19d1c29a843c47a2bdd0
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f29be78508a07fb73240b11ffa7167d4632163ad60416d94009458f2a159529
3
  size 722165393
model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67447}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67447}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3b864d526ba5e96bbeec05edbe0fb74e2fce15748318e767dedff344b89f04bf
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb6e5067d265b6d0ac7d38a072eaf79353dd3359b87c14179ec86e37964a2578
3
  size 377909911
model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67447}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4d3efe634557e655c20c363dfd2ec88502e52a1a2fe24613e53efd8f78e3b4d7
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:506537a02628941466ae681cb8b9d6d5f1a22f228b010fe75395d0621cb8473b
3
  size 14503
model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67447}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:890e7963e056aec6747d41244e17966edf587e5f50edd32309235c4d785e577b
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87e2d1745c681c91159da5acdbfd8bc474c3ecf40e467f65b9d5603d68c91173
3
  size 559
model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67447}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd58c4ff7961fcf183c96f47a9e93eb6ea21e444f358cbb7cda92b3831704324
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:754bdca3fd4dc1dd416ee07d27446b6b442be519e3b7e30a419105bdc19f4b05
3
  size 623
model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67447}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
- "best_metric": 0.18588425381903642,
3
- "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-60729",
4
- "epoch": 538.995983935743,
5
- "global_step": 66949,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -184542,11 +184542,647 @@
184542
  "eval_steps_per_second": 0.679,
184543
  "eval_wer": 0.20303119051105578,
184544
  "step": 66949
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
184545
  }
184546
  ],
184547
  "max_steps": 620000,
184548
  "num_train_epochs": 5000,
184549
- "total_flos": 1.8839674795561802e+20,
184550
  "trial_name": null,
184551
  "trial_params": null
184552
  }
 
1
  {
2
+ "best_metric": 0.1855363713557883,
3
+ "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-67447",
4
+ "epoch": 542.995983935743,
5
+ "global_step": 67447,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
184542
  "eval_steps_per_second": 0.679,
184543
  "eval_wer": 0.20303119051105578,
184544
  "step": 66949
184545
+ },
184546
+ {
184547
+ "epoch": 539.01,
184548
+ "learning_rate": 8.943509615384616e-06,
184549
+ "loss": 0.5425,
184550
+ "step": 66950
184551
+ },
184552
+ {
184553
+ "epoch": 539.05,
184554
+ "learning_rate": 8.943429487179488e-06,
184555
+ "loss": 0.3015,
184556
+ "step": 66955
184557
+ },
184558
+ {
184559
+ "epoch": 539.09,
184560
+ "learning_rate": 8.943349358974359e-06,
184561
+ "loss": 0.2871,
184562
+ "step": 66960
184563
+ },
184564
+ {
184565
+ "epoch": 539.13,
184566
+ "learning_rate": 8.943269230769232e-06,
184567
+ "loss": 0.3098,
184568
+ "step": 66965
184569
+ },
184570
+ {
184571
+ "epoch": 539.17,
184572
+ "learning_rate": 8.943189102564103e-06,
184573
+ "loss": 0.5809,
184574
+ "step": 66970
184575
+ },
184576
+ {
184577
+ "epoch": 539.21,
184578
+ "learning_rate": 8.943108974358975e-06,
184579
+ "loss": 1.1017,
184580
+ "step": 66975
184581
+ },
184582
+ {
184583
+ "epoch": 539.25,
184584
+ "learning_rate": 8.943028846153848e-06,
184585
+ "loss": 0.3202,
184586
+ "step": 66980
184587
+ },
184588
+ {
184589
+ "epoch": 539.29,
184590
+ "learning_rate": 8.94294871794872e-06,
184591
+ "loss": 0.3306,
184592
+ "step": 66985
184593
+ },
184594
+ {
184595
+ "epoch": 539.33,
184596
+ "learning_rate": 8.94286858974359e-06,
184597
+ "loss": 0.3722,
184598
+ "step": 66990
184599
+ },
184600
+ {
184601
+ "epoch": 539.37,
184602
+ "learning_rate": 8.942788461538462e-06,
184603
+ "loss": 0.5289,
184604
+ "step": 66995
184605
+ },
184606
+ {
184607
+ "epoch": 539.41,
184608
+ "learning_rate": 8.942708333333335e-06,
184609
+ "loss": 1.2412,
184610
+ "step": 67000
184611
+ },
184612
+ {
184613
+ "epoch": 539.45,
184614
+ "learning_rate": 8.942628205128205e-06,
184615
+ "loss": 0.3389,
184616
+ "step": 67005
184617
+ },
184618
+ {
184619
+ "epoch": 539.49,
184620
+ "learning_rate": 8.942548076923078e-06,
184621
+ "loss": 0.3951,
184622
+ "step": 67010
184623
+ },
184624
+ {
184625
+ "epoch": 539.53,
184626
+ "learning_rate": 8.94246794871795e-06,
184627
+ "loss": 0.3677,
184628
+ "step": 67015
184629
+ },
184630
+ {
184631
+ "epoch": 539.57,
184632
+ "learning_rate": 8.94238782051282e-06,
184633
+ "loss": 0.4942,
184634
+ "step": 67020
184635
+ },
184636
+ {
184637
+ "epoch": 539.61,
184638
+ "learning_rate": 8.942307692307693e-06,
184639
+ "loss": 1.1607,
184640
+ "step": 67025
184641
+ },
184642
+ {
184643
+ "epoch": 539.65,
184644
+ "learning_rate": 8.942227564102565e-06,
184645
+ "loss": 0.3204,
184646
+ "step": 67030
184647
+ },
184648
+ {
184649
+ "epoch": 539.69,
184650
+ "learning_rate": 8.942147435897436e-06,
184651
+ "loss": 0.2899,
184652
+ "step": 67035
184653
+ },
184654
+ {
184655
+ "epoch": 539.73,
184656
+ "learning_rate": 8.942067307692308e-06,
184657
+ "loss": 0.5907,
184658
+ "step": 67040
184659
+ },
184660
+ {
184661
+ "epoch": 539.77,
184662
+ "learning_rate": 8.94198717948718e-06,
184663
+ "loss": 0.5997,
184664
+ "step": 67045
184665
+ },
184666
+ {
184667
+ "epoch": 539.81,
184668
+ "learning_rate": 8.941907051282052e-06,
184669
+ "loss": 1.2436,
184670
+ "step": 67050
184671
+ },
184672
+ {
184673
+ "epoch": 539.85,
184674
+ "learning_rate": 8.941826923076923e-06,
184675
+ "loss": 0.372,
184676
+ "step": 67055
184677
+ },
184678
+ {
184679
+ "epoch": 539.9,
184680
+ "learning_rate": 8.941746794871795e-06,
184681
+ "loss": 0.3978,
184682
+ "step": 67060
184683
+ },
184684
+ {
184685
+ "epoch": 539.94,
184686
+ "learning_rate": 8.941666666666668e-06,
184687
+ "loss": 0.3908,
184688
+ "step": 67065
184689
+ },
184690
+ {
184691
+ "epoch": 539.98,
184692
+ "learning_rate": 8.941586538461539e-06,
184693
+ "loss": 0.7207,
184694
+ "step": 67070
184695
+ },
184696
+ {
184697
+ "epoch": 540.0,
184698
+ "eval_loss": 0.42999500036239624,
184699
+ "eval_runtime": 39.6376,
184700
+ "eval_samples_per_second": 21.116,
184701
+ "eval_steps_per_second": 0.681,
184702
+ "eval_wer": 0.19431621465666474,
184703
+ "step": 67073
184704
+ },
184705
+ {
184706
+ "epoch": 536.02,
184707
+ "learning_rate": 8.94150641025641e-06,
184708
+ "loss": 0.4822,
184709
+ "step": 67075
184710
+ },
184711
+ {
184712
+ "epoch": 536.06,
184713
+ "learning_rate": 8.941426282051284e-06,
184714
+ "loss": 0.2812,
184715
+ "step": 67080
184716
+ },
184717
+ {
184718
+ "epoch": 536.1,
184719
+ "learning_rate": 8.941346153846155e-06,
184720
+ "loss": 0.3065,
184721
+ "step": 67085
184722
+ },
184723
+ {
184724
+ "epoch": 536.14,
184725
+ "learning_rate": 8.941266025641026e-06,
184726
+ "loss": 0.3642,
184727
+ "step": 67090
184728
+ },
184729
+ {
184730
+ "epoch": 536.18,
184731
+ "learning_rate": 8.941185897435898e-06,
184732
+ "loss": 0.6448,
184733
+ "step": 67095
184734
+ },
184735
+ {
184736
+ "epoch": 536.22,
184737
+ "learning_rate": 8.94110576923077e-06,
184738
+ "loss": 1.0078,
184739
+ "step": 67100
184740
+ },
184741
+ {
184742
+ "epoch": 536.26,
184743
+ "learning_rate": 8.941025641025642e-06,
184744
+ "loss": 0.3443,
184745
+ "step": 67105
184746
+ },
184747
+ {
184748
+ "epoch": 536.3,
184749
+ "learning_rate": 8.940945512820513e-06,
184750
+ "loss": 0.387,
184751
+ "step": 67110
184752
+ },
184753
+ {
184754
+ "epoch": 536.34,
184755
+ "learning_rate": 8.940865384615386e-06,
184756
+ "loss": 0.3822,
184757
+ "step": 67115
184758
+ },
184759
+ {
184760
+ "epoch": 536.38,
184761
+ "learning_rate": 8.940785256410258e-06,
184762
+ "loss": 0.6696,
184763
+ "step": 67120
184764
+ },
184765
+ {
184766
+ "epoch": 536.42,
184767
+ "learning_rate": 8.940705128205129e-06,
184768
+ "loss": 1.0635,
184769
+ "step": 67125
184770
+ },
184771
+ {
184772
+ "epoch": 536.46,
184773
+ "learning_rate": 8.940625e-06,
184774
+ "loss": 0.2825,
184775
+ "step": 67130
184776
+ },
184777
+ {
184778
+ "epoch": 536.5,
184779
+ "learning_rate": 8.940544871794874e-06,
184780
+ "loss": 0.3859,
184781
+ "step": 67135
184782
+ },
184783
+ {
184784
+ "epoch": 536.54,
184785
+ "learning_rate": 8.940464743589743e-06,
184786
+ "loss": 0.4542,
184787
+ "step": 67140
184788
+ },
184789
+ {
184790
+ "epoch": 536.58,
184791
+ "learning_rate": 8.940384615384616e-06,
184792
+ "loss": 0.7055,
184793
+ "step": 67145
184794
+ },
184795
+ {
184796
+ "epoch": 536.62,
184797
+ "learning_rate": 8.940304487179488e-06,
184798
+ "loss": 0.9673,
184799
+ "step": 67150
184800
+ },
184801
+ {
184802
+ "epoch": 536.66,
184803
+ "learning_rate": 8.940224358974359e-06,
184804
+ "loss": 0.3229,
184805
+ "step": 67155
184806
+ },
184807
+ {
184808
+ "epoch": 536.7,
184809
+ "learning_rate": 8.94014423076923e-06,
184810
+ "loss": 0.3374,
184811
+ "step": 67160
184812
+ },
184813
+ {
184814
+ "epoch": 536.74,
184815
+ "learning_rate": 8.940064102564103e-06,
184816
+ "loss": 0.4204,
184817
+ "step": 67165
184818
+ },
184819
+ {
184820
+ "epoch": 536.78,
184821
+ "learning_rate": 8.939983974358975e-06,
184822
+ "loss": 0.7498,
184823
+ "step": 67170
184824
+ },
184825
+ {
184826
+ "epoch": 536.82,
184827
+ "learning_rate": 8.939903846153846e-06,
184828
+ "loss": 1.0866,
184829
+ "step": 67175
184830
+ },
184831
+ {
184832
+ "epoch": 536.86,
184833
+ "learning_rate": 8.939823717948719e-06,
184834
+ "loss": 0.3682,
184835
+ "step": 67180
184836
+ },
184837
+ {
184838
+ "epoch": 536.9,
184839
+ "learning_rate": 8.93974358974359e-06,
184840
+ "loss": 0.3961,
184841
+ "step": 67185
184842
+ },
184843
+ {
184844
+ "epoch": 536.94,
184845
+ "learning_rate": 8.939663461538462e-06,
184846
+ "loss": 0.3728,
184847
+ "step": 67190
184848
+ },
184849
+ {
184850
+ "epoch": 536.98,
184851
+ "learning_rate": 8.939583333333333e-06,
184852
+ "loss": 0.607,
184853
+ "step": 67195
184854
+ },
184855
+ {
184856
+ "epoch": 537.0,
184857
+ "eval_loss": 0.4263405501842499,
184858
+ "eval_runtime": 39.1689,
184859
+ "eval_samples_per_second": 21.369,
184860
+ "eval_steps_per_second": 0.689,
184861
+ "eval_wer": 0.18978423601937472,
184862
+ "step": 67198
184863
+ },
184864
+ {
184865
+ "epoch": 537.02,
184866
+ "learning_rate": 8.939503205128206e-06,
184867
+ "loss": 0.3609,
184868
+ "step": 67200
184869
+ },
184870
+ {
184871
+ "epoch": 537.06,
184872
+ "learning_rate": 8.939423076923078e-06,
184873
+ "loss": 0.3246,
184874
+ "step": 67205
184875
+ },
184876
+ {
184877
+ "epoch": 537.1,
184878
+ "learning_rate": 8.939342948717949e-06,
184879
+ "loss": 0.3176,
184880
+ "step": 67210
184881
+ },
184882
+ {
184883
+ "epoch": 537.14,
184884
+ "learning_rate": 8.939262820512822e-06,
184885
+ "loss": 0.4736,
184886
+ "step": 67215
184887
+ },
184888
+ {
184889
+ "epoch": 537.18,
184890
+ "learning_rate": 8.939182692307693e-06,
184891
+ "loss": 0.6372,
184892
+ "step": 67220
184893
+ },
184894
+ {
184895
+ "epoch": 537.22,
184896
+ "learning_rate": 8.939102564102565e-06,
184897
+ "loss": 1.0112,
184898
+ "step": 67225
184899
+ },
184900
+ {
184901
+ "epoch": 537.26,
184902
+ "learning_rate": 8.939022435897436e-06,
184903
+ "loss": 0.4262,
184904
+ "step": 67230
184905
+ },
184906
+ {
184907
+ "epoch": 537.3,
184908
+ "learning_rate": 8.93894230769231e-06,
184909
+ "loss": 0.3433,
184910
+ "step": 67235
184911
+ },
184912
+ {
184913
+ "epoch": 537.34,
184914
+ "learning_rate": 8.93886217948718e-06,
184915
+ "loss": 0.363,
184916
+ "step": 67240
184917
+ },
184918
+ {
184919
+ "epoch": 537.38,
184920
+ "learning_rate": 8.938782051282052e-06,
184921
+ "loss": 0.6827,
184922
+ "step": 67245
184923
+ },
184924
+ {
184925
+ "epoch": 537.42,
184926
+ "learning_rate": 8.938701923076923e-06,
184927
+ "loss": 0.9992,
184928
+ "step": 67250
184929
+ },
184930
+ {
184931
+ "epoch": 537.46,
184932
+ "learning_rate": 8.938621794871796e-06,
184933
+ "loss": 0.3153,
184934
+ "step": 67255
184935
+ },
184936
+ {
184937
+ "epoch": 537.5,
184938
+ "learning_rate": 8.938541666666668e-06,
184939
+ "loss": 0.2932,
184940
+ "step": 67260
184941
+ },
184942
+ {
184943
+ "epoch": 537.54,
184944
+ "learning_rate": 8.938461538461539e-06,
184945
+ "loss": 0.3589,
184946
+ "step": 67265
184947
+ },
184948
+ {
184949
+ "epoch": 537.58,
184950
+ "learning_rate": 8.938381410256412e-06,
184951
+ "loss": 0.7378,
184952
+ "step": 67270
184953
+ },
184954
+ {
184955
+ "epoch": 537.62,
184956
+ "learning_rate": 8.938301282051283e-06,
184957
+ "loss": 1.2153,
184958
+ "step": 67275
184959
+ },
184960
+ {
184961
+ "epoch": 537.66,
184962
+ "learning_rate": 8.938221153846155e-06,
184963
+ "loss": 0.3399,
184964
+ "step": 67280
184965
+ },
184966
+ {
184967
+ "epoch": 537.7,
184968
+ "learning_rate": 8.938141025641026e-06,
184969
+ "loss": 0.3661,
184970
+ "step": 67285
184971
+ },
184972
+ {
184973
+ "epoch": 537.74,
184974
+ "learning_rate": 8.9380608974359e-06,
184975
+ "loss": 0.3713,
184976
+ "step": 67290
184977
+ },
184978
+ {
184979
+ "epoch": 537.78,
184980
+ "learning_rate": 8.937980769230769e-06,
184981
+ "loss": 0.6155,
184982
+ "step": 67295
184983
+ },
184984
+ {
184985
+ "epoch": 537.82,
184986
+ "learning_rate": 8.937900641025642e-06,
184987
+ "loss": 1.0946,
184988
+ "step": 67300
184989
+ },
184990
+ {
184991
+ "epoch": 537.86,
184992
+ "learning_rate": 8.937820512820513e-06,
184993
+ "loss": 0.3214,
184994
+ "step": 67305
184995
+ },
184996
+ {
184997
+ "epoch": 537.9,
184998
+ "learning_rate": 8.937740384615385e-06,
184999
+ "loss": 0.3359,
185000
+ "step": 67310
185001
+ },
185002
+ {
185003
+ "epoch": 537.94,
185004
+ "learning_rate": 8.937660256410258e-06,
185005
+ "loss": 0.3965,
185006
+ "step": 67315
185007
+ },
185008
+ {
185009
+ "epoch": 537.98,
185010
+ "learning_rate": 8.937580128205129e-06,
185011
+ "loss": 0.7338,
185012
+ "step": 67320
185013
+ },
185014
+ {
185015
+ "epoch": 538.0,
185016
+ "eval_loss": 0.46577340364456177,
185017
+ "eval_runtime": 38.9427,
185018
+ "eval_samples_per_second": 21.544,
185019
+ "eval_steps_per_second": 0.693,
185020
+ "eval_wer": 0.19503469986406238,
185021
+ "step": 67323
185022
+ },
185023
+ {
185024
+ "epoch": 542.02,
185025
+ "learning_rate": 8.9375e-06,
185026
+ "loss": 0.4422,
185027
+ "step": 67325
185028
+ },
185029
+ {
185030
+ "epoch": 542.06,
185031
+ "learning_rate": 8.937419871794872e-06,
185032
+ "loss": 0.3339,
185033
+ "step": 67330
185034
+ },
185035
+ {
185036
+ "epoch": 542.1,
185037
+ "learning_rate": 8.937339743589745e-06,
185038
+ "loss": 0.3232,
185039
+ "step": 67335
185040
+ },
185041
+ {
185042
+ "epoch": 542.14,
185043
+ "learning_rate": 8.937259615384616e-06,
185044
+ "loss": 0.3989,
185045
+ "step": 67340
185046
+ },
185047
+ {
185048
+ "epoch": 542.18,
185049
+ "learning_rate": 8.937179487179488e-06,
185050
+ "loss": 0.7477,
185051
+ "step": 67345
185052
+ },
185053
+ {
185054
+ "epoch": 542.22,
185055
+ "learning_rate": 8.937099358974359e-06,
185056
+ "loss": 1.1284,
185057
+ "step": 67350
185058
+ },
185059
+ {
185060
+ "epoch": 542.26,
185061
+ "learning_rate": 8.937019230769232e-06,
185062
+ "loss": 0.2786,
185063
+ "step": 67355
185064
+ },
185065
+ {
185066
+ "epoch": 542.3,
185067
+ "learning_rate": 8.936939102564103e-06,
185068
+ "loss": 0.3625,
185069
+ "step": 67360
185070
+ },
185071
+ {
185072
+ "epoch": 542.34,
185073
+ "learning_rate": 8.936858974358975e-06,
185074
+ "loss": 0.4097,
185075
+ "step": 67365
185076
+ },
185077
+ {
185078
+ "epoch": 542.38,
185079
+ "learning_rate": 8.936778846153848e-06,
185080
+ "loss": 0.8628,
185081
+ "step": 67370
185082
+ },
185083
+ {
185084
+ "epoch": 542.42,
185085
+ "learning_rate": 8.936698717948719e-06,
185086
+ "loss": 1.0227,
185087
+ "step": 67375
185088
+ },
185089
+ {
185090
+ "epoch": 542.46,
185091
+ "learning_rate": 8.93661858974359e-06,
185092
+ "loss": 0.2886,
185093
+ "step": 67380
185094
+ },
185095
+ {
185096
+ "epoch": 542.5,
185097
+ "learning_rate": 8.936538461538462e-06,
185098
+ "loss": 0.283,
185099
+ "step": 67385
185100
+ },
185101
+ {
185102
+ "epoch": 542.54,
185103
+ "learning_rate": 8.936458333333335e-06,
185104
+ "loss": 0.3514,
185105
+ "step": 67390
185106
+ },
185107
+ {
185108
+ "epoch": 542.58,
185109
+ "learning_rate": 8.936378205128206e-06,
185110
+ "loss": 0.674,
185111
+ "step": 67395
185112
+ },
185113
+ {
185114
+ "epoch": 542.62,
185115
+ "learning_rate": 8.936298076923078e-06,
185116
+ "loss": 1.0492,
185117
+ "step": 67400
185118
+ },
185119
+ {
185120
+ "epoch": 542.66,
185121
+ "learning_rate": 8.936217948717949e-06,
185122
+ "loss": 0.4139,
185123
+ "step": 67405
185124
+ },
185125
+ {
185126
+ "epoch": 542.7,
185127
+ "learning_rate": 8.936137820512822e-06,
185128
+ "loss": 0.3585,
185129
+ "step": 67410
185130
+ },
185131
+ {
185132
+ "epoch": 542.74,
185133
+ "learning_rate": 8.936057692307693e-06,
185134
+ "loss": 0.3681,
185135
+ "step": 67415
185136
+ },
185137
+ {
185138
+ "epoch": 542.78,
185139
+ "learning_rate": 8.935977564102565e-06,
185140
+ "loss": 0.6159,
185141
+ "step": 67420
185142
+ },
185143
+ {
185144
+ "epoch": 542.82,
185145
+ "learning_rate": 8.935897435897438e-06,
185146
+ "loss": 1.0444,
185147
+ "step": 67425
185148
+ },
185149
+ {
185150
+ "epoch": 542.86,
185151
+ "learning_rate": 8.935817307692309e-06,
185152
+ "loss": 0.3311,
185153
+ "step": 67430
185154
+ },
185155
+ {
185156
+ "epoch": 542.9,
185157
+ "learning_rate": 8.93573717948718e-06,
185158
+ "loss": 0.377,
185159
+ "step": 67435
185160
+ },
185161
+ {
185162
+ "epoch": 542.94,
185163
+ "learning_rate": 8.935657051282052e-06,
185164
+ "loss": 0.3752,
185165
+ "step": 67440
185166
+ },
185167
+ {
185168
+ "epoch": 542.98,
185169
+ "learning_rate": 8.935576923076925e-06,
185170
+ "loss": 0.5475,
185171
+ "step": 67445
185172
+ },
185173
+ {
185174
+ "epoch": 543.0,
185175
+ "eval_loss": 0.35275718569755554,
185176
+ "eval_runtime": 40.6597,
185177
+ "eval_samples_per_second": 20.635,
185178
+ "eval_steps_per_second": 0.664,
185179
+ "eval_wer": 0.1855363713557883,
185180
+ "step": 67447
185181
  }
185182
  ],
185183
  "max_steps": 620000,
185184
  "num_train_epochs": 5000,
185185
+ "total_flos": 1.897958418039559e+20,
185186
  "trial_name": null,
185187
  "trial_params": null
185188
  }
model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67447}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67572}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67572}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0406650db1393fb9692f8d43b2e0d3fd7ac1ccba07e5597f49b7e009002820b6
3
- size 722165009
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cad4194b8552407741480b0d7cfec2bbde02181f458dbb03ec6afcfecdd08026
3
+ size 722165393
model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67572}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67572}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2839b98fe4b47ad20f5b8b1bae2552bd34df8084d7cb7ec7a9efadbef996fd76
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43eaa79891cd0d56e96b0fc5b84927dae77481a7bceab190190510d141ccdbf1
3
  size 377909911
model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67572}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e4b858c2d8d13cecbbedfc9bcb6fd9b338c4dab35127f6a04749c2971bdf13f5
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:412a8b45bf8449c4c28a263b3f9f3c718a03c421d731fa385f8caba49e11373c
3
  size 14503
model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67572}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8726c92bb7ef60e388546e9fd18baec8ae7c170a0be794a916a6e8ed8249c65c
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b996ed8da366dae97a015b33368c0ecf067f550c8dd285673547842926659440
3
  size 559
model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67572}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:84754f1abc95954e6ecb719c67cba56880d4f0682e71be6ded68f83ed242eea3
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f3db58c27f09014af98cc064e08602c5304eb789e6d0bbd26dfcf9aa36490f9
3
  size 623
model-bin/finetune/base/{checkpoint-60729 β†’ checkpoint-67572}/trainer_state.json RENAMED
The diff for this file is too large to render. See raw diff
 
model-bin/finetune/base/{checkpoint-66949 β†’ checkpoint-67572}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629834135.9336932/events.out.tfevents.1629834135.c435e1c5ee04.920.211 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b150f7b0958b1971f10d638538ea0ff8131c4ffb78d9e0d29a30e05ea5c52e8
3
+ size 4194
model-bin/finetune/base/log/1629834756.205052/events.out.tfevents.1629834756.c435e1c5ee04.920.213 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b348d2848eadf4eec5cf10700dcda79d2f33f7a21030d5e409d510173556f0cb
3
+ size 4194
model-bin/finetune/base/log/1629835507.9034107/events.out.tfevents.1629835507.c435e1c5ee04.920.215 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90a811cd4ee486b70c250cef428dcc8561cfcdfb386529528d2a36845aa3f4d0
3
+ size 4194
model-bin/finetune/base/log/1629836153.2164783/events.out.tfevents.1629836153.c435e1c5ee04.920.217 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e81be56ed4378a407beed856272baa7fc03446b3b2d110ff221d5a9466f9e29
3
+ size 4194
model-bin/finetune/base/log/1629836799.1113734/events.out.tfevents.1629836799.c435e1c5ee04.920.219 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9240c94e3aeed04b7627a11dcbce62cfe96fe7a1ef7303fe2a4719c4010644fb
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629834135.c435e1c5ee04.920.210 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:630db7e7ef2a4b28bb5ed551f2405e52100e6a95680cfedb7f7e52c850f61cab
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629834756.c435e1c5ee04.920.212 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1482b5d3d281d592a7479c3d20581ff59351aabc483c6c8017660f1e9b7be927
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629835507.c435e1c5ee04.920.214 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:189290b987e9b5fd88e38a209ed29192b4462584bcb030abfe8804f45d1f6220
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629836153.c435e1c5ee04.920.216 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aabad5cc92ba027d433b2116f6f914a5b311adf64157de040def01c7c40781f8
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629836799.c435e1c5ee04.920.218 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e66806bcc547d4ad407e6b0a5a655f795f3facbc7f5cebcc52173308532ef49
3
+ size 8622