Check commited on
Commit
290aa60
Β·
1 Parent(s): d23c680

"auto-commit"

Browse files
Files changed (28) hide show
  1. model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52515}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52515}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52515}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52515}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52515}/rng_state.pth +2 -2
  6. model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52515}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52515}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52515}/trainer_state.json +642 -6
  9. model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52515}/training_args.bin +0 -0
  10. model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52639}/config.json +0 -0
  11. model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52639}/optimizer.pt +1 -1
  12. model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52639}/preprocessor_config.json +0 -0
  13. model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52639}/pytorch_model.bin +1 -1
  14. model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52639}/rng_state.pth +1 -1
  15. model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52639}/scaler.pt +1 -1
  16. model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52639}/scheduler.pt +1 -1
  17. model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52639}/trainer_state.json +0 -0
  18. model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52639}/training_args.bin +0 -0
  19. model-bin/finetune/base/log/1629752462.5473433/events.out.tfevents.1629752462.74272264b15c.932.223 +3 -0
  20. model-bin/finetune/base/log/1629753109.9928777/events.out.tfevents.1629753109.74272264b15c.932.225 +3 -0
  21. model-bin/finetune/base/log/1629753750.9052882/events.out.tfevents.1629753750.74272264b15c.932.227 +3 -0
  22. model-bin/finetune/base/log/1629754385.4896328/events.out.tfevents.1629754385.74272264b15c.932.229 +3 -0
  23. model-bin/finetune/base/log/1629755043.4480946/events.out.tfevents.1629755043.74272264b15c.932.231 +3 -0
  24. model-bin/finetune/base/log/events.out.tfevents.1629752462.74272264b15c.932.222 +3 -0
  25. model-bin/finetune/base/log/events.out.tfevents.1629753109.74272264b15c.932.224 +3 -0
  26. model-bin/finetune/base/log/events.out.tfevents.1629753750.74272264b15c.932.226 +3 -0
  27. model-bin/finetune/base/log/events.out.tfevents.1629754385.74272264b15c.932.228 +3 -0
  28. model-bin/finetune/base/log/events.out.tfevents.1629755043.74272264b15c.932.230 +3 -0
model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52515}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52515}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c4bb48e532c12ad5b1093fd7ebbde27ae407ccd5ced529b170bf89c64883bd4
3
  size 722165009
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07ad5e8e80074832caa0a98b5c08f323f432edcdb3b2a365d123371f20e8beda
3
  size 722165009
model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52515}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52515}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ccd0de1e9539b726c0716fdb3a195e2d50475b80cff249324c9c6298512f047c
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ad7bfa19a307825777f0d973e4d3894ebe0ab9f5e474e9337514096826b79b2
3
  size 377909911
model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52515}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6e05861741a7b5f0d8dde4975b7b4ab4ba0db65702615da6c6340b05b2596ea5
3
- size 14439
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df6b787e2064a641feeeec680cf7d039f7d81c7e51db0009bb53c63dd07dabf9
3
+ size 14567
model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52515}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:83aaac26ff07408c2a26c36f15750d57617e8958aad261fe178da94ede075a9a
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a85bd7e9d0b1ebc11f5f5c5604858f3dbf9d33e447f24534a89df3b0a46a7a1
3
  size 559
model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52515}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:643acba45503adeafd2d3b2a47595b06e56df01c8c7a9df6f473756252f75276
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:675296120f77ac95e3ef252eb671e40c28127966b4388d2ca4ab32f9c73d4f1a
3
  size 623
model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52515}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
- "best_metric": 0.19748327029386092,
3
- "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-46666",
4
- "epoch": 418.99598393574297,
5
- "global_step": 52015,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -165546,11 +165546,647 @@
165546
  "eval_steps_per_second": 0.666,
165547
  "eval_wer": 0.20132061628760087,
165548
  "step": 52015
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
165549
  }
165550
  ],
165551
- "max_steps": 620000,
165552
  "num_train_epochs": 5000,
165553
- "total_flos": 1.463606633594667e+20,
165554
  "trial_name": null,
165555
  "trial_params": null
165556
  }
 
1
  {
2
+ "best_metric": 0.18992848189928482,
3
+ "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-52515",
4
+ "epoch": 420.0,
5
+ "global_step": 52515,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
165546
  "eval_steps_per_second": 0.666,
165547
  "eval_wer": 0.20132061628760087,
165548
  "step": 52015
165549
+ },
165550
+ {
165551
+ "epoch": 416.04,
165552
+ "learning_rate": 9.182676282051282e-06,
165553
+ "loss": 0.3158,
165554
+ "step": 52020
165555
+ },
165556
+ {
165557
+ "epoch": 416.08,
165558
+ "learning_rate": 9.182596153846155e-06,
165559
+ "loss": 0.3036,
165560
+ "step": 52025
165561
+ },
165562
+ {
165563
+ "epoch": 416.12,
165564
+ "learning_rate": 9.182516025641027e-06,
165565
+ "loss": 0.339,
165566
+ "step": 52030
165567
+ },
165568
+ {
165569
+ "epoch": 416.16,
165570
+ "learning_rate": 9.182435897435898e-06,
165571
+ "loss": 0.4548,
165572
+ "step": 52035
165573
+ },
165574
+ {
165575
+ "epoch": 416.2,
165576
+ "learning_rate": 9.18235576923077e-06,
165577
+ "loss": 1.1986,
165578
+ "step": 52040
165579
+ },
165580
+ {
165581
+ "epoch": 416.24,
165582
+ "learning_rate": 9.182275641025642e-06,
165583
+ "loss": 0.4446,
165584
+ "step": 52045
165585
+ },
165586
+ {
165587
+ "epoch": 416.28,
165588
+ "learning_rate": 9.182195512820514e-06,
165589
+ "loss": 0.3353,
165590
+ "step": 52050
165591
+ },
165592
+ {
165593
+ "epoch": 416.32,
165594
+ "learning_rate": 9.182115384615385e-06,
165595
+ "loss": 0.3124,
165596
+ "step": 52055
165597
+ },
165598
+ {
165599
+ "epoch": 416.36,
165600
+ "learning_rate": 9.182035256410257e-06,
165601
+ "loss": 0.5198,
165602
+ "step": 52060
165603
+ },
165604
+ {
165605
+ "epoch": 416.4,
165606
+ "learning_rate": 9.18195512820513e-06,
165607
+ "loss": 1.2901,
165608
+ "step": 52065
165609
+ },
165610
+ {
165611
+ "epoch": 416.44,
165612
+ "learning_rate": 9.181875000000001e-06,
165613
+ "loss": 0.3501,
165614
+ "step": 52070
165615
+ },
165616
+ {
165617
+ "epoch": 416.48,
165618
+ "learning_rate": 9.181794871794872e-06,
165619
+ "loss": 0.3317,
165620
+ "step": 52075
165621
+ },
165622
+ {
165623
+ "epoch": 416.52,
165624
+ "learning_rate": 9.181714743589745e-06,
165625
+ "loss": 0.3237,
165626
+ "step": 52080
165627
+ },
165628
+ {
165629
+ "epoch": 416.56,
165630
+ "learning_rate": 9.181634615384617e-06,
165631
+ "loss": 0.5543,
165632
+ "step": 52085
165633
+ },
165634
+ {
165635
+ "epoch": 416.6,
165636
+ "learning_rate": 9.181554487179488e-06,
165637
+ "loss": 1.2754,
165638
+ "step": 52090
165639
+ },
165640
+ {
165641
+ "epoch": 416.64,
165642
+ "learning_rate": 9.18147435897436e-06,
165643
+ "loss": 0.3846,
165644
+ "step": 52095
165645
+ },
165646
+ {
165647
+ "epoch": 416.68,
165648
+ "learning_rate": 9.181394230769232e-06,
165649
+ "loss": 0.3265,
165650
+ "step": 52100
165651
+ },
165652
+ {
165653
+ "epoch": 416.72,
165654
+ "learning_rate": 9.181314102564102e-06,
165655
+ "loss": 0.3491,
165656
+ "step": 52105
165657
+ },
165658
+ {
165659
+ "epoch": 416.76,
165660
+ "learning_rate": 9.181233974358975e-06,
165661
+ "loss": 0.4898,
165662
+ "step": 52110
165663
+ },
165664
+ {
165665
+ "epoch": 416.8,
165666
+ "learning_rate": 9.181153846153848e-06,
165667
+ "loss": 1.1767,
165668
+ "step": 52115
165669
+ },
165670
+ {
165671
+ "epoch": 416.84,
165672
+ "learning_rate": 9.181073717948718e-06,
165673
+ "loss": 0.3844,
165674
+ "step": 52120
165675
+ },
165676
+ {
165677
+ "epoch": 416.88,
165678
+ "learning_rate": 9.180993589743591e-06,
165679
+ "loss": 0.3551,
165680
+ "step": 52125
165681
+ },
165682
+ {
165683
+ "epoch": 416.92,
165684
+ "learning_rate": 9.180913461538462e-06,
165685
+ "loss": 0.4132,
165686
+ "step": 52130
165687
+ },
165688
+ {
165689
+ "epoch": 416.96,
165690
+ "learning_rate": 9.180833333333334e-06,
165691
+ "loss": 0.5226,
165692
+ "step": 52135
165693
+ },
165694
+ {
165695
+ "epoch": 417.0,
165696
+ "learning_rate": 9.180753205128205e-06,
165697
+ "loss": 1.3741,
165698
+ "step": 52140
165699
+ },
165700
+ {
165701
+ "epoch": 417.0,
165702
+ "eval_loss": 0.45200973749160767,
165703
+ "eval_runtime": 39.6524,
165704
+ "eval_samples_per_second": 21.184,
165705
+ "eval_steps_per_second": 0.681,
165706
+ "eval_wer": 0.20303938043398845,
165707
+ "step": 52140
165708
+ },
165709
+ {
165710
+ "epoch": 417.04,
165711
+ "learning_rate": 9.180673076923078e-06,
165712
+ "loss": 0.4385,
165713
+ "step": 52145
165714
+ },
165715
+ {
165716
+ "epoch": 417.08,
165717
+ "learning_rate": 9.18059294871795e-06,
165718
+ "loss": 0.3828,
165719
+ "step": 52150
165720
+ },
165721
+ {
165722
+ "epoch": 417.12,
165723
+ "learning_rate": 9.18051282051282e-06,
165724
+ "loss": 0.2904,
165725
+ "step": 52155
165726
+ },
165727
+ {
165728
+ "epoch": 417.16,
165729
+ "learning_rate": 9.180432692307692e-06,
165730
+ "loss": 0.5224,
165731
+ "step": 52160
165732
+ },
165733
+ {
165734
+ "epoch": 417.2,
165735
+ "learning_rate": 9.180352564102565e-06,
165736
+ "loss": 1.3969,
165737
+ "step": 52165
165738
+ },
165739
+ {
165740
+ "epoch": 417.24,
165741
+ "learning_rate": 9.180272435897437e-06,
165742
+ "loss": 0.3446,
165743
+ "step": 52170
165744
+ },
165745
+ {
165746
+ "epoch": 417.28,
165747
+ "learning_rate": 9.180192307692308e-06,
165748
+ "loss": 0.3816,
165749
+ "step": 52175
165750
+ },
165751
+ {
165752
+ "epoch": 417.32,
165753
+ "learning_rate": 9.180112179487181e-06,
165754
+ "loss": 0.4264,
165755
+ "step": 52180
165756
+ },
165757
+ {
165758
+ "epoch": 417.36,
165759
+ "learning_rate": 9.180032051282052e-06,
165760
+ "loss": 0.4443,
165761
+ "step": 52185
165762
+ },
165763
+ {
165764
+ "epoch": 417.4,
165765
+ "learning_rate": 9.179951923076924e-06,
165766
+ "loss": 1.2463,
165767
+ "step": 52190
165768
+ },
165769
+ {
165770
+ "epoch": 417.44,
165771
+ "learning_rate": 9.179871794871795e-06,
165772
+ "loss": 0.4601,
165773
+ "step": 52195
165774
+ },
165775
+ {
165776
+ "epoch": 417.48,
165777
+ "learning_rate": 9.179791666666668e-06,
165778
+ "loss": 0.3195,
165779
+ "step": 52200
165780
+ },
165781
+ {
165782
+ "epoch": 417.52,
165783
+ "learning_rate": 9.17971153846154e-06,
165784
+ "loss": 0.3873,
165785
+ "step": 52205
165786
+ },
165787
+ {
165788
+ "epoch": 417.56,
165789
+ "learning_rate": 9.17963141025641e-06,
165790
+ "loss": 0.4696,
165791
+ "step": 52210
165792
+ },
165793
+ {
165794
+ "epoch": 417.6,
165795
+ "learning_rate": 9.179551282051284e-06,
165796
+ "loss": 1.338,
165797
+ "step": 52215
165798
+ },
165799
+ {
165800
+ "epoch": 417.64,
165801
+ "learning_rate": 9.179471153846155e-06,
165802
+ "loss": 0.3751,
165803
+ "step": 52220
165804
+ },
165805
+ {
165806
+ "epoch": 417.68,
165807
+ "learning_rate": 9.179391025641027e-06,
165808
+ "loss": 0.2995,
165809
+ "step": 52225
165810
+ },
165811
+ {
165812
+ "epoch": 417.72,
165813
+ "learning_rate": 9.179310897435898e-06,
165814
+ "loss": 0.4102,
165815
+ "step": 52230
165816
+ },
165817
+ {
165818
+ "epoch": 417.76,
165819
+ "learning_rate": 9.179230769230771e-06,
165820
+ "loss": 0.5121,
165821
+ "step": 52235
165822
+ },
165823
+ {
165824
+ "epoch": 417.8,
165825
+ "learning_rate": 9.17915064102564e-06,
165826
+ "loss": 1.2505,
165827
+ "step": 52240
165828
+ },
165829
+ {
165830
+ "epoch": 417.84,
165831
+ "learning_rate": 9.179070512820514e-06,
165832
+ "loss": 0.428,
165833
+ "step": 52245
165834
+ },
165835
+ {
165836
+ "epoch": 417.88,
165837
+ "learning_rate": 9.178990384615385e-06,
165838
+ "loss": 0.3032,
165839
+ "step": 52250
165840
+ },
165841
+ {
165842
+ "epoch": 417.92,
165843
+ "learning_rate": 9.178910256410256e-06,
165844
+ "loss": 0.3661,
165845
+ "step": 52255
165846
+ },
165847
+ {
165848
+ "epoch": 417.96,
165849
+ "learning_rate": 9.178830128205128e-06,
165850
+ "loss": 0.5023,
165851
+ "step": 52260
165852
+ },
165853
+ {
165854
+ "epoch": 418.0,
165855
+ "learning_rate": 9.17875e-06,
165856
+ "loss": 1.5163,
165857
+ "step": 52265
165858
+ },
165859
+ {
165860
+ "epoch": 418.0,
165861
+ "eval_loss": 0.421318918466568,
165862
+ "eval_runtime": 37.5508,
165863
+ "eval_samples_per_second": 22.37,
165864
+ "eval_steps_per_second": 0.719,
165865
+ "eval_wer": 0.1988814317673378,
165866
+ "step": 52265
165867
+ },
165868
+ {
165869
+ "epoch": 418.04,
165870
+ "learning_rate": 9.178669871794872e-06,
165871
+ "loss": 0.4194,
165872
+ "step": 52270
165873
+ },
165874
+ {
165875
+ "epoch": 418.08,
165876
+ "learning_rate": 9.178589743589744e-06,
165877
+ "loss": 0.3268,
165878
+ "step": 52275
165879
+ },
165880
+ {
165881
+ "epoch": 418.12,
165882
+ "learning_rate": 9.178509615384617e-06,
165883
+ "loss": 0.3918,
165884
+ "step": 52280
165885
+ },
165886
+ {
165887
+ "epoch": 418.16,
165888
+ "learning_rate": 9.178429487179488e-06,
165889
+ "loss": 0.5551,
165890
+ "step": 52285
165891
+ },
165892
+ {
165893
+ "epoch": 418.2,
165894
+ "learning_rate": 9.17834935897436e-06,
165895
+ "loss": 1.2501,
165896
+ "step": 52290
165897
+ },
165898
+ {
165899
+ "epoch": 418.24,
165900
+ "learning_rate": 9.17826923076923e-06,
165901
+ "loss": 0.4417,
165902
+ "step": 52295
165903
+ },
165904
+ {
165905
+ "epoch": 418.28,
165906
+ "learning_rate": 9.178189102564104e-06,
165907
+ "loss": 0.3786,
165908
+ "step": 52300
165909
+ },
165910
+ {
165911
+ "epoch": 418.32,
165912
+ "learning_rate": 9.178108974358975e-06,
165913
+ "loss": 0.3834,
165914
+ "step": 52305
165915
+ },
165916
+ {
165917
+ "epoch": 418.36,
165918
+ "learning_rate": 9.178028846153846e-06,
165919
+ "loss": 0.4986,
165920
+ "step": 52310
165921
+ },
165922
+ {
165923
+ "epoch": 418.4,
165924
+ "learning_rate": 9.17794871794872e-06,
165925
+ "loss": 1.1479,
165926
+ "step": 52315
165927
+ },
165928
+ {
165929
+ "epoch": 418.44,
165930
+ "learning_rate": 9.177868589743591e-06,
165931
+ "loss": 0.3841,
165932
+ "step": 52320
165933
+ },
165934
+ {
165935
+ "epoch": 418.48,
165936
+ "learning_rate": 9.177788461538462e-06,
165937
+ "loss": 0.3993,
165938
+ "step": 52325
165939
+ },
165940
+ {
165941
+ "epoch": 418.52,
165942
+ "learning_rate": 9.177708333333334e-06,
165943
+ "loss": 0.3836,
165944
+ "step": 52330
165945
+ },
165946
+ {
165947
+ "epoch": 418.56,
165948
+ "learning_rate": 9.177628205128207e-06,
165949
+ "loss": 0.5145,
165950
+ "step": 52335
165951
+ },
165952
+ {
165953
+ "epoch": 418.6,
165954
+ "learning_rate": 9.177548076923078e-06,
165955
+ "loss": 1.4892,
165956
+ "step": 52340
165957
+ },
165958
+ {
165959
+ "epoch": 418.64,
165960
+ "learning_rate": 9.17746794871795e-06,
165961
+ "loss": 0.3854,
165962
+ "step": 52345
165963
+ },
165964
+ {
165965
+ "epoch": 418.68,
165966
+ "learning_rate": 9.17738782051282e-06,
165967
+ "loss": 0.5018,
165968
+ "step": 52350
165969
+ },
165970
+ {
165971
+ "epoch": 418.72,
165972
+ "learning_rate": 9.177307692307694e-06,
165973
+ "loss": 0.379,
165974
+ "step": 52355
165975
+ },
165976
+ {
165977
+ "epoch": 418.76,
165978
+ "learning_rate": 9.177227564102565e-06,
165979
+ "loss": 0.4675,
165980
+ "step": 52360
165981
+ },
165982
+ {
165983
+ "epoch": 418.8,
165984
+ "learning_rate": 9.177147435897436e-06,
165985
+ "loss": 1.1614,
165986
+ "step": 52365
165987
+ },
165988
+ {
165989
+ "epoch": 418.84,
165990
+ "learning_rate": 9.17706730769231e-06,
165991
+ "loss": 0.6507,
165992
+ "step": 52370
165993
+ },
165994
+ {
165995
+ "epoch": 418.88,
165996
+ "learning_rate": 9.176987179487181e-06,
165997
+ "loss": 0.357,
165998
+ "step": 52375
165999
+ },
166000
+ {
166001
+ "epoch": 418.92,
166002
+ "learning_rate": 9.176907051282052e-06,
166003
+ "loss": 0.3412,
166004
+ "step": 52380
166005
+ },
166006
+ {
166007
+ "epoch": 418.96,
166008
+ "learning_rate": 9.176826923076924e-06,
166009
+ "loss": 0.5797,
166010
+ "step": 52385
166011
+ },
166012
+ {
166013
+ "epoch": 419.0,
166014
+ "learning_rate": 9.176746794871797e-06,
166015
+ "loss": 1.4001,
166016
+ "step": 52390
166017
+ },
166018
+ {
166019
+ "epoch": 419.0,
166020
+ "eval_loss": 0.4143487215042114,
166021
+ "eval_runtime": 39.3714,
166022
+ "eval_samples_per_second": 21.335,
166023
+ "eval_steps_per_second": 0.686,
166024
+ "eval_wer": 0.2074479579341641,
166025
+ "step": 52390
166026
+ },
166027
+ {
166028
+ "epoch": 419.04,
166029
+ "learning_rate": 9.176666666666666e-06,
166030
+ "loss": 0.4064,
166031
+ "step": 52395
166032
+ },
166033
+ {
166034
+ "epoch": 419.08,
166035
+ "learning_rate": 9.17658653846154e-06,
166036
+ "loss": 0.4067,
166037
+ "step": 52400
166038
+ },
166039
+ {
166040
+ "epoch": 419.12,
166041
+ "learning_rate": 9.17650641025641e-06,
166042
+ "loss": 0.3486,
166043
+ "step": 52405
166044
+ },
166045
+ {
166046
+ "epoch": 419.16,
166047
+ "learning_rate": 9.176426282051282e-06,
166048
+ "loss": 0.4744,
166049
+ "step": 52410
166050
+ },
166051
+ {
166052
+ "epoch": 419.2,
166053
+ "learning_rate": 9.176346153846155e-06,
166054
+ "loss": 1.2905,
166055
+ "step": 52415
166056
+ },
166057
+ {
166058
+ "epoch": 419.24,
166059
+ "learning_rate": 9.176266025641027e-06,
166060
+ "loss": 0.3739,
166061
+ "step": 52420
166062
+ },
166063
+ {
166064
+ "epoch": 419.28,
166065
+ "learning_rate": 9.176185897435898e-06,
166066
+ "loss": 0.3682,
166067
+ "step": 52425
166068
+ },
166069
+ {
166070
+ "epoch": 419.32,
166071
+ "learning_rate": 9.17610576923077e-06,
166072
+ "loss": 0.3172,
166073
+ "step": 52430
166074
+ },
166075
+ {
166076
+ "epoch": 419.36,
166077
+ "learning_rate": 9.176025641025642e-06,
166078
+ "loss": 0.5588,
166079
+ "step": 52435
166080
+ },
166081
+ {
166082
+ "epoch": 419.4,
166083
+ "learning_rate": 9.175945512820514e-06,
166084
+ "loss": 1.3292,
166085
+ "step": 52440
166086
+ },
166087
+ {
166088
+ "epoch": 419.44,
166089
+ "learning_rate": 9.175865384615385e-06,
166090
+ "loss": 0.3841,
166091
+ "step": 52445
166092
+ },
166093
+ {
166094
+ "epoch": 419.48,
166095
+ "learning_rate": 9.175785256410256e-06,
166096
+ "loss": 0.3479,
166097
+ "step": 52450
166098
+ },
166099
+ {
166100
+ "epoch": 419.52,
166101
+ "learning_rate": 9.17570512820513e-06,
166102
+ "loss": 0.3379,
166103
+ "step": 52455
166104
+ },
166105
+ {
166106
+ "epoch": 419.56,
166107
+ "learning_rate": 9.175625e-06,
166108
+ "loss": 0.5808,
166109
+ "step": 52460
166110
+ },
166111
+ {
166112
+ "epoch": 419.6,
166113
+ "learning_rate": 9.175544871794872e-06,
166114
+ "loss": 1.3315,
166115
+ "step": 52465
166116
+ },
166117
+ {
166118
+ "epoch": 419.64,
166119
+ "learning_rate": 9.175464743589745e-06,
166120
+ "loss": 0.5561,
166121
+ "step": 52470
166122
+ },
166123
+ {
166124
+ "epoch": 419.68,
166125
+ "learning_rate": 9.175384615384617e-06,
166126
+ "loss": 0.3038,
166127
+ "step": 52475
166128
+ },
166129
+ {
166130
+ "epoch": 419.72,
166131
+ "learning_rate": 9.175304487179488e-06,
166132
+ "loss": 0.2846,
166133
+ "step": 52480
166134
+ },
166135
+ {
166136
+ "epoch": 419.76,
166137
+ "learning_rate": 9.17522435897436e-06,
166138
+ "loss": 0.5519,
166139
+ "step": 52485
166140
+ },
166141
+ {
166142
+ "epoch": 419.8,
166143
+ "learning_rate": 9.175144230769232e-06,
166144
+ "loss": 1.2826,
166145
+ "step": 52490
166146
+ },
166147
+ {
166148
+ "epoch": 419.84,
166149
+ "learning_rate": 9.175064102564104e-06,
166150
+ "loss": 0.4148,
166151
+ "step": 52495
166152
+ },
166153
+ {
166154
+ "epoch": 419.88,
166155
+ "learning_rate": 9.174983974358975e-06,
166156
+ "loss": 0.3156,
166157
+ "step": 52500
166158
+ },
166159
+ {
166160
+ "epoch": 419.92,
166161
+ "learning_rate": 9.174903846153846e-06,
166162
+ "loss": 0.3902,
166163
+ "step": 52505
166164
+ },
166165
+ {
166166
+ "epoch": 419.96,
166167
+ "learning_rate": 9.17482371794872e-06,
166168
+ "loss": 0.571,
166169
+ "step": 52510
166170
+ },
166171
+ {
166172
+ "epoch": 420.0,
166173
+ "learning_rate": 9.17474358974359e-06,
166174
+ "loss": 1.6161,
166175
+ "step": 52515
166176
+ },
166177
+ {
166178
+ "epoch": 420.0,
166179
+ "eval_loss": 0.40050163865089417,
166180
+ "eval_runtime": 38.5955,
166181
+ "eval_samples_per_second": 21.764,
166182
+ "eval_steps_per_second": 0.7,
166183
+ "eval_wer": 0.18992848189928482,
166184
+ "step": 52515
166185
  }
166186
  ],
166187
+ "max_steps": 625000,
166188
  "num_train_epochs": 5000,
166189
+ "total_flos": 1.4776861135717535e+20,
166190
  "trial_name": null,
166191
  "trial_params": null
166192
  }
model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52515}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52639}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52639}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44b00d4f58219b4d318df10c240cc6442e9022f96f2d6f17ab28ad2f843e3d57
3
  size 722165009
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e6078b1a25bbfb93877b366a383e3230a0601ad5cee0d6d81e8efa4520c3950
3
  size 722165009
model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52639}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52639}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:db8ab23d963db3ed53e208e477e1a7c184c0310e51be6d499da983ba29245b8e
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c21d02d9671d703b466f4b6d46fb6b63007714a088428656792c5d6dd669f30
3
  size 377909911
model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52639}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dd830621b75a7c006bbc6271fd476471660a439ea04437fd320d620d0c7c6a0f
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31b6a4c636c3176580ba417b7373c6cfc166eb27e1434eb9258492ab9053129e
3
  size 14503
model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52639}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dab950eca4068770617d1edbb459049e304099815892cae618787c2f1d967411
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec8225ca5a3e0ec39b9289e48fc0dc88985a398292b7dfa2abaa0c27d66f24e5
3
  size 559
model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52639}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8b0222eb9c2e85654b83c859b959cbaa02d9327f4d42f81bbeea8e5e0ff98e3a
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8186c12820d92b2b7071f4326ea5609e06d88ce375b7275fe1a132a598badd19
3
  size 623
model-bin/finetune/base/{checkpoint-46666 β†’ checkpoint-52639}/trainer_state.json RENAMED
The diff for this file is too large to render. See raw diff
 
model-bin/finetune/base/{checkpoint-52015 β†’ checkpoint-52639}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629752462.5473433/events.out.tfevents.1629752462.74272264b15c.932.223 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c489f511d82915b4e7f88bf4eb3211dc8aae892ff0910bd9ff5ee8a195dacadd
3
+ size 4194
model-bin/finetune/base/log/1629753109.9928777/events.out.tfevents.1629753109.74272264b15c.932.225 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e26ccb6cb1eeea2c72c6474688ab33f0b5ae0a460f8f963c103e5a876c73149a
3
+ size 4194
model-bin/finetune/base/log/1629753750.9052882/events.out.tfevents.1629753750.74272264b15c.932.227 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b209c39007ce8023b09151816d066f42d4cb4348b104fc2923d04a11fe3de6e
3
+ size 4194
model-bin/finetune/base/log/1629754385.4896328/events.out.tfevents.1629754385.74272264b15c.932.229 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b0beb0b33aca7231ecea9d3317ed34eec395d17dc2436a0b32b4e3ffbf6a94c
3
+ size 4194
model-bin/finetune/base/log/1629755043.4480946/events.out.tfevents.1629755043.74272264b15c.932.231 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd61fe9a26c7e4bc4d3b9c61e42fc30f39d40646f8b419baee69f34f6b430824
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629752462.74272264b15c.932.222 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:755a878e8436f6d5c0096545a510104544a3393d3db96b16f5e99824eeb87cd4
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629753109.74272264b15c.932.224 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad11b22666286ca1d9937d8e0d88244eb72bb28b6aaa8f29ba17ca2b614dd526
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629753750.74272264b15c.932.226 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38e53c9f9d3830c0ac7b059a91db4c4b8433d5eb61fa8cc0eecfa53b3401c7ee
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629754385.74272264b15c.932.228 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:645c61ea40c25a843bf499a1a4de5d29800abb44c1a3baf40c5046e3159abf49
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629755043.74272264b15c.932.230 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d0fe88ffe689f196442935fd701efe5d0c8313eba581eea66cd7ff052819ad9
3
+ size 8462