Check commited on
Commit
3d3655e
Β·
1 Parent(s): cd1c869

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/trainer_state.json +792 -3
  9. model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629925935.501896/events.out.tfevents.1629925935.7e498afd5545.7645.155 +3 -0
  11. model-bin/finetune/base/log/1629926415.8598046/events.out.tfevents.1629926415.7e498afd5545.7645.157 +3 -0
  12. model-bin/finetune/base/log/1629926892.2024448/events.out.tfevents.1629926894.7e498afd5545.7645.159 +3 -0
  13. model-bin/finetune/base/log/1629927362.1056397/events.out.tfevents.1629927362.7e498afd5545.7645.161 +3 -0
  14. model-bin/finetune/base/log/1629927831.7978618/events.out.tfevents.1629927831.7e498afd5545.7645.163 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629925935.7e498afd5545.7645.154 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629926415.7e498afd5545.7645.156 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629926892.7e498afd5545.7645.158 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629927362.7e498afd5545.7645.160 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629927831.7e498afd5545.7645.162 +3 -0
model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8cb828e60d114ad9dc3ceadb37368203bd7eecf6bfd146372b381c4c22e7bfd6
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8bc3686dbe43b18d158600bbc73da971d3107e47c128d0290505faf633a2279
3
  size 722165393
model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7210d43a584dc7110bb013457c8bde2d07cac2ead64b2d96061683da7c32196f
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90f24c971fd277add21080d696ac93bf42778470903be4116a7d7279bc76d5ba
3
  size 377909911
model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:483bbb4122a05e42b5f275d32a45398659861e698779087c7b6c1925b5ee3053
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cd302c43a1d5fefe9e23746f7da370464720b5481aac68cd427d03344fe3abc
3
  size 14503
model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d7124f58fe067811067c4ed41a5b8e4aa29a7baa624a2afb0e6cbdd6a7f276b1
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:366b40c92d99c2539fd4abe0fa5bd065289f2be7017a7878394de8391d6acb41
3
  size 559
model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8a9aaf9f33c3d8265d80f1e4b5cf1209f542c0719612971f697bb6b8e1783bdb
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a32b27d201506de22e1468d6811196605d441be5f1ee480339d3343ad95f2cf3
3
  size 623
model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1822989349981638,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-79148",
4
- "epoch": 692.0,
5
- "global_step": 85866,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -208614,11 +208614,800 @@
208614
  "eval_steps_per_second": 0.652,
208615
  "eval_wer": 0.1846578526805172,
208616
  "step": 85866
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
208617
  }
208618
  ],
208619
  "max_steps": 620000,
208620
  "num_train_epochs": 5000,
208621
- "total_flos": 2.4164723863623284e+20,
208622
  "trial_name": null,
208623
  "trial_params": null
208624
  }
 
1
  {
2
  "best_metric": 0.1822989349981638,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-79148",
4
+ "epoch": 696.995983935743,
5
+ "global_step": 86488,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
208614
  "eval_steps_per_second": 0.652,
208615
  "eval_wer": 0.1846578526805172,
208616
  "step": 85866
208617
+ },
208618
+ {
208619
+ "epoch": 692.03,
208620
+ "learning_rate": 8.629466882067852e-06,
208621
+ "loss": 0.3787,
208622
+ "step": 85870
208623
+ },
208624
+ {
208625
+ "epoch": 692.07,
208626
+ "learning_rate": 8.629386106623587e-06,
208627
+ "loss": 0.2704,
208628
+ "step": 85875
208629
+ },
208630
+ {
208631
+ "epoch": 692.11,
208632
+ "learning_rate": 8.629305331179322e-06,
208633
+ "loss": 0.2762,
208634
+ "step": 85880
208635
+ },
208636
+ {
208637
+ "epoch": 692.15,
208638
+ "learning_rate": 8.629224555735057e-06,
208639
+ "loss": 0.4332,
208640
+ "step": 85885
208641
+ },
208642
+ {
208643
+ "epoch": 692.19,
208644
+ "learning_rate": 8.629143780290792e-06,
208645
+ "loss": 1.0391,
208646
+ "step": 85890
208647
+ },
208648
+ {
208649
+ "epoch": 692.23,
208650
+ "learning_rate": 8.629063004846527e-06,
208651
+ "loss": 0.718,
208652
+ "step": 85895
208653
+ },
208654
+ {
208655
+ "epoch": 692.27,
208656
+ "learning_rate": 8.628982229402262e-06,
208657
+ "loss": 0.2963,
208658
+ "step": 85900
208659
+ },
208660
+ {
208661
+ "epoch": 692.31,
208662
+ "learning_rate": 8.628901453957997e-06,
208663
+ "loss": 0.3739,
208664
+ "step": 85905
208665
+ },
208666
+ {
208667
+ "epoch": 692.35,
208668
+ "learning_rate": 8.628820678513732e-06,
208669
+ "loss": 0.4223,
208670
+ "step": 85910
208671
+ },
208672
+ {
208673
+ "epoch": 692.39,
208674
+ "learning_rate": 8.628739903069468e-06,
208675
+ "loss": 0.8857,
208676
+ "step": 85915
208677
+ },
208678
+ {
208679
+ "epoch": 692.43,
208680
+ "learning_rate": 8.628659127625202e-06,
208681
+ "loss": 0.6445,
208682
+ "step": 85920
208683
+ },
208684
+ {
208685
+ "epoch": 692.47,
208686
+ "learning_rate": 8.628578352180938e-06,
208687
+ "loss": 0.3072,
208688
+ "step": 85925
208689
+ },
208690
+ {
208691
+ "epoch": 692.51,
208692
+ "learning_rate": 8.628497576736672e-06,
208693
+ "loss": 0.3156,
208694
+ "step": 85930
208695
+ },
208696
+ {
208697
+ "epoch": 692.55,
208698
+ "learning_rate": 8.628416801292408e-06,
208699
+ "loss": 0.4384,
208700
+ "step": 85935
208701
+ },
208702
+ {
208703
+ "epoch": 692.59,
208704
+ "learning_rate": 8.628336025848142e-06,
208705
+ "loss": 0.9746,
208706
+ "step": 85940
208707
+ },
208708
+ {
208709
+ "epoch": 692.63,
208710
+ "learning_rate": 8.628255250403878e-06,
208711
+ "loss": 0.6991,
208712
+ "step": 85945
208713
+ },
208714
+ {
208715
+ "epoch": 692.67,
208716
+ "learning_rate": 8.628174474959612e-06,
208717
+ "loss": 0.2924,
208718
+ "step": 85950
208719
+ },
208720
+ {
208721
+ "epoch": 692.71,
208722
+ "learning_rate": 8.628093699515348e-06,
208723
+ "loss": 0.271,
208724
+ "step": 85955
208725
+ },
208726
+ {
208727
+ "epoch": 692.76,
208728
+ "learning_rate": 8.628012924071082e-06,
208729
+ "loss": 0.4228,
208730
+ "step": 85960
208731
+ },
208732
+ {
208733
+ "epoch": 692.8,
208734
+ "learning_rate": 8.627932148626818e-06,
208735
+ "loss": 0.9741,
208736
+ "step": 85965
208737
+ },
208738
+ {
208739
+ "epoch": 692.84,
208740
+ "learning_rate": 8.627851373182552e-06,
208741
+ "loss": 0.5785,
208742
+ "step": 85970
208743
+ },
208744
+ {
208745
+ "epoch": 692.88,
208746
+ "learning_rate": 8.627770597738288e-06,
208747
+ "loss": 0.2988,
208748
+ "step": 85975
208749
+ },
208750
+ {
208751
+ "epoch": 692.92,
208752
+ "learning_rate": 8.627689822294024e-06,
208753
+ "loss": 0.3336,
208754
+ "step": 85980
208755
+ },
208756
+ {
208757
+ "epoch": 692.96,
208758
+ "learning_rate": 8.627609046849758e-06,
208759
+ "loss": 0.373,
208760
+ "step": 85985
208761
+ },
208762
+ {
208763
+ "epoch": 693.0,
208764
+ "learning_rate": 8.627528271405494e-06,
208765
+ "loss": 0.9426,
208766
+ "step": 85990
208767
+ },
208768
+ {
208769
+ "epoch": 693.0,
208770
+ "eval_loss": 0.33618369698524475,
208771
+ "eval_runtime": 45.8818,
208772
+ "eval_samples_per_second": 18.308,
208773
+ "eval_steps_per_second": 0.588,
208774
+ "eval_wer": 0.1844791889769401,
208775
+ "step": 85990
208776
+ },
208777
+ {
208778
+ "epoch": 687.04,
208779
+ "learning_rate": 8.627447495961228e-06,
208780
+ "loss": 0.345,
208781
+ "step": 85995
208782
+ },
208783
+ {
208784
+ "epoch": 687.08,
208785
+ "learning_rate": 8.627366720516964e-06,
208786
+ "loss": 0.3009,
208787
+ "step": 86000
208788
+ },
208789
+ {
208790
+ "epoch": 687.12,
208791
+ "learning_rate": 8.627285945072698e-06,
208792
+ "loss": 0.3208,
208793
+ "step": 86005
208794
+ },
208795
+ {
208796
+ "epoch": 687.16,
208797
+ "learning_rate": 8.627205169628434e-06,
208798
+ "loss": 0.5082,
208799
+ "step": 86010
208800
+ },
208801
+ {
208802
+ "epoch": 687.2,
208803
+ "learning_rate": 8.627124394184168e-06,
208804
+ "loss": 1.2031,
208805
+ "step": 86015
208806
+ },
208807
+ {
208808
+ "epoch": 687.24,
208809
+ "learning_rate": 8.627043618739904e-06,
208810
+ "loss": 0.3518,
208811
+ "step": 86020
208812
+ },
208813
+ {
208814
+ "epoch": 687.28,
208815
+ "learning_rate": 8.626962843295638e-06,
208816
+ "loss": 0.3394,
208817
+ "step": 86025
208818
+ },
208819
+ {
208820
+ "epoch": 687.32,
208821
+ "learning_rate": 8.626882067851374e-06,
208822
+ "loss": 0.3621,
208823
+ "step": 86030
208824
+ },
208825
+ {
208826
+ "epoch": 687.36,
208827
+ "learning_rate": 8.626801292407108e-06,
208828
+ "loss": 0.457,
208829
+ "step": 86035
208830
+ },
208831
+ {
208832
+ "epoch": 687.4,
208833
+ "learning_rate": 8.626720516962844e-06,
208834
+ "loss": 1.3182,
208835
+ "step": 86040
208836
+ },
208837
+ {
208838
+ "epoch": 687.44,
208839
+ "learning_rate": 8.62663974151858e-06,
208840
+ "loss": 0.3442,
208841
+ "step": 86045
208842
+ },
208843
+ {
208844
+ "epoch": 687.48,
208845
+ "learning_rate": 8.626558966074314e-06,
208846
+ "loss": 0.3368,
208847
+ "step": 86050
208848
+ },
208849
+ {
208850
+ "epoch": 687.52,
208851
+ "learning_rate": 8.62647819063005e-06,
208852
+ "loss": 0.3164,
208853
+ "step": 86055
208854
+ },
208855
+ {
208856
+ "epoch": 687.56,
208857
+ "learning_rate": 8.626397415185784e-06,
208858
+ "loss": 0.462,
208859
+ "step": 86060
208860
+ },
208861
+ {
208862
+ "epoch": 687.6,
208863
+ "learning_rate": 8.62631663974152e-06,
208864
+ "loss": 1.2451,
208865
+ "step": 86065
208866
+ },
208867
+ {
208868
+ "epoch": 687.64,
208869
+ "learning_rate": 8.626235864297254e-06,
208870
+ "loss": 0.3596,
208871
+ "step": 86070
208872
+ },
208873
+ {
208874
+ "epoch": 687.68,
208875
+ "learning_rate": 8.62615508885299e-06,
208876
+ "loss": 0.3069,
208877
+ "step": 86075
208878
+ },
208879
+ {
208880
+ "epoch": 687.72,
208881
+ "learning_rate": 8.626074313408724e-06,
208882
+ "loss": 0.3492,
208883
+ "step": 86080
208884
+ },
208885
+ {
208886
+ "epoch": 687.76,
208887
+ "learning_rate": 8.62599353796446e-06,
208888
+ "loss": 0.4954,
208889
+ "step": 86085
208890
+ },
208891
+ {
208892
+ "epoch": 687.8,
208893
+ "learning_rate": 8.625912762520194e-06,
208894
+ "loss": 1.1525,
208895
+ "step": 86090
208896
+ },
208897
+ {
208898
+ "epoch": 687.84,
208899
+ "learning_rate": 8.62583198707593e-06,
208900
+ "loss": 0.3123,
208901
+ "step": 86095
208902
+ },
208903
+ {
208904
+ "epoch": 687.88,
208905
+ "learning_rate": 8.625751211631664e-06,
208906
+ "loss": 0.2866,
208907
+ "step": 86100
208908
+ },
208909
+ {
208910
+ "epoch": 687.92,
208911
+ "learning_rate": 8.6256704361874e-06,
208912
+ "loss": 0.3291,
208913
+ "step": 86105
208914
+ },
208915
+ {
208916
+ "epoch": 687.96,
208917
+ "learning_rate": 8.625589660743134e-06,
208918
+ "loss": 0.6972,
208919
+ "step": 86110
208920
+ },
208921
+ {
208922
+ "epoch": 688.0,
208923
+ "learning_rate": 8.62550888529887e-06,
208924
+ "loss": 1.366,
208925
+ "step": 86115
208926
+ },
208927
+ {
208928
+ "epoch": 688.0,
208929
+ "eval_loss": 0.32431867718696594,
208930
+ "eval_runtime": 41.5952,
208931
+ "eval_samples_per_second": 20.171,
208932
+ "eval_steps_per_second": 0.649,
208933
+ "eval_wer": 0.18960425280567042,
208934
+ "step": 86115
208935
+ },
208936
+ {
208937
+ "epoch": 694.04,
208938
+ "learning_rate": 8.625428109854605e-06,
208939
+ "loss": 0.3579,
208940
+ "step": 86120
208941
+ },
208942
+ {
208943
+ "epoch": 694.08,
208944
+ "learning_rate": 8.62534733441034e-06,
208945
+ "loss": 0.2554,
208946
+ "step": 86125
208947
+ },
208948
+ {
208949
+ "epoch": 694.12,
208950
+ "learning_rate": 8.625266558966075e-06,
208951
+ "loss": 0.3368,
208952
+ "step": 86130
208953
+ },
208954
+ {
208955
+ "epoch": 694.16,
208956
+ "learning_rate": 8.62518578352181e-06,
208957
+ "loss": 0.3954,
208958
+ "step": 86135
208959
+ },
208960
+ {
208961
+ "epoch": 694.2,
208962
+ "learning_rate": 8.625105008077545e-06,
208963
+ "loss": 1.1405,
208964
+ "step": 86140
208965
+ },
208966
+ {
208967
+ "epoch": 694.24,
208968
+ "learning_rate": 8.62502423263328e-06,
208969
+ "loss": 0.3505,
208970
+ "step": 86145
208971
+ },
208972
+ {
208973
+ "epoch": 694.28,
208974
+ "learning_rate": 8.624943457189015e-06,
208975
+ "loss": 0.3632,
208976
+ "step": 86150
208977
+ },
208978
+ {
208979
+ "epoch": 694.32,
208980
+ "learning_rate": 8.62486268174475e-06,
208981
+ "loss": 0.2912,
208982
+ "step": 86155
208983
+ },
208984
+ {
208985
+ "epoch": 694.36,
208986
+ "learning_rate": 8.624781906300485e-06,
208987
+ "loss": 0.4371,
208988
+ "step": 86160
208989
+ },
208990
+ {
208991
+ "epoch": 694.4,
208992
+ "learning_rate": 8.62470113085622e-06,
208993
+ "loss": 1.0906,
208994
+ "step": 86165
208995
+ },
208996
+ {
208997
+ "epoch": 694.44,
208998
+ "learning_rate": 8.624620355411955e-06,
208999
+ "loss": 0.3488,
209000
+ "step": 86170
209001
+ },
209002
+ {
209003
+ "epoch": 694.48,
209004
+ "learning_rate": 8.62453957996769e-06,
209005
+ "loss": 0.281,
209006
+ "step": 86175
209007
+ },
209008
+ {
209009
+ "epoch": 694.52,
209010
+ "learning_rate": 8.624458804523425e-06,
209011
+ "loss": 0.3886,
209012
+ "step": 86180
209013
+ },
209014
+ {
209015
+ "epoch": 694.56,
209016
+ "learning_rate": 8.624378029079161e-06,
209017
+ "loss": 0.4976,
209018
+ "step": 86185
209019
+ },
209020
+ {
209021
+ "epoch": 694.6,
209022
+ "learning_rate": 8.624297253634895e-06,
209023
+ "loss": 1.2062,
209024
+ "step": 86190
209025
+ },
209026
+ {
209027
+ "epoch": 694.64,
209028
+ "learning_rate": 8.624216478190631e-06,
209029
+ "loss": 0.3087,
209030
+ "step": 86195
209031
+ },
209032
+ {
209033
+ "epoch": 694.68,
209034
+ "learning_rate": 8.624135702746365e-06,
209035
+ "loss": 0.3002,
209036
+ "step": 86200
209037
+ },
209038
+ {
209039
+ "epoch": 694.72,
209040
+ "learning_rate": 8.624054927302101e-06,
209041
+ "loss": 0.314,
209042
+ "step": 86205
209043
+ },
209044
+ {
209045
+ "epoch": 694.76,
209046
+ "learning_rate": 8.623974151857835e-06,
209047
+ "loss": 0.4838,
209048
+ "step": 86210
209049
+ },
209050
+ {
209051
+ "epoch": 694.8,
209052
+ "learning_rate": 8.623893376413571e-06,
209053
+ "loss": 1.1279,
209054
+ "step": 86215
209055
+ },
209056
+ {
209057
+ "epoch": 694.84,
209058
+ "learning_rate": 8.623812600969305e-06,
209059
+ "loss": 0.3075,
209060
+ "step": 86220
209061
+ },
209062
+ {
209063
+ "epoch": 694.88,
209064
+ "learning_rate": 8.62373182552504e-06,
209065
+ "loss": 0.3139,
209066
+ "step": 86225
209067
+ },
209068
+ {
209069
+ "epoch": 694.92,
209070
+ "learning_rate": 8.623651050080775e-06,
209071
+ "loss": 0.3913,
209072
+ "step": 86230
209073
+ },
209074
+ {
209075
+ "epoch": 694.96,
209076
+ "learning_rate": 8.62357027463651e-06,
209077
+ "loss": 0.4897,
209078
+ "step": 86235
209079
+ },
209080
+ {
209081
+ "epoch": 695.0,
209082
+ "eval_loss": 0.3379378020763397,
209083
+ "eval_runtime": 41.768,
209084
+ "eval_samples_per_second": 20.087,
209085
+ "eval_steps_per_second": 0.646,
209086
+ "eval_wer": 0.18313218183132182,
209087
+ "step": 86239
209088
+ },
209089
+ {
209090
+ "epoch": 689.01,
209091
+ "learning_rate": 8.623489499192245e-06,
209092
+ "loss": 0.28,
209093
+ "step": 86240
209094
+ },
209095
+ {
209096
+ "epoch": 689.05,
209097
+ "learning_rate": 8.62340872374798e-06,
209098
+ "loss": 0.5216,
209099
+ "step": 86245
209100
+ },
209101
+ {
209102
+ "epoch": 689.09,
209103
+ "learning_rate": 8.623327948303717e-06,
209104
+ "loss": 0.3201,
209105
+ "step": 86250
209106
+ },
209107
+ {
209108
+ "epoch": 689.13,
209109
+ "learning_rate": 8.62324717285945e-06,
209110
+ "loss": 0.342,
209111
+ "step": 86255
209112
+ },
209113
+ {
209114
+ "epoch": 689.17,
209115
+ "learning_rate": 8.623166397415187e-06,
209116
+ "loss": 0.5204,
209117
+ "step": 86260
209118
+ },
209119
+ {
209120
+ "epoch": 689.21,
209121
+ "learning_rate": 8.62308562197092e-06,
209122
+ "loss": 1.0585,
209123
+ "step": 86265
209124
+ },
209125
+ {
209126
+ "epoch": 689.25,
209127
+ "learning_rate": 8.623004846526657e-06,
209128
+ "loss": 0.33,
209129
+ "step": 86270
209130
+ },
209131
+ {
209132
+ "epoch": 689.29,
209133
+ "learning_rate": 8.62292407108239e-06,
209134
+ "loss": 0.3383,
209135
+ "step": 86275
209136
+ },
209137
+ {
209138
+ "epoch": 689.33,
209139
+ "learning_rate": 8.622843295638127e-06,
209140
+ "loss": 0.3156,
209141
+ "step": 86280
209142
+ },
209143
+ {
209144
+ "epoch": 689.37,
209145
+ "learning_rate": 8.62276252019386e-06,
209146
+ "loss": 0.5836,
209147
+ "step": 86285
209148
+ },
209149
+ {
209150
+ "epoch": 689.41,
209151
+ "learning_rate": 8.622681744749597e-06,
209152
+ "loss": 1.1502,
209153
+ "step": 86290
209154
+ },
209155
+ {
209156
+ "epoch": 689.45,
209157
+ "learning_rate": 8.62260096930533e-06,
209158
+ "loss": 0.3082,
209159
+ "step": 86295
209160
+ },
209161
+ {
209162
+ "epoch": 689.49,
209163
+ "learning_rate": 8.622520193861066e-06,
209164
+ "loss": 0.2888,
209165
+ "step": 86300
209166
+ },
209167
+ {
209168
+ "epoch": 689.53,
209169
+ "learning_rate": 8.622439418416802e-06,
209170
+ "loss": 0.3934,
209171
+ "step": 86305
209172
+ },
209173
+ {
209174
+ "epoch": 689.57,
209175
+ "learning_rate": 8.622358642972536e-06,
209176
+ "loss": 0.6293,
209177
+ "step": 86310
209178
+ },
209179
+ {
209180
+ "epoch": 689.61,
209181
+ "learning_rate": 8.622277867528272e-06,
209182
+ "loss": 1.0595,
209183
+ "step": 86315
209184
+ },
209185
+ {
209186
+ "epoch": 689.65,
209187
+ "learning_rate": 8.622197092084006e-06,
209188
+ "loss": 0.3427,
209189
+ "step": 86320
209190
+ },
209191
+ {
209192
+ "epoch": 689.69,
209193
+ "learning_rate": 8.622116316639742e-06,
209194
+ "loss": 0.3128,
209195
+ "step": 86325
209196
+ },
209197
+ {
209198
+ "epoch": 689.73,
209199
+ "learning_rate": 8.622035541195476e-06,
209200
+ "loss": 0.3138,
209201
+ "step": 86330
209202
+ },
209203
+ {
209204
+ "epoch": 689.77,
209205
+ "learning_rate": 8.621954765751212e-06,
209206
+ "loss": 0.4887,
209207
+ "step": 86335
209208
+ },
209209
+ {
209210
+ "epoch": 689.81,
209211
+ "learning_rate": 8.621873990306946e-06,
209212
+ "loss": 1.1533,
209213
+ "step": 86340
209214
+ },
209215
+ {
209216
+ "epoch": 689.85,
209217
+ "learning_rate": 8.621793214862682e-06,
209218
+ "loss": 0.3736,
209219
+ "step": 86345
209220
+ },
209221
+ {
209222
+ "epoch": 689.89,
209223
+ "learning_rate": 8.621712439418416e-06,
209224
+ "loss": 0.3076,
209225
+ "step": 86350
209226
+ },
209227
+ {
209228
+ "epoch": 689.93,
209229
+ "learning_rate": 8.621631663974152e-06,
209230
+ "loss": 0.3878,
209231
+ "step": 86355
209232
+ },
209233
+ {
209234
+ "epoch": 689.97,
209235
+ "learning_rate": 8.621550888529886e-06,
209236
+ "loss": 0.5588,
209237
+ "step": 86360
209238
+ },
209239
+ {
209240
+ "epoch": 690.0,
209241
+ "eval_loss": 0.4273076355457306,
209242
+ "eval_runtime": 43.4189,
209243
+ "eval_samples_per_second": 19.323,
209244
+ "eval_steps_per_second": 0.622,
209245
+ "eval_wer": 0.18854415274463007,
209246
+ "step": 86364
209247
+ },
209248
+ {
209249
+ "epoch": 696.01,
209250
+ "learning_rate": 8.621470113085622e-06,
209251
+ "loss": 0.3859,
209252
+ "step": 86365
209253
+ },
209254
+ {
209255
+ "epoch": 696.05,
209256
+ "learning_rate": 8.621389337641358e-06,
209257
+ "loss": 0.3438,
209258
+ "step": 86370
209259
+ },
209260
+ {
209261
+ "epoch": 696.09,
209262
+ "learning_rate": 8.621308562197092e-06,
209263
+ "loss": 0.2927,
209264
+ "step": 86375
209265
+ },
209266
+ {
209267
+ "epoch": 696.13,
209268
+ "learning_rate": 8.621227786752828e-06,
209269
+ "loss": 0.3259,
209270
+ "step": 86380
209271
+ },
209272
+ {
209273
+ "epoch": 696.17,
209274
+ "learning_rate": 8.621147011308562e-06,
209275
+ "loss": 0.5074,
209276
+ "step": 86385
209277
+ },
209278
+ {
209279
+ "epoch": 696.21,
209280
+ "learning_rate": 8.621066235864298e-06,
209281
+ "loss": 1.1853,
209282
+ "step": 86390
209283
+ },
209284
+ {
209285
+ "epoch": 696.25,
209286
+ "learning_rate": 8.620985460420032e-06,
209287
+ "loss": 0.3457,
209288
+ "step": 86395
209289
+ },
209290
+ {
209291
+ "epoch": 696.29,
209292
+ "learning_rate": 8.620904684975768e-06,
209293
+ "loss": 1.0157,
209294
+ "step": 86400
209295
+ },
209296
+ {
209297
+ "epoch": 696.33,
209298
+ "learning_rate": 8.620823909531502e-06,
209299
+ "loss": 0.375,
209300
+ "step": 86405
209301
+ },
209302
+ {
209303
+ "epoch": 696.37,
209304
+ "learning_rate": 8.620743134087238e-06,
209305
+ "loss": 0.4831,
209306
+ "step": 86410
209307
+ },
209308
+ {
209309
+ "epoch": 696.41,
209310
+ "learning_rate": 8.620662358642972e-06,
209311
+ "loss": 1.0284,
209312
+ "step": 86415
209313
+ },
209314
+ {
209315
+ "epoch": 696.45,
209316
+ "learning_rate": 8.620581583198708e-06,
209317
+ "loss": 0.3448,
209318
+ "step": 86420
209319
+ },
209320
+ {
209321
+ "epoch": 696.49,
209322
+ "learning_rate": 8.620500807754444e-06,
209323
+ "loss": 0.2996,
209324
+ "step": 86425
209325
+ },
209326
+ {
209327
+ "epoch": 696.53,
209328
+ "learning_rate": 8.620420032310178e-06,
209329
+ "loss": 0.3335,
209330
+ "step": 86430
209331
+ },
209332
+ {
209333
+ "epoch": 696.57,
209334
+ "learning_rate": 8.620339256865914e-06,
209335
+ "loss": 0.499,
209336
+ "step": 86435
209337
+ },
209338
+ {
209339
+ "epoch": 696.61,
209340
+ "learning_rate": 8.620258481421648e-06,
209341
+ "loss": 1.0738,
209342
+ "step": 86440
209343
+ },
209344
+ {
209345
+ "epoch": 696.65,
209346
+ "learning_rate": 8.620177705977384e-06,
209347
+ "loss": 0.3173,
209348
+ "step": 86445
209349
+ },
209350
+ {
209351
+ "epoch": 696.69,
209352
+ "learning_rate": 8.620096930533118e-06,
209353
+ "loss": 0.2994,
209354
+ "step": 86450
209355
+ },
209356
+ {
209357
+ "epoch": 696.73,
209358
+ "learning_rate": 8.620016155088854e-06,
209359
+ "loss": 0.3495,
209360
+ "step": 86455
209361
+ },
209362
+ {
209363
+ "epoch": 696.77,
209364
+ "learning_rate": 8.619935379644588e-06,
209365
+ "loss": 0.5834,
209366
+ "step": 86460
209367
+ },
209368
+ {
209369
+ "epoch": 696.81,
209370
+ "learning_rate": 8.619854604200324e-06,
209371
+ "loss": 1.2537,
209372
+ "step": 86465
209373
+ },
209374
+ {
209375
+ "epoch": 696.85,
209376
+ "learning_rate": 8.619773828756058e-06,
209377
+ "loss": 0.3,
209378
+ "step": 86470
209379
+ },
209380
+ {
209381
+ "epoch": 696.89,
209382
+ "learning_rate": 8.619693053311794e-06,
209383
+ "loss": 0.3382,
209384
+ "step": 86475
209385
+ },
209386
+ {
209387
+ "epoch": 696.93,
209388
+ "learning_rate": 8.61961227786753e-06,
209389
+ "loss": 0.3447,
209390
+ "step": 86480
209391
+ },
209392
+ {
209393
+ "epoch": 696.97,
209394
+ "learning_rate": 8.619531502423264e-06,
209395
+ "loss": 0.6205,
209396
+ "step": 86485
209397
+ },
209398
+ {
209399
+ "epoch": 697.0,
209400
+ "eval_loss": 0.3951117694377899,
209401
+ "eval_runtime": 41.9031,
209402
+ "eval_samples_per_second": 20.022,
209403
+ "eval_steps_per_second": 0.644,
209404
+ "eval_wer": 0.18736416098328712,
209405
+ "step": 86488
209406
  }
209407
  ],
209408
  "max_steps": 620000,
209409
  "num_train_epochs": 5000,
209410
+ "total_flos": 2.434054730572205e+20,
209411
  "trial_name": null,
209412
  "trial_params": null
209413
  }
model-bin/finetune/base/{checkpoint-85866 β†’ checkpoint-86488}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629925935.501896/events.out.tfevents.1629925935.7e498afd5545.7645.155 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10d35973eca77c60db745e002bb5a631b5bf7d5676ad910bc1835a11ffd27a01
3
+ size 4194
model-bin/finetune/base/log/1629926415.8598046/events.out.tfevents.1629926415.7e498afd5545.7645.157 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:438fd695c26b7b472fd72afe695ec74b4fb7aba6ff66ce7fe3cc518c6c69d2e3
3
+ size 4194
model-bin/finetune/base/log/1629926892.2024448/events.out.tfevents.1629926894.7e498afd5545.7645.159 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44dad85a7a60b622c5daf5bf80aefba6937bf41fb14e5460227e0a67a70901c0
3
+ size 4194
model-bin/finetune/base/log/1629927362.1056397/events.out.tfevents.1629927362.7e498afd5545.7645.161 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:916c557b6636c8b3a4942acff7e70d28baa9a540ffe1b43049c06f5757066b51
3
+ size 4194
model-bin/finetune/base/log/1629927831.7978618/events.out.tfevents.1629927831.7e498afd5545.7645.163 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c861c7ca7554ac3a20d873fed86fad3abc57cafab6550c4112487b57b2d45a2
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629925935.7e498afd5545.7645.154 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:668fcc8b68cca7bef549e72b974308869eaeb3b0faa0be94f699d9c41cae17f6
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629926415.7e498afd5545.7645.156 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a451b4bdf900516fc66d812eff3dc384b66ca340255f264acf1f18dc033c6dc6
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629926892.7e498afd5545.7645.158 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:996bb611b82d79219d30f5e750ba4f0cf8dee9f09c4c493f667c145a1dbaa5f4
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1629927362.7e498afd5545.7645.160 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61720124e9f46598b64ca4b443547bc2f1268d14ad24a8c4405128bbf08c1eb6
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629927831.7e498afd5545.7645.162 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63c815b96007b09b21368cd671bfef0e48268901c1ad0c1e7c877d6741652bc7
3
+ size 8622