Check commited on
Commit
f27793c
Β·
1 Parent(s): e05ae20

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/trainer_state.json +799 -4
  9. model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630140557.4831562/events.out.tfevents.1630140557.86bb0ddabf9b.4092.91 +3 -0
  11. model-bin/finetune/base/log/1630140950.614884/events.out.tfevents.1630140950.86bb0ddabf9b.4092.93 +3 -0
  12. model-bin/finetune/base/log/1630141343.6640568/events.out.tfevents.1630141343.86bb0ddabf9b.4092.95 +3 -0
  13. model-bin/finetune/base/log/1630141737.0182173/events.out.tfevents.1630141737.86bb0ddabf9b.4092.97 +3 -0
  14. model-bin/finetune/base/log/1630142244.9424796/events.out.tfevents.1630142244.86bb0ddabf9b.4092.99 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630140557.86bb0ddabf9b.4092.90 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630140950.86bb0ddabf9b.4092.92 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630141343.86bb0ddabf9b.4092.94 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630141737.86bb0ddabf9b.4092.96 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630142244.86bb0ddabf9b.4092.98 +3 -0
model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6acc0eac329fac04e079ada3a4e4ba51754b947dcadf089d81de73e171291f61
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:444e2c1f4325b4b11162c9a02d9ac033783454f04b0ebac59f29f3f36dacd94e
3
  size 722165393
model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cf7df59b02cc7ed9e88e2cd669353a23deeef6a74308cc4e7167ad41676ee351
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c99be2cb9f6bfd0d130822c68453cf6d1ebfc7d18545903ae81217c345b550ed
3
  size 377909911
model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d886e18d042453eba71f225443945e4c5e10aa1d554d7f2496aa448ac824c95f
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:588ff1d2630a26b0738ac1935a67831bad436083ccce50a276fe4bac0f0ef39e
3
  size 14503
model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:653c9a365fa09a2ba870fc34c869e859ff2c2d0e6cac59c1f2aa8a0082f4764f
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45ec721615ea8aa2c0f6eb839aa7fff0833136cc34604aa9f2ae17982ad6c3dd
3
  size 559
model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d6782817a4fd5e153056fb3435c568a9bfe0113ce6e0ca3c04dd844ec1e2e1d6
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4693177c98f9aa50f7adffdb4243c309abaede1dcf51469b33e334fd32bdf34
3
  size 623
model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1735723097017633,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-123203",
4
- "epoch": 988.0,
5
- "global_step": 123577,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -256593,11 +256593,806 @@
256593
  "eval_steps_per_second": 0.738,
256594
  "eval_wer": 0.17904090475141368,
256595
  "step": 123577
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
256596
  }
256597
  ],
256598
- "max_steps": 625000,
256599
  "num_train_epochs": 5000,
256600
- "total_flos": 3.4776844471530645e+20,
256601
  "trial_name": null,
256602
  "trial_params": null
256603
  }
 
1
  {
2
  "best_metric": 0.1735723097017633,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-123203",
4
+ "epoch": 1000.995983935743,
5
+ "global_step": 124200,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
256593
  "eval_steps_per_second": 0.738,
256594
  "eval_wer": 0.17904090475141368,
256595
  "step": 123577
256596
+ },
256597
+ {
256598
+ "epoch": 996.02,
256599
+ "learning_rate": 8.03636217948718e-06,
256600
+ "loss": 0.2697,
256601
+ "step": 123580
256602
+ },
256603
+ {
256604
+ "epoch": 996.06,
256605
+ "learning_rate": 8.036282051282052e-06,
256606
+ "loss": 0.2864,
256607
+ "step": 123585
256608
+ },
256609
+ {
256610
+ "epoch": 996.1,
256611
+ "learning_rate": 8.036201923076925e-06,
256612
+ "loss": 0.2807,
256613
+ "step": 123590
256614
+ },
256615
+ {
256616
+ "epoch": 996.15,
256617
+ "learning_rate": 8.036121794871794e-06,
256618
+ "loss": 0.3349,
256619
+ "step": 123595
256620
+ },
256621
+ {
256622
+ "epoch": 996.19,
256623
+ "learning_rate": 8.036041666666667e-06,
256624
+ "loss": 0.7529,
256625
+ "step": 123600
256626
+ },
256627
+ {
256628
+ "epoch": 996.23,
256629
+ "learning_rate": 8.035961538461539e-06,
256630
+ "loss": 0.8701,
256631
+ "step": 123605
256632
+ },
256633
+ {
256634
+ "epoch": 996.27,
256635
+ "learning_rate": 8.03588141025641e-06,
256636
+ "loss": 0.2396,
256637
+ "step": 123610
256638
+ },
256639
+ {
256640
+ "epoch": 996.31,
256641
+ "learning_rate": 8.035801282051283e-06,
256642
+ "loss": 0.3098,
256643
+ "step": 123615
256644
+ },
256645
+ {
256646
+ "epoch": 996.35,
256647
+ "learning_rate": 8.035721153846155e-06,
256648
+ "loss": 0.3529,
256649
+ "step": 123620
256650
+ },
256651
+ {
256652
+ "epoch": 996.39,
256653
+ "learning_rate": 8.035641025641026e-06,
256654
+ "loss": 0.7107,
256655
+ "step": 123625
256656
+ },
256657
+ {
256658
+ "epoch": 996.43,
256659
+ "learning_rate": 8.035560897435897e-06,
256660
+ "loss": 0.8661,
256661
+ "step": 123630
256662
+ },
256663
+ {
256664
+ "epoch": 996.47,
256665
+ "learning_rate": 8.03548076923077e-06,
256666
+ "loss": 0.2525,
256667
+ "step": 123635
256668
+ },
256669
+ {
256670
+ "epoch": 996.51,
256671
+ "learning_rate": 8.035400641025642e-06,
256672
+ "loss": 0.277,
256673
+ "step": 123640
256674
+ },
256675
+ {
256676
+ "epoch": 996.55,
256677
+ "learning_rate": 8.035320512820513e-06,
256678
+ "loss": 0.3795,
256679
+ "step": 123645
256680
+ },
256681
+ {
256682
+ "epoch": 996.59,
256683
+ "learning_rate": 8.035240384615384e-06,
256684
+ "loss": 0.6803,
256685
+ "step": 123650
256686
+ },
256687
+ {
256688
+ "epoch": 996.63,
256689
+ "learning_rate": 8.035160256410257e-06,
256690
+ "loss": 0.8381,
256691
+ "step": 123655
256692
+ },
256693
+ {
256694
+ "epoch": 996.67,
256695
+ "learning_rate": 8.035080128205129e-06,
256696
+ "loss": 0.2627,
256697
+ "step": 123660
256698
+ },
256699
+ {
256700
+ "epoch": 996.71,
256701
+ "learning_rate": 8.035e-06,
256702
+ "loss": 0.2651,
256703
+ "step": 123665
256704
+ },
256705
+ {
256706
+ "epoch": 996.75,
256707
+ "learning_rate": 8.034919871794873e-06,
256708
+ "loss": 0.3103,
256709
+ "step": 123670
256710
+ },
256711
+ {
256712
+ "epoch": 996.79,
256713
+ "learning_rate": 8.034839743589745e-06,
256714
+ "loss": 0.7412,
256715
+ "step": 123675
256716
+ },
256717
+ {
256718
+ "epoch": 996.83,
256719
+ "learning_rate": 8.034759615384616e-06,
256720
+ "loss": 0.816,
256721
+ "step": 123680
256722
+ },
256723
+ {
256724
+ "epoch": 996.87,
256725
+ "learning_rate": 8.034679487179487e-06,
256726
+ "loss": 0.2746,
256727
+ "step": 123685
256728
+ },
256729
+ {
256730
+ "epoch": 996.91,
256731
+ "learning_rate": 8.03459935897436e-06,
256732
+ "loss": 0.3075,
256733
+ "step": 123690
256734
+ },
256735
+ {
256736
+ "epoch": 996.95,
256737
+ "learning_rate": 8.034519230769232e-06,
256738
+ "loss": 0.4634,
256739
+ "step": 123695
256740
+ },
256741
+ {
256742
+ "epoch": 996.99,
256743
+ "learning_rate": 8.034439102564103e-06,
256744
+ "loss": 0.9143,
256745
+ "step": 123700
256746
+ },
256747
+ {
256748
+ "epoch": 997.0,
256749
+ "eval_loss": 0.35271042585372925,
256750
+ "eval_runtime": 37.4395,
256751
+ "eval_samples_per_second": 22.49,
256752
+ "eval_steps_per_second": 0.721,
256753
+ "eval_wer": 0.18502449265220433,
256754
+ "step": 123701
256755
+ },
256756
+ {
256757
+ "epoch": 989.03,
256758
+ "learning_rate": 8.034358974358974e-06,
256759
+ "loss": 0.3226,
256760
+ "step": 123705
256761
+ },
256762
+ {
256763
+ "epoch": 989.07,
256764
+ "learning_rate": 8.034278846153847e-06,
256765
+ "loss": 0.3369,
256766
+ "step": 123710
256767
+ },
256768
+ {
256769
+ "epoch": 989.11,
256770
+ "learning_rate": 8.034198717948719e-06,
256771
+ "loss": 0.2795,
256772
+ "step": 123715
256773
+ },
256774
+ {
256775
+ "epoch": 989.15,
256776
+ "learning_rate": 8.03411858974359e-06,
256777
+ "loss": 0.407,
256778
+ "step": 123720
256779
+ },
256780
+ {
256781
+ "epoch": 989.19,
256782
+ "learning_rate": 8.034038461538463e-06,
256783
+ "loss": 0.9145,
256784
+ "step": 123725
256785
+ },
256786
+ {
256787
+ "epoch": 989.23,
256788
+ "learning_rate": 8.033958333333335e-06,
256789
+ "loss": 0.6577,
256790
+ "step": 123730
256791
+ },
256792
+ {
256793
+ "epoch": 989.27,
256794
+ "learning_rate": 8.033878205128206e-06,
256795
+ "loss": 0.2938,
256796
+ "step": 123735
256797
+ },
256798
+ {
256799
+ "epoch": 989.31,
256800
+ "learning_rate": 8.033798076923077e-06,
256801
+ "loss": 0.3101,
256802
+ "step": 123740
256803
+ },
256804
+ {
256805
+ "epoch": 989.35,
256806
+ "learning_rate": 8.03371794871795e-06,
256807
+ "loss": 0.3811,
256808
+ "step": 123745
256809
+ },
256810
+ {
256811
+ "epoch": 989.39,
256812
+ "learning_rate": 8.03363782051282e-06,
256813
+ "loss": 0.8656,
256814
+ "step": 123750
256815
+ },
256816
+ {
256817
+ "epoch": 989.43,
256818
+ "learning_rate": 8.033557692307693e-06,
256819
+ "loss": 0.5815,
256820
+ "step": 123755
256821
+ },
256822
+ {
256823
+ "epoch": 989.47,
256824
+ "learning_rate": 8.033477564102566e-06,
256825
+ "loss": 0.2756,
256826
+ "step": 123760
256827
+ },
256828
+ {
256829
+ "epoch": 989.51,
256830
+ "learning_rate": 8.033397435897436e-06,
256831
+ "loss": 0.3436,
256832
+ "step": 123765
256833
+ },
256834
+ {
256835
+ "epoch": 989.55,
256836
+ "learning_rate": 8.033317307692309e-06,
256837
+ "loss": 0.444,
256838
+ "step": 123770
256839
+ },
256840
+ {
256841
+ "epoch": 989.59,
256842
+ "learning_rate": 8.03323717948718e-06,
256843
+ "loss": 0.8274,
256844
+ "step": 123775
256845
+ },
256846
+ {
256847
+ "epoch": 989.63,
256848
+ "learning_rate": 8.033157051282052e-06,
256849
+ "loss": 0.6641,
256850
+ "step": 123780
256851
+ },
256852
+ {
256853
+ "epoch": 989.67,
256854
+ "learning_rate": 8.033076923076923e-06,
256855
+ "loss": 0.2923,
256856
+ "step": 123785
256857
+ },
256858
+ {
256859
+ "epoch": 989.71,
256860
+ "learning_rate": 8.032996794871796e-06,
256861
+ "loss": 0.3045,
256862
+ "step": 123790
256863
+ },
256864
+ {
256865
+ "epoch": 989.75,
256866
+ "learning_rate": 8.032916666666667e-06,
256867
+ "loss": 0.5925,
256868
+ "step": 123795
256869
+ },
256870
+ {
256871
+ "epoch": 989.79,
256872
+ "learning_rate": 8.032836538461539e-06,
256873
+ "loss": 0.9411,
256874
+ "step": 123800
256875
+ },
256876
+ {
256877
+ "epoch": 989.83,
256878
+ "learning_rate": 8.03275641025641e-06,
256879
+ "loss": 0.6954,
256880
+ "step": 123805
256881
+ },
256882
+ {
256883
+ "epoch": 989.87,
256884
+ "learning_rate": 8.032676282051283e-06,
256885
+ "loss": 0.2363,
256886
+ "step": 123810
256887
+ },
256888
+ {
256889
+ "epoch": 989.91,
256890
+ "learning_rate": 8.032596153846154e-06,
256891
+ "loss": 0.3087,
256892
+ "step": 123815
256893
+ },
256894
+ {
256895
+ "epoch": 989.95,
256896
+ "learning_rate": 8.032516025641026e-06,
256897
+ "loss": 0.4986,
256898
+ "step": 123820
256899
+ },
256900
+ {
256901
+ "epoch": 989.99,
256902
+ "learning_rate": 8.032435897435899e-06,
256903
+ "loss": 0.9063,
256904
+ "step": 123825
256905
+ },
256906
+ {
256907
+ "epoch": 990.0,
256908
+ "eval_loss": 0.37780308723449707,
256909
+ "eval_runtime": 36.1129,
256910
+ "eval_samples_per_second": 23.316,
256911
+ "eval_steps_per_second": 0.748,
256912
+ "eval_wer": 0.18303048065650646,
256913
+ "step": 123826
256914
+ },
256915
+ {
256916
+ "epoch": 990.03,
256917
+ "learning_rate": 8.03235576923077e-06,
256918
+ "loss": 0.3242,
256919
+ "step": 123830
256920
+ },
256921
+ {
256922
+ "epoch": 990.07,
256923
+ "learning_rate": 8.032275641025642e-06,
256924
+ "loss": 0.2932,
256925
+ "step": 123835
256926
+ },
256927
+ {
256928
+ "epoch": 990.11,
256929
+ "learning_rate": 8.032195512820513e-06,
256930
+ "loss": 0.334,
256931
+ "step": 123840
256932
+ },
256933
+ {
256934
+ "epoch": 990.15,
256935
+ "learning_rate": 8.032115384615386e-06,
256936
+ "loss": 0.3641,
256937
+ "step": 123845
256938
+ },
256939
+ {
256940
+ "epoch": 990.19,
256941
+ "learning_rate": 8.032035256410257e-06,
256942
+ "loss": 0.9662,
256943
+ "step": 123850
256944
+ },
256945
+ {
256946
+ "epoch": 990.23,
256947
+ "learning_rate": 8.031955128205129e-06,
256948
+ "loss": 0.7018,
256949
+ "step": 123855
256950
+ },
256951
+ {
256952
+ "epoch": 990.27,
256953
+ "learning_rate": 8.031875000000002e-06,
256954
+ "loss": 0.2948,
256955
+ "step": 123860
256956
+ },
256957
+ {
256958
+ "epoch": 990.31,
256959
+ "learning_rate": 8.031794871794873e-06,
256960
+ "loss": 0.3453,
256961
+ "step": 123865
256962
+ },
256963
+ {
256964
+ "epoch": 990.35,
256965
+ "learning_rate": 8.031714743589744e-06,
256966
+ "loss": 0.439,
256967
+ "step": 123870
256968
+ },
256969
+ {
256970
+ "epoch": 990.39,
256971
+ "learning_rate": 8.031634615384616e-06,
256972
+ "loss": 1.012,
256973
+ "step": 123875
256974
+ },
256975
+ {
256976
+ "epoch": 990.43,
256977
+ "learning_rate": 8.031554487179489e-06,
256978
+ "loss": 0.6446,
256979
+ "step": 123880
256980
+ },
256981
+ {
256982
+ "epoch": 990.47,
256983
+ "learning_rate": 8.031474358974359e-06,
256984
+ "loss": 0.2602,
256985
+ "step": 123885
256986
+ },
256987
+ {
256988
+ "epoch": 990.51,
256989
+ "learning_rate": 8.031394230769232e-06,
256990
+ "loss": 0.3188,
256991
+ "step": 123890
256992
+ },
256993
+ {
256994
+ "epoch": 990.55,
256995
+ "learning_rate": 8.031314102564103e-06,
256996
+ "loss": 0.4144,
256997
+ "step": 123895
256998
+ },
256999
+ {
257000
+ "epoch": 990.59,
257001
+ "learning_rate": 8.031233974358974e-06,
257002
+ "loss": 0.7588,
257003
+ "step": 123900
257004
+ },
257005
+ {
257006
+ "epoch": 990.63,
257007
+ "learning_rate": 8.031153846153846e-06,
257008
+ "loss": 0.7133,
257009
+ "step": 123905
257010
+ },
257011
+ {
257012
+ "epoch": 990.67,
257013
+ "learning_rate": 8.031073717948719e-06,
257014
+ "loss": 0.2741,
257015
+ "step": 123910
257016
+ },
257017
+ {
257018
+ "epoch": 990.71,
257019
+ "learning_rate": 8.03099358974359e-06,
257020
+ "loss": 0.385,
257021
+ "step": 123915
257022
+ },
257023
+ {
257024
+ "epoch": 990.75,
257025
+ "learning_rate": 8.030913461538461e-06,
257026
+ "loss": 0.4035,
257027
+ "step": 123920
257028
+ },
257029
+ {
257030
+ "epoch": 990.79,
257031
+ "learning_rate": 8.030833333333335e-06,
257032
+ "loss": 0.9382,
257033
+ "step": 123925
257034
+ },
257035
+ {
257036
+ "epoch": 990.83,
257037
+ "learning_rate": 8.030753205128206e-06,
257038
+ "loss": 0.6219,
257039
+ "step": 123930
257040
+ },
257041
+ {
257042
+ "epoch": 990.87,
257043
+ "learning_rate": 8.030673076923077e-06,
257044
+ "loss": 0.2894,
257045
+ "step": 123935
257046
+ },
257047
+ {
257048
+ "epoch": 990.91,
257049
+ "learning_rate": 8.030592948717949e-06,
257050
+ "loss": 0.3362,
257051
+ "step": 123940
257052
+ },
257053
+ {
257054
+ "epoch": 990.95,
257055
+ "learning_rate": 8.030512820512822e-06,
257056
+ "loss": 0.4499,
257057
+ "step": 123945
257058
+ },
257059
+ {
257060
+ "epoch": 990.99,
257061
+ "learning_rate": 8.030432692307693e-06,
257062
+ "loss": 0.9499,
257063
+ "step": 123950
257064
+ },
257065
+ {
257066
+ "epoch": 991.0,
257067
+ "eval_loss": 0.38879507780075073,
257068
+ "eval_runtime": 36.3541,
257069
+ "eval_samples_per_second": 23.134,
257070
+ "eval_steps_per_second": 0.743,
257071
+ "eval_wer": 0.17872121256714002,
257072
+ "step": 123951
257073
+ },
257074
+ {
257075
+ "epoch": 991.03,
257076
+ "learning_rate": 8.030352564102564e-06,
257077
+ "loss": 0.3227,
257078
+ "step": 123955
257079
+ },
257080
+ {
257081
+ "epoch": 991.07,
257082
+ "learning_rate": 8.030272435897437e-06,
257083
+ "loss": 0.275,
257084
+ "step": 123960
257085
+ },
257086
+ {
257087
+ "epoch": 991.11,
257088
+ "learning_rate": 8.030192307692309e-06,
257089
+ "loss": 0.3181,
257090
+ "step": 123965
257091
+ },
257092
+ {
257093
+ "epoch": 991.15,
257094
+ "learning_rate": 8.03011217948718e-06,
257095
+ "loss": 0.3956,
257096
+ "step": 123970
257097
+ },
257098
+ {
257099
+ "epoch": 991.19,
257100
+ "learning_rate": 8.030032051282051e-06,
257101
+ "loss": 0.9103,
257102
+ "step": 123975
257103
+ },
257104
+ {
257105
+ "epoch": 991.23,
257106
+ "learning_rate": 8.029951923076925e-06,
257107
+ "loss": 0.6729,
257108
+ "step": 123980
257109
+ },
257110
+ {
257111
+ "epoch": 991.27,
257112
+ "learning_rate": 8.029871794871796e-06,
257113
+ "loss": 0.2789,
257114
+ "step": 123985
257115
+ },
257116
+ {
257117
+ "epoch": 991.31,
257118
+ "learning_rate": 8.029791666666667e-06,
257119
+ "loss": 0.3114,
257120
+ "step": 123990
257121
+ },
257122
+ {
257123
+ "epoch": 991.35,
257124
+ "learning_rate": 8.029711538461539e-06,
257125
+ "loss": 0.3959,
257126
+ "step": 123995
257127
+ },
257128
+ {
257129
+ "epoch": 991.39,
257130
+ "learning_rate": 8.029631410256412e-06,
257131
+ "loss": 0.9762,
257132
+ "step": 124000
257133
+ },
257134
+ {
257135
+ "epoch": 991.43,
257136
+ "learning_rate": 8.029551282051283e-06,
257137
+ "loss": 0.6035,
257138
+ "step": 124005
257139
+ },
257140
+ {
257141
+ "epoch": 991.47,
257142
+ "learning_rate": 8.029471153846154e-06,
257143
+ "loss": 0.3083,
257144
+ "step": 124010
257145
+ },
257146
+ {
257147
+ "epoch": 991.51,
257148
+ "learning_rate": 8.029391025641027e-06,
257149
+ "loss": 0.3193,
257150
+ "step": 124015
257151
+ },
257152
+ {
257153
+ "epoch": 991.55,
257154
+ "learning_rate": 8.029310897435899e-06,
257155
+ "loss": 0.3466,
257156
+ "step": 124020
257157
+ },
257158
+ {
257159
+ "epoch": 991.59,
257160
+ "learning_rate": 8.02923076923077e-06,
257161
+ "loss": 0.9612,
257162
+ "step": 124025
257163
+ },
257164
+ {
257165
+ "epoch": 991.63,
257166
+ "learning_rate": 8.029150641025642e-06,
257167
+ "loss": 0.6535,
257168
+ "step": 124030
257169
+ },
257170
+ {
257171
+ "epoch": 991.67,
257172
+ "learning_rate": 8.029070512820515e-06,
257173
+ "loss": 0.2786,
257174
+ "step": 124035
257175
+ },
257176
+ {
257177
+ "epoch": 991.71,
257178
+ "learning_rate": 8.028990384615384e-06,
257179
+ "loss": 0.2491,
257180
+ "step": 124040
257181
+ },
257182
+ {
257183
+ "epoch": 991.75,
257184
+ "learning_rate": 8.028910256410257e-06,
257185
+ "loss": 0.4216,
257186
+ "step": 124045
257187
+ },
257188
+ {
257189
+ "epoch": 991.79,
257190
+ "learning_rate": 8.028830128205129e-06,
257191
+ "loss": 0.9788,
257192
+ "step": 124050
257193
+ },
257194
+ {
257195
+ "epoch": 991.83,
257196
+ "learning_rate": 8.02875e-06,
257197
+ "loss": 0.5756,
257198
+ "step": 124055
257199
+ },
257200
+ {
257201
+ "epoch": 991.87,
257202
+ "learning_rate": 8.028669871794873e-06,
257203
+ "loss": 0.2539,
257204
+ "step": 124060
257205
+ },
257206
+ {
257207
+ "epoch": 991.91,
257208
+ "learning_rate": 8.028589743589744e-06,
257209
+ "loss": 0.3347,
257210
+ "step": 124065
257211
+ },
257212
+ {
257213
+ "epoch": 991.95,
257214
+ "learning_rate": 8.028509615384616e-06,
257215
+ "loss": 0.4351,
257216
+ "step": 124070
257217
+ },
257218
+ {
257219
+ "epoch": 991.99,
257220
+ "learning_rate": 8.028429487179487e-06,
257221
+ "loss": 0.8815,
257222
+ "step": 124075
257223
+ },
257224
+ {
257225
+ "epoch": 992.0,
257226
+ "eval_loss": 0.39232033491134644,
257227
+ "eval_runtime": 47.6402,
257228
+ "eval_samples_per_second": 17.653,
257229
+ "eval_steps_per_second": 0.567,
257230
+ "eval_wer": 0.18814771566194716,
257231
+ "step": 124076
257232
+ },
257233
+ {
257234
+ "epoch": 1000.03,
257235
+ "learning_rate": 8.02834935897436e-06,
257236
+ "loss": 0.4001,
257237
+ "step": 124080
257238
+ },
257239
+ {
257240
+ "epoch": 1000.07,
257241
+ "learning_rate": 8.028269230769232e-06,
257242
+ "loss": 0.2938,
257243
+ "step": 124085
257244
+ },
257245
+ {
257246
+ "epoch": 1000.11,
257247
+ "learning_rate": 8.028189102564103e-06,
257248
+ "loss": 0.3341,
257249
+ "step": 124090
257250
+ },
257251
+ {
257252
+ "epoch": 1000.15,
257253
+ "learning_rate": 8.028108974358974e-06,
257254
+ "loss": 0.3691,
257255
+ "step": 124095
257256
+ },
257257
+ {
257258
+ "epoch": 1000.19,
257259
+ "learning_rate": 8.028028846153847e-06,
257260
+ "loss": 0.7611,
257261
+ "step": 124100
257262
+ },
257263
+ {
257264
+ "epoch": 1000.23,
257265
+ "learning_rate": 8.027948717948719e-06,
257266
+ "loss": 0.8048,
257267
+ "step": 124105
257268
+ },
257269
+ {
257270
+ "epoch": 1000.27,
257271
+ "learning_rate": 8.02786858974359e-06,
257272
+ "loss": 0.2299,
257273
+ "step": 124110
257274
+ },
257275
+ {
257276
+ "epoch": 1000.31,
257277
+ "learning_rate": 8.027788461538463e-06,
257278
+ "loss": 0.2911,
257279
+ "step": 124115
257280
+ },
257281
+ {
257282
+ "epoch": 1000.35,
257283
+ "learning_rate": 8.027708333333334e-06,
257284
+ "loss": 0.367,
257285
+ "step": 124120
257286
+ },
257287
+ {
257288
+ "epoch": 1000.39,
257289
+ "learning_rate": 8.027628205128206e-06,
257290
+ "loss": 1.2227,
257291
+ "step": 124125
257292
+ },
257293
+ {
257294
+ "epoch": 1000.43,
257295
+ "learning_rate": 8.027548076923077e-06,
257296
+ "loss": 0.6543,
257297
+ "step": 124130
257298
+ },
257299
+ {
257300
+ "epoch": 1000.47,
257301
+ "learning_rate": 8.02746794871795e-06,
257302
+ "loss": 0.2603,
257303
+ "step": 124135
257304
+ },
257305
+ {
257306
+ "epoch": 1000.51,
257307
+ "learning_rate": 8.027387820512822e-06,
257308
+ "loss": 0.3135,
257309
+ "step": 124140
257310
+ },
257311
+ {
257312
+ "epoch": 1000.55,
257313
+ "learning_rate": 8.027307692307693e-06,
257314
+ "loss": 0.4096,
257315
+ "step": 124145
257316
+ },
257317
+ {
257318
+ "epoch": 1000.59,
257319
+ "learning_rate": 8.027227564102564e-06,
257320
+ "loss": 0.8186,
257321
+ "step": 124150
257322
+ },
257323
+ {
257324
+ "epoch": 1000.63,
257325
+ "learning_rate": 8.027147435897437e-06,
257326
+ "loss": 0.5708,
257327
+ "step": 124155
257328
+ },
257329
+ {
257330
+ "epoch": 1000.67,
257331
+ "learning_rate": 8.027067307692307e-06,
257332
+ "loss": 0.2524,
257333
+ "step": 124160
257334
+ },
257335
+ {
257336
+ "epoch": 1000.71,
257337
+ "learning_rate": 8.02698717948718e-06,
257338
+ "loss": 0.3512,
257339
+ "step": 124165
257340
+ },
257341
+ {
257342
+ "epoch": 1000.76,
257343
+ "learning_rate": 8.026907051282053e-06,
257344
+ "loss": 0.3847,
257345
+ "step": 124170
257346
+ },
257347
+ {
257348
+ "epoch": 1000.8,
257349
+ "learning_rate": 8.026826923076923e-06,
257350
+ "loss": 0.8454,
257351
+ "step": 124175
257352
+ },
257353
+ {
257354
+ "epoch": 1000.84,
257355
+ "learning_rate": 8.026746794871796e-06,
257356
+ "loss": 0.6479,
257357
+ "step": 124180
257358
+ },
257359
+ {
257360
+ "epoch": 1000.88,
257361
+ "learning_rate": 8.026666666666667e-06,
257362
+ "loss": 0.3131,
257363
+ "step": 124185
257364
+ },
257365
+ {
257366
+ "epoch": 1000.92,
257367
+ "learning_rate": 8.026586538461539e-06,
257368
+ "loss": 0.3086,
257369
+ "step": 124190
257370
+ },
257371
+ {
257372
+ "epoch": 1000.96,
257373
+ "learning_rate": 8.02650641025641e-06,
257374
+ "loss": 0.4573,
257375
+ "step": 124195
257376
+ },
257377
+ {
257378
+ "epoch": 1001.0,
257379
+ "learning_rate": 8.026426282051283e-06,
257380
+ "loss": 0.9595,
257381
+ "step": 124200
257382
+ },
257383
+ {
257384
+ "epoch": 1001.0,
257385
+ "eval_loss": 0.3651779294013977,
257386
+ "eval_runtime": 38.1351,
257387
+ "eval_samples_per_second": 22.027,
257388
+ "eval_steps_per_second": 0.708,
257389
+ "eval_wer": 0.17809599194070663,
257390
+ "step": 124200
257391
  }
257392
  ],
257393
+ "max_steps": 620000,
257394
  "num_train_epochs": 5000,
257395
+ "total_flos": 3.495167016128345e+20,
257396
  "trial_name": null,
257397
  "trial_params": null
257398
  }
model-bin/finetune/base/{checkpoint-123577 β†’ checkpoint-124200}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630140557.4831562/events.out.tfevents.1630140557.86bb0ddabf9b.4092.91 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d73248dcc2b7910acfc0a677dde4c151d7899c2286efd1aac34ee6d974a413b
3
+ size 4194
model-bin/finetune/base/log/1630140950.614884/events.out.tfevents.1630140950.86bb0ddabf9b.4092.93 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffa9ec9c1181f602da573c19d0a0493f728604b421e88c7131d0109bcb7b0ce4
3
+ size 4194
model-bin/finetune/base/log/1630141343.6640568/events.out.tfevents.1630141343.86bb0ddabf9b.4092.95 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f13ddb00d1f9568e3ae74524e018c20610803841c833f78e5775d7a5b7fc713
3
+ size 4194
model-bin/finetune/base/log/1630141737.0182173/events.out.tfevents.1630141737.86bb0ddabf9b.4092.97 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5147694ce0091023478759fcfb13569e248993596d086c758a5ee1343ca93f9
3
+ size 4194
model-bin/finetune/base/log/1630142244.9424796/events.out.tfevents.1630142244.86bb0ddabf9b.4092.99 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1e885954ef1e20bce9cfe8d13c1518ff5c67664c98f7fe4d14eaa71e4708ad6
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630140557.86bb0ddabf9b.4092.90 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7203831ac8ca7cc0616e15630080885e158dd0d69e7f87cd2a67ae3afed5d97c
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630140950.86bb0ddabf9b.4092.92 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e571256d6df9f671e427227f6da6533abb61847f2c28a568e74ca57fa2f34976
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630141343.86bb0ddabf9b.4092.94 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:889f2ddb22b58feb268eb865b8883ea8ed638ad0c711a484e4cda48112694e78
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630141737.86bb0ddabf9b.4092.96 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02726508ab3fb5220414baabc518871a21c605f9170023db3ba530c086f7f352
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630142244.86bb0ddabf9b.4092.98 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:849522741e53cdb2c7c0a09411191446657a52978759b04b9dac7701a37019fe
3
+ size 8622