Check commited on
Commit
3b30d05
Β·
1 Parent(s): aab15b1

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/rng_state.pth +2 -2
  6. model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/trainer_state.json +799 -4
  9. model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630130593.3816462/events.out.tfevents.1630130593.86bb0ddabf9b.4092.41 +3 -0
  11. model-bin/finetune/base/log/1630130973.7216606/events.out.tfevents.1630130973.86bb0ddabf9b.4092.43 +3 -0
  12. model-bin/finetune/base/log/1630131357.8881845/events.out.tfevents.1630131357.86bb0ddabf9b.4092.45 +3 -0
  13. model-bin/finetune/base/log/1630131736.953837/events.out.tfevents.1630131736.86bb0ddabf9b.4092.47 +3 -0
  14. model-bin/finetune/base/log/1630132125.1505942/events.out.tfevents.1630132125.86bb0ddabf9b.4092.49 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630130593.86bb0ddabf9b.4092.40 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630130973.86bb0ddabf9b.4092.42 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630131357.86bb0ddabf9b.4092.44 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630131736.86bb0ddabf9b.4092.46 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630132125.86bb0ddabf9b.4092.48 +3 -0
model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:02d754e3e35b5bd4dcad1290ee437bc10dda642925d72fb3e1f864b94c718255
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:092b700951279c90023bb11c835e0f5f54e15c88589d7fae37929377088a3adf
3
  size 722165393
model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b0255865cadd0645b960c336eea8a46b1f0f5c05aedb2a71da5a163adb3073fb
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7b657e06bef8e9094828e09219b103fa0568ba25f4c6a79a87c063cf5d83325
3
  size 377909911
model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aa7c7af68a6f1073c4850ff28a3ebe67c89df0455e85db0e2f01209cdcfbc270
3
- size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1c37a1cae1c9c87ffcb6d013494288d4405456b0beb4882c39271505234ed5f
3
+ size 14567
model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0fb121336be771fd94b174473516525f5106a0131cfb91565c4039ac35f3471c
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ebeea6250e483ecd40494e28df8405c19ff072bd9538b8863c97fc25c728f7f
3
  size 559
model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0154ab06f351cde3389d68a9fd4cd15b57e588da1a39ef8fca1b9e44cd8cf2ee
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3eefa8eb85bed9a407060ff02d546b477b03e6766a73c07eae1e5b9ca1444db3
3
  size 623
model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
- "epoch": 970.995983935743,
5
- "global_step": 120459,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -252624,11 +252624,806 @@
252624
  "eval_steps_per_second": 0.764,
252625
  "eval_wer": 0.18568594312587258,
252626
  "step": 120459
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
252627
  }
252628
  ],
252629
- "max_steps": 620000,
252630
  "num_train_epochs": 5000,
252631
- "total_flos": 3.389937279198161e+20,
252632
  "trial_name": null,
252633
  "trial_params": null
252634
  }
 
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
+ "epoch": 967.9960159362549,
5
+ "global_step": 121084,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
252624
  "eval_steps_per_second": 0.764,
252625
  "eval_wer": 0.18568594312587258,
252626
  "step": 120459
252627
+ },
252628
+ {
252629
+ "epoch": 963.01,
252630
+ "learning_rate": 8.08633012820513e-06,
252631
+ "loss": 0.3063,
252632
+ "step": 120460
252633
+ },
252634
+ {
252635
+ "epoch": 963.05,
252636
+ "learning_rate": 8.08625e-06,
252637
+ "loss": 0.3186,
252638
+ "step": 120465
252639
+ },
252640
+ {
252641
+ "epoch": 963.09,
252642
+ "learning_rate": 8.086169871794872e-06,
252643
+ "loss": 0.2936,
252644
+ "step": 120470
252645
+ },
252646
+ {
252647
+ "epoch": 963.13,
252648
+ "learning_rate": 8.086089743589745e-06,
252649
+ "loss": 0.3874,
252650
+ "step": 120475
252651
+ },
252652
+ {
252653
+ "epoch": 963.17,
252654
+ "learning_rate": 8.086009615384616e-06,
252655
+ "loss": 0.538,
252656
+ "step": 120480
252657
+ },
252658
+ {
252659
+ "epoch": 963.21,
252660
+ "learning_rate": 8.085929487179488e-06,
252661
+ "loss": 1.1838,
252662
+ "step": 120485
252663
+ },
252664
+ {
252665
+ "epoch": 963.25,
252666
+ "learning_rate": 8.08584935897436e-06,
252667
+ "loss": 0.2738,
252668
+ "step": 120490
252669
+ },
252670
+ {
252671
+ "epoch": 963.29,
252672
+ "learning_rate": 8.085769230769232e-06,
252673
+ "loss": 0.2775,
252674
+ "step": 120495
252675
+ },
252676
+ {
252677
+ "epoch": 963.33,
252678
+ "learning_rate": 8.085689102564104e-06,
252679
+ "loss": 0.3505,
252680
+ "step": 120500
252681
+ },
252682
+ {
252683
+ "epoch": 963.37,
252684
+ "learning_rate": 8.085608974358975e-06,
252685
+ "loss": 0.5326,
252686
+ "step": 120505
252687
+ },
252688
+ {
252689
+ "epoch": 963.41,
252690
+ "learning_rate": 8.085528846153848e-06,
252691
+ "loss": 1.2263,
252692
+ "step": 120510
252693
+ },
252694
+ {
252695
+ "epoch": 963.45,
252696
+ "learning_rate": 8.085448717948718e-06,
252697
+ "loss": 0.2965,
252698
+ "step": 120515
252699
+ },
252700
+ {
252701
+ "epoch": 963.49,
252702
+ "learning_rate": 8.08536858974359e-06,
252703
+ "loss": 0.275,
252704
+ "step": 120520
252705
+ },
252706
+ {
252707
+ "epoch": 963.53,
252708
+ "learning_rate": 8.085288461538462e-06,
252709
+ "loss": 0.5078,
252710
+ "step": 120525
252711
+ },
252712
+ {
252713
+ "epoch": 963.57,
252714
+ "learning_rate": 8.085208333333333e-06,
252715
+ "loss": 0.5536,
252716
+ "step": 120530
252717
+ },
252718
+ {
252719
+ "epoch": 963.61,
252720
+ "learning_rate": 8.085128205128205e-06,
252721
+ "loss": 1.1345,
252722
+ "step": 120535
252723
+ },
252724
+ {
252725
+ "epoch": 963.65,
252726
+ "learning_rate": 8.085048076923078e-06,
252727
+ "loss": 0.3,
252728
+ "step": 120540
252729
+ },
252730
+ {
252731
+ "epoch": 963.69,
252732
+ "learning_rate": 8.08496794871795e-06,
252733
+ "loss": 0.2699,
252734
+ "step": 120545
252735
+ },
252736
+ {
252737
+ "epoch": 963.73,
252738
+ "learning_rate": 8.08488782051282e-06,
252739
+ "loss": 0.3185,
252740
+ "step": 120550
252741
+ },
252742
+ {
252743
+ "epoch": 963.77,
252744
+ "learning_rate": 8.084807692307694e-06,
252745
+ "loss": 0.5594,
252746
+ "step": 120555
252747
+ },
252748
+ {
252749
+ "epoch": 963.81,
252750
+ "learning_rate": 8.084727564102565e-06,
252751
+ "loss": 1.1854,
252752
+ "step": 120560
252753
+ },
252754
+ {
252755
+ "epoch": 963.85,
252756
+ "learning_rate": 8.084647435897436e-06,
252757
+ "loss": 0.3462,
252758
+ "step": 120565
252759
+ },
252760
+ {
252761
+ "epoch": 963.89,
252762
+ "learning_rate": 8.084567307692308e-06,
252763
+ "loss": 0.2817,
252764
+ "step": 120570
252765
+ },
252766
+ {
252767
+ "epoch": 963.93,
252768
+ "learning_rate": 8.08448717948718e-06,
252769
+ "loss": 0.3001,
252770
+ "step": 120575
252771
+ },
252772
+ {
252773
+ "epoch": 963.97,
252774
+ "learning_rate": 8.084407051282052e-06,
252775
+ "loss": 0.5627,
252776
+ "step": 120580
252777
+ },
252778
+ {
252779
+ "epoch": 964.0,
252780
+ "eval_loss": 0.3748638331890106,
252781
+ "eval_runtime": 35.9106,
252782
+ "eval_samples_per_second": 23.447,
252783
+ "eval_steps_per_second": 0.752,
252784
+ "eval_wer": 0.18153778365278694,
252785
+ "step": 120584
252786
+ },
252787
+ {
252788
+ "epoch": 964.01,
252789
+ "learning_rate": 8.084326923076923e-06,
252790
+ "loss": 0.4757,
252791
+ "step": 120585
252792
+ },
252793
+ {
252794
+ "epoch": 964.05,
252795
+ "learning_rate": 8.084246794871795e-06,
252796
+ "loss": 0.2449,
252797
+ "step": 120590
252798
+ },
252799
+ {
252800
+ "epoch": 964.09,
252801
+ "learning_rate": 8.084166666666668e-06,
252802
+ "loss": 0.2717,
252803
+ "step": 120595
252804
+ },
252805
+ {
252806
+ "epoch": 964.13,
252807
+ "learning_rate": 8.08408653846154e-06,
252808
+ "loss": 0.3517,
252809
+ "step": 120600
252810
+ },
252811
+ {
252812
+ "epoch": 964.17,
252813
+ "learning_rate": 8.08400641025641e-06,
252814
+ "loss": 0.5087,
252815
+ "step": 120605
252816
+ },
252817
+ {
252818
+ "epoch": 964.21,
252819
+ "learning_rate": 8.083926282051284e-06,
252820
+ "loss": 1.0103,
252821
+ "step": 120610
252822
+ },
252823
+ {
252824
+ "epoch": 964.25,
252825
+ "learning_rate": 8.083846153846155e-06,
252826
+ "loss": 0.3475,
252827
+ "step": 120615
252828
+ },
252829
+ {
252830
+ "epoch": 964.29,
252831
+ "learning_rate": 8.083766025641026e-06,
252832
+ "loss": 0.2822,
252833
+ "step": 120620
252834
+ },
252835
+ {
252836
+ "epoch": 964.33,
252837
+ "learning_rate": 8.083685897435898e-06,
252838
+ "loss": 0.2866,
252839
+ "step": 120625
252840
+ },
252841
+ {
252842
+ "epoch": 964.37,
252843
+ "learning_rate": 8.08360576923077e-06,
252844
+ "loss": 0.4979,
252845
+ "step": 120630
252846
+ },
252847
+ {
252848
+ "epoch": 964.41,
252849
+ "learning_rate": 8.08352564102564e-06,
252850
+ "loss": 1.1757,
252851
+ "step": 120635
252852
+ },
252853
+ {
252854
+ "epoch": 964.45,
252855
+ "learning_rate": 8.083445512820514e-06,
252856
+ "loss": 0.3024,
252857
+ "step": 120640
252858
+ },
252859
+ {
252860
+ "epoch": 964.49,
252861
+ "learning_rate": 8.083365384615385e-06,
252862
+ "loss": 0.2628,
252863
+ "step": 120645
252864
+ },
252865
+ {
252866
+ "epoch": 964.53,
252867
+ "learning_rate": 8.083285256410256e-06,
252868
+ "loss": 0.3263,
252869
+ "step": 120650
252870
+ },
252871
+ {
252872
+ "epoch": 964.57,
252873
+ "learning_rate": 8.08320512820513e-06,
252874
+ "loss": 0.5977,
252875
+ "step": 120655
252876
+ },
252877
+ {
252878
+ "epoch": 964.61,
252879
+ "learning_rate": 8.083125e-06,
252880
+ "loss": 1.1081,
252881
+ "step": 120660
252882
+ },
252883
+ {
252884
+ "epoch": 964.65,
252885
+ "learning_rate": 8.083044871794874e-06,
252886
+ "loss": 0.2912,
252887
+ "step": 120665
252888
+ },
252889
+ {
252890
+ "epoch": 964.69,
252891
+ "learning_rate": 8.082964743589743e-06,
252892
+ "loss": 0.2956,
252893
+ "step": 120670
252894
+ },
252895
+ {
252896
+ "epoch": 964.73,
252897
+ "learning_rate": 8.082884615384616e-06,
252898
+ "loss": 0.255,
252899
+ "step": 120675
252900
+ },
252901
+ {
252902
+ "epoch": 964.77,
252903
+ "learning_rate": 8.082804487179488e-06,
252904
+ "loss": 0.4588,
252905
+ "step": 120680
252906
+ },
252907
+ {
252908
+ "epoch": 964.81,
252909
+ "learning_rate": 8.082724358974359e-06,
252910
+ "loss": 1.0802,
252911
+ "step": 120685
252912
+ },
252913
+ {
252914
+ "epoch": 964.85,
252915
+ "learning_rate": 8.08264423076923e-06,
252916
+ "loss": 0.2736,
252917
+ "step": 120690
252918
+ },
252919
+ {
252920
+ "epoch": 964.89,
252921
+ "learning_rate": 8.082564102564104e-06,
252922
+ "loss": 0.2829,
252923
+ "step": 120695
252924
+ },
252925
+ {
252926
+ "epoch": 964.93,
252927
+ "learning_rate": 8.082483974358975e-06,
252928
+ "loss": 0.3132,
252929
+ "step": 120700
252930
+ },
252931
+ {
252932
+ "epoch": 964.97,
252933
+ "learning_rate": 8.082403846153846e-06,
252934
+ "loss": 0.5069,
252935
+ "step": 120705
252936
+ },
252937
+ {
252938
+ "epoch": 965.0,
252939
+ "eval_loss": 0.3655751347541809,
252940
+ "eval_runtime": 37.0544,
252941
+ "eval_samples_per_second": 22.723,
252942
+ "eval_steps_per_second": 0.729,
252943
+ "eval_wer": 0.18698390482855143,
252944
+ "step": 120709
252945
+ },
252946
+ {
252947
+ "epoch": 965.01,
252948
+ "learning_rate": 8.08232371794872e-06,
252949
+ "loss": 0.4841,
252950
+ "step": 120710
252951
+ },
252952
+ {
252953
+ "epoch": 965.05,
252954
+ "learning_rate": 8.08224358974359e-06,
252955
+ "loss": 0.2581,
252956
+ "step": 120715
252957
+ },
252958
+ {
252959
+ "epoch": 965.09,
252960
+ "learning_rate": 8.082163461538462e-06,
252961
+ "loss": 0.269,
252962
+ "step": 120720
252963
+ },
252964
+ {
252965
+ "epoch": 965.13,
252966
+ "learning_rate": 8.082083333333333e-06,
252967
+ "loss": 0.3394,
252968
+ "step": 120725
252969
+ },
252970
+ {
252971
+ "epoch": 965.17,
252972
+ "learning_rate": 8.082003205128206e-06,
252973
+ "loss": 0.5367,
252974
+ "step": 120730
252975
+ },
252976
+ {
252977
+ "epoch": 965.21,
252978
+ "learning_rate": 8.081923076923078e-06,
252979
+ "loss": 1.1198,
252980
+ "step": 120735
252981
+ },
252982
+ {
252983
+ "epoch": 965.25,
252984
+ "learning_rate": 8.08184294871795e-06,
252985
+ "loss": 0.2783,
252986
+ "step": 120740
252987
+ },
252988
+ {
252989
+ "epoch": 965.29,
252990
+ "learning_rate": 8.08176282051282e-06,
252991
+ "loss": 0.2792,
252992
+ "step": 120745
252993
+ },
252994
+ {
252995
+ "epoch": 965.33,
252996
+ "learning_rate": 8.081682692307694e-06,
252997
+ "loss": 0.2745,
252998
+ "step": 120750
252999
+ },
253000
+ {
253001
+ "epoch": 965.37,
253002
+ "learning_rate": 8.081602564102565e-06,
253003
+ "loss": 0.5469,
253004
+ "step": 120755
253005
+ },
253006
+ {
253007
+ "epoch": 965.41,
253008
+ "learning_rate": 8.081522435897436e-06,
253009
+ "loss": 1.0953,
253010
+ "step": 120760
253011
+ },
253012
+ {
253013
+ "epoch": 965.45,
253014
+ "learning_rate": 8.08144230769231e-06,
253015
+ "loss": 0.285,
253016
+ "step": 120765
253017
+ },
253018
+ {
253019
+ "epoch": 965.49,
253020
+ "learning_rate": 8.08136217948718e-06,
253021
+ "loss": 0.2642,
253022
+ "step": 120770
253023
+ },
253024
+ {
253025
+ "epoch": 965.53,
253026
+ "learning_rate": 8.081282051282052e-06,
253027
+ "loss": 0.3305,
253028
+ "step": 120775
253029
+ },
253030
+ {
253031
+ "epoch": 965.57,
253032
+ "learning_rate": 8.081201923076923e-06,
253033
+ "loss": 0.5227,
253034
+ "step": 120780
253035
+ },
253036
+ {
253037
+ "epoch": 965.61,
253038
+ "learning_rate": 8.081121794871796e-06,
253039
+ "loss": 1.1465,
253040
+ "step": 120785
253041
+ },
253042
+ {
253043
+ "epoch": 965.65,
253044
+ "learning_rate": 8.081041666666666e-06,
253045
+ "loss": 0.2476,
253046
+ "step": 120790
253047
+ },
253048
+ {
253049
+ "epoch": 965.69,
253050
+ "learning_rate": 8.08096153846154e-06,
253051
+ "loss": 0.2343,
253052
+ "step": 120795
253053
+ },
253054
+ {
253055
+ "epoch": 965.73,
253056
+ "learning_rate": 8.08088141025641e-06,
253057
+ "loss": 0.3151,
253058
+ "step": 120800
253059
+ },
253060
+ {
253061
+ "epoch": 965.77,
253062
+ "learning_rate": 8.080801282051282e-06,
253063
+ "loss": 0.5144,
253064
+ "step": 120805
253065
+ },
253066
+ {
253067
+ "epoch": 965.81,
253068
+ "learning_rate": 8.080721153846155e-06,
253069
+ "loss": 1.0767,
253070
+ "step": 120810
253071
+ },
253072
+ {
253073
+ "epoch": 965.85,
253074
+ "learning_rate": 8.080641025641026e-06,
253075
+ "loss": 0.3412,
253076
+ "step": 120815
253077
+ },
253078
+ {
253079
+ "epoch": 965.89,
253080
+ "learning_rate": 8.080560897435898e-06,
253081
+ "loss": 0.3241,
253082
+ "step": 120820
253083
+ },
253084
+ {
253085
+ "epoch": 965.93,
253086
+ "learning_rate": 8.080480769230769e-06,
253087
+ "loss": 0.363,
253088
+ "step": 120825
253089
+ },
253090
+ {
253091
+ "epoch": 965.97,
253092
+ "learning_rate": 8.080400641025642e-06,
253093
+ "loss": 0.5905,
253094
+ "step": 120830
253095
+ },
253096
+ {
253097
+ "epoch": 966.0,
253098
+ "eval_loss": 0.38227519392967224,
253099
+ "eval_runtime": 35.5959,
253100
+ "eval_samples_per_second": 23.654,
253101
+ "eval_steps_per_second": 0.759,
253102
+ "eval_wer": 0.18383645955451347,
253103
+ "step": 120834
253104
+ },
253105
+ {
253106
+ "epoch": 966.01,
253107
+ "learning_rate": 8.080320512820513e-06,
253108
+ "loss": 0.3745,
253109
+ "step": 120835
253110
+ },
253111
+ {
253112
+ "epoch": 966.05,
253113
+ "learning_rate": 8.080240384615385e-06,
253114
+ "loss": 0.2932,
253115
+ "step": 120840
253116
+ },
253117
+ {
253118
+ "epoch": 966.09,
253119
+ "learning_rate": 8.080160256410256e-06,
253120
+ "loss": 0.3262,
253121
+ "step": 120845
253122
+ },
253123
+ {
253124
+ "epoch": 966.13,
253125
+ "learning_rate": 8.08008012820513e-06,
253126
+ "loss": 0.3332,
253127
+ "step": 120850
253128
+ },
253129
+ {
253130
+ "epoch": 966.17,
253131
+ "learning_rate": 8.08e-06,
253132
+ "loss": 0.5078,
253133
+ "step": 120855
253134
+ },
253135
+ {
253136
+ "epoch": 966.21,
253137
+ "learning_rate": 8.079919871794872e-06,
253138
+ "loss": 1.1597,
253139
+ "step": 120860
253140
+ },
253141
+ {
253142
+ "epoch": 966.25,
253143
+ "learning_rate": 8.079839743589745e-06,
253144
+ "loss": 0.2896,
253145
+ "step": 120865
253146
+ },
253147
+ {
253148
+ "epoch": 966.29,
253149
+ "learning_rate": 8.079759615384616e-06,
253150
+ "loss": 0.2589,
253151
+ "step": 120870
253152
+ },
253153
+ {
253154
+ "epoch": 966.33,
253155
+ "learning_rate": 8.079679487179488e-06,
253156
+ "loss": 0.3202,
253157
+ "step": 120875
253158
+ },
253159
+ {
253160
+ "epoch": 966.37,
253161
+ "learning_rate": 8.079599358974359e-06,
253162
+ "loss": 0.4251,
253163
+ "step": 120880
253164
+ },
253165
+ {
253166
+ "epoch": 966.41,
253167
+ "learning_rate": 8.079519230769232e-06,
253168
+ "loss": 1.1363,
253169
+ "step": 120885
253170
+ },
253171
+ {
253172
+ "epoch": 966.45,
253173
+ "learning_rate": 8.079439102564103e-06,
253174
+ "loss": 0.3064,
253175
+ "step": 120890
253176
+ },
253177
+ {
253178
+ "epoch": 966.49,
253179
+ "learning_rate": 8.079358974358975e-06,
253180
+ "loss": 0.2289,
253181
+ "step": 120895
253182
+ },
253183
+ {
253184
+ "epoch": 966.53,
253185
+ "learning_rate": 8.079278846153846e-06,
253186
+ "loss": 0.3267,
253187
+ "step": 120900
253188
+ },
253189
+ {
253190
+ "epoch": 966.57,
253191
+ "learning_rate": 8.07919871794872e-06,
253192
+ "loss": 0.5204,
253193
+ "step": 120905
253194
+ },
253195
+ {
253196
+ "epoch": 966.61,
253197
+ "learning_rate": 8.07911858974359e-06,
253198
+ "loss": 1.2683,
253199
+ "step": 120910
253200
+ },
253201
+ {
253202
+ "epoch": 966.65,
253203
+ "learning_rate": 8.079038461538462e-06,
253204
+ "loss": 0.2852,
253205
+ "step": 120915
253206
+ },
253207
+ {
253208
+ "epoch": 966.69,
253209
+ "learning_rate": 8.078958333333335e-06,
253210
+ "loss": 0.2778,
253211
+ "step": 120920
253212
+ },
253213
+ {
253214
+ "epoch": 966.73,
253215
+ "learning_rate": 8.078878205128206e-06,
253216
+ "loss": 0.3523,
253217
+ "step": 120925
253218
+ },
253219
+ {
253220
+ "epoch": 966.77,
253221
+ "learning_rate": 8.078798076923078e-06,
253222
+ "loss": 0.5476,
253223
+ "step": 120930
253224
+ },
253225
+ {
253226
+ "epoch": 966.81,
253227
+ "learning_rate": 8.078717948717949e-06,
253228
+ "loss": 0.9729,
253229
+ "step": 120935
253230
+ },
253231
+ {
253232
+ "epoch": 966.85,
253233
+ "learning_rate": 8.078637820512822e-06,
253234
+ "loss": 0.5187,
253235
+ "step": 120940
253236
+ },
253237
+ {
253238
+ "epoch": 966.89,
253239
+ "learning_rate": 8.078557692307692e-06,
253240
+ "loss": 0.3767,
253241
+ "step": 120945
253242
+ },
253243
+ {
253244
+ "epoch": 966.93,
253245
+ "learning_rate": 8.078477564102565e-06,
253246
+ "loss": 0.3119,
253247
+ "step": 120950
253248
+ },
253249
+ {
253250
+ "epoch": 966.97,
253251
+ "learning_rate": 8.078397435897438e-06,
253252
+ "loss": 0.5822,
253253
+ "step": 120955
253254
+ },
253255
+ {
253256
+ "epoch": 967.0,
253257
+ "eval_loss": 0.39696982502937317,
253258
+ "eval_runtime": 35.5761,
253259
+ "eval_samples_per_second": 23.639,
253260
+ "eval_steps_per_second": 0.759,
253261
+ "eval_wer": 0.17967772790817452,
253262
+ "step": 120959
253263
+ },
253264
+ {
253265
+ "epoch": 967.01,
253266
+ "learning_rate": 8.078317307692308e-06,
253267
+ "loss": 0.3628,
253268
+ "step": 120960
253269
+ },
253270
+ {
253271
+ "epoch": 967.05,
253272
+ "learning_rate": 8.07823717948718e-06,
253273
+ "loss": 0.2908,
253274
+ "step": 120965
253275
+ },
253276
+ {
253277
+ "epoch": 967.09,
253278
+ "learning_rate": 8.078157051282052e-06,
253279
+ "loss": 0.2979,
253280
+ "step": 120970
253281
+ },
253282
+ {
253283
+ "epoch": 967.13,
253284
+ "learning_rate": 8.078076923076923e-06,
253285
+ "loss": 0.3033,
253286
+ "step": 120975
253287
+ },
253288
+ {
253289
+ "epoch": 967.17,
253290
+ "learning_rate": 8.077996794871795e-06,
253291
+ "loss": 0.6087,
253292
+ "step": 120980
253293
+ },
253294
+ {
253295
+ "epoch": 967.21,
253296
+ "learning_rate": 8.077916666666668e-06,
253297
+ "loss": 1.1057,
253298
+ "step": 120985
253299
+ },
253300
+ {
253301
+ "epoch": 967.25,
253302
+ "learning_rate": 8.077836538461539e-06,
253303
+ "loss": 0.2821,
253304
+ "step": 120990
253305
+ },
253306
+ {
253307
+ "epoch": 967.29,
253308
+ "learning_rate": 8.07775641025641e-06,
253309
+ "loss": 0.2461,
253310
+ "step": 120995
253311
+ },
253312
+ {
253313
+ "epoch": 967.33,
253314
+ "learning_rate": 8.077676282051282e-06,
253315
+ "loss": 0.5646,
253316
+ "step": 121000
253317
+ },
253318
+ {
253319
+ "epoch": 967.37,
253320
+ "learning_rate": 8.077596153846155e-06,
253321
+ "loss": 0.563,
253322
+ "step": 121005
253323
+ },
253324
+ {
253325
+ "epoch": 967.41,
253326
+ "learning_rate": 8.077516025641026e-06,
253327
+ "loss": 0.9909,
253328
+ "step": 121010
253329
+ },
253330
+ {
253331
+ "epoch": 967.45,
253332
+ "learning_rate": 8.077435897435898e-06,
253333
+ "loss": 0.2902,
253334
+ "step": 121015
253335
+ },
253336
+ {
253337
+ "epoch": 967.49,
253338
+ "learning_rate": 8.07735576923077e-06,
253339
+ "loss": 0.3619,
253340
+ "step": 121020
253341
+ },
253342
+ {
253343
+ "epoch": 967.53,
253344
+ "learning_rate": 8.077275641025642e-06,
253345
+ "loss": 0.3563,
253346
+ "step": 121025
253347
+ },
253348
+ {
253349
+ "epoch": 967.57,
253350
+ "learning_rate": 8.077195512820513e-06,
253351
+ "loss": 0.5795,
253352
+ "step": 121030
253353
+ },
253354
+ {
253355
+ "epoch": 967.61,
253356
+ "learning_rate": 8.077115384615385e-06,
253357
+ "loss": 1.1314,
253358
+ "step": 121035
253359
+ },
253360
+ {
253361
+ "epoch": 967.65,
253362
+ "learning_rate": 8.077035256410258e-06,
253363
+ "loss": 0.3236,
253364
+ "step": 121040
253365
+ },
253366
+ {
253367
+ "epoch": 967.69,
253368
+ "learning_rate": 8.076955128205129e-06,
253369
+ "loss": 0.2958,
253370
+ "step": 121045
253371
+ },
253372
+ {
253373
+ "epoch": 967.73,
253374
+ "learning_rate": 8.076875e-06,
253375
+ "loss": 0.2668,
253376
+ "step": 121050
253377
+ },
253378
+ {
253379
+ "epoch": 967.76,
253380
+ "learning_rate": 8.076794871794874e-06,
253381
+ "loss": 0.505,
253382
+ "step": 121055
253383
+ },
253384
+ {
253385
+ "epoch": 967.8,
253386
+ "learning_rate": 8.076714743589745e-06,
253387
+ "loss": 1.1167,
253388
+ "step": 121060
253389
+ },
253390
+ {
253391
+ "epoch": 967.84,
253392
+ "learning_rate": 8.076634615384616e-06,
253393
+ "loss": 0.2652,
253394
+ "step": 121065
253395
+ },
253396
+ {
253397
+ "epoch": 967.88,
253398
+ "learning_rate": 8.076554487179488e-06,
253399
+ "loss": 0.2603,
253400
+ "step": 121070
253401
+ },
253402
+ {
253403
+ "epoch": 967.92,
253404
+ "learning_rate": 8.07647435897436e-06,
253405
+ "loss": 0.3676,
253406
+ "step": 121075
253407
+ },
253408
+ {
253409
+ "epoch": 967.96,
253410
+ "learning_rate": 8.07639423076923e-06,
253411
+ "loss": 0.5594,
253412
+ "step": 121080
253413
+ },
253414
+ {
253415
+ "epoch": 968.0,
253416
+ "eval_loss": 0.4073833227157593,
253417
+ "eval_runtime": 35.3932,
253418
+ "eval_samples_per_second": 23.762,
253419
+ "eval_steps_per_second": 0.763,
253420
+ "eval_wer": 0.18734491315136476,
253421
+ "step": 121084
253422
  }
253423
  ],
253424
+ "max_steps": 625000,
253425
  "num_train_epochs": 5000,
253426
+ "total_flos": 3.407525538155879e+20,
253427
  "trial_name": null,
253428
  "trial_params": null
253429
  }
model-bin/finetune/base/{checkpoint-120459 β†’ checkpoint-121084}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630130593.3816462/events.out.tfevents.1630130593.86bb0ddabf9b.4092.41 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc0e8a00a2ef51be7fb8787b9cb0133f9e18b8ad343f39d8f2c93b28efaac501
3
+ size 4194
model-bin/finetune/base/log/1630130973.7216606/events.out.tfevents.1630130973.86bb0ddabf9b.4092.43 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eeca7e6fd5d53b7cdd7d5863d9dbb703f942bec14f7ee8859b7efcd89baef2a5
3
+ size 4194
model-bin/finetune/base/log/1630131357.8881845/events.out.tfevents.1630131357.86bb0ddabf9b.4092.45 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1282b30922d1aa33658b6b050f893881ddbbc23b21a48855cc34fb6341277c3
3
+ size 4194
model-bin/finetune/base/log/1630131736.953837/events.out.tfevents.1630131736.86bb0ddabf9b.4092.47 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8aba7cc3ad6a97807502e9d8fc7b0c7c70e15f4979a3ebc86f63d131e1e13e6
3
+ size 4194
model-bin/finetune/base/log/1630132125.1505942/events.out.tfevents.1630132125.86bb0ddabf9b.4092.49 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ffc4fb9d8b30d998c5540f0504c9aef5a93a60eaf845b88aa54b819568ab5e3
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630130593.86bb0ddabf9b.4092.40 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a2d0088ebc8ac48a43f684f224120bf5b960fc7ef754c5c8b2abd2d45fdca2c
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630130973.86bb0ddabf9b.4092.42 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:475f78cc7c7aa32efa9485e9be0866cc8714e715d64fccdddea1d93322a47379
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630131357.86bb0ddabf9b.4092.44 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d7e8e554fbf5a56b03696543205a58d3a7beefced2144c10c09b66c6e8d8f5c
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630131736.86bb0ddabf9b.4092.46 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37ea119a2e37164b37951258112f60443eabd98b0f8f91628b794e2ccb6c7056
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630132125.86bb0ddabf9b.4092.48 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49550e53dd6617feb12ed6a891c0dd77d7f2436ba592cf395b0b41130e854f9e
3
+ size 8622