Check commited on
Commit
52ffbad
Β·
1 Parent(s): e4a37f8

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/rng_state.pth +2 -2
  6. model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/trainer_state.json +799 -4
  9. model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630241766.1674721/events.out.tfevents.1630241766.cc93b136ebf5.1086.169 +3 -0
  11. model-bin/finetune/base/log/1630242210.9386408/events.out.tfevents.1630242210.cc93b136ebf5.1086.171 +3 -0
  12. model-bin/finetune/base/log/1630242652.9526577/events.out.tfevents.1630242652.cc93b136ebf5.1086.173 +3 -0
  13. model-bin/finetune/base/log/1630243093.5265884/events.out.tfevents.1630243093.cc93b136ebf5.1086.175 +3 -0
  14. model-bin/finetune/base/log/1630243533.7269483/events.out.tfevents.1630243533.cc93b136ebf5.1086.177 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630241766.cc93b136ebf5.1086.168 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630242210.cc93b136ebf5.1086.170 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630242652.cc93b136ebf5.1086.172 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630243093.cc93b136ebf5.1086.174 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630243533.cc93b136ebf5.1086.176 +3 -0
model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:207277e446e0ca60dd83abf90aab7a2af3700e64df25f7d5a5d03195c838cc31
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5341661b50f4e596b6fe02c28dfc7ba32504ab70fca5178ea2952aac4a6cbdf1
3
  size 722165393
model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:04af34e30722560d517d6ae868955c6953393b4b21776c8eb7ccd0753bf1f2ea
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2831ef8aa83687ecb9c8fa7338b9766386e38377b04dd35c04a8c8c06f8148f2
3
  size 377909911
model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:953f51e4a6b744d9cf78d43abb72507dea35603620d29a905184d67154569c48
3
- size 14567
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0626ef614583f5ab57c8ac6ae463c7a0f18875e587996d7e4c49f829145e7e88
3
+ size 14503
model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48a37d54355b773f8b00614fabd924720d9d3f3bbbb3a062ffefc6c699f6469c
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3540207233bcabcb86e1d4ad4f38a1c87a6c2cdf020124389fd687b8b31fe0dd
3
  size 559
model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0dc231f0815a529224131eede510f2dcb3545c8b449267f729dc8e7d82e5f194
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63f3f86c60dbd72222ca372243fb135bc1a49b12c2bc13b3077d16a8db5290be
3
  size 623
model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1668011442822563,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
4
- "epoch": 1196.0,
5
- "global_step": 149584,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -289680,11 +289680,806 @@
289680
  "eval_steps_per_second": 0.654,
289681
  "eval_wer": 0.17671063178605262,
289682
  "step": 149584
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
289683
  }
289684
  ],
289685
- "max_steps": 625000,
289686
  "num_train_epochs": 5000,
289687
- "total_flos": 4.209383244536534e+20,
289688
  "trial_name": null,
289689
  "trial_params": null
289690
  }
 
1
  {
2
  "best_metric": 0.1668011442822563,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
4
+ "epoch": 1211.0,
5
+ "global_step": 150205,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
289680
  "eval_steps_per_second": 0.654,
289681
  "eval_wer": 0.17671063178605262,
289682
  "step": 149584
289683
+ },
289684
+ {
289685
+ "epoch": 1206.01,
289686
+ "learning_rate": 7.600549273021002e-06,
289687
+ "loss": 0.4851,
289688
+ "step": 149585
289689
+ },
289690
+ {
289691
+ "epoch": 1206.05,
289692
+ "learning_rate": 7.600468497576737e-06,
289693
+ "loss": 0.279,
289694
+ "step": 149590
289695
+ },
289696
+ {
289697
+ "epoch": 1206.09,
289698
+ "learning_rate": 7.600387722132473e-06,
289699
+ "loss": 0.3121,
289700
+ "step": 149595
289701
+ },
289702
+ {
289703
+ "epoch": 1206.13,
289704
+ "learning_rate": 7.600306946688207e-06,
289705
+ "loss": 0.3105,
289706
+ "step": 149600
289707
+ },
289708
+ {
289709
+ "epoch": 1206.17,
289710
+ "learning_rate": 7.600226171243943e-06,
289711
+ "loss": 0.5346,
289712
+ "step": 149605
289713
+ },
289714
+ {
289715
+ "epoch": 1206.21,
289716
+ "learning_rate": 7.600145395799678e-06,
289717
+ "loss": 1.0911,
289718
+ "step": 149610
289719
+ },
289720
+ {
289721
+ "epoch": 1206.25,
289722
+ "learning_rate": 7.600064620355413e-06,
289723
+ "loss": 0.2746,
289724
+ "step": 149615
289725
+ },
289726
+ {
289727
+ "epoch": 1206.29,
289728
+ "learning_rate": 7.599983844911148e-06,
289729
+ "loss": 0.3218,
289730
+ "step": 149620
289731
+ },
289732
+ {
289733
+ "epoch": 1206.33,
289734
+ "learning_rate": 7.599903069466883e-06,
289735
+ "loss": 0.3109,
289736
+ "step": 149625
289737
+ },
289738
+ {
289739
+ "epoch": 1206.37,
289740
+ "learning_rate": 7.599822294022618e-06,
289741
+ "loss": 0.4662,
289742
+ "step": 149630
289743
+ },
289744
+ {
289745
+ "epoch": 1206.41,
289746
+ "learning_rate": 7.599741518578353e-06,
289747
+ "loss": 1.0338,
289748
+ "step": 149635
289749
+ },
289750
+ {
289751
+ "epoch": 1206.45,
289752
+ "learning_rate": 7.599660743134088e-06,
289753
+ "loss": 0.2714,
289754
+ "step": 149640
289755
+ },
289756
+ {
289757
+ "epoch": 1206.49,
289758
+ "learning_rate": 7.599579967689823e-06,
289759
+ "loss": 0.2534,
289760
+ "step": 149645
289761
+ },
289762
+ {
289763
+ "epoch": 1206.53,
289764
+ "learning_rate": 7.599499192245558e-06,
289765
+ "loss": 0.3401,
289766
+ "step": 149650
289767
+ },
289768
+ {
289769
+ "epoch": 1206.57,
289770
+ "learning_rate": 7.599418416801293e-06,
289771
+ "loss": 0.5341,
289772
+ "step": 149655
289773
+ },
289774
+ {
289775
+ "epoch": 1206.61,
289776
+ "learning_rate": 7.5993376413570285e-06,
289777
+ "loss": 1.1729,
289778
+ "step": 149660
289779
+ },
289780
+ {
289781
+ "epoch": 1206.65,
289782
+ "learning_rate": 7.599256865912763e-06,
289783
+ "loss": 0.2943,
289784
+ "step": 149665
289785
+ },
289786
+ {
289787
+ "epoch": 1206.69,
289788
+ "learning_rate": 7.5991760904684985e-06,
289789
+ "loss": 0.3001,
289790
+ "step": 149670
289791
+ },
289792
+ {
289793
+ "epoch": 1206.73,
289794
+ "learning_rate": 7.599095315024233e-06,
289795
+ "loss": 0.3681,
289796
+ "step": 149675
289797
+ },
289798
+ {
289799
+ "epoch": 1206.77,
289800
+ "learning_rate": 7.5990145395799685e-06,
289801
+ "loss": 0.54,
289802
+ "step": 149680
289803
+ },
289804
+ {
289805
+ "epoch": 1206.81,
289806
+ "learning_rate": 7.5989337641357034e-06,
289807
+ "loss": 1.0087,
289808
+ "step": 149685
289809
+ },
289810
+ {
289811
+ "epoch": 1206.85,
289812
+ "learning_rate": 7.5988529886914384e-06,
289813
+ "loss": 0.2777,
289814
+ "step": 149690
289815
+ },
289816
+ {
289817
+ "epoch": 1206.89,
289818
+ "learning_rate": 7.598772213247173e-06,
289819
+ "loss": 0.2576,
289820
+ "step": 149695
289821
+ },
289822
+ {
289823
+ "epoch": 1206.93,
289824
+ "learning_rate": 7.598691437802908e-06,
289825
+ "loss": 0.3037,
289826
+ "step": 149700
289827
+ },
289828
+ {
289829
+ "epoch": 1206.97,
289830
+ "learning_rate": 7.598610662358643e-06,
289831
+ "loss": 0.5127,
289832
+ "step": 149705
289833
+ },
289834
+ {
289835
+ "epoch": 1207.0,
289836
+ "eval_loss": 0.40235501527786255,
289837
+ "eval_runtime": 42.5624,
289838
+ "eval_samples_per_second": 19.736,
289839
+ "eval_steps_per_second": 0.634,
289840
+ "eval_wer": 0.17382999044890163,
289841
+ "step": 149708
289842
+ },
289843
+ {
289844
+ "epoch": 1207.02,
289845
+ "learning_rate": 7.598529886914378e-06,
289846
+ "loss": 0.3289,
289847
+ "step": 149710
289848
+ },
289849
+ {
289850
+ "epoch": 1207.06,
289851
+ "learning_rate": 7.598449111470114e-06,
289852
+ "loss": 0.258,
289853
+ "step": 149715
289854
+ },
289855
+ {
289856
+ "epoch": 1207.1,
289857
+ "learning_rate": 7.598368336025848e-06,
289858
+ "loss": 0.322,
289859
+ "step": 149720
289860
+ },
289861
+ {
289862
+ "epoch": 1207.14,
289863
+ "learning_rate": 7.598287560581584e-06,
289864
+ "loss": 0.377,
289865
+ "step": 149725
289866
+ },
289867
+ {
289868
+ "epoch": 1207.18,
289869
+ "learning_rate": 7.598206785137318e-06,
289870
+ "loss": 0.6225,
289871
+ "step": 149730
289872
+ },
289873
+ {
289874
+ "epoch": 1207.22,
289875
+ "learning_rate": 7.598126009693054e-06,
289876
+ "loss": 0.8557,
289877
+ "step": 149735
289878
+ },
289879
+ {
289880
+ "epoch": 1207.26,
289881
+ "learning_rate": 7.598045234248788e-06,
289882
+ "loss": 0.2603,
289883
+ "step": 149740
289884
+ },
289885
+ {
289886
+ "epoch": 1207.3,
289887
+ "learning_rate": 7.597964458804524e-06,
289888
+ "loss": 0.2477,
289889
+ "step": 149745
289890
+ },
289891
+ {
289892
+ "epoch": 1207.34,
289893
+ "learning_rate": 7.597883683360259e-06,
289894
+ "loss": 0.304,
289895
+ "step": 149750
289896
+ },
289897
+ {
289898
+ "epoch": 1207.38,
289899
+ "learning_rate": 7.597802907915994e-06,
289900
+ "loss": 0.6338,
289901
+ "step": 149755
289902
+ },
289903
+ {
289904
+ "epoch": 1207.42,
289905
+ "learning_rate": 7.597722132471729e-06,
289906
+ "loss": 0.9901,
289907
+ "step": 149760
289908
+ },
289909
+ {
289910
+ "epoch": 1207.46,
289911
+ "learning_rate": 7.597641357027464e-06,
289912
+ "loss": 0.3567,
289913
+ "step": 149765
289914
+ },
289915
+ {
289916
+ "epoch": 1207.5,
289917
+ "learning_rate": 7.5975605815832e-06,
289918
+ "loss": 0.2428,
289919
+ "step": 149770
289920
+ },
289921
+ {
289922
+ "epoch": 1207.54,
289923
+ "learning_rate": 7.597479806138934e-06,
289924
+ "loss": 0.3402,
289925
+ "step": 149775
289926
+ },
289927
+ {
289928
+ "epoch": 1207.58,
289929
+ "learning_rate": 7.59739903069467e-06,
289930
+ "loss": 0.4664,
289931
+ "step": 149780
289932
+ },
289933
+ {
289934
+ "epoch": 1207.62,
289935
+ "learning_rate": 7.597318255250404e-06,
289936
+ "loss": 0.8885,
289937
+ "step": 149785
289938
+ },
289939
+ {
289940
+ "epoch": 1207.66,
289941
+ "learning_rate": 7.59723747980614e-06,
289942
+ "loss": 0.2265,
289943
+ "step": 149790
289944
+ },
289945
+ {
289946
+ "epoch": 1207.7,
289947
+ "learning_rate": 7.597156704361874e-06,
289948
+ "loss": 0.3005,
289949
+ "step": 149795
289950
+ },
289951
+ {
289952
+ "epoch": 1207.74,
289953
+ "learning_rate": 7.59707592891761e-06,
289954
+ "loss": 0.3373,
289955
+ "step": 149800
289956
+ },
289957
+ {
289958
+ "epoch": 1207.78,
289959
+ "learning_rate": 7.596995153473344e-06,
289960
+ "loss": 0.6167,
289961
+ "step": 149805
289962
+ },
289963
+ {
289964
+ "epoch": 1207.82,
289965
+ "learning_rate": 7.59691437802908e-06,
289966
+ "loss": 0.9362,
289967
+ "step": 149810
289968
+ },
289969
+ {
289970
+ "epoch": 1207.86,
289971
+ "learning_rate": 7.596833602584815e-06,
289972
+ "loss": 0.2685,
289973
+ "step": 149815
289974
+ },
289975
+ {
289976
+ "epoch": 1207.9,
289977
+ "learning_rate": 7.59675282714055e-06,
289978
+ "loss": 0.275,
289979
+ "step": 149820
289980
+ },
289981
+ {
289982
+ "epoch": 1207.94,
289983
+ "learning_rate": 7.596672051696285e-06,
289984
+ "loss": 0.3284,
289985
+ "step": 149825
289986
+ },
289987
+ {
289988
+ "epoch": 1207.98,
289989
+ "learning_rate": 7.59659127625202e-06,
289990
+ "loss": 0.8461,
289991
+ "step": 149830
289992
+ },
289993
+ {
289994
+ "epoch": 1208.0,
289995
+ "eval_loss": 0.370609313249588,
289996
+ "eval_runtime": 41.783,
289997
+ "eval_samples_per_second": 20.104,
289998
+ "eval_steps_per_second": 0.646,
289999
+ "eval_wer": 0.17362270450751252,
290000
+ "step": 149832
290001
+ },
290002
+ {
290003
+ "epoch": 1208.02,
290004
+ "learning_rate": 7.596510500807756e-06,
290005
+ "loss": 0.2909,
290006
+ "step": 149835
290007
+ },
290008
+ {
290009
+ "epoch": 1208.06,
290010
+ "learning_rate": 7.59642972536349e-06,
290011
+ "loss": 0.2565,
290012
+ "step": 149840
290013
+ },
290014
+ {
290015
+ "epoch": 1208.1,
290016
+ "learning_rate": 7.596348949919226e-06,
290017
+ "loss": 0.2966,
290018
+ "step": 149845
290019
+ },
290020
+ {
290021
+ "epoch": 1208.14,
290022
+ "learning_rate": 7.59626817447496e-06,
290023
+ "loss": 0.3049,
290024
+ "step": 149850
290025
+ },
290026
+ {
290027
+ "epoch": 1208.18,
290028
+ "learning_rate": 7.5961873990306956e-06,
290029
+ "loss": 0.7274,
290030
+ "step": 149855
290031
+ },
290032
+ {
290033
+ "epoch": 1208.22,
290034
+ "learning_rate": 7.59610662358643e-06,
290035
+ "loss": 0.7799,
290036
+ "step": 149860
290037
+ },
290038
+ {
290039
+ "epoch": 1208.27,
290040
+ "learning_rate": 7.5960258481421655e-06,
290041
+ "loss": 0.2418,
290042
+ "step": 149865
290043
+ },
290044
+ {
290045
+ "epoch": 1208.31,
290046
+ "learning_rate": 7.5959450726979e-06,
290047
+ "loss": 0.2956,
290048
+ "step": 149870
290049
+ },
290050
+ {
290051
+ "epoch": 1208.35,
290052
+ "learning_rate": 7.5958642972536355e-06,
290053
+ "loss": 0.3961,
290054
+ "step": 149875
290055
+ },
290056
+ {
290057
+ "epoch": 1208.39,
290058
+ "learning_rate": 7.59578352180937e-06,
290059
+ "loss": 0.7961,
290060
+ "step": 149880
290061
+ },
290062
+ {
290063
+ "epoch": 1208.43,
290064
+ "learning_rate": 7.5957027463651055e-06,
290065
+ "loss": 0.88,
290066
+ "step": 149885
290067
+ },
290068
+ {
290069
+ "epoch": 1208.47,
290070
+ "learning_rate": 7.595621970920841e-06,
290071
+ "loss": 0.337,
290072
+ "step": 149890
290073
+ },
290074
+ {
290075
+ "epoch": 1208.51,
290076
+ "learning_rate": 7.5955411954765755e-06,
290077
+ "loss": 0.2649,
290078
+ "step": 149895
290079
+ },
290080
+ {
290081
+ "epoch": 1208.55,
290082
+ "learning_rate": 7.595460420032311e-06,
290083
+ "loss": 0.2962,
290084
+ "step": 149900
290085
+ },
290086
+ {
290087
+ "epoch": 1208.59,
290088
+ "learning_rate": 7.5953796445880454e-06,
290089
+ "loss": 0.7462,
290090
+ "step": 149905
290091
+ },
290092
+ {
290093
+ "epoch": 1208.63,
290094
+ "learning_rate": 7.595298869143781e-06,
290095
+ "loss": 1.0426,
290096
+ "step": 149910
290097
+ },
290098
+ {
290099
+ "epoch": 1208.67,
290100
+ "learning_rate": 7.595218093699515e-06,
290101
+ "loss": 0.2865,
290102
+ "step": 149915
290103
+ },
290104
+ {
290105
+ "epoch": 1208.71,
290106
+ "learning_rate": 7.595137318255251e-06,
290107
+ "loss": 0.2116,
290108
+ "step": 149920
290109
+ },
290110
+ {
290111
+ "epoch": 1208.75,
290112
+ "learning_rate": 7.595056542810985e-06,
290113
+ "loss": 0.4093,
290114
+ "step": 149925
290115
+ },
290116
+ {
290117
+ "epoch": 1208.79,
290118
+ "learning_rate": 7.594975767366721e-06,
290119
+ "loss": 0.7037,
290120
+ "step": 149930
290121
+ },
290122
+ {
290123
+ "epoch": 1208.83,
290124
+ "learning_rate": 7.594894991922455e-06,
290125
+ "loss": 0.8359,
290126
+ "step": 149935
290127
+ },
290128
+ {
290129
+ "epoch": 1208.87,
290130
+ "learning_rate": 7.594814216478191e-06,
290131
+ "loss": 0.2893,
290132
+ "step": 149940
290133
+ },
290134
+ {
290135
+ "epoch": 1208.91,
290136
+ "learning_rate": 7.594733441033927e-06,
290137
+ "loss": 0.293,
290138
+ "step": 149945
290139
+ },
290140
+ {
290141
+ "epoch": 1208.95,
290142
+ "learning_rate": 7.594652665589661e-06,
290143
+ "loss": 0.3372,
290144
+ "step": 149950
290145
+ },
290146
+ {
290147
+ "epoch": 1208.99,
290148
+ "learning_rate": 7.594571890145397e-06,
290149
+ "loss": 0.8728,
290150
+ "step": 149955
290151
+ },
290152
+ {
290153
+ "epoch": 1209.0,
290154
+ "eval_loss": 0.3431926369667053,
290155
+ "eval_runtime": 42.3561,
290156
+ "eval_samples_per_second": 19.855,
290157
+ "eval_steps_per_second": 0.637,
290158
+ "eval_wer": 0.17415083477259644,
290159
+ "step": 149956
290160
+ },
290161
+ {
290162
+ "epoch": 1199.03,
290163
+ "learning_rate": 7.594491114701131e-06,
290164
+ "loss": 0.2865,
290165
+ "step": 149960
290166
+ },
290167
+ {
290168
+ "epoch": 1199.07,
290169
+ "learning_rate": 7.594410339256867e-06,
290170
+ "loss": 0.324,
290171
+ "step": 149965
290172
+ },
290173
+ {
290174
+ "epoch": 1199.11,
290175
+ "learning_rate": 7.594329563812601e-06,
290176
+ "loss": 0.3592,
290177
+ "step": 149970
290178
+ },
290179
+ {
290180
+ "epoch": 1199.15,
290181
+ "learning_rate": 7.594248788368337e-06,
290182
+ "loss": 0.4291,
290183
+ "step": 149975
290184
+ },
290185
+ {
290186
+ "epoch": 1199.19,
290187
+ "learning_rate": 7.594168012924071e-06,
290188
+ "loss": 0.8308,
290189
+ "step": 149980
290190
+ },
290191
+ {
290192
+ "epoch": 1199.23,
290193
+ "learning_rate": 7.594087237479807e-06,
290194
+ "loss": 0.6,
290195
+ "step": 149985
290196
+ },
290197
+ {
290198
+ "epoch": 1199.27,
290199
+ "learning_rate": 7.594006462035541e-06,
290200
+ "loss": 0.2941,
290201
+ "step": 149990
290202
+ },
290203
+ {
290204
+ "epoch": 1199.31,
290205
+ "learning_rate": 7.593925686591277e-06,
290206
+ "loss": 0.2726,
290207
+ "step": 149995
290208
+ },
290209
+ {
290210
+ "epoch": 1199.35,
290211
+ "learning_rate": 7.593844911147011e-06,
290212
+ "loss": 0.349,
290213
+ "step": 150000
290214
+ },
290215
+ {
290216
+ "epoch": 1199.39,
290217
+ "learning_rate": 7.593764135702747e-06,
290218
+ "loss": 0.8815,
290219
+ "step": 150005
290220
+ },
290221
+ {
290222
+ "epoch": 1199.43,
290223
+ "learning_rate": 7.593683360258483e-06,
290224
+ "loss": 0.7146,
290225
+ "step": 150010
290226
+ },
290227
+ {
290228
+ "epoch": 1199.47,
290229
+ "learning_rate": 7.593602584814217e-06,
290230
+ "loss": 0.2515,
290231
+ "step": 150015
290232
+ },
290233
+ {
290234
+ "epoch": 1199.51,
290235
+ "learning_rate": 7.593521809369953e-06,
290236
+ "loss": 0.271,
290237
+ "step": 150020
290238
+ },
290239
+ {
290240
+ "epoch": 1199.55,
290241
+ "learning_rate": 7.593441033925687e-06,
290242
+ "loss": 0.3378,
290243
+ "step": 150025
290244
+ },
290245
+ {
290246
+ "epoch": 1199.59,
290247
+ "learning_rate": 7.593360258481423e-06,
290248
+ "loss": 0.8722,
290249
+ "step": 150030
290250
+ },
290251
+ {
290252
+ "epoch": 1199.63,
290253
+ "learning_rate": 7.593279483037157e-06,
290254
+ "loss": 0.5991,
290255
+ "step": 150035
290256
+ },
290257
+ {
290258
+ "epoch": 1199.67,
290259
+ "learning_rate": 7.593198707592893e-06,
290260
+ "loss": 0.3126,
290261
+ "step": 150040
290262
+ },
290263
+ {
290264
+ "epoch": 1199.71,
290265
+ "learning_rate": 7.593117932148627e-06,
290266
+ "loss": 0.2889,
290267
+ "step": 150045
290268
+ },
290269
+ {
290270
+ "epoch": 1199.75,
290271
+ "learning_rate": 7.593037156704363e-06,
290272
+ "loss": 0.4038,
290273
+ "step": 150050
290274
+ },
290275
+ {
290276
+ "epoch": 1199.79,
290277
+ "learning_rate": 7.592956381260097e-06,
290278
+ "loss": 0.9201,
290279
+ "step": 150055
290280
+ },
290281
+ {
290282
+ "epoch": 1199.83,
290283
+ "learning_rate": 7.592875605815833e-06,
290284
+ "loss": 0.6433,
290285
+ "step": 150060
290286
+ },
290287
+ {
290288
+ "epoch": 1199.87,
290289
+ "learning_rate": 7.592794830371568e-06,
290290
+ "loss": 0.2426,
290291
+ "step": 150065
290292
+ },
290293
+ {
290294
+ "epoch": 1199.91,
290295
+ "learning_rate": 7.5927140549273026e-06,
290296
+ "loss": 0.2838,
290297
+ "step": 150070
290298
+ },
290299
+ {
290300
+ "epoch": 1199.95,
290301
+ "learning_rate": 7.592633279483038e-06,
290302
+ "loss": 0.3511,
290303
+ "step": 150075
290304
+ },
290305
+ {
290306
+ "epoch": 1199.99,
290307
+ "learning_rate": 7.5925525040387725e-06,
290308
+ "loss": 1.0652,
290309
+ "step": 150080
290310
+ },
290311
+ {
290312
+ "epoch": 1200.0,
290313
+ "eval_loss": 0.38891738653182983,
290314
+ "eval_runtime": 42.3061,
290315
+ "eval_samples_per_second": 19.879,
290316
+ "eval_steps_per_second": 0.638,
290317
+ "eval_wer": 0.1781133181851499,
290318
+ "step": 150081
290319
+ },
290320
+ {
290321
+ "epoch": 1210.03,
290322
+ "learning_rate": 7.592471728594508e-06,
290323
+ "loss": 0.2999,
290324
+ "step": 150085
290325
+ },
290326
+ {
290327
+ "epoch": 1210.07,
290328
+ "learning_rate": 7.5923909531502425e-06,
290329
+ "loss": 0.2679,
290330
+ "step": 150090
290331
+ },
290332
+ {
290333
+ "epoch": 1210.11,
290334
+ "learning_rate": 7.592310177705978e-06,
290335
+ "loss": 0.2975,
290336
+ "step": 150095
290337
+ },
290338
+ {
290339
+ "epoch": 1210.15,
290340
+ "learning_rate": 7.5922294022617125e-06,
290341
+ "loss": 0.3568,
290342
+ "step": 150100
290343
+ },
290344
+ {
290345
+ "epoch": 1210.19,
290346
+ "learning_rate": 7.592148626817448e-06,
290347
+ "loss": 0.8083,
290348
+ "step": 150105
290349
+ },
290350
+ {
290351
+ "epoch": 1210.23,
290352
+ "learning_rate": 7.5920678513731825e-06,
290353
+ "loss": 0.6369,
290354
+ "step": 150110
290355
+ },
290356
+ {
290357
+ "epoch": 1210.27,
290358
+ "learning_rate": 7.591987075928918e-06,
290359
+ "loss": 0.2588,
290360
+ "step": 150115
290361
+ },
290362
+ {
290363
+ "epoch": 1210.31,
290364
+ "learning_rate": 7.591906300484654e-06,
290365
+ "loss": 0.2907,
290366
+ "step": 150120
290367
+ },
290368
+ {
290369
+ "epoch": 1210.35,
290370
+ "learning_rate": 7.591825525040388e-06,
290371
+ "loss": 0.332,
290372
+ "step": 150125
290373
+ },
290374
+ {
290375
+ "epoch": 1210.4,
290376
+ "learning_rate": 7.591744749596124e-06,
290377
+ "loss": 0.8868,
290378
+ "step": 150130
290379
+ },
290380
+ {
290381
+ "epoch": 1210.44,
290382
+ "learning_rate": 7.591663974151858e-06,
290383
+ "loss": 0.5735,
290384
+ "step": 150135
290385
+ },
290386
+ {
290387
+ "epoch": 1210.48,
290388
+ "learning_rate": 7.591583198707594e-06,
290389
+ "loss": 0.2798,
290390
+ "step": 150140
290391
+ },
290392
+ {
290393
+ "epoch": 1210.52,
290394
+ "learning_rate": 7.591502423263328e-06,
290395
+ "loss": 0.3032,
290396
+ "step": 150145
290397
+ },
290398
+ {
290399
+ "epoch": 1210.56,
290400
+ "learning_rate": 7.591421647819064e-06,
290401
+ "loss": 0.42,
290402
+ "step": 150150
290403
+ },
290404
+ {
290405
+ "epoch": 1210.6,
290406
+ "learning_rate": 7.591340872374798e-06,
290407
+ "loss": 0.8859,
290408
+ "step": 150155
290409
+ },
290410
+ {
290411
+ "epoch": 1210.64,
290412
+ "learning_rate": 7.591260096930534e-06,
290413
+ "loss": 0.593,
290414
+ "step": 150160
290415
+ },
290416
+ {
290417
+ "epoch": 1210.68,
290418
+ "learning_rate": 7.591179321486268e-06,
290419
+ "loss": 0.2568,
290420
+ "step": 150165
290421
+ },
290422
+ {
290423
+ "epoch": 1210.72,
290424
+ "learning_rate": 7.591098546042004e-06,
290425
+ "loss": 0.295,
290426
+ "step": 150170
290427
+ },
290428
+ {
290429
+ "epoch": 1210.76,
290430
+ "learning_rate": 7.591017770597738e-06,
290431
+ "loss": 0.3517,
290432
+ "step": 150175
290433
+ },
290434
+ {
290435
+ "epoch": 1210.8,
290436
+ "learning_rate": 7.590936995153474e-06,
290437
+ "loss": 0.9385,
290438
+ "step": 150180
290439
+ },
290440
+ {
290441
+ "epoch": 1210.84,
290442
+ "learning_rate": 7.59085621970921e-06,
290443
+ "loss": 0.6471,
290444
+ "step": 150185
290445
+ },
290446
+ {
290447
+ "epoch": 1210.88,
290448
+ "learning_rate": 7.590775444264944e-06,
290449
+ "loss": 0.2162,
290450
+ "step": 150190
290451
+ },
290452
+ {
290453
+ "epoch": 1210.92,
290454
+ "learning_rate": 7.59069466882068e-06,
290455
+ "loss": 0.2966,
290456
+ "step": 150195
290457
+ },
290458
+ {
290459
+ "epoch": 1210.96,
290460
+ "learning_rate": 7.590613893376414e-06,
290461
+ "loss": 0.4473,
290462
+ "step": 150200
290463
+ },
290464
+ {
290465
+ "epoch": 1211.0,
290466
+ "learning_rate": 7.59053311793215e-06,
290467
+ "loss": 1.2014,
290468
+ "step": 150205
290469
+ },
290470
+ {
290471
+ "epoch": 1211.0,
290472
+ "eval_loss": 0.3155873119831085,
290473
+ "eval_runtime": 41.9259,
290474
+ "eval_samples_per_second": 20.059,
290475
+ "eval_steps_per_second": 0.644,
290476
+ "eval_wer": 0.17267648552564754,
290477
+ "step": 150205
290478
  }
290479
  ],
290480
+ "max_steps": 620000,
290481
  "num_train_epochs": 5000,
290482
+ "total_flos": 4.226898230914657e+20,
290483
  "trial_name": null,
290484
  "trial_params": null
290485
  }
model-bin/finetune/base/{checkpoint-149584 β†’ checkpoint-150205}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630241766.1674721/events.out.tfevents.1630241766.cc93b136ebf5.1086.169 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a8e839a63f1044cdd3733e6d1902b9a273438784bee6f325e386f0b3586fc66
3
+ size 4194
model-bin/finetune/base/log/1630242210.9386408/events.out.tfevents.1630242210.cc93b136ebf5.1086.171 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84cec864ad98be8751dd931572737d3082faa656b62041daab2c9a739ce3b6ea
3
+ size 4194
model-bin/finetune/base/log/1630242652.9526577/events.out.tfevents.1630242652.cc93b136ebf5.1086.173 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f065a191ea7a66fef5f98843dd7ebd9e1d383540d5907ac328552393f3bf6ea
3
+ size 4194
model-bin/finetune/base/log/1630243093.5265884/events.out.tfevents.1630243093.cc93b136ebf5.1086.175 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98cb2b24b76763ede43b591c8a759dd0542a095509a8f38257e99523b07e9bef
3
+ size 4194
model-bin/finetune/base/log/1630243533.7269483/events.out.tfevents.1630243533.cc93b136ebf5.1086.177 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bf56e0a8edf9659a6ca25eee11f9791f25f55e4e3afbfee3befb4f9cc0dc0d2
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630241766.cc93b136ebf5.1086.168 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d37d442dac7faeb1f442af710999b3add3e950fda2f57ffd9ffc909c07dd4b3
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630242210.cc93b136ebf5.1086.170 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bac685984e447ab85c1e5e90ffaf8956acfa0a048401576ac463f1c896e28726
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630242652.cc93b136ebf5.1086.172 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1469d3ae4ad89e67e6efdb48f9759769145fb48aedebdc5644e6d86e7f741cc7
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630243093.cc93b136ebf5.1086.174 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:088b40b88599462f943492215da07c29f168d2b553b786062366b001377e6f31
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630243533.cc93b136ebf5.1086.176 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c77cd55108a8f9d2cb1905ea48e27f7e0979770117b25faa9ec77ac6e38f6997
3
+ size 8622