Check commited on
Commit
cb2f2d4
Β·
1 Parent(s): 5881b1b

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/trainer_state.json +793 -4
  9. model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629901744.9403872/events.out.tfevents.1629901744.7e498afd5545.7645.55 +3 -0
  11. model-bin/finetune/base/log/1629902211.333175/events.out.tfevents.1629902211.7e498afd5545.7645.57 +3 -0
  12. model-bin/finetune/base/log/1629902676.4913468/events.out.tfevents.1629902676.7e498afd5545.7645.59 +3 -0
  13. model-bin/finetune/base/log/1629903142.829317/events.out.tfevents.1629903142.7e498afd5545.7645.61 +3 -0
  14. model-bin/finetune/base/log/1629903613.046621/events.out.tfevents.1629903613.7e498afd5545.7645.63 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629901744.7e498afd5545.7645.54 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629902211.7e498afd5545.7645.56 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629902675.7e498afd5545.7645.58 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629903142.7e498afd5545.7645.60 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629903613.7e498afd5545.7645.62 +3 -0
model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ae74db1cc9558598af05d635223b27c5d6ed48437987e04f589dad401869e06a
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2402c050872a1b6c2132f1354c8235dbc30ffdaec42a7c3e2fe40a3fa3597e6e
3
  size 722165393
model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:32306985457fff3e0a5e10efdbee69e3dce658c6399512f2b5d72e5fe51ad2fb
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ba89a49bd0701ea0db537127d26ed94e4e5fa6eed33bd9ae9d90248a8d473ba
3
  size 377909911
model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:db22fbe7889c01e9608ac0cdcc29148f4582a1bf9b603a12a8aa4bf492d79ef0
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea9d23665821b123a85d6ac5512b6a51253cf13d4f317b3d68b6de999fb06ba7
3
  size 14503
model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c625136fb9e282630f07105a0aae0239f4f5b2ec21da675a9e1b62b4a98867ec
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2edae0f6397edeecd1aedd6821f5f0301b0061a47a55585acacade7161821b94
3
  size 559
model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2fafede2354a24b85129916900c59469f87b31b303b25ff5163560e339b7c4cc
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ca9b1c1ac3a3f74bba9515f59a166eb16c9179459f7d8dfba604bf995fff1f5
3
  size 623
model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1822989349981638,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-79148",
4
- "epoch": 641.995983935743,
5
- "global_step": 79647,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -200700,11 +200700,800 @@
200700
  "eval_steps_per_second": 0.643,
200701
  "eval_wer": 0.19413337186619464,
200702
  "step": 79647
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
200703
  }
200704
  ],
200705
- "max_steps": 620000,
200706
  "num_train_epochs": 5000,
200707
- "total_flos": 2.241393263792259e+20,
200708
  "trial_name": null,
200709
  "trial_params": null
200710
  }
 
1
  {
2
  "best_metric": 0.1822989349981638,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-79148",
4
+ "epoch": 642.0,
5
+ "global_step": 80269,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
200700
  "eval_steps_per_second": 0.643,
200701
  "eval_wer": 0.19413337186619464,
200702
  "step": 79647
200703
+ },
200704
+ {
200705
+ "epoch": 642.02,
200706
+ "learning_rate": 8.729903069466882e-06,
200707
+ "loss": 0.3702,
200708
+ "step": 79650
200709
+ },
200710
+ {
200711
+ "epoch": 642.06,
200712
+ "learning_rate": 8.729822294022618e-06,
200713
+ "loss": 0.3383,
200714
+ "step": 79655
200715
+ },
200716
+ {
200717
+ "epoch": 642.1,
200718
+ "learning_rate": 8.729741518578354e-06,
200719
+ "loss": 0.4167,
200720
+ "step": 79660
200721
+ },
200722
+ {
200723
+ "epoch": 642.14,
200724
+ "learning_rate": 8.729660743134088e-06,
200725
+ "loss": 0.3496,
200726
+ "step": 79665
200727
+ },
200728
+ {
200729
+ "epoch": 642.18,
200730
+ "learning_rate": 8.729579967689824e-06,
200731
+ "loss": 0.7132,
200732
+ "step": 79670
200733
+ },
200734
+ {
200735
+ "epoch": 642.22,
200736
+ "learning_rate": 8.729499192245558e-06,
200737
+ "loss": 0.7956,
200738
+ "step": 79675
200739
+ },
200740
+ {
200741
+ "epoch": 642.27,
200742
+ "learning_rate": 8.729418416801294e-06,
200743
+ "loss": 0.3064,
200744
+ "step": 79680
200745
+ },
200746
+ {
200747
+ "epoch": 642.31,
200748
+ "learning_rate": 8.729337641357028e-06,
200749
+ "loss": 0.3004,
200750
+ "step": 79685
200751
+ },
200752
+ {
200753
+ "epoch": 642.35,
200754
+ "learning_rate": 8.729256865912764e-06,
200755
+ "loss": 0.3636,
200756
+ "step": 79690
200757
+ },
200758
+ {
200759
+ "epoch": 642.39,
200760
+ "learning_rate": 8.729176090468498e-06,
200761
+ "loss": 0.7391,
200762
+ "step": 79695
200763
+ },
200764
+ {
200765
+ "epoch": 642.43,
200766
+ "learning_rate": 8.729095315024234e-06,
200767
+ "loss": 0.7999,
200768
+ "step": 79700
200769
+ },
200770
+ {
200771
+ "epoch": 642.47,
200772
+ "learning_rate": 8.729014539579968e-06,
200773
+ "loss": 0.3359,
200774
+ "step": 79705
200775
+ },
200776
+ {
200777
+ "epoch": 642.51,
200778
+ "learning_rate": 8.728933764135704e-06,
200779
+ "loss": 0.2725,
200780
+ "step": 79710
200781
+ },
200782
+ {
200783
+ "epoch": 642.55,
200784
+ "learning_rate": 8.728852988691438e-06,
200785
+ "loss": 0.3936,
200786
+ "step": 79715
200787
+ },
200788
+ {
200789
+ "epoch": 642.59,
200790
+ "learning_rate": 8.728772213247174e-06,
200791
+ "loss": 0.8775,
200792
+ "step": 79720
200793
+ },
200794
+ {
200795
+ "epoch": 642.63,
200796
+ "learning_rate": 8.72869143780291e-06,
200797
+ "loss": 1.0067,
200798
+ "step": 79725
200799
+ },
200800
+ {
200801
+ "epoch": 642.67,
200802
+ "learning_rate": 8.728610662358644e-06,
200803
+ "loss": 0.3226,
200804
+ "step": 79730
200805
+ },
200806
+ {
200807
+ "epoch": 642.71,
200808
+ "learning_rate": 8.72852988691438e-06,
200809
+ "loss": 0.3211,
200810
+ "step": 79735
200811
+ },
200812
+ {
200813
+ "epoch": 642.75,
200814
+ "learning_rate": 8.728449111470114e-06,
200815
+ "loss": 0.4379,
200816
+ "step": 79740
200817
+ },
200818
+ {
200819
+ "epoch": 642.79,
200820
+ "learning_rate": 8.72836833602585e-06,
200821
+ "loss": 0.7525,
200822
+ "step": 79745
200823
+ },
200824
+ {
200825
+ "epoch": 642.83,
200826
+ "learning_rate": 8.728287560581584e-06,
200827
+ "loss": 0.8896,
200828
+ "step": 79750
200829
+ },
200830
+ {
200831
+ "epoch": 642.87,
200832
+ "learning_rate": 8.72820678513732e-06,
200833
+ "loss": 0.5549,
200834
+ "step": 79755
200835
+ },
200836
+ {
200837
+ "epoch": 642.91,
200838
+ "learning_rate": 8.728126009693054e-06,
200839
+ "loss": 0.3093,
200840
+ "step": 79760
200841
+ },
200842
+ {
200843
+ "epoch": 642.95,
200844
+ "learning_rate": 8.72804523424879e-06,
200845
+ "loss": 0.3845,
200846
+ "step": 79765
200847
+ },
200848
+ {
200849
+ "epoch": 642.99,
200850
+ "learning_rate": 8.727964458804523e-06,
200851
+ "loss": 0.8744,
200852
+ "step": 79770
200853
+ },
200854
+ {
200855
+ "epoch": 643.0,
200856
+ "eval_loss": 0.3736218214035034,
200857
+ "eval_runtime": 43.1561,
200858
+ "eval_samples_per_second": 19.487,
200859
+ "eval_steps_per_second": 0.626,
200860
+ "eval_wer": 0.18469290878551084,
200861
+ "step": 79771
200862
+ },
200863
+ {
200864
+ "epoch": 643.03,
200865
+ "learning_rate": 8.72788368336026e-06,
200866
+ "loss": 0.3615,
200867
+ "step": 79775
200868
+ },
200869
+ {
200870
+ "epoch": 643.07,
200871
+ "learning_rate": 8.727802907915993e-06,
200872
+ "loss": 0.3373,
200873
+ "step": 79780
200874
+ },
200875
+ {
200876
+ "epoch": 643.11,
200877
+ "learning_rate": 8.72772213247173e-06,
200878
+ "loss": 0.3368,
200879
+ "step": 79785
200880
+ },
200881
+ {
200882
+ "epoch": 643.15,
200883
+ "learning_rate": 8.727641357027465e-06,
200884
+ "loss": 0.412,
200885
+ "step": 79790
200886
+ },
200887
+ {
200888
+ "epoch": 643.19,
200889
+ "learning_rate": 8.7275605815832e-06,
200890
+ "loss": 0.9684,
200891
+ "step": 79795
200892
+ },
200893
+ {
200894
+ "epoch": 643.23,
200895
+ "learning_rate": 8.727479806138935e-06,
200896
+ "loss": 0.7489,
200897
+ "step": 79800
200898
+ },
200899
+ {
200900
+ "epoch": 643.27,
200901
+ "learning_rate": 8.72739903069467e-06,
200902
+ "loss": 0.3246,
200903
+ "step": 79805
200904
+ },
200905
+ {
200906
+ "epoch": 643.31,
200907
+ "learning_rate": 8.727318255250405e-06,
200908
+ "loss": 0.3444,
200909
+ "step": 79810
200910
+ },
200911
+ {
200912
+ "epoch": 643.35,
200913
+ "learning_rate": 8.72723747980614e-06,
200914
+ "loss": 0.4126,
200915
+ "step": 79815
200916
+ },
200917
+ {
200918
+ "epoch": 643.39,
200919
+ "learning_rate": 8.727156704361875e-06,
200920
+ "loss": 0.865,
200921
+ "step": 79820
200922
+ },
200923
+ {
200924
+ "epoch": 643.43,
200925
+ "learning_rate": 8.72707592891761e-06,
200926
+ "loss": 0.743,
200927
+ "step": 79825
200928
+ },
200929
+ {
200930
+ "epoch": 643.47,
200931
+ "learning_rate": 8.726995153473345e-06,
200932
+ "loss": 0.3948,
200933
+ "step": 79830
200934
+ },
200935
+ {
200936
+ "epoch": 643.51,
200937
+ "learning_rate": 8.72691437802908e-06,
200938
+ "loss": 0.3934,
200939
+ "step": 79835
200940
+ },
200941
+ {
200942
+ "epoch": 643.55,
200943
+ "learning_rate": 8.726833602584815e-06,
200944
+ "loss": 0.4111,
200945
+ "step": 79840
200946
+ },
200947
+ {
200948
+ "epoch": 643.59,
200949
+ "learning_rate": 8.726752827140549e-06,
200950
+ "loss": 1.0506,
200951
+ "step": 79845
200952
+ },
200953
+ {
200954
+ "epoch": 643.63,
200955
+ "learning_rate": 8.726672051696285e-06,
200956
+ "loss": 0.6835,
200957
+ "step": 79850
200958
+ },
200959
+ {
200960
+ "epoch": 643.67,
200961
+ "learning_rate": 8.726591276252019e-06,
200962
+ "loss": 0.2641,
200963
+ "step": 79855
200964
+ },
200965
+ {
200966
+ "epoch": 643.71,
200967
+ "learning_rate": 8.726510500807755e-06,
200968
+ "loss": 0.7048,
200969
+ "step": 79860
200970
+ },
200971
+ {
200972
+ "epoch": 643.76,
200973
+ "learning_rate": 8.72642972536349e-06,
200974
+ "loss": 0.3964,
200975
+ "step": 79865
200976
+ },
200977
+ {
200978
+ "epoch": 643.8,
200979
+ "learning_rate": 8.726348949919225e-06,
200980
+ "loss": 1.0147,
200981
+ "step": 79870
200982
+ },
200983
+ {
200984
+ "epoch": 643.84,
200985
+ "learning_rate": 8.72626817447496e-06,
200986
+ "loss": 0.7456,
200987
+ "step": 79875
200988
+ },
200989
+ {
200990
+ "epoch": 643.88,
200991
+ "learning_rate": 8.726187399030695e-06,
200992
+ "loss": 0.3131,
200993
+ "step": 79880
200994
+ },
200995
+ {
200996
+ "epoch": 643.92,
200997
+ "learning_rate": 8.72610662358643e-06,
200998
+ "loss": 0.3797,
200999
+ "step": 79885
201000
+ },
201001
+ {
201002
+ "epoch": 643.96,
201003
+ "learning_rate": 8.726025848142165e-06,
201004
+ "loss": 0.4774,
201005
+ "step": 79890
201006
+ },
201007
+ {
201008
+ "epoch": 644.0,
201009
+ "learning_rate": 8.7259450726979e-06,
201010
+ "loss": 1.1313,
201011
+ "step": 79895
201012
+ },
201013
+ {
201014
+ "epoch": 644.0,
201015
+ "eval_loss": 0.4221903383731842,
201016
+ "eval_runtime": 41.5264,
201017
+ "eval_samples_per_second": 20.252,
201018
+ "eval_steps_per_second": 0.65,
201019
+ "eval_wer": 0.1915672351950809,
201020
+ "step": 79895
201021
+ },
201022
+ {
201023
+ "epoch": 639.04,
201024
+ "learning_rate": 8.725864297253635e-06,
201025
+ "loss": 0.3224,
201026
+ "step": 79900
201027
+ },
201028
+ {
201029
+ "epoch": 639.08,
201030
+ "learning_rate": 8.72578352180937e-06,
201031
+ "loss": 0.2609,
201032
+ "step": 79905
201033
+ },
201034
+ {
201035
+ "epoch": 639.12,
201036
+ "learning_rate": 8.725702746365105e-06,
201037
+ "loss": 0.3024,
201038
+ "step": 79910
201039
+ },
201040
+ {
201041
+ "epoch": 639.16,
201042
+ "learning_rate": 8.72562197092084e-06,
201043
+ "loss": 0.4686,
201044
+ "step": 79915
201045
+ },
201046
+ {
201047
+ "epoch": 639.2,
201048
+ "learning_rate": 8.725541195476575e-06,
201049
+ "loss": 1.3208,
201050
+ "step": 79920
201051
+ },
201052
+ {
201053
+ "epoch": 639.24,
201054
+ "learning_rate": 8.72546042003231e-06,
201055
+ "loss": 0.3217,
201056
+ "step": 79925
201057
+ },
201058
+ {
201059
+ "epoch": 639.28,
201060
+ "learning_rate": 8.725379644588046e-06,
201061
+ "loss": 0.3378,
201062
+ "step": 79930
201063
+ },
201064
+ {
201065
+ "epoch": 639.32,
201066
+ "learning_rate": 8.72529886914378e-06,
201067
+ "loss": 0.3434,
201068
+ "step": 79935
201069
+ },
201070
+ {
201071
+ "epoch": 639.36,
201072
+ "learning_rate": 8.725218093699516e-06,
201073
+ "loss": 0.4547,
201074
+ "step": 79940
201075
+ },
201076
+ {
201077
+ "epoch": 639.4,
201078
+ "learning_rate": 8.72513731825525e-06,
201079
+ "loss": 1.2898,
201080
+ "step": 79945
201081
+ },
201082
+ {
201083
+ "epoch": 639.44,
201084
+ "learning_rate": 8.725056542810986e-06,
201085
+ "loss": 0.3957,
201086
+ "step": 79950
201087
+ },
201088
+ {
201089
+ "epoch": 639.48,
201090
+ "learning_rate": 8.72497576736672e-06,
201091
+ "loss": 0.3162,
201092
+ "step": 79955
201093
+ },
201094
+ {
201095
+ "epoch": 639.52,
201096
+ "learning_rate": 8.724894991922456e-06,
201097
+ "loss": 0.3148,
201098
+ "step": 79960
201099
+ },
201100
+ {
201101
+ "epoch": 639.56,
201102
+ "learning_rate": 8.72481421647819e-06,
201103
+ "loss": 0.5403,
201104
+ "step": 79965
201105
+ },
201106
+ {
201107
+ "epoch": 639.6,
201108
+ "learning_rate": 8.724733441033926e-06,
201109
+ "loss": 1.2322,
201110
+ "step": 79970
201111
+ },
201112
+ {
201113
+ "epoch": 639.64,
201114
+ "learning_rate": 8.72465266558966e-06,
201115
+ "loss": 0.3714,
201116
+ "step": 79975
201117
+ },
201118
+ {
201119
+ "epoch": 639.68,
201120
+ "learning_rate": 8.724571890145396e-06,
201121
+ "loss": 0.2936,
201122
+ "step": 79980
201123
+ },
201124
+ {
201125
+ "epoch": 639.72,
201126
+ "learning_rate": 8.72449111470113e-06,
201127
+ "loss": 0.3196,
201128
+ "step": 79985
201129
+ },
201130
+ {
201131
+ "epoch": 639.76,
201132
+ "learning_rate": 8.724410339256866e-06,
201133
+ "loss": 0.4734,
201134
+ "step": 79990
201135
+ },
201136
+ {
201137
+ "epoch": 639.8,
201138
+ "learning_rate": 8.724329563812602e-06,
201139
+ "loss": 1.2245,
201140
+ "step": 79995
201141
+ },
201142
+ {
201143
+ "epoch": 639.84,
201144
+ "learning_rate": 8.724248788368336e-06,
201145
+ "loss": 0.3541,
201146
+ "step": 80000
201147
+ },
201148
+ {
201149
+ "epoch": 639.88,
201150
+ "learning_rate": 8.724168012924072e-06,
201151
+ "loss": 0.3572,
201152
+ "step": 80005
201153
+ },
201154
+ {
201155
+ "epoch": 639.92,
201156
+ "learning_rate": 8.724087237479806e-06,
201157
+ "loss": 0.3127,
201158
+ "step": 80010
201159
+ },
201160
+ {
201161
+ "epoch": 639.96,
201162
+ "learning_rate": 8.724006462035542e-06,
201163
+ "loss": 0.482,
201164
+ "step": 80015
201165
+ },
201166
+ {
201167
+ "epoch": 640.0,
201168
+ "learning_rate": 8.723925686591276e-06,
201169
+ "loss": 1.1057,
201170
+ "step": 80020
201171
+ },
201172
+ {
201173
+ "epoch": 640.0,
201174
+ "eval_loss": 0.40086424350738525,
201175
+ "eval_runtime": 43.143,
201176
+ "eval_samples_per_second": 19.493,
201177
+ "eval_steps_per_second": 0.626,
201178
+ "eval_wer": 0.19523465703971118,
201179
+ "step": 80020
201180
+ },
201181
+ {
201182
+ "epoch": 645.04,
201183
+ "learning_rate": 8.723844911147012e-06,
201184
+ "loss": 0.3555,
201185
+ "step": 80025
201186
+ },
201187
+ {
201188
+ "epoch": 645.08,
201189
+ "learning_rate": 8.723764135702746e-06,
201190
+ "loss": 0.2825,
201191
+ "step": 80030
201192
+ },
201193
+ {
201194
+ "epoch": 645.12,
201195
+ "learning_rate": 8.723683360258482e-06,
201196
+ "loss": 0.3342,
201197
+ "step": 80035
201198
+ },
201199
+ {
201200
+ "epoch": 645.16,
201201
+ "learning_rate": 8.723602584814216e-06,
201202
+ "loss": 0.5168,
201203
+ "step": 80040
201204
+ },
201205
+ {
201206
+ "epoch": 645.2,
201207
+ "learning_rate": 8.723521809369952e-06,
201208
+ "loss": 1.1547,
201209
+ "step": 80045
201210
+ },
201211
+ {
201212
+ "epoch": 645.24,
201213
+ "learning_rate": 8.723441033925686e-06,
201214
+ "loss": 0.331,
201215
+ "step": 80050
201216
+ },
201217
+ {
201218
+ "epoch": 645.28,
201219
+ "learning_rate": 8.723360258481422e-06,
201220
+ "loss": 0.3295,
201221
+ "step": 80055
201222
+ },
201223
+ {
201224
+ "epoch": 645.32,
201225
+ "learning_rate": 8.723279483037156e-06,
201226
+ "loss": 0.3176,
201227
+ "step": 80060
201228
+ },
201229
+ {
201230
+ "epoch": 645.36,
201231
+ "learning_rate": 8.723198707592892e-06,
201232
+ "loss": 0.5193,
201233
+ "step": 80065
201234
+ },
201235
+ {
201236
+ "epoch": 645.4,
201237
+ "learning_rate": 8.723117932148628e-06,
201238
+ "loss": 1.2665,
201239
+ "step": 80070
201240
+ },
201241
+ {
201242
+ "epoch": 645.44,
201243
+ "learning_rate": 8.723037156704362e-06,
201244
+ "loss": 0.306,
201245
+ "step": 80075
201246
+ },
201247
+ {
201248
+ "epoch": 645.48,
201249
+ "learning_rate": 8.722956381260098e-06,
201250
+ "loss": 0.3392,
201251
+ "step": 80080
201252
+ },
201253
+ {
201254
+ "epoch": 645.52,
201255
+ "learning_rate": 8.722875605815832e-06,
201256
+ "loss": 0.3802,
201257
+ "step": 80085
201258
+ },
201259
+ {
201260
+ "epoch": 645.56,
201261
+ "learning_rate": 8.722794830371568e-06,
201262
+ "loss": 0.4104,
201263
+ "step": 80090
201264
+ },
201265
+ {
201266
+ "epoch": 645.6,
201267
+ "learning_rate": 8.722714054927302e-06,
201268
+ "loss": 1.2221,
201269
+ "step": 80095
201270
+ },
201271
+ {
201272
+ "epoch": 645.64,
201273
+ "learning_rate": 8.722633279483038e-06,
201274
+ "loss": 0.3854,
201275
+ "step": 80100
201276
+ },
201277
+ {
201278
+ "epoch": 645.68,
201279
+ "learning_rate": 8.722552504038772e-06,
201280
+ "loss": 0.357,
201281
+ "step": 80105
201282
+ },
201283
+ {
201284
+ "epoch": 645.72,
201285
+ "learning_rate": 8.722471728594508e-06,
201286
+ "loss": 0.5174,
201287
+ "step": 80110
201288
+ },
201289
+ {
201290
+ "epoch": 645.76,
201291
+ "learning_rate": 8.722390953150242e-06,
201292
+ "loss": 0.5564,
201293
+ "step": 80115
201294
+ },
201295
+ {
201296
+ "epoch": 645.8,
201297
+ "learning_rate": 8.722310177705978e-06,
201298
+ "loss": 1.1946,
201299
+ "step": 80120
201300
+ },
201301
+ {
201302
+ "epoch": 645.84,
201303
+ "learning_rate": 8.722229402261712e-06,
201304
+ "loss": 0.3119,
201305
+ "step": 80125
201306
+ },
201307
+ {
201308
+ "epoch": 645.88,
201309
+ "learning_rate": 8.722148626817448e-06,
201310
+ "loss": 0.305,
201311
+ "step": 80130
201312
+ },
201313
+ {
201314
+ "epoch": 645.92,
201315
+ "learning_rate": 8.722067851373184e-06,
201316
+ "loss": 0.4067,
201317
+ "step": 80135
201318
+ },
201319
+ {
201320
+ "epoch": 645.96,
201321
+ "learning_rate": 8.721987075928918e-06,
201322
+ "loss": 0.4924,
201323
+ "step": 80140
201324
+ },
201325
+ {
201326
+ "epoch": 646.0,
201327
+ "eval_loss": 0.40166351199150085,
201328
+ "eval_runtime": 41.5534,
201329
+ "eval_samples_per_second": 20.239,
201330
+ "eval_steps_per_second": 0.65,
201331
+ "eval_wer": 0.1887832837553508,
201332
+ "step": 80144
201333
+ },
201334
+ {
201335
+ "epoch": 641.01,
201336
+ "learning_rate": 8.721906300484653e-06,
201337
+ "loss": 0.3404,
201338
+ "step": 80145
201339
+ },
201340
+ {
201341
+ "epoch": 641.05,
201342
+ "learning_rate": 8.721825525040388e-06,
201343
+ "loss": 0.325,
201344
+ "step": 80150
201345
+ },
201346
+ {
201347
+ "epoch": 641.09,
201348
+ "learning_rate": 8.721744749596123e-06,
201349
+ "loss": 0.3356,
201350
+ "step": 80155
201351
+ },
201352
+ {
201353
+ "epoch": 641.13,
201354
+ "learning_rate": 8.721663974151858e-06,
201355
+ "loss": 0.3636,
201356
+ "step": 80160
201357
+ },
201358
+ {
201359
+ "epoch": 641.17,
201360
+ "learning_rate": 8.721583198707593e-06,
201361
+ "loss": 0.7166,
201362
+ "step": 80165
201363
+ },
201364
+ {
201365
+ "epoch": 641.21,
201366
+ "learning_rate": 8.721502423263328e-06,
201367
+ "loss": 1.2031,
201368
+ "step": 80170
201369
+ },
201370
+ {
201371
+ "epoch": 641.25,
201372
+ "learning_rate": 8.721421647819063e-06,
201373
+ "loss": 0.2894,
201374
+ "step": 80175
201375
+ },
201376
+ {
201377
+ "epoch": 641.29,
201378
+ "learning_rate": 8.721340872374798e-06,
201379
+ "loss": 0.3588,
201380
+ "step": 80180
201381
+ },
201382
+ {
201383
+ "epoch": 641.33,
201384
+ "learning_rate": 8.721260096930533e-06,
201385
+ "loss": 0.344,
201386
+ "step": 80185
201387
+ },
201388
+ {
201389
+ "epoch": 641.37,
201390
+ "learning_rate": 8.721179321486268e-06,
201391
+ "loss": 0.5297,
201392
+ "step": 80190
201393
+ },
201394
+ {
201395
+ "epoch": 641.41,
201396
+ "learning_rate": 8.721098546042003e-06,
201397
+ "loss": 1.0916,
201398
+ "step": 80195
201399
+ },
201400
+ {
201401
+ "epoch": 641.45,
201402
+ "learning_rate": 8.72101777059774e-06,
201403
+ "loss": 0.2758,
201404
+ "step": 80200
201405
+ },
201406
+ {
201407
+ "epoch": 641.49,
201408
+ "learning_rate": 8.720936995153473e-06,
201409
+ "loss": 0.2893,
201410
+ "step": 80205
201411
+ },
201412
+ {
201413
+ "epoch": 641.53,
201414
+ "learning_rate": 8.72085621970921e-06,
201415
+ "loss": 0.3141,
201416
+ "step": 80210
201417
+ },
201418
+ {
201419
+ "epoch": 641.57,
201420
+ "learning_rate": 8.720775444264943e-06,
201421
+ "loss": 0.5601,
201422
+ "step": 80215
201423
+ },
201424
+ {
201425
+ "epoch": 641.61,
201426
+ "learning_rate": 8.720694668820679e-06,
201427
+ "loss": 1.1339,
201428
+ "step": 80220
201429
+ },
201430
+ {
201431
+ "epoch": 641.65,
201432
+ "learning_rate": 8.720613893376413e-06,
201433
+ "loss": 0.2912,
201434
+ "step": 80225
201435
+ },
201436
+ {
201437
+ "epoch": 641.69,
201438
+ "learning_rate": 8.720533117932149e-06,
201439
+ "loss": 0.2561,
201440
+ "step": 80230
201441
+ },
201442
+ {
201443
+ "epoch": 641.73,
201444
+ "learning_rate": 8.720452342487883e-06,
201445
+ "loss": 0.3522,
201446
+ "step": 80235
201447
+ },
201448
+ {
201449
+ "epoch": 641.77,
201450
+ "learning_rate": 8.720371567043619e-06,
201451
+ "loss": 0.5565,
201452
+ "step": 80240
201453
+ },
201454
+ {
201455
+ "epoch": 641.81,
201456
+ "learning_rate": 8.720290791599353e-06,
201457
+ "loss": 1.2682,
201458
+ "step": 80245
201459
+ },
201460
+ {
201461
+ "epoch": 641.85,
201462
+ "learning_rate": 8.720210016155089e-06,
201463
+ "loss": 0.2933,
201464
+ "step": 80250
201465
+ },
201466
+ {
201467
+ "epoch": 641.89,
201468
+ "learning_rate": 8.720129240710825e-06,
201469
+ "loss": 0.3283,
201470
+ "step": 80255
201471
+ },
201472
+ {
201473
+ "epoch": 641.93,
201474
+ "learning_rate": 8.720048465266559e-06,
201475
+ "loss": 0.3409,
201476
+ "step": 80260
201477
+ },
201478
+ {
201479
+ "epoch": 641.97,
201480
+ "learning_rate": 8.719967689822295e-06,
201481
+ "loss": 0.5759,
201482
+ "step": 80265
201483
+ },
201484
+ {
201485
+ "epoch": 642.0,
201486
+ "eval_loss": 0.43214529752731323,
201487
+ "eval_runtime": 43.1264,
201488
+ "eval_samples_per_second": 19.501,
201489
+ "eval_steps_per_second": 0.626,
201490
+ "eval_wer": 0.18905362543335547,
201491
+ "step": 80269
201492
  }
201493
  ],
201494
+ "max_steps": 625000,
201495
  "num_train_epochs": 5000,
201496
+ "total_flos": 2.258889246771459e+20,
201497
  "trial_name": null,
201498
  "trial_params": null
201499
  }
model-bin/finetune/base/{checkpoint-79647 β†’ checkpoint-80269}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629901744.9403872/events.out.tfevents.1629901744.7e498afd5545.7645.55 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34dc2a50ea0adcb3166e140d46a097ce8d11caf48de16acd58516039a1925944
3
+ size 4194
model-bin/finetune/base/log/1629902211.333175/events.out.tfevents.1629902211.7e498afd5545.7645.57 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d64ee7baa7bf328f599e5a902ce7aff625fd640180a7b6ae1da505d91481609
3
+ size 4194
model-bin/finetune/base/log/1629902676.4913468/events.out.tfevents.1629902676.7e498afd5545.7645.59 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae7fa1866134c5a1d49d534d7f3268678df59831d6c3683e29b7bf8ce8d37eab
3
+ size 4194
model-bin/finetune/base/log/1629903142.829317/events.out.tfevents.1629903142.7e498afd5545.7645.61 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a41a8cf3d2fae8ea06f813a49c44c9057db63837b1fb2c4dff7b5c666089a920
3
+ size 4194
model-bin/finetune/base/log/1629903613.046621/events.out.tfevents.1629903613.7e498afd5545.7645.63 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a89bea88e97e0a0919727e772bdf5bf53103d2ff9e0cb5bee7d87ca28fb28dc
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629901744.7e498afd5545.7645.54 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e841d0dda1f95d21f31e00fc53b7c264fb3de7fcf0abbf4dda9d1a368708ec1d
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629902211.7e498afd5545.7645.56 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01089992c000a2e2684bfc1abab5ba7dd58b1ccd453c7a0d87fe2df937261c42
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629902675.7e498afd5545.7645.58 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cdb53525493a599bbbc2df21d433acacc8dd297d8772df0924c31b6d03c9bafd
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629903142.7e498afd5545.7645.60 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad2c155d804bffe59b4cdde03efb02cc4ab51680477529c24ca48a27997986c3
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1629903613.7e498afd5545.7645.62 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3607d09f1eb06a16f8978600368ee506d04af1ecee9ef47d24458f03718a15c
3
+ size 8622