Check commited on
Commit
4dbb8f1
Β·
1 Parent(s): c3ec61b

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/trainer_state.json +792 -3
  9. model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629830879.2497265/events.out.tfevents.1629830879.c435e1c5ee04.920.201 +3 -0
  11. model-bin/finetune/base/log/1629831513.1270049/events.out.tfevents.1629831513.c435e1c5ee04.920.203 +3 -0
  12. model-bin/finetune/base/log/1629832142.3883011/events.out.tfevents.1629832142.c435e1c5ee04.920.205 +3 -0
  13. model-bin/finetune/base/log/1629832782.2191467/events.out.tfevents.1629832782.c435e1c5ee04.920.207 +3 -0
  14. model-bin/finetune/base/log/1629833409.5438683/events.out.tfevents.1629833409.c435e1c5ee04.920.209 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629830879.c435e1c5ee04.920.200 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629831513.c435e1c5ee04.920.202 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629832142.c435e1c5ee04.920.204 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629832782.c435e1c5ee04.920.206 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629833409.c435e1c5ee04.920.208 +3 -0
model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:85ea8ef6c64908f2115fcc83c3c7ba096536f425df00a3c9f1cbfc2255f20245
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4305b79063344e8dcfa70fc08a440bf763d7a5555ef19d1c29a843c47a2bdd0
3
  size 722165393
model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:989a241331db9f31c999d496faab12f6ba17620c47350944896a193165b684aa
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b864d526ba5e96bbeec05edbe0fb74e2fce15748318e767dedff344b89f04bf
3
  size 377909911
model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:06ffb88ff1db1b23a37760fd519fbca7feaf5222271d9129ca3900755d9cb924
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d3efe634557e655c20c363dfd2ec88502e52a1a2fe24613e53efd8f78e3b4d7
3
  size 14503
model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1ae34bb3b632ce84a747aa987dc0f72facb05243dbdbd4095b0555d0a6525518
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:890e7963e056aec6747d41244e17966edf587e5f50edd32309235c4d785e577b
3
  size 559
model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:13a19303204503e39131fe6893ee9e90595d5035f7349c77cb80c388b6b78e06
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84754f1abc95954e6ecb719c67cba56880d4f0682e71be6ded68f83ed242eea3
3
  size 623
model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.18588425381903642,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-60729",
4
- "epoch": 533.995983935743,
5
- "global_step": 66326,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -183753,11 +183753,800 @@
183753
  "eval_steps_per_second": 0.677,
183754
  "eval_wer": 0.19705408178220724,
183755
  "step": 66326
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
183756
  }
183757
  ],
183758
  "max_steps": 620000,
183759
  "num_train_epochs": 5000,
183760
- "total_flos": 1.866428893671497e+20,
183761
  "trial_name": null,
183762
  "trial_params": null
183763
  }
 
1
  {
2
  "best_metric": 0.18588425381903642,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-60729",
4
+ "epoch": 538.995983935743,
5
+ "global_step": 66949,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
183753
  "eval_steps_per_second": 0.677,
183754
  "eval_wer": 0.19705408178220724,
183755
  "step": 66326
183756
+ },
183757
+ {
183758
+ "epoch": 530.03,
183759
+ "learning_rate": 8.953445512820514e-06,
183760
+ "loss": 0.396,
183761
+ "step": 66330
183762
+ },
183763
+ {
183764
+ "epoch": 530.07,
183765
+ "learning_rate": 8.953365384615385e-06,
183766
+ "loss": 0.3424,
183767
+ "step": 66335
183768
+ },
183769
+ {
183770
+ "epoch": 530.11,
183771
+ "learning_rate": 8.953285256410256e-06,
183772
+ "loss": 0.3231,
183773
+ "step": 66340
183774
+ },
183775
+ {
183776
+ "epoch": 530.15,
183777
+ "learning_rate": 8.95320512820513e-06,
183778
+ "loss": 0.4369,
183779
+ "step": 66345
183780
+ },
183781
+ {
183782
+ "epoch": 530.19,
183783
+ "learning_rate": 8.953125e-06,
183784
+ "loss": 0.9247,
183785
+ "step": 66350
183786
+ },
183787
+ {
183788
+ "epoch": 530.23,
183789
+ "learning_rate": 8.953044871794872e-06,
183790
+ "loss": 0.8341,
183791
+ "step": 66355
183792
+ },
183793
+ {
183794
+ "epoch": 530.27,
183795
+ "learning_rate": 8.952964743589743e-06,
183796
+ "loss": 0.2963,
183797
+ "step": 66360
183798
+ },
183799
+ {
183800
+ "epoch": 530.31,
183801
+ "learning_rate": 8.952884615384617e-06,
183802
+ "loss": 0.3888,
183803
+ "step": 66365
183804
+ },
183805
+ {
183806
+ "epoch": 530.35,
183807
+ "learning_rate": 8.952804487179488e-06,
183808
+ "loss": 0.4379,
183809
+ "step": 66370
183810
+ },
183811
+ {
183812
+ "epoch": 530.39,
183813
+ "learning_rate": 8.95272435897436e-06,
183814
+ "loss": 0.9638,
183815
+ "step": 66375
183816
+ },
183817
+ {
183818
+ "epoch": 530.43,
183819
+ "learning_rate": 8.952644230769232e-06,
183820
+ "loss": 0.9032,
183821
+ "step": 66380
183822
+ },
183823
+ {
183824
+ "epoch": 530.47,
183825
+ "learning_rate": 8.952564102564104e-06,
183826
+ "loss": 0.299,
183827
+ "step": 66385
183828
+ },
183829
+ {
183830
+ "epoch": 530.51,
183831
+ "learning_rate": 8.952483974358975e-06,
183832
+ "loss": 0.4344,
183833
+ "step": 66390
183834
+ },
183835
+ {
183836
+ "epoch": 530.55,
183837
+ "learning_rate": 8.952403846153846e-06,
183838
+ "loss": 0.4741,
183839
+ "step": 66395
183840
+ },
183841
+ {
183842
+ "epoch": 530.59,
183843
+ "learning_rate": 8.95232371794872e-06,
183844
+ "loss": 0.968,
183845
+ "step": 66400
183846
+ },
183847
+ {
183848
+ "epoch": 530.63,
183849
+ "learning_rate": 8.95224358974359e-06,
183850
+ "loss": 0.6931,
183851
+ "step": 66405
183852
+ },
183853
+ {
183854
+ "epoch": 530.67,
183855
+ "learning_rate": 8.952163461538462e-06,
183856
+ "loss": 0.3257,
183857
+ "step": 66410
183858
+ },
183859
+ {
183860
+ "epoch": 530.71,
183861
+ "learning_rate": 8.952083333333335e-06,
183862
+ "loss": 0.3857,
183863
+ "step": 66415
183864
+ },
183865
+ {
183866
+ "epoch": 530.75,
183867
+ "learning_rate": 8.952003205128207e-06,
183868
+ "loss": 0.4299,
183869
+ "step": 66420
183870
+ },
183871
+ {
183872
+ "epoch": 530.79,
183873
+ "learning_rate": 8.951923076923078e-06,
183874
+ "loss": 1.1066,
183875
+ "step": 66425
183876
+ },
183877
+ {
183878
+ "epoch": 530.83,
183879
+ "learning_rate": 8.95184294871795e-06,
183880
+ "loss": 0.9173,
183881
+ "step": 66430
183882
+ },
183883
+ {
183884
+ "epoch": 530.87,
183885
+ "learning_rate": 8.951762820512822e-06,
183886
+ "loss": 0.2734,
183887
+ "step": 66435
183888
+ },
183889
+ {
183890
+ "epoch": 530.91,
183891
+ "learning_rate": 8.951682692307692e-06,
183892
+ "loss": 0.3241,
183893
+ "step": 66440
183894
+ },
183895
+ {
183896
+ "epoch": 530.95,
183897
+ "learning_rate": 8.951602564102565e-06,
183898
+ "loss": 0.4544,
183899
+ "step": 66445
183900
+ },
183901
+ {
183902
+ "epoch": 530.99,
183903
+ "learning_rate": 8.951522435897436e-06,
183904
+ "loss": 1.0191,
183905
+ "step": 66450
183906
+ },
183907
+ {
183908
+ "epoch": 531.0,
183909
+ "eval_loss": 0.406460165977478,
183910
+ "eval_runtime": 40.1796,
183911
+ "eval_samples_per_second": 20.831,
183912
+ "eval_steps_per_second": 0.672,
183913
+ "eval_wer": 0.19212436752951528,
183914
+ "step": 66451
183915
+ },
183916
+ {
183917
+ "epoch": 531.03,
183918
+ "learning_rate": 8.951442307692308e-06,
183919
+ "loss": 0.6888,
183920
+ "step": 66455
183921
+ },
183922
+ {
183923
+ "epoch": 531.07,
183924
+ "learning_rate": 8.951362179487179e-06,
183925
+ "loss": 0.3343,
183926
+ "step": 66460
183927
+ },
183928
+ {
183929
+ "epoch": 531.11,
183930
+ "learning_rate": 8.951282051282052e-06,
183931
+ "loss": 0.3268,
183932
+ "step": 66465
183933
+ },
183934
+ {
183935
+ "epoch": 531.15,
183936
+ "learning_rate": 8.951201923076924e-06,
183937
+ "loss": 0.3481,
183938
+ "step": 66470
183939
+ },
183940
+ {
183941
+ "epoch": 531.19,
183942
+ "learning_rate": 8.951121794871795e-06,
183943
+ "loss": 0.9568,
183944
+ "step": 66475
183945
+ },
183946
+ {
183947
+ "epoch": 531.23,
183948
+ "learning_rate": 8.951041666666668e-06,
183949
+ "loss": 0.6338,
183950
+ "step": 66480
183951
+ },
183952
+ {
183953
+ "epoch": 531.27,
183954
+ "learning_rate": 8.95096153846154e-06,
183955
+ "loss": 0.3239,
183956
+ "step": 66485
183957
+ },
183958
+ {
183959
+ "epoch": 531.31,
183960
+ "learning_rate": 8.95088141025641e-06,
183961
+ "loss": 0.3678,
183962
+ "step": 66490
183963
+ },
183964
+ {
183965
+ "epoch": 531.35,
183966
+ "learning_rate": 8.950801282051282e-06,
183967
+ "loss": 0.4088,
183968
+ "step": 66495
183969
+ },
183970
+ {
183971
+ "epoch": 531.39,
183972
+ "learning_rate": 8.950721153846155e-06,
183973
+ "loss": 0.9323,
183974
+ "step": 66500
183975
+ },
183976
+ {
183977
+ "epoch": 531.43,
183978
+ "learning_rate": 8.950641025641026e-06,
183979
+ "loss": 0.6159,
183980
+ "step": 66505
183981
+ },
183982
+ {
183983
+ "epoch": 531.47,
183984
+ "learning_rate": 8.950560897435898e-06,
183985
+ "loss": 0.3322,
183986
+ "step": 66510
183987
+ },
183988
+ {
183989
+ "epoch": 531.51,
183990
+ "learning_rate": 8.950480769230771e-06,
183991
+ "loss": 0.3244,
183992
+ "step": 66515
183993
+ },
183994
+ {
183995
+ "epoch": 531.55,
183996
+ "learning_rate": 8.950400641025642e-06,
183997
+ "loss": 0.4675,
183998
+ "step": 66520
183999
+ },
184000
+ {
184001
+ "epoch": 531.59,
184002
+ "learning_rate": 8.950320512820514e-06,
184003
+ "loss": 0.8916,
184004
+ "step": 66525
184005
+ },
184006
+ {
184007
+ "epoch": 531.63,
184008
+ "learning_rate": 8.950240384615385e-06,
184009
+ "loss": 0.7,
184010
+ "step": 66530
184011
+ },
184012
+ {
184013
+ "epoch": 531.67,
184014
+ "learning_rate": 8.950160256410258e-06,
184015
+ "loss": 0.2958,
184016
+ "step": 66535
184017
+ },
184018
+ {
184019
+ "epoch": 531.71,
184020
+ "learning_rate": 8.95008012820513e-06,
184021
+ "loss": 0.3249,
184022
+ "step": 66540
184023
+ },
184024
+ {
184025
+ "epoch": 531.75,
184026
+ "learning_rate": 8.95e-06,
184027
+ "loss": 0.4388,
184028
+ "step": 66545
184029
+ },
184030
+ {
184031
+ "epoch": 531.79,
184032
+ "learning_rate": 8.949919871794872e-06,
184033
+ "loss": 0.877,
184034
+ "step": 66550
184035
+ },
184036
+ {
184037
+ "epoch": 531.83,
184038
+ "learning_rate": 8.949839743589745e-06,
184039
+ "loss": 0.6828,
184040
+ "step": 66555
184041
+ },
184042
+ {
184043
+ "epoch": 531.87,
184044
+ "learning_rate": 8.949759615384615e-06,
184045
+ "loss": 0.3371,
184046
+ "step": 66560
184047
+ },
184048
+ {
184049
+ "epoch": 531.91,
184050
+ "learning_rate": 8.949679487179488e-06,
184051
+ "loss": 0.344,
184052
+ "step": 66565
184053
+ },
184054
+ {
184055
+ "epoch": 531.95,
184056
+ "learning_rate": 8.949599358974361e-06,
184057
+ "loss": 0.4809,
184058
+ "step": 66570
184059
+ },
184060
+ {
184061
+ "epoch": 531.99,
184062
+ "learning_rate": 8.94951923076923e-06,
184063
+ "loss": 1.0281,
184064
+ "step": 66575
184065
+ },
184066
+ {
184067
+ "epoch": 532.0,
184068
+ "eval_loss": 0.4383997619152069,
184069
+ "eval_runtime": 39.6825,
184070
+ "eval_samples_per_second": 21.092,
184071
+ "eval_steps_per_second": 0.68,
184072
+ "eval_wer": 0.20101062643977113,
184073
+ "step": 66576
184074
+ },
184075
+ {
184076
+ "epoch": 532.03,
184077
+ "learning_rate": 8.949439102564104e-06,
184078
+ "loss": 0.3595,
184079
+ "step": 66580
184080
+ },
184081
+ {
184082
+ "epoch": 532.07,
184083
+ "learning_rate": 8.949358974358975e-06,
184084
+ "loss": 0.3205,
184085
+ "step": 66585
184086
+ },
184087
+ {
184088
+ "epoch": 532.11,
184089
+ "learning_rate": 8.949278846153846e-06,
184090
+ "loss": 0.2612,
184091
+ "step": 66590
184092
+ },
184093
+ {
184094
+ "epoch": 532.15,
184095
+ "learning_rate": 8.949198717948718e-06,
184096
+ "loss": 0.4,
184097
+ "step": 66595
184098
+ },
184099
+ {
184100
+ "epoch": 532.19,
184101
+ "learning_rate": 8.94911858974359e-06,
184102
+ "loss": 0.9481,
184103
+ "step": 66600
184104
+ },
184105
+ {
184106
+ "epoch": 532.23,
184107
+ "learning_rate": 8.949038461538462e-06,
184108
+ "loss": 0.7302,
184109
+ "step": 66605
184110
+ },
184111
+ {
184112
+ "epoch": 532.27,
184113
+ "learning_rate": 8.948958333333333e-06,
184114
+ "loss": 0.3677,
184115
+ "step": 66610
184116
+ },
184117
+ {
184118
+ "epoch": 532.31,
184119
+ "learning_rate": 8.948878205128207e-06,
184120
+ "loss": 0.3763,
184121
+ "step": 66615
184122
+ },
184123
+ {
184124
+ "epoch": 532.35,
184125
+ "learning_rate": 8.948798076923078e-06,
184126
+ "loss": 0.4758,
184127
+ "step": 66620
184128
+ },
184129
+ {
184130
+ "epoch": 532.39,
184131
+ "learning_rate": 8.94871794871795e-06,
184132
+ "loss": 0.9979,
184133
+ "step": 66625
184134
+ },
184135
+ {
184136
+ "epoch": 532.43,
184137
+ "learning_rate": 8.94863782051282e-06,
184138
+ "loss": 0.7855,
184139
+ "step": 66630
184140
+ },
184141
+ {
184142
+ "epoch": 532.47,
184143
+ "learning_rate": 8.948557692307694e-06,
184144
+ "loss": 0.318,
184145
+ "step": 66635
184146
+ },
184147
+ {
184148
+ "epoch": 532.51,
184149
+ "learning_rate": 8.948477564102565e-06,
184150
+ "loss": 0.3541,
184151
+ "step": 66640
184152
+ },
184153
+ {
184154
+ "epoch": 532.55,
184155
+ "learning_rate": 8.948397435897436e-06,
184156
+ "loss": 0.406,
184157
+ "step": 66645
184158
+ },
184159
+ {
184160
+ "epoch": 532.59,
184161
+ "learning_rate": 8.948317307692308e-06,
184162
+ "loss": 1.0155,
184163
+ "step": 66650
184164
+ },
184165
+ {
184166
+ "epoch": 532.63,
184167
+ "learning_rate": 8.94823717948718e-06,
184168
+ "loss": 0.8103,
184169
+ "step": 66655
184170
+ },
184171
+ {
184172
+ "epoch": 532.67,
184173
+ "learning_rate": 8.948157051282052e-06,
184174
+ "loss": 0.3239,
184175
+ "step": 66660
184176
+ },
184177
+ {
184178
+ "epoch": 532.71,
184179
+ "learning_rate": 8.948076923076923e-06,
184180
+ "loss": 0.3431,
184181
+ "step": 66665
184182
+ },
184183
+ {
184184
+ "epoch": 532.75,
184185
+ "learning_rate": 8.947996794871797e-06,
184186
+ "loss": 0.4256,
184187
+ "step": 66670
184188
+ },
184189
+ {
184190
+ "epoch": 532.79,
184191
+ "learning_rate": 8.947916666666668e-06,
184192
+ "loss": 0.8556,
184193
+ "step": 66675
184194
+ },
184195
+ {
184196
+ "epoch": 532.83,
184197
+ "learning_rate": 8.94783653846154e-06,
184198
+ "loss": 0.6975,
184199
+ "step": 66680
184200
+ },
184201
+ {
184202
+ "epoch": 532.87,
184203
+ "learning_rate": 8.94775641025641e-06,
184204
+ "loss": 0.2859,
184205
+ "step": 66685
184206
+ },
184207
+ {
184208
+ "epoch": 532.91,
184209
+ "learning_rate": 8.947676282051284e-06,
184210
+ "loss": 0.3586,
184211
+ "step": 66690
184212
+ },
184213
+ {
184214
+ "epoch": 532.95,
184215
+ "learning_rate": 8.947596153846155e-06,
184216
+ "loss": 0.4585,
184217
+ "step": 66695
184218
+ },
184219
+ {
184220
+ "epoch": 532.99,
184221
+ "learning_rate": 8.947516025641026e-06,
184222
+ "loss": 0.9764,
184223
+ "step": 66700
184224
+ },
184225
+ {
184226
+ "epoch": 533.0,
184227
+ "eval_loss": 0.3940429985523224,
184228
+ "eval_runtime": 40.3665,
184229
+ "eval_samples_per_second": 20.71,
184230
+ "eval_steps_per_second": 0.669,
184231
+ "eval_wer": 0.18733183041233364,
184232
+ "step": 66701
184233
+ },
184234
+ {
184235
+ "epoch": 537.03,
184236
+ "learning_rate": 8.947435897435898e-06,
184237
+ "loss": 0.3689,
184238
+ "step": 66705
184239
+ },
184240
+ {
184241
+ "epoch": 537.07,
184242
+ "learning_rate": 8.94735576923077e-06,
184243
+ "loss": 0.2814,
184244
+ "step": 66710
184245
+ },
184246
+ {
184247
+ "epoch": 537.11,
184248
+ "learning_rate": 8.947275641025642e-06,
184249
+ "loss": 0.3363,
184250
+ "step": 66715
184251
+ },
184252
+ {
184253
+ "epoch": 537.15,
184254
+ "learning_rate": 8.947195512820514e-06,
184255
+ "loss": 0.4299,
184256
+ "step": 66720
184257
+ },
184258
+ {
184259
+ "epoch": 537.19,
184260
+ "learning_rate": 8.947115384615387e-06,
184261
+ "loss": 1.1418,
184262
+ "step": 66725
184263
+ },
184264
+ {
184265
+ "epoch": 537.23,
184266
+ "learning_rate": 8.947035256410256e-06,
184267
+ "loss": 0.6762,
184268
+ "step": 66730
184269
+ },
184270
+ {
184271
+ "epoch": 537.27,
184272
+ "learning_rate": 8.94695512820513e-06,
184273
+ "loss": 0.2862,
184274
+ "step": 66735
184275
+ },
184276
+ {
184277
+ "epoch": 537.31,
184278
+ "learning_rate": 8.946875e-06,
184279
+ "loss": 0.3495,
184280
+ "step": 66740
184281
+ },
184282
+ {
184283
+ "epoch": 537.35,
184284
+ "learning_rate": 8.946794871794872e-06,
184285
+ "loss": 0.4347,
184286
+ "step": 66745
184287
+ },
184288
+ {
184289
+ "epoch": 537.39,
184290
+ "learning_rate": 8.946714743589743e-06,
184291
+ "loss": 0.9885,
184292
+ "step": 66750
184293
+ },
184294
+ {
184295
+ "epoch": 537.43,
184296
+ "learning_rate": 8.946634615384616e-06,
184297
+ "loss": 0.8602,
184298
+ "step": 66755
184299
+ },
184300
+ {
184301
+ "epoch": 537.47,
184302
+ "learning_rate": 8.946554487179488e-06,
184303
+ "loss": 0.3301,
184304
+ "step": 66760
184305
+ },
184306
+ {
184307
+ "epoch": 537.51,
184308
+ "learning_rate": 8.946474358974359e-06,
184309
+ "loss": 0.3302,
184310
+ "step": 66765
184311
+ },
184312
+ {
184313
+ "epoch": 537.55,
184314
+ "learning_rate": 8.946394230769232e-06,
184315
+ "loss": 0.4187,
184316
+ "step": 66770
184317
+ },
184318
+ {
184319
+ "epoch": 537.59,
184320
+ "learning_rate": 8.946314102564104e-06,
184321
+ "loss": 1.0443,
184322
+ "step": 66775
184323
+ },
184324
+ {
184325
+ "epoch": 537.63,
184326
+ "learning_rate": 8.946233974358975e-06,
184327
+ "loss": 0.6987,
184328
+ "step": 66780
184329
+ },
184330
+ {
184331
+ "epoch": 537.67,
184332
+ "learning_rate": 8.946153846153846e-06,
184333
+ "loss": 0.2443,
184334
+ "step": 66785
184335
+ },
184336
+ {
184337
+ "epoch": 537.71,
184338
+ "learning_rate": 8.94607371794872e-06,
184339
+ "loss": 0.3445,
184340
+ "step": 66790
184341
+ },
184342
+ {
184343
+ "epoch": 537.76,
184344
+ "learning_rate": 8.94599358974359e-06,
184345
+ "loss": 0.4564,
184346
+ "step": 66795
184347
+ },
184348
+ {
184349
+ "epoch": 537.8,
184350
+ "learning_rate": 8.945913461538462e-06,
184351
+ "loss": 0.9662,
184352
+ "step": 66800
184353
+ },
184354
+ {
184355
+ "epoch": 537.84,
184356
+ "learning_rate": 8.945833333333333e-06,
184357
+ "loss": 0.7671,
184358
+ "step": 66805
184359
+ },
184360
+ {
184361
+ "epoch": 537.88,
184362
+ "learning_rate": 8.945753205128206e-06,
184363
+ "loss": 0.3258,
184364
+ "step": 66810
184365
+ },
184366
+ {
184367
+ "epoch": 537.92,
184368
+ "learning_rate": 8.945673076923078e-06,
184369
+ "loss": 0.3372,
184370
+ "step": 66815
184371
+ },
184372
+ {
184373
+ "epoch": 537.96,
184374
+ "learning_rate": 8.945592948717949e-06,
184375
+ "loss": 0.4195,
184376
+ "step": 66820
184377
+ },
184378
+ {
184379
+ "epoch": 538.0,
184380
+ "learning_rate": 8.945512820512822e-06,
184381
+ "loss": 1.161,
184382
+ "step": 66825
184383
+ },
184384
+ {
184385
+ "epoch": 538.0,
184386
+ "eval_loss": 0.3947887122631073,
184387
+ "eval_runtime": 39.1445,
184388
+ "eval_samples_per_second": 21.357,
184389
+ "eval_steps_per_second": 0.69,
184390
+ "eval_wer": 0.19173909813535397,
184391
+ "step": 66825
184392
+ },
184393
+ {
184394
+ "epoch": 538.04,
184395
+ "learning_rate": 8.945432692307694e-06,
184396
+ "loss": 0.3732,
184397
+ "step": 66830
184398
+ },
184399
+ {
184400
+ "epoch": 538.08,
184401
+ "learning_rate": 8.945352564102565e-06,
184402
+ "loss": 0.3088,
184403
+ "step": 66835
184404
+ },
184405
+ {
184406
+ "epoch": 538.12,
184407
+ "learning_rate": 8.945272435897436e-06,
184408
+ "loss": 0.3213,
184409
+ "step": 66840
184410
+ },
184411
+ {
184412
+ "epoch": 538.16,
184413
+ "learning_rate": 8.94519230769231e-06,
184414
+ "loss": 0.5747,
184415
+ "step": 66845
184416
+ },
184417
+ {
184418
+ "epoch": 538.2,
184419
+ "learning_rate": 8.945112179487179e-06,
184420
+ "loss": 1.2296,
184421
+ "step": 66850
184422
+ },
184423
+ {
184424
+ "epoch": 538.24,
184425
+ "learning_rate": 8.945032051282052e-06,
184426
+ "loss": 0.3106,
184427
+ "step": 66855
184428
+ },
184429
+ {
184430
+ "epoch": 538.28,
184431
+ "learning_rate": 8.944951923076923e-06,
184432
+ "loss": 0.2994,
184433
+ "step": 66860
184434
+ },
184435
+ {
184436
+ "epoch": 538.32,
184437
+ "learning_rate": 8.944871794871795e-06,
184438
+ "loss": 0.3399,
184439
+ "step": 66865
184440
+ },
184441
+ {
184442
+ "epoch": 538.36,
184443
+ "learning_rate": 8.944791666666668e-06,
184444
+ "loss": 0.4754,
184445
+ "step": 66870
184446
+ },
184447
+ {
184448
+ "epoch": 538.4,
184449
+ "learning_rate": 8.94471153846154e-06,
184450
+ "loss": 1.396,
184451
+ "step": 66875
184452
+ },
184453
+ {
184454
+ "epoch": 538.44,
184455
+ "learning_rate": 8.94463141025641e-06,
184456
+ "loss": 0.3324,
184457
+ "step": 66880
184458
+ },
184459
+ {
184460
+ "epoch": 538.48,
184461
+ "learning_rate": 8.944551282051282e-06,
184462
+ "loss": 0.2814,
184463
+ "step": 66885
184464
+ },
184465
+ {
184466
+ "epoch": 538.52,
184467
+ "learning_rate": 8.944471153846155e-06,
184468
+ "loss": 0.3601,
184469
+ "step": 66890
184470
+ },
184471
+ {
184472
+ "epoch": 538.56,
184473
+ "learning_rate": 8.944391025641026e-06,
184474
+ "loss": 0.5061,
184475
+ "step": 66895
184476
+ },
184477
+ {
184478
+ "epoch": 538.6,
184479
+ "learning_rate": 8.944310897435898e-06,
184480
+ "loss": 1.241,
184481
+ "step": 66900
184482
+ },
184483
+ {
184484
+ "epoch": 538.64,
184485
+ "learning_rate": 8.944230769230769e-06,
184486
+ "loss": 0.3642,
184487
+ "step": 66905
184488
+ },
184489
+ {
184490
+ "epoch": 538.68,
184491
+ "learning_rate": 8.944150641025642e-06,
184492
+ "loss": 0.3311,
184493
+ "step": 66910
184494
+ },
184495
+ {
184496
+ "epoch": 538.72,
184497
+ "learning_rate": 8.944070512820513e-06,
184498
+ "loss": 0.3155,
184499
+ "step": 66915
184500
+ },
184501
+ {
184502
+ "epoch": 538.76,
184503
+ "learning_rate": 8.943990384615385e-06,
184504
+ "loss": 0.4952,
184505
+ "step": 66920
184506
+ },
184507
+ {
184508
+ "epoch": 538.8,
184509
+ "learning_rate": 8.943910256410258e-06,
184510
+ "loss": 1.324,
184511
+ "step": 66925
184512
+ },
184513
+ {
184514
+ "epoch": 538.84,
184515
+ "learning_rate": 8.94383012820513e-06,
184516
+ "loss": 0.3858,
184517
+ "step": 66930
184518
+ },
184519
+ {
184520
+ "epoch": 538.88,
184521
+ "learning_rate": 8.94375e-06,
184522
+ "loss": 0.3126,
184523
+ "step": 66935
184524
+ },
184525
+ {
184526
+ "epoch": 538.92,
184527
+ "learning_rate": 8.943669871794872e-06,
184528
+ "loss": 0.4153,
184529
+ "step": 66940
184530
+ },
184531
+ {
184532
+ "epoch": 538.96,
184533
+ "learning_rate": 8.943589743589745e-06,
184534
+ "loss": 0.4908,
184535
+ "step": 66945
184536
+ },
184537
+ {
184538
+ "epoch": 539.0,
184539
+ "eval_loss": 0.4666649401187897,
184540
+ "eval_runtime": 39.7494,
184541
+ "eval_samples_per_second": 21.032,
184542
+ "eval_steps_per_second": 0.679,
184543
+ "eval_wer": 0.20303119051105578,
184544
+ "step": 66949
184545
  }
184546
  ],
184547
  "max_steps": 620000,
184548
  "num_train_epochs": 5000,
184549
+ "total_flos": 1.8839674795561802e+20,
184550
  "trial_name": null,
184551
  "trial_params": null
184552
  }
model-bin/finetune/base/{checkpoint-66326 β†’ checkpoint-66949}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629830879.2497265/events.out.tfevents.1629830879.c435e1c5ee04.920.201 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b75676a0a25dc896e7207461989b0632eafc614c13a359f7aaef1cf96c72192
3
+ size 4194
model-bin/finetune/base/log/1629831513.1270049/events.out.tfevents.1629831513.c435e1c5ee04.920.203 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95578367bfe0c98bf978e589caeb48f0ebcf8a828aad2c575fd43f3fd2b30c20
3
+ size 4194
model-bin/finetune/base/log/1629832142.3883011/events.out.tfevents.1629832142.c435e1c5ee04.920.205 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e86a97f637df3a6ec1ece5aafd69d2db481f19c1cd9058e34b9dc464299cc6d4
3
+ size 4194
model-bin/finetune/base/log/1629832782.2191467/events.out.tfevents.1629832782.c435e1c5ee04.920.207 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c93212efcd7f1a4db36f67a8d30d53dbac379210090f1a86635a30adef0a2d56
3
+ size 4194
model-bin/finetune/base/log/1629833409.5438683/events.out.tfevents.1629833409.c435e1c5ee04.920.209 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e6ef65c3871d02573b75fa0b9d0d3435d29ce67c10cfbce24f39e9a0b091ff1
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629830879.c435e1c5ee04.920.200 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25291100f3c75456226310622eb0631232e9a2e31acc0b8637445bcc4e620426
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629831513.c435e1c5ee04.920.202 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01f1bbbe487a27e4ced6f72dd38a0e369926931e3eedf287f6df34593805cf2d
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629832142.c435e1c5ee04.920.204 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6a824752ad2615bbf723f157a7bd782eb94bbdd1ad3c483b26b63c499a87e58
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629832782.c435e1c5ee04.920.206 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8371499d74a7d558a9e71f461c29889769a702010dc01b552c2da272d678e078
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629833409.c435e1c5ee04.920.208 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d32682a66d7ceb182068ad415fb4e5a85eea63b4994c01e34fb4be2a1fc4fc6e
3
+ size 8462