Check commited on
Commit
4ada2a4
Β·
1 Parent(s): b9edd56

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/rng_state.pth +2 -2
  6. model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/trainer_state.json +793 -4
  9. model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630148780.943968/events.out.tfevents.1630148780.86bb0ddabf9b.4092.131 +3 -0
  11. model-bin/finetune/base/log/1630149168.480171/events.out.tfevents.1630149168.86bb0ddabf9b.4092.133 +3 -0
  12. model-bin/finetune/base/log/1630149553.0608108/events.out.tfevents.1630149553.86bb0ddabf9b.4092.135 +3 -0
  13. model-bin/finetune/base/log/1630149942.7762349/events.out.tfevents.1630149942.86bb0ddabf9b.4092.137 +3 -0
  14. model-bin/finetune/base/log/1630150333.0672135/events.out.tfevents.1630150333.86bb0ddabf9b.4092.139 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630148780.86bb0ddabf9b.4092.130 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630149168.86bb0ddabf9b.4092.132 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630149553.86bb0ddabf9b.4092.134 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630149942.86bb0ddabf9b.4092.136 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630150333.86bb0ddabf9b.4092.138 +3 -0
model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:16b005e9fb69b2e7c668f7d9d1ddcf9502c134575ee756c2b4dd2a09e4f3dfc7
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3845dc63451d0a1c44498d69f0b827119c42d6e557cf97a7f9735c52705644bd
3
  size 722165393
model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2cfc721d81f03e5e16e7e7c9c5142ddddbe2c39f6609dd7fe74fbe15bffd4bf7
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69b6d80c964bf8059c1020726fed10f992c3b777c5f499f82bdede9d7ca4ced7
3
  size 377909911
model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:499abeb83f7ca59edfcad7e7d1e42dccc904e7481abcc4da5c71266773800888
3
- size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dfcfe608f0a02dd951bb37028335198611a4d183fa2e67bc1b1aed79918029b
3
+ size 14567
model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:683ccffd7adf3cb2b0b65666beebc8e51b2ae5066df9b1b2e36f844178028c85
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f7595729f26a675659d761b684c35ca3b092bd8547ff8d0cfbbdd1d52fc1473
3
  size 559
model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:149ab5492f8d36be18b46436a0461744c2fef1b202cb12894209dd371443a41d
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4e46dd1d0e522e3ca1340174ec5347c27a555d6146d2d14b88e929a99348346
3
  size 623
model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.17162025681719809,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-124947",
4
- "epoch": 1008.0,
5
- "global_step": 126068,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -259761,11 +259761,800 @@
259761
  "eval_steps_per_second": 0.739,
259762
  "eval_wer": 0.18030087568295786,
259763
  "step": 126068
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
259764
  }
259765
  ],
259766
- "max_steps": 625000,
259767
  "num_train_epochs": 5000,
259768
- "total_flos": 3.547605377048925e+20,
259769
  "trial_name": null,
259770
  "trial_params": null
259771
  }
 
1
  {
2
  "best_metric": 0.17162025681719809,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-124947",
4
+ "epoch": 1020.995983935743,
5
+ "global_step": 126689,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
259761
  "eval_steps_per_second": 0.739,
259762
  "eval_wer": 0.18030087568295786,
259763
  "step": 126068
259764
+ },
259765
+ {
259766
+ "epoch": 1016.02,
259767
+ "learning_rate": 7.996458333333334e-06,
259768
+ "loss": 0.3939,
259769
+ "step": 126070
259770
+ },
259771
+ {
259772
+ "epoch": 1016.06,
259773
+ "learning_rate": 7.996378205128205e-06,
259774
+ "loss": 0.2661,
259775
+ "step": 126075
259776
+ },
259777
+ {
259778
+ "epoch": 1016.1,
259779
+ "learning_rate": 7.996298076923078e-06,
259780
+ "loss": 0.2543,
259781
+ "step": 126080
259782
+ },
259783
+ {
259784
+ "epoch": 1016.14,
259785
+ "learning_rate": 7.99621794871795e-06,
259786
+ "loss": 0.3273,
259787
+ "step": 126085
259788
+ },
259789
+ {
259790
+ "epoch": 1016.18,
259791
+ "learning_rate": 7.99613782051282e-06,
259792
+ "loss": 0.5625,
259793
+ "step": 126090
259794
+ },
259795
+ {
259796
+ "epoch": 1016.22,
259797
+ "learning_rate": 7.996057692307692e-06,
259798
+ "loss": 1.0146,
259799
+ "step": 126095
259800
+ },
259801
+ {
259802
+ "epoch": 1016.26,
259803
+ "learning_rate": 7.995977564102565e-06,
259804
+ "loss": 0.4505,
259805
+ "step": 126100
259806
+ },
259807
+ {
259808
+ "epoch": 1016.3,
259809
+ "learning_rate": 7.995897435897437e-06,
259810
+ "loss": 0.3086,
259811
+ "step": 126105
259812
+ },
259813
+ {
259814
+ "epoch": 1016.34,
259815
+ "learning_rate": 7.995817307692308e-06,
259816
+ "loss": 0.3847,
259817
+ "step": 126110
259818
+ },
259819
+ {
259820
+ "epoch": 1016.38,
259821
+ "learning_rate": 7.99573717948718e-06,
259822
+ "loss": 0.5578,
259823
+ "step": 126115
259824
+ },
259825
+ {
259826
+ "epoch": 1016.42,
259827
+ "learning_rate": 7.995657051282052e-06,
259828
+ "loss": 0.9584,
259829
+ "step": 126120
259830
+ },
259831
+ {
259832
+ "epoch": 1016.46,
259833
+ "learning_rate": 7.995576923076924e-06,
259834
+ "loss": 0.307,
259835
+ "step": 126125
259836
+ },
259837
+ {
259838
+ "epoch": 1016.5,
259839
+ "learning_rate": 7.995496794871795e-06,
259840
+ "loss": 0.267,
259841
+ "step": 126130
259842
+ },
259843
+ {
259844
+ "epoch": 1016.54,
259845
+ "learning_rate": 7.995416666666668e-06,
259846
+ "loss": 0.2988,
259847
+ "step": 126135
259848
+ },
259849
+ {
259850
+ "epoch": 1016.58,
259851
+ "learning_rate": 7.99533653846154e-06,
259852
+ "loss": 0.6387,
259853
+ "step": 126140
259854
+ },
259855
+ {
259856
+ "epoch": 1016.62,
259857
+ "learning_rate": 7.99525641025641e-06,
259858
+ "loss": 1.0086,
259859
+ "step": 126145
259860
+ },
259861
+ {
259862
+ "epoch": 1016.66,
259863
+ "learning_rate": 7.995176282051282e-06,
259864
+ "loss": 0.2825,
259865
+ "step": 126150
259866
+ },
259867
+ {
259868
+ "epoch": 1016.7,
259869
+ "learning_rate": 7.995096153846155e-06,
259870
+ "loss": 0.2735,
259871
+ "step": 126155
259872
+ },
259873
+ {
259874
+ "epoch": 1016.74,
259875
+ "learning_rate": 7.995016025641027e-06,
259876
+ "loss": 0.3921,
259877
+ "step": 126160
259878
+ },
259879
+ {
259880
+ "epoch": 1016.78,
259881
+ "learning_rate": 7.994935897435898e-06,
259882
+ "loss": 0.5328,
259883
+ "step": 126165
259884
+ },
259885
+ {
259886
+ "epoch": 1016.82,
259887
+ "learning_rate": 7.99485576923077e-06,
259888
+ "loss": 0.9327,
259889
+ "step": 126170
259890
+ },
259891
+ {
259892
+ "epoch": 1016.86,
259893
+ "learning_rate": 7.994775641025642e-06,
259894
+ "loss": 0.2889,
259895
+ "step": 126175
259896
+ },
259897
+ {
259898
+ "epoch": 1016.9,
259899
+ "learning_rate": 7.994695512820514e-06,
259900
+ "loss": 0.2814,
259901
+ "step": 126180
259902
+ },
259903
+ {
259904
+ "epoch": 1016.94,
259905
+ "learning_rate": 7.994615384615385e-06,
259906
+ "loss": 0.3506,
259907
+ "step": 126185
259908
+ },
259909
+ {
259910
+ "epoch": 1016.98,
259911
+ "learning_rate": 7.994535256410258e-06,
259912
+ "loss": 0.7309,
259913
+ "step": 126190
259914
+ },
259915
+ {
259916
+ "epoch": 1017.0,
259917
+ "eval_loss": 0.31492879986763,
259918
+ "eval_runtime": 36.0348,
259919
+ "eval_samples_per_second": 23.339,
259920
+ "eval_steps_per_second": 0.749,
259921
+ "eval_wer": 0.17529711375212223,
259922
+ "step": 126192
259923
+ },
259924
+ {
259925
+ "epoch": 1017.02,
259926
+ "learning_rate": 7.994455128205128e-06,
259927
+ "loss": 0.3466,
259928
+ "step": 126195
259929
+ },
259930
+ {
259931
+ "epoch": 1017.06,
259932
+ "learning_rate": 7.994375000000001e-06,
259933
+ "loss": 0.2535,
259934
+ "step": 126200
259935
+ },
259936
+ {
259937
+ "epoch": 1017.1,
259938
+ "learning_rate": 7.994294871794872e-06,
259939
+ "loss": 0.2802,
259940
+ "step": 126205
259941
+ },
259942
+ {
259943
+ "epoch": 1017.14,
259944
+ "learning_rate": 7.994214743589744e-06,
259945
+ "loss": 0.3687,
259946
+ "step": 126210
259947
+ },
259948
+ {
259949
+ "epoch": 1017.18,
259950
+ "learning_rate": 7.994134615384615e-06,
259951
+ "loss": 0.811,
259952
+ "step": 126215
259953
+ },
259954
+ {
259955
+ "epoch": 1017.22,
259956
+ "learning_rate": 7.994054487179488e-06,
259957
+ "loss": 0.7224,
259958
+ "step": 126220
259959
+ },
259960
+ {
259961
+ "epoch": 1017.27,
259962
+ "learning_rate": 7.99397435897436e-06,
259963
+ "loss": 0.2839,
259964
+ "step": 126225
259965
+ },
259966
+ {
259967
+ "epoch": 1017.31,
259968
+ "learning_rate": 7.99389423076923e-06,
259969
+ "loss": 0.2759,
259970
+ "step": 126230
259971
+ },
259972
+ {
259973
+ "epoch": 1017.35,
259974
+ "learning_rate": 7.993814102564104e-06,
259975
+ "loss": 0.4042,
259976
+ "step": 126235
259977
+ },
259978
+ {
259979
+ "epoch": 1017.39,
259980
+ "learning_rate": 7.993733974358975e-06,
259981
+ "loss": 0.7433,
259982
+ "step": 126240
259983
+ },
259984
+ {
259985
+ "epoch": 1017.43,
259986
+ "learning_rate": 7.993653846153846e-06,
259987
+ "loss": 0.84,
259988
+ "step": 126245
259989
+ },
259990
+ {
259991
+ "epoch": 1017.47,
259992
+ "learning_rate": 7.993573717948718e-06,
259993
+ "loss": 0.2865,
259994
+ "step": 126250
259995
+ },
259996
+ {
259997
+ "epoch": 1017.51,
259998
+ "learning_rate": 7.993493589743591e-06,
259999
+ "loss": 0.3096,
260000
+ "step": 126255
260001
+ },
260002
+ {
260003
+ "epoch": 1017.55,
260004
+ "learning_rate": 7.993413461538462e-06,
260005
+ "loss": 0.4599,
260006
+ "step": 126260
260007
+ },
260008
+ {
260009
+ "epoch": 1017.59,
260010
+ "learning_rate": 7.993333333333334e-06,
260011
+ "loss": 0.7373,
260012
+ "step": 126265
260013
+ },
260014
+ {
260015
+ "epoch": 1017.63,
260016
+ "learning_rate": 7.993253205128205e-06,
260017
+ "loss": 0.9646,
260018
+ "step": 126270
260019
+ },
260020
+ {
260021
+ "epoch": 1017.67,
260022
+ "learning_rate": 7.993173076923078e-06,
260023
+ "loss": 0.2937,
260024
+ "step": 126275
260025
+ },
260026
+ {
260027
+ "epoch": 1017.71,
260028
+ "learning_rate": 7.99309294871795e-06,
260029
+ "loss": 0.2543,
260030
+ "step": 126280
260031
+ },
260032
+ {
260033
+ "epoch": 1017.75,
260034
+ "learning_rate": 7.99301282051282e-06,
260035
+ "loss": 0.3509,
260036
+ "step": 126285
260037
+ },
260038
+ {
260039
+ "epoch": 1017.79,
260040
+ "learning_rate": 7.992932692307694e-06,
260041
+ "loss": 0.7394,
260042
+ "step": 126290
260043
+ },
260044
+ {
260045
+ "epoch": 1017.83,
260046
+ "learning_rate": 7.992852564102565e-06,
260047
+ "loss": 0.8648,
260048
+ "step": 126295
260049
+ },
260050
+ {
260051
+ "epoch": 1017.87,
260052
+ "learning_rate": 7.992772435897436e-06,
260053
+ "loss": 0.4316,
260054
+ "step": 126300
260055
+ },
260056
+ {
260057
+ "epoch": 1017.91,
260058
+ "learning_rate": 7.992692307692308e-06,
260059
+ "loss": 0.307,
260060
+ "step": 126305
260061
+ },
260062
+ {
260063
+ "epoch": 1017.95,
260064
+ "learning_rate": 7.992612179487181e-06,
260065
+ "loss": 0.4597,
260066
+ "step": 126310
260067
+ },
260068
+ {
260069
+ "epoch": 1017.99,
260070
+ "learning_rate": 7.992532051282052e-06,
260071
+ "loss": 0.8193,
260072
+ "step": 126315
260073
+ },
260074
+ {
260075
+ "epoch": 1018.0,
260076
+ "eval_loss": 0.4346490502357483,
260077
+ "eval_runtime": 35.2379,
260078
+ "eval_samples_per_second": 23.866,
260079
+ "eval_steps_per_second": 0.766,
260080
+ "eval_wer": 0.18682317304051496,
260081
+ "step": 126316
260082
+ },
260083
+ {
260084
+ "epoch": 1018.03,
260085
+ "learning_rate": 7.992451923076924e-06,
260086
+ "loss": 0.3603,
260087
+ "step": 126320
260088
+ },
260089
+ {
260090
+ "epoch": 1018.07,
260091
+ "learning_rate": 7.992371794871797e-06,
260092
+ "loss": 0.3009,
260093
+ "step": 126325
260094
+ },
260095
+ {
260096
+ "epoch": 1018.11,
260097
+ "learning_rate": 7.992291666666668e-06,
260098
+ "loss": 0.3143,
260099
+ "step": 126330
260100
+ },
260101
+ {
260102
+ "epoch": 1018.15,
260103
+ "learning_rate": 7.99221153846154e-06,
260104
+ "loss": 0.3757,
260105
+ "step": 126335
260106
+ },
260107
+ {
260108
+ "epoch": 1018.19,
260109
+ "learning_rate": 7.99213141025641e-06,
260110
+ "loss": 0.8593,
260111
+ "step": 126340
260112
+ },
260113
+ {
260114
+ "epoch": 1018.23,
260115
+ "learning_rate": 7.992051282051284e-06,
260116
+ "loss": 0.644,
260117
+ "step": 126345
260118
+ },
260119
+ {
260120
+ "epoch": 1018.27,
260121
+ "learning_rate": 7.991971153846153e-06,
260122
+ "loss": 0.377,
260123
+ "step": 126350
260124
+ },
260125
+ {
260126
+ "epoch": 1018.31,
260127
+ "learning_rate": 7.991891025641027e-06,
260128
+ "loss": 0.3072,
260129
+ "step": 126355
260130
+ },
260131
+ {
260132
+ "epoch": 1018.35,
260133
+ "learning_rate": 7.991810897435898e-06,
260134
+ "loss": 0.3897,
260135
+ "step": 126360
260136
+ },
260137
+ {
260138
+ "epoch": 1018.39,
260139
+ "learning_rate": 7.99173076923077e-06,
260140
+ "loss": 0.9711,
260141
+ "step": 126365
260142
+ },
260143
+ {
260144
+ "epoch": 1018.43,
260145
+ "learning_rate": 7.99165064102564e-06,
260146
+ "loss": 0.6407,
260147
+ "step": 126370
260148
+ },
260149
+ {
260150
+ "epoch": 1018.47,
260151
+ "learning_rate": 7.991570512820514e-06,
260152
+ "loss": 0.3422,
260153
+ "step": 126375
260154
+ },
260155
+ {
260156
+ "epoch": 1018.51,
260157
+ "learning_rate": 7.991490384615385e-06,
260158
+ "loss": 0.3255,
260159
+ "step": 126380
260160
+ },
260161
+ {
260162
+ "epoch": 1018.55,
260163
+ "learning_rate": 7.991410256410256e-06,
260164
+ "loss": 0.4262,
260165
+ "step": 126385
260166
+ },
260167
+ {
260168
+ "epoch": 1018.59,
260169
+ "learning_rate": 7.99133012820513e-06,
260170
+ "loss": 1.022,
260171
+ "step": 126390
260172
+ },
260173
+ {
260174
+ "epoch": 1018.63,
260175
+ "learning_rate": 7.99125e-06,
260176
+ "loss": 0.6429,
260177
+ "step": 126395
260178
+ },
260179
+ {
260180
+ "epoch": 1018.67,
260181
+ "learning_rate": 7.991169871794872e-06,
260182
+ "loss": 0.2113,
260183
+ "step": 126400
260184
+ },
260185
+ {
260186
+ "epoch": 1018.71,
260187
+ "learning_rate": 7.991089743589743e-06,
260188
+ "loss": 0.3158,
260189
+ "step": 126405
260190
+ },
260191
+ {
260192
+ "epoch": 1018.76,
260193
+ "learning_rate": 7.991009615384617e-06,
260194
+ "loss": 0.4313,
260195
+ "step": 126410
260196
+ },
260197
+ {
260198
+ "epoch": 1018.8,
260199
+ "learning_rate": 7.990929487179488e-06,
260200
+ "loss": 0.879,
260201
+ "step": 126415
260202
+ },
260203
+ {
260204
+ "epoch": 1018.84,
260205
+ "learning_rate": 7.99084935897436e-06,
260206
+ "loss": 0.7075,
260207
+ "step": 126420
260208
+ },
260209
+ {
260210
+ "epoch": 1018.88,
260211
+ "learning_rate": 7.990769230769232e-06,
260212
+ "loss": 0.262,
260213
+ "step": 126425
260214
+ },
260215
+ {
260216
+ "epoch": 1018.92,
260217
+ "learning_rate": 7.990689102564104e-06,
260218
+ "loss": 0.2746,
260219
+ "step": 126430
260220
+ },
260221
+ {
260222
+ "epoch": 1018.96,
260223
+ "learning_rate": 7.990608974358975e-06,
260224
+ "loss": 0.4399,
260225
+ "step": 126435
260226
+ },
260227
+ {
260228
+ "epoch": 1019.0,
260229
+ "learning_rate": 7.990528846153846e-06,
260230
+ "loss": 0.9204,
260231
+ "step": 126440
260232
+ },
260233
+ {
260234
+ "epoch": 1019.0,
260235
+ "eval_loss": 0.4004644453525543,
260236
+ "eval_runtime": 36.4199,
260237
+ "eval_samples_per_second": 23.092,
260238
+ "eval_steps_per_second": 0.741,
260239
+ "eval_wer": 0.1846421831905937,
260240
+ "step": 126440
260241
+ },
260242
+ {
260243
+ "epoch": 1011.04,
260244
+ "learning_rate": 7.99044871794872e-06,
260245
+ "loss": 0.3103,
260246
+ "step": 126445
260247
+ },
260248
+ {
260249
+ "epoch": 1011.08,
260250
+ "learning_rate": 7.99036858974359e-06,
260251
+ "loss": 0.2886,
260252
+ "step": 126450
260253
+ },
260254
+ {
260255
+ "epoch": 1011.12,
260256
+ "learning_rate": 7.990288461538462e-06,
260257
+ "loss": 0.3135,
260258
+ "step": 126455
260259
+ },
260260
+ {
260261
+ "epoch": 1011.16,
260262
+ "learning_rate": 7.990208333333334e-06,
260263
+ "loss": 0.5276,
260264
+ "step": 126460
260265
+ },
260266
+ {
260267
+ "epoch": 1011.2,
260268
+ "learning_rate": 7.990128205128207e-06,
260269
+ "loss": 1.2226,
260270
+ "step": 126465
260271
+ },
260272
+ {
260273
+ "epoch": 1011.24,
260274
+ "learning_rate": 7.990048076923076e-06,
260275
+ "loss": 0.3239,
260276
+ "step": 126470
260277
+ },
260278
+ {
260279
+ "epoch": 1011.28,
260280
+ "learning_rate": 7.98996794871795e-06,
260281
+ "loss": 0.301,
260282
+ "step": 126475
260283
+ },
260284
+ {
260285
+ "epoch": 1011.32,
260286
+ "learning_rate": 7.989887820512822e-06,
260287
+ "loss": 0.3185,
260288
+ "step": 126480
260289
+ },
260290
+ {
260291
+ "epoch": 1011.36,
260292
+ "learning_rate": 7.989807692307692e-06,
260293
+ "loss": 0.4219,
260294
+ "step": 126485
260295
+ },
260296
+ {
260297
+ "epoch": 1011.4,
260298
+ "learning_rate": 7.989727564102565e-06,
260299
+ "loss": 1.1383,
260300
+ "step": 126490
260301
+ },
260302
+ {
260303
+ "epoch": 1011.44,
260304
+ "learning_rate": 7.989647435897436e-06,
260305
+ "loss": 0.2869,
260306
+ "step": 126495
260307
+ },
260308
+ {
260309
+ "epoch": 1011.48,
260310
+ "learning_rate": 7.989567307692308e-06,
260311
+ "loss": 0.3177,
260312
+ "step": 126500
260313
+ },
260314
+ {
260315
+ "epoch": 1011.52,
260316
+ "learning_rate": 7.989487179487179e-06,
260317
+ "loss": 0.2936,
260318
+ "step": 126505
260319
+ },
260320
+ {
260321
+ "epoch": 1011.56,
260322
+ "learning_rate": 7.989407051282052e-06,
260323
+ "loss": 0.4247,
260324
+ "step": 126510
260325
+ },
260326
+ {
260327
+ "epoch": 1011.6,
260328
+ "learning_rate": 7.989326923076924e-06,
260329
+ "loss": 1.1716,
260330
+ "step": 126515
260331
+ },
260332
+ {
260333
+ "epoch": 1011.64,
260334
+ "learning_rate": 7.989246794871795e-06,
260335
+ "loss": 0.392,
260336
+ "step": 126520
260337
+ },
260338
+ {
260339
+ "epoch": 1011.68,
260340
+ "learning_rate": 7.989166666666668e-06,
260341
+ "loss": 0.2561,
260342
+ "step": 126525
260343
+ },
260344
+ {
260345
+ "epoch": 1011.72,
260346
+ "learning_rate": 7.98908653846154e-06,
260347
+ "loss": 0.2848,
260348
+ "step": 126530
260349
+ },
260350
+ {
260351
+ "epoch": 1011.76,
260352
+ "learning_rate": 7.98900641025641e-06,
260353
+ "loss": 0.5091,
260354
+ "step": 126535
260355
+ },
260356
+ {
260357
+ "epoch": 1011.8,
260358
+ "learning_rate": 7.988926282051282e-06,
260359
+ "loss": 1.127,
260360
+ "step": 126540
260361
+ },
260362
+ {
260363
+ "epoch": 1011.84,
260364
+ "learning_rate": 7.988846153846155e-06,
260365
+ "loss": 0.2767,
260366
+ "step": 126545
260367
+ },
260368
+ {
260369
+ "epoch": 1011.88,
260370
+ "learning_rate": 7.988766025641026e-06,
260371
+ "loss": 0.265,
260372
+ "step": 126550
260373
+ },
260374
+ {
260375
+ "epoch": 1011.92,
260376
+ "learning_rate": 7.988685897435898e-06,
260377
+ "loss": 0.2979,
260378
+ "step": 126555
260379
+ },
260380
+ {
260381
+ "epoch": 1011.96,
260382
+ "learning_rate": 7.98860576923077e-06,
260383
+ "loss": 0.3722,
260384
+ "step": 126560
260385
+ },
260386
+ {
260387
+ "epoch": 1012.0,
260388
+ "learning_rate": 7.988525641025642e-06,
260389
+ "loss": 1.2953,
260390
+ "step": 126565
260391
+ },
260392
+ {
260393
+ "epoch": 1012.0,
260394
+ "eval_loss": 0.3975546061992645,
260395
+ "eval_runtime": 36.063,
260396
+ "eval_samples_per_second": 23.32,
260397
+ "eval_steps_per_second": 0.749,
260398
+ "eval_wer": 0.17799635701275046,
260399
+ "step": 126565
260400
+ },
260401
+ {
260402
+ "epoch": 1020.04,
260403
+ "learning_rate": 7.988445512820514e-06,
260404
+ "loss": 0.2682,
260405
+ "step": 126570
260406
+ },
260407
+ {
260408
+ "epoch": 1020.08,
260409
+ "learning_rate": 7.988365384615385e-06,
260410
+ "loss": 0.3308,
260411
+ "step": 126575
260412
+ },
260413
+ {
260414
+ "epoch": 1020.12,
260415
+ "learning_rate": 7.988285256410258e-06,
260416
+ "loss": 0.306,
260417
+ "step": 126580
260418
+ },
260419
+ {
260420
+ "epoch": 1020.16,
260421
+ "learning_rate": 7.98820512820513e-06,
260422
+ "loss": 0.4581,
260423
+ "step": 126585
260424
+ },
260425
+ {
260426
+ "epoch": 1020.2,
260427
+ "learning_rate": 7.988125e-06,
260428
+ "loss": 1.1796,
260429
+ "step": 126590
260430
+ },
260431
+ {
260432
+ "epoch": 1020.24,
260433
+ "learning_rate": 7.988044871794872e-06,
260434
+ "loss": 0.3497,
260435
+ "step": 126595
260436
+ },
260437
+ {
260438
+ "epoch": 1020.28,
260439
+ "learning_rate": 7.987964743589745e-06,
260440
+ "loss": 0.2608,
260441
+ "step": 126600
260442
+ },
260443
+ {
260444
+ "epoch": 1020.32,
260445
+ "learning_rate": 7.987884615384616e-06,
260446
+ "loss": 0.3339,
260447
+ "step": 126605
260448
+ },
260449
+ {
260450
+ "epoch": 1020.36,
260451
+ "learning_rate": 7.987804487179488e-06,
260452
+ "loss": 0.4798,
260453
+ "step": 126610
260454
+ },
260455
+ {
260456
+ "epoch": 1020.4,
260457
+ "learning_rate": 7.98772435897436e-06,
260458
+ "loss": 1.2394,
260459
+ "step": 126615
260460
+ },
260461
+ {
260462
+ "epoch": 1020.44,
260463
+ "learning_rate": 7.987644230769232e-06,
260464
+ "loss": 0.3557,
260465
+ "step": 126620
260466
+ },
260467
+ {
260468
+ "epoch": 1020.48,
260469
+ "learning_rate": 7.987564102564104e-06,
260470
+ "loss": 0.2716,
260471
+ "step": 126625
260472
+ },
260473
+ {
260474
+ "epoch": 1020.52,
260475
+ "learning_rate": 7.987483974358975e-06,
260476
+ "loss": 0.3483,
260477
+ "step": 126630
260478
+ },
260479
+ {
260480
+ "epoch": 1020.56,
260481
+ "learning_rate": 7.987403846153848e-06,
260482
+ "loss": 0.4007,
260483
+ "step": 126635
260484
+ },
260485
+ {
260486
+ "epoch": 1020.6,
260487
+ "learning_rate": 7.987323717948718e-06,
260488
+ "loss": 1.1896,
260489
+ "step": 126640
260490
+ },
260491
+ {
260492
+ "epoch": 1020.64,
260493
+ "learning_rate": 7.98724358974359e-06,
260494
+ "loss": 0.3013,
260495
+ "step": 126645
260496
+ },
260497
+ {
260498
+ "epoch": 1020.68,
260499
+ "learning_rate": 7.987163461538462e-06,
260500
+ "loss": 0.3619,
260501
+ "step": 126650
260502
+ },
260503
+ {
260504
+ "epoch": 1020.72,
260505
+ "learning_rate": 7.987083333333333e-06,
260506
+ "loss": 0.3159,
260507
+ "step": 126655
260508
+ },
260509
+ {
260510
+ "epoch": 1020.76,
260511
+ "learning_rate": 7.987003205128205e-06,
260512
+ "loss": 0.5123,
260513
+ "step": 126660
260514
+ },
260515
+ {
260516
+ "epoch": 1020.8,
260517
+ "learning_rate": 7.986923076923078e-06,
260518
+ "loss": 1.0823,
260519
+ "step": 126665
260520
+ },
260521
+ {
260522
+ "epoch": 1020.84,
260523
+ "learning_rate": 7.98684294871795e-06,
260524
+ "loss": 0.4015,
260525
+ "step": 126670
260526
+ },
260527
+ {
260528
+ "epoch": 1020.88,
260529
+ "learning_rate": 7.98676282051282e-06,
260530
+ "loss": 0.282,
260531
+ "step": 126675
260532
+ },
260533
+ {
260534
+ "epoch": 1020.92,
260535
+ "learning_rate": 7.986682692307694e-06,
260536
+ "loss": 0.2633,
260537
+ "step": 126680
260538
+ },
260539
+ {
260540
+ "epoch": 1020.96,
260541
+ "learning_rate": 7.986602564102565e-06,
260542
+ "loss": 0.5204,
260543
+ "step": 126685
260544
+ },
260545
+ {
260546
+ "epoch": 1021.0,
260547
+ "eval_loss": 0.48318079113960266,
260548
+ "eval_runtime": 36.3732,
260549
+ "eval_samples_per_second": 23.121,
260550
+ "eval_steps_per_second": 0.742,
260551
+ "eval_wer": 0.17306996599377758,
260552
+ "step": 126689
260553
  }
260554
  ],
260555
+ "max_steps": 620000,
260556
  "num_train_epochs": 5000,
260557
+ "total_flos": 3.565038460292019e+20,
260558
  "trial_name": null,
260559
  "trial_params": null
260560
  }
model-bin/finetune/base/{checkpoint-126068 β†’ checkpoint-126689}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630148780.943968/events.out.tfevents.1630148780.86bb0ddabf9b.4092.131 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ab44b6e5c35156b4ad476574b19d4935688b65613900c4cd88a73d613d37011
3
+ size 4194
model-bin/finetune/base/log/1630149168.480171/events.out.tfevents.1630149168.86bb0ddabf9b.4092.133 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94896acafd410a6524def8fbde908e3695d87dc1fdba3d87fdecd2e7bf2d5c97
3
+ size 4194
model-bin/finetune/base/log/1630149553.0608108/events.out.tfevents.1630149553.86bb0ddabf9b.4092.135 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:833c89ce3560863ef1517cce9f72e99ae495783f0489457aa4194d5f504edc98
3
+ size 4194
model-bin/finetune/base/log/1630149942.7762349/events.out.tfevents.1630149942.86bb0ddabf9b.4092.137 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b58d267d6bdac191c76983bb0f51d62f719d3bbbc94d369e42d273f1bf6d79ef
3
+ size 4194
model-bin/finetune/base/log/1630150333.0672135/events.out.tfevents.1630150333.86bb0ddabf9b.4092.139 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0ead56c0eb2b95b220c1eccc5eba92f37d7152145b86b84a7e7c9c0db293c85
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630148780.86bb0ddabf9b.4092.130 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00048b58100d6130feb62b3377cf7b29b7917adac51c8fbb83895e49947cb01c
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630149168.86bb0ddabf9b.4092.132 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74f26fc93c6313721154ee0e726336b59c3cda1c4fc9581c1175daf4005edd28
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630149553.86bb0ddabf9b.4092.134 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:236afb18a29aeb252568902992854ea8ecb721fdfb487077654c8691d1ddb1f6
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630149942.86bb0ddabf9b.4092.136 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1294791834390db7b3bc786993a9abcc101b0e705f9079616c8f8f3002ec5659
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630150333.86bb0ddabf9b.4092.138 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de79b511c4f8041a051940b71d66e3d1c442e9d5899f28fb484327a5ec5d45d5
3
+ size 8462