Check commited on
Commit
c3ec61b
Β·
1 Parent(s): 384fb00

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/trainer_state.json +793 -4
  9. model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629827645.253879/events.out.tfevents.1629827645.c435e1c5ee04.920.191 +3 -0
  11. model-bin/finetune/base/log/1629828281.4169908/events.out.tfevents.1629828281.c435e1c5ee04.920.193 +3 -0
  12. model-bin/finetune/base/log/1629828923.8657014/events.out.tfevents.1629828923.c435e1c5ee04.920.195 +3 -0
  13. model-bin/finetune/base/log/1629829555.331946/events.out.tfevents.1629829555.c435e1c5ee04.920.197 +3 -0
  14. model-bin/finetune/base/log/1629830197.8430533/events.out.tfevents.1629830197.c435e1c5ee04.920.199 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629827645.c435e1c5ee04.920.190 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629828281.c435e1c5ee04.920.192 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629828923.c435e1c5ee04.920.194 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629829555.c435e1c5ee04.920.196 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629830197.c435e1c5ee04.920.198 +3 -0
model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dd06fe640e23974d0f6a0907f379f9ed0b9d1d39aa9e878b427221d68e27fb0c
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85ea8ef6c64908f2115fcc83c3c7ba096536f425df00a3c9f1cbfc2255f20245
3
  size 722165393
model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9ca37dc64921b379bc597984828ccb1ffca37d2bfba4dc65a0bf832d7fb196fc
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:989a241331db9f31c999d496faab12f6ba17620c47350944896a193165b684aa
3
  size 377909911
model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:21a9d2cdf11457adf20e12a9253bdd90f36c56e3655c2a2506c52983a83e6230
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06ffb88ff1db1b23a37760fd519fbca7feaf5222271d9129ca3900755d9cb924
3
  size 14503
model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d7219f96e098583795db3c9787cf815cf9738bb70a1274c9695498cf865dfb3c
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ae34bb3b632ce84a747aa987dc0f72facb05243dbdbd4095b0555d0a6525518
3
  size 559
model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ba2fa47ca503d1240381ecfce0ecd5428d683eb16b4f83b6f890ff832b6f62f
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13a19303204503e39131fe6893ee9e90595d5035f7349c77cb80c388b6b78e06
3
  size 623
model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.18588425381903642,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-60729",
4
- "epoch": 525.0,
5
- "global_step": 65705,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -182964,11 +182964,800 @@
182964
  "eval_steps_per_second": 0.677,
182965
  "eval_wer": 0.18776816231546797,
182966
  "step": 65705
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
182967
  }
182968
  ],
182969
- "max_steps": 625000,
182970
  "num_train_epochs": 5000,
182971
- "total_flos": 1.8489006770031778e+20,
182972
  "trial_name": null,
182973
  "trial_params": null
182974
  }
 
1
  {
2
  "best_metric": 0.18588425381903642,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-60729",
4
+ "epoch": 533.995983935743,
5
+ "global_step": 66326,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
182964
  "eval_steps_per_second": 0.677,
182965
  "eval_wer": 0.18776816231546797,
182966
  "step": 65705
182967
+ },
182968
+ {
182969
+ "epoch": 529.04,
182970
+ "learning_rate": 8.963381410256411e-06,
182971
+ "loss": 0.3291,
182972
+ "step": 65710
182973
+ },
182974
+ {
182975
+ "epoch": 529.08,
182976
+ "learning_rate": 8.963301282051284e-06,
182977
+ "loss": 0.325,
182978
+ "step": 65715
182979
+ },
182980
+ {
182981
+ "epoch": 529.12,
182982
+ "learning_rate": 8.963221153846154e-06,
182983
+ "loss": 0.3205,
182984
+ "step": 65720
182985
+ },
182986
+ {
182987
+ "epoch": 529.16,
182988
+ "learning_rate": 8.963141025641027e-06,
182989
+ "loss": 0.4565,
182990
+ "step": 65725
182991
+ },
182992
+ {
182993
+ "epoch": 529.2,
182994
+ "learning_rate": 8.963060897435898e-06,
182995
+ "loss": 1.14,
182996
+ "step": 65730
182997
+ },
182998
+ {
182999
+ "epoch": 529.24,
183000
+ "learning_rate": 8.96298076923077e-06,
183001
+ "loss": 0.2933,
183002
+ "step": 65735
183003
+ },
183004
+ {
183005
+ "epoch": 529.28,
183006
+ "learning_rate": 8.96290064102564e-06,
183007
+ "loss": 0.3287,
183008
+ "step": 65740
183009
+ },
183010
+ {
183011
+ "epoch": 529.32,
183012
+ "learning_rate": 8.962820512820514e-06,
183013
+ "loss": 0.3612,
183014
+ "step": 65745
183015
+ },
183016
+ {
183017
+ "epoch": 529.36,
183018
+ "learning_rate": 8.962740384615385e-06,
183019
+ "loss": 0.505,
183020
+ "step": 65750
183021
+ },
183022
+ {
183023
+ "epoch": 529.4,
183024
+ "learning_rate": 8.962660256410257e-06,
183025
+ "loss": 1.2446,
183026
+ "step": 65755
183027
+ },
183028
+ {
183029
+ "epoch": 529.44,
183030
+ "learning_rate": 8.96258012820513e-06,
183031
+ "loss": 0.4099,
183032
+ "step": 65760
183033
+ },
183034
+ {
183035
+ "epoch": 529.48,
183036
+ "learning_rate": 8.962500000000001e-06,
183037
+ "loss": 0.4326,
183038
+ "step": 65765
183039
+ },
183040
+ {
183041
+ "epoch": 529.52,
183042
+ "learning_rate": 8.962419871794872e-06,
183043
+ "loss": 0.29,
183044
+ "step": 65770
183045
+ },
183046
+ {
183047
+ "epoch": 529.56,
183048
+ "learning_rate": 8.962339743589744e-06,
183049
+ "loss": 0.501,
183050
+ "step": 65775
183051
+ },
183052
+ {
183053
+ "epoch": 529.6,
183054
+ "learning_rate": 8.962259615384617e-06,
183055
+ "loss": 1.2461,
183056
+ "step": 65780
183057
+ },
183058
+ {
183059
+ "epoch": 529.64,
183060
+ "learning_rate": 8.962179487179488e-06,
183061
+ "loss": 0.366,
183062
+ "step": 65785
183063
+ },
183064
+ {
183065
+ "epoch": 529.68,
183066
+ "learning_rate": 8.96209935897436e-06,
183067
+ "loss": 0.3331,
183068
+ "step": 65790
183069
+ },
183070
+ {
183071
+ "epoch": 529.72,
183072
+ "learning_rate": 8.962019230769231e-06,
183073
+ "loss": 0.3664,
183074
+ "step": 65795
183075
+ },
183076
+ {
183077
+ "epoch": 529.76,
183078
+ "learning_rate": 8.961939102564104e-06,
183079
+ "loss": 0.595,
183080
+ "step": 65800
183081
+ },
183082
+ {
183083
+ "epoch": 529.8,
183084
+ "learning_rate": 8.961858974358975e-06,
183085
+ "loss": 1.1922,
183086
+ "step": 65805
183087
+ },
183088
+ {
183089
+ "epoch": 529.84,
183090
+ "learning_rate": 8.961778846153847e-06,
183091
+ "loss": 0.3499,
183092
+ "step": 65810
183093
+ },
183094
+ {
183095
+ "epoch": 529.88,
183096
+ "learning_rate": 8.96169871794872e-06,
183097
+ "loss": 0.2735,
183098
+ "step": 65815
183099
+ },
183100
+ {
183101
+ "epoch": 529.92,
183102
+ "learning_rate": 8.961618589743591e-06,
183103
+ "loss": 0.3509,
183104
+ "step": 65820
183105
+ },
183106
+ {
183107
+ "epoch": 529.96,
183108
+ "learning_rate": 8.961538461538462e-06,
183109
+ "loss": 0.4441,
183110
+ "step": 65825
183111
+ },
183112
+ {
183113
+ "epoch": 530.0,
183114
+ "eval_loss": 0.38648757338523865,
183115
+ "eval_runtime": 40.1418,
183116
+ "eval_samples_per_second": 20.826,
183117
+ "eval_steps_per_second": 0.673,
183118
+ "eval_wer": 0.19017903573285788,
183119
+ "step": 65829
183120
+ },
183121
+ {
183122
+ "epoch": 530.01,
183123
+ "learning_rate": 8.961458333333334e-06,
183124
+ "loss": 0.4862,
183125
+ "step": 65830
183126
+ },
183127
+ {
183128
+ "epoch": 530.05,
183129
+ "learning_rate": 8.961378205128207e-06,
183130
+ "loss": 0.4615,
183131
+ "step": 65835
183132
+ },
183133
+ {
183134
+ "epoch": 530.09,
183135
+ "learning_rate": 8.961298076923076e-06,
183136
+ "loss": 0.3626,
183137
+ "step": 65840
183138
+ },
183139
+ {
183140
+ "epoch": 530.13,
183141
+ "learning_rate": 8.96121794871795e-06,
183142
+ "loss": 0.41,
183143
+ "step": 65845
183144
+ },
183145
+ {
183146
+ "epoch": 530.17,
183147
+ "learning_rate": 8.961137820512821e-06,
183148
+ "loss": 0.5743,
183149
+ "step": 65850
183150
+ },
183151
+ {
183152
+ "epoch": 530.21,
183153
+ "learning_rate": 8.961057692307692e-06,
183154
+ "loss": 1.2365,
183155
+ "step": 65855
183156
+ },
183157
+ {
183158
+ "epoch": 530.25,
183159
+ "learning_rate": 8.960977564102565e-06,
183160
+ "loss": 0.3229,
183161
+ "step": 65860
183162
+ },
183163
+ {
183164
+ "epoch": 530.29,
183165
+ "learning_rate": 8.960897435897437e-06,
183166
+ "loss": 0.3577,
183167
+ "step": 65865
183168
+ },
183169
+ {
183170
+ "epoch": 530.33,
183171
+ "learning_rate": 8.960817307692308e-06,
183172
+ "loss": 0.3767,
183173
+ "step": 65870
183174
+ },
183175
+ {
183176
+ "epoch": 530.37,
183177
+ "learning_rate": 8.96073717948718e-06,
183178
+ "loss": 0.4945,
183179
+ "step": 65875
183180
+ },
183181
+ {
183182
+ "epoch": 530.41,
183183
+ "learning_rate": 8.960657051282052e-06,
183184
+ "loss": 1.2582,
183185
+ "step": 65880
183186
+ },
183187
+ {
183188
+ "epoch": 530.45,
183189
+ "learning_rate": 8.960576923076924e-06,
183190
+ "loss": 0.3046,
183191
+ "step": 65885
183192
+ },
183193
+ {
183194
+ "epoch": 530.49,
183195
+ "learning_rate": 8.960496794871795e-06,
183196
+ "loss": 0.2987,
183197
+ "step": 65890
183198
+ },
183199
+ {
183200
+ "epoch": 530.53,
183201
+ "learning_rate": 8.960416666666666e-06,
183202
+ "loss": 0.3325,
183203
+ "step": 65895
183204
+ },
183205
+ {
183206
+ "epoch": 530.57,
183207
+ "learning_rate": 8.96033653846154e-06,
183208
+ "loss": 0.5671,
183209
+ "step": 65900
183210
+ },
183211
+ {
183212
+ "epoch": 530.61,
183213
+ "learning_rate": 8.960256410256411e-06,
183214
+ "loss": 1.1719,
183215
+ "step": 65905
183216
+ },
183217
+ {
183218
+ "epoch": 530.65,
183219
+ "learning_rate": 8.960176282051282e-06,
183220
+ "loss": 0.3255,
183221
+ "step": 65910
183222
+ },
183223
+ {
183224
+ "epoch": 530.69,
183225
+ "learning_rate": 8.960096153846155e-06,
183226
+ "loss": 0.3273,
183227
+ "step": 65915
183228
+ },
183229
+ {
183230
+ "epoch": 530.73,
183231
+ "learning_rate": 8.960016025641027e-06,
183232
+ "loss": 0.4071,
183233
+ "step": 65920
183234
+ },
183235
+ {
183236
+ "epoch": 530.77,
183237
+ "learning_rate": 8.959935897435898e-06,
183238
+ "loss": 0.4977,
183239
+ "step": 65925
183240
+ },
183241
+ {
183242
+ "epoch": 530.81,
183243
+ "learning_rate": 8.95985576923077e-06,
183244
+ "loss": 1.1886,
183245
+ "step": 65930
183246
+ },
183247
+ {
183248
+ "epoch": 530.85,
183249
+ "learning_rate": 8.959775641025642e-06,
183250
+ "loss": 0.3619,
183251
+ "step": 65935
183252
+ },
183253
+ {
183254
+ "epoch": 530.89,
183255
+ "learning_rate": 8.959695512820514e-06,
183256
+ "loss": 0.296,
183257
+ "step": 65940
183258
+ },
183259
+ {
183260
+ "epoch": 530.93,
183261
+ "learning_rate": 8.959615384615385e-06,
183262
+ "loss": 0.3815,
183263
+ "step": 65945
183264
+ },
183265
+ {
183266
+ "epoch": 530.97,
183267
+ "learning_rate": 8.959535256410257e-06,
183268
+ "loss": 0.5965,
183269
+ "step": 65950
183270
+ },
183271
+ {
183272
+ "epoch": 531.0,
183273
+ "eval_loss": 0.4443589448928833,
183274
+ "eval_runtime": 40.3659,
183275
+ "eval_samples_per_second": 20.711,
183276
+ "eval_steps_per_second": 0.669,
183277
+ "eval_wer": 0.2028847561868502,
183278
+ "step": 65953
183279
+ },
183280
+ {
183281
+ "epoch": 527.02,
183282
+ "learning_rate": 8.95945512820513e-06,
183283
+ "loss": 0.4132,
183284
+ "step": 65955
183285
+ },
183286
+ {
183287
+ "epoch": 527.06,
183288
+ "learning_rate": 8.959375000000001e-06,
183289
+ "loss": 0.3103,
183290
+ "step": 65960
183291
+ },
183292
+ {
183293
+ "epoch": 527.1,
183294
+ "learning_rate": 8.959294871794872e-06,
183295
+ "loss": 0.3276,
183296
+ "step": 65965
183297
+ },
183298
+ {
183299
+ "epoch": 527.14,
183300
+ "learning_rate": 8.959214743589745e-06,
183301
+ "loss": 0.4584,
183302
+ "step": 65970
183303
+ },
183304
+ {
183305
+ "epoch": 527.18,
183306
+ "learning_rate": 8.959134615384617e-06,
183307
+ "loss": 0.6109,
183308
+ "step": 65975
183309
+ },
183310
+ {
183311
+ "epoch": 527.22,
183312
+ "learning_rate": 8.959054487179488e-06,
183313
+ "loss": 1.1637,
183314
+ "step": 65980
183315
+ },
183316
+ {
183317
+ "epoch": 527.26,
183318
+ "learning_rate": 8.95897435897436e-06,
183319
+ "loss": 0.2906,
183320
+ "step": 65985
183321
+ },
183322
+ {
183323
+ "epoch": 527.3,
183324
+ "learning_rate": 8.958894230769232e-06,
183325
+ "loss": 0.3802,
183326
+ "step": 65990
183327
+ },
183328
+ {
183329
+ "epoch": 527.34,
183330
+ "learning_rate": 8.958814102564102e-06,
183331
+ "loss": 0.3634,
183332
+ "step": 65995
183333
+ },
183334
+ {
183335
+ "epoch": 527.38,
183336
+ "learning_rate": 8.958733974358975e-06,
183337
+ "loss": 0.6906,
183338
+ "step": 66000
183339
+ },
183340
+ {
183341
+ "epoch": 527.42,
183342
+ "learning_rate": 8.958653846153847e-06,
183343
+ "loss": 1.0945,
183344
+ "step": 66005
183345
+ },
183346
+ {
183347
+ "epoch": 527.46,
183348
+ "learning_rate": 8.958573717948718e-06,
183349
+ "loss": 0.3012,
183350
+ "step": 66010
183351
+ },
183352
+ {
183353
+ "epoch": 527.5,
183354
+ "learning_rate": 8.958493589743591e-06,
183355
+ "loss": 0.3196,
183356
+ "step": 66015
183357
+ },
183358
+ {
183359
+ "epoch": 527.54,
183360
+ "learning_rate": 8.958413461538462e-06,
183361
+ "loss": 0.3637,
183362
+ "step": 66020
183363
+ },
183364
+ {
183365
+ "epoch": 527.58,
183366
+ "learning_rate": 8.958333333333334e-06,
183367
+ "loss": 0.6842,
183368
+ "step": 66025
183369
+ },
183370
+ {
183371
+ "epoch": 527.62,
183372
+ "learning_rate": 8.958253205128205e-06,
183373
+ "loss": 0.9815,
183374
+ "step": 66030
183375
+ },
183376
+ {
183377
+ "epoch": 527.66,
183378
+ "learning_rate": 8.958173076923078e-06,
183379
+ "loss": 0.2745,
183380
+ "step": 66035
183381
+ },
183382
+ {
183383
+ "epoch": 527.7,
183384
+ "learning_rate": 8.95809294871795e-06,
183385
+ "loss": 0.4007,
183386
+ "step": 66040
183387
+ },
183388
+ {
183389
+ "epoch": 527.74,
183390
+ "learning_rate": 8.95801282051282e-06,
183391
+ "loss": 0.3229,
183392
+ "step": 66045
183393
+ },
183394
+ {
183395
+ "epoch": 527.78,
183396
+ "learning_rate": 8.957932692307692e-06,
183397
+ "loss": 0.7089,
183398
+ "step": 66050
183399
+ },
183400
+ {
183401
+ "epoch": 527.82,
183402
+ "learning_rate": 8.957852564102565e-06,
183403
+ "loss": 1.0354,
183404
+ "step": 66055
183405
+ },
183406
+ {
183407
+ "epoch": 527.86,
183408
+ "learning_rate": 8.957772435897437e-06,
183409
+ "loss": 0.3453,
183410
+ "step": 66060
183411
+ },
183412
+ {
183413
+ "epoch": 527.9,
183414
+ "learning_rate": 8.957692307692308e-06,
183415
+ "loss": 0.2743,
183416
+ "step": 66065
183417
+ },
183418
+ {
183419
+ "epoch": 527.94,
183420
+ "learning_rate": 8.957612179487181e-06,
183421
+ "loss": 0.4068,
183422
+ "step": 66070
183423
+ },
183424
+ {
183425
+ "epoch": 527.98,
183426
+ "learning_rate": 8.957532051282052e-06,
183427
+ "loss": 0.7323,
183428
+ "step": 66075
183429
+ },
183430
+ {
183431
+ "epoch": 528.0,
183432
+ "eval_loss": 0.41770240664482117,
183433
+ "eval_runtime": 39.8837,
183434
+ "eval_samples_per_second": 20.986,
183435
+ "eval_steps_per_second": 0.677,
183436
+ "eval_wer": 0.19590305828043855,
183437
+ "step": 66078
183438
+ },
183439
+ {
183440
+ "epoch": 532.02,
183441
+ "learning_rate": 8.957451923076924e-06,
183442
+ "loss": 0.3931,
183443
+ "step": 66080
183444
+ },
183445
+ {
183446
+ "epoch": 532.06,
183447
+ "learning_rate": 8.957371794871795e-06,
183448
+ "loss": 0.3411,
183449
+ "step": 66085
183450
+ },
183451
+ {
183452
+ "epoch": 532.1,
183453
+ "learning_rate": 8.957291666666668e-06,
183454
+ "loss": 0.3569,
183455
+ "step": 66090
183456
+ },
183457
+ {
183458
+ "epoch": 532.14,
183459
+ "learning_rate": 8.95721153846154e-06,
183460
+ "loss": 0.3359,
183461
+ "step": 66095
183462
+ },
183463
+ {
183464
+ "epoch": 532.18,
183465
+ "learning_rate": 8.95713141025641e-06,
183466
+ "loss": 0.7029,
183467
+ "step": 66100
183468
+ },
183469
+ {
183470
+ "epoch": 532.22,
183471
+ "learning_rate": 8.957051282051282e-06,
183472
+ "loss": 1.1022,
183473
+ "step": 66105
183474
+ },
183475
+ {
183476
+ "epoch": 532.26,
183477
+ "learning_rate": 8.956971153846155e-06,
183478
+ "loss": 0.3636,
183479
+ "step": 66110
183480
+ },
183481
+ {
183482
+ "epoch": 532.3,
183483
+ "learning_rate": 8.956891025641027e-06,
183484
+ "loss": 0.3155,
183485
+ "step": 66115
183486
+ },
183487
+ {
183488
+ "epoch": 532.34,
183489
+ "learning_rate": 8.956810897435898e-06,
183490
+ "loss": 0.4378,
183491
+ "step": 66120
183492
+ },
183493
+ {
183494
+ "epoch": 532.38,
183495
+ "learning_rate": 8.956730769230771e-06,
183496
+ "loss": 0.6137,
183497
+ "step": 66125
183498
+ },
183499
+ {
183500
+ "epoch": 532.42,
183501
+ "learning_rate": 8.95665064102564e-06,
183502
+ "loss": 1.0293,
183503
+ "step": 66130
183504
+ },
183505
+ {
183506
+ "epoch": 532.46,
183507
+ "learning_rate": 8.956570512820514e-06,
183508
+ "loss": 0.3335,
183509
+ "step": 66135
183510
+ },
183511
+ {
183512
+ "epoch": 532.5,
183513
+ "learning_rate": 8.956490384615385e-06,
183514
+ "loss": 0.3447,
183515
+ "step": 66140
183516
+ },
183517
+ {
183518
+ "epoch": 532.54,
183519
+ "learning_rate": 8.956410256410258e-06,
183520
+ "loss": 0.3888,
183521
+ "step": 66145
183522
+ },
183523
+ {
183524
+ "epoch": 532.58,
183525
+ "learning_rate": 8.956330128205128e-06,
183526
+ "loss": 0.6986,
183527
+ "step": 66150
183528
+ },
183529
+ {
183530
+ "epoch": 532.62,
183531
+ "learning_rate": 8.956250000000001e-06,
183532
+ "loss": 0.9883,
183533
+ "step": 66155
183534
+ },
183535
+ {
183536
+ "epoch": 532.66,
183537
+ "learning_rate": 8.956169871794874e-06,
183538
+ "loss": 0.3206,
183539
+ "step": 66160
183540
+ },
183541
+ {
183542
+ "epoch": 532.7,
183543
+ "learning_rate": 8.956089743589744e-06,
183544
+ "loss": 0.3433,
183545
+ "step": 66165
183546
+ },
183547
+ {
183548
+ "epoch": 532.74,
183549
+ "learning_rate": 8.956009615384617e-06,
183550
+ "loss": 0.4522,
183551
+ "step": 66170
183552
+ },
183553
+ {
183554
+ "epoch": 532.78,
183555
+ "learning_rate": 8.955929487179488e-06,
183556
+ "loss": 0.7144,
183557
+ "step": 66175
183558
+ },
183559
+ {
183560
+ "epoch": 532.82,
183561
+ "learning_rate": 8.95584935897436e-06,
183562
+ "loss": 1.0969,
183563
+ "step": 66180
183564
+ },
183565
+ {
183566
+ "epoch": 532.86,
183567
+ "learning_rate": 8.95576923076923e-06,
183568
+ "loss": 0.2837,
183569
+ "step": 66185
183570
+ },
183571
+ {
183572
+ "epoch": 532.9,
183573
+ "learning_rate": 8.955689102564104e-06,
183574
+ "loss": 0.2845,
183575
+ "step": 66190
183576
+ },
183577
+ {
183578
+ "epoch": 532.94,
183579
+ "learning_rate": 8.955608974358975e-06,
183580
+ "loss": 0.4446,
183581
+ "step": 66195
183582
+ },
183583
+ {
183584
+ "epoch": 532.98,
183585
+ "learning_rate": 8.955528846153846e-06,
183586
+ "loss": 0.6778,
183587
+ "step": 66200
183588
+ },
183589
+ {
183590
+ "epoch": 533.0,
183591
+ "eval_loss": 0.4082199037075043,
183592
+ "eval_runtime": 39.496,
183593
+ "eval_samples_per_second": 21.192,
183594
+ "eval_steps_per_second": 0.684,
183595
+ "eval_wer": 0.190224570673712,
183596
+ "step": 66202
183597
+ },
183598
+ {
183599
+ "epoch": 533.02,
183600
+ "learning_rate": 8.955448717948718e-06,
183601
+ "loss": 0.3439,
183602
+ "step": 66205
183603
+ },
183604
+ {
183605
+ "epoch": 533.06,
183606
+ "learning_rate": 8.955368589743591e-06,
183607
+ "loss": 0.2805,
183608
+ "step": 66210
183609
+ },
183610
+ {
183611
+ "epoch": 533.1,
183612
+ "learning_rate": 8.955288461538462e-06,
183613
+ "loss": 0.3489,
183614
+ "step": 66215
183615
+ },
183616
+ {
183617
+ "epoch": 533.14,
183618
+ "learning_rate": 8.955208333333334e-06,
183619
+ "loss": 0.3674,
183620
+ "step": 66220
183621
+ },
183622
+ {
183623
+ "epoch": 533.18,
183624
+ "learning_rate": 8.955128205128207e-06,
183625
+ "loss": 0.7902,
183626
+ "step": 66225
183627
+ },
183628
+ {
183629
+ "epoch": 533.22,
183630
+ "learning_rate": 8.955048076923078e-06,
183631
+ "loss": 0.8694,
183632
+ "step": 66230
183633
+ },
183634
+ {
183635
+ "epoch": 533.27,
183636
+ "learning_rate": 8.95496794871795e-06,
183637
+ "loss": 0.3494,
183638
+ "step": 66235
183639
+ },
183640
+ {
183641
+ "epoch": 533.31,
183642
+ "learning_rate": 8.95488782051282e-06,
183643
+ "loss": 0.3249,
183644
+ "step": 66240
183645
+ },
183646
+ {
183647
+ "epoch": 533.35,
183648
+ "learning_rate": 8.954807692307694e-06,
183649
+ "loss": 0.3455,
183650
+ "step": 66245
183651
+ },
183652
+ {
183653
+ "epoch": 533.39,
183654
+ "learning_rate": 8.954727564102565e-06,
183655
+ "loss": 0.7674,
183656
+ "step": 66250
183657
+ },
183658
+ {
183659
+ "epoch": 533.43,
183660
+ "learning_rate": 8.954647435897436e-06,
183661
+ "loss": 0.9707,
183662
+ "step": 66255
183663
+ },
183664
+ {
183665
+ "epoch": 533.47,
183666
+ "learning_rate": 8.95456730769231e-06,
183667
+ "loss": 0.3004,
183668
+ "step": 66260
183669
+ },
183670
+ {
183671
+ "epoch": 533.51,
183672
+ "learning_rate": 8.954487179487181e-06,
183673
+ "loss": 0.2967,
183674
+ "step": 66265
183675
+ },
183676
+ {
183677
+ "epoch": 533.55,
183678
+ "learning_rate": 8.954407051282052e-06,
183679
+ "loss": 0.412,
183680
+ "step": 66270
183681
+ },
183682
+ {
183683
+ "epoch": 533.59,
183684
+ "learning_rate": 8.954326923076924e-06,
183685
+ "loss": 0.7517,
183686
+ "step": 66275
183687
+ },
183688
+ {
183689
+ "epoch": 533.63,
183690
+ "learning_rate": 8.954246794871797e-06,
183691
+ "loss": 0.7993,
183692
+ "step": 66280
183693
+ },
183694
+ {
183695
+ "epoch": 533.67,
183696
+ "learning_rate": 8.954166666666666e-06,
183697
+ "loss": 0.3173,
183698
+ "step": 66285
183699
+ },
183700
+ {
183701
+ "epoch": 533.71,
183702
+ "learning_rate": 8.95408653846154e-06,
183703
+ "loss": 0.3012,
183704
+ "step": 66290
183705
+ },
183706
+ {
183707
+ "epoch": 533.75,
183708
+ "learning_rate": 8.95400641025641e-06,
183709
+ "loss": 0.4471,
183710
+ "step": 66295
183711
+ },
183712
+ {
183713
+ "epoch": 533.79,
183714
+ "learning_rate": 8.953926282051282e-06,
183715
+ "loss": 0.7729,
183716
+ "step": 66300
183717
+ },
183718
+ {
183719
+ "epoch": 533.83,
183720
+ "learning_rate": 8.953846153846153e-06,
183721
+ "loss": 0.8857,
183722
+ "step": 66305
183723
+ },
183724
+ {
183725
+ "epoch": 533.87,
183726
+ "learning_rate": 8.953766025641027e-06,
183727
+ "loss": 0.3209,
183728
+ "step": 66310
183729
+ },
183730
+ {
183731
+ "epoch": 533.91,
183732
+ "learning_rate": 8.953685897435898e-06,
183733
+ "loss": 0.3262,
183734
+ "step": 66315
183735
+ },
183736
+ {
183737
+ "epoch": 533.95,
183738
+ "learning_rate": 8.95360576923077e-06,
183739
+ "loss": 0.4383,
183740
+ "step": 66320
183741
+ },
183742
+ {
183743
+ "epoch": 533.99,
183744
+ "learning_rate": 8.953525641025642e-06,
183745
+ "loss": 0.9138,
183746
+ "step": 66325
183747
+ },
183748
+ {
183749
+ "epoch": 534.0,
183750
+ "eval_loss": 0.3869166672229767,
183751
+ "eval_runtime": 39.8868,
183752
+ "eval_samples_per_second": 20.984,
183753
+ "eval_steps_per_second": 0.677,
183754
+ "eval_wer": 0.19705408178220724,
183755
+ "step": 66326
183756
  }
183757
  ],
183758
+ "max_steps": 620000,
183759
  "num_train_epochs": 5000,
183760
+ "total_flos": 1.866428893671497e+20,
183761
  "trial_name": null,
183762
  "trial_params": null
183763
  }
model-bin/finetune/base/{checkpoint-65705 β†’ checkpoint-66326}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629827645.253879/events.out.tfevents.1629827645.c435e1c5ee04.920.191 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:872c038887a161a464f61dfc1bd7fb3419cf63fe36b08086a4be082eeea4b456
3
+ size 4194
model-bin/finetune/base/log/1629828281.4169908/events.out.tfevents.1629828281.c435e1c5ee04.920.193 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38cc1698243689a4e07596c20d358915a55461f481aa14dbf3a6aa3f68d4412b
3
+ size 4194
model-bin/finetune/base/log/1629828923.8657014/events.out.tfevents.1629828923.c435e1c5ee04.920.195 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c2834245a6b5685d069560b77886515d3846cd31764fafd250f4c67f885ec68
3
+ size 4194
model-bin/finetune/base/log/1629829555.331946/events.out.tfevents.1629829555.c435e1c5ee04.920.197 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3f3cf89b5faea2d9873af60bceb7caf01fd9fb18a1b596c30476ddd252b9758
3
+ size 4194
model-bin/finetune/base/log/1629830197.8430533/events.out.tfevents.1629830197.c435e1c5ee04.920.199 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:935bbeba559dcbc45d49319c829d68c04f8a629414eb471bbf5108379f88da1c
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629827645.c435e1c5ee04.920.190 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c9db1b9680624628069187065587324bf6e43d3dec54350e4e84aecc3bc0f17
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1629828281.c435e1c5ee04.920.192 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c89f6ad475e5e639d42a3763c7212290b119596b392a58764c8515aacb20a05a
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629828923.c435e1c5ee04.920.194 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83f9437a06d2c9b772fb5c7b77a362493630d69a53034dc7802d1d022a6d1128
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629829555.c435e1c5ee04.920.196 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2eb0097ac856f24766d524b13034bca29928832090f5611adb8c6adcd0bced64
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629830197.c435e1c5ee04.920.198 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3ae473a659ee7593575954eb1040e59cf7ccd057b8594dc6fece1639d9365af
3
+ size 8622