Check commited on
Commit
a2ab101
Β·
1 Parent(s): e0502e4

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/trainer_state.json +798 -3
  9. model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629729537.2079122/events.out.tfevents.1629729537.74272264b15c.932.153 +3 -0
  11. model-bin/finetune/base/log/1629730178.309412/events.out.tfevents.1629730178.74272264b15c.932.155 +3 -0
  12. model-bin/finetune/base/log/1629730824.9113455/events.out.tfevents.1629730824.74272264b15c.932.157 +3 -0
  13. model-bin/finetune/base/log/1629731477.0083396/events.out.tfevents.1629731477.74272264b15c.932.159 +3 -0
  14. model-bin/finetune/base/log/1629732120.6295552/events.out.tfevents.1629732120.74272264b15c.932.161 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629729537.74272264b15c.932.152 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629730178.74272264b15c.932.154 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629730824.74272264b15c.932.156 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629731477.74272264b15c.932.158 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629732120.74272264b15c.932.160 +3 -0
model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c6f1253e46cde7d178f39e6b17a8972a3aad4111872e5bf27d7b6dca137c72c2
3
  size 722165009
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b6ae36f1470151fcec1ad3152a33a1a67095e63b48c5439b20bcf248749fd62
3
  size 722165009
model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e40873fa8da493e92e672079326593cb7f57a26d199c8587c88979165d3b405d
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26445c09a29688f4d145954a54e6e0b49beda1836a932a18751c83fe7ab4143d
3
  size 377909911
model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8655ea7d9fd7f1ddbb7f3c0285d61ea693bed9b3fcbbe176e4c78f2515296b7f
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e4fe6d0d49281984fdb840043efd745377fb7293293dd4c999cf5b86fcb8cc9
3
  size 14503
model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f26c7b4da20f1cc9a4ffd34476828968f627bc9309919fb52d579c167c07ee6a
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ff79ea670b032ae76eb17263d0356be4bded002f7ffdf7b390d21305eda812a
3
  size 559
model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:31131468d789f262e6e19dc1f7e277df2e3b8527cff32367b76129cd61996678
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9f8dfa58cc10ed162b1196ed5f37a00019f0b2ee7112b758926599597261135
3
  size 623
model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.19748327029386092,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-46666",
4
- "epoch": 381.0,
5
- "global_step": 47663,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -160005,11 +160005,806 @@
160005
  "eval_steps_per_second": 0.655,
160006
  "eval_wer": 0.20148385219668316,
160007
  "step": 47663
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
160008
  }
160009
  ],
160010
  "max_steps": 625000,
160011
  "num_train_epochs": 5000,
160012
- "total_flos": 1.3410923324603713e+20,
160013
  "trial_name": null,
160014
  "trial_params": null
160015
  }
 
1
  {
2
  "best_metric": 0.19748327029386092,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-46666",
4
+ "epoch": 386.0,
5
+ "global_step": 48288,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
160005
  "eval_steps_per_second": 0.655,
160006
  "eval_wer": 0.20148385219668316,
160007
  "step": 47663
160008
+ },
160009
+ {
160010
+ "epoch": 381.02,
160011
+ "learning_rate": 9.252451923076923e-06,
160012
+ "loss": 0.3971,
160013
+ "step": 47665
160014
+ },
160015
+ {
160016
+ "epoch": 381.06,
160017
+ "learning_rate": 9.252371794871796e-06,
160018
+ "loss": 0.3389,
160019
+ "step": 47670
160020
+ },
160021
+ {
160022
+ "epoch": 381.1,
160023
+ "learning_rate": 9.252291666666668e-06,
160024
+ "loss": 0.323,
160025
+ "step": 47675
160026
+ },
160027
+ {
160028
+ "epoch": 381.14,
160029
+ "learning_rate": 9.25221153846154e-06,
160030
+ "loss": 0.4617,
160031
+ "step": 47680
160032
+ },
160033
+ {
160034
+ "epoch": 381.18,
160035
+ "learning_rate": 9.25213141025641e-06,
160036
+ "loss": 0.7376,
160037
+ "step": 47685
160038
+ },
160039
+ {
160040
+ "epoch": 381.22,
160041
+ "learning_rate": 9.252051282051284e-06,
160042
+ "loss": 1.0757,
160043
+ "step": 47690
160044
+ },
160045
+ {
160046
+ "epoch": 381.26,
160047
+ "learning_rate": 9.251971153846155e-06,
160048
+ "loss": 0.2988,
160049
+ "step": 47695
160050
+ },
160051
+ {
160052
+ "epoch": 381.3,
160053
+ "learning_rate": 9.251891025641026e-06,
160054
+ "loss": 0.3414,
160055
+ "step": 47700
160056
+ },
160057
+ {
160058
+ "epoch": 381.34,
160059
+ "learning_rate": 9.251810897435898e-06,
160060
+ "loss": 0.3587,
160061
+ "step": 47705
160062
+ },
160063
+ {
160064
+ "epoch": 381.38,
160065
+ "learning_rate": 9.25173076923077e-06,
160066
+ "loss": 0.6515,
160067
+ "step": 47710
160068
+ },
160069
+ {
160070
+ "epoch": 381.42,
160071
+ "learning_rate": 9.251650641025642e-06,
160072
+ "loss": 1.1233,
160073
+ "step": 47715
160074
+ },
160075
+ {
160076
+ "epoch": 381.46,
160077
+ "learning_rate": 9.251570512820513e-06,
160078
+ "loss": 0.4138,
160079
+ "step": 47720
160080
+ },
160081
+ {
160082
+ "epoch": 381.5,
160083
+ "learning_rate": 9.251490384615386e-06,
160084
+ "loss": 0.3441,
160085
+ "step": 47725
160086
+ },
160087
+ {
160088
+ "epoch": 381.54,
160089
+ "learning_rate": 9.251410256410258e-06,
160090
+ "loss": 0.4364,
160091
+ "step": 47730
160092
+ },
160093
+ {
160094
+ "epoch": 381.58,
160095
+ "learning_rate": 9.25133012820513e-06,
160096
+ "loss": 0.7155,
160097
+ "step": 47735
160098
+ },
160099
+ {
160100
+ "epoch": 381.62,
160101
+ "learning_rate": 9.25125e-06,
160102
+ "loss": 1.246,
160103
+ "step": 47740
160104
+ },
160105
+ {
160106
+ "epoch": 381.66,
160107
+ "learning_rate": 9.251169871794874e-06,
160108
+ "loss": 0.3348,
160109
+ "step": 47745
160110
+ },
160111
+ {
160112
+ "epoch": 381.7,
160113
+ "learning_rate": 9.251089743589743e-06,
160114
+ "loss": 0.3715,
160115
+ "step": 47750
160116
+ },
160117
+ {
160118
+ "epoch": 381.74,
160119
+ "learning_rate": 9.251009615384616e-06,
160120
+ "loss": 0.37,
160121
+ "step": 47755
160122
+ },
160123
+ {
160124
+ "epoch": 381.78,
160125
+ "learning_rate": 9.250929487179488e-06,
160126
+ "loss": 0.6945,
160127
+ "step": 47760
160128
+ },
160129
+ {
160130
+ "epoch": 381.82,
160131
+ "learning_rate": 9.250849358974359e-06,
160132
+ "loss": 1.1434,
160133
+ "step": 47765
160134
+ },
160135
+ {
160136
+ "epoch": 381.86,
160137
+ "learning_rate": 9.250769230769232e-06,
160138
+ "loss": 0.3241,
160139
+ "step": 47770
160140
+ },
160141
+ {
160142
+ "epoch": 381.9,
160143
+ "learning_rate": 9.250689102564103e-06,
160144
+ "loss": 0.3328,
160145
+ "step": 47775
160146
+ },
160147
+ {
160148
+ "epoch": 381.94,
160149
+ "learning_rate": 9.250608974358975e-06,
160150
+ "loss": 0.4408,
160151
+ "step": 47780
160152
+ },
160153
+ {
160154
+ "epoch": 381.98,
160155
+ "learning_rate": 9.250528846153846e-06,
160156
+ "loss": 0.7996,
160157
+ "step": 47785
160158
+ },
160159
+ {
160160
+ "epoch": 382.0,
160161
+ "eval_loss": 0.4474487900733948,
160162
+ "eval_runtime": 39.2412,
160163
+ "eval_samples_per_second": 21.279,
160164
+ "eval_steps_per_second": 0.688,
160165
+ "eval_wer": 0.20729635920537626,
160166
+ "step": 47788
160167
+ },
160168
+ {
160169
+ "epoch": 382.02,
160170
+ "learning_rate": 9.25044871794872e-06,
160171
+ "loss": 0.4093,
160172
+ "step": 47790
160173
+ },
160174
+ {
160175
+ "epoch": 382.06,
160176
+ "learning_rate": 9.25036858974359e-06,
160177
+ "loss": 0.3645,
160178
+ "step": 47795
160179
+ },
160180
+ {
160181
+ "epoch": 382.1,
160182
+ "learning_rate": 9.250288461538462e-06,
160183
+ "loss": 0.3569,
160184
+ "step": 47800
160185
+ },
160186
+ {
160187
+ "epoch": 382.14,
160188
+ "learning_rate": 9.250208333333333e-06,
160189
+ "loss": 0.361,
160190
+ "step": 47805
160191
+ },
160192
+ {
160193
+ "epoch": 382.18,
160194
+ "learning_rate": 9.250128205128206e-06,
160195
+ "loss": 0.606,
160196
+ "step": 47810
160197
+ },
160198
+ {
160199
+ "epoch": 382.22,
160200
+ "learning_rate": 9.250048076923078e-06,
160201
+ "loss": 1.1471,
160202
+ "step": 47815
160203
+ },
160204
+ {
160205
+ "epoch": 382.26,
160206
+ "learning_rate": 9.249967948717949e-06,
160207
+ "loss": 0.3549,
160208
+ "step": 47820
160209
+ },
160210
+ {
160211
+ "epoch": 382.3,
160212
+ "learning_rate": 9.249887820512822e-06,
160213
+ "loss": 0.2992,
160214
+ "step": 47825
160215
+ },
160216
+ {
160217
+ "epoch": 382.34,
160218
+ "learning_rate": 9.249807692307693e-06,
160219
+ "loss": 0.4853,
160220
+ "step": 47830
160221
+ },
160222
+ {
160223
+ "epoch": 382.38,
160224
+ "learning_rate": 9.249727564102565e-06,
160225
+ "loss": 0.7708,
160226
+ "step": 47835
160227
+ },
160228
+ {
160229
+ "epoch": 382.42,
160230
+ "learning_rate": 9.249647435897436e-06,
160231
+ "loss": 1.1267,
160232
+ "step": 47840
160233
+ },
160234
+ {
160235
+ "epoch": 382.46,
160236
+ "learning_rate": 9.24956730769231e-06,
160237
+ "loss": 0.3346,
160238
+ "step": 47845
160239
+ },
160240
+ {
160241
+ "epoch": 382.5,
160242
+ "learning_rate": 9.24948717948718e-06,
160243
+ "loss": 0.3632,
160244
+ "step": 47850
160245
+ },
160246
+ {
160247
+ "epoch": 382.54,
160248
+ "learning_rate": 9.249407051282052e-06,
160249
+ "loss": 0.3969,
160250
+ "step": 47855
160251
+ },
160252
+ {
160253
+ "epoch": 382.58,
160254
+ "learning_rate": 9.249326923076923e-06,
160255
+ "loss": 0.7573,
160256
+ "step": 47860
160257
+ },
160258
+ {
160259
+ "epoch": 382.62,
160260
+ "learning_rate": 9.249246794871796e-06,
160261
+ "loss": 1.1858,
160262
+ "step": 47865
160263
+ },
160264
+ {
160265
+ "epoch": 382.66,
160266
+ "learning_rate": 9.249166666666668e-06,
160267
+ "loss": 0.3861,
160268
+ "step": 47870
160269
+ },
160270
+ {
160271
+ "epoch": 382.7,
160272
+ "learning_rate": 9.249086538461539e-06,
160273
+ "loss": 0.3233,
160274
+ "step": 47875
160275
+ },
160276
+ {
160277
+ "epoch": 382.74,
160278
+ "learning_rate": 9.249006410256412e-06,
160279
+ "loss": 0.4369,
160280
+ "step": 47880
160281
+ },
160282
+ {
160283
+ "epoch": 382.78,
160284
+ "learning_rate": 9.248926282051282e-06,
160285
+ "loss": 0.6703,
160286
+ "step": 47885
160287
+ },
160288
+ {
160289
+ "epoch": 382.82,
160290
+ "learning_rate": 9.248846153846155e-06,
160291
+ "loss": 1.1758,
160292
+ "step": 47890
160293
+ },
160294
+ {
160295
+ "epoch": 382.86,
160296
+ "learning_rate": 9.248766025641026e-06,
160297
+ "loss": 0.2987,
160298
+ "step": 47895
160299
+ },
160300
+ {
160301
+ "epoch": 382.9,
160302
+ "learning_rate": 9.248685897435898e-06,
160303
+ "loss": 0.3724,
160304
+ "step": 47900
160305
+ },
160306
+ {
160307
+ "epoch": 382.94,
160308
+ "learning_rate": 9.248605769230769e-06,
160309
+ "loss": 0.4235,
160310
+ "step": 47905
160311
+ },
160312
+ {
160313
+ "epoch": 382.98,
160314
+ "learning_rate": 9.248525641025642e-06,
160315
+ "loss": 0.6528,
160316
+ "step": 47910
160317
+ },
160318
+ {
160319
+ "epoch": 383.0,
160320
+ "eval_loss": 0.4338654577732086,
160321
+ "eval_runtime": 39.3369,
160322
+ "eval_samples_per_second": 21.227,
160323
+ "eval_steps_per_second": 0.686,
160324
+ "eval_wer": 0.21281085154483798,
160325
+ "step": 47913
160326
+ },
160327
+ {
160328
+ "epoch": 383.02,
160329
+ "learning_rate": 9.248445512820513e-06,
160330
+ "loss": 0.3667,
160331
+ "step": 47915
160332
+ },
160333
+ {
160334
+ "epoch": 383.06,
160335
+ "learning_rate": 9.248365384615385e-06,
160336
+ "loss": 0.3526,
160337
+ "step": 47920
160338
+ },
160339
+ {
160340
+ "epoch": 383.1,
160341
+ "learning_rate": 9.248285256410258e-06,
160342
+ "loss": 0.3728,
160343
+ "step": 47925
160344
+ },
160345
+ {
160346
+ "epoch": 383.14,
160347
+ "learning_rate": 9.248205128205129e-06,
160348
+ "loss": 0.3497,
160349
+ "step": 47930
160350
+ },
160351
+ {
160352
+ "epoch": 383.18,
160353
+ "learning_rate": 9.248125e-06,
160354
+ "loss": 0.7086,
160355
+ "step": 47935
160356
+ },
160357
+ {
160358
+ "epoch": 383.22,
160359
+ "learning_rate": 9.248044871794872e-06,
160360
+ "loss": 1.2526,
160361
+ "step": 47940
160362
+ },
160363
+ {
160364
+ "epoch": 383.26,
160365
+ "learning_rate": 9.247964743589745e-06,
160366
+ "loss": 0.3158,
160367
+ "step": 47945
160368
+ },
160369
+ {
160370
+ "epoch": 383.3,
160371
+ "learning_rate": 9.247884615384616e-06,
160372
+ "loss": 0.3663,
160373
+ "step": 47950
160374
+ },
160375
+ {
160376
+ "epoch": 383.34,
160377
+ "learning_rate": 9.247804487179488e-06,
160378
+ "loss": 0.4438,
160379
+ "step": 47955
160380
+ },
160381
+ {
160382
+ "epoch": 383.38,
160383
+ "learning_rate": 9.247724358974359e-06,
160384
+ "loss": 0.726,
160385
+ "step": 47960
160386
+ },
160387
+ {
160388
+ "epoch": 383.42,
160389
+ "learning_rate": 9.247644230769232e-06,
160390
+ "loss": 1.2397,
160391
+ "step": 47965
160392
+ },
160393
+ {
160394
+ "epoch": 383.46,
160395
+ "learning_rate": 9.247564102564103e-06,
160396
+ "loss": 0.3575,
160397
+ "step": 47970
160398
+ },
160399
+ {
160400
+ "epoch": 383.5,
160401
+ "learning_rate": 9.247483974358975e-06,
160402
+ "loss": 0.3929,
160403
+ "step": 47975
160404
+ },
160405
+ {
160406
+ "epoch": 383.54,
160407
+ "learning_rate": 9.247403846153848e-06,
160408
+ "loss": 0.3886,
160409
+ "step": 47980
160410
+ },
160411
+ {
160412
+ "epoch": 383.58,
160413
+ "learning_rate": 9.24732371794872e-06,
160414
+ "loss": 0.6418,
160415
+ "step": 47985
160416
+ },
160417
+ {
160418
+ "epoch": 383.62,
160419
+ "learning_rate": 9.24724358974359e-06,
160420
+ "loss": 1.1219,
160421
+ "step": 47990
160422
+ },
160423
+ {
160424
+ "epoch": 383.66,
160425
+ "learning_rate": 9.247163461538462e-06,
160426
+ "loss": 0.3736,
160427
+ "step": 47995
160428
+ },
160429
+ {
160430
+ "epoch": 383.7,
160431
+ "learning_rate": 9.247083333333335e-06,
160432
+ "loss": 0.3584,
160433
+ "step": 48000
160434
+ },
160435
+ {
160436
+ "epoch": 383.74,
160437
+ "learning_rate": 9.247003205128206e-06,
160438
+ "loss": 0.4055,
160439
+ "step": 48005
160440
+ },
160441
+ {
160442
+ "epoch": 383.78,
160443
+ "learning_rate": 9.246923076923078e-06,
160444
+ "loss": 0.7403,
160445
+ "step": 48010
160446
+ },
160447
+ {
160448
+ "epoch": 383.82,
160449
+ "learning_rate": 9.246842948717949e-06,
160450
+ "loss": 1.0718,
160451
+ "step": 48015
160452
+ },
160453
+ {
160454
+ "epoch": 383.86,
160455
+ "learning_rate": 9.246762820512822e-06,
160456
+ "loss": 0.3165,
160457
+ "step": 48020
160458
+ },
160459
+ {
160460
+ "epoch": 383.9,
160461
+ "learning_rate": 9.246682692307693e-06,
160462
+ "loss": 0.3457,
160463
+ "step": 48025
160464
+ },
160465
+ {
160466
+ "epoch": 383.94,
160467
+ "learning_rate": 9.246602564102565e-06,
160468
+ "loss": 0.3643,
160469
+ "step": 48030
160470
+ },
160471
+ {
160472
+ "epoch": 383.98,
160473
+ "learning_rate": 9.246522435897438e-06,
160474
+ "loss": 0.6838,
160475
+ "step": 48035
160476
+ },
160477
+ {
160478
+ "epoch": 384.0,
160479
+ "eval_loss": 0.4506882131099701,
160480
+ "eval_runtime": 40.789,
160481
+ "eval_samples_per_second": 20.471,
160482
+ "eval_steps_per_second": 0.662,
160483
+ "eval_wer": 0.20609548167092925,
160484
+ "step": 48038
160485
+ },
160486
+ {
160487
+ "epoch": 384.02,
160488
+ "learning_rate": 9.246442307692307e-06,
160489
+ "loss": 0.4569,
160490
+ "step": 48040
160491
+ },
160492
+ {
160493
+ "epoch": 384.06,
160494
+ "learning_rate": 9.24636217948718e-06,
160495
+ "loss": 0.3582,
160496
+ "step": 48045
160497
+ },
160498
+ {
160499
+ "epoch": 384.1,
160500
+ "learning_rate": 9.246282051282052e-06,
160501
+ "loss": 0.334,
160502
+ "step": 48050
160503
+ },
160504
+ {
160505
+ "epoch": 384.14,
160506
+ "learning_rate": 9.246201923076923e-06,
160507
+ "loss": 0.4399,
160508
+ "step": 48055
160509
+ },
160510
+ {
160511
+ "epoch": 384.18,
160512
+ "learning_rate": 9.246121794871795e-06,
160513
+ "loss": 0.6171,
160514
+ "step": 48060
160515
+ },
160516
+ {
160517
+ "epoch": 384.22,
160518
+ "learning_rate": 9.246041666666668e-06,
160519
+ "loss": 1.0357,
160520
+ "step": 48065
160521
+ },
160522
+ {
160523
+ "epoch": 384.26,
160524
+ "learning_rate": 9.245961538461539e-06,
160525
+ "loss": 0.3912,
160526
+ "step": 48070
160527
+ },
160528
+ {
160529
+ "epoch": 384.3,
160530
+ "learning_rate": 9.24588141025641e-06,
160531
+ "loss": 0.438,
160532
+ "step": 48075
160533
+ },
160534
+ {
160535
+ "epoch": 384.34,
160536
+ "learning_rate": 9.245801282051283e-06,
160537
+ "loss": 0.402,
160538
+ "step": 48080
160539
+ },
160540
+ {
160541
+ "epoch": 384.38,
160542
+ "learning_rate": 9.245721153846155e-06,
160543
+ "loss": 0.6041,
160544
+ "step": 48085
160545
+ },
160546
+ {
160547
+ "epoch": 384.42,
160548
+ "learning_rate": 9.245641025641026e-06,
160549
+ "loss": 1.202,
160550
+ "step": 48090
160551
+ },
160552
+ {
160553
+ "epoch": 384.46,
160554
+ "learning_rate": 9.245560897435898e-06,
160555
+ "loss": 0.3859,
160556
+ "step": 48095
160557
+ },
160558
+ {
160559
+ "epoch": 384.5,
160560
+ "learning_rate": 9.24548076923077e-06,
160561
+ "loss": 0.3481,
160562
+ "step": 48100
160563
+ },
160564
+ {
160565
+ "epoch": 384.54,
160566
+ "learning_rate": 9.245400641025642e-06,
160567
+ "loss": 0.4098,
160568
+ "step": 48105
160569
+ },
160570
+ {
160571
+ "epoch": 384.58,
160572
+ "learning_rate": 9.245320512820513e-06,
160573
+ "loss": 0.7351,
160574
+ "step": 48110
160575
+ },
160576
+ {
160577
+ "epoch": 384.62,
160578
+ "learning_rate": 9.245240384615385e-06,
160579
+ "loss": 1.0795,
160580
+ "step": 48115
160581
+ },
160582
+ {
160583
+ "epoch": 384.66,
160584
+ "learning_rate": 9.245160256410258e-06,
160585
+ "loss": 0.3069,
160586
+ "step": 48120
160587
+ },
160588
+ {
160589
+ "epoch": 384.7,
160590
+ "learning_rate": 9.245080128205129e-06,
160591
+ "loss": 0.3185,
160592
+ "step": 48125
160593
+ },
160594
+ {
160595
+ "epoch": 384.74,
160596
+ "learning_rate": 9.245e-06,
160597
+ "loss": 0.4013,
160598
+ "step": 48130
160599
+ },
160600
+ {
160601
+ "epoch": 384.78,
160602
+ "learning_rate": 9.244919871794873e-06,
160603
+ "loss": 0.6792,
160604
+ "step": 48135
160605
+ },
160606
+ {
160607
+ "epoch": 384.82,
160608
+ "learning_rate": 9.244839743589745e-06,
160609
+ "loss": 1.153,
160610
+ "step": 48140
160611
+ },
160612
+ {
160613
+ "epoch": 384.86,
160614
+ "learning_rate": 9.244759615384616e-06,
160615
+ "loss": 0.2832,
160616
+ "step": 48145
160617
+ },
160618
+ {
160619
+ "epoch": 384.9,
160620
+ "learning_rate": 9.244679487179488e-06,
160621
+ "loss": 0.4042,
160622
+ "step": 48150
160623
+ },
160624
+ {
160625
+ "epoch": 384.94,
160626
+ "learning_rate": 9.24459935897436e-06,
160627
+ "loss": 0.4558,
160628
+ "step": 48155
160629
+ },
160630
+ {
160631
+ "epoch": 384.98,
160632
+ "learning_rate": 9.24451923076923e-06,
160633
+ "loss": 0.7045,
160634
+ "step": 48160
160635
+ },
160636
+ {
160637
+ "epoch": 385.0,
160638
+ "eval_loss": 0.487787127494812,
160639
+ "eval_runtime": 39.5507,
160640
+ "eval_samples_per_second": 21.137,
160641
+ "eval_steps_per_second": 0.683,
160642
+ "eval_wer": 0.2028522617544381,
160643
+ "step": 48163
160644
+ },
160645
+ {
160646
+ "epoch": 385.02,
160647
+ "learning_rate": 9.244439102564103e-06,
160648
+ "loss": 0.3803,
160649
+ "step": 48165
160650
+ },
160651
+ {
160652
+ "epoch": 385.06,
160653
+ "learning_rate": 9.244358974358976e-06,
160654
+ "loss": 0.6276,
160655
+ "step": 48170
160656
+ },
160657
+ {
160658
+ "epoch": 385.1,
160659
+ "learning_rate": 9.244278846153846e-06,
160660
+ "loss": 0.3612,
160661
+ "step": 48175
160662
+ },
160663
+ {
160664
+ "epoch": 385.14,
160665
+ "learning_rate": 9.244198717948719e-06,
160666
+ "loss": 0.4303,
160667
+ "step": 48180
160668
+ },
160669
+ {
160670
+ "epoch": 385.18,
160671
+ "learning_rate": 9.24411858974359e-06,
160672
+ "loss": 0.7484,
160673
+ "step": 48185
160674
+ },
160675
+ {
160676
+ "epoch": 385.22,
160677
+ "learning_rate": 9.244038461538462e-06,
160678
+ "loss": 1.0026,
160679
+ "step": 48190
160680
+ },
160681
+ {
160682
+ "epoch": 385.26,
160683
+ "learning_rate": 9.243958333333333e-06,
160684
+ "loss": 0.3777,
160685
+ "step": 48195
160686
+ },
160687
+ {
160688
+ "epoch": 385.3,
160689
+ "learning_rate": 9.243878205128206e-06,
160690
+ "loss": 0.3769,
160691
+ "step": 48200
160692
+ },
160693
+ {
160694
+ "epoch": 385.34,
160695
+ "learning_rate": 9.243798076923078e-06,
160696
+ "loss": 0.3786,
160697
+ "step": 48205
160698
+ },
160699
+ {
160700
+ "epoch": 385.38,
160701
+ "learning_rate": 9.243717948717949e-06,
160702
+ "loss": 0.8006,
160703
+ "step": 48210
160704
+ },
160705
+ {
160706
+ "epoch": 385.42,
160707
+ "learning_rate": 9.24363782051282e-06,
160708
+ "loss": 1.2286,
160709
+ "step": 48215
160710
+ },
160711
+ {
160712
+ "epoch": 385.46,
160713
+ "learning_rate": 9.243557692307693e-06,
160714
+ "loss": 0.3582,
160715
+ "step": 48220
160716
+ },
160717
+ {
160718
+ "epoch": 385.5,
160719
+ "learning_rate": 9.243477564102565e-06,
160720
+ "loss": 0.4509,
160721
+ "step": 48225
160722
+ },
160723
+ {
160724
+ "epoch": 385.54,
160725
+ "learning_rate": 9.243397435897436e-06,
160726
+ "loss": 0.4411,
160727
+ "step": 48230
160728
+ },
160729
+ {
160730
+ "epoch": 385.58,
160731
+ "learning_rate": 9.243317307692309e-06,
160732
+ "loss": 0.9501,
160733
+ "step": 48235
160734
+ },
160735
+ {
160736
+ "epoch": 385.62,
160737
+ "learning_rate": 9.24323717948718e-06,
160738
+ "loss": 1.132,
160739
+ "step": 48240
160740
+ },
160741
+ {
160742
+ "epoch": 385.66,
160743
+ "learning_rate": 9.243157051282052e-06,
160744
+ "loss": 0.3727,
160745
+ "step": 48245
160746
+ },
160747
+ {
160748
+ "epoch": 385.7,
160749
+ "learning_rate": 9.243076923076923e-06,
160750
+ "loss": 0.349,
160751
+ "step": 48250
160752
+ },
160753
+ {
160754
+ "epoch": 385.74,
160755
+ "learning_rate": 9.242996794871796e-06,
160756
+ "loss": 0.3987,
160757
+ "step": 48255
160758
+ },
160759
+ {
160760
+ "epoch": 385.78,
160761
+ "learning_rate": 9.242916666666668e-06,
160762
+ "loss": 0.6941,
160763
+ "step": 48260
160764
+ },
160765
+ {
160766
+ "epoch": 385.82,
160767
+ "learning_rate": 9.242836538461539e-06,
160768
+ "loss": 1.2013,
160769
+ "step": 48265
160770
+ },
160771
+ {
160772
+ "epoch": 385.86,
160773
+ "learning_rate": 9.24275641025641e-06,
160774
+ "loss": 0.3445,
160775
+ "step": 48270
160776
+ },
160777
+ {
160778
+ "epoch": 385.9,
160779
+ "learning_rate": 9.242676282051283e-06,
160780
+ "loss": 0.4103,
160781
+ "step": 48275
160782
+ },
160783
+ {
160784
+ "epoch": 385.94,
160785
+ "learning_rate": 9.242596153846155e-06,
160786
+ "loss": 0.3933,
160787
+ "step": 48280
160788
+ },
160789
+ {
160790
+ "epoch": 385.98,
160791
+ "learning_rate": 9.242516025641026e-06,
160792
+ "loss": 0.7593,
160793
+ "step": 48285
160794
+ },
160795
+ {
160796
+ "epoch": 386.0,
160797
+ "eval_loss": 0.43986836075782776,
160798
+ "eval_runtime": 42.306,
160799
+ "eval_samples_per_second": 19.761,
160800
+ "eval_steps_per_second": 0.638,
160801
+ "eval_wer": 0.21344501408043903,
160802
+ "step": 48288
160803
  }
160804
  ],
160805
  "max_steps": 625000,
160806
  "num_train_epochs": 5000,
160807
+ "total_flos": 1.3586798852174347e+20,
160808
  "trial_name": null,
160809
  "trial_params": null
160810
  }
model-bin/finetune/base/{checkpoint-47663 β†’ checkpoint-48288}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629729537.2079122/events.out.tfevents.1629729537.74272264b15c.932.153 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:657a136ddcc4499ea27567347ebe2644f8af1c6172d51209aa7d9f1afd5234f2
3
+ size 4194
model-bin/finetune/base/log/1629730178.309412/events.out.tfevents.1629730178.74272264b15c.932.155 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3636cd11adfc23bec228de24951199d9184dce52516bc83e1def29ee958e610
3
+ size 4194
model-bin/finetune/base/log/1629730824.9113455/events.out.tfevents.1629730824.74272264b15c.932.157 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1fdd45dfe899c2a80fa2732458736b1f6bfd7216dfa6317d1507c410aa37accb
3
+ size 4194
model-bin/finetune/base/log/1629731477.0083396/events.out.tfevents.1629731477.74272264b15c.932.159 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:432a4a4291df8feb956101dd81997cb49db02a729ba2dc6661ec43bb2c43872b
3
+ size 4194
model-bin/finetune/base/log/1629732120.6295552/events.out.tfevents.1629732120.74272264b15c.932.161 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a7f4c731a57cc22e079308f9ff1b2ea6e0efcf0d7d49fb5f5784bc07fc60890
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629729537.74272264b15c.932.152 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fa03af12aac867931e68772572da097d641158f5792d54f3699b993f372e9c8
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629730178.74272264b15c.932.154 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2d6c61d8e9be86d165815c4208a5ceb4459fe1a29e8ae544d373b891b096cdc
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629730824.74272264b15c.932.156 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:708d537067aa55f754951c51ce323249dde515b8a154107449dc039375296f39
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629731477.74272264b15c.932.158 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c69aa0234d4e6b4c46ea7eeb5e5ebb8eea5e78d5c5a08152c02397a5418345bd
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629732120.74272264b15c.932.160 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0ee6172e7980f77301b88fe7e6f9a1c84039a3df2e22026fe0b7a1323ed3bcc
3
+ size 8622