Check commited on
Commit
5cffbbe
Β·
1 Parent(s): 6ed1a9d

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/trainer_state.json +793 -4
  9. model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630237299.1184735/events.out.tfevents.1630237299.cc93b136ebf5.1086.149 +3 -0
  11. model-bin/finetune/base/log/1630237737.6447396/events.out.tfevents.1630237737.cc93b136ebf5.1086.151 +3 -0
  12. model-bin/finetune/base/log/1630238176.5216558/events.out.tfevents.1630238176.cc93b136ebf5.1086.153 +3 -0
  13. model-bin/finetune/base/log/1630238608.102783/events.out.tfevents.1630238608.cc93b136ebf5.1086.155 +3 -0
  14. model-bin/finetune/base/log/1630239044.210316/events.out.tfevents.1630239044.cc93b136ebf5.1086.157 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630237299.cc93b136ebf5.1086.148 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630237737.cc93b136ebf5.1086.150 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630238176.cc93b136ebf5.1086.152 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630238608.cc93b136ebf5.1086.154 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630239044.cc93b136ebf5.1086.156 +3 -0
model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca053a21f43a42ef9039777b97798ad4984877959c3297c934cb79b5f9280b8d
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f93d5bc8ae226e1c1aee31035ac894051b19c4bbe441659018307488de04cd85
3
  size 722165393
model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f6a1ad1b6bdf94b092b4dbaadb21109adb1abf08755cc31224793164fd555d6
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93a42bd5cb554963f46e2895d26148d57b1cf04e1e671d3cc0e6a201be5624e6
3
  size 377909911
model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc771b6e0d0fae529a3ac50ef6ed92b0fc4e02d966c231e4ea817bf981f64442
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4057531845a0717c33153a0d43d59bbf8df52909b80221bb96290592a2c80c25
3
  size 14503
model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b54d1b931ef6d5d353a2168a55b89d51f96d90c49dc5e37c1e18abd323651d0d
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61520e9eb8eb51b6c2259c437be709f2dfc17a4076a25563394233f7ae24e463
3
  size 559
model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:352fbfc64d23a3b022c4a65026575cbcb8c45fcce17227e15ee0f0fca4bb6241
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cdb9e88c57bc6cc6a5791bc7ab637007af04603717e93fe539b17454fd741ec
3
  size 623
model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1668011442822563,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
4
- "epoch": 1195.995983935743,
5
- "global_step": 148340,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -288102,11 +288102,800 @@
288102
  "eval_steps_per_second": 0.633,
288103
  "eval_wer": 0.1786735277301315,
288104
  "step": 148340
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
288105
  }
288106
  ],
288107
- "max_steps": 620000,
288108
  "num_train_epochs": 5000,
288109
- "total_flos": 4.1744157123227523e+20,
288110
  "trial_name": null,
288111
  "trial_params": null
288112
  }
 
1
  {
2
  "best_metric": 0.1668011442822563,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
4
+ "epoch": 1191.0,
5
+ "global_step": 148960,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
288102
  "eval_steps_per_second": 0.633,
288103
  "eval_wer": 0.1786735277301315,
288104
  "step": 148340
288105
+ },
288106
+ {
288107
+ "epoch": 1196.04,
288108
+ "learning_rate": 7.620581583198708e-06,
288109
+ "loss": 0.2904,
288110
+ "step": 148345
288111
+ },
288112
+ {
288113
+ "epoch": 1196.08,
288114
+ "learning_rate": 7.620500807754443e-06,
288115
+ "loss": 0.2934,
288116
+ "step": 148350
288117
+ },
288118
+ {
288119
+ "epoch": 1196.12,
288120
+ "learning_rate": 7.620420032310178e-06,
288121
+ "loss": 0.2976,
288122
+ "step": 148355
288123
+ },
288124
+ {
288125
+ "epoch": 1196.16,
288126
+ "learning_rate": 7.620339256865913e-06,
288127
+ "loss": 0.4439,
288128
+ "step": 148360
288129
+ },
288130
+ {
288131
+ "epoch": 1196.2,
288132
+ "learning_rate": 7.620258481421648e-06,
288133
+ "loss": 1.1461,
288134
+ "step": 148365
288135
+ },
288136
+ {
288137
+ "epoch": 1196.24,
288138
+ "learning_rate": 7.620177705977384e-06,
288139
+ "loss": 0.3897,
288140
+ "step": 148370
288141
+ },
288142
+ {
288143
+ "epoch": 1196.28,
288144
+ "learning_rate": 7.620096930533119e-06,
288145
+ "loss": 0.2371,
288146
+ "step": 148375
288147
+ },
288148
+ {
288149
+ "epoch": 1196.32,
288150
+ "learning_rate": 7.620016155088854e-06,
288151
+ "loss": 0.2599,
288152
+ "step": 148380
288153
+ },
288154
+ {
288155
+ "epoch": 1196.36,
288156
+ "learning_rate": 7.619935379644589e-06,
288157
+ "loss": 0.4827,
288158
+ "step": 148385
288159
+ },
288160
+ {
288161
+ "epoch": 1196.4,
288162
+ "learning_rate": 7.619854604200324e-06,
288163
+ "loss": 1.2766,
288164
+ "step": 148390
288165
+ },
288166
+ {
288167
+ "epoch": 1196.44,
288168
+ "learning_rate": 7.619773828756059e-06,
288169
+ "loss": 0.3409,
288170
+ "step": 148395
288171
+ },
288172
+ {
288173
+ "epoch": 1196.48,
288174
+ "learning_rate": 7.619693053311794e-06,
288175
+ "loss": 0.2787,
288176
+ "step": 148400
288177
+ },
288178
+ {
288179
+ "epoch": 1196.52,
288180
+ "learning_rate": 7.619612277867529e-06,
288181
+ "loss": 0.2924,
288182
+ "step": 148405
288183
+ },
288184
+ {
288185
+ "epoch": 1196.56,
288186
+ "learning_rate": 7.619531502423264e-06,
288187
+ "loss": 0.396,
288188
+ "step": 148410
288189
+ },
288190
+ {
288191
+ "epoch": 1196.6,
288192
+ "learning_rate": 7.619450726978999e-06,
288193
+ "loss": 1.151,
288194
+ "step": 148415
288195
+ },
288196
+ {
288197
+ "epoch": 1196.64,
288198
+ "learning_rate": 7.619369951534734e-06,
288199
+ "loss": 0.2854,
288200
+ "step": 148420
288201
+ },
288202
+ {
288203
+ "epoch": 1196.68,
288204
+ "learning_rate": 7.619289176090469e-06,
288205
+ "loss": 0.269,
288206
+ "step": 148425
288207
+ },
288208
+ {
288209
+ "epoch": 1196.72,
288210
+ "learning_rate": 7.619208400646204e-06,
288211
+ "loss": 0.2801,
288212
+ "step": 148430
288213
+ },
288214
+ {
288215
+ "epoch": 1196.76,
288216
+ "learning_rate": 7.61912762520194e-06,
288217
+ "loss": 0.4201,
288218
+ "step": 148435
288219
+ },
288220
+ {
288221
+ "epoch": 1196.8,
288222
+ "learning_rate": 7.619046849757674e-06,
288223
+ "loss": 1.1717,
288224
+ "step": 148440
288225
+ },
288226
+ {
288227
+ "epoch": 1196.84,
288228
+ "learning_rate": 7.61896607431341e-06,
288229
+ "loss": 0.299,
288230
+ "step": 148445
288231
+ },
288232
+ {
288233
+ "epoch": 1196.88,
288234
+ "learning_rate": 7.618885298869145e-06,
288235
+ "loss": 0.2666,
288236
+ "step": 148450
288237
+ },
288238
+ {
288239
+ "epoch": 1196.92,
288240
+ "learning_rate": 7.6188045234248796e-06,
288241
+ "loss": 0.2868,
288242
+ "step": 148455
288243
+ },
288244
+ {
288245
+ "epoch": 1196.96,
288246
+ "learning_rate": 7.6187237479806146e-06,
288247
+ "loss": 0.4595,
288248
+ "step": 148460
288249
+ },
288250
+ {
288251
+ "epoch": 1197.0,
288252
+ "eval_loss": 0.3382641077041626,
288253
+ "eval_runtime": 42.1842,
288254
+ "eval_samples_per_second": 19.936,
288255
+ "eval_steps_per_second": 0.64,
288256
+ "eval_wer": 0.17046594982078853,
288257
+ "step": 148464
288258
+ },
288259
+ {
288260
+ "epoch": 1197.01,
288261
+ "learning_rate": 7.6186429725363495e-06,
288262
+ "loss": 0.4864,
288263
+ "step": 148465
288264
+ },
288265
+ {
288266
+ "epoch": 1197.05,
288267
+ "learning_rate": 7.6185621970920845e-06,
288268
+ "loss": 0.2608,
288269
+ "step": 148470
288270
+ },
288271
+ {
288272
+ "epoch": 1197.09,
288273
+ "learning_rate": 7.6184814216478195e-06,
288274
+ "loss": 0.2822,
288275
+ "step": 148475
288276
+ },
288277
+ {
288278
+ "epoch": 1197.13,
288279
+ "learning_rate": 7.6184006462035545e-06,
288280
+ "loss": 0.3627,
288281
+ "step": 148480
288282
+ },
288283
+ {
288284
+ "epoch": 1197.17,
288285
+ "learning_rate": 7.6183198707592895e-06,
288286
+ "loss": 0.5249,
288287
+ "step": 148485
288288
+ },
288289
+ {
288290
+ "epoch": 1197.21,
288291
+ "learning_rate": 7.618239095315025e-06,
288292
+ "loss": 1.042,
288293
+ "step": 148490
288294
+ },
288295
+ {
288296
+ "epoch": 1197.25,
288297
+ "learning_rate": 7.6181583198707595e-06,
288298
+ "loss": 0.2996,
288299
+ "step": 148495
288300
+ },
288301
+ {
288302
+ "epoch": 1197.29,
288303
+ "learning_rate": 7.618077544426495e-06,
288304
+ "loss": 0.2382,
288305
+ "step": 148500
288306
+ },
288307
+ {
288308
+ "epoch": 1197.33,
288309
+ "learning_rate": 7.6179967689822294e-06,
288310
+ "loss": 0.3243,
288311
+ "step": 148505
288312
+ },
288313
+ {
288314
+ "epoch": 1197.37,
288315
+ "learning_rate": 7.617915993537965e-06,
288316
+ "loss": 0.4499,
288317
+ "step": 148510
288318
+ },
288319
+ {
288320
+ "epoch": 1197.41,
288321
+ "learning_rate": 7.6178352180937e-06,
288322
+ "loss": 1.0114,
288323
+ "step": 148515
288324
+ },
288325
+ {
288326
+ "epoch": 1197.45,
288327
+ "learning_rate": 7.617754442649435e-06,
288328
+ "loss": 0.2706,
288329
+ "step": 148520
288330
+ },
288331
+ {
288332
+ "epoch": 1197.49,
288333
+ "learning_rate": 7.61767366720517e-06,
288334
+ "loss": 0.2452,
288335
+ "step": 148525
288336
+ },
288337
+ {
288338
+ "epoch": 1197.53,
288339
+ "learning_rate": 7.617592891760905e-06,
288340
+ "loss": 0.325,
288341
+ "step": 148530
288342
+ },
288343
+ {
288344
+ "epoch": 1197.57,
288345
+ "learning_rate": 7.61751211631664e-06,
288346
+ "loss": 0.5252,
288347
+ "step": 148535
288348
+ },
288349
+ {
288350
+ "epoch": 1197.61,
288351
+ "learning_rate": 7.617431340872375e-06,
288352
+ "loss": 1.137,
288353
+ "step": 148540
288354
+ },
288355
+ {
288356
+ "epoch": 1197.65,
288357
+ "learning_rate": 7.617350565428111e-06,
288358
+ "loss": 0.3361,
288359
+ "step": 148545
288360
+ },
288361
+ {
288362
+ "epoch": 1197.69,
288363
+ "learning_rate": 7.617269789983845e-06,
288364
+ "loss": 0.3604,
288365
+ "step": 148550
288366
+ },
288367
+ {
288368
+ "epoch": 1197.73,
288369
+ "learning_rate": 7.617189014539581e-06,
288370
+ "loss": 0.3392,
288371
+ "step": 148555
288372
+ },
288373
+ {
288374
+ "epoch": 1197.77,
288375
+ "learning_rate": 7.617108239095315e-06,
288376
+ "loss": 0.5506,
288377
+ "step": 148560
288378
+ },
288379
+ {
288380
+ "epoch": 1197.81,
288381
+ "learning_rate": 7.617027463651051e-06,
288382
+ "loss": 1.1279,
288383
+ "step": 148565
288384
+ },
288385
+ {
288386
+ "epoch": 1197.85,
288387
+ "learning_rate": 7.616946688206785e-06,
288388
+ "loss": 0.3246,
288389
+ "step": 148570
288390
+ },
288391
+ {
288392
+ "epoch": 1197.89,
288393
+ "learning_rate": 7.616865912762521e-06,
288394
+ "loss": 0.2687,
288395
+ "step": 148575
288396
+ },
288397
+ {
288398
+ "epoch": 1197.93,
288399
+ "learning_rate": 7.616785137318256e-06,
288400
+ "loss": 0.2504,
288401
+ "step": 148580
288402
+ },
288403
+ {
288404
+ "epoch": 1197.97,
288405
+ "learning_rate": 7.616704361873991e-06,
288406
+ "loss": 0.6615,
288407
+ "step": 148585
288408
+ },
288409
+ {
288410
+ "epoch": 1198.0,
288411
+ "eval_loss": 0.3838162422180176,
288412
+ "eval_runtime": 40.8275,
288413
+ "eval_samples_per_second": 20.574,
288414
+ "eval_steps_per_second": 0.661,
288415
+ "eval_wer": 0.17510748378634408,
288416
+ "step": 148588
288417
+ },
288418
+ {
288419
+ "epoch": 1208.02,
288420
+ "learning_rate": 7.616623586429726e-06,
288421
+ "loss": 0.3853,
288422
+ "step": 148590
288423
+ },
288424
+ {
288425
+ "epoch": 1208.06,
288426
+ "learning_rate": 7.616542810985461e-06,
288427
+ "loss": 0.2698,
288428
+ "step": 148595
288429
+ },
288430
+ {
288431
+ "epoch": 1208.1,
288432
+ "learning_rate": 7.616462035541196e-06,
288433
+ "loss": 0.2669,
288434
+ "step": 148600
288435
+ },
288436
+ {
288437
+ "epoch": 1208.14,
288438
+ "learning_rate": 7.616381260096931e-06,
288439
+ "loss": 0.3389,
288440
+ "step": 148605
288441
+ },
288442
+ {
288443
+ "epoch": 1208.18,
288444
+ "learning_rate": 7.616300484652667e-06,
288445
+ "loss": 0.5707,
288446
+ "step": 148610
288447
+ },
288448
+ {
288449
+ "epoch": 1208.22,
288450
+ "learning_rate": 7.616219709208401e-06,
288451
+ "loss": 0.8938,
288452
+ "step": 148615
288453
+ },
288454
+ {
288455
+ "epoch": 1208.26,
288456
+ "learning_rate": 7.616138933764137e-06,
288457
+ "loss": 0.2657,
288458
+ "step": 148620
288459
+ },
288460
+ {
288461
+ "epoch": 1208.3,
288462
+ "learning_rate": 7.616058158319871e-06,
288463
+ "loss": 0.2442,
288464
+ "step": 148625
288465
+ },
288466
+ {
288467
+ "epoch": 1208.34,
288468
+ "learning_rate": 7.615977382875607e-06,
288469
+ "loss": 0.2851,
288470
+ "step": 148630
288471
+ },
288472
+ {
288473
+ "epoch": 1208.38,
288474
+ "learning_rate": 7.615896607431341e-06,
288475
+ "loss": 0.535,
288476
+ "step": 148635
288477
+ },
288478
+ {
288479
+ "epoch": 1208.42,
288480
+ "learning_rate": 7.615815831987077e-06,
288481
+ "loss": 0.8672,
288482
+ "step": 148640
288483
+ },
288484
+ {
288485
+ "epoch": 1208.46,
288486
+ "learning_rate": 7.615735056542811e-06,
288487
+ "loss": 0.2416,
288488
+ "step": 148645
288489
+ },
288490
+ {
288491
+ "epoch": 1208.5,
288492
+ "learning_rate": 7.615654281098547e-06,
288493
+ "loss": 0.3156,
288494
+ "step": 148650
288495
+ },
288496
+ {
288497
+ "epoch": 1208.54,
288498
+ "learning_rate": 7.615573505654282e-06,
288499
+ "loss": 0.3208,
288500
+ "step": 148655
288501
+ },
288502
+ {
288503
+ "epoch": 1208.58,
288504
+ "learning_rate": 7.615492730210017e-06,
288505
+ "loss": 0.73,
288506
+ "step": 148660
288507
+ },
288508
+ {
288509
+ "epoch": 1208.62,
288510
+ "learning_rate": 7.6154119547657524e-06,
288511
+ "loss": 1.0101,
288512
+ "step": 148665
288513
+ },
288514
+ {
288515
+ "epoch": 1208.66,
288516
+ "learning_rate": 7.6153311793214866e-06,
288517
+ "loss": 0.3275,
288518
+ "step": 148670
288519
+ },
288520
+ {
288521
+ "epoch": 1208.7,
288522
+ "learning_rate": 7.615250403877222e-06,
288523
+ "loss": 0.2837,
288524
+ "step": 148675
288525
+ },
288526
+ {
288527
+ "epoch": 1208.74,
288528
+ "learning_rate": 7.6151696284329565e-06,
288529
+ "loss": 0.3495,
288530
+ "step": 148680
288531
+ },
288532
+ {
288533
+ "epoch": 1208.79,
288534
+ "learning_rate": 7.615088852988692e-06,
288535
+ "loss": 0.5637,
288536
+ "step": 148685
288537
+ },
288538
+ {
288539
+ "epoch": 1208.83,
288540
+ "learning_rate": 7.6150080775444265e-06,
288541
+ "loss": 0.9843,
288542
+ "step": 148690
288543
+ },
288544
+ {
288545
+ "epoch": 1208.87,
288546
+ "learning_rate": 7.614927302100162e-06,
288547
+ "loss": 0.2692,
288548
+ "step": 148695
288549
+ },
288550
+ {
288551
+ "epoch": 1208.91,
288552
+ "learning_rate": 7.6148465266558965e-06,
288553
+ "loss": 0.2533,
288554
+ "step": 148700
288555
+ },
288556
+ {
288557
+ "epoch": 1208.95,
288558
+ "learning_rate": 7.614765751211632e-06,
288559
+ "loss": 0.3418,
288560
+ "step": 148705
288561
+ },
288562
+ {
288563
+ "epoch": 1208.99,
288564
+ "learning_rate": 7.6146849757673665e-06,
288565
+ "loss": 0.7517,
288566
+ "step": 148710
288567
+ },
288568
+ {
288569
+ "epoch": 1209.0,
288570
+ "eval_loss": 0.3921215534210205,
288571
+ "eval_runtime": 42.6819,
288572
+ "eval_samples_per_second": 19.68,
288573
+ "eval_steps_per_second": 0.633,
288574
+ "eval_wer": 0.1728014106237602,
288575
+ "step": 148711
288576
+ },
288577
+ {
288578
+ "epoch": 1199.03,
288579
+ "learning_rate": 7.614604200323102e-06,
288580
+ "loss": 0.2649,
288581
+ "step": 148715
288582
+ },
288583
+ {
288584
+ "epoch": 1199.07,
288585
+ "learning_rate": 7.614523424878838e-06,
288586
+ "loss": 0.3106,
288587
+ "step": 148720
288588
+ },
288589
+ {
288590
+ "epoch": 1199.11,
288591
+ "learning_rate": 7.614442649434572e-06,
288592
+ "loss": 0.2858,
288593
+ "step": 148725
288594
+ },
288595
+ {
288596
+ "epoch": 1199.15,
288597
+ "learning_rate": 7.614361873990308e-06,
288598
+ "loss": 0.3628,
288599
+ "step": 148730
288600
+ },
288601
+ {
288602
+ "epoch": 1199.19,
288603
+ "learning_rate": 7.614281098546042e-06,
288604
+ "loss": 0.8461,
288605
+ "step": 148735
288606
+ },
288607
+ {
288608
+ "epoch": 1199.23,
288609
+ "learning_rate": 7.614200323101778e-06,
288610
+ "loss": 0.5783,
288611
+ "step": 148740
288612
+ },
288613
+ {
288614
+ "epoch": 1199.27,
288615
+ "learning_rate": 7.614119547657512e-06,
288616
+ "loss": 0.5042,
288617
+ "step": 148745
288618
+ },
288619
+ {
288620
+ "epoch": 1199.31,
288621
+ "learning_rate": 7.614038772213248e-06,
288622
+ "loss": 0.2598,
288623
+ "step": 148750
288624
+ },
288625
+ {
288626
+ "epoch": 1199.35,
288627
+ "learning_rate": 7.613957996768982e-06,
288628
+ "loss": 0.3811,
288629
+ "step": 148755
288630
+ },
288631
+ {
288632
+ "epoch": 1199.39,
288633
+ "learning_rate": 7.613877221324718e-06,
288634
+ "loss": 0.9186,
288635
+ "step": 148760
288636
+ },
288637
+ {
288638
+ "epoch": 1199.43,
288639
+ "learning_rate": 7.613796445880452e-06,
288640
+ "loss": 0.6704,
288641
+ "step": 148765
288642
+ },
288643
+ {
288644
+ "epoch": 1199.47,
288645
+ "learning_rate": 7.613715670436188e-06,
288646
+ "loss": 0.2681,
288647
+ "step": 148770
288648
+ },
288649
+ {
288650
+ "epoch": 1199.51,
288651
+ "learning_rate": 7.613634894991922e-06,
288652
+ "loss": 0.2969,
288653
+ "step": 148775
288654
+ },
288655
+ {
288656
+ "epoch": 1199.55,
288657
+ "learning_rate": 7.613554119547658e-06,
288658
+ "loss": 0.3976,
288659
+ "step": 148780
288660
+ },
288661
+ {
288662
+ "epoch": 1199.59,
288663
+ "learning_rate": 7.613473344103394e-06,
288664
+ "loss": 0.969,
288665
+ "step": 148785
288666
+ },
288667
+ {
288668
+ "epoch": 1199.63,
288669
+ "learning_rate": 7.613392568659128e-06,
288670
+ "loss": 0.6605,
288671
+ "step": 148790
288672
+ },
288673
+ {
288674
+ "epoch": 1199.67,
288675
+ "learning_rate": 7.613311793214864e-06,
288676
+ "loss": 0.3244,
288677
+ "step": 148795
288678
+ },
288679
+ {
288680
+ "epoch": 1199.71,
288681
+ "learning_rate": 7.613231017770598e-06,
288682
+ "loss": 0.2454,
288683
+ "step": 148800
288684
+ },
288685
+ {
288686
+ "epoch": 1199.76,
288687
+ "learning_rate": 7.613150242326334e-06,
288688
+ "loss": 0.3354,
288689
+ "step": 148805
288690
+ },
288691
+ {
288692
+ "epoch": 1199.8,
288693
+ "learning_rate": 7.613069466882068e-06,
288694
+ "loss": 0.9782,
288695
+ "step": 148810
288696
+ },
288697
+ {
288698
+ "epoch": 1199.84,
288699
+ "learning_rate": 7.612988691437804e-06,
288700
+ "loss": 0.6457,
288701
+ "step": 148815
288702
+ },
288703
+ {
288704
+ "epoch": 1199.88,
288705
+ "learning_rate": 7.612907915993538e-06,
288706
+ "loss": 0.3441,
288707
+ "step": 148820
288708
+ },
288709
+ {
288710
+ "epoch": 1199.92,
288711
+ "learning_rate": 7.612827140549274e-06,
288712
+ "loss": 0.2688,
288713
+ "step": 148825
288714
+ },
288715
+ {
288716
+ "epoch": 1199.96,
288717
+ "learning_rate": 7.612746365105008e-06,
288718
+ "loss": 0.3786,
288719
+ "step": 148830
288720
+ },
288721
+ {
288722
+ "epoch": 1200.0,
288723
+ "learning_rate": 7.612665589660744e-06,
288724
+ "loss": 0.9727,
288725
+ "step": 148835
288726
+ },
288727
+ {
288728
+ "epoch": 1200.0,
288729
+ "eval_loss": 0.3765946924686432,
288730
+ "eval_runtime": 41.4291,
288731
+ "eval_samples_per_second": 20.276,
288732
+ "eval_steps_per_second": 0.652,
288733
+ "eval_wer": 0.17325978079407708,
288734
+ "step": 148835
288735
+ },
288736
+ {
288737
+ "epoch": 1190.04,
288738
+ "learning_rate": 7.6125848142164795e-06,
288739
+ "loss": 0.3049,
288740
+ "step": 148840
288741
+ },
288742
+ {
288743
+ "epoch": 1190.08,
288744
+ "learning_rate": 7.612504038772214e-06,
288745
+ "loss": 0.244,
288746
+ "step": 148845
288747
+ },
288748
+ {
288749
+ "epoch": 1190.12,
288750
+ "learning_rate": 7.6124232633279495e-06,
288751
+ "loss": 0.2753,
288752
+ "step": 148850
288753
+ },
288754
+ {
288755
+ "epoch": 1190.16,
288756
+ "learning_rate": 7.612342487883684e-06,
288757
+ "loss": 0.3841,
288758
+ "step": 148855
288759
+ },
288760
+ {
288761
+ "epoch": 1190.2,
288762
+ "learning_rate": 7.6122617124394195e-06,
288763
+ "loss": 0.9727,
288764
+ "step": 148860
288765
+ },
288766
+ {
288767
+ "epoch": 1190.24,
288768
+ "learning_rate": 7.612180936995154e-06,
288769
+ "loss": 0.3428,
288770
+ "step": 148865
288771
+ },
288772
+ {
288773
+ "epoch": 1190.28,
288774
+ "learning_rate": 7.6121001615508895e-06,
288775
+ "loss": 0.2587,
288776
+ "step": 148870
288777
+ },
288778
+ {
288779
+ "epoch": 1190.32,
288780
+ "learning_rate": 7.612019386106624e-06,
288781
+ "loss": 0.2912,
288782
+ "step": 148875
288783
+ },
288784
+ {
288785
+ "epoch": 1190.36,
288786
+ "learning_rate": 7.6119386106623594e-06,
288787
+ "loss": 0.4598,
288788
+ "step": 148880
288789
+ },
288790
+ {
288791
+ "epoch": 1190.4,
288792
+ "learning_rate": 7.611857835218094e-06,
288793
+ "loss": 1.111,
288794
+ "step": 148885
288795
+ },
288796
+ {
288797
+ "epoch": 1190.44,
288798
+ "learning_rate": 7.611777059773829e-06,
288799
+ "loss": 0.295,
288800
+ "step": 148890
288801
+ },
288802
+ {
288803
+ "epoch": 1190.48,
288804
+ "learning_rate": 7.611696284329565e-06,
288805
+ "loss": 0.2636,
288806
+ "step": 148895
288807
+ },
288808
+ {
288809
+ "epoch": 1190.52,
288810
+ "learning_rate": 7.611615508885299e-06,
288811
+ "loss": 0.3022,
288812
+ "step": 148900
288813
+ },
288814
+ {
288815
+ "epoch": 1190.56,
288816
+ "learning_rate": 7.611534733441035e-06,
288817
+ "loss": 0.4243,
288818
+ "step": 148905
288819
+ },
288820
+ {
288821
+ "epoch": 1190.6,
288822
+ "learning_rate": 7.611453957996769e-06,
288823
+ "loss": 1.1794,
288824
+ "step": 148910
288825
+ },
288826
+ {
288827
+ "epoch": 1190.64,
288828
+ "learning_rate": 7.611373182552505e-06,
288829
+ "loss": 0.3277,
288830
+ "step": 148915
288831
+ },
288832
+ {
288833
+ "epoch": 1190.68,
288834
+ "learning_rate": 7.611292407108239e-06,
288835
+ "loss": 0.2458,
288836
+ "step": 148920
288837
+ },
288838
+ {
288839
+ "epoch": 1190.72,
288840
+ "learning_rate": 7.611211631663975e-06,
288841
+ "loss": 0.3399,
288842
+ "step": 148925
288843
+ },
288844
+ {
288845
+ "epoch": 1190.76,
288846
+ "learning_rate": 7.611130856219709e-06,
288847
+ "loss": 0.4827,
288848
+ "step": 148930
288849
+ },
288850
+ {
288851
+ "epoch": 1190.8,
288852
+ "learning_rate": 7.611050080775445e-06,
288853
+ "loss": 1.2963,
288854
+ "step": 148935
288855
+ },
288856
+ {
288857
+ "epoch": 1190.84,
288858
+ "learning_rate": 7.610969305331179e-06,
288859
+ "loss": 0.3011,
288860
+ "step": 148940
288861
+ },
288862
+ {
288863
+ "epoch": 1190.88,
288864
+ "learning_rate": 7.610888529886915e-06,
288865
+ "loss": 0.2551,
288866
+ "step": 148945
288867
+ },
288868
+ {
288869
+ "epoch": 1190.92,
288870
+ "learning_rate": 7.610807754442649e-06,
288871
+ "loss": 0.2753,
288872
+ "step": 148950
288873
+ },
288874
+ {
288875
+ "epoch": 1190.96,
288876
+ "learning_rate": 7.610726978998385e-06,
288877
+ "loss": 0.409,
288878
+ "step": 148955
288879
+ },
288880
+ {
288881
+ "epoch": 1191.0,
288882
+ "learning_rate": 7.610646203554121e-06,
288883
+ "loss": 1.1083,
288884
+ "step": 148960
288885
+ },
288886
+ {
288887
+ "epoch": 1191.0,
288888
+ "eval_loss": 0.3409165143966675,
288889
+ "eval_runtime": 41.855,
288890
+ "eval_samples_per_second": 20.093,
288891
+ "eval_steps_per_second": 0.645,
288892
+ "eval_wer": 0.17429476108232586,
288893
+ "step": 148960
288894
  }
288895
  ],
288896
+ "max_steps": 625000,
288897
  "num_train_epochs": 5000,
288898
+ "total_flos": 4.1919118003071346e+20,
288899
  "trial_name": null,
288900
  "trial_params": null
288901
  }
model-bin/finetune/base/{checkpoint-148340 β†’ checkpoint-148960}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630237299.1184735/events.out.tfevents.1630237299.cc93b136ebf5.1086.149 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:055281116f10c036024a1c591328c8686c68d52ebebf1c09411dbb557a2d3b20
3
+ size 4194
model-bin/finetune/base/log/1630237737.6447396/events.out.tfevents.1630237737.cc93b136ebf5.1086.151 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68674c462a5a2d4abd34805035e7a648cfd8fb5ea855bb6a2530d65bbdc535cc
3
+ size 4194
model-bin/finetune/base/log/1630238176.5216558/events.out.tfevents.1630238176.cc93b136ebf5.1086.153 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c02f1a64f951ac3de63f2c2f42f5023db15ab13380ede957d260304a16f2a4a
3
+ size 4194
model-bin/finetune/base/log/1630238608.102783/events.out.tfevents.1630238608.cc93b136ebf5.1086.155 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9991109f76ad167908d19792327fcc959dd4093a639c032198f6aaceca8d89ff
3
+ size 4194
model-bin/finetune/base/log/1630239044.210316/events.out.tfevents.1630239044.cc93b136ebf5.1086.157 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff0d81a3769965a1487bdcfab204f85e0435df7cfbbcb78b2944c162b0056976
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630237299.cc93b136ebf5.1086.148 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2acd2d728b970db419ea96f68ccb0c0fc0538d7bd57261950be955e0c72dab20
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1630237737.cc93b136ebf5.1086.150 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b15ba1e3ce4911b844aadc4448753febcee107b492d9bed71fc8a1bf6f4883c4
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630238176.cc93b136ebf5.1086.152 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4e38775e10091e47e180161166c1a47622a06e6349bacafb3e41b6e6813042e
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630238608.cc93b136ebf5.1086.154 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fde57f985d0dd5de8e7954a18587d2ea779ae74850ddeb7191395a3e81c971c
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630239044.cc93b136ebf5.1086.156 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d4f53e1b364a692f61dc000d469a055b64babd41998b702fadb88e5e19e7b50
3
+ size 8622