Check commited on
Commit
063b9a9
Β·
1 Parent(s): 1ab40fe

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/trainer_state.json +793 -4
  9. model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629758991.8110738/events.out.tfevents.1629758991.74272264b15c.932.243 +3 -0
  11. model-bin/finetune/base/log/1629759638.1886127/events.out.tfevents.1629759638.74272264b15c.932.245 +3 -0
  12. model-bin/finetune/base/log/1629760286.003077/events.out.tfevents.1629760286.74272264b15c.932.247 +3 -0
  13. model-bin/finetune/base/log/1629760955.239782/events.out.tfevents.1629760955.74272264b15c.932.249 +3 -0
  14. model-bin/finetune/base/log/1629761608.7262728/events.out.tfevents.1629761608.74272264b15c.932.251 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629758991.74272264b15c.932.242 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629759638.74272264b15c.932.244 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629760285.74272264b15c.932.246 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629760955.74272264b15c.932.248 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629761608.74272264b15c.932.250 +3 -0
model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:626b3f50af4abb59b8ff7f6804e06113fb806913351046af8543ea49da10c813
3
  size 722165009
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54045d565c66c3e5b8018fc3afcc2c41061b9918603d0883df91a655d33fee22
3
  size 722165009
model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d352b6d04a55702a9681be729eddc009d2ea5243b18a85abd21f667f39d49109
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d25122eb873e0266a53d0c13a8c157bfbcf84c44543a60b0d7a792d97aeb34d
3
  size 377909911
model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dbb325d6f8bea8b4f4941152996b4536b066af419a1d56f45a2e5ec2f1f384ba
3
  size 14439
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:303bcbfcafc202ca744ae4ed8003b741b08ac41373010724ddcd8ab1d77d651a
3
  size 14439
model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a219303497a68a287afddd4ed7c81f20be58bcdfd986786df2e94c5cc582f93f
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:517b8e9913beb30b2ae33d6974fd6fee1e6f87a69ee2028115af6866d0917bf4
3
  size 559
model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cfc84a5ed027e14acf0de020be6f00867fccccd474c7cf821d3fac65edf38749
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41f48a461a4300556f9a200eddd4d70732e267633a4abe25cef3f92554b7b8c0
3
  size 623
model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.18992848189928482,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-52515",
4
- "epoch": 429.0,
5
- "global_step": 53261,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -167130,11 +167130,800 @@
167130
  "eval_steps_per_second": 0.69,
167131
  "eval_wer": 0.19746457867263237,
167132
  "step": 53261
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
167133
  }
167134
  ],
167135
- "max_steps": 620000,
167136
  "num_train_epochs": 5000,
167137
- "total_flos": 1.498716444928101e+20,
167138
  "trial_name": null,
167139
  "trial_params": null
167140
  }
 
1
  {
2
  "best_metric": 0.18992848189928482,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-52515",
4
+ "epoch": 431.0,
5
+ "global_step": 53883,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
167130
  "eval_steps_per_second": 0.69,
167131
  "eval_wer": 0.19746457867263237,
167132
  "step": 53261
167133
+ },
167134
+ {
167135
+ "epoch": 429.03,
167136
+ "learning_rate": 9.162724358974359e-06,
167137
+ "loss": 0.4907,
167138
+ "step": 53265
167139
+ },
167140
+ {
167141
+ "epoch": 429.07,
167142
+ "learning_rate": 9.162644230769232e-06,
167143
+ "loss": 0.413,
167144
+ "step": 53270
167145
+ },
167146
+ {
167147
+ "epoch": 429.11,
167148
+ "learning_rate": 9.162564102564103e-06,
167149
+ "loss": 0.3893,
167150
+ "step": 53275
167151
+ },
167152
+ {
167153
+ "epoch": 429.15,
167154
+ "learning_rate": 9.162483974358975e-06,
167155
+ "loss": 0.4385,
167156
+ "step": 53280
167157
+ },
167158
+ {
167159
+ "epoch": 429.19,
167160
+ "learning_rate": 9.162403846153846e-06,
167161
+ "loss": 1.2454,
167162
+ "step": 53285
167163
+ },
167164
+ {
167165
+ "epoch": 429.23,
167166
+ "learning_rate": 9.162323717948719e-06,
167167
+ "loss": 0.7755,
167168
+ "step": 53290
167169
+ },
167170
+ {
167171
+ "epoch": 429.27,
167172
+ "learning_rate": 9.16224358974359e-06,
167173
+ "loss": 0.3312,
167174
+ "step": 53295
167175
+ },
167176
+ {
167177
+ "epoch": 429.31,
167178
+ "learning_rate": 9.162163461538462e-06,
167179
+ "loss": 0.3771,
167180
+ "step": 53300
167181
+ },
167182
+ {
167183
+ "epoch": 429.35,
167184
+ "learning_rate": 9.162083333333333e-06,
167185
+ "loss": 0.4456,
167186
+ "step": 53305
167187
+ },
167188
+ {
167189
+ "epoch": 429.39,
167190
+ "learning_rate": 9.162003205128206e-06,
167191
+ "loss": 0.9052,
167192
+ "step": 53310
167193
+ },
167194
+ {
167195
+ "epoch": 429.43,
167196
+ "learning_rate": 9.161923076923078e-06,
167197
+ "loss": 0.7603,
167198
+ "step": 53315
167199
+ },
167200
+ {
167201
+ "epoch": 429.47,
167202
+ "learning_rate": 9.161842948717949e-06,
167203
+ "loss": 0.4158,
167204
+ "step": 53320
167205
+ },
167206
+ {
167207
+ "epoch": 429.51,
167208
+ "learning_rate": 9.161762820512822e-06,
167209
+ "loss": 0.3428,
167210
+ "step": 53325
167211
+ },
167212
+ {
167213
+ "epoch": 429.55,
167214
+ "learning_rate": 9.161682692307693e-06,
167215
+ "loss": 0.475,
167216
+ "step": 53330
167217
+ },
167218
+ {
167219
+ "epoch": 429.59,
167220
+ "learning_rate": 9.161602564102565e-06,
167221
+ "loss": 1.054,
167222
+ "step": 53335
167223
+ },
167224
+ {
167225
+ "epoch": 429.63,
167226
+ "learning_rate": 9.161522435897436e-06,
167227
+ "loss": 0.6962,
167228
+ "step": 53340
167229
+ },
167230
+ {
167231
+ "epoch": 429.67,
167232
+ "learning_rate": 9.16144230769231e-06,
167233
+ "loss": 0.3644,
167234
+ "step": 53345
167235
+ },
167236
+ {
167237
+ "epoch": 429.71,
167238
+ "learning_rate": 9.16136217948718e-06,
167239
+ "loss": 0.4176,
167240
+ "step": 53350
167241
+ },
167242
+ {
167243
+ "epoch": 429.76,
167244
+ "learning_rate": 9.161282051282052e-06,
167245
+ "loss": 0.48,
167246
+ "step": 53355
167247
+ },
167248
+ {
167249
+ "epoch": 429.8,
167250
+ "learning_rate": 9.161201923076923e-06,
167251
+ "loss": 0.9505,
167252
+ "step": 53360
167253
+ },
167254
+ {
167255
+ "epoch": 429.84,
167256
+ "learning_rate": 9.161121794871796e-06,
167257
+ "loss": 0.81,
167258
+ "step": 53365
167259
+ },
167260
+ {
167261
+ "epoch": 429.88,
167262
+ "learning_rate": 9.161041666666668e-06,
167263
+ "loss": 0.3829,
167264
+ "step": 53370
167265
+ },
167266
+ {
167267
+ "epoch": 429.92,
167268
+ "learning_rate": 9.160961538461539e-06,
167269
+ "loss": 0.3667,
167270
+ "step": 53375
167271
+ },
167272
+ {
167273
+ "epoch": 429.96,
167274
+ "learning_rate": 9.160881410256412e-06,
167275
+ "loss": 0.4709,
167276
+ "step": 53380
167277
+ },
167278
+ {
167279
+ "epoch": 430.0,
167280
+ "learning_rate": 9.160801282051283e-06,
167281
+ "loss": 1.0048,
167282
+ "step": 53385
167283
+ },
167284
+ {
167285
+ "epoch": 430.0,
167286
+ "eval_loss": 0.4243323802947998,
167287
+ "eval_runtime": 39.8421,
167288
+ "eval_samples_per_second": 21.083,
167289
+ "eval_steps_per_second": 0.678,
167290
+ "eval_wer": 0.20694374414751854,
167291
+ "step": 53385
167292
+ },
167293
+ {
167294
+ "epoch": 430.04,
167295
+ "learning_rate": 9.160721153846155e-06,
167296
+ "loss": 0.3929,
167297
+ "step": 53390
167298
+ },
167299
+ {
167300
+ "epoch": 430.08,
167301
+ "learning_rate": 9.160641025641026e-06,
167302
+ "loss": 0.2955,
167303
+ "step": 53395
167304
+ },
167305
+ {
167306
+ "epoch": 430.12,
167307
+ "learning_rate": 9.1605608974359e-06,
167308
+ "loss": 0.3902,
167309
+ "step": 53400
167310
+ },
167311
+ {
167312
+ "epoch": 430.16,
167313
+ "learning_rate": 9.160480769230769e-06,
167314
+ "loss": 0.4729,
167315
+ "step": 53405
167316
+ },
167317
+ {
167318
+ "epoch": 430.2,
167319
+ "learning_rate": 9.160400641025642e-06,
167320
+ "loss": 1.3877,
167321
+ "step": 53410
167322
+ },
167323
+ {
167324
+ "epoch": 430.24,
167325
+ "learning_rate": 9.160320512820513e-06,
167326
+ "loss": 0.4524,
167327
+ "step": 53415
167328
+ },
167329
+ {
167330
+ "epoch": 430.28,
167331
+ "learning_rate": 9.160240384615385e-06,
167332
+ "loss": 0.3233,
167333
+ "step": 53420
167334
+ },
167335
+ {
167336
+ "epoch": 430.32,
167337
+ "learning_rate": 9.160160256410258e-06,
167338
+ "loss": 0.4162,
167339
+ "step": 53425
167340
+ },
167341
+ {
167342
+ "epoch": 430.36,
167343
+ "learning_rate": 9.160080128205129e-06,
167344
+ "loss": 0.4888,
167345
+ "step": 53430
167346
+ },
167347
+ {
167348
+ "epoch": 430.4,
167349
+ "learning_rate": 9.16e-06,
167350
+ "loss": 1.219,
167351
+ "step": 53435
167352
+ },
167353
+ {
167354
+ "epoch": 430.44,
167355
+ "learning_rate": 9.159919871794872e-06,
167356
+ "loss": 0.3253,
167357
+ "step": 53440
167358
+ },
167359
+ {
167360
+ "epoch": 430.48,
167361
+ "learning_rate": 9.159839743589745e-06,
167362
+ "loss": 0.3146,
167363
+ "step": 53445
167364
+ },
167365
+ {
167366
+ "epoch": 430.52,
167367
+ "learning_rate": 9.159759615384616e-06,
167368
+ "loss": 0.3463,
167369
+ "step": 53450
167370
+ },
167371
+ {
167372
+ "epoch": 430.56,
167373
+ "learning_rate": 9.159679487179488e-06,
167374
+ "loss": 0.4703,
167375
+ "step": 53455
167376
+ },
167377
+ {
167378
+ "epoch": 430.6,
167379
+ "learning_rate": 9.159599358974359e-06,
167380
+ "loss": 1.2574,
167381
+ "step": 53460
167382
+ },
167383
+ {
167384
+ "epoch": 430.64,
167385
+ "learning_rate": 9.159519230769232e-06,
167386
+ "loss": 0.3584,
167387
+ "step": 53465
167388
+ },
167389
+ {
167390
+ "epoch": 430.68,
167391
+ "learning_rate": 9.159439102564103e-06,
167392
+ "loss": 0.3258,
167393
+ "step": 53470
167394
+ },
167395
+ {
167396
+ "epoch": 430.72,
167397
+ "learning_rate": 9.159358974358975e-06,
167398
+ "loss": 0.3546,
167399
+ "step": 53475
167400
+ },
167401
+ {
167402
+ "epoch": 430.76,
167403
+ "learning_rate": 9.159278846153848e-06,
167404
+ "loss": 0.4754,
167405
+ "step": 53480
167406
+ },
167407
+ {
167408
+ "epoch": 430.8,
167409
+ "learning_rate": 9.159198717948719e-06,
167410
+ "loss": 1.2657,
167411
+ "step": 53485
167412
+ },
167413
+ {
167414
+ "epoch": 430.84,
167415
+ "learning_rate": 9.15911858974359e-06,
167416
+ "loss": 0.3547,
167417
+ "step": 53490
167418
+ },
167419
+ {
167420
+ "epoch": 430.88,
167421
+ "learning_rate": 9.159038461538462e-06,
167422
+ "loss": 0.3136,
167423
+ "step": 53495
167424
+ },
167425
+ {
167426
+ "epoch": 430.92,
167427
+ "learning_rate": 9.158958333333335e-06,
167428
+ "loss": 0.397,
167429
+ "step": 53500
167430
+ },
167431
+ {
167432
+ "epoch": 430.96,
167433
+ "learning_rate": 9.158878205128206e-06,
167434
+ "loss": 0.5967,
167435
+ "step": 53505
167436
+ },
167437
+ {
167438
+ "epoch": 431.0,
167439
+ "eval_loss": 0.3771889805793762,
167440
+ "eval_runtime": 39.5555,
167441
+ "eval_samples_per_second": 21.236,
167442
+ "eval_steps_per_second": 0.683,
167443
+ "eval_wer": 0.19051832190518322,
167444
+ "step": 53509
167445
+ },
167446
+ {
167447
+ "epoch": 428.01,
167448
+ "learning_rate": 9.158798076923078e-06,
167449
+ "loss": 0.3878,
167450
+ "step": 53510
167451
+ },
167452
+ {
167453
+ "epoch": 428.05,
167454
+ "learning_rate": 9.158717948717949e-06,
167455
+ "loss": 0.3302,
167456
+ "step": 53515
167457
+ },
167458
+ {
167459
+ "epoch": 428.09,
167460
+ "learning_rate": 9.158637820512822e-06,
167461
+ "loss": 0.3114,
167462
+ "step": 53520
167463
+ },
167464
+ {
167465
+ "epoch": 428.13,
167466
+ "learning_rate": 9.158557692307693e-06,
167467
+ "loss": 0.3538,
167468
+ "step": 53525
167469
+ },
167470
+ {
167471
+ "epoch": 428.17,
167472
+ "learning_rate": 9.158477564102565e-06,
167473
+ "loss": 0.6061,
167474
+ "step": 53530
167475
+ },
167476
+ {
167477
+ "epoch": 428.21,
167478
+ "learning_rate": 9.158397435897438e-06,
167479
+ "loss": 1.3689,
167480
+ "step": 53535
167481
+ },
167482
+ {
167483
+ "epoch": 428.25,
167484
+ "learning_rate": 9.158317307692307e-06,
167485
+ "loss": 0.5001,
167486
+ "step": 53540
167487
+ },
167488
+ {
167489
+ "epoch": 428.29,
167490
+ "learning_rate": 9.15823717948718e-06,
167491
+ "loss": 0.3581,
167492
+ "step": 53545
167493
+ },
167494
+ {
167495
+ "epoch": 428.33,
167496
+ "learning_rate": 9.158157051282052e-06,
167497
+ "loss": 0.335,
167498
+ "step": 53550
167499
+ },
167500
+ {
167501
+ "epoch": 428.37,
167502
+ "learning_rate": 9.158076923076923e-06,
167503
+ "loss": 0.6611,
167504
+ "step": 53555
167505
+ },
167506
+ {
167507
+ "epoch": 428.41,
167508
+ "learning_rate": 9.157996794871795e-06,
167509
+ "loss": 1.2383,
167510
+ "step": 53560
167511
+ },
167512
+ {
167513
+ "epoch": 428.45,
167514
+ "learning_rate": 9.157916666666668e-06,
167515
+ "loss": 0.3176,
167516
+ "step": 53565
167517
+ },
167518
+ {
167519
+ "epoch": 428.49,
167520
+ "learning_rate": 9.157836538461539e-06,
167521
+ "loss": 0.2605,
167522
+ "step": 53570
167523
+ },
167524
+ {
167525
+ "epoch": 428.53,
167526
+ "learning_rate": 9.15775641025641e-06,
167527
+ "loss": 0.3196,
167528
+ "step": 53575
167529
+ },
167530
+ {
167531
+ "epoch": 428.57,
167532
+ "learning_rate": 9.157676282051283e-06,
167533
+ "loss": 0.6539,
167534
+ "step": 53580
167535
+ },
167536
+ {
167537
+ "epoch": 428.61,
167538
+ "learning_rate": 9.157596153846155e-06,
167539
+ "loss": 1.1682,
167540
+ "step": 53585
167541
+ },
167542
+ {
167543
+ "epoch": 428.65,
167544
+ "learning_rate": 9.157516025641026e-06,
167545
+ "loss": 0.338,
167546
+ "step": 53590
167547
+ },
167548
+ {
167549
+ "epoch": 428.69,
167550
+ "learning_rate": 9.157435897435897e-06,
167551
+ "loss": 0.3081,
167552
+ "step": 53595
167553
+ },
167554
+ {
167555
+ "epoch": 428.73,
167556
+ "learning_rate": 9.15735576923077e-06,
167557
+ "loss": 0.3158,
167558
+ "step": 53600
167559
+ },
167560
+ {
167561
+ "epoch": 428.77,
167562
+ "learning_rate": 9.157275641025642e-06,
167563
+ "loss": 0.5656,
167564
+ "step": 53605
167565
+ },
167566
+ {
167567
+ "epoch": 428.81,
167568
+ "learning_rate": 9.157195512820513e-06,
167569
+ "loss": 1.0353,
167570
+ "step": 53610
167571
+ },
167572
+ {
167573
+ "epoch": 428.85,
167574
+ "learning_rate": 9.157115384615385e-06,
167575
+ "loss": 0.3622,
167576
+ "step": 53615
167577
+ },
167578
+ {
167579
+ "epoch": 428.89,
167580
+ "learning_rate": 9.157035256410258e-06,
167581
+ "loss": 0.3066,
167582
+ "step": 53620
167583
+ },
167584
+ {
167585
+ "epoch": 428.93,
167586
+ "learning_rate": 9.156955128205129e-06,
167587
+ "loss": 0.4568,
167588
+ "step": 53625
167589
+ },
167590
+ {
167591
+ "epoch": 428.97,
167592
+ "learning_rate": 9.156875e-06,
167593
+ "loss": 0.5429,
167594
+ "step": 53630
167595
+ },
167596
+ {
167597
+ "epoch": 429.0,
167598
+ "eval_loss": 0.4089108407497406,
167599
+ "eval_runtime": 39.7496,
167600
+ "eval_samples_per_second": 21.132,
167601
+ "eval_steps_per_second": 0.679,
167602
+ "eval_wer": 0.2008918780612618,
167603
+ "step": 53634
167604
+ },
167605
+ {
167606
+ "epoch": 432.01,
167607
+ "learning_rate": 9.156794871794873e-06,
167608
+ "loss": 0.462,
167609
+ "step": 53635
167610
+ },
167611
+ {
167612
+ "epoch": 432.05,
167613
+ "learning_rate": 9.156714743589745e-06,
167614
+ "loss": 0.3044,
167615
+ "step": 53640
167616
+ },
167617
+ {
167618
+ "epoch": 432.09,
167619
+ "learning_rate": 9.156634615384616e-06,
167620
+ "loss": 0.3536,
167621
+ "step": 53645
167622
+ },
167623
+ {
167624
+ "epoch": 432.13,
167625
+ "learning_rate": 9.156554487179487e-06,
167626
+ "loss": 0.3502,
167627
+ "step": 53650
167628
+ },
167629
+ {
167630
+ "epoch": 432.17,
167631
+ "learning_rate": 9.15647435897436e-06,
167632
+ "loss": 0.5653,
167633
+ "step": 53655
167634
+ },
167635
+ {
167636
+ "epoch": 432.21,
167637
+ "learning_rate": 9.156394230769232e-06,
167638
+ "loss": 1.2789,
167639
+ "step": 53660
167640
+ },
167641
+ {
167642
+ "epoch": 432.25,
167643
+ "learning_rate": 9.156314102564103e-06,
167644
+ "loss": 0.3512,
167645
+ "step": 53665
167646
+ },
167647
+ {
167648
+ "epoch": 432.29,
167649
+ "learning_rate": 9.156233974358976e-06,
167650
+ "loss": 0.3391,
167651
+ "step": 53670
167652
+ },
167653
+ {
167654
+ "epoch": 432.33,
167655
+ "learning_rate": 9.156153846153848e-06,
167656
+ "loss": 0.3438,
167657
+ "step": 53675
167658
+ },
167659
+ {
167660
+ "epoch": 432.37,
167661
+ "learning_rate": 9.156073717948719e-06,
167662
+ "loss": 0.5578,
167663
+ "step": 53680
167664
+ },
167665
+ {
167666
+ "epoch": 432.41,
167667
+ "learning_rate": 9.15599358974359e-06,
167668
+ "loss": 1.1817,
167669
+ "step": 53685
167670
+ },
167671
+ {
167672
+ "epoch": 432.45,
167673
+ "learning_rate": 9.155913461538463e-06,
167674
+ "loss": 0.3402,
167675
+ "step": 53690
167676
+ },
167677
+ {
167678
+ "epoch": 432.49,
167679
+ "learning_rate": 9.155833333333333e-06,
167680
+ "loss": 0.2861,
167681
+ "step": 53695
167682
+ },
167683
+ {
167684
+ "epoch": 432.53,
167685
+ "learning_rate": 9.155753205128206e-06,
167686
+ "loss": 0.2979,
167687
+ "step": 53700
167688
+ },
167689
+ {
167690
+ "epoch": 432.57,
167691
+ "learning_rate": 9.155673076923077e-06,
167692
+ "loss": 0.5803,
167693
+ "step": 53705
167694
+ },
167695
+ {
167696
+ "epoch": 432.61,
167697
+ "learning_rate": 9.155592948717949e-06,
167698
+ "loss": 1.1293,
167699
+ "step": 53710
167700
+ },
167701
+ {
167702
+ "epoch": 432.65,
167703
+ "learning_rate": 9.15551282051282e-06,
167704
+ "loss": 0.4187,
167705
+ "step": 53715
167706
+ },
167707
+ {
167708
+ "epoch": 432.69,
167709
+ "learning_rate": 9.155432692307693e-06,
167710
+ "loss": 0.3486,
167711
+ "step": 53720
167712
+ },
167713
+ {
167714
+ "epoch": 432.73,
167715
+ "learning_rate": 9.155352564102565e-06,
167716
+ "loss": 0.4529,
167717
+ "step": 53725
167718
+ },
167719
+ {
167720
+ "epoch": 432.77,
167721
+ "learning_rate": 9.155272435897436e-06,
167722
+ "loss": 0.6183,
167723
+ "step": 53730
167724
+ },
167725
+ {
167726
+ "epoch": 432.81,
167727
+ "learning_rate": 9.155192307692309e-06,
167728
+ "loss": 1.154,
167729
+ "step": 53735
167730
+ },
167731
+ {
167732
+ "epoch": 432.85,
167733
+ "learning_rate": 9.15511217948718e-06,
167734
+ "loss": 0.3461,
167735
+ "step": 53740
167736
+ },
167737
+ {
167738
+ "epoch": 432.9,
167739
+ "learning_rate": 9.155032051282052e-06,
167740
+ "loss": 0.3242,
167741
+ "step": 53745
167742
+ },
167743
+ {
167744
+ "epoch": 432.94,
167745
+ "learning_rate": 9.154951923076923e-06,
167746
+ "loss": 0.4448,
167747
+ "step": 53750
167748
+ },
167749
+ {
167750
+ "epoch": 432.98,
167751
+ "learning_rate": 9.154871794871796e-06,
167752
+ "loss": 0.6214,
167753
+ "step": 53755
167754
+ },
167755
+ {
167756
+ "epoch": 433.0,
167757
+ "eval_loss": 0.45085570216178894,
167758
+ "eval_runtime": 39.3642,
167759
+ "eval_samples_per_second": 21.339,
167760
+ "eval_steps_per_second": 0.686,
167761
+ "eval_wer": 0.2042432392051397,
167762
+ "step": 53758
167763
+ },
167764
+ {
167765
+ "epoch": 430.02,
167766
+ "learning_rate": 9.154791666666667e-06,
167767
+ "loss": 0.3722,
167768
+ "step": 53760
167769
+ },
167770
+ {
167771
+ "epoch": 430.06,
167772
+ "learning_rate": 9.154711538461539e-06,
167773
+ "loss": 0.2811,
167774
+ "step": 53765
167775
+ },
167776
+ {
167777
+ "epoch": 430.1,
167778
+ "learning_rate": 9.154631410256412e-06,
167779
+ "loss": 0.3527,
167780
+ "step": 53770
167781
+ },
167782
+ {
167783
+ "epoch": 430.14,
167784
+ "learning_rate": 9.154551282051283e-06,
167785
+ "loss": 0.4463,
167786
+ "step": 53775
167787
+ },
167788
+ {
167789
+ "epoch": 430.18,
167790
+ "learning_rate": 9.154471153846155e-06,
167791
+ "loss": 0.7272,
167792
+ "step": 53780
167793
+ },
167794
+ {
167795
+ "epoch": 430.22,
167796
+ "learning_rate": 9.154391025641026e-06,
167797
+ "loss": 1.0655,
167798
+ "step": 53785
167799
+ },
167800
+ {
167801
+ "epoch": 430.26,
167802
+ "learning_rate": 9.154310897435899e-06,
167803
+ "loss": 0.3984,
167804
+ "step": 53790
167805
+ },
167806
+ {
167807
+ "epoch": 430.3,
167808
+ "learning_rate": 9.15423076923077e-06,
167809
+ "loss": 0.3087,
167810
+ "step": 53795
167811
+ },
167812
+ {
167813
+ "epoch": 430.34,
167814
+ "learning_rate": 9.154150641025642e-06,
167815
+ "loss": 0.3548,
167816
+ "step": 53800
167817
+ },
167818
+ {
167819
+ "epoch": 430.38,
167820
+ "learning_rate": 9.154070512820513e-06,
167821
+ "loss": 0.6702,
167822
+ "step": 53805
167823
+ },
167824
+ {
167825
+ "epoch": 430.42,
167826
+ "learning_rate": 9.153990384615386e-06,
167827
+ "loss": 1.0987,
167828
+ "step": 53810
167829
+ },
167830
+ {
167831
+ "epoch": 430.46,
167832
+ "learning_rate": 9.153910256410256e-06,
167833
+ "loss": 0.2952,
167834
+ "step": 53815
167835
+ },
167836
+ {
167837
+ "epoch": 430.5,
167838
+ "learning_rate": 9.153830128205129e-06,
167839
+ "loss": 0.3559,
167840
+ "step": 53820
167841
+ },
167842
+ {
167843
+ "epoch": 430.54,
167844
+ "learning_rate": 9.153750000000002e-06,
167845
+ "loss": 0.4175,
167846
+ "step": 53825
167847
+ },
167848
+ {
167849
+ "epoch": 430.58,
167850
+ "learning_rate": 9.153669871794872e-06,
167851
+ "loss": 0.6887,
167852
+ "step": 53830
167853
+ },
167854
+ {
167855
+ "epoch": 430.62,
167856
+ "learning_rate": 9.153589743589745e-06,
167857
+ "loss": 1.087,
167858
+ "step": 53835
167859
+ },
167860
+ {
167861
+ "epoch": 430.66,
167862
+ "learning_rate": 9.153509615384616e-06,
167863
+ "loss": 0.3551,
167864
+ "step": 53840
167865
+ },
167866
+ {
167867
+ "epoch": 430.7,
167868
+ "learning_rate": 9.153429487179487e-06,
167869
+ "loss": 0.4347,
167870
+ "step": 53845
167871
+ },
167872
+ {
167873
+ "epoch": 430.74,
167874
+ "learning_rate": 9.153349358974359e-06,
167875
+ "loss": 0.3289,
167876
+ "step": 53850
167877
+ },
167878
+ {
167879
+ "epoch": 430.78,
167880
+ "learning_rate": 9.153269230769232e-06,
167881
+ "loss": 0.7435,
167882
+ "step": 53855
167883
+ },
167884
+ {
167885
+ "epoch": 430.82,
167886
+ "learning_rate": 9.153189102564103e-06,
167887
+ "loss": 1.0853,
167888
+ "step": 53860
167889
+ },
167890
+ {
167891
+ "epoch": 430.86,
167892
+ "learning_rate": 9.153108974358974e-06,
167893
+ "loss": 0.3564,
167894
+ "step": 53865
167895
+ },
167896
+ {
167897
+ "epoch": 430.9,
167898
+ "learning_rate": 9.153028846153846e-06,
167899
+ "loss": 0.3129,
167900
+ "step": 53870
167901
+ },
167902
+ {
167903
+ "epoch": 430.94,
167904
+ "learning_rate": 9.152948717948719e-06,
167905
+ "loss": 0.392,
167906
+ "step": 53875
167907
+ },
167908
+ {
167909
+ "epoch": 430.98,
167910
+ "learning_rate": 9.15286858974359e-06,
167911
+ "loss": 0.6987,
167912
+ "step": 53880
167913
+ },
167914
+ {
167915
+ "epoch": 431.0,
167916
+ "eval_loss": 0.4261249005794525,
167917
+ "eval_runtime": 40.6123,
167918
+ "eval_samples_per_second": 20.659,
167919
+ "eval_steps_per_second": 0.665,
167920
+ "eval_wer": 0.20484629294755877,
167921
+ "step": 53883
167922
  }
167923
  ],
167924
+ "max_steps": 625000,
167925
  "num_train_epochs": 5000,
167926
+ "total_flos": 1.5162302896695742e+20,
167927
  "trial_name": null,
167928
  "trial_params": null
167929
  }
model-bin/finetune/base/{checkpoint-53261 β†’ checkpoint-53883}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629758991.8110738/events.out.tfevents.1629758991.74272264b15c.932.243 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f0fc3ca94b527b56044fb11636857cb312846428b7d1c25b83e74e109074f54
3
+ size 4194
model-bin/finetune/base/log/1629759638.1886127/events.out.tfevents.1629759638.74272264b15c.932.245 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f6359b4ca72e068cd143972392b8861fb25bb26fff76fad9f39e9ac10e40a02
3
+ size 4194
model-bin/finetune/base/log/1629760286.003077/events.out.tfevents.1629760286.74272264b15c.932.247 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef32d9aba34e34b5e0f7cc8a3c906b2dc71fcd8341702a6b588aeccab8106973
3
+ size 4194
model-bin/finetune/base/log/1629760955.239782/events.out.tfevents.1629760955.74272264b15c.932.249 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f1da4b68b9d4aefb7159d273cd7ec0e3b69f54b12f5804637195205ee331f63
3
+ size 4194
model-bin/finetune/base/log/1629761608.7262728/events.out.tfevents.1629761608.74272264b15c.932.251 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ea57a4f3dc82d4c084a699c376f70748c4c1bbe875f50b1cbaa6c55cf6647e2
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629758991.74272264b15c.932.242 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:178ab82b89560feef777c069a0f7b3726e6f12a3b5599864d43de710700cc8da
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629759638.74272264b15c.932.244 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8a51922dc61333f94c718cad295656d8b7c3f913e9cf0b82b99642cb70c2dcc
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1629760285.74272264b15c.932.246 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18426d2d0c2f1b79173d109598f1c9a8059024fc5e4fadf82d5733417fafa301
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629760955.74272264b15c.932.248 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:156683756fdd361565d13aba4dd180679aa27b81624d9bd9a5645718f956c3ec
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629761608.74272264b15c.932.250 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee1f37541d338cead8ce83ac5eeed746f1b61263f11f9e905fdfcbbf1a96750f
3
+ size 8622