Check commited on
Commit
e0502e4
Β·
1 Parent(s): 429c63e

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/trainer_state.json +799 -4
  9. model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629726271.057744/events.out.tfevents.1629726271.74272264b15c.932.143 +3 -0
  11. model-bin/finetune/base/log/1629726936.3025346/events.out.tfevents.1629726936.74272264b15c.932.145 +3 -0
  12. model-bin/finetune/base/log/1629727569.1115878/events.out.tfevents.1629727569.74272264b15c.932.147 +3 -0
  13. model-bin/finetune/base/log/1629728226.273403/events.out.tfevents.1629728226.74272264b15c.932.149 +3 -0
  14. model-bin/finetune/base/log/1629728868.313585/events.out.tfevents.1629728868.74272264b15c.932.151 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629726271.74272264b15c.932.142 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629726936.74272264b15c.932.144 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629727569.74272264b15c.932.146 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629728226.74272264b15c.932.148 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629728868.74272264b15c.932.150 +3 -0
model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f416f5d36f988a6620623de1a64b755b296384c4bac29fc47591a2e219967a4
3
  size 722165009
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6f1253e46cde7d178f39e6b17a8972a3aad4111872e5bf27d7b6dca137c72c2
3
  size 722165009
model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:280e523434905e5e2e0729e163d74f5165b7a5d3e455d469d7ad26838ee0f41b
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e40873fa8da493e92e672079326593cb7f57a26d199c8587c88979165d3b405d
3
  size 377909911
model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8c6812e674dae21fda5a108e32f336d8e8647314d881174e7505b176858b6aaf
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8655ea7d9fd7f1ddbb7f3c0285d61ea693bed9b3fcbbe176e4c78f2515296b7f
3
  size 14503
model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c7beac282102164b8797bb861a3ef7dd67f499924d975234a76a921e8e6e17be
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f26c7b4da20f1cc9a4ffd34476828968f627bc9309919fb52d579c167c07ee6a
3
  size 559
model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5bb395051453173d13b4e26bb7c0389dd00c26c9126ed17d59ddbd6f76a6d50a
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31131468d789f262e6e19dc1f7e277df2e3b8527cff32367b76129cd61996678
3
  size 623
model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.19748327029386092,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-46666",
4
- "epoch": 378.99598393574297,
5
- "global_step": 47039,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -159210,11 +159210,806 @@
159210
  "eval_steps_per_second": 0.658,
159211
  "eval_wer": 0.20545609548167093,
159212
  "step": 47039
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
159213
  }
159214
  ],
159215
- "max_steps": 620000,
159216
  "num_train_epochs": 5000,
159217
- "total_flos": 1.3235134764537184e+20,
159218
  "trial_name": null,
159219
  "trial_params": null
159220
  }
 
1
  {
2
  "best_metric": 0.19748327029386092,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-46666",
4
+ "epoch": 381.0,
5
+ "global_step": 47663,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
159210
  "eval_steps_per_second": 0.658,
159211
  "eval_wer": 0.20545609548167093,
159212
  "step": 47039
159213
+ },
159214
+ {
159215
+ "epoch": 376.01,
159216
+ "learning_rate": 9.262451923076923e-06,
159217
+ "loss": 0.4411,
159218
+ "step": 47040
159219
+ },
159220
+ {
159221
+ "epoch": 376.05,
159222
+ "learning_rate": 9.262371794871796e-06,
159223
+ "loss": 0.3758,
159224
+ "step": 47045
159225
+ },
159226
+ {
159227
+ "epoch": 376.09,
159228
+ "learning_rate": 9.262291666666668e-06,
159229
+ "loss": 0.3242,
159230
+ "step": 47050
159231
+ },
159232
+ {
159233
+ "epoch": 376.13,
159234
+ "learning_rate": 9.262211538461539e-06,
159235
+ "loss": 0.3592,
159236
+ "step": 47055
159237
+ },
159238
+ {
159239
+ "epoch": 376.17,
159240
+ "learning_rate": 9.26213141025641e-06,
159241
+ "loss": 0.5519,
159242
+ "step": 47060
159243
+ },
159244
+ {
159245
+ "epoch": 376.21,
159246
+ "learning_rate": 9.262051282051284e-06,
159247
+ "loss": 1.1892,
159248
+ "step": 47065
159249
+ },
159250
+ {
159251
+ "epoch": 376.25,
159252
+ "learning_rate": 9.261971153846155e-06,
159253
+ "loss": 0.3381,
159254
+ "step": 47070
159255
+ },
159256
+ {
159257
+ "epoch": 376.29,
159258
+ "learning_rate": 9.261891025641026e-06,
159259
+ "loss": 0.3679,
159260
+ "step": 47075
159261
+ },
159262
+ {
159263
+ "epoch": 376.33,
159264
+ "learning_rate": 9.261810897435898e-06,
159265
+ "loss": 0.3559,
159266
+ "step": 47080
159267
+ },
159268
+ {
159269
+ "epoch": 376.37,
159270
+ "learning_rate": 9.26173076923077e-06,
159271
+ "loss": 0.768,
159272
+ "step": 47085
159273
+ },
159274
+ {
159275
+ "epoch": 376.41,
159276
+ "learning_rate": 9.261650641025642e-06,
159277
+ "loss": 1.2828,
159278
+ "step": 47090
159279
+ },
159280
+ {
159281
+ "epoch": 376.45,
159282
+ "learning_rate": 9.261570512820513e-06,
159283
+ "loss": 0.3981,
159284
+ "step": 47095
159285
+ },
159286
+ {
159287
+ "epoch": 376.49,
159288
+ "learning_rate": 9.261490384615386e-06,
159289
+ "loss": 0.3617,
159290
+ "step": 47100
159291
+ },
159292
+ {
159293
+ "epoch": 376.53,
159294
+ "learning_rate": 9.261410256410258e-06,
159295
+ "loss": 0.3778,
159296
+ "step": 47105
159297
+ },
159298
+ {
159299
+ "epoch": 376.57,
159300
+ "learning_rate": 9.261330128205129e-06,
159301
+ "loss": 0.5657,
159302
+ "step": 47110
159303
+ },
159304
+ {
159305
+ "epoch": 376.61,
159306
+ "learning_rate": 9.26125e-06,
159307
+ "loss": 1.2655,
159308
+ "step": 47115
159309
+ },
159310
+ {
159311
+ "epoch": 376.65,
159312
+ "learning_rate": 9.261169871794874e-06,
159313
+ "loss": 0.3939,
159314
+ "step": 47120
159315
+ },
159316
+ {
159317
+ "epoch": 376.69,
159318
+ "learning_rate": 9.261089743589743e-06,
159319
+ "loss": 0.3814,
159320
+ "step": 47125
159321
+ },
159322
+ {
159323
+ "epoch": 376.73,
159324
+ "learning_rate": 9.261009615384616e-06,
159325
+ "loss": 0.4081,
159326
+ "step": 47130
159327
+ },
159328
+ {
159329
+ "epoch": 376.77,
159330
+ "learning_rate": 9.260929487179488e-06,
159331
+ "loss": 0.6094,
159332
+ "step": 47135
159333
+ },
159334
+ {
159335
+ "epoch": 376.81,
159336
+ "learning_rate": 9.260849358974359e-06,
159337
+ "loss": 1.2367,
159338
+ "step": 47140
159339
+ },
159340
+ {
159341
+ "epoch": 376.85,
159342
+ "learning_rate": 9.260769230769232e-06,
159343
+ "loss": 0.3702,
159344
+ "step": 47145
159345
+ },
159346
+ {
159347
+ "epoch": 376.89,
159348
+ "learning_rate": 9.260689102564103e-06,
159349
+ "loss": 0.3565,
159350
+ "step": 47150
159351
+ },
159352
+ {
159353
+ "epoch": 376.93,
159354
+ "learning_rate": 9.260608974358975e-06,
159355
+ "loss": 0.3359,
159356
+ "step": 47155
159357
+ },
159358
+ {
159359
+ "epoch": 376.97,
159360
+ "learning_rate": 9.260528846153846e-06,
159361
+ "loss": 0.6193,
159362
+ "step": 47160
159363
+ },
159364
+ {
159365
+ "epoch": 377.0,
159366
+ "eval_loss": 0.482885479927063,
159367
+ "eval_runtime": 40.1976,
159368
+ "eval_samples_per_second": 20.797,
159369
+ "eval_steps_per_second": 0.672,
159370
+ "eval_wer": 0.20530342419965833,
159371
+ "step": 47164
159372
+ },
159373
+ {
159374
+ "epoch": 377.01,
159375
+ "learning_rate": 9.260448717948719e-06,
159376
+ "loss": 0.4196,
159377
+ "step": 47165
159378
+ },
159379
+ {
159380
+ "epoch": 377.05,
159381
+ "learning_rate": 9.26036858974359e-06,
159382
+ "loss": 0.3283,
159383
+ "step": 47170
159384
+ },
159385
+ {
159386
+ "epoch": 377.09,
159387
+ "learning_rate": 9.260288461538462e-06,
159388
+ "loss": 0.3014,
159389
+ "step": 47175
159390
+ },
159391
+ {
159392
+ "epoch": 377.13,
159393
+ "learning_rate": 9.260208333333333e-06,
159394
+ "loss": 0.4168,
159395
+ "step": 47180
159396
+ },
159397
+ {
159398
+ "epoch": 377.17,
159399
+ "learning_rate": 9.260128205128206e-06,
159400
+ "loss": 0.5062,
159401
+ "step": 47185
159402
+ },
159403
+ {
159404
+ "epoch": 377.21,
159405
+ "learning_rate": 9.260048076923078e-06,
159406
+ "loss": 1.4582,
159407
+ "step": 47190
159408
+ },
159409
+ {
159410
+ "epoch": 377.25,
159411
+ "learning_rate": 9.259967948717949e-06,
159412
+ "loss": 0.3753,
159413
+ "step": 47195
159414
+ },
159415
+ {
159416
+ "epoch": 377.29,
159417
+ "learning_rate": 9.259887820512822e-06,
159418
+ "loss": 0.3808,
159419
+ "step": 47200
159420
+ },
159421
+ {
159422
+ "epoch": 377.33,
159423
+ "learning_rate": 9.259807692307693e-06,
159424
+ "loss": 0.5026,
159425
+ "step": 47205
159426
+ },
159427
+ {
159428
+ "epoch": 377.37,
159429
+ "learning_rate": 9.259727564102565e-06,
159430
+ "loss": 0.6164,
159431
+ "step": 47210
159432
+ },
159433
+ {
159434
+ "epoch": 377.41,
159435
+ "learning_rate": 9.259647435897436e-06,
159436
+ "loss": 1.2348,
159437
+ "step": 47215
159438
+ },
159439
+ {
159440
+ "epoch": 377.45,
159441
+ "learning_rate": 9.25956730769231e-06,
159442
+ "loss": 0.3612,
159443
+ "step": 47220
159444
+ },
159445
+ {
159446
+ "epoch": 377.49,
159447
+ "learning_rate": 9.25948717948718e-06,
159448
+ "loss": 0.3119,
159449
+ "step": 47225
159450
+ },
159451
+ {
159452
+ "epoch": 377.53,
159453
+ "learning_rate": 9.259407051282052e-06,
159454
+ "loss": 0.3822,
159455
+ "step": 47230
159456
+ },
159457
+ {
159458
+ "epoch": 377.57,
159459
+ "learning_rate": 9.259326923076923e-06,
159460
+ "loss": 0.5769,
159461
+ "step": 47235
159462
+ },
159463
+ {
159464
+ "epoch": 377.61,
159465
+ "learning_rate": 9.259246794871796e-06,
159466
+ "loss": 1.225,
159467
+ "step": 47240
159468
+ },
159469
+ {
159470
+ "epoch": 377.65,
159471
+ "learning_rate": 9.259166666666668e-06,
159472
+ "loss": 0.3264,
159473
+ "step": 47245
159474
+ },
159475
+ {
159476
+ "epoch": 377.69,
159477
+ "learning_rate": 9.259086538461539e-06,
159478
+ "loss": 0.401,
159479
+ "step": 47250
159480
+ },
159481
+ {
159482
+ "epoch": 377.73,
159483
+ "learning_rate": 9.259006410256412e-06,
159484
+ "loss": 0.35,
159485
+ "step": 47255
159486
+ },
159487
+ {
159488
+ "epoch": 377.77,
159489
+ "learning_rate": 9.258926282051283e-06,
159490
+ "loss": 0.5652,
159491
+ "step": 47260
159492
+ },
159493
+ {
159494
+ "epoch": 377.81,
159495
+ "learning_rate": 9.258846153846155e-06,
159496
+ "loss": 1.2207,
159497
+ "step": 47265
159498
+ },
159499
+ {
159500
+ "epoch": 377.85,
159501
+ "learning_rate": 9.258766025641026e-06,
159502
+ "loss": 0.3268,
159503
+ "step": 47270
159504
+ },
159505
+ {
159506
+ "epoch": 377.89,
159507
+ "learning_rate": 9.2586858974359e-06,
159508
+ "loss": 0.3757,
159509
+ "step": 47275
159510
+ },
159511
+ {
159512
+ "epoch": 377.93,
159513
+ "learning_rate": 9.258605769230769e-06,
159514
+ "loss": 0.9012,
159515
+ "step": 47280
159516
+ },
159517
+ {
159518
+ "epoch": 377.97,
159519
+ "learning_rate": 9.258525641025642e-06,
159520
+ "loss": 0.7584,
159521
+ "step": 47285
159522
+ },
159523
+ {
159524
+ "epoch": 378.0,
159525
+ "eval_loss": 0.4493987560272217,
159526
+ "eval_runtime": 40.9245,
159527
+ "eval_samples_per_second": 20.428,
159528
+ "eval_steps_per_second": 0.66,
159529
+ "eval_wer": 0.21467253953354032,
159530
+ "step": 47289
159531
+ },
159532
+ {
159533
+ "epoch": 378.01,
159534
+ "learning_rate": 9.25846153846154e-06,
159535
+ "loss": 0.5,
159536
+ "step": 47290
159537
+ },
159538
+ {
159539
+ "epoch": 378.05,
159540
+ "learning_rate": 9.25838141025641e-06,
159541
+ "loss": 0.3277,
159542
+ "step": 47295
159543
+ },
159544
+ {
159545
+ "epoch": 378.09,
159546
+ "learning_rate": 9.258301282051284e-06,
159547
+ "loss": 0.3122,
159548
+ "step": 47300
159549
+ },
159550
+ {
159551
+ "epoch": 378.13,
159552
+ "learning_rate": 9.258221153846153e-06,
159553
+ "loss": 0.4403,
159554
+ "step": 47305
159555
+ },
159556
+ {
159557
+ "epoch": 378.17,
159558
+ "learning_rate": 9.258141025641026e-06,
159559
+ "loss": 0.604,
159560
+ "step": 47310
159561
+ },
159562
+ {
159563
+ "epoch": 378.21,
159564
+ "learning_rate": 9.258060897435898e-06,
159565
+ "loss": 1.1431,
159566
+ "step": 47315
159567
+ },
159568
+ {
159569
+ "epoch": 378.25,
159570
+ "learning_rate": 9.25798076923077e-06,
159571
+ "loss": 0.3546,
159572
+ "step": 47320
159573
+ },
159574
+ {
159575
+ "epoch": 378.29,
159576
+ "learning_rate": 9.257900641025642e-06,
159577
+ "loss": 0.3311,
159578
+ "step": 47325
159579
+ },
159580
+ {
159581
+ "epoch": 378.33,
159582
+ "learning_rate": 9.257820512820514e-06,
159583
+ "loss": 0.3303,
159584
+ "step": 47330
159585
+ },
159586
+ {
159587
+ "epoch": 378.37,
159588
+ "learning_rate": 9.257740384615385e-06,
159589
+ "loss": 0.638,
159590
+ "step": 47335
159591
+ },
159592
+ {
159593
+ "epoch": 378.41,
159594
+ "learning_rate": 9.257660256410256e-06,
159595
+ "loss": 1.1426,
159596
+ "step": 47340
159597
+ },
159598
+ {
159599
+ "epoch": 378.45,
159600
+ "learning_rate": 9.25758012820513e-06,
159601
+ "loss": 0.3966,
159602
+ "step": 47345
159603
+ },
159604
+ {
159605
+ "epoch": 378.49,
159606
+ "learning_rate": 9.2575e-06,
159607
+ "loss": 0.3601,
159608
+ "step": 47350
159609
+ },
159610
+ {
159611
+ "epoch": 378.53,
159612
+ "learning_rate": 9.257419871794872e-06,
159613
+ "loss": 0.3695,
159614
+ "step": 47355
159615
+ },
159616
+ {
159617
+ "epoch": 378.57,
159618
+ "learning_rate": 9.257339743589743e-06,
159619
+ "loss": 0.7932,
159620
+ "step": 47360
159621
+ },
159622
+ {
159623
+ "epoch": 378.61,
159624
+ "learning_rate": 9.257259615384616e-06,
159625
+ "loss": 1.4832,
159626
+ "step": 47365
159627
+ },
159628
+ {
159629
+ "epoch": 378.65,
159630
+ "learning_rate": 9.257179487179488e-06,
159631
+ "loss": 0.3544,
159632
+ "step": 47370
159633
+ },
159634
+ {
159635
+ "epoch": 378.69,
159636
+ "learning_rate": 9.25709935897436e-06,
159637
+ "loss": 0.3521,
159638
+ "step": 47375
159639
+ },
159640
+ {
159641
+ "epoch": 378.73,
159642
+ "learning_rate": 9.257019230769232e-06,
159643
+ "loss": 0.3812,
159644
+ "step": 47380
159645
+ },
159646
+ {
159647
+ "epoch": 378.77,
159648
+ "learning_rate": 9.256939102564104e-06,
159649
+ "loss": 0.5812,
159650
+ "step": 47385
159651
+ },
159652
+ {
159653
+ "epoch": 378.81,
159654
+ "learning_rate": 9.256858974358975e-06,
159655
+ "loss": 1.0635,
159656
+ "step": 47390
159657
+ },
159658
+ {
159659
+ "epoch": 378.85,
159660
+ "learning_rate": 9.256778846153846e-06,
159661
+ "loss": 0.3511,
159662
+ "step": 47395
159663
+ },
159664
+ {
159665
+ "epoch": 378.89,
159666
+ "learning_rate": 9.25669871794872e-06,
159667
+ "loss": 0.4201,
159668
+ "step": 47400
159669
+ },
159670
+ {
159671
+ "epoch": 378.93,
159672
+ "learning_rate": 9.25661858974359e-06,
159673
+ "loss": 0.3906,
159674
+ "step": 47405
159675
+ },
159676
+ {
159677
+ "epoch": 378.97,
159678
+ "learning_rate": 9.256538461538462e-06,
159679
+ "loss": 0.5379,
159680
+ "step": 47410
159681
+ },
159682
+ {
159683
+ "epoch": 379.0,
159684
+ "eval_loss": 0.41098281741142273,
159685
+ "eval_runtime": 41.7422,
159686
+ "eval_samples_per_second": 20.028,
159687
+ "eval_steps_per_second": 0.647,
159688
+ "eval_wer": 0.20962497290266638,
159689
+ "step": 47414
159690
+ },
159691
+ {
159692
+ "epoch": 382.01,
159693
+ "learning_rate": 9.256458333333333e-06,
159694
+ "loss": 0.4079,
159695
+ "step": 47415
159696
+ },
159697
+ {
159698
+ "epoch": 382.05,
159699
+ "learning_rate": 9.256378205128207e-06,
159700
+ "loss": 0.4051,
159701
+ "step": 47420
159702
+ },
159703
+ {
159704
+ "epoch": 382.09,
159705
+ "learning_rate": 9.256298076923078e-06,
159706
+ "loss": 0.3244,
159707
+ "step": 47425
159708
+ },
159709
+ {
159710
+ "epoch": 382.13,
159711
+ "learning_rate": 9.25621794871795e-06,
159712
+ "loss": 0.3739,
159713
+ "step": 47430
159714
+ },
159715
+ {
159716
+ "epoch": 382.17,
159717
+ "learning_rate": 9.256137820512822e-06,
159718
+ "loss": 0.5677,
159719
+ "step": 47435
159720
+ },
159721
+ {
159722
+ "epoch": 382.21,
159723
+ "learning_rate": 9.256057692307692e-06,
159724
+ "loss": 1.2913,
159725
+ "step": 47440
159726
+ },
159727
+ {
159728
+ "epoch": 382.25,
159729
+ "learning_rate": 9.255977564102565e-06,
159730
+ "loss": 0.3349,
159731
+ "step": 47445
159732
+ },
159733
+ {
159734
+ "epoch": 382.29,
159735
+ "learning_rate": 9.255897435897436e-06,
159736
+ "loss": 0.3091,
159737
+ "step": 47450
159738
+ },
159739
+ {
159740
+ "epoch": 382.33,
159741
+ "learning_rate": 9.255817307692308e-06,
159742
+ "loss": 0.3982,
159743
+ "step": 47455
159744
+ },
159745
+ {
159746
+ "epoch": 382.37,
159747
+ "learning_rate": 9.255737179487179e-06,
159748
+ "loss": 0.7085,
159749
+ "step": 47460
159750
+ },
159751
+ {
159752
+ "epoch": 382.41,
159753
+ "learning_rate": 9.255657051282052e-06,
159754
+ "loss": 1.1537,
159755
+ "step": 47465
159756
+ },
159757
+ {
159758
+ "epoch": 382.45,
159759
+ "learning_rate": 9.255576923076925e-06,
159760
+ "loss": 0.3985,
159761
+ "step": 47470
159762
+ },
159763
+ {
159764
+ "epoch": 382.49,
159765
+ "learning_rate": 9.255496794871795e-06,
159766
+ "loss": 0.3551,
159767
+ "step": 47475
159768
+ },
159769
+ {
159770
+ "epoch": 382.53,
159771
+ "learning_rate": 9.255416666666668e-06,
159772
+ "loss": 0.3884,
159773
+ "step": 47480
159774
+ },
159775
+ {
159776
+ "epoch": 382.57,
159777
+ "learning_rate": 9.25533653846154e-06,
159778
+ "loss": 0.5744,
159779
+ "step": 47485
159780
+ },
159781
+ {
159782
+ "epoch": 382.61,
159783
+ "learning_rate": 9.25525641025641e-06,
159784
+ "loss": 1.2079,
159785
+ "step": 47490
159786
+ },
159787
+ {
159788
+ "epoch": 382.65,
159789
+ "learning_rate": 9.255176282051282e-06,
159790
+ "loss": 0.3573,
159791
+ "step": 47495
159792
+ },
159793
+ {
159794
+ "epoch": 382.69,
159795
+ "learning_rate": 9.255096153846155e-06,
159796
+ "loss": 0.3638,
159797
+ "step": 47500
159798
+ },
159799
+ {
159800
+ "epoch": 382.73,
159801
+ "learning_rate": 9.255016025641026e-06,
159802
+ "loss": 0.373,
159803
+ "step": 47505
159804
+ },
159805
+ {
159806
+ "epoch": 382.77,
159807
+ "learning_rate": 9.254935897435898e-06,
159808
+ "loss": 0.5704,
159809
+ "step": 47510
159810
+ },
159811
+ {
159812
+ "epoch": 382.81,
159813
+ "learning_rate": 9.254855769230769e-06,
159814
+ "loss": 1.3223,
159815
+ "step": 47515
159816
+ },
159817
+ {
159818
+ "epoch": 382.85,
159819
+ "learning_rate": 9.254775641025642e-06,
159820
+ "loss": 0.4135,
159821
+ "step": 47520
159822
+ },
159823
+ {
159824
+ "epoch": 382.89,
159825
+ "learning_rate": 9.254695512820514e-06,
159826
+ "loss": 0.3511,
159827
+ "step": 47525
159828
+ },
159829
+ {
159830
+ "epoch": 382.93,
159831
+ "learning_rate": 9.254615384615385e-06,
159832
+ "loss": 0.443,
159833
+ "step": 47530
159834
+ },
159835
+ {
159836
+ "epoch": 382.97,
159837
+ "learning_rate": 9.254535256410258e-06,
159838
+ "loss": 0.843,
159839
+ "step": 47535
159840
+ },
159841
+ {
159842
+ "epoch": 383.0,
159843
+ "eval_loss": 0.40662431716918945,
159844
+ "eval_runtime": 40.2188,
159845
+ "eval_samples_per_second": 20.786,
159846
+ "eval_steps_per_second": 0.671,
159847
+ "eval_wer": 0.20183823529411765,
159848
+ "step": 47538
159849
+ },
159850
+ {
159851
+ "epoch": 380.02,
159852
+ "learning_rate": 9.25445512820513e-06,
159853
+ "loss": 0.3762,
159854
+ "step": 47540
159855
+ },
159856
+ {
159857
+ "epoch": 380.06,
159858
+ "learning_rate": 9.254375e-06,
159859
+ "loss": 0.2878,
159860
+ "step": 47545
159861
+ },
159862
+ {
159863
+ "epoch": 380.1,
159864
+ "learning_rate": 9.254294871794872e-06,
159865
+ "loss": 0.3521,
159866
+ "step": 47550
159867
+ },
159868
+ {
159869
+ "epoch": 380.14,
159870
+ "learning_rate": 9.254214743589745e-06,
159871
+ "loss": 0.4816,
159872
+ "step": 47555
159873
+ },
159874
+ {
159875
+ "epoch": 380.18,
159876
+ "learning_rate": 9.254134615384616e-06,
159877
+ "loss": 0.6926,
159878
+ "step": 47560
159879
+ },
159880
+ {
159881
+ "epoch": 380.22,
159882
+ "learning_rate": 9.254054487179488e-06,
159883
+ "loss": 1.1028,
159884
+ "step": 47565
159885
+ },
159886
+ {
159887
+ "epoch": 380.26,
159888
+ "learning_rate": 9.25397435897436e-06,
159889
+ "loss": 0.333,
159890
+ "step": 47570
159891
+ },
159892
+ {
159893
+ "epoch": 380.3,
159894
+ "learning_rate": 9.253894230769232e-06,
159895
+ "loss": 0.3393,
159896
+ "step": 47575
159897
+ },
159898
+ {
159899
+ "epoch": 380.34,
159900
+ "learning_rate": 9.253814102564104e-06,
159901
+ "loss": 0.3715,
159902
+ "step": 47580
159903
+ },
159904
+ {
159905
+ "epoch": 380.38,
159906
+ "learning_rate": 9.253733974358975e-06,
159907
+ "loss": 0.7846,
159908
+ "step": 47585
159909
+ },
159910
+ {
159911
+ "epoch": 380.42,
159912
+ "learning_rate": 9.253653846153848e-06,
159913
+ "loss": 1.1017,
159914
+ "step": 47590
159915
+ },
159916
+ {
159917
+ "epoch": 380.46,
159918
+ "learning_rate": 9.253573717948718e-06,
159919
+ "loss": 0.3653,
159920
+ "step": 47595
159921
+ },
159922
+ {
159923
+ "epoch": 380.5,
159924
+ "learning_rate": 9.25349358974359e-06,
159925
+ "loss": 0.3604,
159926
+ "step": 47600
159927
+ },
159928
+ {
159929
+ "epoch": 380.54,
159930
+ "learning_rate": 9.253413461538462e-06,
159931
+ "loss": 0.4365,
159932
+ "step": 47605
159933
+ },
159934
+ {
159935
+ "epoch": 380.58,
159936
+ "learning_rate": 9.253333333333333e-06,
159937
+ "loss": 0.6824,
159938
+ "step": 47610
159939
+ },
159940
+ {
159941
+ "epoch": 380.62,
159942
+ "learning_rate": 9.253253205128205e-06,
159943
+ "loss": 1.1141,
159944
+ "step": 47615
159945
+ },
159946
+ {
159947
+ "epoch": 380.66,
159948
+ "learning_rate": 9.253173076923078e-06,
159949
+ "loss": 0.4078,
159950
+ "step": 47620
159951
+ },
159952
+ {
159953
+ "epoch": 380.7,
159954
+ "learning_rate": 9.25309294871795e-06,
159955
+ "loss": 0.3401,
159956
+ "step": 47625
159957
+ },
159958
+ {
159959
+ "epoch": 380.74,
159960
+ "learning_rate": 9.25301282051282e-06,
159961
+ "loss": 0.4202,
159962
+ "step": 47630
159963
+ },
159964
+ {
159965
+ "epoch": 380.78,
159966
+ "learning_rate": 9.252932692307694e-06,
159967
+ "loss": 0.8217,
159968
+ "step": 47635
159969
+ },
159970
+ {
159971
+ "epoch": 380.82,
159972
+ "learning_rate": 9.252852564102565e-06,
159973
+ "loss": 1.1662,
159974
+ "step": 47640
159975
+ },
159976
+ {
159977
+ "epoch": 380.86,
159978
+ "learning_rate": 9.252772435897436e-06,
159979
+ "loss": 0.384,
159980
+ "step": 47645
159981
+ },
159982
+ {
159983
+ "epoch": 380.9,
159984
+ "learning_rate": 9.252692307692308e-06,
159985
+ "loss": 0.3267,
159986
+ "step": 47650
159987
+ },
159988
+ {
159989
+ "epoch": 380.94,
159990
+ "learning_rate": 9.25261217948718e-06,
159991
+ "loss": 0.4336,
159992
+ "step": 47655
159993
+ },
159994
+ {
159995
+ "epoch": 380.98,
159996
+ "learning_rate": 9.252532051282052e-06,
159997
+ "loss": 0.8251,
159998
+ "step": 47660
159999
+ },
160000
+ {
160001
+ "epoch": 381.0,
160002
+ "eval_loss": 0.4215746521949768,
160003
+ "eval_runtime": 41.2472,
160004
+ "eval_samples_per_second": 20.268,
160005
+ "eval_steps_per_second": 0.655,
160006
+ "eval_wer": 0.20148385219668316,
160007
+ "step": 47663
160008
  }
160009
  ],
160010
+ "max_steps": 625000,
160011
  "num_train_epochs": 5000,
160012
+ "total_flos": 1.3410923324603713e+20,
160013
  "trial_name": null,
160014
  "trial_params": null
160015
  }
model-bin/finetune/base/{checkpoint-47039 β†’ checkpoint-47663}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629726271.057744/events.out.tfevents.1629726271.74272264b15c.932.143 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9db3e9d3ea351bc26cb8b9983127bb0f702a9106cac7ea3dc41b274977c333a
3
+ size 4194
model-bin/finetune/base/log/1629726936.3025346/events.out.tfevents.1629726936.74272264b15c.932.145 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:165e9315c50ee5d69f05a748092d591eedf130eefcb7df44fc318cd85193aede
3
+ size 4194
model-bin/finetune/base/log/1629727569.1115878/events.out.tfevents.1629727569.74272264b15c.932.147 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10f156f766741d9487a186c530690042f45e38b1470071e7c67342fc2f4de57e
3
+ size 4194
model-bin/finetune/base/log/1629728226.273403/events.out.tfevents.1629728226.74272264b15c.932.149 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1c9dc7390405e3fe4eccecef2b743e161f9072b9190fa500eadc2be3b6d7a01
3
+ size 4194
model-bin/finetune/base/log/1629728868.313585/events.out.tfevents.1629728868.74272264b15c.932.151 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c64cddb3bc7b984e1ab64dd8d370fdd804d9667df449b32078a4e5821bd879f
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629726271.74272264b15c.932.142 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03b96f74365f459c84063de933519e6e22eed7bd3f11ef3bff26358ce5781c45
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629726936.74272264b15c.932.144 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f8cad111a8dc44b3e0e2e3199a42f6d4778ea382a280132cd13a1a32431c443
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629727569.74272264b15c.932.146 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68b30bee97f18985d11b998fbada5fda3910081917a0d4e3297fea859b2dba95
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629728226.74272264b15c.932.148 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22eefdf23fa9bf2fa42d1587ed28ef33d72d72209393e4b6d6cf5b6b8edc19a5
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629728868.74272264b15c.932.150 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54c523d74d675d81de49b234b5990db7bd9e205ec6bd89e2b2413b51fcdf1e5b
3
+ size 8622