Check commited on
Commit
68bb7c3
Β·
1 Parent(s): ccccc13

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/trainer_state.json +792 -3
  9. model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630134518.566483/events.out.tfevents.1630134518.86bb0ddabf9b.4092.61 +3 -0
  11. model-bin/finetune/base/log/1630134911.0423136/events.out.tfevents.1630134911.86bb0ddabf9b.4092.63 +3 -0
  12. model-bin/finetune/base/log/1630135301.2668922/events.out.tfevents.1630135301.86bb0ddabf9b.4092.65 +3 -0
  13. model-bin/finetune/base/log/1630135690.2425787/events.out.tfevents.1630135690.86bb0ddabf9b.4092.67 +3 -0
  14. model-bin/finetune/base/log/1630136081.7873354/events.out.tfevents.1630136081.86bb0ddabf9b.4092.69 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630134518.86bb0ddabf9b.4092.60 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630134911.86bb0ddabf9b.4092.62 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630135301.86bb0ddabf9b.4092.64 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630135690.86bb0ddabf9b.4092.66 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630136081.86bb0ddabf9b.4092.68 +3 -0
model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:51fcbfe601df65783aad9ed90b143196bb9116032626928bb4e87c5585f1c83a
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:636ee02e97153db492c1e25610a1815b762f603251f71ecf979797a912931ee1
3
  size 722165393
model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d3904af36b18b4d095fb0124e026e85f54816e0865cc1484799d44b454c39030
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6276c6dd9dc2aaa4f75d517ca8508a8d43751e6445d190d706be9592e9b68088
3
  size 377909911
model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c1fa1ccd4d088fa375b9751da87ad6f087ce1ff3bd1852b863df97283f7f7a5c
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32115bf128a0979d07acf3ff9faa7174ce3f4d18ad456f6b27cb05a856a9246b
3
  size 14503
model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2e6d50a58f036c734ac486a1b8b8de8f2a72641d68799526e48e85c9dc1db370
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:145db2a6bd6fad33acea1d837e65e5d9dc1ffb472af1f46c6cced784aa941fb5
3
  size 559
model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ecb97a7d6b086ed775dfd796d92ea4da8cacd42ec572a22c69cec1b8c8c780d8
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fee945b453a68e33846c1e2b7e09fdb2acfb604bb05d5631e98a77b7387ea20
3
  size 623
model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
- "epoch": 973.0,
5
- "global_step": 121707,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -254214,11 +254214,800 @@
254214
  "eval_steps_per_second": 0.728,
254215
  "eval_wer": 0.18579426172148356,
254216
  "step": 121707
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
254217
  }
254218
  ],
254219
  "max_steps": 625000,
254220
  "num_train_epochs": 5000,
254221
- "total_flos": 3.425064251533899e+20,
254222
  "trial_name": null,
254223
  "trial_params": null
254224
  }
 
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
+ "epoch": 978.0,
5
+ "global_step": 122329,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
254214
  "eval_steps_per_second": 0.728,
254215
  "eval_wer": 0.18579426172148356,
254216
  "step": 121707
254217
+ },
254218
+ {
254219
+ "epoch": 981.02,
254220
+ "learning_rate": 8.066314102564103e-06,
254221
+ "loss": 0.3619,
254222
+ "step": 121710
254223
+ },
254224
+ {
254225
+ "epoch": 981.06,
254226
+ "learning_rate": 8.066233974358975e-06,
254227
+ "loss": 0.2627,
254228
+ "step": 121715
254229
+ },
254230
+ {
254231
+ "epoch": 981.1,
254232
+ "learning_rate": 8.066153846153846e-06,
254233
+ "loss": 0.3249,
254234
+ "step": 121720
254235
+ },
254236
+ {
254237
+ "epoch": 981.14,
254238
+ "learning_rate": 8.066073717948718e-06,
254239
+ "loss": 0.3249,
254240
+ "step": 121725
254241
+ },
254242
+ {
254243
+ "epoch": 981.18,
254244
+ "learning_rate": 8.06599358974359e-06,
254245
+ "loss": 0.7026,
254246
+ "step": 121730
254247
+ },
254248
+ {
254249
+ "epoch": 981.22,
254250
+ "learning_rate": 8.065913461538462e-06,
254251
+ "loss": 0.8091,
254252
+ "step": 121735
254253
+ },
254254
+ {
254255
+ "epoch": 981.27,
254256
+ "learning_rate": 8.065833333333333e-06,
254257
+ "loss": 0.2785,
254258
+ "step": 121740
254259
+ },
254260
+ {
254261
+ "epoch": 981.31,
254262
+ "learning_rate": 8.065753205128206e-06,
254263
+ "loss": 0.2591,
254264
+ "step": 121745
254265
+ },
254266
+ {
254267
+ "epoch": 981.35,
254268
+ "learning_rate": 8.065673076923078e-06,
254269
+ "loss": 0.2905,
254270
+ "step": 121750
254271
+ },
254272
+ {
254273
+ "epoch": 981.39,
254274
+ "learning_rate": 8.065592948717949e-06,
254275
+ "loss": 0.709,
254276
+ "step": 121755
254277
+ },
254278
+ {
254279
+ "epoch": 981.43,
254280
+ "learning_rate": 8.06551282051282e-06,
254281
+ "loss": 0.7867,
254282
+ "step": 121760
254283
+ },
254284
+ {
254285
+ "epoch": 981.47,
254286
+ "learning_rate": 8.065432692307694e-06,
254287
+ "loss": 0.247,
254288
+ "step": 121765
254289
+ },
254290
+ {
254291
+ "epoch": 981.51,
254292
+ "learning_rate": 8.065352564102565e-06,
254293
+ "loss": 0.3309,
254294
+ "step": 121770
254295
+ },
254296
+ {
254297
+ "epoch": 981.55,
254298
+ "learning_rate": 8.065272435897436e-06,
254299
+ "loss": 0.3943,
254300
+ "step": 121775
254301
+ },
254302
+ {
254303
+ "epoch": 981.59,
254304
+ "learning_rate": 8.06519230769231e-06,
254305
+ "loss": 0.6668,
254306
+ "step": 121780
254307
+ },
254308
+ {
254309
+ "epoch": 981.63,
254310
+ "learning_rate": 8.06511217948718e-06,
254311
+ "loss": 0.8405,
254312
+ "step": 121785
254313
+ },
254314
+ {
254315
+ "epoch": 981.67,
254316
+ "learning_rate": 8.065032051282052e-06,
254317
+ "loss": 0.3276,
254318
+ "step": 121790
254319
+ },
254320
+ {
254321
+ "epoch": 981.71,
254322
+ "learning_rate": 8.064951923076923e-06,
254323
+ "loss": 0.3355,
254324
+ "step": 121795
254325
+ },
254326
+ {
254327
+ "epoch": 981.75,
254328
+ "learning_rate": 8.064871794871796e-06,
254329
+ "loss": 0.3806,
254330
+ "step": 121800
254331
+ },
254332
+ {
254333
+ "epoch": 981.79,
254334
+ "learning_rate": 8.064791666666668e-06,
254335
+ "loss": 0.9231,
254336
+ "step": 121805
254337
+ },
254338
+ {
254339
+ "epoch": 981.83,
254340
+ "learning_rate": 8.064711538461539e-06,
254341
+ "loss": 0.8035,
254342
+ "step": 121810
254343
+ },
254344
+ {
254345
+ "epoch": 981.87,
254346
+ "learning_rate": 8.06463141025641e-06,
254347
+ "loss": 0.286,
254348
+ "step": 121815
254349
+ },
254350
+ {
254351
+ "epoch": 981.91,
254352
+ "learning_rate": 8.064551282051284e-06,
254353
+ "loss": 0.2576,
254354
+ "step": 121820
254355
+ },
254356
+ {
254357
+ "epoch": 981.95,
254358
+ "learning_rate": 8.064471153846153e-06,
254359
+ "loss": 0.3875,
254360
+ "step": 121825
254361
+ },
254362
+ {
254363
+ "epoch": 981.99,
254364
+ "learning_rate": 8.064391025641026e-06,
254365
+ "loss": 0.9652,
254366
+ "step": 121830
254367
+ },
254368
+ {
254369
+ "epoch": 982.0,
254370
+ "eval_loss": 0.38985419273376465,
254371
+ "eval_runtime": 36.1301,
254372
+ "eval_samples_per_second": 23.305,
254373
+ "eval_steps_per_second": 0.747,
254374
+ "eval_wer": 0.183689917936694,
254375
+ "step": 121831
254376
+ },
254377
+ {
254378
+ "epoch": 974.03,
254379
+ "learning_rate": 8.0643108974359e-06,
254380
+ "loss": 0.3079,
254381
+ "step": 121835
254382
+ },
254383
+ {
254384
+ "epoch": 974.07,
254385
+ "learning_rate": 8.064230769230769e-06,
254386
+ "loss": 0.2465,
254387
+ "step": 121840
254388
+ },
254389
+ {
254390
+ "epoch": 974.11,
254391
+ "learning_rate": 8.064150641025642e-06,
254392
+ "loss": 0.3261,
254393
+ "step": 121845
254394
+ },
254395
+ {
254396
+ "epoch": 974.15,
254397
+ "learning_rate": 8.064070512820513e-06,
254398
+ "loss": 0.4024,
254399
+ "step": 121850
254400
+ },
254401
+ {
254402
+ "epoch": 974.19,
254403
+ "learning_rate": 8.063990384615385e-06,
254404
+ "loss": 0.931,
254405
+ "step": 121855
254406
+ },
254407
+ {
254408
+ "epoch": 974.23,
254409
+ "learning_rate": 8.063910256410256e-06,
254410
+ "loss": 0.7664,
254411
+ "step": 121860
254412
+ },
254413
+ {
254414
+ "epoch": 974.27,
254415
+ "learning_rate": 8.063830128205129e-06,
254416
+ "loss": 0.268,
254417
+ "step": 121865
254418
+ },
254419
+ {
254420
+ "epoch": 974.31,
254421
+ "learning_rate": 8.06375e-06,
254422
+ "loss": 0.3031,
254423
+ "step": 121870
254424
+ },
254425
+ {
254426
+ "epoch": 974.35,
254427
+ "learning_rate": 8.063669871794872e-06,
254428
+ "loss": 0.3893,
254429
+ "step": 121875
254430
+ },
254431
+ {
254432
+ "epoch": 974.39,
254433
+ "learning_rate": 8.063589743589745e-06,
254434
+ "loss": 0.9955,
254435
+ "step": 121880
254436
+ },
254437
+ {
254438
+ "epoch": 974.43,
254439
+ "learning_rate": 8.063509615384616e-06,
254440
+ "loss": 0.6156,
254441
+ "step": 121885
254442
+ },
254443
+ {
254444
+ "epoch": 974.47,
254445
+ "learning_rate": 8.063429487179488e-06,
254446
+ "loss": 0.2728,
254447
+ "step": 121890
254448
+ },
254449
+ {
254450
+ "epoch": 974.51,
254451
+ "learning_rate": 8.063349358974359e-06,
254452
+ "loss": 0.2574,
254453
+ "step": 121895
254454
+ },
254455
+ {
254456
+ "epoch": 974.55,
254457
+ "learning_rate": 8.063269230769232e-06,
254458
+ "loss": 0.3403,
254459
+ "step": 121900
254460
+ },
254461
+ {
254462
+ "epoch": 974.59,
254463
+ "learning_rate": 8.063189102564103e-06,
254464
+ "loss": 0.9278,
254465
+ "step": 121905
254466
+ },
254467
+ {
254468
+ "epoch": 974.63,
254469
+ "learning_rate": 8.063108974358975e-06,
254470
+ "loss": 0.6179,
254471
+ "step": 121910
254472
+ },
254473
+ {
254474
+ "epoch": 974.67,
254475
+ "learning_rate": 8.063028846153846e-06,
254476
+ "loss": 0.2698,
254477
+ "step": 121915
254478
+ },
254479
+ {
254480
+ "epoch": 974.71,
254481
+ "learning_rate": 8.06294871794872e-06,
254482
+ "loss": 0.2849,
254483
+ "step": 121920
254484
+ },
254485
+ {
254486
+ "epoch": 974.75,
254487
+ "learning_rate": 8.06286858974359e-06,
254488
+ "loss": 0.382,
254489
+ "step": 121925
254490
+ },
254491
+ {
254492
+ "epoch": 974.79,
254493
+ "learning_rate": 8.062788461538462e-06,
254494
+ "loss": 0.8746,
254495
+ "step": 121930
254496
+ },
254497
+ {
254498
+ "epoch": 974.83,
254499
+ "learning_rate": 8.062708333333335e-06,
254500
+ "loss": 0.6217,
254501
+ "step": 121935
254502
+ },
254503
+ {
254504
+ "epoch": 974.87,
254505
+ "learning_rate": 8.062628205128206e-06,
254506
+ "loss": 0.2796,
254507
+ "step": 121940
254508
+ },
254509
+ {
254510
+ "epoch": 974.91,
254511
+ "learning_rate": 8.062548076923078e-06,
254512
+ "loss": 0.3042,
254513
+ "step": 121945
254514
+ },
254515
+ {
254516
+ "epoch": 974.95,
254517
+ "learning_rate": 8.062467948717949e-06,
254518
+ "loss": 0.4009,
254519
+ "step": 121950
254520
+ },
254521
+ {
254522
+ "epoch": 974.99,
254523
+ "learning_rate": 8.062387820512822e-06,
254524
+ "loss": 1.0557,
254525
+ "step": 121955
254526
+ },
254527
+ {
254528
+ "epoch": 975.0,
254529
+ "eval_loss": 0.4002327620983124,
254530
+ "eval_runtime": 36.1482,
254531
+ "eval_samples_per_second": 23.265,
254532
+ "eval_steps_per_second": 0.747,
254533
+ "eval_wer": 0.18004561842395703,
254534
+ "step": 121956
254535
+ },
254536
+ {
254537
+ "epoch": 983.03,
254538
+ "learning_rate": 8.062307692307693e-06,
254539
+ "loss": 0.3859,
254540
+ "step": 121960
254541
+ },
254542
+ {
254543
+ "epoch": 983.07,
254544
+ "learning_rate": 8.062227564102565e-06,
254545
+ "loss": 0.2278,
254546
+ "step": 121965
254547
+ },
254548
+ {
254549
+ "epoch": 983.11,
254550
+ "learning_rate": 8.062147435897436e-06,
254551
+ "loss": 0.2952,
254552
+ "step": 121970
254553
+ },
254554
+ {
254555
+ "epoch": 983.15,
254556
+ "learning_rate": 8.06206730769231e-06,
254557
+ "loss": 0.337,
254558
+ "step": 121975
254559
+ },
254560
+ {
254561
+ "epoch": 983.19,
254562
+ "learning_rate": 8.06198717948718e-06,
254563
+ "loss": 1.0278,
254564
+ "step": 121980
254565
+ },
254566
+ {
254567
+ "epoch": 983.23,
254568
+ "learning_rate": 8.061907051282052e-06,
254569
+ "loss": 0.6666,
254570
+ "step": 121985
254571
+ },
254572
+ {
254573
+ "epoch": 983.27,
254574
+ "learning_rate": 8.061826923076925e-06,
254575
+ "loss": 0.2537,
254576
+ "step": 121990
254577
+ },
254578
+ {
254579
+ "epoch": 983.31,
254580
+ "learning_rate": 8.061746794871795e-06,
254581
+ "loss": 0.336,
254582
+ "step": 121995
254583
+ },
254584
+ {
254585
+ "epoch": 983.35,
254586
+ "learning_rate": 8.061666666666668e-06,
254587
+ "loss": 0.4248,
254588
+ "step": 122000
254589
+ },
254590
+ {
254591
+ "epoch": 983.4,
254592
+ "learning_rate": 8.061586538461539e-06,
254593
+ "loss": 0.8619,
254594
+ "step": 122005
254595
+ },
254596
+ {
254597
+ "epoch": 983.44,
254598
+ "learning_rate": 8.06150641025641e-06,
254599
+ "loss": 0.578,
254600
+ "step": 122010
254601
+ },
254602
+ {
254603
+ "epoch": 983.48,
254604
+ "learning_rate": 8.061426282051282e-06,
254605
+ "loss": 0.2749,
254606
+ "step": 122015
254607
+ },
254608
+ {
254609
+ "epoch": 983.52,
254610
+ "learning_rate": 8.061346153846155e-06,
254611
+ "loss": 0.2588,
254612
+ "step": 122020
254613
+ },
254614
+ {
254615
+ "epoch": 983.56,
254616
+ "learning_rate": 8.061266025641026e-06,
254617
+ "loss": 0.4149,
254618
+ "step": 122025
254619
+ },
254620
+ {
254621
+ "epoch": 983.6,
254622
+ "learning_rate": 8.061185897435898e-06,
254623
+ "loss": 0.9949,
254624
+ "step": 122030
254625
+ },
254626
+ {
254627
+ "epoch": 983.64,
254628
+ "learning_rate": 8.06110576923077e-06,
254629
+ "loss": 0.7007,
254630
+ "step": 122035
254631
+ },
254632
+ {
254633
+ "epoch": 983.68,
254634
+ "learning_rate": 8.061025641025642e-06,
254635
+ "loss": 0.2401,
254636
+ "step": 122040
254637
+ },
254638
+ {
254639
+ "epoch": 983.72,
254640
+ "learning_rate": 8.060945512820513e-06,
254641
+ "loss": 0.3475,
254642
+ "step": 122045
254643
+ },
254644
+ {
254645
+ "epoch": 983.76,
254646
+ "learning_rate": 8.060865384615385e-06,
254647
+ "loss": 0.409,
254648
+ "step": 122050
254649
+ },
254650
+ {
254651
+ "epoch": 983.8,
254652
+ "learning_rate": 8.060785256410258e-06,
254653
+ "loss": 0.8656,
254654
+ "step": 122055
254655
+ },
254656
+ {
254657
+ "epoch": 983.84,
254658
+ "learning_rate": 8.060705128205129e-06,
254659
+ "loss": 0.7616,
254660
+ "step": 122060
254661
+ },
254662
+ {
254663
+ "epoch": 983.88,
254664
+ "learning_rate": 8.060625e-06,
254665
+ "loss": 0.2378,
254666
+ "step": 122065
254667
+ },
254668
+ {
254669
+ "epoch": 983.92,
254670
+ "learning_rate": 8.060544871794872e-06,
254671
+ "loss": 0.341,
254672
+ "step": 122070
254673
+ },
254674
+ {
254675
+ "epoch": 983.96,
254676
+ "learning_rate": 8.060464743589745e-06,
254677
+ "loss": 0.4632,
254678
+ "step": 122075
254679
+ },
254680
+ {
254681
+ "epoch": 984.0,
254682
+ "learning_rate": 8.060384615384616e-06,
254683
+ "loss": 1.1266,
254684
+ "step": 122080
254685
+ },
254686
+ {
254687
+ "epoch": 984.0,
254688
+ "eval_loss": 0.42599642276763916,
254689
+ "eval_runtime": 36.0836,
254690
+ "eval_samples_per_second": 23.307,
254691
+ "eval_steps_per_second": 0.748,
254692
+ "eval_wer": 0.1875638592906145,
254693
+ "step": 122080
254694
+ },
254695
+ {
254696
+ "epoch": 984.04,
254697
+ "learning_rate": 8.060304487179488e-06,
254698
+ "loss": 0.3444,
254699
+ "step": 122085
254700
+ },
254701
+ {
254702
+ "epoch": 984.08,
254703
+ "learning_rate": 8.06022435897436e-06,
254704
+ "loss": 0.32,
254705
+ "step": 122090
254706
+ },
254707
+ {
254708
+ "epoch": 984.12,
254709
+ "learning_rate": 8.060144230769232e-06,
254710
+ "loss": 0.2597,
254711
+ "step": 122095
254712
+ },
254713
+ {
254714
+ "epoch": 984.16,
254715
+ "learning_rate": 8.060064102564103e-06,
254716
+ "loss": 0.4258,
254717
+ "step": 122100
254718
+ },
254719
+ {
254720
+ "epoch": 984.2,
254721
+ "learning_rate": 8.059983974358975e-06,
254722
+ "loss": 1.2592,
254723
+ "step": 122105
254724
+ },
254725
+ {
254726
+ "epoch": 984.24,
254727
+ "learning_rate": 8.059903846153848e-06,
254728
+ "loss": 0.319,
254729
+ "step": 122110
254730
+ },
254731
+ {
254732
+ "epoch": 984.28,
254733
+ "learning_rate": 8.059823717948717e-06,
254734
+ "loss": 0.2761,
254735
+ "step": 122115
254736
+ },
254737
+ {
254738
+ "epoch": 984.32,
254739
+ "learning_rate": 8.05974358974359e-06,
254740
+ "loss": 0.2952,
254741
+ "step": 122120
254742
+ },
254743
+ {
254744
+ "epoch": 984.36,
254745
+ "learning_rate": 8.059663461538462e-06,
254746
+ "loss": 0.4896,
254747
+ "step": 122125
254748
+ },
254749
+ {
254750
+ "epoch": 984.4,
254751
+ "learning_rate": 8.059583333333333e-06,
254752
+ "loss": 1.2025,
254753
+ "step": 122130
254754
+ },
254755
+ {
254756
+ "epoch": 984.44,
254757
+ "learning_rate": 8.059503205128206e-06,
254758
+ "loss": 0.3081,
254759
+ "step": 122135
254760
+ },
254761
+ {
254762
+ "epoch": 984.48,
254763
+ "learning_rate": 8.059423076923078e-06,
254764
+ "loss": 0.2566,
254765
+ "step": 122140
254766
+ },
254767
+ {
254768
+ "epoch": 984.52,
254769
+ "learning_rate": 8.059342948717949e-06,
254770
+ "loss": 0.2943,
254771
+ "step": 122145
254772
+ },
254773
+ {
254774
+ "epoch": 984.56,
254775
+ "learning_rate": 8.05926282051282e-06,
254776
+ "loss": 0.432,
254777
+ "step": 122150
254778
+ },
254779
+ {
254780
+ "epoch": 984.6,
254781
+ "learning_rate": 8.059182692307693e-06,
254782
+ "loss": 1.2402,
254783
+ "step": 122155
254784
+ },
254785
+ {
254786
+ "epoch": 984.65,
254787
+ "learning_rate": 8.059102564102565e-06,
254788
+ "loss": 0.3292,
254789
+ "step": 122160
254790
+ },
254791
+ {
254792
+ "epoch": 984.69,
254793
+ "learning_rate": 8.059022435897436e-06,
254794
+ "loss": 0.3257,
254795
+ "step": 122165
254796
+ },
254797
+ {
254798
+ "epoch": 984.73,
254799
+ "learning_rate": 8.058942307692307e-06,
254800
+ "loss": 0.2744,
254801
+ "step": 122170
254802
+ },
254803
+ {
254804
+ "epoch": 984.77,
254805
+ "learning_rate": 8.05886217948718e-06,
254806
+ "loss": 0.5238,
254807
+ "step": 122175
254808
+ },
254809
+ {
254810
+ "epoch": 984.81,
254811
+ "learning_rate": 8.058782051282052e-06,
254812
+ "loss": 1.13,
254813
+ "step": 122180
254814
+ },
254815
+ {
254816
+ "epoch": 984.85,
254817
+ "learning_rate": 8.058701923076923e-06,
254818
+ "loss": 0.3126,
254819
+ "step": 122185
254820
+ },
254821
+ {
254822
+ "epoch": 984.89,
254823
+ "learning_rate": 8.058621794871796e-06,
254824
+ "loss": 0.2472,
254825
+ "step": 122190
254826
+ },
254827
+ {
254828
+ "epoch": 984.93,
254829
+ "learning_rate": 8.058541666666668e-06,
254830
+ "loss": 0.2963,
254831
+ "step": 122195
254832
+ },
254833
+ {
254834
+ "epoch": 984.97,
254835
+ "learning_rate": 8.058461538461539e-06,
254836
+ "loss": 0.5364,
254837
+ "step": 122200
254838
+ },
254839
+ {
254840
+ "epoch": 985.0,
254841
+ "eval_loss": 0.44340452551841736,
254842
+ "eval_runtime": 36.0415,
254843
+ "eval_samples_per_second": 23.362,
254844
+ "eval_steps_per_second": 0.749,
254845
+ "eval_wer": 0.1748058921703795,
254846
+ "step": 122204
254847
+ },
254848
+ {
254849
+ "epoch": 977.01,
254850
+ "learning_rate": 8.05838141025641e-06,
254851
+ "loss": 0.3177,
254852
+ "step": 122205
254853
+ },
254854
+ {
254855
+ "epoch": 977.05,
254856
+ "learning_rate": 8.058301282051283e-06,
254857
+ "loss": 0.2794,
254858
+ "step": 122210
254859
+ },
254860
+ {
254861
+ "epoch": 977.09,
254862
+ "learning_rate": 8.058221153846155e-06,
254863
+ "loss": 0.2729,
254864
+ "step": 122215
254865
+ },
254866
+ {
254867
+ "epoch": 977.13,
254868
+ "learning_rate": 8.058141025641026e-06,
254869
+ "loss": 0.3982,
254870
+ "step": 122220
254871
+ },
254872
+ {
254873
+ "epoch": 977.17,
254874
+ "learning_rate": 8.058060897435897e-06,
254875
+ "loss": 0.5194,
254876
+ "step": 122225
254877
+ },
254878
+ {
254879
+ "epoch": 977.21,
254880
+ "learning_rate": 8.05798076923077e-06,
254881
+ "loss": 1.0576,
254882
+ "step": 122230
254883
+ },
254884
+ {
254885
+ "epoch": 977.25,
254886
+ "learning_rate": 8.057900641025642e-06,
254887
+ "loss": 0.3235,
254888
+ "step": 122235
254889
+ },
254890
+ {
254891
+ "epoch": 977.29,
254892
+ "learning_rate": 8.057820512820513e-06,
254893
+ "loss": 0.3022,
254894
+ "step": 122240
254895
+ },
254896
+ {
254897
+ "epoch": 977.33,
254898
+ "learning_rate": 8.057740384615386e-06,
254899
+ "loss": 0.4027,
254900
+ "step": 122245
254901
+ },
254902
+ {
254903
+ "epoch": 977.37,
254904
+ "learning_rate": 8.057660256410258e-06,
254905
+ "loss": 0.5891,
254906
+ "step": 122250
254907
+ },
254908
+ {
254909
+ "epoch": 977.41,
254910
+ "learning_rate": 8.057580128205129e-06,
254911
+ "loss": 0.9342,
254912
+ "step": 122255
254913
+ },
254914
+ {
254915
+ "epoch": 977.45,
254916
+ "learning_rate": 8.0575e-06,
254917
+ "loss": 0.2948,
254918
+ "step": 122260
254919
+ },
254920
+ {
254921
+ "epoch": 977.49,
254922
+ "learning_rate": 8.057419871794873e-06,
254923
+ "loss": 0.3486,
254924
+ "step": 122265
254925
+ },
254926
+ {
254927
+ "epoch": 977.53,
254928
+ "learning_rate": 8.057339743589743e-06,
254929
+ "loss": 0.3783,
254930
+ "step": 122270
254931
+ },
254932
+ {
254933
+ "epoch": 977.57,
254934
+ "learning_rate": 8.057259615384616e-06,
254935
+ "loss": 0.5175,
254936
+ "step": 122275
254937
+ },
254938
+ {
254939
+ "epoch": 977.61,
254940
+ "learning_rate": 8.05717948717949e-06,
254941
+ "loss": 1.0765,
254942
+ "step": 122280
254943
+ },
254944
+ {
254945
+ "epoch": 977.65,
254946
+ "learning_rate": 8.057099358974359e-06,
254947
+ "loss": 0.303,
254948
+ "step": 122285
254949
+ },
254950
+ {
254951
+ "epoch": 977.69,
254952
+ "learning_rate": 8.057019230769232e-06,
254953
+ "loss": 0.2748,
254954
+ "step": 122290
254955
+ },
254956
+ {
254957
+ "epoch": 977.73,
254958
+ "learning_rate": 8.056939102564103e-06,
254959
+ "loss": 0.3423,
254960
+ "step": 122295
254961
+ },
254962
+ {
254963
+ "epoch": 977.77,
254964
+ "learning_rate": 8.056858974358975e-06,
254965
+ "loss": 0.6243,
254966
+ "step": 122300
254967
+ },
254968
+ {
254969
+ "epoch": 977.81,
254970
+ "learning_rate": 8.056778846153846e-06,
254971
+ "loss": 1.0672,
254972
+ "step": 122305
254973
+ },
254974
+ {
254975
+ "epoch": 977.85,
254976
+ "learning_rate": 8.056698717948719e-06,
254977
+ "loss": 0.3127,
254978
+ "step": 122310
254979
+ },
254980
+ {
254981
+ "epoch": 977.89,
254982
+ "learning_rate": 8.05661858974359e-06,
254983
+ "loss": 0.3113,
254984
+ "step": 122315
254985
+ },
254986
+ {
254987
+ "epoch": 977.93,
254988
+ "learning_rate": 8.056538461538462e-06,
254989
+ "loss": 0.2507,
254990
+ "step": 122320
254991
+ },
254992
+ {
254993
+ "epoch": 977.97,
254994
+ "learning_rate": 8.056458333333333e-06,
254995
+ "loss": 0.5587,
254996
+ "step": 122325
254997
+ },
254998
+ {
254999
+ "epoch": 978.0,
255000
+ "eval_loss": 0.35215896368026733,
255001
+ "eval_runtime": 35.884,
255002
+ "eval_samples_per_second": 23.465,
255003
+ "eval_steps_per_second": 0.752,
255004
+ "eval_wer": 0.18008964719491036,
255005
+ "step": 122329
255006
  }
255007
  ],
255008
  "max_steps": 625000,
255009
  "num_train_epochs": 5000,
255010
+ "total_flos": 3.442586939742868e+20,
255011
  "trial_name": null,
255012
  "trial_params": null
255013
  }
model-bin/finetune/base/{checkpoint-121707 β†’ checkpoint-122329}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630134518.566483/events.out.tfevents.1630134518.86bb0ddabf9b.4092.61 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e612e579b37ed3f331d52a2c89d46b56a6cb14229212b5523e27c975350b9ac2
3
+ size 4194
model-bin/finetune/base/log/1630134911.0423136/events.out.tfevents.1630134911.86bb0ddabf9b.4092.63 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:437150c6a57d50409e9dd83654e814b93fbd96cbc247badda53f8484f9649b4f
3
+ size 4194
model-bin/finetune/base/log/1630135301.2668922/events.out.tfevents.1630135301.86bb0ddabf9b.4092.65 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1bbb96a36c4e01cb227a07f6311fdaeb0a5cac9c8ecc28604dd66045d5567a4
3
+ size 4194
model-bin/finetune/base/log/1630135690.2425787/events.out.tfevents.1630135690.86bb0ddabf9b.4092.67 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae1f703a01b139b6bfcf93aff706fec547dad2da4865432048c6da6585ace025
3
+ size 4194
model-bin/finetune/base/log/1630136081.7873354/events.out.tfevents.1630136081.86bb0ddabf9b.4092.69 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f35a49bc3d7eaa575e4ebf872706afa6ee96b401d163af9f4804c1b6603c6c37
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630134518.86bb0ddabf9b.4092.60 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:835f087cf9e7906ca11f62e4d8edc8a144ac59419fccc5c116ae560110d5c9e5
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630134911.86bb0ddabf9b.4092.62 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:758f52c60958962f54f9c344ee09e16b3f9c014d5329b12e18f2cc2991a589b3
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630135301.86bb0ddabf9b.4092.64 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0da5df6c8fc06509ae3f8711a1c3109fbad0fec9510e894e05ea780fac1a6c4
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630135690.86bb0ddabf9b.4092.66 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66c2d786875e9b986a2b2895342a2759d4591e5ab2cb2eea2c56bf327e595daa
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1630136081.86bb0ddabf9b.4092.68 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d97cd85878dee2434036a92620b62dc8daa43d16fc730ef9a67636ab629857f
3
+ size 8622