Check commited on
Commit
69dcd5b
Β·
1 Parent(s): 58df76a

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/rng_state.pth +2 -2
  6. model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/trainer_state.json +799 -4
  9. model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629837488.925602/events.out.tfevents.1629837488.c435e1c5ee04.920.221 +3 -0
  11. model-bin/finetune/base/log/1629838113.24123/events.out.tfevents.1629838113.c435e1c5ee04.920.223 +3 -0
  12. model-bin/finetune/base/log/1629838737.3102496/events.out.tfevents.1629838737.c435e1c5ee04.920.225 +3 -0
  13. model-bin/finetune/base/log/1629839378.965553/events.out.tfevents.1629839378.c435e1c5ee04.920.227 +3 -0
  14. model-bin/finetune/base/log/1629840015.2190616/events.out.tfevents.1629840015.c435e1c5ee04.920.229 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629837488.c435e1c5ee04.920.220 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629838113.c435e1c5ee04.920.222 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629838737.c435e1c5ee04.920.224 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629839378.c435e1c5ee04.920.226 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629840015.c435e1c5ee04.920.228 +3 -0
model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cad4194b8552407741480b0d7cfec2bbde02181f458dbb03ec6afcfecdd08026
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bf43d968dbe02ae7a22a40fea3f3b35c92eea6da15a59b6a0bf2b3a511f4915
3
  size 722165393
model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:43eaa79891cd0d56e96b0fc5b84927dae77481a7bceab190190510d141ccdbf1
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb02808d8842791dc4116b2aee9127299d59ea802c1a61ba6ba3092f36c849da
3
  size 377909911
model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:412a8b45bf8449c4c28a263b3f9f3c718a03c421d731fa385f8caba49e11373c
3
- size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e38e660e121c61cfbe53b56b9b19b73618319ff2d4b4f0c08beadc29f3062f73
3
+ size 14567
model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b996ed8da366dae97a015b33368c0ecf067f550c8dd285673547842926659440
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6497b3560f4dce675de0441236126359cb27ad7919d73715dda52b9f6fbfb5a6
3
  size 559
model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1f3db58c27f09014af98cc064e08602c5304eb789e6d0bbd26dfcf9aa36490f9
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ca10994f54dd3af978a50aed2d65798b9b5d3bd7145e83c529f6a03bd9f42bf
3
  size 623
model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1855363713557883,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-67447",
4
- "epoch": 540.0,
5
- "global_step": 67572,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -185337,11 +185337,806 @@
185337
  "eval_steps_per_second": 0.705,
185338
  "eval_wer": 0.192304851931172,
185339
  "step": 67572
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
185340
  }
185341
  ],
185342
- "max_steps": 625000,
185343
  "num_train_epochs": 5000,
185344
- "total_flos": 1.901450150090872e+20,
185345
  "trial_name": null,
185346
  "trial_params": null
185347
  }
 
1
  {
2
  "best_metric": 0.1855363713557883,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-67447",
4
+ "epoch": 548.995983935743,
5
+ "global_step": 68195,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
185337
  "eval_steps_per_second": 0.705,
185338
  "eval_wer": 0.192304851931172,
185339
  "step": 67572
185340
+ },
185341
+ {
185342
+ "epoch": 540.02,
185343
+ "learning_rate": 8.93349358974359e-06,
185344
+ "loss": 0.3757,
185345
+ "step": 67575
185346
+ },
185347
+ {
185348
+ "epoch": 540.06,
185349
+ "learning_rate": 8.933413461538463e-06,
185350
+ "loss": 0.3056,
185351
+ "step": 67580
185352
+ },
185353
+ {
185354
+ "epoch": 540.1,
185355
+ "learning_rate": 8.933333333333333e-06,
185356
+ "loss": 0.3568,
185357
+ "step": 67585
185358
+ },
185359
+ {
185360
+ "epoch": 540.14,
185361
+ "learning_rate": 8.933253205128206e-06,
185362
+ "loss": 0.4253,
185363
+ "step": 67590
185364
+ },
185365
+ {
185366
+ "epoch": 540.18,
185367
+ "learning_rate": 8.933173076923077e-06,
185368
+ "loss": 0.9211,
185369
+ "step": 67595
185370
+ },
185371
+ {
185372
+ "epoch": 540.22,
185373
+ "learning_rate": 8.933092948717949e-06,
185374
+ "loss": 1.0014,
185375
+ "step": 67600
185376
+ },
185377
+ {
185378
+ "epoch": 540.26,
185379
+ "learning_rate": 8.93301282051282e-06,
185380
+ "loss": 0.5676,
185381
+ "step": 67605
185382
+ },
185383
+ {
185384
+ "epoch": 540.3,
185385
+ "learning_rate": 8.932932692307693e-06,
185386
+ "loss": 0.3648,
185387
+ "step": 67610
185388
+ },
185389
+ {
185390
+ "epoch": 540.34,
185391
+ "learning_rate": 8.932852564102565e-06,
185392
+ "loss": 0.4003,
185393
+ "step": 67615
185394
+ },
185395
+ {
185396
+ "epoch": 540.38,
185397
+ "learning_rate": 8.932772435897436e-06,
185398
+ "loss": 0.8113,
185399
+ "step": 67620
185400
+ },
185401
+ {
185402
+ "epoch": 540.42,
185403
+ "learning_rate": 8.932692307692309e-06,
185404
+ "loss": 1.0352,
185405
+ "step": 67625
185406
+ },
185407
+ {
185408
+ "epoch": 540.46,
185409
+ "learning_rate": 8.93261217948718e-06,
185410
+ "loss": 0.3178,
185411
+ "step": 67630
185412
+ },
185413
+ {
185414
+ "epoch": 540.5,
185415
+ "learning_rate": 8.932532051282052e-06,
185416
+ "loss": 0.5683,
185417
+ "step": 67635
185418
+ },
185419
+ {
185420
+ "epoch": 540.54,
185421
+ "learning_rate": 8.932451923076923e-06,
185422
+ "loss": 0.3428,
185423
+ "step": 67640
185424
+ },
185425
+ {
185426
+ "epoch": 540.58,
185427
+ "learning_rate": 8.932371794871796e-06,
185428
+ "loss": 0.7413,
185429
+ "step": 67645
185430
+ },
185431
+ {
185432
+ "epoch": 540.62,
185433
+ "learning_rate": 8.932291666666668e-06,
185434
+ "loss": 0.8567,
185435
+ "step": 67650
185436
+ },
185437
+ {
185438
+ "epoch": 540.66,
185439
+ "learning_rate": 8.932211538461539e-06,
185440
+ "loss": 0.314,
185441
+ "step": 67655
185442
+ },
185443
+ {
185444
+ "epoch": 540.7,
185445
+ "learning_rate": 8.93213141025641e-06,
185446
+ "loss": 0.2992,
185447
+ "step": 67660
185448
+ },
185449
+ {
185450
+ "epoch": 540.74,
185451
+ "learning_rate": 8.932051282051283e-06,
185452
+ "loss": 0.3365,
185453
+ "step": 67665
185454
+ },
185455
+ {
185456
+ "epoch": 540.78,
185457
+ "learning_rate": 8.931971153846155e-06,
185458
+ "loss": 0.7477,
185459
+ "step": 67670
185460
+ },
185461
+ {
185462
+ "epoch": 540.82,
185463
+ "learning_rate": 8.931891025641026e-06,
185464
+ "loss": 0.9346,
185465
+ "step": 67675
185466
+ },
185467
+ {
185468
+ "epoch": 540.86,
185469
+ "learning_rate": 8.931810897435899e-06,
185470
+ "loss": 0.3394,
185471
+ "step": 67680
185472
+ },
185473
+ {
185474
+ "epoch": 540.9,
185475
+ "learning_rate": 8.93173076923077e-06,
185476
+ "loss": 0.2948,
185477
+ "step": 67685
185478
+ },
185479
+ {
185480
+ "epoch": 540.94,
185481
+ "learning_rate": 8.931650641025642e-06,
185482
+ "loss": 0.4479,
185483
+ "step": 67690
185484
+ },
185485
+ {
185486
+ "epoch": 540.98,
185487
+ "learning_rate": 8.931570512820513e-06,
185488
+ "loss": 0.7604,
185489
+ "step": 67695
185490
+ },
185491
+ {
185492
+ "epoch": 541.0,
185493
+ "eval_loss": 0.44643455743789673,
185494
+ "eval_runtime": 38.3882,
185495
+ "eval_samples_per_second": 21.856,
185496
+ "eval_steps_per_second": 0.703,
185497
+ "eval_wer": 0.1991250455705432,
185498
+ "step": 67697
185499
+ },
185500
+ {
185501
+ "epoch": 541.02,
185502
+ "learning_rate": 8.931490384615386e-06,
185503
+ "loss": 0.3752,
185504
+ "step": 67700
185505
+ },
185506
+ {
185507
+ "epoch": 541.06,
185508
+ "learning_rate": 8.931410256410258e-06,
185509
+ "loss": 0.2947,
185510
+ "step": 67705
185511
+ },
185512
+ {
185513
+ "epoch": 541.1,
185514
+ "learning_rate": 8.931330128205129e-06,
185515
+ "loss": 0.3437,
185516
+ "step": 67710
185517
+ },
185518
+ {
185519
+ "epoch": 541.14,
185520
+ "learning_rate": 8.93125e-06,
185521
+ "loss": 0.3961,
185522
+ "step": 67715
185523
+ },
185524
+ {
185525
+ "epoch": 541.18,
185526
+ "learning_rate": 8.931169871794873e-06,
185527
+ "loss": 0.7735,
185528
+ "step": 67720
185529
+ },
185530
+ {
185531
+ "epoch": 541.22,
185532
+ "learning_rate": 8.931089743589745e-06,
185533
+ "loss": 1.0214,
185534
+ "step": 67725
185535
+ },
185536
+ {
185537
+ "epoch": 541.26,
185538
+ "learning_rate": 8.931009615384616e-06,
185539
+ "loss": 0.2891,
185540
+ "step": 67730
185541
+ },
185542
+ {
185543
+ "epoch": 541.3,
185544
+ "learning_rate": 8.930929487179489e-06,
185545
+ "loss": 0.2835,
185546
+ "step": 67735
185547
+ },
185548
+ {
185549
+ "epoch": 541.34,
185550
+ "learning_rate": 8.930849358974359e-06,
185551
+ "loss": 0.3882,
185552
+ "step": 67740
185553
+ },
185554
+ {
185555
+ "epoch": 541.38,
185556
+ "learning_rate": 8.930769230769232e-06,
185557
+ "loss": 0.769,
185558
+ "step": 67745
185559
+ },
185560
+ {
185561
+ "epoch": 541.42,
185562
+ "learning_rate": 8.930689102564103e-06,
185563
+ "loss": 0.7533,
185564
+ "step": 67750
185565
+ },
185566
+ {
185567
+ "epoch": 541.46,
185568
+ "learning_rate": 8.930608974358975e-06,
185569
+ "loss": 0.2833,
185570
+ "step": 67755
185571
+ },
185572
+ {
185573
+ "epoch": 541.5,
185574
+ "learning_rate": 8.930528846153846e-06,
185575
+ "loss": 0.2899,
185576
+ "step": 67760
185577
+ },
185578
+ {
185579
+ "epoch": 541.54,
185580
+ "learning_rate": 8.930448717948719e-06,
185581
+ "loss": 0.4753,
185582
+ "step": 67765
185583
+ },
185584
+ {
185585
+ "epoch": 541.58,
185586
+ "learning_rate": 8.93036858974359e-06,
185587
+ "loss": 0.7755,
185588
+ "step": 67770
185589
+ },
185590
+ {
185591
+ "epoch": 541.62,
185592
+ "learning_rate": 8.930288461538462e-06,
185593
+ "loss": 0.8648,
185594
+ "step": 67775
185595
+ },
185596
+ {
185597
+ "epoch": 541.66,
185598
+ "learning_rate": 8.930208333333335e-06,
185599
+ "loss": 0.2927,
185600
+ "step": 67780
185601
+ },
185602
+ {
185603
+ "epoch": 541.7,
185604
+ "learning_rate": 8.930128205128206e-06,
185605
+ "loss": 0.3099,
185606
+ "step": 67785
185607
+ },
185608
+ {
185609
+ "epoch": 541.74,
185610
+ "learning_rate": 8.930048076923077e-06,
185611
+ "loss": 0.3615,
185612
+ "step": 67790
185613
+ },
185614
+ {
185615
+ "epoch": 541.78,
185616
+ "learning_rate": 8.929967948717949e-06,
185617
+ "loss": 0.8285,
185618
+ "step": 67795
185619
+ },
185620
+ {
185621
+ "epoch": 541.82,
185622
+ "learning_rate": 8.929887820512822e-06,
185623
+ "loss": 0.8921,
185624
+ "step": 67800
185625
+ },
185626
+ {
185627
+ "epoch": 541.86,
185628
+ "learning_rate": 8.929807692307693e-06,
185629
+ "loss": 0.3251,
185630
+ "step": 67805
185631
+ },
185632
+ {
185633
+ "epoch": 541.9,
185634
+ "learning_rate": 8.929727564102565e-06,
185635
+ "loss": 0.338,
185636
+ "step": 67810
185637
+ },
185638
+ {
185639
+ "epoch": 541.94,
185640
+ "learning_rate": 8.929647435897436e-06,
185641
+ "loss": 0.408,
185642
+ "step": 67815
185643
+ },
185644
+ {
185645
+ "epoch": 541.98,
185646
+ "learning_rate": 8.929567307692309e-06,
185647
+ "loss": 0.7369,
185648
+ "step": 67820
185649
+ },
185650
+ {
185651
+ "epoch": 542.0,
185652
+ "eval_loss": 0.3617897629737854,
185653
+ "eval_runtime": 39.0831,
185654
+ "eval_samples_per_second": 21.441,
185655
+ "eval_steps_per_second": 0.691,
185656
+ "eval_wer": 0.19498813760379596,
185657
+ "step": 67822
185658
+ },
185659
+ {
185660
+ "epoch": 542.02,
185661
+ "learning_rate": 8.92948717948718e-06,
185662
+ "loss": 0.482,
185663
+ "step": 67825
185664
+ },
185665
+ {
185666
+ "epoch": 542.06,
185667
+ "learning_rate": 8.929407051282052e-06,
185668
+ "loss": 0.3239,
185669
+ "step": 67830
185670
+ },
185671
+ {
185672
+ "epoch": 542.1,
185673
+ "learning_rate": 8.929326923076925e-06,
185674
+ "loss": 0.2679,
185675
+ "step": 67835
185676
+ },
185677
+ {
185678
+ "epoch": 542.14,
185679
+ "learning_rate": 8.929246794871796e-06,
185680
+ "loss": 0.3645,
185681
+ "step": 67840
185682
+ },
185683
+ {
185684
+ "epoch": 542.18,
185685
+ "learning_rate": 8.929166666666667e-06,
185686
+ "loss": 0.8296,
185687
+ "step": 67845
185688
+ },
185689
+ {
185690
+ "epoch": 542.22,
185691
+ "learning_rate": 8.929086538461539e-06,
185692
+ "loss": 0.8821,
185693
+ "step": 67850
185694
+ },
185695
+ {
185696
+ "epoch": 542.26,
185697
+ "learning_rate": 8.929006410256412e-06,
185698
+ "loss": 0.3484,
185699
+ "step": 67855
185700
+ },
185701
+ {
185702
+ "epoch": 542.3,
185703
+ "learning_rate": 8.928926282051282e-06,
185704
+ "loss": 0.4008,
185705
+ "step": 67860
185706
+ },
185707
+ {
185708
+ "epoch": 542.34,
185709
+ "learning_rate": 8.928846153846155e-06,
185710
+ "loss": 0.3902,
185711
+ "step": 67865
185712
+ },
185713
+ {
185714
+ "epoch": 542.38,
185715
+ "learning_rate": 8.928766025641028e-06,
185716
+ "loss": 0.7359,
185717
+ "step": 67870
185718
+ },
185719
+ {
185720
+ "epoch": 542.42,
185721
+ "learning_rate": 8.928685897435897e-06,
185722
+ "loss": 1.0072,
185723
+ "step": 67875
185724
+ },
185725
+ {
185726
+ "epoch": 542.46,
185727
+ "learning_rate": 8.92860576923077e-06,
185728
+ "loss": 0.2912,
185729
+ "step": 67880
185730
+ },
185731
+ {
185732
+ "epoch": 542.5,
185733
+ "learning_rate": 8.928525641025642e-06,
185734
+ "loss": 0.3905,
185735
+ "step": 67885
185736
+ },
185737
+ {
185738
+ "epoch": 542.54,
185739
+ "learning_rate": 8.928445512820513e-06,
185740
+ "loss": 0.381,
185741
+ "step": 67890
185742
+ },
185743
+ {
185744
+ "epoch": 542.58,
185745
+ "learning_rate": 8.928365384615384e-06,
185746
+ "loss": 0.7648,
185747
+ "step": 67895
185748
+ },
185749
+ {
185750
+ "epoch": 542.62,
185751
+ "learning_rate": 8.928285256410257e-06,
185752
+ "loss": 0.8711,
185753
+ "step": 67900
185754
+ },
185755
+ {
185756
+ "epoch": 542.66,
185757
+ "learning_rate": 8.928205128205129e-06,
185758
+ "loss": 0.2766,
185759
+ "step": 67905
185760
+ },
185761
+ {
185762
+ "epoch": 542.7,
185763
+ "learning_rate": 8.928125e-06,
185764
+ "loss": 0.3258,
185765
+ "step": 67910
185766
+ },
185767
+ {
185768
+ "epoch": 542.74,
185769
+ "learning_rate": 8.928044871794872e-06,
185770
+ "loss": 0.3602,
185771
+ "step": 67915
185772
+ },
185773
+ {
185774
+ "epoch": 542.78,
185775
+ "learning_rate": 8.927964743589745e-06,
185776
+ "loss": 0.7964,
185777
+ "step": 67920
185778
+ },
185779
+ {
185780
+ "epoch": 542.82,
185781
+ "learning_rate": 8.927884615384616e-06,
185782
+ "loss": 1.025,
185783
+ "step": 67925
185784
+ },
185785
+ {
185786
+ "epoch": 542.86,
185787
+ "learning_rate": 8.927804487179487e-06,
185788
+ "loss": 0.2855,
185789
+ "step": 67930
185790
+ },
185791
+ {
185792
+ "epoch": 542.9,
185793
+ "learning_rate": 8.92772435897436e-06,
185794
+ "loss": 0.3043,
185795
+ "step": 67935
185796
+ },
185797
+ {
185798
+ "epoch": 542.94,
185799
+ "learning_rate": 8.927644230769232e-06,
185800
+ "loss": 0.3756,
185801
+ "step": 67940
185802
+ },
185803
+ {
185804
+ "epoch": 542.98,
185805
+ "learning_rate": 8.927564102564103e-06,
185806
+ "loss": 0.8851,
185807
+ "step": 67945
185808
+ },
185809
+ {
185810
+ "epoch": 543.0,
185811
+ "eval_loss": 0.3958238661289215,
185812
+ "eval_runtime": 39.4545,
185813
+ "eval_samples_per_second": 21.24,
185814
+ "eval_steps_per_second": 0.684,
185815
+ "eval_wer": 0.19044464075382803,
185816
+ "step": 67947
185817
+ },
185818
+ {
185819
+ "epoch": 547.02,
185820
+ "learning_rate": 8.927483974358974e-06,
185821
+ "loss": 0.3812,
185822
+ "step": 67950
185823
+ },
185824
+ {
185825
+ "epoch": 547.06,
185826
+ "learning_rate": 8.927403846153847e-06,
185827
+ "loss": 0.3034,
185828
+ "step": 67955
185829
+ },
185830
+ {
185831
+ "epoch": 547.1,
185832
+ "learning_rate": 8.927323717948719e-06,
185833
+ "loss": 0.3513,
185834
+ "step": 67960
185835
+ },
185836
+ {
185837
+ "epoch": 547.14,
185838
+ "learning_rate": 8.92724358974359e-06,
185839
+ "loss": 0.4248,
185840
+ "step": 67965
185841
+ },
185842
+ {
185843
+ "epoch": 547.18,
185844
+ "learning_rate": 8.927163461538463e-06,
185845
+ "loss": 0.579,
185846
+ "step": 67970
185847
+ },
185848
+ {
185849
+ "epoch": 547.22,
185850
+ "learning_rate": 8.927083333333335e-06,
185851
+ "loss": 0.9471,
185852
+ "step": 67975
185853
+ },
185854
+ {
185855
+ "epoch": 547.27,
185856
+ "learning_rate": 8.927003205128206e-06,
185857
+ "loss": 0.3662,
185858
+ "step": 67980
185859
+ },
185860
+ {
185861
+ "epoch": 547.31,
185862
+ "learning_rate": 8.926923076923077e-06,
185863
+ "loss": 0.3769,
185864
+ "step": 67985
185865
+ },
185866
+ {
185867
+ "epoch": 547.35,
185868
+ "learning_rate": 8.92684294871795e-06,
185869
+ "loss": 0.4178,
185870
+ "step": 67990
185871
+ },
185872
+ {
185873
+ "epoch": 547.39,
185874
+ "learning_rate": 8.926762820512822e-06,
185875
+ "loss": 0.703,
185876
+ "step": 67995
185877
+ },
185878
+ {
185879
+ "epoch": 547.43,
185880
+ "learning_rate": 8.926682692307693e-06,
185881
+ "loss": 1.0705,
185882
+ "step": 68000
185883
+ },
185884
+ {
185885
+ "epoch": 547.47,
185886
+ "learning_rate": 8.926602564102564e-06,
185887
+ "loss": 0.3506,
185888
+ "step": 68005
185889
+ },
185890
+ {
185891
+ "epoch": 547.51,
185892
+ "learning_rate": 8.926522435897438e-06,
185893
+ "loss": 0.3535,
185894
+ "step": 68010
185895
+ },
185896
+ {
185897
+ "epoch": 547.55,
185898
+ "learning_rate": 8.926442307692307e-06,
185899
+ "loss": 0.3949,
185900
+ "step": 68015
185901
+ },
185902
+ {
185903
+ "epoch": 547.59,
185904
+ "learning_rate": 8.92636217948718e-06,
185905
+ "loss": 0.7904,
185906
+ "step": 68020
185907
+ },
185908
+ {
185909
+ "epoch": 547.63,
185910
+ "learning_rate": 8.926282051282053e-06,
185911
+ "loss": 0.9079,
185912
+ "step": 68025
185913
+ },
185914
+ {
185915
+ "epoch": 547.67,
185916
+ "learning_rate": 8.926201923076923e-06,
185917
+ "loss": 0.4741,
185918
+ "step": 68030
185919
+ },
185920
+ {
185921
+ "epoch": 547.71,
185922
+ "learning_rate": 8.926121794871796e-06,
185923
+ "loss": 0.3193,
185924
+ "step": 68035
185925
+ },
185926
+ {
185927
+ "epoch": 547.75,
185928
+ "learning_rate": 8.926041666666667e-06,
185929
+ "loss": 0.3734,
185930
+ "step": 68040
185931
+ },
185932
+ {
185933
+ "epoch": 547.79,
185934
+ "learning_rate": 8.925961538461539e-06,
185935
+ "loss": 0.7489,
185936
+ "step": 68045
185937
+ },
185938
+ {
185939
+ "epoch": 547.83,
185940
+ "learning_rate": 8.92588141025641e-06,
185941
+ "loss": 1.1998,
185942
+ "step": 68050
185943
+ },
185944
+ {
185945
+ "epoch": 547.87,
185946
+ "learning_rate": 8.925801282051283e-06,
185947
+ "loss": 0.321,
185948
+ "step": 68055
185949
+ },
185950
+ {
185951
+ "epoch": 547.91,
185952
+ "learning_rate": 8.925721153846154e-06,
185953
+ "loss": 0.3205,
185954
+ "step": 68060
185955
+ },
185956
+ {
185957
+ "epoch": 547.95,
185958
+ "learning_rate": 8.925641025641026e-06,
185959
+ "loss": 0.4711,
185960
+ "step": 68065
185961
+ },
185962
+ {
185963
+ "epoch": 547.99,
185964
+ "learning_rate": 8.925560897435899e-06,
185965
+ "loss": 0.897,
185966
+ "step": 68070
185967
+ },
185968
+ {
185969
+ "epoch": 548.0,
185970
+ "eval_loss": 0.3655170798301697,
185971
+ "eval_runtime": 38.9264,
185972
+ "eval_samples_per_second": 21.553,
185973
+ "eval_steps_per_second": 0.694,
185974
+ "eval_wer": 0.19745080500894455,
185975
+ "step": 68071
185976
+ },
185977
+ {
185978
+ "epoch": 548.03,
185979
+ "learning_rate": 8.92548076923077e-06,
185980
+ "loss": 0.3636,
185981
+ "step": 68075
185982
+ },
185983
+ {
185984
+ "epoch": 548.07,
185985
+ "learning_rate": 8.925400641025642e-06,
185986
+ "loss": 0.3824,
185987
+ "step": 68080
185988
+ },
185989
+ {
185990
+ "epoch": 548.11,
185991
+ "learning_rate": 8.925320512820513e-06,
185992
+ "loss": 0.3019,
185993
+ "step": 68085
185994
+ },
185995
+ {
185996
+ "epoch": 548.15,
185997
+ "learning_rate": 8.925240384615386e-06,
185998
+ "loss": 0.3797,
185999
+ "step": 68090
186000
+ },
186001
+ {
186002
+ "epoch": 548.19,
186003
+ "learning_rate": 8.925160256410257e-06,
186004
+ "loss": 0.9859,
186005
+ "step": 68095
186006
+ },
186007
+ {
186008
+ "epoch": 548.23,
186009
+ "learning_rate": 8.925080128205129e-06,
186010
+ "loss": 0.7557,
186011
+ "step": 68100
186012
+ },
186013
+ {
186014
+ "epoch": 548.27,
186015
+ "learning_rate": 8.925e-06,
186016
+ "loss": 0.359,
186017
+ "step": 68105
186018
+ },
186019
+ {
186020
+ "epoch": 548.31,
186021
+ "learning_rate": 8.924919871794873e-06,
186022
+ "loss": 0.273,
186023
+ "step": 68110
186024
+ },
186025
+ {
186026
+ "epoch": 548.35,
186027
+ "learning_rate": 8.924839743589745e-06,
186028
+ "loss": 0.4286,
186029
+ "step": 68115
186030
+ },
186031
+ {
186032
+ "epoch": 548.39,
186033
+ "learning_rate": 8.924759615384616e-06,
186034
+ "loss": 0.8112,
186035
+ "step": 68120
186036
+ },
186037
+ {
186038
+ "epoch": 548.43,
186039
+ "learning_rate": 8.924679487179489e-06,
186040
+ "loss": 0.6704,
186041
+ "step": 68125
186042
+ },
186043
+ {
186044
+ "epoch": 548.47,
186045
+ "learning_rate": 8.92459935897436e-06,
186046
+ "loss": 0.3017,
186047
+ "step": 68130
186048
+ },
186049
+ {
186050
+ "epoch": 548.51,
186051
+ "learning_rate": 8.924519230769232e-06,
186052
+ "loss": 0.3403,
186053
+ "step": 68135
186054
+ },
186055
+ {
186056
+ "epoch": 548.55,
186057
+ "learning_rate": 8.924439102564103e-06,
186058
+ "loss": 0.4607,
186059
+ "step": 68140
186060
+ },
186061
+ {
186062
+ "epoch": 548.59,
186063
+ "learning_rate": 8.924358974358976e-06,
186064
+ "loss": 1.0333,
186065
+ "step": 68145
186066
+ },
186067
+ {
186068
+ "epoch": 548.63,
186069
+ "learning_rate": 8.924278846153846e-06,
186070
+ "loss": 0.8082,
186071
+ "step": 68150
186072
+ },
186073
+ {
186074
+ "epoch": 548.67,
186075
+ "learning_rate": 8.924198717948719e-06,
186076
+ "loss": 0.2737,
186077
+ "step": 68155
186078
+ },
186079
+ {
186080
+ "epoch": 548.71,
186081
+ "learning_rate": 8.92411858974359e-06,
186082
+ "loss": 0.283,
186083
+ "step": 68160
186084
+ },
186085
+ {
186086
+ "epoch": 548.76,
186087
+ "learning_rate": 8.924038461538461e-06,
186088
+ "loss": 0.5159,
186089
+ "step": 68165
186090
+ },
186091
+ {
186092
+ "epoch": 548.8,
186093
+ "learning_rate": 8.923958333333335e-06,
186094
+ "loss": 1.1066,
186095
+ "step": 68170
186096
+ },
186097
+ {
186098
+ "epoch": 548.84,
186099
+ "learning_rate": 8.923878205128206e-06,
186100
+ "loss": 0.7521,
186101
+ "step": 68175
186102
+ },
186103
+ {
186104
+ "epoch": 548.88,
186105
+ "learning_rate": 8.923798076923077e-06,
186106
+ "loss": 0.3451,
186107
+ "step": 68180
186108
+ },
186109
+ {
186110
+ "epoch": 548.92,
186111
+ "learning_rate": 8.923717948717949e-06,
186112
+ "loss": 0.3402,
186113
+ "step": 68185
186114
+ },
186115
+ {
186116
+ "epoch": 548.96,
186117
+ "learning_rate": 8.923637820512822e-06,
186118
+ "loss": 0.4594,
186119
+ "step": 68190
186120
+ },
186121
+ {
186122
+ "epoch": 549.0,
186123
+ "learning_rate": 8.923573717948719e-06,
186124
+ "loss": 1.3483,
186125
+ "step": 68195
186126
+ },
186127
+ {
186128
+ "epoch": 549.0,
186129
+ "eval_loss": 0.38058263063430786,
186130
+ "eval_runtime": 39.1105,
186131
+ "eval_samples_per_second": 21.452,
186132
+ "eval_steps_per_second": 0.69,
186133
+ "eval_wer": 0.1986700537955768,
186134
+ "step": 68195
186135
  }
186136
  ],
186137
+ "max_steps": 620000,
186138
  "num_train_epochs": 5000,
186139
+ "total_flos": 1.919002060140202e+20,
186140
  "trial_name": null,
186141
  "trial_params": null
186142
  }
model-bin/finetune/base/{checkpoint-67572 β†’ checkpoint-68195}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629837488.925602/events.out.tfevents.1629837488.c435e1c5ee04.920.221 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11a21fc8f147677a8204ea6fd68167f58938d658f3f6c9d85a072bf833a47153
3
+ size 4194
model-bin/finetune/base/log/1629838113.24123/events.out.tfevents.1629838113.c435e1c5ee04.920.223 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c87c23309dcbd740181b93f110e946100c7e5c1d6e79d8ed9ec5fdf67c127beb
3
+ size 4194
model-bin/finetune/base/log/1629838737.3102496/events.out.tfevents.1629838737.c435e1c5ee04.920.225 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:361864c826c35b2fc42f117bbb96788d0dc929c5248438f1c561cdc439022b27
3
+ size 4194
model-bin/finetune/base/log/1629839378.965553/events.out.tfevents.1629839378.c435e1c5ee04.920.227 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38efd8537eb21083b0de52e22cb6825508891dcfc3e1937d91b899af6bf030e4
3
+ size 4194
model-bin/finetune/base/log/1629840015.2190616/events.out.tfevents.1629840015.c435e1c5ee04.920.229 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5f973dd5e9cb6c871e2e118ff1419d9451dd9cc081e892d90a6195c7e3a0921
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629837488.c435e1c5ee04.920.220 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:618e987937fec56b32ef0ac288968c32fef6b3f06c07c5895e99df6432aa907f
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629838113.c435e1c5ee04.920.222 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2d243d2392ff1b3a2498ac040ea1fbde1159401225f1325087750bc2a099f77
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629838737.c435e1c5ee04.920.224 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff9e714dc0ae26ecf20d6efa0cf08919b8511cd95c487a9f8cb023380f51d40c
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629839378.c435e1c5ee04.920.226 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97109b02856def0906fd893d520582d1b4c154e4d3264c9f47d105068a5140d5
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629840015.c435e1c5ee04.920.228 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e2e1336f165fb7fa488885abb82e9ab78ead84f17ca64909350e74604045451
3
+ size 8622