Check commited on
Commit
612c9e4
Β·
1 Parent(s): b87ce57

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/rng_state.pth +2 -2
  6. model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/trainer_state.json +792 -3
  9. model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630212588.3931246/events.out.tfevents.1630212588.cc93b136ebf5.1086.39 +3 -0
  11. model-bin/finetune/base/log/1630213036.9606094/events.out.tfevents.1630213036.cc93b136ebf5.1086.41 +3 -0
  12. model-bin/finetune/base/log/1630213481.1394386/events.out.tfevents.1630213481.cc93b136ebf5.1086.43 +3 -0
  13. model-bin/finetune/base/log/1630213924.587934/events.out.tfevents.1630213924.cc93b136ebf5.1086.45 +3 -0
  14. model-bin/finetune/base/log/1630214361.273169/events.out.tfevents.1630214361.cc93b136ebf5.1086.47 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630212588.cc93b136ebf5.1086.38 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630213036.cc93b136ebf5.1086.40 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630213481.cc93b136ebf5.1086.42 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630213924.cc93b136ebf5.1086.44 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630214361.cc93b136ebf5.1086.46 +3 -0
model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7772525640a71b94879c52f0a8df49e5016ee6aa63e7d2a2617e69741a553f6a
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fabeaa58e79bc630b4651cb5de62d3becf8f3d011d2a72a8769bd12ad8ec2b2d
3
  size 722165393
model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f2c46c95f08599e97c7b8af2fcb752dc29b3c24da837b1bf54096ea462fa1b1d
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8ca5c08add489f83f8fbbe54f7967a3267d0c0a25ffcaf253ac3312f999a598
3
  size 377909911
model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ab856d5c1a50208d454e8f6bea1f46dc6469b0f803326d076dac2a148f7479be
3
- size 14439
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ead055a10b6a20dc53b007fa47951821da9955300fa2f176322d948a4b8ef3b
3
+ size 14503
model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:24c524215320dbd91b7fda70e307ff61c64b39d6458057cd5b75b3d4be7f9c2c
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:435c7949f76b27d1892fd86718a853997f446c51f752754e7e9bab37c34db055
3
  size 559
model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eeb9204627a0bfb8dbb675ece7e93c026210f3a03315b8bc633c50705545ebfd
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e654bbb2cd6ccd57472ee266ae110834ec44311eb0593d1db337b55a5107ef7
3
  size 623
model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1689111747851003,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
4
- "epoch": 1140.995983935743,
5
- "global_step": 141498,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -279393,11 +279393,800 @@
279393
  "eval_steps_per_second": 0.636,
279394
  "eval_wer": 0.17664390945023356,
279395
  "step": 141498
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
279396
  }
279397
  ],
279398
  "max_steps": 620000,
279399
  "num_train_epochs": 5000,
279400
- "total_flos": 3.9816733700847796e+20,
279401
  "trial_name": null,
279402
  "trial_params": null
279403
  }
 
1
  {
2
  "best_metric": 0.1689111747851003,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
4
+ "epoch": 1146.0,
5
+ "global_step": 142119,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
279393
  "eval_steps_per_second": 0.636,
279394
  "eval_wer": 0.17664390945023356,
279395
  "step": 141498
279396
+ },
279397
+ {
279398
+ "epoch": 1131.02,
279399
+ "learning_rate": 7.731098546042004e-06,
279400
+ "loss": 0.4259,
279401
+ "step": 141500
279402
+ },
279403
+ {
279404
+ "epoch": 1131.06,
279405
+ "learning_rate": 7.731017770597739e-06,
279406
+ "loss": 0.2644,
279407
+ "step": 141505
279408
+ },
279409
+ {
279410
+ "epoch": 1131.1,
279411
+ "learning_rate": 7.730936995153474e-06,
279412
+ "loss": 0.2419,
279413
+ "step": 141510
279414
+ },
279415
+ {
279416
+ "epoch": 1131.14,
279417
+ "learning_rate": 7.730856219709208e-06,
279418
+ "loss": 0.3695,
279419
+ "step": 141515
279420
+ },
279421
+ {
279422
+ "epoch": 1131.18,
279423
+ "learning_rate": 7.730775444264944e-06,
279424
+ "loss": 0.5446,
279425
+ "step": 141520
279426
+ },
279427
+ {
279428
+ "epoch": 1131.22,
279429
+ "learning_rate": 7.73069466882068e-06,
279430
+ "loss": 0.948,
279431
+ "step": 141525
279432
+ },
279433
+ {
279434
+ "epoch": 1131.25,
279435
+ "learning_rate": 7.730613893376414e-06,
279436
+ "loss": 0.277,
279437
+ "step": 141530
279438
+ },
279439
+ {
279440
+ "epoch": 1131.29,
279441
+ "learning_rate": 7.73053311793215e-06,
279442
+ "loss": 0.2816,
279443
+ "step": 141535
279444
+ },
279445
+ {
279446
+ "epoch": 1131.33,
279447
+ "learning_rate": 7.730452342487884e-06,
279448
+ "loss": 0.3233,
279449
+ "step": 141540
279450
+ },
279451
+ {
279452
+ "epoch": 1131.37,
279453
+ "learning_rate": 7.73037156704362e-06,
279454
+ "loss": 0.6062,
279455
+ "step": 141545
279456
+ },
279457
+ {
279458
+ "epoch": 1131.41,
279459
+ "learning_rate": 7.730290791599354e-06,
279460
+ "loss": 1.098,
279461
+ "step": 141550
279462
+ },
279463
+ {
279464
+ "epoch": 1131.45,
279465
+ "learning_rate": 7.73021001615509e-06,
279466
+ "loss": 0.2661,
279467
+ "step": 141555
279468
+ },
279469
+ {
279470
+ "epoch": 1131.49,
279471
+ "learning_rate": 7.730129240710824e-06,
279472
+ "loss": 0.3398,
279473
+ "step": 141560
279474
+ },
279475
+ {
279476
+ "epoch": 1131.53,
279477
+ "learning_rate": 7.73004846526656e-06,
279478
+ "loss": 0.329,
279479
+ "step": 141565
279480
+ },
279481
+ {
279482
+ "epoch": 1131.57,
279483
+ "learning_rate": 7.729967689822294e-06,
279484
+ "loss": 0.6156,
279485
+ "step": 141570
279486
+ },
279487
+ {
279488
+ "epoch": 1131.61,
279489
+ "learning_rate": 7.72988691437803e-06,
279490
+ "loss": 1.1026,
279491
+ "step": 141575
279492
+ },
279493
+ {
279494
+ "epoch": 1131.65,
279495
+ "learning_rate": 7.729806138933766e-06,
279496
+ "loss": 0.2513,
279497
+ "step": 141580
279498
+ },
279499
+ {
279500
+ "epoch": 1131.69,
279501
+ "learning_rate": 7.7297253634895e-06,
279502
+ "loss": 0.284,
279503
+ "step": 141585
279504
+ },
279505
+ {
279506
+ "epoch": 1131.73,
279507
+ "learning_rate": 7.729644588045236e-06,
279508
+ "loss": 0.3635,
279509
+ "step": 141590
279510
+ },
279511
+ {
279512
+ "epoch": 1131.77,
279513
+ "learning_rate": 7.72956381260097e-06,
279514
+ "loss": 0.5597,
279515
+ "step": 141595
279516
+ },
279517
+ {
279518
+ "epoch": 1131.81,
279519
+ "learning_rate": 7.729483037156706e-06,
279520
+ "loss": 0.9859,
279521
+ "step": 141600
279522
+ },
279523
+ {
279524
+ "epoch": 1131.85,
279525
+ "learning_rate": 7.72940226171244e-06,
279526
+ "loss": 0.247,
279527
+ "step": 141605
279528
+ },
279529
+ {
279530
+ "epoch": 1131.89,
279531
+ "learning_rate": 7.729321486268176e-06,
279532
+ "loss": 0.2809,
279533
+ "step": 141610
279534
+ },
279535
+ {
279536
+ "epoch": 1131.93,
279537
+ "learning_rate": 7.72924071082391e-06,
279538
+ "loss": 0.3225,
279539
+ "step": 141615
279540
+ },
279541
+ {
279542
+ "epoch": 1131.97,
279543
+ "learning_rate": 7.729159935379646e-06,
279544
+ "loss": 0.6184,
279545
+ "step": 141620
279546
+ },
279547
+ {
279548
+ "epoch": 1132.0,
279549
+ "eval_loss": 0.39268848299980164,
279550
+ "eval_runtime": 42.1482,
279551
+ "eval_samples_per_second": 19.858,
279552
+ "eval_steps_per_second": 0.641,
279553
+ "eval_wer": 0.18137723230690087,
279554
+ "step": 141623
279555
+ },
279556
+ {
279557
+ "epoch": 1142.02,
279558
+ "learning_rate": 7.72907915993538e-06,
279559
+ "loss": 0.3626,
279560
+ "step": 141625
279561
+ },
279562
+ {
279563
+ "epoch": 1142.06,
279564
+ "learning_rate": 7.728998384491116e-06,
279565
+ "loss": 0.3136,
279566
+ "step": 141630
279567
+ },
279568
+ {
279569
+ "epoch": 1142.1,
279570
+ "learning_rate": 7.72891760904685e-06,
279571
+ "loss": 0.2508,
279572
+ "step": 141635
279573
+ },
279574
+ {
279575
+ "epoch": 1142.14,
279576
+ "learning_rate": 7.728836833602586e-06,
279577
+ "loss": 0.3972,
279578
+ "step": 141640
279579
+ },
279580
+ {
279581
+ "epoch": 1142.18,
279582
+ "learning_rate": 7.728756058158322e-06,
279583
+ "loss": 0.6265,
279584
+ "step": 141645
279585
+ },
279586
+ {
279587
+ "epoch": 1142.22,
279588
+ "learning_rate": 7.728675282714056e-06,
279589
+ "loss": 0.9015,
279590
+ "step": 141650
279591
+ },
279592
+ {
279593
+ "epoch": 1142.26,
279594
+ "learning_rate": 7.728594507269792e-06,
279595
+ "loss": 0.2899,
279596
+ "step": 141655
279597
+ },
279598
+ {
279599
+ "epoch": 1142.3,
279600
+ "learning_rate": 7.728513731825526e-06,
279601
+ "loss": 0.3102,
279602
+ "step": 141660
279603
+ },
279604
+ {
279605
+ "epoch": 1142.34,
279606
+ "learning_rate": 7.728432956381261e-06,
279607
+ "loss": 0.3435,
279608
+ "step": 141665
279609
+ },
279610
+ {
279611
+ "epoch": 1142.38,
279612
+ "learning_rate": 7.728352180936996e-06,
279613
+ "loss": 0.6388,
279614
+ "step": 141670
279615
+ },
279616
+ {
279617
+ "epoch": 1142.42,
279618
+ "learning_rate": 7.728271405492731e-06,
279619
+ "loss": 0.9743,
279620
+ "step": 141675
279621
+ },
279622
+ {
279623
+ "epoch": 1142.46,
279624
+ "learning_rate": 7.728190630048466e-06,
279625
+ "loss": 0.2384,
279626
+ "step": 141680
279627
+ },
279628
+ {
279629
+ "epoch": 1142.5,
279630
+ "learning_rate": 7.728109854604201e-06,
279631
+ "loss": 0.4136,
279632
+ "step": 141685
279633
+ },
279634
+ {
279635
+ "epoch": 1142.54,
279636
+ "learning_rate": 7.728029079159936e-06,
279637
+ "loss": 0.3579,
279638
+ "step": 141690
279639
+ },
279640
+ {
279641
+ "epoch": 1142.58,
279642
+ "learning_rate": 7.727948303715671e-06,
279643
+ "loss": 0.5829,
279644
+ "step": 141695
279645
+ },
279646
+ {
279647
+ "epoch": 1142.62,
279648
+ "learning_rate": 7.727867528271407e-06,
279649
+ "loss": 0.9261,
279650
+ "step": 141700
279651
+ },
279652
+ {
279653
+ "epoch": 1142.66,
279654
+ "learning_rate": 7.727786752827141e-06,
279655
+ "loss": 0.2647,
279656
+ "step": 141705
279657
+ },
279658
+ {
279659
+ "epoch": 1142.7,
279660
+ "learning_rate": 7.727705977382877e-06,
279661
+ "loss": 0.2991,
279662
+ "step": 141710
279663
+ },
279664
+ {
279665
+ "epoch": 1142.74,
279666
+ "learning_rate": 7.727625201938611e-06,
279667
+ "loss": 0.3412,
279668
+ "step": 141715
279669
+ },
279670
+ {
279671
+ "epoch": 1142.78,
279672
+ "learning_rate": 7.727544426494347e-06,
279673
+ "loss": 0.5808,
279674
+ "step": 141720
279675
+ },
279676
+ {
279677
+ "epoch": 1142.82,
279678
+ "learning_rate": 7.727463651050081e-06,
279679
+ "loss": 0.9713,
279680
+ "step": 141725
279681
+ },
279682
+ {
279683
+ "epoch": 1142.86,
279684
+ "learning_rate": 7.727382875605817e-06,
279685
+ "loss": 0.2904,
279686
+ "step": 141730
279687
+ },
279688
+ {
279689
+ "epoch": 1142.9,
279690
+ "learning_rate": 7.727302100161551e-06,
279691
+ "loss": 0.2996,
279692
+ "step": 141735
279693
+ },
279694
+ {
279695
+ "epoch": 1142.94,
279696
+ "learning_rate": 7.727221324717287e-06,
279697
+ "loss": 0.3428,
279698
+ "step": 141740
279699
+ },
279700
+ {
279701
+ "epoch": 1142.98,
279702
+ "learning_rate": 7.727140549273021e-06,
279703
+ "loss": 0.6433,
279704
+ "step": 141745
279705
+ },
279706
+ {
279707
+ "epoch": 1143.0,
279708
+ "eval_loss": 0.3831433951854706,
279709
+ "eval_runtime": 41.1179,
279710
+ "eval_samples_per_second": 20.356,
279711
+ "eval_steps_per_second": 0.657,
279712
+ "eval_wer": 0.19271791099668997,
279713
+ "step": 141747
279714
+ },
279715
+ {
279716
+ "epoch": 1143.02,
279717
+ "learning_rate": 7.727059773828757e-06,
279718
+ "loss": 0.3706,
279719
+ "step": 141750
279720
+ },
279721
+ {
279722
+ "epoch": 1143.06,
279723
+ "learning_rate": 7.726978998384493e-06,
279724
+ "loss": 0.2449,
279725
+ "step": 141755
279726
+ },
279727
+ {
279728
+ "epoch": 1143.1,
279729
+ "learning_rate": 7.726898222940227e-06,
279730
+ "loss": 0.2845,
279731
+ "step": 141760
279732
+ },
279733
+ {
279734
+ "epoch": 1143.14,
279735
+ "learning_rate": 7.726817447495963e-06,
279736
+ "loss": 0.3353,
279737
+ "step": 141765
279738
+ },
279739
+ {
279740
+ "epoch": 1143.18,
279741
+ "learning_rate": 7.726736672051697e-06,
279742
+ "loss": 0.7831,
279743
+ "step": 141770
279744
+ },
279745
+ {
279746
+ "epoch": 1143.22,
279747
+ "learning_rate": 7.726655896607433e-06,
279748
+ "loss": 0.9689,
279749
+ "step": 141775
279750
+ },
279751
+ {
279752
+ "epoch": 1143.27,
279753
+ "learning_rate": 7.726575121163167e-06,
279754
+ "loss": 0.3101,
279755
+ "step": 141780
279756
+ },
279757
+ {
279758
+ "epoch": 1143.31,
279759
+ "learning_rate": 7.726494345718903e-06,
279760
+ "loss": 0.2913,
279761
+ "step": 141785
279762
+ },
279763
+ {
279764
+ "epoch": 1143.35,
279765
+ "learning_rate": 7.726413570274637e-06,
279766
+ "loss": 0.3821,
279767
+ "step": 141790
279768
+ },
279769
+ {
279770
+ "epoch": 1143.39,
279771
+ "learning_rate": 7.726332794830373e-06,
279772
+ "loss": 0.7098,
279773
+ "step": 141795
279774
+ },
279775
+ {
279776
+ "epoch": 1143.43,
279777
+ "learning_rate": 7.726252019386107e-06,
279778
+ "loss": 0.8485,
279779
+ "step": 141800
279780
+ },
279781
+ {
279782
+ "epoch": 1143.47,
279783
+ "learning_rate": 7.726171243941843e-06,
279784
+ "loss": 0.2593,
279785
+ "step": 141805
279786
+ },
279787
+ {
279788
+ "epoch": 1143.51,
279789
+ "learning_rate": 7.726090468497577e-06,
279790
+ "loss": 0.2944,
279791
+ "step": 141810
279792
+ },
279793
+ {
279794
+ "epoch": 1143.55,
279795
+ "learning_rate": 7.726009693053313e-06,
279796
+ "loss": 0.2959,
279797
+ "step": 141815
279798
+ },
279799
+ {
279800
+ "epoch": 1143.59,
279801
+ "learning_rate": 7.725928917609049e-06,
279802
+ "loss": 0.7449,
279803
+ "step": 141820
279804
+ },
279805
+ {
279806
+ "epoch": 1143.63,
279807
+ "learning_rate": 7.725848142164783e-06,
279808
+ "loss": 0.7156,
279809
+ "step": 141825
279810
+ },
279811
+ {
279812
+ "epoch": 1143.67,
279813
+ "learning_rate": 7.725767366720519e-06,
279814
+ "loss": 0.3058,
279815
+ "step": 141830
279816
+ },
279817
+ {
279818
+ "epoch": 1143.71,
279819
+ "learning_rate": 7.725686591276253e-06,
279820
+ "loss": 0.2875,
279821
+ "step": 141835
279822
+ },
279823
+ {
279824
+ "epoch": 1143.75,
279825
+ "learning_rate": 7.725605815831989e-06,
279826
+ "loss": 0.2819,
279827
+ "step": 141840
279828
+ },
279829
+ {
279830
+ "epoch": 1143.79,
279831
+ "learning_rate": 7.725525040387723e-06,
279832
+ "loss": 0.6082,
279833
+ "step": 141845
279834
+ },
279835
+ {
279836
+ "epoch": 1143.83,
279837
+ "learning_rate": 7.725444264943459e-06,
279838
+ "loss": 0.7674,
279839
+ "step": 141850
279840
+ },
279841
+ {
279842
+ "epoch": 1143.87,
279843
+ "learning_rate": 7.725363489499193e-06,
279844
+ "loss": 0.2629,
279845
+ "step": 141855
279846
+ },
279847
+ {
279848
+ "epoch": 1143.91,
279849
+ "learning_rate": 7.725282714054929e-06,
279850
+ "loss": 0.2882,
279851
+ "step": 141860
279852
+ },
279853
+ {
279854
+ "epoch": 1143.95,
279855
+ "learning_rate": 7.725201938610663e-06,
279856
+ "loss": 0.3376,
279857
+ "step": 141865
279858
+ },
279859
+ {
279860
+ "epoch": 1143.99,
279861
+ "learning_rate": 7.725121163166399e-06,
279862
+ "loss": 0.7656,
279863
+ "step": 141870
279864
+ },
279865
+ {
279866
+ "epoch": 1144.0,
279867
+ "eval_loss": 0.32772520184516907,
279868
+ "eval_runtime": 41.4072,
279869
+ "eval_samples_per_second": 20.214,
279870
+ "eval_steps_per_second": 0.652,
279871
+ "eval_wer": 0.18056568012829857,
279872
+ "step": 141871
279873
+ },
279874
+ {
279875
+ "epoch": 1144.03,
279876
+ "learning_rate": 7.725040387722134e-06,
279877
+ "loss": 0.2765,
279878
+ "step": 141875
279879
+ },
279880
+ {
279881
+ "epoch": 1144.07,
279882
+ "learning_rate": 7.724959612277868e-06,
279883
+ "loss": 0.2441,
279884
+ "step": 141880
279885
+ },
279886
+ {
279887
+ "epoch": 1144.11,
279888
+ "learning_rate": 7.724878836833604e-06,
279889
+ "loss": 0.311,
279890
+ "step": 141885
279891
+ },
279892
+ {
279893
+ "epoch": 1144.15,
279894
+ "learning_rate": 7.724798061389338e-06,
279895
+ "loss": 0.3957,
279896
+ "step": 141890
279897
+ },
279898
+ {
279899
+ "epoch": 1144.19,
279900
+ "learning_rate": 7.724717285945074e-06,
279901
+ "loss": 0.8882,
279902
+ "step": 141895
279903
+ },
279904
+ {
279905
+ "epoch": 1144.23,
279906
+ "learning_rate": 7.724636510500808e-06,
279907
+ "loss": 0.5858,
279908
+ "step": 141900
279909
+ },
279910
+ {
279911
+ "epoch": 1144.27,
279912
+ "learning_rate": 7.724555735056544e-06,
279913
+ "loss": 0.2691,
279914
+ "step": 141905
279915
+ },
279916
+ {
279917
+ "epoch": 1144.31,
279918
+ "learning_rate": 7.724474959612278e-06,
279919
+ "loss": 0.2628,
279920
+ "step": 141910
279921
+ },
279922
+ {
279923
+ "epoch": 1144.35,
279924
+ "learning_rate": 7.724394184168014e-06,
279925
+ "loss": 0.3303,
279926
+ "step": 141915
279927
+ },
279928
+ {
279929
+ "epoch": 1144.39,
279930
+ "learning_rate": 7.724313408723748e-06,
279931
+ "loss": 0.979,
279932
+ "step": 141920
279933
+ },
279934
+ {
279935
+ "epoch": 1144.43,
279936
+ "learning_rate": 7.724232633279484e-06,
279937
+ "loss": 1.0384,
279938
+ "step": 141925
279939
+ },
279940
+ {
279941
+ "epoch": 1144.47,
279942
+ "learning_rate": 7.72415185783522e-06,
279943
+ "loss": 0.2295,
279944
+ "step": 141930
279945
+ },
279946
+ {
279947
+ "epoch": 1144.51,
279948
+ "learning_rate": 7.724071082390954e-06,
279949
+ "loss": 0.3122,
279950
+ "step": 141935
279951
+ },
279952
+ {
279953
+ "epoch": 1144.55,
279954
+ "learning_rate": 7.72399030694669e-06,
279955
+ "loss": 0.3926,
279956
+ "step": 141940
279957
+ },
279958
+ {
279959
+ "epoch": 1144.59,
279960
+ "learning_rate": 7.723909531502424e-06,
279961
+ "loss": 0.7148,
279962
+ "step": 141945
279963
+ },
279964
+ {
279965
+ "epoch": 1144.63,
279966
+ "learning_rate": 7.72382875605816e-06,
279967
+ "loss": 0.591,
279968
+ "step": 141950
279969
+ },
279970
+ {
279971
+ "epoch": 1144.67,
279972
+ "learning_rate": 7.723747980613894e-06,
279973
+ "loss": 0.2606,
279974
+ "step": 141955
279975
+ },
279976
+ {
279977
+ "epoch": 1144.71,
279978
+ "learning_rate": 7.72366720516963e-06,
279979
+ "loss": 0.272,
279980
+ "step": 141960
279981
+ },
279982
+ {
279983
+ "epoch": 1144.76,
279984
+ "learning_rate": 7.723586429725364e-06,
279985
+ "loss": 0.402,
279986
+ "step": 141965
279987
+ },
279988
+ {
279989
+ "epoch": 1144.8,
279990
+ "learning_rate": 7.7235056542811e-06,
279991
+ "loss": 0.7544,
279992
+ "step": 141970
279993
+ },
279994
+ {
279995
+ "epoch": 1144.84,
279996
+ "learning_rate": 7.723424878836834e-06,
279997
+ "loss": 0.6907,
279998
+ "step": 141975
279999
+ },
280000
+ {
280001
+ "epoch": 1144.88,
280002
+ "learning_rate": 7.72334410339257e-06,
280003
+ "loss": 0.3154,
280004
+ "step": 141980
280005
+ },
280006
+ {
280007
+ "epoch": 1144.92,
280008
+ "learning_rate": 7.723263327948304e-06,
280009
+ "loss": 0.314,
280010
+ "step": 141985
280011
+ },
280012
+ {
280013
+ "epoch": 1144.96,
280014
+ "learning_rate": 7.72318255250404e-06,
280015
+ "loss": 0.4343,
280016
+ "step": 141990
280017
+ },
280018
+ {
280019
+ "epoch": 1145.0,
280020
+ "learning_rate": 7.723101777059776e-06,
280021
+ "loss": 1.0367,
280022
+ "step": 141995
280023
+ },
280024
+ {
280025
+ "epoch": 1145.0,
280026
+ "eval_loss": 0.36632317304611206,
280027
+ "eval_runtime": 41.4842,
280028
+ "eval_samples_per_second": 20.2,
280029
+ "eval_steps_per_second": 0.651,
280030
+ "eval_wer": 0.17683603283780785,
280031
+ "step": 141995
280032
+ },
280033
+ {
280034
+ "epoch": 1145.04,
280035
+ "learning_rate": 7.72302100161551e-06,
280036
+ "loss": 0.3223,
280037
+ "step": 142000
280038
+ },
280039
+ {
280040
+ "epoch": 1145.08,
280041
+ "learning_rate": 7.722940226171246e-06,
280042
+ "loss": 0.3433,
280043
+ "step": 142005
280044
+ },
280045
+ {
280046
+ "epoch": 1145.12,
280047
+ "learning_rate": 7.72285945072698e-06,
280048
+ "loss": 0.3129,
280049
+ "step": 142010
280050
+ },
280051
+ {
280052
+ "epoch": 1145.16,
280053
+ "learning_rate": 7.722778675282716e-06,
280054
+ "loss": 0.3968,
280055
+ "step": 142015
280056
+ },
280057
+ {
280058
+ "epoch": 1145.2,
280059
+ "learning_rate": 7.72269789983845e-06,
280060
+ "loss": 1.1603,
280061
+ "step": 142020
280062
+ },
280063
+ {
280064
+ "epoch": 1145.24,
280065
+ "learning_rate": 7.722617124394186e-06,
280066
+ "loss": 0.2974,
280067
+ "step": 142025
280068
+ },
280069
+ {
280070
+ "epoch": 1145.28,
280071
+ "learning_rate": 7.72253634894992e-06,
280072
+ "loss": 0.3043,
280073
+ "step": 142030
280074
+ },
280075
+ {
280076
+ "epoch": 1145.32,
280077
+ "learning_rate": 7.722455573505656e-06,
280078
+ "loss": 0.2808,
280079
+ "step": 142035
280080
+ },
280081
+ {
280082
+ "epoch": 1145.36,
280083
+ "learning_rate": 7.72237479806139e-06,
280084
+ "loss": 0.418,
280085
+ "step": 142040
280086
+ },
280087
+ {
280088
+ "epoch": 1145.4,
280089
+ "learning_rate": 7.722294022617126e-06,
280090
+ "loss": 1.1368,
280091
+ "step": 142045
280092
+ },
280093
+ {
280094
+ "epoch": 1145.44,
280095
+ "learning_rate": 7.72221324717286e-06,
280096
+ "loss": 0.305,
280097
+ "step": 142050
280098
+ },
280099
+ {
280100
+ "epoch": 1145.48,
280101
+ "learning_rate": 7.722132471728596e-06,
280102
+ "loss": 0.2486,
280103
+ "step": 142055
280104
+ },
280105
+ {
280106
+ "epoch": 1145.52,
280107
+ "learning_rate": 7.722051696284331e-06,
280108
+ "loss": 0.276,
280109
+ "step": 142060
280110
+ },
280111
+ {
280112
+ "epoch": 1145.56,
280113
+ "learning_rate": 7.721970920840066e-06,
280114
+ "loss": 0.3786,
280115
+ "step": 142065
280116
+ },
280117
+ {
280118
+ "epoch": 1145.6,
280119
+ "learning_rate": 7.721890145395801e-06,
280120
+ "loss": 1.0858,
280121
+ "step": 142070
280122
+ },
280123
+ {
280124
+ "epoch": 1145.65,
280125
+ "learning_rate": 7.721809369951536e-06,
280126
+ "loss": 0.3283,
280127
+ "step": 142075
280128
+ },
280129
+ {
280130
+ "epoch": 1145.69,
280131
+ "learning_rate": 7.721728594507271e-06,
280132
+ "loss": 0.269,
280133
+ "step": 142080
280134
+ },
280135
+ {
280136
+ "epoch": 1145.73,
280137
+ "learning_rate": 7.721647819063006e-06,
280138
+ "loss": 0.3025,
280139
+ "step": 142085
280140
+ },
280141
+ {
280142
+ "epoch": 1145.77,
280143
+ "learning_rate": 7.721567043618741e-06,
280144
+ "loss": 0.459,
280145
+ "step": 142090
280146
+ },
280147
+ {
280148
+ "epoch": 1145.81,
280149
+ "learning_rate": 7.721486268174475e-06,
280150
+ "loss": 1.3132,
280151
+ "step": 142095
280152
+ },
280153
+ {
280154
+ "epoch": 1145.85,
280155
+ "learning_rate": 7.721405492730211e-06,
280156
+ "loss": 0.5061,
280157
+ "step": 142100
280158
+ },
280159
+ {
280160
+ "epoch": 1145.89,
280161
+ "learning_rate": 7.721324717285945e-06,
280162
+ "loss": 0.3975,
280163
+ "step": 142105
280164
+ },
280165
+ {
280166
+ "epoch": 1145.93,
280167
+ "learning_rate": 7.721243941841681e-06,
280168
+ "loss": 0.2731,
280169
+ "step": 142110
280170
+ },
280171
+ {
280172
+ "epoch": 1145.97,
280173
+ "learning_rate": 7.721163166397415e-06,
280174
+ "loss": 0.4579,
280175
+ "step": 142115
280176
+ },
280177
+ {
280178
+ "epoch": 1146.0,
280179
+ "eval_loss": 0.3085727393627167,
280180
+ "eval_runtime": 41.3912,
280181
+ "eval_samples_per_second": 20.246,
280182
+ "eval_steps_per_second": 0.652,
280183
+ "eval_wer": 0.18064516129032257,
280184
+ "step": 142119
280185
  }
280186
  ],
280187
  "max_steps": 620000,
280188
  "num_train_epochs": 5000,
280189
+ "total_flos": 3.999266636171066e+20,
280190
  "trial_name": null,
280191
  "trial_params": null
280192
  }
model-bin/finetune/base/{checkpoint-141498 β†’ checkpoint-142119}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630212588.3931246/events.out.tfevents.1630212588.cc93b136ebf5.1086.39 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c5e861e922cdc7f53de2c00486af0f74996318b37fc1d9a494d8bf793692a91
3
+ size 4194
model-bin/finetune/base/log/1630213036.9606094/events.out.tfevents.1630213036.cc93b136ebf5.1086.41 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:021df7086f0e001322c25f1ee0bfc9a11e6cc059a7a4bb73d75890a874f6cdff
3
+ size 4194
model-bin/finetune/base/log/1630213481.1394386/events.out.tfevents.1630213481.cc93b136ebf5.1086.43 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d78ef73847fcf4c078b856415372a2e0d6576f9df97aa1685ddaf3c9f50c4817
3
+ size 4194
model-bin/finetune/base/log/1630213924.587934/events.out.tfevents.1630213924.cc93b136ebf5.1086.45 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:553de5a9a6a00d490406270cb5e07813c91d013e4de78cf533963a3c61620cd3
3
+ size 4194
model-bin/finetune/base/log/1630214361.273169/events.out.tfevents.1630214361.cc93b136ebf5.1086.47 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e16b44e2fa2d95584801fdcfbc14d177de5fd3b3d3a763c5ef239f82dc43921
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630212588.cc93b136ebf5.1086.38 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:503b8170234cacdf6c83700bcb0029868147641f33e3037d43424ab2d405a5ec
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630213036.cc93b136ebf5.1086.40 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55d2b14d9c8d7738fa05f69362035ff2b99c1dd41be2078b1cd300bd8297638f
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630213481.cc93b136ebf5.1086.42 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff7c368ef5da476643bc4e149343cae1f0d21f86a868ba6b7cf452073289150f
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630213924.cc93b136ebf5.1086.44 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:128d43720ab01b9d11b9073331af63a176f4d66056fbe6d2548778e1103e271b
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630214361.cc93b136ebf5.1086.46 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:812e218f1a368ab0f3db92f2f3fcae7cef4c84f56b8baab5cb1d5f6eb0ed0510
3
+ size 8462