Check commited on
Commit
fe55e3e
Β·
1 Parent(s): ad3332a

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/trainer_state.json +792 -3
  9. model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629892093.865816/events.out.tfevents.1629892093.7e498afd5545.7645.15 +3 -0
  11. model-bin/finetune/base/log/1629892567.0899503/events.out.tfevents.1629892567.7e498afd5545.7645.17 +3 -0
  12. model-bin/finetune/base/log/1629893045.2545307/events.out.tfevents.1629893045.7e498afd5545.7645.19 +3 -0
  13. model-bin/finetune/base/log/1629893510.1288207/events.out.tfevents.1629893510.7e498afd5545.7645.21 +3 -0
  14. model-bin/finetune/base/log/1629893973.3909373/events.out.tfevents.1629893973.7e498afd5545.7645.23 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629892093.7e498afd5545.7645.14 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629892566.7e498afd5545.7645.16 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629893045.7e498afd5545.7645.18 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629893510.7e498afd5545.7645.20 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629893973.7e498afd5545.7645.22 +3 -0
model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e71760057eeaf95512c73bda8c72702fdcd07ff2723d1f896c7d0a878caeb038
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f7011fd213a00609d063c7de52646f933c02885b6729b9df100cc88653c0770
3
  size 722165393
model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:609b28af0de8e6654be054f551fe06d6ac03c459c4482a5a4552b7a7f7c75d86
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e61021c5f6956a83f28794793bf05ffaa3847dfe43fd62b5cb1b5e70078ceae4
3
  size 377909911
model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b164785d44f4581870a7236a75a3744bb07d4b54e54ff053a662784001b7aba
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b43bab0cd2e069d88f2cb00182ba42d4d4428e663533bd0454a533cf361690e2
3
  size 14503
model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:86fe6f6b7182e301e95d9029914dff71904a56f874a0ac7549c42dca4b397c8c
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55268cedf33cbd9726c9ca2afafd5dc42f5d2a582af18496d5e36eaf0ef58f59
3
  size 559
model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5591339c9f642aace34467285535f0b0980daab6c067438026e89ebdc4a46cf7
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f21781a404675bc1af4e71bb00027419ad339e8a09ed59aa2536f2b39085f87
3
  size 623
model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.18412114350410416,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
4
- "epoch": 621.995983935743,
5
- "global_step": 77155,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -197532,11 +197532,800 @@
197532
  "eval_steps_per_second": 0.624,
197533
  "eval_wer": 0.18589790672052883,
197534
  "step": 77155
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
197535
  }
197536
  ],
197537
  "max_steps": 620000,
197538
  "num_train_epochs": 5000,
197539
- "total_flos": 2.1712747856892222e+20,
197540
  "trial_name": null,
197541
  "trial_params": null
197542
  }
 
1
  {
2
  "best_metric": 0.18412114350410416,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
4
+ "epoch": 626.995983935743,
5
+ "global_step": 77779,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
197532
  "eval_steps_per_second": 0.624,
197533
  "eval_wer": 0.18589790672052883,
197534
  "step": 77155
197535
+ },
197536
+ {
197537
+ "epoch": 617.04,
197538
+ "learning_rate": 8.770096930533118e-06,
197539
+ "loss": 0.4089,
197540
+ "step": 77160
197541
+ },
197542
+ {
197543
+ "epoch": 617.08,
197544
+ "learning_rate": 8.770016155088854e-06,
197545
+ "loss": 0.2633,
197546
+ "step": 77165
197547
+ },
197548
+ {
197549
+ "epoch": 617.12,
197550
+ "learning_rate": 8.769935379644588e-06,
197551
+ "loss": 0.362,
197552
+ "step": 77170
197553
+ },
197554
+ {
197555
+ "epoch": 617.16,
197556
+ "learning_rate": 8.769854604200324e-06,
197557
+ "loss": 0.4616,
197558
+ "step": 77175
197559
+ },
197560
+ {
197561
+ "epoch": 617.2,
197562
+ "learning_rate": 8.769773828756058e-06,
197563
+ "loss": 1.1666,
197564
+ "step": 77180
197565
+ },
197566
+ {
197567
+ "epoch": 617.24,
197568
+ "learning_rate": 8.769693053311794e-06,
197569
+ "loss": 0.3367,
197570
+ "step": 77185
197571
+ },
197572
+ {
197573
+ "epoch": 617.28,
197574
+ "learning_rate": 8.769612277867528e-06,
197575
+ "loss": 0.328,
197576
+ "step": 77190
197577
+ },
197578
+ {
197579
+ "epoch": 617.32,
197580
+ "learning_rate": 8.769531502423264e-06,
197581
+ "loss": 0.3628,
197582
+ "step": 77195
197583
+ },
197584
+ {
197585
+ "epoch": 617.36,
197586
+ "learning_rate": 8.769450726978998e-06,
197587
+ "loss": 0.514,
197588
+ "step": 77200
197589
+ },
197590
+ {
197591
+ "epoch": 617.4,
197592
+ "learning_rate": 8.769369951534734e-06,
197593
+ "loss": 1.3437,
197594
+ "step": 77205
197595
+ },
197596
+ {
197597
+ "epoch": 617.44,
197598
+ "learning_rate": 8.76928917609047e-06,
197599
+ "loss": 0.3451,
197600
+ "step": 77210
197601
+ },
197602
+ {
197603
+ "epoch": 617.48,
197604
+ "learning_rate": 8.769208400646204e-06,
197605
+ "loss": 0.2851,
197606
+ "step": 77215
197607
+ },
197608
+ {
197609
+ "epoch": 617.52,
197610
+ "learning_rate": 8.76912762520194e-06,
197611
+ "loss": 0.3446,
197612
+ "step": 77220
197613
+ },
197614
+ {
197615
+ "epoch": 617.56,
197616
+ "learning_rate": 8.769046849757674e-06,
197617
+ "loss": 0.4927,
197618
+ "step": 77225
197619
+ },
197620
+ {
197621
+ "epoch": 617.6,
197622
+ "learning_rate": 8.76896607431341e-06,
197623
+ "loss": 1.215,
197624
+ "step": 77230
197625
+ },
197626
+ {
197627
+ "epoch": 617.64,
197628
+ "learning_rate": 8.768885298869144e-06,
197629
+ "loss": 0.2961,
197630
+ "step": 77235
197631
+ },
197632
+ {
197633
+ "epoch": 617.68,
197634
+ "learning_rate": 8.76880452342488e-06,
197635
+ "loss": 0.2855,
197636
+ "step": 77240
197637
+ },
197638
+ {
197639
+ "epoch": 617.72,
197640
+ "learning_rate": 8.768723747980614e-06,
197641
+ "loss": 0.3748,
197642
+ "step": 77245
197643
+ },
197644
+ {
197645
+ "epoch": 617.76,
197646
+ "learning_rate": 8.76864297253635e-06,
197647
+ "loss": 0.5142,
197648
+ "step": 77250
197649
+ },
197650
+ {
197651
+ "epoch": 617.8,
197652
+ "learning_rate": 8.768562197092084e-06,
197653
+ "loss": 1.1865,
197654
+ "step": 77255
197655
+ },
197656
+ {
197657
+ "epoch": 617.84,
197658
+ "learning_rate": 8.76848142164782e-06,
197659
+ "loss": 0.3083,
197660
+ "step": 77260
197661
+ },
197662
+ {
197663
+ "epoch": 617.88,
197664
+ "learning_rate": 8.768400646203555e-06,
197665
+ "loss": 0.2774,
197666
+ "step": 77265
197667
+ },
197668
+ {
197669
+ "epoch": 617.92,
197670
+ "learning_rate": 8.76831987075929e-06,
197671
+ "loss": 0.3284,
197672
+ "step": 77270
197673
+ },
197674
+ {
197675
+ "epoch": 617.96,
197676
+ "learning_rate": 8.768239095315025e-06,
197677
+ "loss": 0.5053,
197678
+ "step": 77275
197679
+ },
197680
+ {
197681
+ "epoch": 618.0,
197682
+ "learning_rate": 8.76815831987076e-06,
197683
+ "loss": 1.2728,
197684
+ "step": 77280
197685
+ },
197686
+ {
197687
+ "epoch": 618.0,
197688
+ "eval_loss": 0.37362736463546753,
197689
+ "eval_runtime": 41.9714,
197690
+ "eval_samples_per_second": 19.918,
197691
+ "eval_steps_per_second": 0.643,
197692
+ "eval_wer": 0.1899724677582959,
197693
+ "step": 77280
197694
+ },
197695
+ {
197696
+ "epoch": 618.04,
197697
+ "learning_rate": 8.768077544426495e-06,
197698
+ "loss": 0.4184,
197699
+ "step": 77285
197700
+ },
197701
+ {
197702
+ "epoch": 618.08,
197703
+ "learning_rate": 8.76799676898223e-06,
197704
+ "loss": 0.3041,
197705
+ "step": 77290
197706
+ },
197707
+ {
197708
+ "epoch": 618.12,
197709
+ "learning_rate": 8.767915993537965e-06,
197710
+ "loss": 0.371,
197711
+ "step": 77295
197712
+ },
197713
+ {
197714
+ "epoch": 618.16,
197715
+ "learning_rate": 8.7678352180937e-06,
197716
+ "loss": 0.516,
197717
+ "step": 77300
197718
+ },
197719
+ {
197720
+ "epoch": 618.2,
197721
+ "learning_rate": 8.767754442649435e-06,
197722
+ "loss": 1.1994,
197723
+ "step": 77305
197724
+ },
197725
+ {
197726
+ "epoch": 618.24,
197727
+ "learning_rate": 8.76767366720517e-06,
197728
+ "loss": 0.3946,
197729
+ "step": 77310
197730
+ },
197731
+ {
197732
+ "epoch": 618.28,
197733
+ "learning_rate": 8.767592891760905e-06,
197734
+ "loss": 0.32,
197735
+ "step": 77315
197736
+ },
197737
+ {
197738
+ "epoch": 618.32,
197739
+ "learning_rate": 8.767512116316641e-06,
197740
+ "loss": 0.318,
197741
+ "step": 77320
197742
+ },
197743
+ {
197744
+ "epoch": 618.36,
197745
+ "learning_rate": 8.767431340872375e-06,
197746
+ "loss": 0.4967,
197747
+ "step": 77325
197748
+ },
197749
+ {
197750
+ "epoch": 618.4,
197751
+ "learning_rate": 8.767350565428111e-06,
197752
+ "loss": 1.3815,
197753
+ "step": 77330
197754
+ },
197755
+ {
197756
+ "epoch": 618.44,
197757
+ "learning_rate": 8.767269789983845e-06,
197758
+ "loss": 0.3609,
197759
+ "step": 77335
197760
+ },
197761
+ {
197762
+ "epoch": 618.48,
197763
+ "learning_rate": 8.767189014539581e-06,
197764
+ "loss": 0.3027,
197765
+ "step": 77340
197766
+ },
197767
+ {
197768
+ "epoch": 618.52,
197769
+ "learning_rate": 8.767108239095315e-06,
197770
+ "loss": 0.3554,
197771
+ "step": 77345
197772
+ },
197773
+ {
197774
+ "epoch": 618.56,
197775
+ "learning_rate": 8.767027463651051e-06,
197776
+ "loss": 0.4742,
197777
+ "step": 77350
197778
+ },
197779
+ {
197780
+ "epoch": 618.6,
197781
+ "learning_rate": 8.766946688206785e-06,
197782
+ "loss": 1.1874,
197783
+ "step": 77355
197784
+ },
197785
+ {
197786
+ "epoch": 618.64,
197787
+ "learning_rate": 8.766865912762521e-06,
197788
+ "loss": 0.3542,
197789
+ "step": 77360
197790
+ },
197791
+ {
197792
+ "epoch": 618.68,
197793
+ "learning_rate": 8.766785137318255e-06,
197794
+ "loss": 0.2895,
197795
+ "step": 77365
197796
+ },
197797
+ {
197798
+ "epoch": 618.72,
197799
+ "learning_rate": 8.766704361873991e-06,
197800
+ "loss": 0.31,
197801
+ "step": 77370
197802
+ },
197803
+ {
197804
+ "epoch": 618.76,
197805
+ "learning_rate": 8.766623586429725e-06,
197806
+ "loss": 0.4415,
197807
+ "step": 77375
197808
+ },
197809
+ {
197810
+ "epoch": 618.8,
197811
+ "learning_rate": 8.766542810985461e-06,
197812
+ "loss": 1.3175,
197813
+ "step": 77380
197814
+ },
197815
+ {
197816
+ "epoch": 618.84,
197817
+ "learning_rate": 8.766462035541197e-06,
197818
+ "loss": 0.3257,
197819
+ "step": 77385
197820
+ },
197821
+ {
197822
+ "epoch": 618.88,
197823
+ "learning_rate": 8.766381260096931e-06,
197824
+ "loss": 0.3193,
197825
+ "step": 77390
197826
+ },
197827
+ {
197828
+ "epoch": 618.92,
197829
+ "learning_rate": 8.766300484652667e-06,
197830
+ "loss": 0.3781,
197831
+ "step": 77395
197832
+ },
197833
+ {
197834
+ "epoch": 618.96,
197835
+ "learning_rate": 8.766219709208401e-06,
197836
+ "loss": 0.5195,
197837
+ "step": 77400
197838
+ },
197839
+ {
197840
+ "epoch": 619.0,
197841
+ "learning_rate": 8.766138933764137e-06,
197842
+ "loss": 1.2633,
197843
+ "step": 77405
197844
+ },
197845
+ {
197846
+ "epoch": 619.0,
197847
+ "eval_loss": 0.38753950595855713,
197848
+ "eval_runtime": 41.8381,
197849
+ "eval_samples_per_second": 19.982,
197850
+ "eval_steps_per_second": 0.645,
197851
+ "eval_wer": 0.1945414847161572,
197852
+ "step": 77405
197853
+ },
197854
+ {
197855
+ "epoch": 619.04,
197856
+ "learning_rate": 8.76605815831987e-06,
197857
+ "loss": 0.3856,
197858
+ "step": 77410
197859
+ },
197860
+ {
197861
+ "epoch": 619.08,
197862
+ "learning_rate": 8.765977382875607e-06,
197863
+ "loss": 0.3124,
197864
+ "step": 77415
197865
+ },
197866
+ {
197867
+ "epoch": 619.12,
197868
+ "learning_rate": 8.76589660743134e-06,
197869
+ "loss": 0.4006,
197870
+ "step": 77420
197871
+ },
197872
+ {
197873
+ "epoch": 619.16,
197874
+ "learning_rate": 8.765815831987077e-06,
197875
+ "loss": 0.4272,
197876
+ "step": 77425
197877
+ },
197878
+ {
197879
+ "epoch": 619.2,
197880
+ "learning_rate": 8.76573505654281e-06,
197881
+ "loss": 1.1045,
197882
+ "step": 77430
197883
+ },
197884
+ {
197885
+ "epoch": 619.24,
197886
+ "learning_rate": 8.765654281098547e-06,
197887
+ "loss": 0.3543,
197888
+ "step": 77435
197889
+ },
197890
+ {
197891
+ "epoch": 619.28,
197892
+ "learning_rate": 8.765573505654282e-06,
197893
+ "loss": 0.388,
197894
+ "step": 77440
197895
+ },
197896
+ {
197897
+ "epoch": 619.32,
197898
+ "learning_rate": 8.765492730210017e-06,
197899
+ "loss": 0.3724,
197900
+ "step": 77445
197901
+ },
197902
+ {
197903
+ "epoch": 619.36,
197904
+ "learning_rate": 8.765411954765752e-06,
197905
+ "loss": 0.3966,
197906
+ "step": 77450
197907
+ },
197908
+ {
197909
+ "epoch": 619.4,
197910
+ "learning_rate": 8.765331179321487e-06,
197911
+ "loss": 1.2215,
197912
+ "step": 77455
197913
+ },
197914
+ {
197915
+ "epoch": 619.44,
197916
+ "learning_rate": 8.765250403877222e-06,
197917
+ "loss": 0.3409,
197918
+ "step": 77460
197919
+ },
197920
+ {
197921
+ "epoch": 619.48,
197922
+ "learning_rate": 8.765169628432957e-06,
197923
+ "loss": 0.3197,
197924
+ "step": 77465
197925
+ },
197926
+ {
197927
+ "epoch": 619.52,
197928
+ "learning_rate": 8.765088852988692e-06,
197929
+ "loss": 0.3963,
197930
+ "step": 77470
197931
+ },
197932
+ {
197933
+ "epoch": 619.56,
197934
+ "learning_rate": 8.765008077544427e-06,
197935
+ "loss": 0.4343,
197936
+ "step": 77475
197937
+ },
197938
+ {
197939
+ "epoch": 619.6,
197940
+ "learning_rate": 8.764927302100162e-06,
197941
+ "loss": 1.3709,
197942
+ "step": 77480
197943
+ },
197944
+ {
197945
+ "epoch": 619.64,
197946
+ "learning_rate": 8.764846526655896e-06,
197947
+ "loss": 0.2994,
197948
+ "step": 77485
197949
+ },
197950
+ {
197951
+ "epoch": 619.68,
197952
+ "learning_rate": 8.764765751211632e-06,
197953
+ "loss": 0.346,
197954
+ "step": 77490
197955
+ },
197956
+ {
197957
+ "epoch": 619.72,
197958
+ "learning_rate": 8.764684975767368e-06,
197959
+ "loss": 0.3852,
197960
+ "step": 77495
197961
+ },
197962
+ {
197963
+ "epoch": 619.76,
197964
+ "learning_rate": 8.764604200323102e-06,
197965
+ "loss": 0.5026,
197966
+ "step": 77500
197967
+ },
197968
+ {
197969
+ "epoch": 619.8,
197970
+ "learning_rate": 8.764523424878838e-06,
197971
+ "loss": 1.3511,
197972
+ "step": 77505
197973
+ },
197974
+ {
197975
+ "epoch": 619.84,
197976
+ "learning_rate": 8.764442649434572e-06,
197977
+ "loss": 0.3118,
197978
+ "step": 77510
197979
+ },
197980
+ {
197981
+ "epoch": 619.88,
197982
+ "learning_rate": 8.764361873990308e-06,
197983
+ "loss": 0.3242,
197984
+ "step": 77515
197985
+ },
197986
+ {
197987
+ "epoch": 619.92,
197988
+ "learning_rate": 8.764281098546042e-06,
197989
+ "loss": 0.3453,
197990
+ "step": 77520
197991
+ },
197992
+ {
197993
+ "epoch": 619.96,
197994
+ "learning_rate": 8.764200323101778e-06,
197995
+ "loss": 0.4682,
197996
+ "step": 77525
197997
+ },
197998
+ {
197999
+ "epoch": 620.0,
198000
+ "learning_rate": 8.764119547657512e-06,
198001
+ "loss": 1.1805,
198002
+ "step": 77530
198003
+ },
198004
+ {
198005
+ "epoch": 620.0,
198006
+ "eval_loss": 0.4079735279083252,
198007
+ "eval_runtime": 43.0502,
198008
+ "eval_samples_per_second": 19.419,
198009
+ "eval_steps_per_second": 0.627,
198010
+ "eval_wer": 0.18670113456843315,
198011
+ "step": 77530
198012
+ },
198013
+ {
198014
+ "epoch": 620.04,
198015
+ "learning_rate": 8.764038772213248e-06,
198016
+ "loss": 0.3656,
198017
+ "step": 77535
198018
+ },
198019
+ {
198020
+ "epoch": 620.08,
198021
+ "learning_rate": 8.763957996768982e-06,
198022
+ "loss": 0.2756,
198023
+ "step": 77540
198024
+ },
198025
+ {
198026
+ "epoch": 620.12,
198027
+ "learning_rate": 8.763877221324718e-06,
198028
+ "loss": 0.4158,
198029
+ "step": 77545
198030
+ },
198031
+ {
198032
+ "epoch": 620.16,
198033
+ "learning_rate": 8.763796445880452e-06,
198034
+ "loss": 0.4729,
198035
+ "step": 77550
198036
+ },
198037
+ {
198038
+ "epoch": 620.2,
198039
+ "learning_rate": 8.763715670436188e-06,
198040
+ "loss": 1.4564,
198041
+ "step": 77555
198042
+ },
198043
+ {
198044
+ "epoch": 620.24,
198045
+ "learning_rate": 8.763634894991924e-06,
198046
+ "loss": 0.3279,
198047
+ "step": 77560
198048
+ },
198049
+ {
198050
+ "epoch": 620.28,
198051
+ "learning_rate": 8.763554119547658e-06,
198052
+ "loss": 0.3054,
198053
+ "step": 77565
198054
+ },
198055
+ {
198056
+ "epoch": 620.32,
198057
+ "learning_rate": 8.763473344103394e-06,
198058
+ "loss": 0.3148,
198059
+ "step": 77570
198060
+ },
198061
+ {
198062
+ "epoch": 620.36,
198063
+ "learning_rate": 8.763392568659128e-06,
198064
+ "loss": 0.4941,
198065
+ "step": 77575
198066
+ },
198067
+ {
198068
+ "epoch": 620.4,
198069
+ "learning_rate": 8.763311793214864e-06,
198070
+ "loss": 1.3589,
198071
+ "step": 77580
198072
+ },
198073
+ {
198074
+ "epoch": 620.44,
198075
+ "learning_rate": 8.763231017770598e-06,
198076
+ "loss": 0.3719,
198077
+ "step": 77585
198078
+ },
198079
+ {
198080
+ "epoch": 620.48,
198081
+ "learning_rate": 8.763150242326334e-06,
198082
+ "loss": 0.2943,
198083
+ "step": 77590
198084
+ },
198085
+ {
198086
+ "epoch": 620.52,
198087
+ "learning_rate": 8.763069466882068e-06,
198088
+ "loss": 0.414,
198089
+ "step": 77595
198090
+ },
198091
+ {
198092
+ "epoch": 620.56,
198093
+ "learning_rate": 8.762988691437804e-06,
198094
+ "loss": 0.3896,
198095
+ "step": 77600
198096
+ },
198097
+ {
198098
+ "epoch": 620.6,
198099
+ "learning_rate": 8.762907915993538e-06,
198100
+ "loss": 1.2367,
198101
+ "step": 77605
198102
+ },
198103
+ {
198104
+ "epoch": 620.64,
198105
+ "learning_rate": 8.762827140549274e-06,
198106
+ "loss": 0.3736,
198107
+ "step": 77610
198108
+ },
198109
+ {
198110
+ "epoch": 620.68,
198111
+ "learning_rate": 8.76274636510501e-06,
198112
+ "loss": 0.3219,
198113
+ "step": 77615
198114
+ },
198115
+ {
198116
+ "epoch": 620.72,
198117
+ "learning_rate": 8.762665589660744e-06,
198118
+ "loss": 0.4256,
198119
+ "step": 77620
198120
+ },
198121
+ {
198122
+ "epoch": 620.76,
198123
+ "learning_rate": 8.76258481421648e-06,
198124
+ "loss": 0.5015,
198125
+ "step": 77625
198126
+ },
198127
+ {
198128
+ "epoch": 620.8,
198129
+ "learning_rate": 8.762504038772214e-06,
198130
+ "loss": 1.3466,
198131
+ "step": 77630
198132
+ },
198133
+ {
198134
+ "epoch": 620.84,
198135
+ "learning_rate": 8.76242326332795e-06,
198136
+ "loss": 0.3469,
198137
+ "step": 77635
198138
+ },
198139
+ {
198140
+ "epoch": 620.88,
198141
+ "learning_rate": 8.762342487883684e-06,
198142
+ "loss": 0.2818,
198143
+ "step": 77640
198144
+ },
198145
+ {
198146
+ "epoch": 620.92,
198147
+ "learning_rate": 8.76226171243942e-06,
198148
+ "loss": 0.3715,
198149
+ "step": 77645
198150
+ },
198151
+ {
198152
+ "epoch": 620.96,
198153
+ "learning_rate": 8.762180936995154e-06,
198154
+ "loss": 0.5095,
198155
+ "step": 77650
198156
+ },
198157
+ {
198158
+ "epoch": 621.0,
198159
+ "learning_rate": 8.76210016155089e-06,
198160
+ "loss": 1.3371,
198161
+ "step": 77655
198162
+ },
198163
+ {
198164
+ "epoch": 621.0,
198165
+ "eval_loss": 0.3949979543685913,
198166
+ "eval_runtime": 41.8082,
198167
+ "eval_samples_per_second": 20.02,
198168
+ "eval_steps_per_second": 0.646,
198169
+ "eval_wer": 0.19575146813601102,
198170
+ "step": 77655
198171
+ },
198172
+ {
198173
+ "epoch": 626.04,
198174
+ "learning_rate": 8.762019386106624e-06,
198175
+ "loss": 0.3413,
198176
+ "step": 77660
198177
+ },
198178
+ {
198179
+ "epoch": 626.08,
198180
+ "learning_rate": 8.76193861066236e-06,
198181
+ "loss": 0.3734,
198182
+ "step": 77665
198183
+ },
198184
+ {
198185
+ "epoch": 626.12,
198186
+ "learning_rate": 8.761857835218095e-06,
198187
+ "loss": 0.3673,
198188
+ "step": 77670
198189
+ },
198190
+ {
198191
+ "epoch": 626.16,
198192
+ "learning_rate": 8.76177705977383e-06,
198193
+ "loss": 0.4349,
198194
+ "step": 77675
198195
+ },
198196
+ {
198197
+ "epoch": 626.2,
198198
+ "learning_rate": 8.761696284329565e-06,
198199
+ "loss": 1.4509,
198200
+ "step": 77680
198201
+ },
198202
+ {
198203
+ "epoch": 626.24,
198204
+ "learning_rate": 8.7616155088853e-06,
198205
+ "loss": 0.3588,
198206
+ "step": 77685
198207
+ },
198208
+ {
198209
+ "epoch": 626.28,
198210
+ "learning_rate": 8.761534733441035e-06,
198211
+ "loss": 0.3109,
198212
+ "step": 77690
198213
+ },
198214
+ {
198215
+ "epoch": 626.32,
198216
+ "learning_rate": 8.76145395799677e-06,
198217
+ "loss": 0.3368,
198218
+ "step": 77695
198219
+ },
198220
+ {
198221
+ "epoch": 626.36,
198222
+ "learning_rate": 8.761373182552505e-06,
198223
+ "loss": 0.5113,
198224
+ "step": 77700
198225
+ },
198226
+ {
198227
+ "epoch": 626.4,
198228
+ "learning_rate": 8.76129240710824e-06,
198229
+ "loss": 1.1518,
198230
+ "step": 77705
198231
+ },
198232
+ {
198233
+ "epoch": 626.44,
198234
+ "learning_rate": 8.761211631663975e-06,
198235
+ "loss": 0.4165,
198236
+ "step": 77710
198237
+ },
198238
+ {
198239
+ "epoch": 626.48,
198240
+ "learning_rate": 8.76113085621971e-06,
198241
+ "loss": 0.3262,
198242
+ "step": 77715
198243
+ },
198244
+ {
198245
+ "epoch": 626.52,
198246
+ "learning_rate": 8.761050080775445e-06,
198247
+ "loss": 0.3053,
198248
+ "step": 77720
198249
+ },
198250
+ {
198251
+ "epoch": 626.56,
198252
+ "learning_rate": 8.76096930533118e-06,
198253
+ "loss": 0.4874,
198254
+ "step": 77725
198255
+ },
198256
+ {
198257
+ "epoch": 626.6,
198258
+ "learning_rate": 8.760888529886915e-06,
198259
+ "loss": 1.3094,
198260
+ "step": 77730
198261
+ },
198262
+ {
198263
+ "epoch": 626.64,
198264
+ "learning_rate": 8.760807754442651e-06,
198265
+ "loss": 0.3594,
198266
+ "step": 77735
198267
+ },
198268
+ {
198269
+ "epoch": 626.68,
198270
+ "learning_rate": 8.760726978998385e-06,
198271
+ "loss": 0.3254,
198272
+ "step": 77740
198273
+ },
198274
+ {
198275
+ "epoch": 626.72,
198276
+ "learning_rate": 8.760646203554121e-06,
198277
+ "loss": 0.3081,
198278
+ "step": 77745
198279
+ },
198280
+ {
198281
+ "epoch": 626.76,
198282
+ "learning_rate": 8.760565428109855e-06,
198283
+ "loss": 0.4125,
198284
+ "step": 77750
198285
+ },
198286
+ {
198287
+ "epoch": 626.8,
198288
+ "learning_rate": 8.760484652665591e-06,
198289
+ "loss": 1.2727,
198290
+ "step": 77755
198291
+ },
198292
+ {
198293
+ "epoch": 626.84,
198294
+ "learning_rate": 8.760403877221325e-06,
198295
+ "loss": 0.3183,
198296
+ "step": 77760
198297
+ },
198298
+ {
198299
+ "epoch": 626.88,
198300
+ "learning_rate": 8.760323101777061e-06,
198301
+ "loss": 0.3672,
198302
+ "step": 77765
198303
+ },
198304
+ {
198305
+ "epoch": 626.92,
198306
+ "learning_rate": 8.760242326332795e-06,
198307
+ "loss": 0.3245,
198308
+ "step": 77770
198309
+ },
198310
+ {
198311
+ "epoch": 626.96,
198312
+ "learning_rate": 8.760161550888531e-06,
198313
+ "loss": 0.4834,
198314
+ "step": 77775
198315
+ },
198316
+ {
198317
+ "epoch": 627.0,
198318
+ "eval_loss": 0.4380192756652832,
198319
+ "eval_runtime": 42.5503,
198320
+ "eval_samples_per_second": 19.671,
198321
+ "eval_steps_per_second": 0.635,
198322
+ "eval_wer": 0.19297325642370214,
198323
+ "step": 77779
198324
  }
198325
  ],
198326
  "max_steps": 620000,
198327
  "num_train_epochs": 5000,
198328
+ "total_flos": 2.1888125289934658e+20,
198329
  "trial_name": null,
198330
  "trial_params": null
198331
  }
model-bin/finetune/base/{checkpoint-77155 β†’ checkpoint-77779}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629892093.865816/events.out.tfevents.1629892093.7e498afd5545.7645.15 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93b6706032dcaef909be43b5fe7a4c558845d351dc9c196ee7a47b6fbe16f68c
3
+ size 4194
model-bin/finetune/base/log/1629892567.0899503/events.out.tfevents.1629892567.7e498afd5545.7645.17 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b3101719cd5afdbb2f7914ee12a139bd70492eed84bf82bdb1dcf857d2527a2
3
+ size 4194
model-bin/finetune/base/log/1629893045.2545307/events.out.tfevents.1629893045.7e498afd5545.7645.19 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0355fbcef8893ca75f858f1841846c42bc4c2b5d1d3b408071860d74efda4f7e
3
+ size 4194
model-bin/finetune/base/log/1629893510.1288207/events.out.tfevents.1629893510.7e498afd5545.7645.21 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca5317449e2f180e69d2ef819442715847aeccfca0f66584ed91a0f8059f7839
3
+ size 4194
model-bin/finetune/base/log/1629893973.3909373/events.out.tfevents.1629893973.7e498afd5545.7645.23 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d13be364e6cd4fcd39d562b3f0a9d11e0f53aa5162c3cfcc76a07ad5dcc44f0
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629892093.7e498afd5545.7645.14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c4ad90a8d0f68de4a48de6f09259d2c3f3da8cc2f22563aee5b7e4836c14eb4
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629892566.7e498afd5545.7645.16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5adf725e71f1292675eddf7ef8d38ef14fe9c7b57c7e462c565c2c82a571eca
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629893045.7e498afd5545.7645.18 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c1771ab9166258a17533cdc266bb1be41a00d8a695c19f0ee7b17c1bb6559aa
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629893510.7e498afd5545.7645.20 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4b988b11d6a85d25d2a05c50594be9524ebf1c8e4daa4030ae5837283db84ec
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629893973.7e498afd5545.7645.22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ed23b1847a5e37f605602b624d41aa91c18bb2180f34333cc53c8aad8696a34
3
+ size 8462