Check commited on
Commit
da5620b
Β·
1 Parent(s): 4ada2a4

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/rng_state.pth +2 -2
  6. model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/trainer_state.json +799 -4
  9. model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630150760.2881565/events.out.tfevents.1630150760.86bb0ddabf9b.4092.141 +3 -0
  11. model-bin/finetune/base/log/1630151150.1171627/events.out.tfevents.1630151150.86bb0ddabf9b.4092.143 +3 -0
  12. model-bin/finetune/base/log/1630151541.9840891/events.out.tfevents.1630151541.86bb0ddabf9b.4092.145 +3 -0
  13. model-bin/finetune/base/log/1630151929.4076536/events.out.tfevents.1630151929.86bb0ddabf9b.4092.147 +3 -0
  14. model-bin/finetune/base/log/1630152323.041462/events.out.tfevents.1630152323.86bb0ddabf9b.4092.149 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630150760.86bb0ddabf9b.4092.140 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630151150.86bb0ddabf9b.4092.142 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630151541.86bb0ddabf9b.4092.144 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630151929.86bb0ddabf9b.4092.146 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630152323.86bb0ddabf9b.4092.148 +3 -0
model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3845dc63451d0a1c44498d69f0b827119c42d6e557cf97a7f9735c52705644bd
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b3156a8f4e201766c4df6ffa74d8abc97daafe531c5db333d2a9a6e1bbb367c
3
  size 722165393
model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:69b6d80c964bf8059c1020726fed10f992c3b777c5f499f82bdede9d7ca4ced7
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:380e4ce8c820adecac0ac117754c22dbcbee91001dd4e41eebe002872dbe631c
3
  size 377909911
model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0dfcfe608f0a02dd951bb37028335198611a4d183fa2e67bc1b1aed79918029b
3
- size 14567
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5dee93488770f745feda0ea9fd7c0a22d4ef4429c854d4af13ea78ee1e76527d
3
+ size 14503
model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f7595729f26a675659d761b684c35ca3b092bd8547ff8d0cfbbdd1d52fc1473
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:703b2c951df4efdfabf698ea4ebf8efc6d144661a2d4c8aee86aa621f308f4ea
3
  size 559
model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e4e46dd1d0e522e3ca1340174ec5347c27a555d6146d2d14b88e929a99348346
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9bc4ee8cc526980e6e93a1f57d2bdc9f5f11496f758041ced6b9b5c7d86ec3e
3
  size 623
model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.17162025681719809,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-124947",
4
- "epoch": 1020.995983935743,
5
- "global_step": 126689,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -260550,11 +260550,806 @@
260550
  "eval_steps_per_second": 0.742,
260551
  "eval_wer": 0.17306996599377758,
260552
  "step": 126689
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
260553
  }
260554
  ],
260555
- "max_steps": 620000,
260556
  "num_train_epochs": 5000,
260557
- "total_flos": 3.565038460292019e+20,
260558
  "trial_name": null,
260559
  "trial_params": null
260560
  }
 
1
  {
2
  "best_metric": 0.17162025681719809,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-124947",
4
+ "epoch": 1018.0,
5
+ "global_step": 127312,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
260550
  "eval_steps_per_second": 0.742,
260551
  "eval_wer": 0.17306996599377758,
260552
  "step": 126689
260553
+ },
260554
+ {
260555
+ "epoch": 1013.01,
260556
+ "learning_rate": 7.986522435897436e-06,
260557
+ "loss": 0.2831,
260558
+ "step": 126690
260559
+ },
260560
+ {
260561
+ "epoch": 1013.05,
260562
+ "learning_rate": 7.986442307692308e-06,
260563
+ "loss": 0.2879,
260564
+ "step": 126695
260565
+ },
260566
+ {
260567
+ "epoch": 1013.09,
260568
+ "learning_rate": 7.98636217948718e-06,
260569
+ "loss": 0.2605,
260570
+ "step": 126700
260571
+ },
260572
+ {
260573
+ "epoch": 1013.13,
260574
+ "learning_rate": 7.986282051282052e-06,
260575
+ "loss": 0.346,
260576
+ "step": 126705
260577
+ },
260578
+ {
260579
+ "epoch": 1013.17,
260580
+ "learning_rate": 7.986201923076923e-06,
260581
+ "loss": 0.5423,
260582
+ "step": 126710
260583
+ },
260584
+ {
260585
+ "epoch": 1013.21,
260586
+ "learning_rate": 7.986121794871795e-06,
260587
+ "loss": 1.0971,
260588
+ "step": 126715
260589
+ },
260590
+ {
260591
+ "epoch": 1013.25,
260592
+ "learning_rate": 7.986041666666668e-06,
260593
+ "loss": 0.2661,
260594
+ "step": 126720
260595
+ },
260596
+ {
260597
+ "epoch": 1013.29,
260598
+ "learning_rate": 7.98596153846154e-06,
260599
+ "loss": 0.2675,
260600
+ "step": 126725
260601
+ },
260602
+ {
260603
+ "epoch": 1013.33,
260604
+ "learning_rate": 7.98588141025641e-06,
260605
+ "loss": 0.3517,
260606
+ "step": 126730
260607
+ },
260608
+ {
260609
+ "epoch": 1013.37,
260610
+ "learning_rate": 7.985801282051284e-06,
260611
+ "loss": 0.5183,
260612
+ "step": 126735
260613
+ },
260614
+ {
260615
+ "epoch": 1013.41,
260616
+ "learning_rate": 7.985721153846155e-06,
260617
+ "loss": 1.2005,
260618
+ "step": 126740
260619
+ },
260620
+ {
260621
+ "epoch": 1013.45,
260622
+ "learning_rate": 7.985641025641026e-06,
260623
+ "loss": 0.2968,
260624
+ "step": 126745
260625
+ },
260626
+ {
260627
+ "epoch": 1013.49,
260628
+ "learning_rate": 7.985560897435898e-06,
260629
+ "loss": 0.2646,
260630
+ "step": 126750
260631
+ },
260632
+ {
260633
+ "epoch": 1013.53,
260634
+ "learning_rate": 7.98548076923077e-06,
260635
+ "loss": 0.3582,
260636
+ "step": 126755
260637
+ },
260638
+ {
260639
+ "epoch": 1013.57,
260640
+ "learning_rate": 7.98540064102564e-06,
260641
+ "loss": 0.4968,
260642
+ "step": 126760
260643
+ },
260644
+ {
260645
+ "epoch": 1013.61,
260646
+ "learning_rate": 7.985320512820514e-06,
260647
+ "loss": 1.1176,
260648
+ "step": 126765
260649
+ },
260650
+ {
260651
+ "epoch": 1013.65,
260652
+ "learning_rate": 7.985240384615385e-06,
260653
+ "loss": 0.2675,
260654
+ "step": 126770
260655
+ },
260656
+ {
260657
+ "epoch": 1013.69,
260658
+ "learning_rate": 7.985160256410256e-06,
260659
+ "loss": 0.2741,
260660
+ "step": 126775
260661
+ },
260662
+ {
260663
+ "epoch": 1013.73,
260664
+ "learning_rate": 7.98508012820513e-06,
260665
+ "loss": 0.2681,
260666
+ "step": 126780
260667
+ },
260668
+ {
260669
+ "epoch": 1013.77,
260670
+ "learning_rate": 7.985e-06,
260671
+ "loss": 0.4765,
260672
+ "step": 126785
260673
+ },
260674
+ {
260675
+ "epoch": 1013.81,
260676
+ "learning_rate": 7.984919871794872e-06,
260677
+ "loss": 1.0048,
260678
+ "step": 126790
260679
+ },
260680
+ {
260681
+ "epoch": 1013.85,
260682
+ "learning_rate": 7.984839743589743e-06,
260683
+ "loss": 0.3135,
260684
+ "step": 126795
260685
+ },
260686
+ {
260687
+ "epoch": 1013.89,
260688
+ "learning_rate": 7.984759615384616e-06,
260689
+ "loss": 0.2559,
260690
+ "step": 126800
260691
+ },
260692
+ {
260693
+ "epoch": 1013.93,
260694
+ "learning_rate": 7.984679487179488e-06,
260695
+ "loss": 0.285,
260696
+ "step": 126805
260697
+ },
260698
+ {
260699
+ "epoch": 1013.97,
260700
+ "learning_rate": 7.984599358974359e-06,
260701
+ "loss": 0.5271,
260702
+ "step": 126810
260703
+ },
260704
+ {
260705
+ "epoch": 1014.0,
260706
+ "eval_loss": 0.39917340874671936,
260707
+ "eval_runtime": 35.7013,
260708
+ "eval_samples_per_second": 23.557,
260709
+ "eval_steps_per_second": 0.756,
260710
+ "eval_wer": 0.17787259084890392,
260711
+ "step": 126814
260712
+ },
260713
+ {
260714
+ "epoch": 1014.01,
260715
+ "learning_rate": 7.984535256410258e-06,
260716
+ "loss": 0.3605,
260717
+ "step": 126815
260718
+ },
260719
+ {
260720
+ "epoch": 1014.05,
260721
+ "learning_rate": 7.984455128205128e-06,
260722
+ "loss": 0.2761,
260723
+ "step": 126820
260724
+ },
260725
+ {
260726
+ "epoch": 1014.09,
260727
+ "learning_rate": 7.984375000000001e-06,
260728
+ "loss": 0.2825,
260729
+ "step": 126825
260730
+ },
260731
+ {
260732
+ "epoch": 1014.13,
260733
+ "learning_rate": 7.984294871794872e-06,
260734
+ "loss": 0.2817,
260735
+ "step": 126830
260736
+ },
260737
+ {
260738
+ "epoch": 1014.17,
260739
+ "learning_rate": 7.984214743589744e-06,
260740
+ "loss": 0.5611,
260741
+ "step": 126835
260742
+ },
260743
+ {
260744
+ "epoch": 1014.21,
260745
+ "learning_rate": 7.984134615384615e-06,
260746
+ "loss": 1.0804,
260747
+ "step": 126840
260748
+ },
260749
+ {
260750
+ "epoch": 1014.25,
260751
+ "learning_rate": 7.984054487179488e-06,
260752
+ "loss": 0.2839,
260753
+ "step": 126845
260754
+ },
260755
+ {
260756
+ "epoch": 1014.29,
260757
+ "learning_rate": 7.98397435897436e-06,
260758
+ "loss": 0.305,
260759
+ "step": 126850
260760
+ },
260761
+ {
260762
+ "epoch": 1014.33,
260763
+ "learning_rate": 7.98389423076923e-06,
260764
+ "loss": 0.3367,
260765
+ "step": 126855
260766
+ },
260767
+ {
260768
+ "epoch": 1014.37,
260769
+ "learning_rate": 7.983814102564104e-06,
260770
+ "loss": 0.5284,
260771
+ "step": 126860
260772
+ },
260773
+ {
260774
+ "epoch": 1014.41,
260775
+ "learning_rate": 7.983733974358975e-06,
260776
+ "loss": 1.2354,
260777
+ "step": 126865
260778
+ },
260779
+ {
260780
+ "epoch": 1014.45,
260781
+ "learning_rate": 7.983653846153847e-06,
260782
+ "loss": 0.2866,
260783
+ "step": 126870
260784
+ },
260785
+ {
260786
+ "epoch": 1014.49,
260787
+ "learning_rate": 7.983573717948718e-06,
260788
+ "loss": 0.2693,
260789
+ "step": 126875
260790
+ },
260791
+ {
260792
+ "epoch": 1014.53,
260793
+ "learning_rate": 7.983493589743591e-06,
260794
+ "loss": 0.2434,
260795
+ "step": 126880
260796
+ },
260797
+ {
260798
+ "epoch": 1014.57,
260799
+ "learning_rate": 7.983413461538462e-06,
260800
+ "loss": 0.6202,
260801
+ "step": 126885
260802
+ },
260803
+ {
260804
+ "epoch": 1014.61,
260805
+ "learning_rate": 7.983333333333334e-06,
260806
+ "loss": 1.2398,
260807
+ "step": 126890
260808
+ },
260809
+ {
260810
+ "epoch": 1014.65,
260811
+ "learning_rate": 7.983253205128205e-06,
260812
+ "loss": 0.3453,
260813
+ "step": 126895
260814
+ },
260815
+ {
260816
+ "epoch": 1014.69,
260817
+ "learning_rate": 7.983173076923078e-06,
260818
+ "loss": 0.297,
260819
+ "step": 126900
260820
+ },
260821
+ {
260822
+ "epoch": 1014.73,
260823
+ "learning_rate": 7.98309294871795e-06,
260824
+ "loss": 0.3543,
260825
+ "step": 126905
260826
+ },
260827
+ {
260828
+ "epoch": 1014.77,
260829
+ "learning_rate": 7.98301282051282e-06,
260830
+ "loss": 0.5697,
260831
+ "step": 126910
260832
+ },
260833
+ {
260834
+ "epoch": 1014.81,
260835
+ "learning_rate": 7.982932692307694e-06,
260836
+ "loss": 1.1563,
260837
+ "step": 126915
260838
+ },
260839
+ {
260840
+ "epoch": 1014.85,
260841
+ "learning_rate": 7.982852564102565e-06,
260842
+ "loss": 0.2929,
260843
+ "step": 126920
260844
+ },
260845
+ {
260846
+ "epoch": 1014.89,
260847
+ "learning_rate": 7.982772435897437e-06,
260848
+ "loss": 0.2657,
260849
+ "step": 126925
260850
+ },
260851
+ {
260852
+ "epoch": 1014.93,
260853
+ "learning_rate": 7.982692307692308e-06,
260854
+ "loss": 0.3184,
260855
+ "step": 126930
260856
+ },
260857
+ {
260858
+ "epoch": 1014.97,
260859
+ "learning_rate": 7.982612179487181e-06,
260860
+ "loss": 0.5176,
260861
+ "step": 126935
260862
+ },
260863
+ {
260864
+ "epoch": 1015.0,
260865
+ "eval_loss": 0.37343114614486694,
260866
+ "eval_runtime": 36.0496,
260867
+ "eval_samples_per_second": 23.329,
260868
+ "eval_steps_per_second": 0.749,
260869
+ "eval_wer": 0.18139708629062384,
260870
+ "step": 126939
260871
+ },
260872
+ {
260873
+ "epoch": 1023.01,
260874
+ "learning_rate": 7.98253205128205e-06,
260875
+ "loss": 0.4567,
260876
+ "step": 126940
260877
+ },
260878
+ {
260879
+ "epoch": 1023.05,
260880
+ "learning_rate": 7.982451923076924e-06,
260881
+ "loss": 0.2387,
260882
+ "step": 126945
260883
+ },
260884
+ {
260885
+ "epoch": 1023.09,
260886
+ "learning_rate": 7.982371794871797e-06,
260887
+ "loss": 0.3367,
260888
+ "step": 126950
260889
+ },
260890
+ {
260891
+ "epoch": 1023.13,
260892
+ "learning_rate": 7.982291666666666e-06,
260893
+ "loss": 0.2928,
260894
+ "step": 126955
260895
+ },
260896
+ {
260897
+ "epoch": 1023.17,
260898
+ "learning_rate": 7.98221153846154e-06,
260899
+ "loss": 0.5915,
260900
+ "step": 126960
260901
+ },
260902
+ {
260903
+ "epoch": 1023.21,
260904
+ "learning_rate": 7.982131410256411e-06,
260905
+ "loss": 1.155,
260906
+ "step": 126965
260907
+ },
260908
+ {
260909
+ "epoch": 1023.25,
260910
+ "learning_rate": 7.982051282051282e-06,
260911
+ "loss": 0.3101,
260912
+ "step": 126970
260913
+ },
260914
+ {
260915
+ "epoch": 1023.29,
260916
+ "learning_rate": 7.981971153846154e-06,
260917
+ "loss": 0.3658,
260918
+ "step": 126975
260919
+ },
260920
+ {
260921
+ "epoch": 1023.33,
260922
+ "learning_rate": 7.981891025641027e-06,
260923
+ "loss": 0.3383,
260924
+ "step": 126980
260925
+ },
260926
+ {
260927
+ "epoch": 1023.37,
260928
+ "learning_rate": 7.981810897435898e-06,
260929
+ "loss": 0.536,
260930
+ "step": 126985
260931
+ },
260932
+ {
260933
+ "epoch": 1023.41,
260934
+ "learning_rate": 7.98173076923077e-06,
260935
+ "loss": 1.1083,
260936
+ "step": 126990
260937
+ },
260938
+ {
260939
+ "epoch": 1023.45,
260940
+ "learning_rate": 7.98165064102564e-06,
260941
+ "loss": 0.2801,
260942
+ "step": 126995
260943
+ },
260944
+ {
260945
+ "epoch": 1023.49,
260946
+ "learning_rate": 7.981570512820514e-06,
260947
+ "loss": 0.3007,
260948
+ "step": 127000
260949
+ },
260950
+ {
260951
+ "epoch": 1023.53,
260952
+ "learning_rate": 7.981490384615385e-06,
260953
+ "loss": 0.3513,
260954
+ "step": 127005
260955
+ },
260956
+ {
260957
+ "epoch": 1023.57,
260958
+ "learning_rate": 7.981410256410256e-06,
260959
+ "loss": 0.6116,
260960
+ "step": 127010
260961
+ },
260962
+ {
260963
+ "epoch": 1023.61,
260964
+ "learning_rate": 7.98133012820513e-06,
260965
+ "loss": 1.1935,
260966
+ "step": 127015
260967
+ },
260968
+ {
260969
+ "epoch": 1023.65,
260970
+ "learning_rate": 7.981250000000001e-06,
260971
+ "loss": 0.3328,
260972
+ "step": 127020
260973
+ },
260974
+ {
260975
+ "epoch": 1023.69,
260976
+ "learning_rate": 7.981169871794872e-06,
260977
+ "loss": 0.2515,
260978
+ "step": 127025
260979
+ },
260980
+ {
260981
+ "epoch": 1023.73,
260982
+ "learning_rate": 7.981089743589744e-06,
260983
+ "loss": 0.2931,
260984
+ "step": 127030
260985
+ },
260986
+ {
260987
+ "epoch": 1023.77,
260988
+ "learning_rate": 7.981009615384617e-06,
260989
+ "loss": 0.499,
260990
+ "step": 127035
260991
+ },
260992
+ {
260993
+ "epoch": 1023.81,
260994
+ "learning_rate": 7.980929487179488e-06,
260995
+ "loss": 1.0857,
260996
+ "step": 127040
260997
+ },
260998
+ {
260999
+ "epoch": 1023.85,
261000
+ "learning_rate": 7.98084935897436e-06,
261001
+ "loss": 0.3193,
261002
+ "step": 127045
261003
+ },
261004
+ {
261005
+ "epoch": 1023.9,
261006
+ "learning_rate": 7.980769230769232e-06,
261007
+ "loss": 0.4392,
261008
+ "step": 127050
261009
+ },
261010
+ {
261011
+ "epoch": 1023.94,
261012
+ "learning_rate": 7.980689102564104e-06,
261013
+ "loss": 0.3994,
261014
+ "step": 127055
261015
+ },
261016
+ {
261017
+ "epoch": 1023.98,
261018
+ "learning_rate": 7.980608974358975e-06,
261019
+ "loss": 0.6098,
261020
+ "step": 127060
261021
+ },
261022
+ {
261023
+ "epoch": 1024.0,
261024
+ "eval_loss": 0.4028739035129547,
261025
+ "eval_runtime": 36.5089,
261026
+ "eval_samples_per_second": 23.036,
261027
+ "eval_steps_per_second": 0.74,
261028
+ "eval_wer": 0.17880398173789386,
261029
+ "step": 127063
261030
+ },
261031
+ {
261032
+ "epoch": 1024.02,
261033
+ "learning_rate": 7.980528846153846e-06,
261034
+ "loss": 0.3767,
261035
+ "step": 127065
261036
+ },
261037
+ {
261038
+ "epoch": 1024.06,
261039
+ "learning_rate": 7.98044871794872e-06,
261040
+ "loss": 0.2873,
261041
+ "step": 127070
261042
+ },
261043
+ {
261044
+ "epoch": 1024.1,
261045
+ "learning_rate": 7.980368589743591e-06,
261046
+ "loss": 0.2949,
261047
+ "step": 127075
261048
+ },
261049
+ {
261050
+ "epoch": 1024.14,
261051
+ "learning_rate": 7.980288461538462e-06,
261052
+ "loss": 0.3629,
261053
+ "step": 127080
261054
+ },
261055
+ {
261056
+ "epoch": 1024.18,
261057
+ "learning_rate": 7.980208333333334e-06,
261058
+ "loss": 0.5691,
261059
+ "step": 127085
261060
+ },
261061
+ {
261062
+ "epoch": 1024.22,
261063
+ "learning_rate": 7.980128205128207e-06,
261064
+ "loss": 1.0522,
261065
+ "step": 127090
261066
+ },
261067
+ {
261068
+ "epoch": 1024.26,
261069
+ "learning_rate": 7.980048076923076e-06,
261070
+ "loss": 0.2721,
261071
+ "step": 127095
261072
+ },
261073
+ {
261074
+ "epoch": 1024.3,
261075
+ "learning_rate": 7.97996794871795e-06,
261076
+ "loss": 0.2598,
261077
+ "step": 127100
261078
+ },
261079
+ {
261080
+ "epoch": 1024.34,
261081
+ "learning_rate": 7.979887820512822e-06,
261082
+ "loss": 0.3676,
261083
+ "step": 127105
261084
+ },
261085
+ {
261086
+ "epoch": 1024.38,
261087
+ "learning_rate": 7.979807692307692e-06,
261088
+ "loss": 0.5932,
261089
+ "step": 127110
261090
+ },
261091
+ {
261092
+ "epoch": 1024.42,
261093
+ "learning_rate": 7.979727564102565e-06,
261094
+ "loss": 0.9843,
261095
+ "step": 127115
261096
+ },
261097
+ {
261098
+ "epoch": 1024.46,
261099
+ "learning_rate": 7.979647435897437e-06,
261100
+ "loss": 0.2915,
261101
+ "step": 127120
261102
+ },
261103
+ {
261104
+ "epoch": 1024.5,
261105
+ "learning_rate": 7.979567307692308e-06,
261106
+ "loss": 0.2842,
261107
+ "step": 127125
261108
+ },
261109
+ {
261110
+ "epoch": 1024.54,
261111
+ "learning_rate": 7.97948717948718e-06,
261112
+ "loss": 0.3159,
261113
+ "step": 127130
261114
+ },
261115
+ {
261116
+ "epoch": 1024.58,
261117
+ "learning_rate": 7.979407051282052e-06,
261118
+ "loss": 0.5888,
261119
+ "step": 127135
261120
+ },
261121
+ {
261122
+ "epoch": 1024.62,
261123
+ "learning_rate": 7.979326923076924e-06,
261124
+ "loss": 0.9817,
261125
+ "step": 127140
261126
+ },
261127
+ {
261128
+ "epoch": 1024.66,
261129
+ "learning_rate": 7.979246794871795e-06,
261130
+ "loss": 0.2712,
261131
+ "step": 127145
261132
+ },
261133
+ {
261134
+ "epoch": 1024.7,
261135
+ "learning_rate": 7.979166666666668e-06,
261136
+ "loss": 0.3096,
261137
+ "step": 127150
261138
+ },
261139
+ {
261140
+ "epoch": 1024.74,
261141
+ "learning_rate": 7.97908653846154e-06,
261142
+ "loss": 0.3049,
261143
+ "step": 127155
261144
+ },
261145
+ {
261146
+ "epoch": 1024.78,
261147
+ "learning_rate": 7.97900641025641e-06,
261148
+ "loss": 0.6031,
261149
+ "step": 127160
261150
+ },
261151
+ {
261152
+ "epoch": 1024.82,
261153
+ "learning_rate": 7.978926282051282e-06,
261154
+ "loss": 0.9333,
261155
+ "step": 127165
261156
+ },
261157
+ {
261158
+ "epoch": 1024.86,
261159
+ "learning_rate": 7.978846153846155e-06,
261160
+ "loss": 0.2702,
261161
+ "step": 127170
261162
+ },
261163
+ {
261164
+ "epoch": 1024.9,
261165
+ "learning_rate": 7.978766025641027e-06,
261166
+ "loss": 0.2849,
261167
+ "step": 127175
261168
+ },
261169
+ {
261170
+ "epoch": 1024.94,
261171
+ "learning_rate": 7.978685897435898e-06,
261172
+ "loss": 0.3361,
261173
+ "step": 127180
261174
+ },
261175
+ {
261176
+ "epoch": 1024.98,
261177
+ "learning_rate": 7.97860576923077e-06,
261178
+ "loss": 0.5515,
261179
+ "step": 127185
261180
+ },
261181
+ {
261182
+ "epoch": 1025.0,
261183
+ "eval_loss": 0.30642202496528625,
261184
+ "eval_runtime": 36.1653,
261185
+ "eval_samples_per_second": 23.254,
261186
+ "eval_steps_per_second": 0.747,
261187
+ "eval_wer": 0.1736700622524052,
261188
+ "step": 127187
261189
+ },
261190
+ {
261191
+ "epoch": 1017.02,
261192
+ "learning_rate": 7.978525641025642e-06,
261193
+ "loss": 0.4288,
261194
+ "step": 127190
261195
+ },
261196
+ {
261197
+ "epoch": 1017.06,
261198
+ "learning_rate": 7.978445512820514e-06,
261199
+ "loss": 0.2895,
261200
+ "step": 127195
261201
+ },
261202
+ {
261203
+ "epoch": 1017.1,
261204
+ "learning_rate": 7.978365384615385e-06,
261205
+ "loss": 0.2966,
261206
+ "step": 127200
261207
+ },
261208
+ {
261209
+ "epoch": 1017.14,
261210
+ "learning_rate": 7.978285256410258e-06,
261211
+ "loss": 0.3754,
261212
+ "step": 127205
261213
+ },
261214
+ {
261215
+ "epoch": 1017.18,
261216
+ "learning_rate": 7.97820512820513e-06,
261217
+ "loss": 0.8517,
261218
+ "step": 127210
261219
+ },
261220
+ {
261221
+ "epoch": 1017.22,
261222
+ "learning_rate": 7.978125e-06,
261223
+ "loss": 0.8469,
261224
+ "step": 127215
261225
+ },
261226
+ {
261227
+ "epoch": 1017.26,
261228
+ "learning_rate": 7.978044871794872e-06,
261229
+ "loss": 0.2694,
261230
+ "step": 127220
261231
+ },
261232
+ {
261233
+ "epoch": 1017.3,
261234
+ "learning_rate": 7.977964743589745e-06,
261235
+ "loss": 0.2834,
261236
+ "step": 127225
261237
+ },
261238
+ {
261239
+ "epoch": 1017.34,
261240
+ "learning_rate": 7.977884615384615e-06,
261241
+ "loss": 0.4049,
261242
+ "step": 127230
261243
+ },
261244
+ {
261245
+ "epoch": 1017.38,
261246
+ "learning_rate": 7.977804487179488e-06,
261247
+ "loss": 0.6497,
261248
+ "step": 127235
261249
+ },
261250
+ {
261251
+ "epoch": 1017.42,
261252
+ "learning_rate": 7.97772435897436e-06,
261253
+ "loss": 0.8937,
261254
+ "step": 127240
261255
+ },
261256
+ {
261257
+ "epoch": 1017.46,
261258
+ "learning_rate": 7.97764423076923e-06,
261259
+ "loss": 0.2865,
261260
+ "step": 127245
261261
+ },
261262
+ {
261263
+ "epoch": 1017.5,
261264
+ "learning_rate": 7.977564102564104e-06,
261265
+ "loss": 0.2933,
261266
+ "step": 127250
261267
+ },
261268
+ {
261269
+ "epoch": 1017.54,
261270
+ "learning_rate": 7.977483974358975e-06,
261271
+ "loss": 0.3423,
261272
+ "step": 127255
261273
+ },
261274
+ {
261275
+ "epoch": 1017.58,
261276
+ "learning_rate": 7.977403846153846e-06,
261277
+ "loss": 0.636,
261278
+ "step": 127260
261279
+ },
261280
+ {
261281
+ "epoch": 1017.62,
261282
+ "learning_rate": 7.977323717948718e-06,
261283
+ "loss": 0.8939,
261284
+ "step": 127265
261285
+ },
261286
+ {
261287
+ "epoch": 1017.66,
261288
+ "learning_rate": 7.97724358974359e-06,
261289
+ "loss": 0.2902,
261290
+ "step": 127270
261291
+ },
261292
+ {
261293
+ "epoch": 1017.7,
261294
+ "learning_rate": 7.977163461538462e-06,
261295
+ "loss": 0.2895,
261296
+ "step": 127275
261297
+ },
261298
+ {
261299
+ "epoch": 1017.74,
261300
+ "learning_rate": 7.977083333333334e-06,
261301
+ "loss": 0.3549,
261302
+ "step": 127280
261303
+ },
261304
+ {
261305
+ "epoch": 1017.78,
261306
+ "learning_rate": 7.977003205128205e-06,
261307
+ "loss": 0.7446,
261308
+ "step": 127285
261309
+ },
261310
+ {
261311
+ "epoch": 1017.82,
261312
+ "learning_rate": 7.976923076923078e-06,
261313
+ "loss": 0.8723,
261314
+ "step": 127290
261315
+ },
261316
+ {
261317
+ "epoch": 1017.86,
261318
+ "learning_rate": 7.97684294871795e-06,
261319
+ "loss": 0.366,
261320
+ "step": 127295
261321
+ },
261322
+ {
261323
+ "epoch": 1017.9,
261324
+ "learning_rate": 7.97676282051282e-06,
261325
+ "loss": 0.2834,
261326
+ "step": 127300
261327
+ },
261328
+ {
261329
+ "epoch": 1017.94,
261330
+ "learning_rate": 7.976682692307694e-06,
261331
+ "loss": 0.5171,
261332
+ "step": 127305
261333
+ },
261334
+ {
261335
+ "epoch": 1017.98,
261336
+ "learning_rate": 7.976602564102565e-06,
261337
+ "loss": 0.754,
261338
+ "step": 127310
261339
+ },
261340
+ {
261341
+ "epoch": 1018.0,
261342
+ "eval_loss": 0.45526769757270813,
261343
+ "eval_runtime": 35.6393,
261344
+ "eval_samples_per_second": 23.598,
261345
+ "eval_steps_per_second": 0.758,
261346
+ "eval_wer": 0.18508140855736463,
261347
+ "step": 127312
261348
  }
261349
  ],
261350
+ "max_steps": 625000,
261351
  "num_train_epochs": 5000,
261352
+ "total_flos": 3.582563971308685e+20,
261353
  "trial_name": null,
261354
  "trial_params": null
261355
  }
model-bin/finetune/base/{checkpoint-126689 β†’ checkpoint-127312}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630150760.2881565/events.out.tfevents.1630150760.86bb0ddabf9b.4092.141 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:095dcf1ab0afdc359a9db9f65eb9d40ab61ba0174dddb16d9eec8f49bbf84f03
3
+ size 4194
model-bin/finetune/base/log/1630151150.1171627/events.out.tfevents.1630151150.86bb0ddabf9b.4092.143 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08ab4fbcb92cb5a41866a164ef988d957e600f50c11af4f07e19989b759eebbd
3
+ size 4194
model-bin/finetune/base/log/1630151541.9840891/events.out.tfevents.1630151541.86bb0ddabf9b.4092.145 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:417ec6e2a2d39968cae6f4aa173a2a8967a5eb715e4a85c1897ae1798e649597
3
+ size 4194
model-bin/finetune/base/log/1630151929.4076536/events.out.tfevents.1630151929.86bb0ddabf9b.4092.147 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5254878b62ded00fa804510340156dc900ed49ee89d328f177190c04e8b77caa
3
+ size 4194
model-bin/finetune/base/log/1630152323.041462/events.out.tfevents.1630152323.86bb0ddabf9b.4092.149 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a496ef544914a1fdb502b6e350fdc6c69eeed216c95e80fc33623f3b72ce3e9
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630150760.86bb0ddabf9b.4092.140 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3481ddc6fe8733f703c18a8ff359eced165cb5682bf077a705b29e57b482d529
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630151150.86bb0ddabf9b.4092.142 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c68c4278c0f61ff4965befa70beccd3ba4bd56725eb8c2916e7a9f29df34bef2
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630151541.86bb0ddabf9b.4092.144 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5aee1d6d4ab9a2cd5fb99ffa3b67bd3bcbad46cc2f6663dde04626e23f94786c
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630151929.86bb0ddabf9b.4092.146 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd372f6de219ef5c36d546f06292c34ad87d8af7573b5e260b8261b332c99c4a
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630152323.86bb0ddabf9b.4092.148 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c43ca82bc199c61deab2c9f2e1f615d8243eb0b1bacefa233818c677f1a88aa
3
+ size 8622