Check commited on
Commit
56680f4
Β·
1 Parent(s): 40b1526

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/rng_state.pth +2 -2
  6. model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/trainer_state.json +792 -3
  9. model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629995693.1376626/events.out.tfevents.1629995693.8e89bd551565.924.251 +3 -0
  11. model-bin/finetune/base/log/1629996128.4661825/events.out.tfevents.1629996129.8e89bd551565.924.253 +3 -0
  12. model-bin/finetune/base/log/1629996648.4800694/events.out.tfevents.1629996648.8e89bd551565.924.255 +3 -0
  13. model-bin/finetune/base/log/1629997104.0112085/events.out.tfevents.1629997104.8e89bd551565.924.257 +3 -0
  14. model-bin/finetune/base/log/1629997541.0580804/events.out.tfevents.1629997541.8e89bd551565.924.259 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629995693.8e89bd551565.924.250 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629996128.8e89bd551565.924.252 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629996648.8e89bd551565.924.254 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629997104.8e89bd551565.924.256 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629997541.8e89bd551565.924.258 +3 -0
model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f887ccc9e530275eabb942742b49945da5e396af357fcb4c6216a8b0e38aaa0
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18d1d0e7b18117085c5c6980f6665a7ddd2e685a7c21bedd75352320eae612e1
3
  size 722165393
model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa66840495953d2dcf8d0f972d70b80e8025406460af3c06559744b0b8e00069
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:491fbf5bec3ceab2b3029b1588b8a53cc7c6dc6f1a3b0f743dd919485a4809fb
3
  size 377909911
model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a1d7ee77a309d4b02f0e0b50a254c8d8994dd350eafd8040d4e8821c92cec5ce
3
- size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:863a206f5996b23822226dcc80d7f21f039bee94d1f6943796990274de49b558
3
+ size 14375
model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a7e782bf5b4614ad2f9ac21555ba4813411b4db165cdaa4d98d6f042e85cb7e
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52590b6b41ae98301e51f804dfdc3089d29db3b95cf917c1c4047b381e792adb
3
  size 559
model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1a6409ad26afd7a4d5286ea7d95c3da331c0a58c248fb8e74926edc29a9ff2b5
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da5ef322dcad50f2827ba3193e02d6f59c8fd27dd174d3cf7d38cbad2827c51d
3
  size 623
model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
- "epoch": 838.0,
5
- "global_step": 103918,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -231579,11 +231579,800 @@
231579
  "eval_steps_per_second": 0.686,
231580
  "eval_wer": 0.18322231169021302,
231581
  "step": 103918
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
231582
  }
231583
  ],
231584
  "max_steps": 620000,
231585
  "num_train_epochs": 5000,
231586
- "total_flos": 2.924503505825161e+20,
231587
  "trial_name": null,
231588
  "trial_params": null
231589
  }
 
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
+ "epoch": 842.995983935743,
5
+ "global_step": 104539,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
231579
  "eval_steps_per_second": 0.686,
231580
  "eval_wer": 0.18322231169021302,
231581
  "step": 103918
231582
+ },
231583
+ {
231584
+ "epoch": 838.02,
231585
+ "learning_rate": 8.351298076923077e-06,
231586
+ "loss": 0.2745,
231587
+ "step": 103920
231588
+ },
231589
+ {
231590
+ "epoch": 838.06,
231591
+ "learning_rate": 8.351217948717949e-06,
231592
+ "loss": 0.2572,
231593
+ "step": 103925
231594
+ },
231595
+ {
231596
+ "epoch": 838.1,
231597
+ "learning_rate": 8.351137820512822e-06,
231598
+ "loss": 0.3528,
231599
+ "step": 103930
231600
+ },
231601
+ {
231602
+ "epoch": 838.14,
231603
+ "learning_rate": 8.351057692307693e-06,
231604
+ "loss": 0.3604,
231605
+ "step": 103935
231606
+ },
231607
+ {
231608
+ "epoch": 838.18,
231609
+ "learning_rate": 8.350977564102564e-06,
231610
+ "loss": 0.7321,
231611
+ "step": 103940
231612
+ },
231613
+ {
231614
+ "epoch": 838.22,
231615
+ "learning_rate": 8.350897435897437e-06,
231616
+ "loss": 0.9929,
231617
+ "step": 103945
231618
+ },
231619
+ {
231620
+ "epoch": 838.26,
231621
+ "learning_rate": 8.350817307692309e-06,
231622
+ "loss": 0.2765,
231623
+ "step": 103950
231624
+ },
231625
+ {
231626
+ "epoch": 838.3,
231627
+ "learning_rate": 8.35073717948718e-06,
231628
+ "loss": 0.2527,
231629
+ "step": 103955
231630
+ },
231631
+ {
231632
+ "epoch": 838.34,
231633
+ "learning_rate": 8.350657051282051e-06,
231634
+ "loss": 0.3403,
231635
+ "step": 103960
231636
+ },
231637
+ {
231638
+ "epoch": 838.38,
231639
+ "learning_rate": 8.350576923076925e-06,
231640
+ "loss": 0.6165,
231641
+ "step": 103965
231642
+ },
231643
+ {
231644
+ "epoch": 838.42,
231645
+ "learning_rate": 8.350496794871796e-06,
231646
+ "loss": 0.9087,
231647
+ "step": 103970
231648
+ },
231649
+ {
231650
+ "epoch": 838.46,
231651
+ "learning_rate": 8.350416666666667e-06,
231652
+ "loss": 0.3288,
231653
+ "step": 103975
231654
+ },
231655
+ {
231656
+ "epoch": 838.5,
231657
+ "learning_rate": 8.350336538461539e-06,
231658
+ "loss": 0.3491,
231659
+ "step": 103980
231660
+ },
231661
+ {
231662
+ "epoch": 838.54,
231663
+ "learning_rate": 8.350256410256412e-06,
231664
+ "loss": 0.3221,
231665
+ "step": 103985
231666
+ },
231667
+ {
231668
+ "epoch": 838.58,
231669
+ "learning_rate": 8.350176282051283e-06,
231670
+ "loss": 0.6123,
231671
+ "step": 103990
231672
+ },
231673
+ {
231674
+ "epoch": 838.62,
231675
+ "learning_rate": 8.350096153846154e-06,
231676
+ "loss": 0.9183,
231677
+ "step": 103995
231678
+ },
231679
+ {
231680
+ "epoch": 838.66,
231681
+ "learning_rate": 8.350016025641027e-06,
231682
+ "loss": 0.3169,
231683
+ "step": 104000
231684
+ },
231685
+ {
231686
+ "epoch": 838.7,
231687
+ "learning_rate": 8.349935897435897e-06,
231688
+ "loss": 0.3932,
231689
+ "step": 104005
231690
+ },
231691
+ {
231692
+ "epoch": 838.74,
231693
+ "learning_rate": 8.34985576923077e-06,
231694
+ "loss": 0.3419,
231695
+ "step": 104010
231696
+ },
231697
+ {
231698
+ "epoch": 838.78,
231699
+ "learning_rate": 8.349775641025642e-06,
231700
+ "loss": 0.638,
231701
+ "step": 104015
231702
+ },
231703
+ {
231704
+ "epoch": 838.82,
231705
+ "learning_rate": 8.349695512820513e-06,
231706
+ "loss": 1.0426,
231707
+ "step": 104020
231708
+ },
231709
+ {
231710
+ "epoch": 838.86,
231711
+ "learning_rate": 8.349615384615384e-06,
231712
+ "loss": 0.2742,
231713
+ "step": 104025
231714
+ },
231715
+ {
231716
+ "epoch": 838.9,
231717
+ "learning_rate": 8.349535256410257e-06,
231718
+ "loss": 0.3108,
231719
+ "step": 104030
231720
+ },
231721
+ {
231722
+ "epoch": 838.94,
231723
+ "learning_rate": 8.349455128205129e-06,
231724
+ "loss": 0.3617,
231725
+ "step": 104035
231726
+ },
231727
+ {
231728
+ "epoch": 838.98,
231729
+ "learning_rate": 8.349375e-06,
231730
+ "loss": 0.6442,
231731
+ "step": 104040
231732
+ },
231733
+ {
231734
+ "epoch": 839.0,
231735
+ "eval_loss": 0.35273829102516174,
231736
+ "eval_runtime": 41.2938,
231737
+ "eval_samples_per_second": 20.342,
231738
+ "eval_steps_per_second": 0.654,
231739
+ "eval_wer": 0.17864609617186936,
231740
+ "step": 104042
231741
+ },
231742
+ {
231743
+ "epoch": 839.02,
231744
+ "learning_rate": 8.349294871794873e-06,
231745
+ "loss": 0.3248,
231746
+ "step": 104045
231747
+ },
231748
+ {
231749
+ "epoch": 839.06,
231750
+ "learning_rate": 8.349214743589744e-06,
231751
+ "loss": 0.2771,
231752
+ "step": 104050
231753
+ },
231754
+ {
231755
+ "epoch": 839.1,
231756
+ "learning_rate": 8.349134615384616e-06,
231757
+ "loss": 0.2696,
231758
+ "step": 104055
231759
+ },
231760
+ {
231761
+ "epoch": 839.14,
231762
+ "learning_rate": 8.349054487179487e-06,
231763
+ "loss": 0.3619,
231764
+ "step": 104060
231765
+ },
231766
+ {
231767
+ "epoch": 839.18,
231768
+ "learning_rate": 8.34897435897436e-06,
231769
+ "loss": 0.7768,
231770
+ "step": 104065
231771
+ },
231772
+ {
231773
+ "epoch": 839.22,
231774
+ "learning_rate": 8.348894230769232e-06,
231775
+ "loss": 0.8096,
231776
+ "step": 104070
231777
+ },
231778
+ {
231779
+ "epoch": 839.27,
231780
+ "learning_rate": 8.348814102564103e-06,
231781
+ "loss": 0.2898,
231782
+ "step": 104075
231783
+ },
231784
+ {
231785
+ "epoch": 839.31,
231786
+ "learning_rate": 8.348733974358974e-06,
231787
+ "loss": 0.2886,
231788
+ "step": 104080
231789
+ },
231790
+ {
231791
+ "epoch": 839.35,
231792
+ "learning_rate": 8.348653846153847e-06,
231793
+ "loss": 0.3345,
231794
+ "step": 104085
231795
+ },
231796
+ {
231797
+ "epoch": 839.39,
231798
+ "learning_rate": 8.348573717948719e-06,
231799
+ "loss": 0.7645,
231800
+ "step": 104090
231801
+ },
231802
+ {
231803
+ "epoch": 839.43,
231804
+ "learning_rate": 8.34849358974359e-06,
231805
+ "loss": 0.9751,
231806
+ "step": 104095
231807
+ },
231808
+ {
231809
+ "epoch": 839.47,
231810
+ "learning_rate": 8.348413461538463e-06,
231811
+ "loss": 0.2766,
231812
+ "step": 104100
231813
+ },
231814
+ {
231815
+ "epoch": 839.51,
231816
+ "learning_rate": 8.348333333333334e-06,
231817
+ "loss": 0.3699,
231818
+ "step": 104105
231819
+ },
231820
+ {
231821
+ "epoch": 839.55,
231822
+ "learning_rate": 8.348253205128206e-06,
231823
+ "loss": 0.4221,
231824
+ "step": 104110
231825
+ },
231826
+ {
231827
+ "epoch": 839.59,
231828
+ "learning_rate": 8.348173076923077e-06,
231829
+ "loss": 0.7732,
231830
+ "step": 104115
231831
+ },
231832
+ {
231833
+ "epoch": 839.63,
231834
+ "learning_rate": 8.34809294871795e-06,
231835
+ "loss": 0.8807,
231836
+ "step": 104120
231837
+ },
231838
+ {
231839
+ "epoch": 839.67,
231840
+ "learning_rate": 8.348012820512822e-06,
231841
+ "loss": 0.2793,
231842
+ "step": 104125
231843
+ },
231844
+ {
231845
+ "epoch": 839.71,
231846
+ "learning_rate": 8.347932692307693e-06,
231847
+ "loss": 0.2918,
231848
+ "step": 104130
231849
+ },
231850
+ {
231851
+ "epoch": 839.75,
231852
+ "learning_rate": 8.347852564102564e-06,
231853
+ "loss": 0.3833,
231854
+ "step": 104135
231855
+ },
231856
+ {
231857
+ "epoch": 839.79,
231858
+ "learning_rate": 8.347772435897437e-06,
231859
+ "loss": 0.7802,
231860
+ "step": 104140
231861
+ },
231862
+ {
231863
+ "epoch": 839.83,
231864
+ "learning_rate": 8.347692307692309e-06,
231865
+ "loss": 0.8053,
231866
+ "step": 104145
231867
+ },
231868
+ {
231869
+ "epoch": 839.87,
231870
+ "learning_rate": 8.34761217948718e-06,
231871
+ "loss": 0.362,
231872
+ "step": 104150
231873
+ },
231874
+ {
231875
+ "epoch": 839.91,
231876
+ "learning_rate": 8.347532051282053e-06,
231877
+ "loss": 0.3359,
231878
+ "step": 104155
231879
+ },
231880
+ {
231881
+ "epoch": 839.95,
231882
+ "learning_rate": 8.347451923076923e-06,
231883
+ "loss": 0.3976,
231884
+ "step": 104160
231885
+ },
231886
+ {
231887
+ "epoch": 839.99,
231888
+ "learning_rate": 8.347371794871796e-06,
231889
+ "loss": 0.8711,
231890
+ "step": 104165
231891
+ },
231892
+ {
231893
+ "epoch": 840.0,
231894
+ "eval_loss": 0.42207542061805725,
231895
+ "eval_runtime": 40.7022,
231896
+ "eval_samples_per_second": 20.638,
231897
+ "eval_steps_per_second": 0.663,
231898
+ "eval_wer": 0.18555353901996371,
231899
+ "step": 104166
231900
+ },
231901
+ {
231902
+ "epoch": 833.03,
231903
+ "learning_rate": 8.347291666666667e-06,
231904
+ "loss": 0.347,
231905
+ "step": 104170
231906
+ },
231907
+ {
231908
+ "epoch": 833.07,
231909
+ "learning_rate": 8.347211538461539e-06,
231910
+ "loss": 0.3221,
231911
+ "step": 104175
231912
+ },
231913
+ {
231914
+ "epoch": 833.11,
231915
+ "learning_rate": 8.34713141025641e-06,
231916
+ "loss": 0.2602,
231917
+ "step": 104180
231918
+ },
231919
+ {
231920
+ "epoch": 833.15,
231921
+ "learning_rate": 8.347051282051283e-06,
231922
+ "loss": 0.3825,
231923
+ "step": 104185
231924
+ },
231925
+ {
231926
+ "epoch": 833.19,
231927
+ "learning_rate": 8.346971153846154e-06,
231928
+ "loss": 0.8882,
231929
+ "step": 104190
231930
+ },
231931
+ {
231932
+ "epoch": 833.23,
231933
+ "learning_rate": 8.346891025641026e-06,
231934
+ "loss": 0.7291,
231935
+ "step": 104195
231936
+ },
231937
+ {
231938
+ "epoch": 833.27,
231939
+ "learning_rate": 8.346810897435899e-06,
231940
+ "loss": 0.312,
231941
+ "step": 104200
231942
+ },
231943
+ {
231944
+ "epoch": 833.31,
231945
+ "learning_rate": 8.34673076923077e-06,
231946
+ "loss": 0.2868,
231947
+ "step": 104205
231948
+ },
231949
+ {
231950
+ "epoch": 833.35,
231951
+ "learning_rate": 8.346650641025641e-06,
231952
+ "loss": 0.4478,
231953
+ "step": 104210
231954
+ },
231955
+ {
231956
+ "epoch": 833.39,
231957
+ "learning_rate": 8.346570512820513e-06,
231958
+ "loss": 0.8855,
231959
+ "step": 104215
231960
+ },
231961
+ {
231962
+ "epoch": 833.43,
231963
+ "learning_rate": 8.346490384615386e-06,
231964
+ "loss": 0.7151,
231965
+ "step": 104220
231966
+ },
231967
+ {
231968
+ "epoch": 833.47,
231969
+ "learning_rate": 8.346410256410257e-06,
231970
+ "loss": 0.3058,
231971
+ "step": 104225
231972
+ },
231973
+ {
231974
+ "epoch": 833.51,
231975
+ "learning_rate": 8.346330128205129e-06,
231976
+ "loss": 0.3211,
231977
+ "step": 104230
231978
+ },
231979
+ {
231980
+ "epoch": 833.55,
231981
+ "learning_rate": 8.34625e-06,
231982
+ "loss": 0.4314,
231983
+ "step": 104235
231984
+ },
231985
+ {
231986
+ "epoch": 833.59,
231987
+ "learning_rate": 8.346169871794873e-06,
231988
+ "loss": 0.9153,
231989
+ "step": 104240
231990
+ },
231991
+ {
231992
+ "epoch": 833.63,
231993
+ "learning_rate": 8.346089743589744e-06,
231994
+ "loss": 0.6461,
231995
+ "step": 104245
231996
+ },
231997
+ {
231998
+ "epoch": 833.67,
231999
+ "learning_rate": 8.346009615384616e-06,
232000
+ "loss": 0.3166,
232001
+ "step": 104250
232002
+ },
232003
+ {
232004
+ "epoch": 833.71,
232005
+ "learning_rate": 8.345929487179489e-06,
232006
+ "loss": 0.3198,
232007
+ "step": 104255
232008
+ },
232009
+ {
232010
+ "epoch": 833.75,
232011
+ "learning_rate": 8.34584935897436e-06,
232012
+ "loss": 0.4173,
232013
+ "step": 104260
232014
+ },
232015
+ {
232016
+ "epoch": 833.79,
232017
+ "learning_rate": 8.345769230769231e-06,
232018
+ "loss": 0.9187,
232019
+ "step": 104265
232020
+ },
232021
+ {
232022
+ "epoch": 833.83,
232023
+ "learning_rate": 8.345689102564103e-06,
232024
+ "loss": 0.6422,
232025
+ "step": 104270
232026
+ },
232027
+ {
232028
+ "epoch": 833.87,
232029
+ "learning_rate": 8.345608974358976e-06,
232030
+ "loss": 0.2888,
232031
+ "step": 104275
232032
+ },
232033
+ {
232034
+ "epoch": 833.91,
232035
+ "learning_rate": 8.345528846153847e-06,
232036
+ "loss": 0.3323,
232037
+ "step": 104280
232038
+ },
232039
+ {
232040
+ "epoch": 833.95,
232041
+ "learning_rate": 8.345448717948719e-06,
232042
+ "loss": 0.4291,
232043
+ "step": 104285
232044
+ },
232045
+ {
232046
+ "epoch": 833.99,
232047
+ "learning_rate": 8.34536858974359e-06,
232048
+ "loss": 1.0071,
232049
+ "step": 104290
232050
+ },
232051
+ {
232052
+ "epoch": 834.0,
232053
+ "eval_loss": 0.3745848536491394,
232054
+ "eval_runtime": 40.2099,
232055
+ "eval_samples_per_second": 20.841,
232056
+ "eval_steps_per_second": 0.671,
232057
+ "eval_wer": 0.18355542539171182,
232058
+ "step": 104291
232059
+ },
232060
+ {
232061
+ "epoch": 841.03,
232062
+ "learning_rate": 8.345288461538463e-06,
232063
+ "loss": 0.3266,
232064
+ "step": 104295
232065
+ },
232066
+ {
232067
+ "epoch": 841.07,
232068
+ "learning_rate": 8.345208333333334e-06,
232069
+ "loss": 0.3229,
232070
+ "step": 104300
232071
+ },
232072
+ {
232073
+ "epoch": 841.11,
232074
+ "learning_rate": 8.345128205128206e-06,
232075
+ "loss": 0.3198,
232076
+ "step": 104305
232077
+ },
232078
+ {
232079
+ "epoch": 841.15,
232080
+ "learning_rate": 8.345048076923079e-06,
232081
+ "loss": 0.3572,
232082
+ "step": 104310
232083
+ },
232084
+ {
232085
+ "epoch": 841.19,
232086
+ "learning_rate": 8.344967948717948e-06,
232087
+ "loss": 0.8764,
232088
+ "step": 104315
232089
+ },
232090
+ {
232091
+ "epoch": 841.23,
232092
+ "learning_rate": 8.344887820512821e-06,
232093
+ "loss": 0.6498,
232094
+ "step": 104320
232095
+ },
232096
+ {
232097
+ "epoch": 841.27,
232098
+ "learning_rate": 8.344807692307693e-06,
232099
+ "loss": 0.2689,
232100
+ "step": 104325
232101
+ },
232102
+ {
232103
+ "epoch": 841.31,
232104
+ "learning_rate": 8.344727564102564e-06,
232105
+ "loss": 0.2865,
232106
+ "step": 104330
232107
+ },
232108
+ {
232109
+ "epoch": 841.35,
232110
+ "learning_rate": 8.344647435897436e-06,
232111
+ "loss": 0.3679,
232112
+ "step": 104335
232113
+ },
232114
+ {
232115
+ "epoch": 841.39,
232116
+ "learning_rate": 8.344567307692309e-06,
232117
+ "loss": 0.9102,
232118
+ "step": 104340
232119
+ },
232120
+ {
232121
+ "epoch": 841.43,
232122
+ "learning_rate": 8.34448717948718e-06,
232123
+ "loss": 0.6518,
232124
+ "step": 104345
232125
+ },
232126
+ {
232127
+ "epoch": 841.47,
232128
+ "learning_rate": 8.344407051282051e-06,
232129
+ "loss": 0.3188,
232130
+ "step": 104350
232131
+ },
232132
+ {
232133
+ "epoch": 841.51,
232134
+ "learning_rate": 8.344326923076924e-06,
232135
+ "loss": 0.2554,
232136
+ "step": 104355
232137
+ },
232138
+ {
232139
+ "epoch": 841.55,
232140
+ "learning_rate": 8.344246794871796e-06,
232141
+ "loss": 0.3999,
232142
+ "step": 104360
232143
+ },
232144
+ {
232145
+ "epoch": 841.59,
232146
+ "learning_rate": 8.344166666666667e-06,
232147
+ "loss": 0.8557,
232148
+ "step": 104365
232149
+ },
232150
+ {
232151
+ "epoch": 841.63,
232152
+ "learning_rate": 8.344086538461538e-06,
232153
+ "loss": 0.6551,
232154
+ "step": 104370
232155
+ },
232156
+ {
232157
+ "epoch": 841.67,
232158
+ "learning_rate": 8.344006410256412e-06,
232159
+ "loss": 0.2773,
232160
+ "step": 104375
232161
+ },
232162
+ {
232163
+ "epoch": 841.71,
232164
+ "learning_rate": 8.343926282051283e-06,
232165
+ "loss": 0.258,
232166
+ "step": 104380
232167
+ },
232168
+ {
232169
+ "epoch": 841.76,
232170
+ "learning_rate": 8.343846153846154e-06,
232171
+ "loss": 0.3869,
232172
+ "step": 104385
232173
+ },
232174
+ {
232175
+ "epoch": 841.8,
232176
+ "learning_rate": 8.343766025641026e-06,
232177
+ "loss": 0.967,
232178
+ "step": 104390
232179
+ },
232180
+ {
232181
+ "epoch": 841.84,
232182
+ "learning_rate": 8.343685897435899e-06,
232183
+ "loss": 0.5489,
232184
+ "step": 104395
232185
+ },
232186
+ {
232187
+ "epoch": 841.88,
232188
+ "learning_rate": 8.34360576923077e-06,
232189
+ "loss": 0.2847,
232190
+ "step": 104400
232191
+ },
232192
+ {
232193
+ "epoch": 841.92,
232194
+ "learning_rate": 8.343525641025641e-06,
232195
+ "loss": 0.3259,
232196
+ "step": 104405
232197
+ },
232198
+ {
232199
+ "epoch": 841.96,
232200
+ "learning_rate": 8.343445512820514e-06,
232201
+ "loss": 0.4115,
232202
+ "step": 104410
232203
+ },
232204
+ {
232205
+ "epoch": 842.0,
232206
+ "learning_rate": 8.343365384615386e-06,
232207
+ "loss": 1.1216,
232208
+ "step": 104415
232209
+ },
232210
+ {
232211
+ "epoch": 842.0,
232212
+ "eval_loss": 0.3493131697177887,
232213
+ "eval_runtime": 38.6328,
232214
+ "eval_samples_per_second": 21.666,
232215
+ "eval_steps_per_second": 0.699,
232216
+ "eval_wer": 0.1871700296521299,
232217
+ "step": 104415
232218
+ },
232219
+ {
232220
+ "epoch": 842.04,
232221
+ "learning_rate": 8.343285256410257e-06,
232222
+ "loss": 0.3497,
232223
+ "step": 104420
232224
+ },
232225
+ {
232226
+ "epoch": 842.08,
232227
+ "learning_rate": 8.343205128205128e-06,
232228
+ "loss": 0.318,
232229
+ "step": 104425
232230
+ },
232231
+ {
232232
+ "epoch": 842.12,
232233
+ "learning_rate": 8.343125000000002e-06,
232234
+ "loss": 0.3279,
232235
+ "step": 104430
232236
+ },
232237
+ {
232238
+ "epoch": 842.16,
232239
+ "learning_rate": 8.343044871794871e-06,
232240
+ "loss": 0.4165,
232241
+ "step": 104435
232242
+ },
232243
+ {
232244
+ "epoch": 842.2,
232245
+ "learning_rate": 8.342964743589744e-06,
232246
+ "loss": 1.2737,
232247
+ "step": 104440
232248
+ },
232249
+ {
232250
+ "epoch": 842.24,
232251
+ "learning_rate": 8.342884615384617e-06,
232252
+ "loss": 0.3312,
232253
+ "step": 104445
232254
+ },
232255
+ {
232256
+ "epoch": 842.28,
232257
+ "learning_rate": 8.342804487179487e-06,
232258
+ "loss": 0.2809,
232259
+ "step": 104450
232260
+ },
232261
+ {
232262
+ "epoch": 842.32,
232263
+ "learning_rate": 8.34272435897436e-06,
232264
+ "loss": 0.2927,
232265
+ "step": 104455
232266
+ },
232267
+ {
232268
+ "epoch": 842.36,
232269
+ "learning_rate": 8.342644230769231e-06,
232270
+ "loss": 0.4408,
232271
+ "step": 104460
232272
+ },
232273
+ {
232274
+ "epoch": 842.4,
232275
+ "learning_rate": 8.342564102564103e-06,
232276
+ "loss": 1.1952,
232277
+ "step": 104465
232278
+ },
232279
+ {
232280
+ "epoch": 842.44,
232281
+ "learning_rate": 8.342483974358974e-06,
232282
+ "loss": 0.5169,
232283
+ "step": 104470
232284
+ },
232285
+ {
232286
+ "epoch": 842.48,
232287
+ "learning_rate": 8.342403846153847e-06,
232288
+ "loss": 0.2839,
232289
+ "step": 104475
232290
+ },
232291
+ {
232292
+ "epoch": 842.52,
232293
+ "learning_rate": 8.342323717948719e-06,
232294
+ "loss": 0.3011,
232295
+ "step": 104480
232296
+ },
232297
+ {
232298
+ "epoch": 842.56,
232299
+ "learning_rate": 8.34224358974359e-06,
232300
+ "loss": 0.4053,
232301
+ "step": 104485
232302
+ },
232303
+ {
232304
+ "epoch": 842.6,
232305
+ "learning_rate": 8.342163461538461e-06,
232306
+ "loss": 1.216,
232307
+ "step": 104490
232308
+ },
232309
+ {
232310
+ "epoch": 842.64,
232311
+ "learning_rate": 8.342083333333334e-06,
232312
+ "loss": 0.3982,
232313
+ "step": 104495
232314
+ },
232315
+ {
232316
+ "epoch": 842.68,
232317
+ "learning_rate": 8.342003205128206e-06,
232318
+ "loss": 0.3052,
232319
+ "step": 104500
232320
+ },
232321
+ {
232322
+ "epoch": 842.72,
232323
+ "learning_rate": 8.341923076923077e-06,
232324
+ "loss": 0.3062,
232325
+ "step": 104505
232326
+ },
232327
+ {
232328
+ "epoch": 842.76,
232329
+ "learning_rate": 8.34184294871795e-06,
232330
+ "loss": 0.4384,
232331
+ "step": 104510
232332
+ },
232333
+ {
232334
+ "epoch": 842.8,
232335
+ "learning_rate": 8.341762820512821e-06,
232336
+ "loss": 1.0651,
232337
+ "step": 104515
232338
+ },
232339
+ {
232340
+ "epoch": 842.84,
232341
+ "learning_rate": 8.341682692307693e-06,
232342
+ "loss": 0.3653,
232343
+ "step": 104520
232344
+ },
232345
+ {
232346
+ "epoch": 842.88,
232347
+ "learning_rate": 8.341602564102564e-06,
232348
+ "loss": 0.2773,
232349
+ "step": 104525
232350
+ },
232351
+ {
232352
+ "epoch": 842.92,
232353
+ "learning_rate": 8.341522435897437e-06,
232354
+ "loss": 0.3293,
232355
+ "step": 104530
232356
+ },
232357
+ {
232358
+ "epoch": 842.96,
232359
+ "learning_rate": 8.341442307692309e-06,
232360
+ "loss": 0.5005,
232361
+ "step": 104535
232362
+ },
232363
+ {
232364
+ "epoch": 843.0,
232365
+ "eval_loss": 0.38071903586387634,
232366
+ "eval_runtime": 39.7203,
232367
+ "eval_samples_per_second": 21.098,
232368
+ "eval_steps_per_second": 0.68,
232369
+ "eval_wer": 0.18273625239569513,
232370
+ "step": 104539
232371
  }
232372
  ],
232373
  "max_steps": 620000,
232374
  "num_train_epochs": 5000,
232375
+ "total_flos": 2.9420435882603774e+20,
232376
  "trial_name": null,
232377
  "trial_params": null
232378
  }
model-bin/finetune/base/{checkpoint-103918 β†’ checkpoint-104539}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629995693.1376626/events.out.tfevents.1629995693.8e89bd551565.924.251 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e4bc3acb124c42d484dbabf88d21e231d725f982e77f48bdbd37d1203e02155
3
+ size 4194
model-bin/finetune/base/log/1629996128.4661825/events.out.tfevents.1629996129.8e89bd551565.924.253 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a613f4d8efe43ba38dcc6ecede800e02f85b5daaab321050c023bcbb2b0a55f
3
+ size 4194
model-bin/finetune/base/log/1629996648.4800694/events.out.tfevents.1629996648.8e89bd551565.924.255 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32fbff97f311ec84719302a8a9eae8ccb82d2e34a2816795fefbc1ac2b6bc286
3
+ size 4194
model-bin/finetune/base/log/1629997104.0112085/events.out.tfevents.1629997104.8e89bd551565.924.257 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36b3da19efc5e1df9126ba79f7c6bd5d788eededd35a568c7b72e39438dc5994
3
+ size 4194
model-bin/finetune/base/log/1629997541.0580804/events.out.tfevents.1629997541.8e89bd551565.924.259 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e7ae8d8e64747f877d97e0ecbbffab661c74155f45d449c94f550f913c79e70
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629995693.8e89bd551565.924.250 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0242544dd702433852891d50f4cc660aa6a8fac0f3ac1abb63c5ddf84bcc6a91
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629996128.8e89bd551565.924.252 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0968a8a9eb32f58a83fd93bea19f510bdaee87a50fd7dcb19e74c4259f13f44
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629996648.8e89bd551565.924.254 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cbf4ade00deacc89491ea814480b14fe1704e8919bda47a88fedcd3651890da
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629997104.8e89bd551565.924.256 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fe8cc646a2439c9d920c71bf82c83b02b1c5b40ad2fbe0525af7ee250246b7d
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629997541.8e89bd551565.924.258 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94dfcbb96d61aec4c4ad0d0460a30fe18b7f938b5f40c7320d730ebb5ca9ef84
3
+ size 8462