Check commited on
Commit
096355b
Β·
1 Parent(s): cac7488

"auto-commit"

Browse files
Files changed (28) hide show
  1. model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88857}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88857}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88857}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88857}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88857}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88857}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88857}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88857}/trainer_state.json +641 -5
  9. model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88857}/training_args.bin +0 -0
  10. model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88981}/config.json +0 -0
  11. model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88981}/optimizer.pt +1 -1
  12. model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88981}/preprocessor_config.json +0 -0
  13. model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88981}/pytorch_model.bin +1 -1
  14. model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88981}/rng_state.pth +2 -2
  15. model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88981}/scaler.pt +1 -1
  16. model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88981}/scheduler.pt +1 -1
  17. model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88981}/trainer_state.json +0 -0
  18. model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88981}/training_args.bin +0 -0
  19. model-bin/finetune/base/log/1629939465.6417572/events.out.tfevents.1629939465.8e89bd551565.924.1 +3 -0
  20. model-bin/finetune/base/log/1629939905.367717/events.out.tfevents.1629939905.8e89bd551565.924.3 +3 -0
  21. model-bin/finetune/base/log/1629940450.426258/events.out.tfevents.1629940450.8e89bd551565.924.5 +3 -0
  22. model-bin/finetune/base/log/1629940922.8454914/events.out.tfevents.1629940922.8e89bd551565.924.7 +3 -0
  23. model-bin/finetune/base/log/1629941350.450414/events.out.tfevents.1629941350.8e89bd551565.924.9 +3 -0
  24. model-bin/finetune/base/log/events.out.tfevents.1629939465.8e89bd551565.924.0 +3 -0
  25. model-bin/finetune/base/log/events.out.tfevents.1629939905.8e89bd551565.924.2 +3 -0
  26. model-bin/finetune/base/log/events.out.tfevents.1629940450.8e89bd551565.924.4 +3 -0
  27. model-bin/finetune/base/log/events.out.tfevents.1629940922.8e89bd551565.924.6 +3 -0
  28. model-bin/finetune/base/log/events.out.tfevents.1629941350.8e89bd551565.924.8 +3 -0
model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88857}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88857}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fac09e3d3892f68c97eba3e9f42969e402ccb83581834ce4220ef00c901e5ba6
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03ab89beca2f81624334757a47196f2700ffb64a09d11ad89369efc03d60446c
3
  size 722165393
model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88857}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88857}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1bfbc399f7e6345794f7e9c2f12f77dc0abb92b9ddca2b8f4a0d98a910f1596a
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f743c4a7f5e90dc62c3a3443673fcb9e237965bd397b6783bf1f839fc38b5c10
3
  size 377909911
model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88857}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:967ade1e5ebd4405f020ecc7951719de991516fbdf76457cd993be421aabfdd1
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12bb072db3e1f88523968032baaced85ff7adeedb4a8a975c73385b0edf9e365
3
  size 14503
model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88857}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da825b21776e1bdfcebabe3acb018ec3e5d206390ffcfc17db0f26cbc8bed857
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4cd92943fc5a165b57777669567f5fddb9a9bb4e6eef2e77618b7c86a10f5fc
3
  size 559
model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88857}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c350197eb47e928c3abd37ad8f5c23f1484357bfcbe5a3ed1bce6c295318afcc
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:320a18356ec6893f73250de1ff44f61371d24c656b47dfc9e100b1a509e1b7ee
3
  size 623
model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88857}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
- "best_metric": 0.1822989349981638,
3
- "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-79148",
4
- "epoch": 711.995983935743,
5
- "global_step": 88359,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -211782,11 +211782,647 @@
211782
  "eval_steps_per_second": 0.656,
211783
  "eval_wer": 0.189195053156867,
211784
  "step": 88359
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
211785
  }
211786
  ],
211787
  "max_steps": 620000,
211788
  "num_train_epochs": 5000,
211789
- "total_flos": 2.486625232987812e+20,
211790
  "trial_name": null,
211791
  "trial_params": null
211792
  }
 
1
  {
2
+ "best_metric": 0.17731766728757703,
3
+ "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-88857",
4
+ "epoch": 715.995983935743,
5
+ "global_step": 88857,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
211782
  "eval_steps_per_second": 0.656,
211783
  "eval_wer": 0.189195053156867,
211784
  "step": 88359
211785
+ },
211786
+ {
211787
+ "epoch": 706.01,
211788
+ "learning_rate": 8.600576923076924e-06,
211789
+ "loss": 0.3692,
211790
+ "step": 88360
211791
+ },
211792
+ {
211793
+ "epoch": 706.05,
211794
+ "learning_rate": 8.600496794871795e-06,
211795
+ "loss": 0.2983,
211796
+ "step": 88365
211797
+ },
211798
+ {
211799
+ "epoch": 706.09,
211800
+ "learning_rate": 8.600416666666668e-06,
211801
+ "loss": 0.3386,
211802
+ "step": 88370
211803
+ },
211804
+ {
211805
+ "epoch": 706.13,
211806
+ "learning_rate": 8.60033653846154e-06,
211807
+ "loss": 0.3999,
211808
+ "step": 88375
211809
+ },
211810
+ {
211811
+ "epoch": 706.17,
211812
+ "learning_rate": 8.600256410256411e-06,
211813
+ "loss": 0.5919,
211814
+ "step": 88380
211815
+ },
211816
+ {
211817
+ "epoch": 706.21,
211818
+ "learning_rate": 8.600176282051282e-06,
211819
+ "loss": 1.0348,
211820
+ "step": 88385
211821
+ },
211822
+ {
211823
+ "epoch": 706.25,
211824
+ "learning_rate": 8.600096153846155e-06,
211825
+ "loss": 0.2688,
211826
+ "step": 88390
211827
+ },
211828
+ {
211829
+ "epoch": 706.29,
211830
+ "learning_rate": 8.600016025641025e-06,
211831
+ "loss": 0.2694,
211832
+ "step": 88395
211833
+ },
211834
+ {
211835
+ "epoch": 706.33,
211836
+ "learning_rate": 8.599935897435898e-06,
211837
+ "loss": 0.2873,
211838
+ "step": 88400
211839
+ },
211840
+ {
211841
+ "epoch": 706.37,
211842
+ "learning_rate": 8.599855769230771e-06,
211843
+ "loss": 0.5365,
211844
+ "step": 88405
211845
+ },
211846
+ {
211847
+ "epoch": 706.41,
211848
+ "learning_rate": 8.599775641025643e-06,
211849
+ "loss": 1.1916,
211850
+ "step": 88410
211851
+ },
211852
+ {
211853
+ "epoch": 706.45,
211854
+ "learning_rate": 8.599695512820514e-06,
211855
+ "loss": 0.3537,
211856
+ "step": 88415
211857
+ },
211858
+ {
211859
+ "epoch": 706.49,
211860
+ "learning_rate": 8.599615384615385e-06,
211861
+ "loss": 0.334,
211862
+ "step": 88420
211863
+ },
211864
+ {
211865
+ "epoch": 706.53,
211866
+ "learning_rate": 8.599535256410258e-06,
211867
+ "loss": 0.3377,
211868
+ "step": 88425
211869
+ },
211870
+ {
211871
+ "epoch": 706.57,
211872
+ "learning_rate": 8.599455128205128e-06,
211873
+ "loss": 0.5144,
211874
+ "step": 88430
211875
+ },
211876
+ {
211877
+ "epoch": 706.61,
211878
+ "learning_rate": 8.599375000000001e-06,
211879
+ "loss": 1.357,
211880
+ "step": 88435
211881
+ },
211882
+ {
211883
+ "epoch": 706.65,
211884
+ "learning_rate": 8.599294871794872e-06,
211885
+ "loss": 0.3338,
211886
+ "step": 88440
211887
+ },
211888
+ {
211889
+ "epoch": 706.69,
211890
+ "learning_rate": 8.599214743589744e-06,
211891
+ "loss": 0.3426,
211892
+ "step": 88445
211893
+ },
211894
+ {
211895
+ "epoch": 706.73,
211896
+ "learning_rate": 8.599134615384615e-06,
211897
+ "loss": 0.3479,
211898
+ "step": 88450
211899
+ },
211900
+ {
211901
+ "epoch": 706.77,
211902
+ "learning_rate": 8.599054487179488e-06,
211903
+ "loss": 0.584,
211904
+ "step": 88455
211905
+ },
211906
+ {
211907
+ "epoch": 706.81,
211908
+ "learning_rate": 8.59897435897436e-06,
211909
+ "loss": 1.4261,
211910
+ "step": 88460
211911
+ },
211912
+ {
211913
+ "epoch": 706.85,
211914
+ "learning_rate": 8.59889423076923e-06,
211915
+ "loss": 0.2559,
211916
+ "step": 88465
211917
+ },
211918
+ {
211919
+ "epoch": 706.89,
211920
+ "learning_rate": 8.598814102564104e-06,
211921
+ "loss": 0.3294,
211922
+ "step": 88470
211923
+ },
211924
+ {
211925
+ "epoch": 706.93,
211926
+ "learning_rate": 8.598733974358975e-06,
211927
+ "loss": 0.3859,
211928
+ "step": 88475
211929
+ },
211930
+ {
211931
+ "epoch": 706.97,
211932
+ "learning_rate": 8.598653846153847e-06,
211933
+ "loss": 0.592,
211934
+ "step": 88480
211935
+ },
211936
+ {
211937
+ "epoch": 707.0,
211938
+ "eval_loss": 0.40724536776542664,
211939
+ "eval_runtime": 38.5536,
211940
+ "eval_samples_per_second": 21.762,
211941
+ "eval_steps_per_second": 0.7,
211942
+ "eval_wer": 0.1872142696545005,
211943
+ "step": 88484
211944
+ },
211945
+ {
211946
+ "epoch": 713.01,
211947
+ "learning_rate": 8.598573717948718e-06,
211948
+ "loss": 0.3631,
211949
+ "step": 88485
211950
+ },
211951
+ {
211952
+ "epoch": 713.05,
211953
+ "learning_rate": 8.598493589743591e-06,
211954
+ "loss": 0.3241,
211955
+ "step": 88490
211956
+ },
211957
+ {
211958
+ "epoch": 713.09,
211959
+ "learning_rate": 8.598413461538462e-06,
211960
+ "loss": 0.3114,
211961
+ "step": 88495
211962
+ },
211963
+ {
211964
+ "epoch": 713.13,
211965
+ "learning_rate": 8.598333333333334e-06,
211966
+ "loss": 0.3329,
211967
+ "step": 88500
211968
+ },
211969
+ {
211970
+ "epoch": 713.17,
211971
+ "learning_rate": 8.598253205128207e-06,
211972
+ "loss": 0.5927,
211973
+ "step": 88505
211974
+ },
211975
+ {
211976
+ "epoch": 713.21,
211977
+ "learning_rate": 8.598173076923078e-06,
211978
+ "loss": 1.2067,
211979
+ "step": 88510
211980
+ },
211981
+ {
211982
+ "epoch": 713.25,
211983
+ "learning_rate": 8.59809294871795e-06,
211984
+ "loss": 0.3103,
211985
+ "step": 88515
211986
+ },
211987
+ {
211988
+ "epoch": 713.29,
211989
+ "learning_rate": 8.598012820512821e-06,
211990
+ "loss": 0.3437,
211991
+ "step": 88520
211992
+ },
211993
+ {
211994
+ "epoch": 713.33,
211995
+ "learning_rate": 8.597932692307694e-06,
211996
+ "loss": 0.3895,
211997
+ "step": 88525
211998
+ },
211999
+ {
212000
+ "epoch": 713.37,
212001
+ "learning_rate": 8.597852564102565e-06,
212002
+ "loss": 0.6844,
212003
+ "step": 88530
212004
+ },
212005
+ {
212006
+ "epoch": 713.41,
212007
+ "learning_rate": 8.597772435897437e-06,
212008
+ "loss": 1.1051,
212009
+ "step": 88535
212010
+ },
212011
+ {
212012
+ "epoch": 713.45,
212013
+ "learning_rate": 8.597692307692308e-06,
212014
+ "loss": 0.2844,
212015
+ "step": 88540
212016
+ },
212017
+ {
212018
+ "epoch": 713.49,
212019
+ "learning_rate": 8.597612179487181e-06,
212020
+ "loss": 0.2973,
212021
+ "step": 88545
212022
+ },
212023
+ {
212024
+ "epoch": 713.53,
212025
+ "learning_rate": 8.59753205128205e-06,
212026
+ "loss": 0.5913,
212027
+ "step": 88550
212028
+ },
212029
+ {
212030
+ "epoch": 713.57,
212031
+ "learning_rate": 8.597451923076924e-06,
212032
+ "loss": 0.5365,
212033
+ "step": 88555
212034
+ },
212035
+ {
212036
+ "epoch": 713.61,
212037
+ "learning_rate": 8.597371794871797e-06,
212038
+ "loss": 1.1011,
212039
+ "step": 88560
212040
+ },
212041
+ {
212042
+ "epoch": 713.65,
212043
+ "learning_rate": 8.597291666666666e-06,
212044
+ "loss": 0.304,
212045
+ "step": 88565
212046
+ },
212047
+ {
212048
+ "epoch": 713.69,
212049
+ "learning_rate": 8.59721153846154e-06,
212050
+ "loss": 0.3349,
212051
+ "step": 88570
212052
+ },
212053
+ {
212054
+ "epoch": 713.73,
212055
+ "learning_rate": 8.597131410256411e-06,
212056
+ "loss": 0.3495,
212057
+ "step": 88575
212058
+ },
212059
+ {
212060
+ "epoch": 713.77,
212061
+ "learning_rate": 8.597051282051282e-06,
212062
+ "loss": 0.5413,
212063
+ "step": 88580
212064
+ },
212065
+ {
212066
+ "epoch": 713.81,
212067
+ "learning_rate": 8.596971153846154e-06,
212068
+ "loss": 1.1389,
212069
+ "step": 88585
212070
+ },
212071
+ {
212072
+ "epoch": 713.85,
212073
+ "learning_rate": 8.596891025641027e-06,
212074
+ "loss": 0.3478,
212075
+ "step": 88590
212076
+ },
212077
+ {
212078
+ "epoch": 713.9,
212079
+ "learning_rate": 8.596810897435898e-06,
212080
+ "loss": 0.3048,
212081
+ "step": 88595
212082
+ },
212083
+ {
212084
+ "epoch": 713.94,
212085
+ "learning_rate": 8.59673076923077e-06,
212086
+ "loss": 0.3369,
212087
+ "step": 88600
212088
+ },
212089
+ {
212090
+ "epoch": 713.98,
212091
+ "learning_rate": 8.59665064102564e-06,
212092
+ "loss": 0.6376,
212093
+ "step": 88605
212094
+ },
212095
+ {
212096
+ "epoch": 714.0,
212097
+ "eval_loss": 0.38512417674064636,
212098
+ "eval_runtime": 39.6255,
212099
+ "eval_samples_per_second": 21.198,
212100
+ "eval_steps_per_second": 0.681,
212101
+ "eval_wer": 0.18599016080475805,
212102
+ "step": 88608
212103
+ },
212104
+ {
212105
+ "epoch": 708.02,
212106
+ "learning_rate": 8.596570512820514e-06,
212107
+ "loss": 0.3381,
212108
+ "step": 88610
212109
+ },
212110
+ {
212111
+ "epoch": 708.06,
212112
+ "learning_rate": 8.596490384615385e-06,
212113
+ "loss": 0.3083,
212114
+ "step": 88615
212115
+ },
212116
+ {
212117
+ "epoch": 708.1,
212118
+ "learning_rate": 8.596410256410257e-06,
212119
+ "loss": 0.3253,
212120
+ "step": 88620
212121
+ },
212122
+ {
212123
+ "epoch": 708.14,
212124
+ "learning_rate": 8.59633012820513e-06,
212125
+ "loss": 0.4044,
212126
+ "step": 88625
212127
+ },
212128
+ {
212129
+ "epoch": 708.18,
212130
+ "learning_rate": 8.596250000000001e-06,
212131
+ "loss": 0.6699,
212132
+ "step": 88630
212133
+ },
212134
+ {
212135
+ "epoch": 708.22,
212136
+ "learning_rate": 8.596169871794872e-06,
212137
+ "loss": 0.958,
212138
+ "step": 88635
212139
+ },
212140
+ {
212141
+ "epoch": 708.26,
212142
+ "learning_rate": 8.596089743589744e-06,
212143
+ "loss": 0.3337,
212144
+ "step": 88640
212145
+ },
212146
+ {
212147
+ "epoch": 708.3,
212148
+ "learning_rate": 8.596009615384617e-06,
212149
+ "loss": 0.3107,
212150
+ "step": 88645
212151
+ },
212152
+ {
212153
+ "epoch": 708.34,
212154
+ "learning_rate": 8.595929487179488e-06,
212155
+ "loss": 0.4837,
212156
+ "step": 88650
212157
+ },
212158
+ {
212159
+ "epoch": 708.38,
212160
+ "learning_rate": 8.59584935897436e-06,
212161
+ "loss": 0.7526,
212162
+ "step": 88655
212163
+ },
212164
+ {
212165
+ "epoch": 708.42,
212166
+ "learning_rate": 8.595769230769232e-06,
212167
+ "loss": 1.1701,
212168
+ "step": 88660
212169
+ },
212170
+ {
212171
+ "epoch": 708.46,
212172
+ "learning_rate": 8.595689102564104e-06,
212173
+ "loss": 0.3475,
212174
+ "step": 88665
212175
+ },
212176
+ {
212177
+ "epoch": 708.5,
212178
+ "learning_rate": 8.595608974358975e-06,
212179
+ "loss": 0.2971,
212180
+ "step": 88670
212181
+ },
212182
+ {
212183
+ "epoch": 708.54,
212184
+ "learning_rate": 8.595528846153847e-06,
212185
+ "loss": 0.3562,
212186
+ "step": 88675
212187
+ },
212188
+ {
212189
+ "epoch": 708.58,
212190
+ "learning_rate": 8.59544871794872e-06,
212191
+ "loss": 0.6971,
212192
+ "step": 88680
212193
+ },
212194
+ {
212195
+ "epoch": 708.62,
212196
+ "learning_rate": 8.595368589743591e-06,
212197
+ "loss": 1.0363,
212198
+ "step": 88685
212199
+ },
212200
+ {
212201
+ "epoch": 708.66,
212202
+ "learning_rate": 8.595288461538462e-06,
212203
+ "loss": 0.3066,
212204
+ "step": 88690
212205
+ },
212206
+ {
212207
+ "epoch": 708.7,
212208
+ "learning_rate": 8.595208333333334e-06,
212209
+ "loss": 0.2913,
212210
+ "step": 88695
212211
+ },
212212
+ {
212213
+ "epoch": 708.74,
212214
+ "learning_rate": 8.595128205128207e-06,
212215
+ "loss": 0.341,
212216
+ "step": 88700
212217
+ },
212218
+ {
212219
+ "epoch": 708.78,
212220
+ "learning_rate": 8.595048076923076e-06,
212221
+ "loss": 0.6731,
212222
+ "step": 88705
212223
+ },
212224
+ {
212225
+ "epoch": 708.82,
212226
+ "learning_rate": 8.59496794871795e-06,
212227
+ "loss": 1.0604,
212228
+ "step": 88710
212229
+ },
212230
+ {
212231
+ "epoch": 708.86,
212232
+ "learning_rate": 8.594887820512822e-06,
212233
+ "loss": 0.2956,
212234
+ "step": 88715
212235
+ },
212236
+ {
212237
+ "epoch": 708.9,
212238
+ "learning_rate": 8.594807692307692e-06,
212239
+ "loss": 0.2839,
212240
+ "step": 88720
212241
+ },
212242
+ {
212243
+ "epoch": 708.94,
212244
+ "learning_rate": 8.594727564102565e-06,
212245
+ "loss": 0.3427,
212246
+ "step": 88725
212247
+ },
212248
+ {
212249
+ "epoch": 708.98,
212250
+ "learning_rate": 8.594647435897437e-06,
212251
+ "loss": 0.7291,
212252
+ "step": 88730
212253
+ },
212254
+ {
212255
+ "epoch": 709.0,
212256
+ "eval_loss": 0.384115993976593,
212257
+ "eval_runtime": 39.7382,
212258
+ "eval_samples_per_second": 21.189,
212259
+ "eval_steps_per_second": 0.679,
212260
+ "eval_wer": 0.18778051252352684,
212261
+ "step": 88733
212262
+ },
212263
+ {
212264
+ "epoch": 715.02,
212265
+ "learning_rate": 8.594567307692308e-06,
212266
+ "loss": 0.3286,
212267
+ "step": 88735
212268
+ },
212269
+ {
212270
+ "epoch": 715.06,
212271
+ "learning_rate": 8.59448717948718e-06,
212272
+ "loss": 0.2882,
212273
+ "step": 88740
212274
+ },
212275
+ {
212276
+ "epoch": 715.1,
212277
+ "learning_rate": 8.594407051282052e-06,
212278
+ "loss": 0.3112,
212279
+ "step": 88745
212280
+ },
212281
+ {
212282
+ "epoch": 715.14,
212283
+ "learning_rate": 8.594326923076924e-06,
212284
+ "loss": 0.3241,
212285
+ "step": 88750
212286
+ },
212287
+ {
212288
+ "epoch": 715.18,
212289
+ "learning_rate": 8.594246794871795e-06,
212290
+ "loss": 0.5784,
212291
+ "step": 88755
212292
+ },
212293
+ {
212294
+ "epoch": 715.22,
212295
+ "learning_rate": 8.594166666666668e-06,
212296
+ "loss": 0.9482,
212297
+ "step": 88760
212298
+ },
212299
+ {
212300
+ "epoch": 715.26,
212301
+ "learning_rate": 8.59408653846154e-06,
212302
+ "loss": 0.2835,
212303
+ "step": 88765
212304
+ },
212305
+ {
212306
+ "epoch": 715.3,
212307
+ "learning_rate": 8.59400641025641e-06,
212308
+ "loss": 0.2901,
212309
+ "step": 88770
212310
+ },
212311
+ {
212312
+ "epoch": 715.34,
212313
+ "learning_rate": 8.593926282051282e-06,
212314
+ "loss": 0.3611,
212315
+ "step": 88775
212316
+ },
212317
+ {
212318
+ "epoch": 715.38,
212319
+ "learning_rate": 8.593846153846155e-06,
212320
+ "loss": 0.6981,
212321
+ "step": 88780
212322
+ },
212323
+ {
212324
+ "epoch": 715.42,
212325
+ "learning_rate": 8.593766025641027e-06,
212326
+ "loss": 0.9617,
212327
+ "step": 88785
212328
+ },
212329
+ {
212330
+ "epoch": 715.46,
212331
+ "learning_rate": 8.593685897435898e-06,
212332
+ "loss": 0.3106,
212333
+ "step": 88790
212334
+ },
212335
+ {
212336
+ "epoch": 715.5,
212337
+ "learning_rate": 8.59360576923077e-06,
212338
+ "loss": 0.2652,
212339
+ "step": 88795
212340
+ },
212341
+ {
212342
+ "epoch": 715.54,
212343
+ "learning_rate": 8.593525641025642e-06,
212344
+ "loss": 0.3255,
212345
+ "step": 88800
212346
+ },
212347
+ {
212348
+ "epoch": 715.58,
212349
+ "learning_rate": 8.593445512820514e-06,
212350
+ "loss": 0.6727,
212351
+ "step": 88805
212352
+ },
212353
+ {
212354
+ "epoch": 715.62,
212355
+ "learning_rate": 8.593365384615385e-06,
212356
+ "loss": 0.9983,
212357
+ "step": 88810
212358
+ },
212359
+ {
212360
+ "epoch": 715.66,
212361
+ "learning_rate": 8.593285256410258e-06,
212362
+ "loss": 0.3482,
212363
+ "step": 88815
212364
+ },
212365
+ {
212366
+ "epoch": 715.7,
212367
+ "learning_rate": 8.59320512820513e-06,
212368
+ "loss": 0.3422,
212369
+ "step": 88820
212370
+ },
212371
+ {
212372
+ "epoch": 715.74,
212373
+ "learning_rate": 8.593125e-06,
212374
+ "loss": 0.3654,
212375
+ "step": 88825
212376
+ },
212377
+ {
212378
+ "epoch": 715.78,
212379
+ "learning_rate": 8.593044871794872e-06,
212380
+ "loss": 0.7087,
212381
+ "step": 88830
212382
+ },
212383
+ {
212384
+ "epoch": 715.82,
212385
+ "learning_rate": 8.592964743589745e-06,
212386
+ "loss": 0.9036,
212387
+ "step": 88835
212388
+ },
212389
+ {
212390
+ "epoch": 715.86,
212391
+ "learning_rate": 8.592884615384615e-06,
212392
+ "loss": 0.3362,
212393
+ "step": 88840
212394
+ },
212395
+ {
212396
+ "epoch": 715.9,
212397
+ "learning_rate": 8.592804487179488e-06,
212398
+ "loss": 0.3543,
212399
+ "step": 88845
212400
+ },
212401
+ {
212402
+ "epoch": 715.94,
212403
+ "learning_rate": 8.59272435897436e-06,
212404
+ "loss": 0.3444,
212405
+ "step": 88850
212406
+ },
212407
+ {
212408
+ "epoch": 715.98,
212409
+ "learning_rate": 8.59264423076923e-06,
212410
+ "loss": 0.7275,
212411
+ "step": 88855
212412
+ },
212413
+ {
212414
+ "epoch": 716.0,
212415
+ "eval_loss": 0.42075592279434204,
212416
+ "eval_runtime": 40.0083,
212417
+ "eval_samples_per_second": 21.046,
212418
+ "eval_steps_per_second": 0.675,
212419
+ "eval_wer": 0.17731766728757703,
212420
+ "step": 88857
212421
  }
212422
  ],
212423
  "max_steps": 620000,
212424
  "num_train_epochs": 5000,
212425
+ "total_flos": 2.5005701371269464e+20,
212426
  "trial_name": null,
212427
  "trial_params": null
212428
  }
model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88857}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88981}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88981}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:02b5f5d438dd7750ddcbd5e03a06bf5da10647a0694f2c80b916bb08f63d65f3
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85223ba53941b432c78346aab6f9972c74939487ac260fa93b1d6b1f037f2caa
3
  size 722165393
model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88981}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88981}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a3d8965d59d5aefe625b01be2b25e1c5475c317dfa971d5c9d78af786af1ca5
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7d0ee63362b6103e3b14d409796c9f7d0158647e1d489337c6a63ebb1785554
3
  size 377909911
model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88981}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1a668946223de84953b525379efbf1f1aafb391fa2b0b6fd2b09337d0053137e
3
- size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa2d10ac6adbf42dc45f693ff91266acd0ee130d5f99d16aabf290292d077da1
3
+ size 14567
model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88981}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed72d3338847ccab1a388d7c3792f1105409f238a637a418c2bbf8290dcbc6b6
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9da1248dae22f22d368d318bd65aa7f85b60c95535b8a8cd70eeacd40df98fbe
3
  size 559
model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88981}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:08370db5ce158b578b395d894b742358ae8d2c0dfd355a9ab2b4606ae3610978
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca8f1a5975684ec4e0bf01320f1c3c47ae2829cdaa18b8d590d05e5f125b3b39
3
  size 623
model-bin/finetune/base/{checkpoint-79148 β†’ checkpoint-88981}/trainer_state.json RENAMED
The diff for this file is too large to render. See raw diff
 
model-bin/finetune/base/{checkpoint-88359 β†’ checkpoint-88981}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629939465.6417572/events.out.tfevents.1629939465.8e89bd551565.924.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e141ce9b4127e1770c55b27424b480f9c16cfa6bdddc9384d374ea0d3366cd5a
3
+ size 4194
model-bin/finetune/base/log/1629939905.367717/events.out.tfevents.1629939905.8e89bd551565.924.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6ad0afe0573af339a0a61f1f61cb45e5da611c5fe3726ee9ae92cdb358076d0
3
+ size 4194
model-bin/finetune/base/log/1629940450.426258/events.out.tfevents.1629940450.8e89bd551565.924.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abe17b1e4d81daacd59ebd5fc55b611e1ecbbe08c365049f1a9dca0363ccbde5
3
+ size 4194
model-bin/finetune/base/log/1629940922.8454914/events.out.tfevents.1629940922.8e89bd551565.924.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3e7148ef8a4bb21a33b79e1754c1ce5d8745bba37d398773eb587100c19a3fb
3
+ size 4194
model-bin/finetune/base/log/1629941350.450414/events.out.tfevents.1629941350.8e89bd551565.924.9 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ee05a6111fcb4b236e602754c722306f188669fd805f372d1ec101cdae2803a
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629939465.8e89bd551565.924.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1ec61a66682433e6a2c4c5ed8f52536e1bdadf14b2be8bf14989e4e914e480a
3
+ size 8630
model-bin/finetune/base/log/events.out.tfevents.1629939905.8e89bd551565.924.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09a4bed9d1b47a2fe49ea61ab22b2e850eadbb3b56f5fbbacf3720eb199c4230
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629940450.8e89bd551565.924.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e6fcceaf09e59a28fc26f7a9881cc55849bdd514a9b0322a9089593c40cfb29
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629940922.8e89bd551565.924.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33c176ebbf4c0baa5db92ea369c5070629ad1717dcf415f0901f471b8243bf9e
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629941350.8e89bd551565.924.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14955db15f07f8b3d9576f0a5e03653799ac0382b970196223a0e10f1ecade56
3
+ size 8622