Check commited on
Commit
40b1526
Β·
1 Parent(s): 1c3eeca

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/trainer_state.json +793 -4
  9. model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629993468.0041914/events.out.tfevents.1629993468.8e89bd551565.924.241 +3 -0
  11. model-bin/finetune/base/log/1629993903.8396595/events.out.tfevents.1629993903.8e89bd551565.924.243 +3 -0
  12. model-bin/finetune/base/log/1629994337.0035079/events.out.tfevents.1629994337.8e89bd551565.924.245 +3 -0
  13. model-bin/finetune/base/log/1629994779.4147024/events.out.tfevents.1629994779.8e89bd551565.924.247 +3 -0
  14. model-bin/finetune/base/log/1629995217.6110506/events.out.tfevents.1629995217.8e89bd551565.924.249 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629993467.8e89bd551565.924.240 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629993901.8e89bd551565.924.242 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629994336.8e89bd551565.924.244 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629994779.8e89bd551565.924.246 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629995217.8e89bd551565.924.248 +3 -0
model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9857b78c463eaa385a304848157348739a41eb8da315dfeeaa2d0a3ec71e75a8
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f887ccc9e530275eabb942742b49945da5e396af357fcb4c6216a8b0e38aaa0
3
  size 722165393
model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:14fb77230a7ee63d8bf5f9f184067800cd9ffc8bb19caf3281f975ac41a0fcb9
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa66840495953d2dcf8d0f972d70b80e8025406460af3c06559744b0b8e00069
3
  size 377909911
model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:af11dafbe5807298fa685ae5cccc7951004d27d1d92f3f39b987c08874b4ccb8
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1d7ee77a309d4b02f0e0b50a254c8d8994dd350eafd8040d4e8821c92cec5ce
3
  size 14503
model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:69d87179115fbc6b93e2b095c017d1bc9093ab62fbae15eb403587041f7292fd
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a7e782bf5b4614ad2f9ac21555ba4813411b4db165cdaa4d98d6f042e85cb7e
3
  size 559
model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d5a33c6409adafb3ae851e601f2030d4c7390f928113bd1e1c69ada32803f1df
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a6409ad26afd7a4d5286ea7d95c3da331c0a58c248fb8e74926edc29a9ff2b5
3
  size 623
model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
- "epoch": 826.0,
5
- "global_step": 103297,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -230790,11 +230790,800 @@
230790
  "eval_steps_per_second": 0.684,
230791
  "eval_wer": 0.18283473714783618,
230792
  "step": 103297
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
230793
  }
230794
  ],
230795
- "max_steps": 625000,
230796
  "num_train_epochs": 5000,
230797
- "total_flos": 2.907039073185294e+20,
230798
  "trial_name": null,
230799
  "trial_params": null
230800
  }
 
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
+ "epoch": 838.0,
5
+ "global_step": 103918,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
230790
  "eval_steps_per_second": 0.684,
230791
  "eval_wer": 0.18283473714783618,
230792
  "step": 103297
230793
+ },
230794
+ {
230795
+ "epoch": 833.02,
230796
+ "learning_rate": 8.361233974358976e-06,
230797
+ "loss": 0.351,
230798
+ "step": 103300
230799
+ },
230800
+ {
230801
+ "epoch": 833.06,
230802
+ "learning_rate": 8.361153846153846e-06,
230803
+ "loss": 0.3207,
230804
+ "step": 103305
230805
+ },
230806
+ {
230807
+ "epoch": 833.1,
230808
+ "learning_rate": 8.361073717948719e-06,
230809
+ "loss": 0.2721,
230810
+ "step": 103310
230811
+ },
230812
+ {
230813
+ "epoch": 833.14,
230814
+ "learning_rate": 8.36099358974359e-06,
230815
+ "loss": 0.3887,
230816
+ "step": 103315
230817
+ },
230818
+ {
230819
+ "epoch": 833.18,
230820
+ "learning_rate": 8.360913461538462e-06,
230821
+ "loss": 0.7524,
230822
+ "step": 103320
230823
+ },
230824
+ {
230825
+ "epoch": 833.22,
230826
+ "learning_rate": 8.360833333333333e-06,
230827
+ "loss": 0.915,
230828
+ "step": 103325
230829
+ },
230830
+ {
230831
+ "epoch": 833.27,
230832
+ "learning_rate": 8.360753205128206e-06,
230833
+ "loss": 0.291,
230834
+ "step": 103330
230835
+ },
230836
+ {
230837
+ "epoch": 833.31,
230838
+ "learning_rate": 8.360673076923077e-06,
230839
+ "loss": 0.3147,
230840
+ "step": 103335
230841
+ },
230842
+ {
230843
+ "epoch": 833.35,
230844
+ "learning_rate": 8.360592948717949e-06,
230845
+ "loss": 0.3544,
230846
+ "step": 103340
230847
+ },
230848
+ {
230849
+ "epoch": 833.39,
230850
+ "learning_rate": 8.360512820512822e-06,
230851
+ "loss": 0.7792,
230852
+ "step": 103345
230853
+ },
230854
+ {
230855
+ "epoch": 833.43,
230856
+ "learning_rate": 8.360432692307693e-06,
230857
+ "loss": 0.8214,
230858
+ "step": 103350
230859
+ },
230860
+ {
230861
+ "epoch": 833.47,
230862
+ "learning_rate": 8.360352564102565e-06,
230863
+ "loss": 0.2678,
230864
+ "step": 103355
230865
+ },
230866
+ {
230867
+ "epoch": 833.51,
230868
+ "learning_rate": 8.360272435897436e-06,
230869
+ "loss": 0.288,
230870
+ "step": 103360
230871
+ },
230872
+ {
230873
+ "epoch": 833.55,
230874
+ "learning_rate": 8.360192307692309e-06,
230875
+ "loss": 0.4139,
230876
+ "step": 103365
230877
+ },
230878
+ {
230879
+ "epoch": 833.59,
230880
+ "learning_rate": 8.36011217948718e-06,
230881
+ "loss": 0.7124,
230882
+ "step": 103370
230883
+ },
230884
+ {
230885
+ "epoch": 833.63,
230886
+ "learning_rate": 8.360032051282052e-06,
230887
+ "loss": 0.8996,
230888
+ "step": 103375
230889
+ },
230890
+ {
230891
+ "epoch": 833.67,
230892
+ "learning_rate": 8.359951923076923e-06,
230893
+ "loss": 0.3268,
230894
+ "step": 103380
230895
+ },
230896
+ {
230897
+ "epoch": 833.71,
230898
+ "learning_rate": 8.359871794871796e-06,
230899
+ "loss": 0.599,
230900
+ "step": 103385
230901
+ },
230902
+ {
230903
+ "epoch": 833.75,
230904
+ "learning_rate": 8.359791666666667e-06,
230905
+ "loss": 0.3569,
230906
+ "step": 103390
230907
+ },
230908
+ {
230909
+ "epoch": 833.79,
230910
+ "learning_rate": 8.359711538461539e-06,
230911
+ "loss": 1.1455,
230912
+ "step": 103395
230913
+ },
230914
+ {
230915
+ "epoch": 833.83,
230916
+ "learning_rate": 8.359631410256412e-06,
230917
+ "loss": 0.8055,
230918
+ "step": 103400
230919
+ },
230920
+ {
230921
+ "epoch": 833.87,
230922
+ "learning_rate": 8.359551282051283e-06,
230923
+ "loss": 0.3046,
230924
+ "step": 103405
230925
+ },
230926
+ {
230927
+ "epoch": 833.91,
230928
+ "learning_rate": 8.359471153846155e-06,
230929
+ "loss": 0.3044,
230930
+ "step": 103410
230931
+ },
230932
+ {
230933
+ "epoch": 833.95,
230934
+ "learning_rate": 8.359391025641026e-06,
230935
+ "loss": 0.3412,
230936
+ "step": 103415
230937
+ },
230938
+ {
230939
+ "epoch": 833.99,
230940
+ "learning_rate": 8.359310897435899e-06,
230941
+ "loss": 0.9576,
230942
+ "step": 103420
230943
+ },
230944
+ {
230945
+ "epoch": 834.0,
230946
+ "eval_loss": 0.3438977897167206,
230947
+ "eval_runtime": 40.5967,
230948
+ "eval_samples_per_second": 20.691,
230949
+ "eval_steps_per_second": 0.665,
230950
+ "eval_wer": 0.18366442655638432,
230951
+ "step": 103421
230952
+ },
230953
+ {
230954
+ "epoch": 827.03,
230955
+ "learning_rate": 8.359230769230769e-06,
230956
+ "loss": 0.4528,
230957
+ "step": 103425
230958
+ },
230959
+ {
230960
+ "epoch": 827.07,
230961
+ "learning_rate": 8.359150641025642e-06,
230962
+ "loss": 0.3116,
230963
+ "step": 103430
230964
+ },
230965
+ {
230966
+ "epoch": 827.11,
230967
+ "learning_rate": 8.359070512820513e-06,
230968
+ "loss": 0.3281,
230969
+ "step": 103435
230970
+ },
230971
+ {
230972
+ "epoch": 827.15,
230973
+ "learning_rate": 8.358990384615384e-06,
230974
+ "loss": 0.3707,
230975
+ "step": 103440
230976
+ },
230977
+ {
230978
+ "epoch": 827.19,
230979
+ "learning_rate": 8.358910256410258e-06,
230980
+ "loss": 0.9188,
230981
+ "step": 103445
230982
+ },
230983
+ {
230984
+ "epoch": 827.23,
230985
+ "learning_rate": 8.358830128205129e-06,
230986
+ "loss": 0.6167,
230987
+ "step": 103450
230988
+ },
230989
+ {
230990
+ "epoch": 827.27,
230991
+ "learning_rate": 8.35875e-06,
230992
+ "loss": 0.3096,
230993
+ "step": 103455
230994
+ },
230995
+ {
230996
+ "epoch": 827.31,
230997
+ "learning_rate": 8.358669871794872e-06,
230998
+ "loss": 0.3597,
230999
+ "step": 103460
231000
+ },
231001
+ {
231002
+ "epoch": 827.35,
231003
+ "learning_rate": 8.358589743589745e-06,
231004
+ "loss": 0.3656,
231005
+ "step": 103465
231006
+ },
231007
+ {
231008
+ "epoch": 827.39,
231009
+ "learning_rate": 8.358509615384616e-06,
231010
+ "loss": 0.9181,
231011
+ "step": 103470
231012
+ },
231013
+ {
231014
+ "epoch": 827.43,
231015
+ "learning_rate": 8.358429487179487e-06,
231016
+ "loss": 0.6144,
231017
+ "step": 103475
231018
+ },
231019
+ {
231020
+ "epoch": 827.47,
231021
+ "learning_rate": 8.358349358974359e-06,
231022
+ "loss": 0.3339,
231023
+ "step": 103480
231024
+ },
231025
+ {
231026
+ "epoch": 827.51,
231027
+ "learning_rate": 8.358269230769232e-06,
231028
+ "loss": 0.3008,
231029
+ "step": 103485
231030
+ },
231031
+ {
231032
+ "epoch": 827.55,
231033
+ "learning_rate": 8.358189102564103e-06,
231034
+ "loss": 0.417,
231035
+ "step": 103490
231036
+ },
231037
+ {
231038
+ "epoch": 827.59,
231039
+ "learning_rate": 8.358108974358974e-06,
231040
+ "loss": 0.9079,
231041
+ "step": 103495
231042
+ },
231043
+ {
231044
+ "epoch": 827.63,
231045
+ "learning_rate": 8.358028846153848e-06,
231046
+ "loss": 0.7081,
231047
+ "step": 103500
231048
+ },
231049
+ {
231050
+ "epoch": 827.67,
231051
+ "learning_rate": 8.357948717948719e-06,
231052
+ "loss": 0.2889,
231053
+ "step": 103505
231054
+ },
231055
+ {
231056
+ "epoch": 827.71,
231057
+ "learning_rate": 8.35786858974359e-06,
231058
+ "loss": 0.3271,
231059
+ "step": 103510
231060
+ },
231061
+ {
231062
+ "epoch": 827.75,
231063
+ "learning_rate": 8.357788461538462e-06,
231064
+ "loss": 0.5678,
231065
+ "step": 103515
231066
+ },
231067
+ {
231068
+ "epoch": 827.79,
231069
+ "learning_rate": 8.357708333333335e-06,
231070
+ "loss": 0.9176,
231071
+ "step": 103520
231072
+ },
231073
+ {
231074
+ "epoch": 827.83,
231075
+ "learning_rate": 8.357628205128206e-06,
231076
+ "loss": 0.7331,
231077
+ "step": 103525
231078
+ },
231079
+ {
231080
+ "epoch": 827.87,
231081
+ "learning_rate": 8.357548076923077e-06,
231082
+ "loss": 0.2808,
231083
+ "step": 103530
231084
+ },
231085
+ {
231086
+ "epoch": 827.91,
231087
+ "learning_rate": 8.357467948717949e-06,
231088
+ "loss": 0.3182,
231089
+ "step": 103535
231090
+ },
231091
+ {
231092
+ "epoch": 827.95,
231093
+ "learning_rate": 8.357387820512822e-06,
231094
+ "loss": 0.43,
231095
+ "step": 103540
231096
+ },
231097
+ {
231098
+ "epoch": 827.99,
231099
+ "learning_rate": 8.357307692307693e-06,
231100
+ "loss": 0.8699,
231101
+ "step": 103545
231102
+ },
231103
+ {
231104
+ "epoch": 828.0,
231105
+ "eval_loss": 0.36451485753059387,
231106
+ "eval_runtime": 40.0413,
231107
+ "eval_samples_per_second": 21.003,
231108
+ "eval_steps_per_second": 0.674,
231109
+ "eval_wer": 0.1745266037871697,
231110
+ "step": 103546
231111
+ },
231112
+ {
231113
+ "epoch": 835.03,
231114
+ "learning_rate": 8.357227564102565e-06,
231115
+ "loss": 0.2803,
231116
+ "step": 103550
231117
+ },
231118
+ {
231119
+ "epoch": 835.07,
231120
+ "learning_rate": 8.357147435897438e-06,
231121
+ "loss": 0.3105,
231122
+ "step": 103555
231123
+ },
231124
+ {
231125
+ "epoch": 835.11,
231126
+ "learning_rate": 8.357067307692309e-06,
231127
+ "loss": 0.2909,
231128
+ "step": 103560
231129
+ },
231130
+ {
231131
+ "epoch": 835.15,
231132
+ "learning_rate": 8.35698717948718e-06,
231133
+ "loss": 0.4187,
231134
+ "step": 103565
231135
+ },
231136
+ {
231137
+ "epoch": 835.19,
231138
+ "learning_rate": 8.356907051282052e-06,
231139
+ "loss": 0.9497,
231140
+ "step": 103570
231141
+ },
231142
+ {
231143
+ "epoch": 835.23,
231144
+ "learning_rate": 8.356826923076925e-06,
231145
+ "loss": 0.734,
231146
+ "step": 103575
231147
+ },
231148
+ {
231149
+ "epoch": 835.27,
231150
+ "learning_rate": 8.356746794871794e-06,
231151
+ "loss": 0.3495,
231152
+ "step": 103580
231153
+ },
231154
+ {
231155
+ "epoch": 835.31,
231156
+ "learning_rate": 8.356666666666667e-06,
231157
+ "loss": 0.3019,
231158
+ "step": 103585
231159
+ },
231160
+ {
231161
+ "epoch": 835.35,
231162
+ "learning_rate": 8.35658653846154e-06,
231163
+ "loss": 0.3799,
231164
+ "step": 103590
231165
+ },
231166
+ {
231167
+ "epoch": 835.39,
231168
+ "learning_rate": 8.35650641025641e-06,
231169
+ "loss": 0.8788,
231170
+ "step": 103595
231171
+ },
231172
+ {
231173
+ "epoch": 835.43,
231174
+ "learning_rate": 8.356426282051283e-06,
231175
+ "loss": 0.6812,
231176
+ "step": 103600
231177
+ },
231178
+ {
231179
+ "epoch": 835.47,
231180
+ "learning_rate": 8.356346153846155e-06,
231181
+ "loss": 0.2686,
231182
+ "step": 103605
231183
+ },
231184
+ {
231185
+ "epoch": 835.51,
231186
+ "learning_rate": 8.356266025641026e-06,
231187
+ "loss": 0.3267,
231188
+ "step": 103610
231189
+ },
231190
+ {
231191
+ "epoch": 835.55,
231192
+ "learning_rate": 8.356185897435897e-06,
231193
+ "loss": 0.4017,
231194
+ "step": 103615
231195
+ },
231196
+ {
231197
+ "epoch": 835.59,
231198
+ "learning_rate": 8.35610576923077e-06,
231199
+ "loss": 0.8509,
231200
+ "step": 103620
231201
+ },
231202
+ {
231203
+ "epoch": 835.63,
231204
+ "learning_rate": 8.356025641025642e-06,
231205
+ "loss": 0.6918,
231206
+ "step": 103625
231207
+ },
231208
+ {
231209
+ "epoch": 835.67,
231210
+ "learning_rate": 8.355945512820513e-06,
231211
+ "loss": 0.2692,
231212
+ "step": 103630
231213
+ },
231214
+ {
231215
+ "epoch": 835.71,
231216
+ "learning_rate": 8.355865384615384e-06,
231217
+ "loss": 0.3076,
231218
+ "step": 103635
231219
+ },
231220
+ {
231221
+ "epoch": 835.76,
231222
+ "learning_rate": 8.355785256410257e-06,
231223
+ "loss": 0.3897,
231224
+ "step": 103640
231225
+ },
231226
+ {
231227
+ "epoch": 835.8,
231228
+ "learning_rate": 8.355705128205129e-06,
231229
+ "loss": 0.8999,
231230
+ "step": 103645
231231
+ },
231232
+ {
231233
+ "epoch": 835.84,
231234
+ "learning_rate": 8.355625e-06,
231235
+ "loss": 0.6748,
231236
+ "step": 103650
231237
+ },
231238
+ {
231239
+ "epoch": 835.88,
231240
+ "learning_rate": 8.355544871794873e-06,
231241
+ "loss": 0.3033,
231242
+ "step": 103655
231243
+ },
231244
+ {
231245
+ "epoch": 835.92,
231246
+ "learning_rate": 8.355464743589745e-06,
231247
+ "loss": 0.3177,
231248
+ "step": 103660
231249
+ },
231250
+ {
231251
+ "epoch": 835.96,
231252
+ "learning_rate": 8.355384615384616e-06,
231253
+ "loss": 0.4767,
231254
+ "step": 103665
231255
+ },
231256
+ {
231257
+ "epoch": 836.0,
231258
+ "learning_rate": 8.355304487179487e-06,
231259
+ "loss": 1.0467,
231260
+ "step": 103670
231261
+ },
231262
+ {
231263
+ "epoch": 836.0,
231264
+ "eval_loss": 0.4262886345386505,
231265
+ "eval_runtime": 39.1875,
231266
+ "eval_samples_per_second": 21.435,
231267
+ "eval_steps_per_second": 0.689,
231268
+ "eval_wer": 0.18302426152605733,
231269
+ "step": 103670
231270
+ },
231271
+ {
231272
+ "epoch": 836.04,
231273
+ "learning_rate": 8.35522435897436e-06,
231274
+ "loss": 0.2928,
231275
+ "step": 103675
231276
+ },
231277
+ {
231278
+ "epoch": 836.08,
231279
+ "learning_rate": 8.355144230769232e-06,
231280
+ "loss": 0.2817,
231281
+ "step": 103680
231282
+ },
231283
+ {
231284
+ "epoch": 836.12,
231285
+ "learning_rate": 8.355064102564103e-06,
231286
+ "loss": 0.3598,
231287
+ "step": 103685
231288
+ },
231289
+ {
231290
+ "epoch": 836.16,
231291
+ "learning_rate": 8.354983974358976e-06,
231292
+ "loss": 0.4622,
231293
+ "step": 103690
231294
+ },
231295
+ {
231296
+ "epoch": 836.2,
231297
+ "learning_rate": 8.354903846153847e-06,
231298
+ "loss": 1.2675,
231299
+ "step": 103695
231300
+ },
231301
+ {
231302
+ "epoch": 836.24,
231303
+ "learning_rate": 8.354823717948719e-06,
231304
+ "loss": 0.3431,
231305
+ "step": 103700
231306
+ },
231307
+ {
231308
+ "epoch": 836.28,
231309
+ "learning_rate": 8.35474358974359e-06,
231310
+ "loss": 0.2755,
231311
+ "step": 103705
231312
+ },
231313
+ {
231314
+ "epoch": 836.32,
231315
+ "learning_rate": 8.354663461538463e-06,
231316
+ "loss": 0.3133,
231317
+ "step": 103710
231318
+ },
231319
+ {
231320
+ "epoch": 836.36,
231321
+ "learning_rate": 8.354583333333333e-06,
231322
+ "loss": 0.4812,
231323
+ "step": 103715
231324
+ },
231325
+ {
231326
+ "epoch": 836.4,
231327
+ "learning_rate": 8.354503205128206e-06,
231328
+ "loss": 1.1412,
231329
+ "step": 103720
231330
+ },
231331
+ {
231332
+ "epoch": 836.44,
231333
+ "learning_rate": 8.354423076923077e-06,
231334
+ "loss": 0.294,
231335
+ "step": 103725
231336
+ },
231337
+ {
231338
+ "epoch": 836.48,
231339
+ "learning_rate": 8.354342948717949e-06,
231340
+ "loss": 0.2575,
231341
+ "step": 103730
231342
+ },
231343
+ {
231344
+ "epoch": 836.52,
231345
+ "learning_rate": 8.35426282051282e-06,
231346
+ "loss": 0.3326,
231347
+ "step": 103735
231348
+ },
231349
+ {
231350
+ "epoch": 836.56,
231351
+ "learning_rate": 8.354182692307693e-06,
231352
+ "loss": 0.5133,
231353
+ "step": 103740
231354
+ },
231355
+ {
231356
+ "epoch": 836.6,
231357
+ "learning_rate": 8.354102564102564e-06,
231358
+ "loss": 1.1794,
231359
+ "step": 103745
231360
+ },
231361
+ {
231362
+ "epoch": 836.65,
231363
+ "learning_rate": 8.354022435897436e-06,
231364
+ "loss": 0.307,
231365
+ "step": 103750
231366
+ },
231367
+ {
231368
+ "epoch": 836.69,
231369
+ "learning_rate": 8.353942307692309e-06,
231370
+ "loss": 0.2519,
231371
+ "step": 103755
231372
+ },
231373
+ {
231374
+ "epoch": 836.73,
231375
+ "learning_rate": 8.35386217948718e-06,
231376
+ "loss": 0.3294,
231377
+ "step": 103760
231378
+ },
231379
+ {
231380
+ "epoch": 836.77,
231381
+ "learning_rate": 8.353782051282052e-06,
231382
+ "loss": 0.432,
231383
+ "step": 103765
231384
+ },
231385
+ {
231386
+ "epoch": 836.81,
231387
+ "learning_rate": 8.353701923076923e-06,
231388
+ "loss": 1.1507,
231389
+ "step": 103770
231390
+ },
231391
+ {
231392
+ "epoch": 836.85,
231393
+ "learning_rate": 8.353621794871796e-06,
231394
+ "loss": 0.3704,
231395
+ "step": 103775
231396
+ },
231397
+ {
231398
+ "epoch": 836.89,
231399
+ "learning_rate": 8.353541666666667e-06,
231400
+ "loss": 0.3221,
231401
+ "step": 103780
231402
+ },
231403
+ {
231404
+ "epoch": 836.93,
231405
+ "learning_rate": 8.353461538461539e-06,
231406
+ "loss": 0.304,
231407
+ "step": 103785
231408
+ },
231409
+ {
231410
+ "epoch": 836.97,
231411
+ "learning_rate": 8.353381410256412e-06,
231412
+ "loss": 0.4899,
231413
+ "step": 103790
231414
+ },
231415
+ {
231416
+ "epoch": 837.0,
231417
+ "eval_loss": 0.3803502321243286,
231418
+ "eval_runtime": 39.27,
231419
+ "eval_samples_per_second": 21.39,
231420
+ "eval_steps_per_second": 0.688,
231421
+ "eval_wer": 0.18735172004744957,
231422
+ "step": 103794
231423
+ },
231424
+ {
231425
+ "epoch": 837.01,
231426
+ "learning_rate": 8.353301282051283e-06,
231427
+ "loss": 0.3926,
231428
+ "step": 103795
231429
+ },
231430
+ {
231431
+ "epoch": 837.05,
231432
+ "learning_rate": 8.353221153846154e-06,
231433
+ "loss": 0.284,
231434
+ "step": 103800
231435
+ },
231436
+ {
231437
+ "epoch": 837.09,
231438
+ "learning_rate": 8.353141025641026e-06,
231439
+ "loss": 0.3139,
231440
+ "step": 103805
231441
+ },
231442
+ {
231443
+ "epoch": 837.13,
231444
+ "learning_rate": 8.353060897435899e-06,
231445
+ "loss": 0.3067,
231446
+ "step": 103810
231447
+ },
231448
+ {
231449
+ "epoch": 837.17,
231450
+ "learning_rate": 8.35298076923077e-06,
231451
+ "loss": 0.689,
231452
+ "step": 103815
231453
+ },
231454
+ {
231455
+ "epoch": 837.21,
231456
+ "learning_rate": 8.352900641025642e-06,
231457
+ "loss": 1.1145,
231458
+ "step": 103820
231459
+ },
231460
+ {
231461
+ "epoch": 837.25,
231462
+ "learning_rate": 8.352820512820513e-06,
231463
+ "loss": 0.288,
231464
+ "step": 103825
231465
+ },
231466
+ {
231467
+ "epoch": 837.29,
231468
+ "learning_rate": 8.352740384615386e-06,
231469
+ "loss": 0.2808,
231470
+ "step": 103830
231471
+ },
231472
+ {
231473
+ "epoch": 837.33,
231474
+ "learning_rate": 8.352660256410257e-06,
231475
+ "loss": 0.3067,
231476
+ "step": 103835
231477
+ },
231478
+ {
231479
+ "epoch": 837.37,
231480
+ "learning_rate": 8.352580128205129e-06,
231481
+ "loss": 0.5511,
231482
+ "step": 103840
231483
+ },
231484
+ {
231485
+ "epoch": 837.41,
231486
+ "learning_rate": 8.352500000000002e-06,
231487
+ "loss": 1.3246,
231488
+ "step": 103845
231489
+ },
231490
+ {
231491
+ "epoch": 837.45,
231492
+ "learning_rate": 8.352419871794873e-06,
231493
+ "loss": 0.2766,
231494
+ "step": 103850
231495
+ },
231496
+ {
231497
+ "epoch": 837.49,
231498
+ "learning_rate": 8.352339743589744e-06,
231499
+ "loss": 0.2609,
231500
+ "step": 103855
231501
+ },
231502
+ {
231503
+ "epoch": 837.53,
231504
+ "learning_rate": 8.352259615384616e-06,
231505
+ "loss": 0.3097,
231506
+ "step": 103860
231507
+ },
231508
+ {
231509
+ "epoch": 837.57,
231510
+ "learning_rate": 8.352179487179489e-06,
231511
+ "loss": 0.5214,
231512
+ "step": 103865
231513
+ },
231514
+ {
231515
+ "epoch": 837.61,
231516
+ "learning_rate": 8.352099358974359e-06,
231517
+ "loss": 1.0392,
231518
+ "step": 103870
231519
+ },
231520
+ {
231521
+ "epoch": 837.65,
231522
+ "learning_rate": 8.352019230769232e-06,
231523
+ "loss": 0.3029,
231524
+ "step": 103875
231525
+ },
231526
+ {
231527
+ "epoch": 837.69,
231528
+ "learning_rate": 8.351939102564103e-06,
231529
+ "loss": 0.3948,
231530
+ "step": 103880
231531
+ },
231532
+ {
231533
+ "epoch": 837.73,
231534
+ "learning_rate": 8.351858974358974e-06,
231535
+ "loss": 0.3793,
231536
+ "step": 103885
231537
+ },
231538
+ {
231539
+ "epoch": 837.77,
231540
+ "learning_rate": 8.351778846153847e-06,
231541
+ "loss": 0.5007,
231542
+ "step": 103890
231543
+ },
231544
+ {
231545
+ "epoch": 837.81,
231546
+ "learning_rate": 8.351698717948719e-06,
231547
+ "loss": 1.1322,
231548
+ "step": 103895
231549
+ },
231550
+ {
231551
+ "epoch": 837.85,
231552
+ "learning_rate": 8.35161858974359e-06,
231553
+ "loss": 0.323,
231554
+ "step": 103900
231555
+ },
231556
+ {
231557
+ "epoch": 837.9,
231558
+ "learning_rate": 8.351538461538461e-06,
231559
+ "loss": 0.3,
231560
+ "step": 103905
231561
+ },
231562
+ {
231563
+ "epoch": 837.94,
231564
+ "learning_rate": 8.351458333333335e-06,
231565
+ "loss": 0.3611,
231566
+ "step": 103910
231567
+ },
231568
+ {
231569
+ "epoch": 837.98,
231570
+ "learning_rate": 8.351378205128206e-06,
231571
+ "loss": 0.6142,
231572
+ "step": 103915
231573
+ },
231574
+ {
231575
+ "epoch": 838.0,
231576
+ "eval_loss": 0.4361937344074249,
231577
+ "eval_runtime": 39.334,
231578
+ "eval_samples_per_second": 21.356,
231579
+ "eval_steps_per_second": 0.686,
231580
+ "eval_wer": 0.18322231169021302,
231581
+ "step": 103918
231582
  }
231583
  ],
231584
+ "max_steps": 620000,
231585
  "num_train_epochs": 5000,
231586
+ "total_flos": 2.924503505825161e+20,
231587
  "trial_name": null,
231588
  "trial_params": null
231589
  }
model-bin/finetune/base/{checkpoint-103297 β†’ checkpoint-103918}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629993468.0041914/events.out.tfevents.1629993468.8e89bd551565.924.241 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28fba00972f7f5bf570ea4d89510c3990af4becc1debc7d6f73a261a90e24e32
3
+ size 4194
model-bin/finetune/base/log/1629993903.8396595/events.out.tfevents.1629993903.8e89bd551565.924.243 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f26487cdcfa5719ab8283301c96055d718938f75c42ab74a4cbfca4503fd9a6f
3
+ size 4194
model-bin/finetune/base/log/1629994337.0035079/events.out.tfevents.1629994337.8e89bd551565.924.245 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f1551ae0163f82fa11f10f2d261376555d92ddad08fb73642842db296a3328a
3
+ size 4194
model-bin/finetune/base/log/1629994779.4147024/events.out.tfevents.1629994779.8e89bd551565.924.247 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:852829eb34335d8026c2ee2937d265a7f770c06c2e7d951fde11f96bb1d64198
3
+ size 4194
model-bin/finetune/base/log/1629995217.6110506/events.out.tfevents.1629995217.8e89bd551565.924.249 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ace0e7105d54f62d60f1d144a5e00d108d00642188814d46ac908f2216fee68
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629993467.8e89bd551565.924.240 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:934eab1a0bf19c364256cdf29ad8b199d2f33cf4c8e3a85df06f64b7a2b16732
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629993901.8e89bd551565.924.242 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a60899e6d9c9122f1bcfc553fe563a0390463894524f6da806a171562a90249
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629994336.8e89bd551565.924.244 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:993ff42bf365cce8cebae0f498ba6509d70e03cbc2654d64581d3431dacfdb9a
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629994779.8e89bd551565.924.246 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19e4b098e0133780d4710aa6766ec212da6268ae34b59ee4ee0a649ae3f9b240
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1629995217.8e89bd551565.924.248 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a072963f9decedf6ba9cdf71a2019d152e4e70bb933ec60083aecb1b1433ce2
3
+ size 8622