Check commited on
Commit
abf8b95
Β·
1 Parent(s): e773e4d

"auto-commit"

Browse files
Files changed (28) hide show
  1. model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60604}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60604}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60604}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60604}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60604}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60604}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60604}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60604}/trainer_state.json +633 -3
  9. model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60604}/training_args.bin +0 -0
  10. model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60729}/config.json +0 -0
  11. model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60729}/optimizer.pt +1 -1
  12. model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60729}/preprocessor_config.json +0 -0
  13. model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60729}/pytorch_model.bin +1 -1
  14. model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60729}/rng_state.pth +1 -1
  15. model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60729}/scaler.pt +1 -1
  16. model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60729}/scheduler.pt +1 -1
  17. model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60729}/trainer_state.json +0 -0
  18. model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60729}/training_args.bin +0 -0
  19. model-bin/finetune/base/log/1629798286.419865/events.out.tfevents.1629798286.c435e1c5ee04.920.101 +3 -0
  20. model-bin/finetune/base/log/1629798930.4451573/events.out.tfevents.1629798930.c435e1c5ee04.920.103 +3 -0
  21. model-bin/finetune/base/log/1629799575.228714/events.out.tfevents.1629799575.c435e1c5ee04.920.105 +3 -0
  22. model-bin/finetune/base/log/1629800218.038527/events.out.tfevents.1629800218.c435e1c5ee04.920.107 +3 -0
  23. model-bin/finetune/base/log/1629800871.1235416/events.out.tfevents.1629800871.c435e1c5ee04.920.109 +3 -0
  24. model-bin/finetune/base/log/events.out.tfevents.1629798286.c435e1c5ee04.920.100 +3 -0
  25. model-bin/finetune/base/log/events.out.tfevents.1629798930.c435e1c5ee04.920.102 +3 -0
  26. model-bin/finetune/base/log/events.out.tfevents.1629799575.c435e1c5ee04.920.104 +3 -0
  27. model-bin/finetune/base/log/events.out.tfevents.1629800218.c435e1c5ee04.920.106 +3 -0
  28. model-bin/finetune/base/log/events.out.tfevents.1629800871.c435e1c5ee04.920.108 +3 -0
model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60604}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60604}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:417134b490a43dadfd6c863fdd32c93e62f5670589c60de518474bbbcb87192e
3
  size 722165009
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54d742d39c8446eace3eb84f293d54d3c92367bac539b2c6aac2610c3b5cd216
3
  size 722165009
model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60604}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60604}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:61f15a7e42709bb323434cf25df9eb7599246900a93a843147941594b17bad97
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5929eb300a5e7bc9be64e52e4b06762401ab913925a2a1b0e17dcaad3e94476
3
  size 377909911
model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60604}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2928313b9be06a9303804ad370f60cd7c0e5a076b0b5be3d27103d3299ee392c
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c1a87fab1e51e42d80278684e7ba62284a23bae14afa60f95e959e29d7d01e1
3
  size 14503
model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60604}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:058991856934dabbf10711a0d6d47759e1ab7bf80455e28bbbd566855d88c31d
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:789e2be1e3c06a3e2bb8e0d942a5bad1e52db387c9c9778d8d055b91e46a0dc3
3
  size 559
model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60604}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:127e8b3b2266ec02eb34759e17b27d11dfed14d46050a41bc1083ae605156cfe
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d917b2672d75e0efb76bc510eaa2d3141b6f755a5d5799309de61ece06544b7f
3
  size 623
model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60604}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.18631571186315712,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-55501",
4
- "epoch": 483.99598393574297,
5
- "global_step": 60105,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -175839,11 +175839,641 @@
175839
  "eval_steps_per_second": 0.68,
175840
  "eval_wer": 0.19872743415211602,
175841
  "step": 60105
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
175842
  }
175843
  ],
175844
  "max_steps": 620000,
175845
  "num_train_epochs": 5000,
175846
- "total_flos": 1.691543804876753e+20,
175847
  "trial_name": null,
175848
  "trial_params": null
175849
  }
 
1
  {
2
  "best_metric": 0.18631571186315712,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-55501",
4
+ "epoch": 488.0,
5
+ "global_step": 60604,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
175839
  "eval_steps_per_second": 0.68,
175840
  "eval_wer": 0.19872743415211602,
175841
  "step": 60105
175842
+ },
175843
+ {
175844
+ "epoch": 480.04,
175845
+ "learning_rate": 9.053076923076923e-06,
175846
+ "loss": 0.3265,
175847
+ "step": 60110
175848
+ },
175849
+ {
175850
+ "epoch": 480.08,
175851
+ "learning_rate": 9.052996794871796e-06,
175852
+ "loss": 0.29,
175853
+ "step": 60115
175854
+ },
175855
+ {
175856
+ "epoch": 480.12,
175857
+ "learning_rate": 9.052916666666668e-06,
175858
+ "loss": 0.3835,
175859
+ "step": 60120
175860
+ },
175861
+ {
175862
+ "epoch": 480.16,
175863
+ "learning_rate": 9.052836538461539e-06,
175864
+ "loss": 0.4613,
175865
+ "step": 60125
175866
+ },
175867
+ {
175868
+ "epoch": 480.2,
175869
+ "learning_rate": 9.052756410256412e-06,
175870
+ "loss": 1.1667,
175871
+ "step": 60130
175872
+ },
175873
+ {
175874
+ "epoch": 480.24,
175875
+ "learning_rate": 9.052676282051282e-06,
175876
+ "loss": 0.3625,
175877
+ "step": 60135
175878
+ },
175879
+ {
175880
+ "epoch": 480.28,
175881
+ "learning_rate": 9.052596153846155e-06,
175882
+ "loss": 0.3146,
175883
+ "step": 60140
175884
+ },
175885
+ {
175886
+ "epoch": 480.32,
175887
+ "learning_rate": 9.052516025641026e-06,
175888
+ "loss": 0.3837,
175889
+ "step": 60145
175890
+ },
175891
+ {
175892
+ "epoch": 480.36,
175893
+ "learning_rate": 9.052435897435898e-06,
175894
+ "loss": 0.5827,
175895
+ "step": 60150
175896
+ },
175897
+ {
175898
+ "epoch": 480.4,
175899
+ "learning_rate": 9.052355769230769e-06,
175900
+ "loss": 1.3282,
175901
+ "step": 60155
175902
+ },
175903
+ {
175904
+ "epoch": 480.44,
175905
+ "learning_rate": 9.052275641025642e-06,
175906
+ "loss": 0.3936,
175907
+ "step": 60160
175908
+ },
175909
+ {
175910
+ "epoch": 480.48,
175911
+ "learning_rate": 9.052195512820513e-06,
175912
+ "loss": 0.3122,
175913
+ "step": 60165
175914
+ },
175915
+ {
175916
+ "epoch": 480.52,
175917
+ "learning_rate": 9.052115384615385e-06,
175918
+ "loss": 0.3587,
175919
+ "step": 60170
175920
+ },
175921
+ {
175922
+ "epoch": 480.56,
175923
+ "learning_rate": 9.052035256410258e-06,
175924
+ "loss": 0.5124,
175925
+ "step": 60175
175926
+ },
175927
+ {
175928
+ "epoch": 480.6,
175929
+ "learning_rate": 9.051955128205129e-06,
175930
+ "loss": 1.1635,
175931
+ "step": 60180
175932
+ },
175933
+ {
175934
+ "epoch": 480.64,
175935
+ "learning_rate": 9.051875e-06,
175936
+ "loss": 0.3471,
175937
+ "step": 60185
175938
+ },
175939
+ {
175940
+ "epoch": 480.68,
175941
+ "learning_rate": 9.051794871794872e-06,
175942
+ "loss": 0.3232,
175943
+ "step": 60190
175944
+ },
175945
+ {
175946
+ "epoch": 480.72,
175947
+ "learning_rate": 9.051714743589745e-06,
175948
+ "loss": 0.2884,
175949
+ "step": 60195
175950
+ },
175951
+ {
175952
+ "epoch": 480.76,
175953
+ "learning_rate": 9.051634615384616e-06,
175954
+ "loss": 0.4555,
175955
+ "step": 60200
175956
+ },
175957
+ {
175958
+ "epoch": 480.8,
175959
+ "learning_rate": 9.051554487179488e-06,
175960
+ "loss": 1.1592,
175961
+ "step": 60205
175962
+ },
175963
+ {
175964
+ "epoch": 480.84,
175965
+ "learning_rate": 9.051474358974359e-06,
175966
+ "loss": 0.3908,
175967
+ "step": 60210
175968
+ },
175969
+ {
175970
+ "epoch": 480.88,
175971
+ "learning_rate": 9.051394230769232e-06,
175972
+ "loss": 0.3136,
175973
+ "step": 60215
175974
+ },
175975
+ {
175976
+ "epoch": 480.92,
175977
+ "learning_rate": 9.051314102564103e-06,
175978
+ "loss": 0.357,
175979
+ "step": 60220
175980
+ },
175981
+ {
175982
+ "epoch": 480.96,
175983
+ "learning_rate": 9.051233974358975e-06,
175984
+ "loss": 0.4976,
175985
+ "step": 60225
175986
+ },
175987
+ {
175988
+ "epoch": 481.0,
175989
+ "learning_rate": 9.051153846153848e-06,
175990
+ "loss": 1.3006,
175991
+ "step": 60230
175992
+ },
175993
+ {
175994
+ "epoch": 481.0,
175995
+ "eval_loss": 0.43538910150527954,
175996
+ "eval_runtime": 40.2473,
175997
+ "eval_samples_per_second": 20.821,
175998
+ "eval_steps_per_second": 0.671,
175999
+ "eval_wer": 0.19678979771328056,
176000
+ "step": 60230
176001
+ },
176002
+ {
176003
+ "epoch": 481.04,
176004
+ "learning_rate": 9.051073717948719e-06,
176005
+ "loss": 0.3275,
176006
+ "step": 60235
176007
+ },
176008
+ {
176009
+ "epoch": 481.08,
176010
+ "learning_rate": 9.05099358974359e-06,
176011
+ "loss": 0.3484,
176012
+ "step": 60240
176013
+ },
176014
+ {
176015
+ "epoch": 481.12,
176016
+ "learning_rate": 9.050913461538462e-06,
176017
+ "loss": 0.3509,
176018
+ "step": 60245
176019
+ },
176020
+ {
176021
+ "epoch": 481.16,
176022
+ "learning_rate": 9.050833333333335e-06,
176023
+ "loss": 0.5427,
176024
+ "step": 60250
176025
+ },
176026
+ {
176027
+ "epoch": 481.2,
176028
+ "learning_rate": 9.050753205128206e-06,
176029
+ "loss": 1.181,
176030
+ "step": 60255
176031
+ },
176032
+ {
176033
+ "epoch": 481.24,
176034
+ "learning_rate": 9.050673076923078e-06,
176035
+ "loss": 0.4515,
176036
+ "step": 60260
176037
+ },
176038
+ {
176039
+ "epoch": 481.28,
176040
+ "learning_rate": 9.050592948717949e-06,
176041
+ "loss": 0.2992,
176042
+ "step": 60265
176043
+ },
176044
+ {
176045
+ "epoch": 481.32,
176046
+ "learning_rate": 9.050512820512822e-06,
176047
+ "loss": 0.398,
176048
+ "step": 60270
176049
+ },
176050
+ {
176051
+ "epoch": 481.36,
176052
+ "learning_rate": 9.050432692307693e-06,
176053
+ "loss": 0.4436,
176054
+ "step": 60275
176055
+ },
176056
+ {
176057
+ "epoch": 481.4,
176058
+ "learning_rate": 9.050352564102565e-06,
176059
+ "loss": 1.1835,
176060
+ "step": 60280
176061
+ },
176062
+ {
176063
+ "epoch": 481.44,
176064
+ "learning_rate": 9.050272435897438e-06,
176065
+ "loss": 0.3682,
176066
+ "step": 60285
176067
+ },
176068
+ {
176069
+ "epoch": 481.48,
176070
+ "learning_rate": 9.050192307692307e-06,
176071
+ "loss": 0.4315,
176072
+ "step": 60290
176073
+ },
176074
+ {
176075
+ "epoch": 481.52,
176076
+ "learning_rate": 9.05011217948718e-06,
176077
+ "loss": 0.3953,
176078
+ "step": 60295
176079
+ },
176080
+ {
176081
+ "epoch": 481.56,
176082
+ "learning_rate": 9.050032051282052e-06,
176083
+ "loss": 0.5403,
176084
+ "step": 60300
176085
+ },
176086
+ {
176087
+ "epoch": 481.6,
176088
+ "learning_rate": 9.049951923076923e-06,
176089
+ "loss": 1.2516,
176090
+ "step": 60305
176091
+ },
176092
+ {
176093
+ "epoch": 481.64,
176094
+ "learning_rate": 9.049871794871795e-06,
176095
+ "loss": 0.3523,
176096
+ "step": 60310
176097
+ },
176098
+ {
176099
+ "epoch": 481.68,
176100
+ "learning_rate": 9.049791666666668e-06,
176101
+ "loss": 0.34,
176102
+ "step": 60315
176103
+ },
176104
+ {
176105
+ "epoch": 481.72,
176106
+ "learning_rate": 9.049711538461539e-06,
176107
+ "loss": 0.3594,
176108
+ "step": 60320
176109
+ },
176110
+ {
176111
+ "epoch": 481.76,
176112
+ "learning_rate": 9.04963141025641e-06,
176113
+ "loss": 0.5865,
176114
+ "step": 60325
176115
+ },
176116
+ {
176117
+ "epoch": 481.8,
176118
+ "learning_rate": 9.049551282051283e-06,
176119
+ "loss": 1.1754,
176120
+ "step": 60330
176121
+ },
176122
+ {
176123
+ "epoch": 481.84,
176124
+ "learning_rate": 9.049471153846155e-06,
176125
+ "loss": 0.3405,
176126
+ "step": 60335
176127
+ },
176128
+ {
176129
+ "epoch": 481.88,
176130
+ "learning_rate": 9.049391025641026e-06,
176131
+ "loss": 0.656,
176132
+ "step": 60340
176133
+ },
176134
+ {
176135
+ "epoch": 481.92,
176136
+ "learning_rate": 9.049310897435897e-06,
176137
+ "loss": 0.342,
176138
+ "step": 60345
176139
+ },
176140
+ {
176141
+ "epoch": 481.96,
176142
+ "learning_rate": 9.04923076923077e-06,
176143
+ "loss": 0.4822,
176144
+ "step": 60350
176145
+ },
176146
+ {
176147
+ "epoch": 482.0,
176148
+ "learning_rate": 9.049150641025642e-06,
176149
+ "loss": 1.2516,
176150
+ "step": 60355
176151
+ },
176152
+ {
176153
+ "epoch": 482.0,
176154
+ "eval_loss": 0.4162551164627075,
176155
+ "eval_runtime": 40.1501,
176156
+ "eval_samples_per_second": 20.872,
176157
+ "eval_steps_per_second": 0.672,
176158
+ "eval_wer": 0.19967876177265095,
176159
+ "step": 60355
176160
+ },
176161
+ {
176162
+ "epoch": 482.04,
176163
+ "learning_rate": 9.049070512820513e-06,
176164
+ "loss": 0.4382,
176165
+ "step": 60360
176166
+ },
176167
+ {
176168
+ "epoch": 482.08,
176169
+ "learning_rate": 9.048990384615385e-06,
176170
+ "loss": 0.3064,
176171
+ "step": 60365
176172
+ },
176173
+ {
176174
+ "epoch": 482.12,
176175
+ "learning_rate": 9.048910256410258e-06,
176176
+ "loss": 0.348,
176177
+ "step": 60370
176178
+ },
176179
+ {
176180
+ "epoch": 482.16,
176181
+ "learning_rate": 9.048830128205129e-06,
176182
+ "loss": 0.4735,
176183
+ "step": 60375
176184
+ },
176185
+ {
176186
+ "epoch": 482.2,
176187
+ "learning_rate": 9.04875e-06,
176188
+ "loss": 1.3066,
176189
+ "step": 60380
176190
+ },
176191
+ {
176192
+ "epoch": 482.24,
176193
+ "learning_rate": 9.048669871794873e-06,
176194
+ "loss": 0.4388,
176195
+ "step": 60385
176196
+ },
176197
+ {
176198
+ "epoch": 482.28,
176199
+ "learning_rate": 9.048589743589745e-06,
176200
+ "loss": 0.3116,
176201
+ "step": 60390
176202
+ },
176203
+ {
176204
+ "epoch": 482.32,
176205
+ "learning_rate": 9.048509615384616e-06,
176206
+ "loss": 0.3442,
176207
+ "step": 60395
176208
+ },
176209
+ {
176210
+ "epoch": 482.36,
176211
+ "learning_rate": 9.048429487179488e-06,
176212
+ "loss": 0.483,
176213
+ "step": 60400
176214
+ },
176215
+ {
176216
+ "epoch": 482.4,
176217
+ "learning_rate": 9.04834935897436e-06,
176218
+ "loss": 1.329,
176219
+ "step": 60405
176220
+ },
176221
+ {
176222
+ "epoch": 482.44,
176223
+ "learning_rate": 9.04826923076923e-06,
176224
+ "loss": 0.3322,
176225
+ "step": 60410
176226
+ },
176227
+ {
176228
+ "epoch": 482.48,
176229
+ "learning_rate": 9.048189102564103e-06,
176230
+ "loss": 0.3371,
176231
+ "step": 60415
176232
+ },
176233
+ {
176234
+ "epoch": 482.52,
176235
+ "learning_rate": 9.048108974358976e-06,
176236
+ "loss": 0.3224,
176237
+ "step": 60420
176238
+ },
176239
+ {
176240
+ "epoch": 482.56,
176241
+ "learning_rate": 9.048028846153846e-06,
176242
+ "loss": 0.5545,
176243
+ "step": 60425
176244
+ },
176245
+ {
176246
+ "epoch": 482.6,
176247
+ "learning_rate": 9.047948717948719e-06,
176248
+ "loss": 1.3621,
176249
+ "step": 60430
176250
+ },
176251
+ {
176252
+ "epoch": 482.64,
176253
+ "learning_rate": 9.04786858974359e-06,
176254
+ "loss": 0.3918,
176255
+ "step": 60435
176256
+ },
176257
+ {
176258
+ "epoch": 482.68,
176259
+ "learning_rate": 9.047788461538462e-06,
176260
+ "loss": 0.3828,
176261
+ "step": 60440
176262
+ },
176263
+ {
176264
+ "epoch": 482.72,
176265
+ "learning_rate": 9.047708333333333e-06,
176266
+ "loss": 0.3523,
176267
+ "step": 60445
176268
+ },
176269
+ {
176270
+ "epoch": 482.76,
176271
+ "learning_rate": 9.047628205128206e-06,
176272
+ "loss": 0.448,
176273
+ "step": 60450
176274
+ },
176275
+ {
176276
+ "epoch": 482.8,
176277
+ "learning_rate": 9.047548076923078e-06,
176278
+ "loss": 1.3147,
176279
+ "step": 60455
176280
+ },
176281
+ {
176282
+ "epoch": 482.84,
176283
+ "learning_rate": 9.047467948717949e-06,
176284
+ "loss": 0.3676,
176285
+ "step": 60460
176286
+ },
176287
+ {
176288
+ "epoch": 482.88,
176289
+ "learning_rate": 9.04738782051282e-06,
176290
+ "loss": 0.3621,
176291
+ "step": 60465
176292
+ },
176293
+ {
176294
+ "epoch": 482.92,
176295
+ "learning_rate": 9.047307692307693e-06,
176296
+ "loss": 0.3729,
176297
+ "step": 60470
176298
+ },
176299
+ {
176300
+ "epoch": 482.96,
176301
+ "learning_rate": 9.047227564102565e-06,
176302
+ "loss": 0.5614,
176303
+ "step": 60475
176304
+ },
176305
+ {
176306
+ "epoch": 483.0,
176307
+ "learning_rate": 9.047147435897436e-06,
176308
+ "loss": 1.4147,
176309
+ "step": 60480
176310
+ },
176311
+ {
176312
+ "epoch": 483.0,
176313
+ "eval_loss": 0.43452778458595276,
176314
+ "eval_runtime": 40.3349,
176315
+ "eval_samples_per_second": 20.776,
176316
+ "eval_steps_per_second": 0.669,
176317
+ "eval_wer": 0.19634636463280963,
176318
+ "step": 60480
176319
+ },
176320
+ {
176321
+ "epoch": 487.04,
176322
+ "learning_rate": 9.047067307692309e-06,
176323
+ "loss": 0.4135,
176324
+ "step": 60485
176325
+ },
176326
+ {
176327
+ "epoch": 487.08,
176328
+ "learning_rate": 9.04698717948718e-06,
176329
+ "loss": 0.347,
176330
+ "step": 60490
176331
+ },
176332
+ {
176333
+ "epoch": 487.12,
176334
+ "learning_rate": 9.046907051282052e-06,
176335
+ "loss": 0.5046,
176336
+ "step": 60495
176337
+ },
176338
+ {
176339
+ "epoch": 487.16,
176340
+ "learning_rate": 9.046826923076923e-06,
176341
+ "loss": 0.4958,
176342
+ "step": 60500
176343
+ },
176344
+ {
176345
+ "epoch": 487.2,
176346
+ "learning_rate": 9.046746794871796e-06,
176347
+ "loss": 1.1384,
176348
+ "step": 60505
176349
+ },
176350
+ {
176351
+ "epoch": 487.24,
176352
+ "learning_rate": 9.046666666666668e-06,
176353
+ "loss": 0.386,
176354
+ "step": 60510
176355
+ },
176356
+ {
176357
+ "epoch": 487.28,
176358
+ "learning_rate": 9.046586538461539e-06,
176359
+ "loss": 0.3101,
176360
+ "step": 60515
176361
+ },
176362
+ {
176363
+ "epoch": 487.32,
176364
+ "learning_rate": 9.046506410256412e-06,
176365
+ "loss": 0.3785,
176366
+ "step": 60520
176367
+ },
176368
+ {
176369
+ "epoch": 487.36,
176370
+ "learning_rate": 9.046426282051283e-06,
176371
+ "loss": 0.5308,
176372
+ "step": 60525
176373
+ },
176374
+ {
176375
+ "epoch": 487.4,
176376
+ "learning_rate": 9.046346153846155e-06,
176377
+ "loss": 1.2602,
176378
+ "step": 60530
176379
+ },
176380
+ {
176381
+ "epoch": 487.44,
176382
+ "learning_rate": 9.046266025641026e-06,
176383
+ "loss": 0.3258,
176384
+ "step": 60535
176385
+ },
176386
+ {
176387
+ "epoch": 487.48,
176388
+ "learning_rate": 9.046185897435899e-06,
176389
+ "loss": 0.305,
176390
+ "step": 60540
176391
+ },
176392
+ {
176393
+ "epoch": 487.52,
176394
+ "learning_rate": 9.04610576923077e-06,
176395
+ "loss": 0.336,
176396
+ "step": 60545
176397
+ },
176398
+ {
176399
+ "epoch": 487.56,
176400
+ "learning_rate": 9.046025641025642e-06,
176401
+ "loss": 0.5471,
176402
+ "step": 60550
176403
+ },
176404
+ {
176405
+ "epoch": 487.6,
176406
+ "learning_rate": 9.045945512820513e-06,
176407
+ "loss": 1.2149,
176408
+ "step": 60555
176409
+ },
176410
+ {
176411
+ "epoch": 487.65,
176412
+ "learning_rate": 9.045865384615386e-06,
176413
+ "loss": 0.3799,
176414
+ "step": 60560
176415
+ },
176416
+ {
176417
+ "epoch": 487.69,
176418
+ "learning_rate": 9.045785256410256e-06,
176419
+ "loss": 0.2942,
176420
+ "step": 60565
176421
+ },
176422
+ {
176423
+ "epoch": 487.73,
176424
+ "learning_rate": 9.045705128205129e-06,
176425
+ "loss": 0.3494,
176426
+ "step": 60570
176427
+ },
176428
+ {
176429
+ "epoch": 487.77,
176430
+ "learning_rate": 9.045625000000002e-06,
176431
+ "loss": 0.4582,
176432
+ "step": 60575
176433
+ },
176434
+ {
176435
+ "epoch": 487.81,
176436
+ "learning_rate": 9.045560897435898e-06,
176437
+ "loss": 1.3935,
176438
+ "step": 60580
176439
+ },
176440
+ {
176441
+ "epoch": 487.85,
176442
+ "learning_rate": 9.04548076923077e-06,
176443
+ "loss": 0.4015,
176444
+ "step": 60585
176445
+ },
176446
+ {
176447
+ "epoch": 487.89,
176448
+ "learning_rate": 9.045400641025642e-06,
176449
+ "loss": 0.2689,
176450
+ "step": 60590
176451
+ },
176452
+ {
176453
+ "epoch": 487.93,
176454
+ "learning_rate": 9.045320512820514e-06,
176455
+ "loss": 0.3782,
176456
+ "step": 60595
176457
+ },
176458
+ {
176459
+ "epoch": 487.97,
176460
+ "learning_rate": 9.045240384615387e-06,
176461
+ "loss": 0.5354,
176462
+ "step": 60600
176463
+ },
176464
+ {
176465
+ "epoch": 488.0,
176466
+ "eval_loss": 0.38258084654808044,
176467
+ "eval_runtime": 39.0437,
176468
+ "eval_samples_per_second": 21.463,
176469
+ "eval_steps_per_second": 0.692,
176470
+ "eval_wer": 0.19932432432432431,
176471
+ "step": 60604
176472
  }
176473
  ],
176474
  "max_steps": 620000,
176475
  "num_train_epochs": 5000,
176476
+ "total_flos": 1.7056102466424242e+20,
176477
  "trial_name": null,
176478
  "trial_params": null
176479
  }
model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60604}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60729}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60729}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:40b3314448fea63402a07aa602956be42c9e740284c01a33022761c3b6ed3b0c
3
  size 722165009
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0406650db1393fb9692f8d43b2e0d3fd7ac1ccba07e5597f49b7e009002820b6
3
  size 722165009
model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60729}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60729}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6666b72c33815713f5ef42633441bdc6ce6b93b09179c37e3c0c9ffe31501f52
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2839b98fe4b47ad20f5b8b1bae2552bd34df8084d7cb7ec7a9efadbef996fd76
3
  size 377909911
model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60729}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b4a1406c8db515233304a48cdeea6ce28b26f307346d9edec4a1f735f2edd837
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4b858c2d8d13cecbbedfc9bcb6fd9b338c4dab35127f6a04749c2971bdf13f5
3
  size 14503
model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60729}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d90403b234e2899cca64e4c4d1ecd5ac4db6b9c472764daea1d674f115d75b81
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8726c92bb7ef60e388546e9fd18baec8ae7c170a0be794a916a6e8ed8249c65c
3
  size 559
model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60729}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c1ac81a5988d8abd42a7dd17e61fda58638a7b0876ba21dfaf0d68642ddcb129
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd58c4ff7961fcf183c96f47a9e93eb6ea21e444f358cbb7cda92b3831704324
3
  size 623
model-bin/finetune/base/{checkpoint-55501 β†’ checkpoint-60729}/trainer_state.json RENAMED
The diff for this file is too large to render. See raw diff
 
model-bin/finetune/base/{checkpoint-60105 β†’ checkpoint-60729}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629798286.419865/events.out.tfevents.1629798286.c435e1c5ee04.920.101 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cdff19b05c5968077560f33eef23f6875ca4369a594138a22f689d4e5110b1f
3
+ size 4194
model-bin/finetune/base/log/1629798930.4451573/events.out.tfevents.1629798930.c435e1c5ee04.920.103 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3c25008b862a6b83c49da8da3e9cf89f99feeaa5e5dd15b7671b3a36095c3e1
3
+ size 4194
model-bin/finetune/base/log/1629799575.228714/events.out.tfevents.1629799575.c435e1c5ee04.920.105 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf181fc102c8aa4d48a39d2da285c497cfaf12a7875db52f035c81bfe74de9f2
3
+ size 4194
model-bin/finetune/base/log/1629800218.038527/events.out.tfevents.1629800218.c435e1c5ee04.920.107 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5f981abdb766a16f2729062859a1afb40cb90685c56de17aa042b8517f669a1
3
+ size 4194
model-bin/finetune/base/log/1629800871.1235416/events.out.tfevents.1629800871.c435e1c5ee04.920.109 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a20fd083fb3c7be7c7944010abc2e1b9fa23a305d56148813f97d5af1bb8c279
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629798286.c435e1c5ee04.920.100 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3f3ed544a80037d3a4a4ff4bf81759482a236f489fb1992f204e3aa98d05195
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629798930.c435e1c5ee04.920.102 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:377c33ce99cb86655a50d4fcd7b766ff3cb1dfce42c4d694049996756a81d123
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629799575.c435e1c5ee04.920.104 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22913b623bf7a65164b7fff2d9e6a4d08d222fd5509d4a2aa46b358922f1bfdf
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629800218.c435e1c5ee04.920.106 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cee26fb6ace669d99253f6d5365bdfff21dc68cef4451a6e5d3730bced531825
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1629800871.c435e1c5ee04.920.108 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbb8fff7bcd951291a1f98c7d9279705af59cd699ed617ad824f6084531a7843
3
+ size 8622