Check commited on
Commit
43ba46a
Β·
1 Parent(s): d905eab

"auto-commit"

Browse files
Files changed (28) hide show
  1. model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143240}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143240}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143240}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143240}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143240}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143240}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143240}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143240}/trainer_state.json +641 -5
  9. model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143240}/training_args.bin +0 -0
  10. model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143365}/config.json +0 -0
  11. model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143365}/optimizer.pt +1 -1
  12. model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143365}/preprocessor_config.json +0 -0
  13. model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143365}/pytorch_model.bin +1 -1
  14. model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143365}/rng_state.pth +1 -1
  15. model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143365}/scaler.pt +1 -1
  16. model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143365}/scheduler.pt +1 -1
  17. model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143365}/trainer_state.json +0 -0
  18. model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143365}/training_args.bin +0 -0
  19. model-bin/finetune/base/log/1630217087.7673898/events.out.tfevents.1630217087.cc93b136ebf5.1086.59 +3 -0
  20. model-bin/finetune/base/log/1630217533.1449723/events.out.tfevents.1630217533.cc93b136ebf5.1086.61 +3 -0
  21. model-bin/finetune/base/log/1630217981.5941153/events.out.tfevents.1630217981.cc93b136ebf5.1086.63 +3 -0
  22. model-bin/finetune/base/log/1630218424.3441827/events.out.tfevents.1630218424.cc93b136ebf5.1086.65 +3 -0
  23. model-bin/finetune/base/log/1630218868.8727193/events.out.tfevents.1630218868.cc93b136ebf5.1086.67 +3 -0
  24. model-bin/finetune/base/log/events.out.tfevents.1630217087.cc93b136ebf5.1086.58 +3 -0
  25. model-bin/finetune/base/log/events.out.tfevents.1630217533.cc93b136ebf5.1086.60 +3 -0
  26. model-bin/finetune/base/log/events.out.tfevents.1630217981.cc93b136ebf5.1086.62 +3 -0
  27. model-bin/finetune/base/log/events.out.tfevents.1630218424.cc93b136ebf5.1086.64 +3 -0
  28. model-bin/finetune/base/log/events.out.tfevents.1630218868.cc93b136ebf5.1086.66 +3 -0
model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143240}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143240}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7646d4658c7fb581271e3f3a4cd882f46c4e9eaff0c1d3ece8c4a2f60d2fe9b3
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b07eb70666c98efa81b1ae25d30d8dc0ef685547e99aa9c605653d13ffc1a26
3
  size 722165393
model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143240}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143240}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:60daade2e90914f426e13eda15e69959b6dfbcf964cfa92071ea99fff71da02f
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:719c64023422102f774c4b415dbe42084ebf730e04e436945dbc23a39747dc12
3
  size 377909911
model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143240}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e42e08f28ef8b1308862c7a353c1fb5d4b225eeb6cba470f033516a5bfc09a8a
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4d59aaa1d1129f5ce7777bdc711ff6665db55d55d5210162fcd2ecfe304ca32
3
  size 14503
model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143240}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:08811403bda30859be9404fa2f69eb552c512d466566eb146d6719d79f4b49f8
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:392bcd906a1be97f8bf6b539921d31e1244067c313fa0a9adb8dd019a4c75fa3
3
  size 559
model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143240}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b87ec296eb9fdd22d42513ded3472eba2b1440f44b4d69537e0b23fe4befe58a
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:317d976693bb649858811cc275bd77488d50936643f55da2776daa72308f421c
3
  size 623
model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143240}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
- "best_metric": 0.1689111747851003,
3
- "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
4
- "epoch": 1141.0,
5
- "global_step": 142742,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -280977,11 +280977,647 @@
280977
  "eval_steps_per_second": 0.655,
280978
  "eval_wer": 0.18999632217727105,
280979
  "step": 142742
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
280980
  }
280981
  ],
280982
  "max_steps": 625000,
280983
  "num_train_epochs": 5000,
280984
- "total_flos": 4.0167170527047575e+20,
280985
  "trial_name": null,
280986
  "trial_params": null
280987
  }
 
1
  {
2
+ "best_metric": 0.16819527695529718,
3
+ "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-143240",
4
+ "epoch": 1145.0,
5
+ "global_step": 143240,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
280977
  "eval_steps_per_second": 0.655,
280978
  "eval_wer": 0.18999632217727105,
280979
  "step": 142742
280980
+ },
280981
+ {
280982
+ "epoch": 1151.02,
280983
+ "learning_rate": 7.711017770597739e-06,
280984
+ "loss": 0.3134,
280985
+ "step": 142745
280986
+ },
280987
+ {
280988
+ "epoch": 1151.06,
280989
+ "learning_rate": 7.710936995153475e-06,
280990
+ "loss": 0.2984,
280991
+ "step": 142750
280992
+ },
280993
+ {
280994
+ "epoch": 1151.1,
280995
+ "learning_rate": 7.710856219709209e-06,
280996
+ "loss": 0.286,
280997
+ "step": 142755
280998
+ },
280999
+ {
281000
+ "epoch": 1151.14,
281001
+ "learning_rate": 7.710775444264945e-06,
281002
+ "loss": 0.2923,
281003
+ "step": 142760
281004
+ },
281005
+ {
281006
+ "epoch": 1151.18,
281007
+ "learning_rate": 7.71069466882068e-06,
281008
+ "loss": 0.7799,
281009
+ "step": 142765
281010
+ },
281011
+ {
281012
+ "epoch": 1151.22,
281013
+ "learning_rate": 7.710613893376414e-06,
281014
+ "loss": 0.8129,
281015
+ "step": 142770
281016
+ },
281017
+ {
281018
+ "epoch": 1151.27,
281019
+ "learning_rate": 7.71053311793215e-06,
281020
+ "loss": 0.3205,
281021
+ "step": 142775
281022
+ },
281023
+ {
281024
+ "epoch": 1151.31,
281025
+ "learning_rate": 7.710452342487884e-06,
281026
+ "loss": 0.2673,
281027
+ "step": 142780
281028
+ },
281029
+ {
281030
+ "epoch": 1151.35,
281031
+ "learning_rate": 7.71037156704362e-06,
281032
+ "loss": 0.3895,
281033
+ "step": 142785
281034
+ },
281035
+ {
281036
+ "epoch": 1151.39,
281037
+ "learning_rate": 7.710290791599354e-06,
281038
+ "loss": 0.6514,
281039
+ "step": 142790
281040
+ },
281041
+ {
281042
+ "epoch": 1151.43,
281043
+ "learning_rate": 7.71021001615509e-06,
281044
+ "loss": 0.8698,
281045
+ "step": 142795
281046
+ },
281047
+ {
281048
+ "epoch": 1151.47,
281049
+ "learning_rate": 7.710129240710824e-06,
281050
+ "loss": 0.283,
281051
+ "step": 142800
281052
+ },
281053
+ {
281054
+ "epoch": 1151.51,
281055
+ "learning_rate": 7.71004846526656e-06,
281056
+ "loss": 0.3581,
281057
+ "step": 142805
281058
+ },
281059
+ {
281060
+ "epoch": 1151.55,
281061
+ "learning_rate": 7.709967689822294e-06,
281062
+ "loss": 0.3568,
281063
+ "step": 142810
281064
+ },
281065
+ {
281066
+ "epoch": 1151.59,
281067
+ "learning_rate": 7.70988691437803e-06,
281068
+ "loss": 0.7716,
281069
+ "step": 142815
281070
+ },
281071
+ {
281072
+ "epoch": 1151.63,
281073
+ "learning_rate": 7.709806138933766e-06,
281074
+ "loss": 0.8879,
281075
+ "step": 142820
281076
+ },
281077
+ {
281078
+ "epoch": 1151.67,
281079
+ "learning_rate": 7.7097253634895e-06,
281080
+ "loss": 0.2742,
281081
+ "step": 142825
281082
+ },
281083
+ {
281084
+ "epoch": 1151.71,
281085
+ "learning_rate": 7.709644588045236e-06,
281086
+ "loss": 0.2852,
281087
+ "step": 142830
281088
+ },
281089
+ {
281090
+ "epoch": 1151.75,
281091
+ "learning_rate": 7.70956381260097e-06,
281092
+ "loss": 0.3155,
281093
+ "step": 142835
281094
+ },
281095
+ {
281096
+ "epoch": 1151.79,
281097
+ "learning_rate": 7.709483037156706e-06,
281098
+ "loss": 0.6991,
281099
+ "step": 142840
281100
+ },
281101
+ {
281102
+ "epoch": 1151.83,
281103
+ "learning_rate": 7.70940226171244e-06,
281104
+ "loss": 0.8038,
281105
+ "step": 142845
281106
+ },
281107
+ {
281108
+ "epoch": 1151.87,
281109
+ "learning_rate": 7.709321486268176e-06,
281110
+ "loss": 0.2299,
281111
+ "step": 142850
281112
+ },
281113
+ {
281114
+ "epoch": 1151.91,
281115
+ "learning_rate": 7.70924071082391e-06,
281116
+ "loss": 0.2772,
281117
+ "step": 142855
281118
+ },
281119
+ {
281120
+ "epoch": 1151.95,
281121
+ "learning_rate": 7.709159935379646e-06,
281122
+ "loss": 0.354,
281123
+ "step": 142860
281124
+ },
281125
+ {
281126
+ "epoch": 1151.99,
281127
+ "learning_rate": 7.70907915993538e-06,
281128
+ "loss": 0.7848,
281129
+ "step": 142865
281130
+ },
281131
+ {
281132
+ "epoch": 1152.0,
281133
+ "eval_loss": 0.3314845860004425,
281134
+ "eval_runtime": 41.7288,
281135
+ "eval_samples_per_second": 20.058,
281136
+ "eval_steps_per_second": 0.647,
281137
+ "eval_wer": 0.18063857705204842,
281138
+ "step": 142866
281139
+ },
281140
+ {
281141
+ "epoch": 1142.03,
281142
+ "learning_rate": 7.708998384491116e-06,
281143
+ "loss": 0.37,
281144
+ "step": 142870
281145
+ },
281146
+ {
281147
+ "epoch": 1142.07,
281148
+ "learning_rate": 7.708917609046852e-06,
281149
+ "loss": 0.268,
281150
+ "step": 142875
281151
+ },
281152
+ {
281153
+ "epoch": 1142.11,
281154
+ "learning_rate": 7.708836833602586e-06,
281155
+ "loss": 0.3045,
281156
+ "step": 142880
281157
+ },
281158
+ {
281159
+ "epoch": 1142.15,
281160
+ "learning_rate": 7.708756058158322e-06,
281161
+ "loss": 0.3686,
281162
+ "step": 142885
281163
+ },
281164
+ {
281165
+ "epoch": 1142.19,
281166
+ "learning_rate": 7.708675282714056e-06,
281167
+ "loss": 0.7276,
281168
+ "step": 142890
281169
+ },
281170
+ {
281171
+ "epoch": 1142.23,
281172
+ "learning_rate": 7.708594507269792e-06,
281173
+ "loss": 0.6929,
281174
+ "step": 142895
281175
+ },
281176
+ {
281177
+ "epoch": 1142.27,
281178
+ "learning_rate": 7.708513731825526e-06,
281179
+ "loss": 0.3062,
281180
+ "step": 142900
281181
+ },
281182
+ {
281183
+ "epoch": 1142.31,
281184
+ "learning_rate": 7.708432956381262e-06,
281185
+ "loss": 0.2826,
281186
+ "step": 142905
281187
+ },
281188
+ {
281189
+ "epoch": 1142.35,
281190
+ "learning_rate": 7.708352180936996e-06,
281191
+ "loss": 0.3328,
281192
+ "step": 142910
281193
+ },
281194
+ {
281195
+ "epoch": 1142.39,
281196
+ "learning_rate": 7.708271405492732e-06,
281197
+ "loss": 0.8634,
281198
+ "step": 142915
281199
+ },
281200
+ {
281201
+ "epoch": 1142.43,
281202
+ "learning_rate": 7.708190630048466e-06,
281203
+ "loss": 0.7105,
281204
+ "step": 142920
281205
+ },
281206
+ {
281207
+ "epoch": 1142.47,
281208
+ "learning_rate": 7.708109854604202e-06,
281209
+ "loss": 0.274,
281210
+ "step": 142925
281211
+ },
281212
+ {
281213
+ "epoch": 1142.51,
281214
+ "learning_rate": 7.708029079159936e-06,
281215
+ "loss": 0.3273,
281216
+ "step": 142930
281217
+ },
281218
+ {
281219
+ "epoch": 1142.55,
281220
+ "learning_rate": 7.707948303715672e-06,
281221
+ "loss": 0.41,
281222
+ "step": 142935
281223
+ },
281224
+ {
281225
+ "epoch": 1142.59,
281226
+ "learning_rate": 7.707867528271407e-06,
281227
+ "loss": 0.8705,
281228
+ "step": 142940
281229
+ },
281230
+ {
281231
+ "epoch": 1142.63,
281232
+ "learning_rate": 7.707786752827142e-06,
281233
+ "loss": 0.6095,
281234
+ "step": 142945
281235
+ },
281236
+ {
281237
+ "epoch": 1142.67,
281238
+ "learning_rate": 7.707705977382877e-06,
281239
+ "loss": 0.3094,
281240
+ "step": 142950
281241
+ },
281242
+ {
281243
+ "epoch": 1142.71,
281244
+ "learning_rate": 7.707625201938612e-06,
281245
+ "loss": 0.3067,
281246
+ "step": 142955
281247
+ },
281248
+ {
281249
+ "epoch": 1142.75,
281250
+ "learning_rate": 7.707544426494347e-06,
281251
+ "loss": 0.3958,
281252
+ "step": 142960
281253
+ },
281254
+ {
281255
+ "epoch": 1142.79,
281256
+ "learning_rate": 7.707463651050082e-06,
281257
+ "loss": 0.9498,
281258
+ "step": 142965
281259
+ },
281260
+ {
281261
+ "epoch": 1142.83,
281262
+ "learning_rate": 7.707382875605817e-06,
281263
+ "loss": 0.6889,
281264
+ "step": 142970
281265
+ },
281266
+ {
281267
+ "epoch": 1142.87,
281268
+ "learning_rate": 7.707302100161552e-06,
281269
+ "loss": 0.2645,
281270
+ "step": 142975
281271
+ },
281272
+ {
281273
+ "epoch": 1142.91,
281274
+ "learning_rate": 7.707221324717287e-06,
281275
+ "loss": 0.291,
281276
+ "step": 142980
281277
+ },
281278
+ {
281279
+ "epoch": 1142.95,
281280
+ "learning_rate": 7.707140549273021e-06,
281281
+ "loss": 0.4395,
281282
+ "step": 142985
281283
+ },
281284
+ {
281285
+ "epoch": 1142.99,
281286
+ "learning_rate": 7.707059773828757e-06,
281287
+ "loss": 1.038,
281288
+ "step": 142990
281289
+ },
281290
+ {
281291
+ "epoch": 1143.0,
281292
+ "eval_loss": 0.3770361542701721,
281293
+ "eval_runtime": 41.6024,
281294
+ "eval_samples_per_second": 20.143,
281295
+ "eval_steps_per_second": 0.649,
281296
+ "eval_wer": 0.176540196731011,
281297
+ "step": 142991
281298
+ },
281299
+ {
281300
+ "epoch": 1153.03,
281301
+ "learning_rate": 7.706978998384491e-06,
281302
+ "loss": 0.2919,
281303
+ "step": 142995
281304
+ },
281305
+ {
281306
+ "epoch": 1153.07,
281307
+ "learning_rate": 7.706898222940227e-06,
281308
+ "loss": 0.243,
281309
+ "step": 143000
281310
+ },
281311
+ {
281312
+ "epoch": 1153.11,
281313
+ "learning_rate": 7.706817447495963e-06,
281314
+ "loss": 0.2585,
281315
+ "step": 143005
281316
+ },
281317
+ {
281318
+ "epoch": 1153.15,
281319
+ "learning_rate": 7.706736672051697e-06,
281320
+ "loss": 0.4033,
281321
+ "step": 143010
281322
+ },
281323
+ {
281324
+ "epoch": 1153.19,
281325
+ "learning_rate": 7.706655896607433e-06,
281326
+ "loss": 0.854,
281327
+ "step": 143015
281328
+ },
281329
+ {
281330
+ "epoch": 1153.23,
281331
+ "learning_rate": 7.706575121163167e-06,
281332
+ "loss": 0.7609,
281333
+ "step": 143020
281334
+ },
281335
+ {
281336
+ "epoch": 1153.27,
281337
+ "learning_rate": 7.706494345718903e-06,
281338
+ "loss": 0.2552,
281339
+ "step": 143025
281340
+ },
281341
+ {
281342
+ "epoch": 1153.31,
281343
+ "learning_rate": 7.706413570274637e-06,
281344
+ "loss": 0.3037,
281345
+ "step": 143030
281346
+ },
281347
+ {
281348
+ "epoch": 1153.35,
281349
+ "learning_rate": 7.706332794830373e-06,
281350
+ "loss": 0.3576,
281351
+ "step": 143035
281352
+ },
281353
+ {
281354
+ "epoch": 1153.39,
281355
+ "learning_rate": 7.706252019386107e-06,
281356
+ "loss": 0.8974,
281357
+ "step": 143040
281358
+ },
281359
+ {
281360
+ "epoch": 1153.43,
281361
+ "learning_rate": 7.706171243941843e-06,
281362
+ "loss": 0.6808,
281363
+ "step": 143045
281364
+ },
281365
+ {
281366
+ "epoch": 1153.47,
281367
+ "learning_rate": 7.706090468497577e-06,
281368
+ "loss": 0.2624,
281369
+ "step": 143050
281370
+ },
281371
+ {
281372
+ "epoch": 1153.51,
281373
+ "learning_rate": 7.706009693053313e-06,
281374
+ "loss": 0.3187,
281375
+ "step": 143055
281376
+ },
281377
+ {
281378
+ "epoch": 1153.55,
281379
+ "learning_rate": 7.705928917609047e-06,
281380
+ "loss": 0.3324,
281381
+ "step": 143060
281382
+ },
281383
+ {
281384
+ "epoch": 1153.59,
281385
+ "learning_rate": 7.705848142164783e-06,
281386
+ "loss": 0.8548,
281387
+ "step": 143065
281388
+ },
281389
+ {
281390
+ "epoch": 1153.63,
281391
+ "learning_rate": 7.705767366720517e-06,
281392
+ "loss": 0.5806,
281393
+ "step": 143070
281394
+ },
281395
+ {
281396
+ "epoch": 1153.67,
281397
+ "learning_rate": 7.705686591276253e-06,
281398
+ "loss": 0.2566,
281399
+ "step": 143075
281400
+ },
281401
+ {
281402
+ "epoch": 1153.71,
281403
+ "learning_rate": 7.705605815831989e-06,
281404
+ "loss": 0.2849,
281405
+ "step": 143080
281406
+ },
281407
+ {
281408
+ "epoch": 1153.76,
281409
+ "learning_rate": 7.705525040387723e-06,
281410
+ "loss": 0.3476,
281411
+ "step": 143085
281412
+ },
281413
+ {
281414
+ "epoch": 1153.8,
281415
+ "learning_rate": 7.705444264943459e-06,
281416
+ "loss": 0.8085,
281417
+ "step": 143090
281418
+ },
281419
+ {
281420
+ "epoch": 1153.84,
281421
+ "learning_rate": 7.705363489499193e-06,
281422
+ "loss": 0.6814,
281423
+ "step": 143095
281424
+ },
281425
+ {
281426
+ "epoch": 1153.88,
281427
+ "learning_rate": 7.705282714054929e-06,
281428
+ "loss": 0.2852,
281429
+ "step": 143100
281430
+ },
281431
+ {
281432
+ "epoch": 1153.92,
281433
+ "learning_rate": 7.705201938610663e-06,
281434
+ "loss": 0.2844,
281435
+ "step": 143105
281436
+ },
281437
+ {
281438
+ "epoch": 1153.96,
281439
+ "learning_rate": 7.705121163166399e-06,
281440
+ "loss": 0.4279,
281441
+ "step": 143110
281442
+ },
281443
+ {
281444
+ "epoch": 1154.0,
281445
+ "learning_rate": 7.705040387722133e-06,
281446
+ "loss": 1.0141,
281447
+ "step": 143115
281448
+ },
281449
+ {
281450
+ "epoch": 1154.0,
281451
+ "eval_loss": 0.31175512075424194,
281452
+ "eval_runtime": 41.2685,
281453
+ "eval_samples_per_second": 20.306,
281454
+ "eval_steps_per_second": 0.654,
281455
+ "eval_wer": 0.1807184750733138,
281456
+ "step": 143115
281457
+ },
281458
+ {
281459
+ "epoch": 1144.04,
281460
+ "learning_rate": 7.704959612277869e-06,
281461
+ "loss": 0.3268,
281462
+ "step": 143120
281463
+ },
281464
+ {
281465
+ "epoch": 1144.08,
281466
+ "learning_rate": 7.704878836833603e-06,
281467
+ "loss": 0.2871,
281468
+ "step": 143125
281469
+ },
281470
+ {
281471
+ "epoch": 1144.12,
281472
+ "learning_rate": 7.704798061389339e-06,
281473
+ "loss": 0.2926,
281474
+ "step": 143130
281475
+ },
281476
+ {
281477
+ "epoch": 1144.16,
281478
+ "learning_rate": 7.704717285945073e-06,
281479
+ "loss": 0.4266,
281480
+ "step": 143135
281481
+ },
281482
+ {
281483
+ "epoch": 1144.2,
281484
+ "learning_rate": 7.704636510500809e-06,
281485
+ "loss": 1.2651,
281486
+ "step": 143140
281487
+ },
281488
+ {
281489
+ "epoch": 1144.24,
281490
+ "learning_rate": 7.704555735056544e-06,
281491
+ "loss": 0.3356,
281492
+ "step": 143145
281493
+ },
281494
+ {
281495
+ "epoch": 1144.28,
281496
+ "learning_rate": 7.704474959612279e-06,
281497
+ "loss": 0.2414,
281498
+ "step": 143150
281499
+ },
281500
+ {
281501
+ "epoch": 1144.32,
281502
+ "learning_rate": 7.704394184168014e-06,
281503
+ "loss": 0.2823,
281504
+ "step": 143155
281505
+ },
281506
+ {
281507
+ "epoch": 1144.36,
281508
+ "learning_rate": 7.704313408723749e-06,
281509
+ "loss": 0.4548,
281510
+ "step": 143160
281511
+ },
281512
+ {
281513
+ "epoch": 1144.4,
281514
+ "learning_rate": 7.704232633279484e-06,
281515
+ "loss": 1.1442,
281516
+ "step": 143165
281517
+ },
281518
+ {
281519
+ "epoch": 1144.44,
281520
+ "learning_rate": 7.704151857835219e-06,
281521
+ "loss": 0.2731,
281522
+ "step": 143170
281523
+ },
281524
+ {
281525
+ "epoch": 1144.48,
281526
+ "learning_rate": 7.704071082390954e-06,
281527
+ "loss": 0.2685,
281528
+ "step": 143175
281529
+ },
281530
+ {
281531
+ "epoch": 1144.52,
281532
+ "learning_rate": 7.703990306946689e-06,
281533
+ "loss": 0.2771,
281534
+ "step": 143180
281535
+ },
281536
+ {
281537
+ "epoch": 1144.56,
281538
+ "learning_rate": 7.703909531502424e-06,
281539
+ "loss": 0.4552,
281540
+ "step": 143185
281541
+ },
281542
+ {
281543
+ "epoch": 1144.6,
281544
+ "learning_rate": 7.703828756058159e-06,
281545
+ "loss": 1.3455,
281546
+ "step": 143190
281547
+ },
281548
+ {
281549
+ "epoch": 1144.64,
281550
+ "learning_rate": 7.703747980613894e-06,
281551
+ "loss": 0.3369,
281552
+ "step": 143195
281553
+ },
281554
+ {
281555
+ "epoch": 1144.68,
281556
+ "learning_rate": 7.703667205169628e-06,
281557
+ "loss": 0.2937,
281558
+ "step": 143200
281559
+ },
281560
+ {
281561
+ "epoch": 1144.72,
281562
+ "learning_rate": 7.703586429725364e-06,
281563
+ "loss": 0.3249,
281564
+ "step": 143205
281565
+ },
281566
+ {
281567
+ "epoch": 1144.76,
281568
+ "learning_rate": 7.7035056542811e-06,
281569
+ "loss": 0.4669,
281570
+ "step": 143210
281571
+ },
281572
+ {
281573
+ "epoch": 1144.8,
281574
+ "learning_rate": 7.703424878836834e-06,
281575
+ "loss": 1.1763,
281576
+ "step": 143215
281577
+ },
281578
+ {
281579
+ "epoch": 1144.84,
281580
+ "learning_rate": 7.70334410339257e-06,
281581
+ "loss": 0.2941,
281582
+ "step": 143220
281583
+ },
281584
+ {
281585
+ "epoch": 1144.88,
281586
+ "learning_rate": 7.703263327948304e-06,
281587
+ "loss": 0.4048,
281588
+ "step": 143225
281589
+ },
281590
+ {
281591
+ "epoch": 1144.92,
281592
+ "learning_rate": 7.70318255250404e-06,
281593
+ "loss": 0.3128,
281594
+ "step": 143230
281595
+ },
281596
+ {
281597
+ "epoch": 1144.96,
281598
+ "learning_rate": 7.703101777059774e-06,
281599
+ "loss": 0.4631,
281600
+ "step": 143235
281601
+ },
281602
+ {
281603
+ "epoch": 1145.0,
281604
+ "learning_rate": 7.70302100161551e-06,
281605
+ "loss": 1.2677,
281606
+ "step": 143240
281607
+ },
281608
+ {
281609
+ "epoch": 1145.0,
281610
+ "eval_loss": 0.35869625210762024,
281611
+ "eval_runtime": 41.8075,
281612
+ "eval_samples_per_second": 20.044,
281613
+ "eval_steps_per_second": 0.646,
281614
+ "eval_wer": 0.16819527695529718,
281615
+ "step": 143240
281616
  }
281617
  ],
281618
  "max_steps": 625000,
281619
  "num_train_epochs": 5000,
281620
+ "total_flos": 4.0307380836057815e+20,
281621
  "trial_name": null,
281622
  "trial_params": null
281623
  }
model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143240}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143365}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143365}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:13943fb05087ba1925fc85cf6d8ccdcd436e67bce7b73387491c31d71a0d9a4d
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b4ae2f824fc44e74d5f922c65c684da85a296c2dd928113161d2e4f8fd01f68
3
  size 722165393
model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143365}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143365}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ab98858ccc2a45c3fd5e7878c3e771f33185de7cf38be497d83aacd345b6bc15
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e1221ef028ae6ba23c0df2fe3cfe15fa48e2d55a99582cb94df564ba0f055f1
3
  size 377909911
model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143365}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fbf964a925c16111bd00a1243d38524b685492690de876b21aa0b98f4a2bdd2b
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67d1a327091bff6c6eb8436fcc27644978a972446a5eebb7ff5c3adc62b63046
3
  size 14503
model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143365}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:72f28b93e2cec7a5a1f715c7d28d98c2d935e7a77f9f7576f9c998aa09c5d067
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbaca6535201eae6dc1d868a0f5be4c8789d1fe82762cfc6215c355aacb74f62
3
  size 559
model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143365}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ef285b8f55ac571b7d259714b492a3bca4226307e36590c40ca749e5758e1e7
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2b05ae9c3a5bf446ba77acdd3c9ade4ed6e50b98322ae9d0848f215de2ab73f
3
  size 623
model-bin/finetune/base/{checkpoint-132910 β†’ checkpoint-143365}/trainer_state.json RENAMED
The diff for this file is too large to render. See raw diff
 
model-bin/finetune/base/{checkpoint-142742 β†’ checkpoint-143365}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630217087.7673898/events.out.tfevents.1630217087.cc93b136ebf5.1086.59 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09b8084a2ac97413aaf76a88b7e3092a97e277067c3796295860c9c4caf09d02
3
+ size 4194
model-bin/finetune/base/log/1630217533.1449723/events.out.tfevents.1630217533.cc93b136ebf5.1086.61 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21212778424bd51abb4aacfd5252daebd31ce1f97f1f09ea1fcce286c46f9c93
3
+ size 4194
model-bin/finetune/base/log/1630217981.5941153/events.out.tfevents.1630217981.cc93b136ebf5.1086.63 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d6b8848122be1124612dcd0db04e18faf74347bacb94c76b956338a454d50ad
3
+ size 4194
model-bin/finetune/base/log/1630218424.3441827/events.out.tfevents.1630218424.cc93b136ebf5.1086.65 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c924e5278bf26844662007c8bd5145ebe4ea3994598afbebe9e83f677faefec
3
+ size 4194
model-bin/finetune/base/log/1630218868.8727193/events.out.tfevents.1630218868.cc93b136ebf5.1086.67 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b19fa218f1be9449a5d32a64885bfecc29770580523ffc10dfeed6ccb873be0
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630217087.cc93b136ebf5.1086.58 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:907cf29111876985d236cafe3f953035e84c59b474c9dbd9dff4e969ab83f1c3
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630217533.cc93b136ebf5.1086.60 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee0d13fac16746b67f94ac3c316b9c00b8b3dbf21d6bcf5e02048b7b1a93ea26
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630217981.cc93b136ebf5.1086.62 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8b3555234e7b0bc800f9349f325108795cf1047ae6183e6ed8e6afb5c0295d9
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630218424.cc93b136ebf5.1086.64 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fccc790c4cb6a6b399c77a5a7fccd5ca694c52a3a290cd8636074e64786e9d37
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630218868.cc93b136ebf5.1086.66 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ae733cf523888acb79e1284050d06ee1259396ee832443898d7c9e8c29a64f9
3
+ size 8622