davidmeikle commited on
Commit
739b760
·
verified ·
1 Parent(s): 36f24a0

Upload folder using huggingface_hub

Browse files
bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7846f689d2b2b333dac0ed54353579e7846b95ff52ec066dabb6a20bb8ece8d2
3
+ size 493882
config.json ADDED
@@ -0,0 +1,280 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "transforms": [
3
+ "huggingface_tokenize"
4
+ ],
5
+ "src_vocab_size": 51865,
6
+ "data": null,
7
+ "tgt_vocab_size": 51865,
8
+ "decoder_start_token": "<|startoftranscript|>",
9
+ "src_vocab": null,
10
+ "n_sample": 0,
11
+ "pad_token": "<|endoftext|>",
12
+ "vocab_size_multiple": 8,
13
+ "save_data": null,
14
+ "bos_token": "<|endoftext|>",
15
+ "tgt_vocab": null,
16
+ "skip_empty_level": "silent",
17
+ "eos_token": "<|endoftext|>",
18
+ "unk_token": "<|endoftext|>",
19
+ "share_vocab": true,
20
+ "training": {
21
+ "normalization": "tokens",
22
+ "batch_type": "tokens",
23
+ "accum_steps": [
24
+ 0
25
+ ],
26
+ "valid_batch_size": 256,
27
+ "quant_type": "",
28
+ "batch_size_multiple": 1,
29
+ "group_size": 0,
30
+ "quant_layers": [],
31
+ "w_bit": 0,
32
+ "batch_size": 896,
33
+ "compute_dtype": "torch.float32",
34
+ "accum_count": [
35
+ 32
36
+ ]
37
+ },
38
+ "model": {
39
+ "sliding_window": 4096,
40
+ "no_timestamps_token_id": 50363,
41
+ "first_k_dense_replace": 0,
42
+ "attn_scaling": null,
43
+ "layer_norm": "standard",
44
+ "layers": 12,
45
+ "num_experts": 0,
46
+ "adapter_bias": false,
47
+ "num_shared_experts": 0,
48
+ "add_final_linear_bias": true,
49
+ "share_decoder_embeddings": true,
50
+ "left_pad": false,
51
+ "num_experts_per_tok": 0,
52
+ "transformer_ff": 3072,
53
+ "architecture": "whisper",
54
+ "suppress_tokens": [
55
+ 1,
56
+ 2,
57
+ 7,
58
+ 8,
59
+ 9,
60
+ 10,
61
+ 14,
62
+ 25,
63
+ 26,
64
+ 27,
65
+ 28,
66
+ 29,
67
+ 31,
68
+ 58,
69
+ 59,
70
+ 60,
71
+ 61,
72
+ 62,
73
+ 63,
74
+ 90,
75
+ 91,
76
+ 92,
77
+ 93,
78
+ 359,
79
+ 503,
80
+ 522,
81
+ 542,
82
+ 873,
83
+ 893,
84
+ 902,
85
+ 918,
86
+ 922,
87
+ 931,
88
+ 1350,
89
+ 1853,
90
+ 1982,
91
+ 2460,
92
+ 2627,
93
+ 3246,
94
+ 3253,
95
+ 3268,
96
+ 3536,
97
+ 3846,
98
+ 3961,
99
+ 4183,
100
+ 4667,
101
+ 6585,
102
+ 6647,
103
+ 7273,
104
+ 9061,
105
+ 9383,
106
+ 10428,
107
+ 10929,
108
+ 11938,
109
+ 12033,
110
+ 12331,
111
+ 12562,
112
+ 13793,
113
+ 14157,
114
+ 14635,
115
+ 15265,
116
+ 15618,
117
+ 16553,
118
+ 16604,
119
+ 18362,
120
+ 18956,
121
+ 20075,
122
+ 21675,
123
+ 22520,
124
+ 26130,
125
+ 26161,
126
+ 26435,
127
+ 28279,
128
+ 29464,
129
+ 31650,
130
+ 32302,
131
+ 32470,
132
+ 36865,
133
+ 42863,
134
+ 47425,
135
+ 49870,
136
+ 50254,
137
+ 50258,
138
+ 50358,
139
+ 50359,
140
+ 50360,
141
+ 50361,
142
+ 50362
143
+ ],
144
+ "parallel_residual": false,
145
+ "position_encoding_type": "Learned",
146
+ "add_ffnbias": true,
147
+ "shared_layer_norm": false,
148
+ "norm_eps": 1e-05,
149
+ "heads": 12,
150
+ "heads_kv": 12,
151
+ "huggingface_model": "openai/whisper-small",
152
+ "mlp_activation_fn": "gelu",
153
+ "hidden_size": 768,
154
+ "add_key_bias": false,
155
+ "word_timestamp_heads": [
156
+ [
157
+ 5,
158
+ 3
159
+ ],
160
+ [
161
+ 5,
162
+ 9
163
+ ],
164
+ [
165
+ 8,
166
+ 0
167
+ ],
168
+ [
169
+ 8,
170
+ 4
171
+ ],
172
+ [
173
+ 8,
174
+ 7
175
+ ],
176
+ [
177
+ 8,
178
+ 8
179
+ ],
180
+ [
181
+ 9,
182
+ 0
183
+ ],
184
+ [
185
+ 9,
186
+ 7
187
+ ],
188
+ [
189
+ 9,
190
+ 9
191
+ ],
192
+ [
193
+ 10,
194
+ 5
195
+ ]
196
+ ],
197
+ "add_qkvbias": true,
198
+ "moe_transformer_ff": null,
199
+ "head_dim": null,
200
+ "begin_suppress_tokens": [
201
+ 220,
202
+ 50257
203
+ ],
204
+ "generator_bias": false,
205
+ "encoder": {
206
+ "src_word_vec_size": 768,
207
+ "norm_eps": 1e-05,
208
+ "sliding_window": 4096,
209
+ "heads": 12,
210
+ "heads_kv": 12,
211
+ "max_source_positions": 1500,
212
+ "encoder_type": "whisper",
213
+ "first_k_dense_replace": 0,
214
+ "mlp_activation_fn": "gelu",
215
+ "hidden_size": 768,
216
+ "attn_scaling": null,
217
+ "add_key_bias": false,
218
+ "add_qkvbias": true,
219
+ "layer_norm": "standard",
220
+ "layers": 12,
221
+ "num_mel_bins": 80,
222
+ "moe_transformer_ff": null,
223
+ "head_dim": null,
224
+ "num_experts": 0,
225
+ "num_shared_experts": 0,
226
+ "add_final_linear_bias": true,
227
+ "num_experts_per_tok": 0,
228
+ "transformer_ff": 3072,
229
+ "position_encoding_type": null,
230
+ "parallel_residual": false,
231
+ "add_ffnbias": true,
232
+ "shared_layer_norm": false
233
+ },
234
+ "embeddings": {
235
+ "src_word_vec_size": 768,
236
+ "tgt_word_vec_size": 768,
237
+ "n_positions": 448,
238
+ "position_encoding_type": "Learned"
239
+ },
240
+ "decoder": {
241
+ "norm_eps": 1e-05,
242
+ "n_positions": 448,
243
+ "sliding_window": 4096,
244
+ "heads": 12,
245
+ "heads_kv": 12,
246
+ "first_k_dense_replace": 0,
247
+ "mlp_activation_fn": "gelu",
248
+ "tgt_word_vec_size": 768,
249
+ "hidden_size": 768,
250
+ "attn_scaling": null,
251
+ "decoder_type": "transformer",
252
+ "add_key_bias": false,
253
+ "add_qkvbias": true,
254
+ "layers": 12,
255
+ "layer_norm": "standard",
256
+ "moe_transformer_ff": null,
257
+ "head_dim": null,
258
+ "num_experts": 0,
259
+ "num_shared_experts": 0,
260
+ "add_final_linear_bias": true,
261
+ "num_experts_per_tok": 0,
262
+ "transformer_ff": 3072,
263
+ "position_encoding_type": "Learned",
264
+ "parallel_residual": false,
265
+ "add_ffnbias": true,
266
+ "shared_layer_norm": false
267
+ }
268
+ },
269
+ "transforms_configs": {
270
+ "huggingface_tokenize": {
271
+ "huggingface_model": "openai/whisper-small",
272
+ "path": "/scratch/models/whisper-small-eole/tokenizer.json"
273
+ }
274
+ },
275
+ "inference": {
276
+ "optional_eos": [],
277
+ "max_length": 448,
278
+ "chat_template": null
279
+ }
280
+ }
generation_config.json ADDED
@@ -0,0 +1,264 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alignment_heads": [
3
+ [
4
+ 5,
5
+ 3
6
+ ],
7
+ [
8
+ 5,
9
+ 9
10
+ ],
11
+ [
12
+ 8,
13
+ 0
14
+ ],
15
+ [
16
+ 8,
17
+ 4
18
+ ],
19
+ [
20
+ 8,
21
+ 7
22
+ ],
23
+ [
24
+ 8,
25
+ 8
26
+ ],
27
+ [
28
+ 9,
29
+ 0
30
+ ],
31
+ [
32
+ 9,
33
+ 7
34
+ ],
35
+ [
36
+ 9,
37
+ 9
38
+ ],
39
+ [
40
+ 10,
41
+ 5
42
+ ]
43
+ ],
44
+ "begin_suppress_tokens": [
45
+ 220,
46
+ 50257
47
+ ],
48
+ "bos_token_id": 50257,
49
+ "decoder_start_token_id": 50258,
50
+ "eos_token_id": 50257,
51
+ "forced_decoder_ids": [
52
+ [
53
+ 1,
54
+ null
55
+ ],
56
+ [
57
+ 2,
58
+ 50359
59
+ ]
60
+ ],
61
+ "is_multilingual": true,
62
+ "lang_to_id": {
63
+ "<|af|>": 50327,
64
+ "<|am|>": 50334,
65
+ "<|ar|>": 50272,
66
+ "<|as|>": 50350,
67
+ "<|az|>": 50304,
68
+ "<|ba|>": 50355,
69
+ "<|be|>": 50330,
70
+ "<|bg|>": 50292,
71
+ "<|bn|>": 50302,
72
+ "<|bo|>": 50347,
73
+ "<|br|>": 50309,
74
+ "<|bs|>": 50315,
75
+ "<|ca|>": 50270,
76
+ "<|cs|>": 50283,
77
+ "<|cy|>": 50297,
78
+ "<|da|>": 50285,
79
+ "<|de|>": 50261,
80
+ "<|el|>": 50281,
81
+ "<|en|>": 50259,
82
+ "<|es|>": 50262,
83
+ "<|et|>": 50307,
84
+ "<|eu|>": 50310,
85
+ "<|fa|>": 50300,
86
+ "<|fi|>": 50277,
87
+ "<|fo|>": 50338,
88
+ "<|fr|>": 50265,
89
+ "<|gl|>": 50319,
90
+ "<|gu|>": 50333,
91
+ "<|haw|>": 50352,
92
+ "<|ha|>": 50354,
93
+ "<|he|>": 50279,
94
+ "<|hi|>": 50276,
95
+ "<|hr|>": 50291,
96
+ "<|ht|>": 50339,
97
+ "<|hu|>": 50286,
98
+ "<|hy|>": 50312,
99
+ "<|id|>": 50275,
100
+ "<|is|>": 50311,
101
+ "<|it|>": 50274,
102
+ "<|ja|>": 50266,
103
+ "<|jw|>": 50356,
104
+ "<|ka|>": 50329,
105
+ "<|kk|>": 50316,
106
+ "<|km|>": 50323,
107
+ "<|kn|>": 50306,
108
+ "<|ko|>": 50264,
109
+ "<|la|>": 50294,
110
+ "<|lb|>": 50345,
111
+ "<|ln|>": 50353,
112
+ "<|lo|>": 50336,
113
+ "<|lt|>": 50293,
114
+ "<|lv|>": 50301,
115
+ "<|mg|>": 50349,
116
+ "<|mi|>": 50295,
117
+ "<|mk|>": 50308,
118
+ "<|ml|>": 50296,
119
+ "<|mn|>": 50314,
120
+ "<|mr|>": 50320,
121
+ "<|ms|>": 50282,
122
+ "<|mt|>": 50343,
123
+ "<|my|>": 50346,
124
+ "<|ne|>": 50313,
125
+ "<|nl|>": 50271,
126
+ "<|nn|>": 50342,
127
+ "<|no|>": 50288,
128
+ "<|oc|>": 50328,
129
+ "<|pa|>": 50321,
130
+ "<|pl|>": 50269,
131
+ "<|ps|>": 50340,
132
+ "<|pt|>": 50267,
133
+ "<|ro|>": 50284,
134
+ "<|ru|>": 50263,
135
+ "<|sa|>": 50344,
136
+ "<|sd|>": 50332,
137
+ "<|si|>": 50322,
138
+ "<|sk|>": 50298,
139
+ "<|sl|>": 50305,
140
+ "<|sn|>": 50324,
141
+ "<|so|>": 50326,
142
+ "<|sq|>": 50317,
143
+ "<|sr|>": 50303,
144
+ "<|su|>": 50357,
145
+ "<|sv|>": 50273,
146
+ "<|sw|>": 50318,
147
+ "<|ta|>": 50287,
148
+ "<|te|>": 50299,
149
+ "<|tg|>": 50331,
150
+ "<|th|>": 50289,
151
+ "<|tk|>": 50341,
152
+ "<|tl|>": 50348,
153
+ "<|tr|>": 50268,
154
+ "<|tt|>": 50351,
155
+ "<|uk|>": 50280,
156
+ "<|ur|>": 50290,
157
+ "<|uz|>": 50337,
158
+ "<|vi|>": 50278,
159
+ "<|yi|>": 50335,
160
+ "<|yo|>": 50325,
161
+ "<|zh|>": 50260
162
+ },
163
+ "max_initial_timestamp_index": 50,
164
+ "max_length": 448,
165
+ "no_timestamps_token_id": 50363,
166
+ "pad_token_id": 50257,
167
+ "prev_sot_token_id": 50361,
168
+ "return_timestamps": false,
169
+ "suppress_tokens": [
170
+ 1,
171
+ 2,
172
+ 7,
173
+ 8,
174
+ 9,
175
+ 10,
176
+ 14,
177
+ 25,
178
+ 26,
179
+ 27,
180
+ 28,
181
+ 29,
182
+ 31,
183
+ 58,
184
+ 59,
185
+ 60,
186
+ 61,
187
+ 62,
188
+ 63,
189
+ 90,
190
+ 91,
191
+ 92,
192
+ 93,
193
+ 359,
194
+ 503,
195
+ 522,
196
+ 542,
197
+ 873,
198
+ 893,
199
+ 902,
200
+ 918,
201
+ 922,
202
+ 931,
203
+ 1350,
204
+ 1853,
205
+ 1982,
206
+ 2460,
207
+ 2627,
208
+ 3246,
209
+ 3253,
210
+ 3268,
211
+ 3536,
212
+ 3846,
213
+ 3961,
214
+ 4183,
215
+ 4667,
216
+ 6585,
217
+ 6647,
218
+ 7273,
219
+ 9061,
220
+ 9383,
221
+ 10428,
222
+ 10929,
223
+ 11938,
224
+ 12033,
225
+ 12331,
226
+ 12562,
227
+ 13793,
228
+ 14157,
229
+ 14635,
230
+ 15265,
231
+ 15618,
232
+ 16553,
233
+ 16604,
234
+ 18362,
235
+ 18956,
236
+ 20075,
237
+ 21675,
238
+ 22520,
239
+ 26130,
240
+ 26161,
241
+ 26435,
242
+ 28279,
243
+ 29464,
244
+ 31650,
245
+ 32302,
246
+ 32470,
247
+ 36865,
248
+ 42863,
249
+ 47425,
250
+ 49870,
251
+ 50254,
252
+ 50258,
253
+ 50358,
254
+ 50359,
255
+ 50360,
256
+ 50361,
257
+ 50362
258
+ ],
259
+ "task_to_id": {
260
+ "transcribe": 50359,
261
+ "translate": 50358
262
+ },
263
+ "transformers_version": "4.31.0.dev0"
264
+ }
model.00.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62ef4c0c1930adb29d1dc517d7d756258398cc69a0db26edf7e135cb4fd1c023
3
+ size 967000200
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d7734884874f1a1513ed9aa760a4f8e97aaa02fd6d93a3a85d27b2ae9ca596b
3
+ size 966995080
special_tokens_map.json ADDED
@@ -0,0 +1,139 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|endoftext|>",
4
+ "<|startoftranscript|>",
5
+ "<|en|>",
6
+ "<|zh|>",
7
+ "<|de|>",
8
+ "<|es|>",
9
+ "<|ru|>",
10
+ "<|ko|>",
11
+ "<|fr|>",
12
+ "<|ja|>",
13
+ "<|pt|>",
14
+ "<|tr|>",
15
+ "<|pl|>",
16
+ "<|ca|>",
17
+ "<|nl|>",
18
+ "<|ar|>",
19
+ "<|sv|>",
20
+ "<|it|>",
21
+ "<|id|>",
22
+ "<|hi|>",
23
+ "<|fi|>",
24
+ "<|vi|>",
25
+ "<|he|>",
26
+ "<|uk|>",
27
+ "<|el|>",
28
+ "<|ms|>",
29
+ "<|cs|>",
30
+ "<|ro|>",
31
+ "<|da|>",
32
+ "<|hu|>",
33
+ "<|ta|>",
34
+ "<|no|>",
35
+ "<|th|>",
36
+ "<|ur|>",
37
+ "<|hr|>",
38
+ "<|bg|>",
39
+ "<|lt|>",
40
+ "<|la|>",
41
+ "<|mi|>",
42
+ "<|ml|>",
43
+ "<|cy|>",
44
+ "<|sk|>",
45
+ "<|te|>",
46
+ "<|fa|>",
47
+ "<|lv|>",
48
+ "<|bn|>",
49
+ "<|sr|>",
50
+ "<|az|>",
51
+ "<|sl|>",
52
+ "<|kn|>",
53
+ "<|et|>",
54
+ "<|mk|>",
55
+ "<|br|>",
56
+ "<|eu|>",
57
+ "<|is|>",
58
+ "<|hy|>",
59
+ "<|ne|>",
60
+ "<|mn|>",
61
+ "<|bs|>",
62
+ "<|kk|>",
63
+ "<|sq|>",
64
+ "<|sw|>",
65
+ "<|gl|>",
66
+ "<|mr|>",
67
+ "<|pa|>",
68
+ "<|si|>",
69
+ "<|km|>",
70
+ "<|sn|>",
71
+ "<|yo|>",
72
+ "<|so|>",
73
+ "<|af|>",
74
+ "<|oc|>",
75
+ "<|ka|>",
76
+ "<|be|>",
77
+ "<|tg|>",
78
+ "<|sd|>",
79
+ "<|gu|>",
80
+ "<|am|>",
81
+ "<|yi|>",
82
+ "<|lo|>",
83
+ "<|uz|>",
84
+ "<|fo|>",
85
+ "<|ht|>",
86
+ "<|ps|>",
87
+ "<|tk|>",
88
+ "<|nn|>",
89
+ "<|mt|>",
90
+ "<|sa|>",
91
+ "<|lb|>",
92
+ "<|my|>",
93
+ "<|bo|>",
94
+ "<|tl|>",
95
+ "<|mg|>",
96
+ "<|as|>",
97
+ "<|tt|>",
98
+ "<|haw|>",
99
+ "<|ln|>",
100
+ "<|ha|>",
101
+ "<|ba|>",
102
+ "<|jw|>",
103
+ "<|su|>",
104
+ "<|translate|>",
105
+ "<|transcribe|>",
106
+ "<|startoflm|>",
107
+ "<|startofprev|>",
108
+ "<|nocaptions|>",
109
+ "<|notimestamps|>"
110
+ ],
111
+ "bos_token": {
112
+ "content": "<|endoftext|>",
113
+ "lstrip": false,
114
+ "normalized": false,
115
+ "rstrip": false,
116
+ "single_word": false
117
+ },
118
+ "eos_token": {
119
+ "content": "<|endoftext|>",
120
+ "lstrip": false,
121
+ "normalized": false,
122
+ "rstrip": false,
123
+ "single_word": false
124
+ },
125
+ "pad_token": {
126
+ "content": "<|endoftext|>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false
131
+ },
132
+ "unk_token": {
133
+ "content": "<|endoftext|>",
134
+ "lstrip": false,
135
+ "normalized": false,
136
+ "rstrip": false,
137
+ "single_word": false
138
+ }
139
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
vocab.txt ADDED
The diff for this file is too large to render. See raw diff