hans00 commited on
Commit
60093f1
·
verified ·
1 Parent(s): bb74452

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,7 +1,8 @@
1
  {
2
- "_name_or_path": "openai/whisper-small",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
 
5
  "architectures": [
6
  "WhisperForConditionalGeneration"
7
  ],
@@ -11,6 +12,7 @@
11
  50257
12
  ],
13
  "bos_token_id": 50257,
 
14
  "d_model": 768,
15
  "decoder_attention_heads": 12,
16
  "decoder_ffn_dim": 3072,
@@ -39,9 +41,16 @@
39
  ],
40
  "init_std": 0.02,
41
  "is_encoder_decoder": true,
 
 
 
 
 
 
42
  "max_length": 448,
43
  "max_source_positions": 1500,
44
  "max_target_positions": 448,
 
45
  "model_type": "whisper",
46
  "num_hidden_layers": 12,
47
  "num_mel_bins": 80,
@@ -135,8 +144,8 @@
135
  50361,
136
  50362
137
  ],
138
- "torch_dtype": "float32",
139
- "transformers_version": "4.27.0.dev0",
140
  "use_cache": true,
 
141
  "vocab_size": 51865
142
  }
 
1
  {
2
+ "_name_or_path": "whisper-small",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
+ "apply_spec_augment": false,
6
  "architectures": [
7
  "WhisperForConditionalGeneration"
8
  ],
 
12
  50257
13
  ],
14
  "bos_token_id": 50257,
15
+ "classifier_proj_size": 256,
16
  "d_model": 768,
17
  "decoder_attention_heads": 12,
18
  "decoder_ffn_dim": 3072,
 
41
  ],
42
  "init_std": 0.02,
43
  "is_encoder_decoder": true,
44
+ "mask_feature_length": 10,
45
+ "mask_feature_min_masks": 0,
46
+ "mask_feature_prob": 0.0,
47
+ "mask_time_length": 10,
48
+ "mask_time_min_masks": 2,
49
+ "mask_time_prob": 0.05,
50
  "max_length": 448,
51
  "max_source_positions": 1500,
52
  "max_target_positions": 448,
53
+ "median_filter_width": 7,
54
  "model_type": "whisper",
55
  "num_hidden_layers": 12,
56
  "num_mel_bins": 80,
 
144
  50361,
145
  50362
146
  ],
147
+ "transformers_version": "4.33.1",
 
148
  "use_cache": true,
149
+ "use_weighted_layer_sum": false,
150
  "vocab_size": 51865
151
  }
generation_config.json CHANGED
@@ -1,4 +1,46 @@
1
  {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  "begin_suppress_tokens": [
3
  220,
4
  50257
@@ -217,5 +259,5 @@
217
  "transcribe": 50359,
218
  "translate": 50358
219
  },
220
- "transformers_version": "4.27.0.dev0"
221
- }
 
1
  {
2
+ "alignment_heads": [
3
+ [
4
+ 5,
5
+ 3
6
+ ],
7
+ [
8
+ 5,
9
+ 9
10
+ ],
11
+ [
12
+ 8,
13
+ 0
14
+ ],
15
+ [
16
+ 8,
17
+ 4
18
+ ],
19
+ [
20
+ 8,
21
+ 7
22
+ ],
23
+ [
24
+ 8,
25
+ 8
26
+ ],
27
+ [
28
+ 9,
29
+ 0
30
+ ],
31
+ [
32
+ 9,
33
+ 7
34
+ ],
35
+ [
36
+ 9,
37
+ 9
38
+ ],
39
+ [
40
+ 10,
41
+ 5
42
+ ]
43
+ ],
44
  "begin_suppress_tokens": [
45
  220,
46
  50257
 
259
  "transcribe": 50359,
260
  "translate": 50358
261
  },
262
+ "transformers_version": "4.33.1"
263
+ }
onnx/decoder_model_merged_int4.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee0f64d9e52f33f641ba7800e156389f63ca55a1717482f2b39eed878bd04cac
3
+ size 257656197
onnx/decoder_model_merged_int8.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:944beba5a09f21b18a9856cdc42f98bba8641c01823ddb7dce77db4bc1925907
3
- size 196598134
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a609e09a9e5a0e6f0e47718ddd1c82afb03e2d5c22619ebdaecb0bf8664ae6f6
3
+ size 315937891
onnx/encoder_model_int4.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d58105d9c47ec6a28d5512ae4b6b3a68159844dcc79e319b643d2c86cf6220b
3
+ size 66103133
onnx/encoder_model_int8.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87ed5a8e6df837185feb613d74de58b068f178d85261dc2e64d1d5340cd95e7e
3
- size 92574888
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:587d41d569087d5b46604189b3210d878233468cb21e5f2a7c49ea10a78408fe
3
+ size 98406132
preprocessor_config.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -9,6 +9,7 @@
9
  "rstrip": false,
10
  "single_word": false
11
  },
 
12
  "eos_token": {
13
  "__type": "AddedToken",
14
  "content": "<|endoftext|>",
@@ -19,12 +20,11 @@
19
  },
20
  "errors": "replace",
21
  "model_max_length": 1024,
22
- "name_or_path": "openai/whisper-small",
23
  "pad_token": null,
24
  "processor_class": "WhisperProcessor",
25
  "return_attention_mask": false,
26
- "special_tokens_map_file": null,
27
  "tokenizer_class": "WhisperTokenizer",
 
28
  "unk_token": {
29
  "__type": "AddedToken",
30
  "content": "<|endoftext|>",
 
9
  "rstrip": false,
10
  "single_word": false
11
  },
12
+ "clean_up_tokenization_spaces": true,
13
  "eos_token": {
14
  "__type": "AddedToken",
15
  "content": "<|endoftext|>",
 
20
  },
21
  "errors": "replace",
22
  "model_max_length": 1024,
 
23
  "pad_token": null,
24
  "processor_class": "WhisperProcessor",
25
  "return_attention_mask": false,
 
26
  "tokenizer_class": "WhisperTokenizer",
27
+ "trust_remote_code": false,
28
  "unk_token": {
29
  "__type": "AddedToken",
30
  "content": "<|endoftext|>",
vocab.json CHANGED
@@ -1,6 +1,5 @@
1
  {
2
  "": 50256,
3
- "<|endoftext|>": 50257,
4
  "!": 0,
5
  "!!": 1432,
6
  "!!!": 4589,
@@ -315,6 +314,7 @@
315
  ";;": 35746,
316
  "<": 27,
317
  "</": 3433,
 
318
  "=": 28,
319
  "=\"": 13114,
320
  "=\"#": 34106,
 
1
  {
2
  "": 50256,
 
3
  "!": 0,
4
  "!!": 1432,
5
  "!!!": 4589,
 
314
  ";;": 35746,
315
  "<": 27,
316
  "</": 3433,
317
+ "<|endoftext|>": 50257,
318
  "=": 28,
319
  "=\"": 13114,
320
  "=\"#": 34106,