Upload folder using huggingface_hub
Browse files
.gitattributes
CHANGED
|
@@ -47,3 +47,4 @@ optimized-whisper-small-both-quantized/encoder_model_quantized.onnx.data filter=
|
|
| 47 |
optimized-whisper-turbo-quantized/decoder_model_quantized.onnx.data filter=lfs diff=lfs merge=lfs -text
|
| 48 |
optimized-whisper-turbo-quantized/decoder_with_past_model_quantized.onnx.data filter=lfs diff=lfs merge=lfs -text
|
| 49 |
optimized-whisper-turbo-quantized/encoder_model_quantized.onnx.data filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 47 |
optimized-whisper-turbo-quantized/decoder_model_quantized.onnx.data filter=lfs diff=lfs merge=lfs -text
|
| 48 |
optimized-whisper-turbo-quantized/decoder_with_past_model_quantized.onnx.data filter=lfs diff=lfs merge=lfs -text
|
| 49 |
optimized-whisper-turbo-quantized/encoder_model_quantized.onnx.data filter=lfs diff=lfs merge=lfs -text
|
| 50 |
+
whisper-turbo/encoder_model.onnx_data filter=lfs diff=lfs merge=lfs -text
|
whisper-turbo/config.json
ADDED
|
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"activation_dropout": 0.0,
|
| 3 |
+
"activation_function": "gelu",
|
| 4 |
+
"apply_spec_augment": false,
|
| 5 |
+
"architectures": [
|
| 6 |
+
"WhisperForConditionalGeneration"
|
| 7 |
+
],
|
| 8 |
+
"attention_dropout": 0.0,
|
| 9 |
+
"begin_suppress_tokens": [
|
| 10 |
+
220,
|
| 11 |
+
50256
|
| 12 |
+
],
|
| 13 |
+
"bos_token_id": 50257,
|
| 14 |
+
"classifier_proj_size": 256,
|
| 15 |
+
"d_model": 1280,
|
| 16 |
+
"decoder_attention_heads": 20,
|
| 17 |
+
"decoder_ffn_dim": 5120,
|
| 18 |
+
"decoder_layerdrop": 0.0,
|
| 19 |
+
"decoder_layers": 4,
|
| 20 |
+
"decoder_start_token_id": 50258,
|
| 21 |
+
"dropout": 0.0,
|
| 22 |
+
"encoder_attention_heads": 20,
|
| 23 |
+
"encoder_ffn_dim": 5120,
|
| 24 |
+
"encoder_layerdrop": 0.0,
|
| 25 |
+
"encoder_layers": 32,
|
| 26 |
+
"eos_token_id": 50257,
|
| 27 |
+
"init_std": 0.02,
|
| 28 |
+
"is_encoder_decoder": true,
|
| 29 |
+
"mask_feature_length": 10,
|
| 30 |
+
"mask_feature_min_masks": 0,
|
| 31 |
+
"mask_feature_prob": 0.0,
|
| 32 |
+
"mask_time_length": 10,
|
| 33 |
+
"mask_time_min_masks": 2,
|
| 34 |
+
"mask_time_prob": 0.05,
|
| 35 |
+
"max_source_positions": 1500,
|
| 36 |
+
"max_target_positions": 448,
|
| 37 |
+
"median_filter_width": 7,
|
| 38 |
+
"model_type": "whisper",
|
| 39 |
+
"num_hidden_layers": 32,
|
| 40 |
+
"num_mel_bins": 128,
|
| 41 |
+
"pad_token_id": 50257,
|
| 42 |
+
"scale_embedding": false,
|
| 43 |
+
"torch_dtype": "float32",
|
| 44 |
+
"transformers_version": "4.51.3",
|
| 45 |
+
"use_cache": true,
|
| 46 |
+
"use_weighted_layer_sum": false,
|
| 47 |
+
"vocab_size": 51866
|
| 48 |
+
}
|
whisper-turbo/decoder_model.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:295ba141f21fc2e478fd2f251ea31a62f0763105f10eb8cc64004774fe82495d
|
| 3 |
+
size 953330447
|
whisper-turbo/decoder_with_past_model.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e73e657e21fd41dfc429dc17d561f82e728087871d5139f7e148911e7d5e8780
|
| 3 |
+
size 900869067
|
whisper-turbo/encoder_model.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:47fd3032a1f3c5a77304ece14a512c84e0244c032ca82ca8c1a9f993f5870a2d
|
| 3 |
+
size 496106
|
whisper-turbo/encoder_model.onnx_data
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:044521792ec063eb3995608b213e34055aa6bc2f251e4d8ee7a787bf7fdfc07b
|
| 3 |
+
size 2547875840
|
whisper-turbo/generation_config.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"alignment_heads": [[2, 4], [2, 11], [3, 3], [3, 6], [3, 11], [3, 14]], "begin_suppress_tokens": [220, 50257], "bos_token_id": 50257, "decoder_start_token_id": 50258, "eos_token_id": 50257, "is_multilingual": true, "lang_to_id": {"<|af|>": 50327, "<|am|>": 50334, "<|ar|>": 50272, "<|as|>": 50350, "<|az|>": 50304, "<|ba|>": 50355, "<|be|>": 50330, "<|bg|>": 50292, "<|bn|>": 50302, "<|bo|>": 50347, "<|br|>": 50309, "<|bs|>": 50315, "<|ca|>": 50270, "<|cs|>": 50283, "<|cy|>": 50297, "<|da|>": 50285, "<|de|>": 50261, "<|el|>": 50281, "<|en|>": 50259, "<|es|>": 50262, "<|et|>": 50307, "<|eu|>": 50310, "<|fa|>": 50300, "<|fi|>": 50277, "<|fo|>": 50338, "<|fr|>": 50265, "<|gl|>": 50319, "<|gu|>": 50333, "<|haw|>": 50352, "<|ha|>": 50354, "<|he|>": 50279, "<|hi|>": 50276, "<|hr|>": 50291, "<|ht|>": 50339, "<|hu|>": 50286, "<|hy|>": 50312, "<|id|>": 50275, "<|is|>": 50311, "<|it|>": 50274, "<|ja|>": 50266, "<|jw|>": 50356, "<|ka|>": 50329, "<|kk|>": 50316, "<|km|>": 50323, "<|kn|>": 50306, "<|ko|>": 50264, "<|la|>": 50294, "<|lb|>": 50345, "<|ln|>": 50353, "<|lo|>": 50336, "<|lt|>": 50293, "<|lv|>": 50301, "<|mg|>": 50349, "<|mi|>": 50295, "<|mk|>": 50308, "<|ml|>": 50296, "<|mn|>": 50314, "<|mr|>": 50320, "<|ms|>": 50282, "<|mt|>": 50343, "<|my|>": 50346, "<|ne|>": 50313, "<|nl|>": 50271, "<|nn|>": 50342, "<|no|>": 50288, "<|oc|>": 50328, "<|pa|>": 50321, "<|pl|>": 50269, "<|ps|>": 50340, "<|pt|>": 50267, "<|ro|>": 50284, "<|ru|>": 50263, "<|sa|>": 50344, "<|sd|>": 50332, "<|si|>": 50322, "<|sk|>": 50298, "<|sl|>": 50305, "<|sn|>": 50324, "<|so|>": 50326, "<|sq|>": 50317, "<|sr|>": 50303, "<|su|>": 50357, "<|sv|>": 50273, "<|sw|>": 50318, "<|ta|>": 50287, "<|te|>": 50299, "<|tg|>": 50331, "<|th|>": 50289, "<|tk|>": 50341, "<|tl|>": 50348, "<|tr|>": 50268, "<|tt|>": 50351, "<|uk|>": 50280, "<|ur|>": 50290, "<|uz|>": 50337, "<|vi|>": 50278, "<|yi|>": 50335, "<|yo|>": 50325, "<|yue|>": 50358, "<|zh|>": 50260}, "max_initial_timestamp_index": 50, "max_length": 448, "no_timestamps_token_id": 50364, "pad_token_id": 50257, "prev_sot_token_id": 50362, "return_timestamps": false, "suppress_tokens": [1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 359, 503, 522, 542, 873, 893, 902, 918, 922, 931, 1350, 1853, 1982, 2460, 2627, 3246, 3253, 3268, 3536, 3846, 3961, 4183, 4667, 6585, 6647, 7273, 9061, 9383, 10428, 10929, 11938, 12033, 12331, 12562, 13793, 14157, 14635, 15265, 15618, 16553, 16604, 18362, 18956, 20075, 21675, 22520, 26130, 26161, 26435, 28279, 29464, 31650, 32302, 32470, 36865, 42863, 47425, 49870, 50254, 50258, 50359, 50360, 50361, 50362, 50363], "task_to_id": {"transcribe": 50360, "translate": 50359}, "transformers_version": "4.51.3"}
|