Xenova HF Staff commited on
Commit
5ee7e58
·
1 Parent(s): 698d234

Upload optimized ONNX model (#1)

Browse files

- Upload optimized ONNX model (b265fcbf76870a62a8ccc244db94f14c21c7ac4e)

.gitattributes CHANGED
@@ -33,3 +33,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ onnx/model.onnx_data filter=lfs diff=lfs merge=lfs -text
37
+ onnx/model_fp16.onnx_data filter=lfs diff=lfs merge=lfs -text
38
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
chat_template.jinja ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {{bos_token}}
2
+ {%- if tools %}
3
+ {{- '<|im_start|>system\n' }}
4
+ {%- if messages[0].role == 'system' %}
5
+ {{- messages[0].content + '\n\n' }}
6
+ {%- endif %}
7
+ {{- "You are a function calling AI model. You are provided with function signature within <tools> </tools> XML tags. You may call one or more functions to assist with the user query. Don't make assumptions about what values to plug into functions.\n<tools>\n" }}
8
+ {%- for tool in tools %}[{{- tool | tojson }}]{%- endfor %}
9
+ {{- "\n</tools>\nFor each function call, return a json object with function name and arguments within <tool_call> </tool_call> tags with the following schema:\n<tool_call>\n{'arguments': <args-dict>, 'name': <function-name>}\n</tool_call>\n" }}
10
+ {%- else %}
11
+ {%- if messages[0].role == 'system' %}
12
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
13
+ {%- endif %}
14
+ {%- endif %}{% for message in messages %}{%- if message.role != 'system' %}{{'<|im_start|>' + message['role'] + '
15
+ ' + message['content'] + '<|im_end|>' + '
16
+ '}}{%- endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
17
+ ' }}{% endif %}
config.json ADDED
@@ -0,0 +1,84 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "FalconH1ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "attention_in_multiplier": 1.0,
8
+ "attention_out_multiplier": 0.0375,
9
+ "attn_layer_indices": null,
10
+ "bos_token_id": 1,
11
+ "dtype": "bfloat16",
12
+ "embedding_multiplier": 5.656854249492381,
13
+ "eos_token_id": 11,
14
+ "head_dim": 32,
15
+ "hidden_act": "silu",
16
+ "hidden_size": 8,
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 64,
19
+ "key_multiplier": 0.011048543456039804,
20
+ "lm_head_multiplier": 0.0078125,
21
+ "mamba_chunk_size": 128,
22
+ "mamba_conv_bias": true,
23
+ "mamba_d_conv": 4,
24
+ "mamba_d_head": 32,
25
+ "mamba_d_ssm": 256,
26
+ "mamba_d_state": 32,
27
+ "mamba_expand": 32,
28
+ "mamba_n_groups": 2,
29
+ "mamba_n_heads": 8,
30
+ "mamba_norm_before_gate": false,
31
+ "mamba_proj_bias": false,
32
+ "mamba_rms_norm": true,
33
+ "mamba_use_mlp": true,
34
+ "max_position_embeddings": 262144,
35
+ "mlp_bias": false,
36
+ "mlp_expansion_factor": 8,
37
+ "mlp_multipliers": [
38
+ 0.1767766952966369,
39
+ 0.011160714285714284
40
+ ],
41
+ "model_type": "falcon_h1",
42
+ "num_attention_heads": 8,
43
+ "num_hidden_layers": 2,
44
+ "num_key_value_heads": 4,
45
+ "num_logits_to_keep": 1,
46
+ "pad_token_id": 0,
47
+ "projectors_bias": false,
48
+ "rms_norm_eps": 1e-05,
49
+ "rope_parameters": {
50
+ "rope_theta": 100000000000.0,
51
+ "rope_type": "default"
52
+ },
53
+ "ssm_in_multiplier": 0.25,
54
+ "ssm_multipliers": [
55
+ 0.3535533905932738,
56
+ 0.25,
57
+ 0.1767766952966369,
58
+ 0.5,
59
+ 0.3535533905932738
60
+ ],
61
+ "ssm_out_multiplier": 0.08838834764831845,
62
+ "tie_word_embeddings": true,
63
+ "time_step_limit": [
64
+ 0.0,
65
+ {
66
+ "__float__": "Infinity"
67
+ }
68
+ ],
69
+ "time_step_max": 0.1,
70
+ "time_step_min": 0.001,
71
+ "transformers_version": "5.3.0.dev0",
72
+ "use_cache": true,
73
+ "vocab_size": 261120,
74
+ "transformers.js_config": {
75
+ "use_external_data_format": {
76
+ "model.onnx": 1,
77
+ "model_fp16.onnx": 1
78
+ },
79
+ "kv_cache_dtype": {
80
+ "q4f16": "float16",
81
+ "fp16": "float16"
82
+ }
83
+ }
84
+ }
generation_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": [
5
+ 11,
6
+ 228
7
+ ],
8
+ "pad_token_id": 0,
9
+ "transformers_version": "5.3.0.dev0"
10
+ }
onnx/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5fae147e7e8b6cfaa79b8ee649abba26374f3b0b2252b6de3b1ff4f79df3f6e
3
+ size 129642
onnx/model.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fb3595543e20b9a609b7ed354f8879b30df2ec6d38a04159db3a52521ad99af
3
+ size 42128896
onnx/model_fp16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b754b1841bcd710946fbe8babeffebffc1519a1d54f7fd4d901f90ee5f031f20
3
+ size 137383
onnx/model_fp16.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eabac72aecfebcd79f755fc00257e4ce954e537a281811dc6371f6b2f189b286
3
+ size 21061888
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c5da8e95031f5e8e6a30980cedfc34b2e3c4714bb5b0bfe52ea9d9d54379842
3
+ size 22442959
tokenizer_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "bos_token": "<|begin_of_text|>",
4
+ "clean_up_tokenization_spaces": true,
5
+ "eos_token": "<|end_of_text|>",
6
+ "is_local": false,
7
+ "model_input_names": [
8
+ "input_ids",
9
+ "attention_mask"
10
+ ],
11
+ "model_max_length": 1000000000000000019884624838656,
12
+ "pad_token": "<|pad|>",
13
+ "padding_side": "left",
14
+ "tokenizer_class": "TokenizersBackend",
15
+ "chat_template": "{{bos_token}}\n{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"You are a function calling AI model. You are provided with function signature within <tools> </tools> XML tags. You may call one or more functions to assist with the user query. Don't make assumptions about what values to plug into functions.\\n<tools>\\n\" }}\n {%- for tool in tools %}[{{- tool | tojson }}]{%- endfor %}\n {{- \"\\n</tools>\\nFor each function call, return a json object with function name and arguments within <tool_call> </tool_call> tags with the following schema:\\n<tool_call>\\n{'arguments': <args-dict>, 'name': <function-name>}\\n</tool_call>\\n\" }}\n{%- else %}\n {%- if messages[0].role == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0].content + '<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}{% for message in messages %}{%- if message.role != 'system' %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{%- endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}"
16
+ }