Xenova HF Staff commited on
Commit
b1ee71d
·
1 Parent(s): affe9ba

Upload optimized ONNX model (#2)

Browse files

- Upload optimized ONNX model (3b2ae266746a9a6236ee5acd951bdc9c23e4765d)

config.json CHANGED
@@ -32,6 +32,8 @@
32
  "rope_theta": 1000000.0,
33
  "rope_type": "default"
34
  },
 
 
35
  "sliding_attention": {
36
  "rope_theta": 10000.0,
37
  "rope_type": "default"
@@ -40,14 +42,17 @@
40
  "sliding_window": 512,
41
  "tie_word_embeddings": true,
42
  "transformers.js_config": {
43
- "kv_cache_dtype": {
44
- "fp16": "float16",
45
- "q4f16": "float16"
46
  },
47
- "use_external_data_format": true
 
 
 
48
  },
49
  "transformers_version": "5.3.0.dev0",
50
  "use_bidirectional_attention": false,
51
  "use_cache": true,
52
  "vocab_size": 262144
53
- }
 
32
  "rope_theta": 1000000.0,
33
  "rope_type": "default"
34
  },
35
+ "rope_theta": null,
36
+ "rope_type": "default",
37
  "sliding_attention": {
38
  "rope_theta": 10000.0,
39
  "rope_type": "default"
 
42
  "sliding_window": 512,
43
  "tie_word_embeddings": true,
44
  "transformers.js_config": {
45
+ "use_external_data_format": {
46
+ "model.onnx": 1,
47
+ "model_fp16.onnx": 1
48
  },
49
+ "kv_cache_dtype": {
50
+ "q4f16": "float16",
51
+ "fp16": "float16"
52
+ }
53
  },
54
  "transformers_version": "5.3.0.dev0",
55
  "use_bidirectional_attention": false,
56
  "use_cache": true,
57
  "vocab_size": 262144
58
+ }
onnx/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e95268234b58d36540236e116554f34d3ba95c6a4ed2cf562bd3e4a38a13da7
3
+ size 30111
onnx/model.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3beb44efa24531572f67714b3a0015487e429e39d75f56918957e1a260a4ced
3
+ size 69353472
onnx/model_fp16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38e091f02c03eb538fae5e3e9b297f4b055215ca15146a728780871f9dfac8e1
3
+ size 39321
onnx/model_fp16.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8afd6d6df73c564dddfd35a5d1b2a44f94a370c4b6e4379256eb27dce8dc17ec
3
+ size 34676736
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:daab2354f8a74e70d70b4d1f804939b68a8c9624dd06cb7858e52dd8970e9726
3
- size 33384567
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b472c7c2d1a2ac607daa53c794575e59504c06b39579a0f2399bb95cea6a4949
3
+ size 20323105
tokenizer_config.json CHANGED
@@ -20,5 +20,6 @@
20
  "spaces_between_special_tokens": false,
21
  "tokenizer_class": "GemmaTokenizer",
22
  "unk_token": "<unk>",
23
- "use_default_system_prompt": false
24
- }
 
 
20
  "spaces_between_special_tokens": false,
21
  "tokenizer_class": "GemmaTokenizer",
22
  "unk_token": "<unk>",
23
+ "use_default_system_prompt": false,
24
+ "chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[0]['content'] is string -%}\n {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n {%- else -%}\n {%- set first_user_prefix = messages[0]['content'][0]['text'] + '\n\n' -%}\n {%- endif -%}\n {%- set loop_messages = messages[1:] -%}\n{%- else -%}\n {%- set first_user_prefix = \"\" -%}\n {%- set loop_messages = messages -%}\n{%- endif -%}\n{%- for message in loop_messages -%}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) -%}\n {{ raise_exception(\"Conversation roles must alternate user/assistant/user/assistant/...\") }}\n {%- endif -%}\n {%- if (message['role'] == 'assistant') -%}\n {%- set role = \"model\" -%}\n {%- else -%}\n {%- set role = message['role'] -%}\n {%- endif -%}\n {{ '<start_of_turn>' + role + '\n' + (first_user_prefix if loop.first else \"\") }}\n {%- if message['content'] is string -%}\n {{ message['content'] | trim }}\n {%- elif message['content'] is iterable -%}\n {%- for item in message['content'] -%}\n {%- if item['type'] == 'image' -%}\n {{ '<start_of_image>' }}\n {%- elif item['type'] == 'text' -%}\n {{ item['text'] | trim }}\n {%- endif -%}\n {%- endfor -%}\n {%- else -%}\n {{ raise_exception(\"Invalid content type\") }}\n {%- endif -%}\n {{ '<end_of_turn>\n' }}\n{%- endfor -%}\n{%- if add_generation_prompt -%}\n {{'<start_of_turn>model\n'}}\n{%- endif -%}\n"
25
+ }