Upload optimized ONNX model

#1
by Xenova HF Staff - opened
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ onnx/model.onnx_data filter=lfs diff=lfs merge=lfs -text
37
+ onnx/model_fp16.onnx_data filter=lfs diff=lfs merge=lfs -text
config.json CHANGED
@@ -55,5 +55,15 @@
55
  "use_cla": false,
56
  "use_qk_norm": true,
57
  "use_rotary_pos_emb": true,
58
- "vocab_size": 120818
59
- }
 
 
 
 
 
 
 
 
 
 
 
55
  "use_cla": false,
56
  "use_qk_norm": true,
57
  "use_rotary_pos_emb": true,
58
+ "vocab_size": 120818,
59
+ "transformers.js_config": {
60
+ "use_external_data_format": {
61
+ "model.onnx": 1,
62
+ "model_fp16.onnx": 1
63
+ },
64
+ "kv_cache_dtype": {
65
+ "q4f16": "float16",
66
+ "fp16": "float16"
67
+ }
68
+ }
69
+ }
onnx/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4b76af2c1be9520fd960b773e65346d805b4783f5ce225196a056f6249743cd
3
+ size 23933
onnx/model.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39ee96150cd3721466c67ab251310f01334c0c006ce67489eada47f7756c546f
3
+ size 48001536
onnx/model_fp16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5244d92f8675934a3c68f56ed157f092554bcd349bb46eff612ed95fe78d128
3
+ size 23399
onnx/model_fp16.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3f4fa048f1f4e0d8a8902208c4b8d1d175dd8311a689a251578fd616c3bf9c6
3
+ size 24000768
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,11 +1,12 @@
1
  {
2
  "backend": "tokenizers",
3
- "bos_token": "<|hy_begin▁of▁sentence|>",
4
  "clean_up_tokenization_spaces": true,
5
- "eos_token": "<|hy_place▁holder▁no▁2|>",
6
  "is_local": false,
7
  "model_max_length": 1000000000000000019884624838656,
8
  "model_specific_special_tokens": {},
9
- "pad_token": "<|hy_▁pad▁|>",
10
- "tokenizer_class": "TokenizersBackend"
11
- }
 
 
1
  {
2
  "backend": "tokenizers",
3
+ "bos_token": "<\uff5chy_begin\u2581of\u2581sentence\uff5c>",
4
  "clean_up_tokenization_spaces": true,
5
+ "eos_token": "<\uff5chy_place\u2581holder\u2581no\u25812\uff5c>",
6
  "is_local": false,
7
  "model_max_length": 1000000000000000019884624838656,
8
  "model_specific_special_tokens": {},
9
+ "pad_token": "<\uff5chy_\u2581pad\u2581\uff5c>",
10
+ "tokenizer_class": "TokenizersBackend",
11
+ "chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}<\uff5chy_begin\u2581of\u2581sentence\uff5c>{{ system_message }}<\uff5chy_place\u2581holder\u2581no\u25813\uff5c>{% else %}{% set loop_messages = messages %}<\uff5chy_begin\u2581of\u2581sentence\uff5c>{% endif %}{% for message in loop_messages %}{% if message['role'] == 'user' %}<\uff5chy_User\uff5c>{{ message['content'] }}{% elif message['role'] == 'assistant' %}<\uff5chy_Assistant\uff5c>{{ message['content'] }}<\uff5chy_place\u2581holder\u2581no\u25812\uff5c>{% endif %}{% endfor %}{% if add_generation_prompt %}<\uff5chy_Assistant\uff5c>{% else %}<\uff5chy_place\u2581holder\u2581no\u25818\uff5c>{% endif %}"
12
+ }