Xenova HF Staff commited on
Commit
5410886
·
verified ·
1 Parent(s): 76eb3a9

Upload optimized ONNX model

Browse files
.gitattributes CHANGED
@@ -34,3 +34,9 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ onnx/decoder_model_merged.onnx_data filter=lfs diff=lfs merge=lfs -text
38
+ onnx/decoder_model_merged_fp16.onnx_data filter=lfs diff=lfs merge=lfs -text
39
+ onnx/embed_tokens.onnx_data filter=lfs diff=lfs merge=lfs -text
40
+ onnx/embed_tokens_fp16.onnx_data filter=lfs diff=lfs merge=lfs -text
41
+ onnx/vision_encoder.onnx_data filter=lfs diff=lfs merge=lfs -text
42
+ onnx/vision_encoder_fp16.onnx_data filter=lfs diff=lfs merge=lfs -text
config.json CHANGED
@@ -44,6 +44,8 @@
44
  "rope_theta": 1000000.0,
45
  "rope_type": "linear"
46
  },
 
 
47
  "sliding_attention": {
48
  "rope_theta": 10000.0,
49
  "rope_type": "default"
@@ -70,5 +72,19 @@
70
  "num_hidden_layers": 2,
71
  "patch_size": 14,
72
  "vision_use_head": false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
73
  }
74
- }
 
44
  "rope_theta": 1000000.0,
45
  "rope_type": "linear"
46
  },
47
+ "rope_theta": null,
48
+ "rope_type": "default",
49
  "sliding_attention": {
50
  "rope_theta": 10000.0,
51
  "rope_type": "default"
 
72
  "num_hidden_layers": 2,
73
  "patch_size": 14,
74
  "vision_use_head": false
75
+ },
76
+ "transformers.js_config": {
77
+ "use_external_data_format": {
78
+ "vision_encoder.onnx": 1,
79
+ "decoder_model_merged.onnx": 1,
80
+ "embed_tokens.onnx": 1,
81
+ "vision_encoder_fp16.onnx": 1,
82
+ "decoder_model_merged_fp16.onnx": 1,
83
+ "embed_tokens_fp16.onnx": 1
84
+ },
85
+ "kv_cache_dtype": {
86
+ "q4f16": "float16",
87
+ "fp16": "float16"
88
+ }
89
  }
90
+ }
onnx/decoder_model_merged.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8773406e8e47a9c42ce80381950f9312ebd8250a964ac94e955c240de6b402b
3
+ size 29675
onnx/decoder_model_merged.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1c61df01ffd7dfb4ce4133fa3274339b3315ec4073bf729a6b709ceca7e03b7
3
+ size 75661312
onnx/decoder_model_merged_fp16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5631035a783afa72b237252bca9f20728261cf26a0e6478f49a79e6d88c9834
3
+ size 39173
onnx/decoder_model_merged_fp16.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fc6bf2a8135e41e673abdbae3aae54c90e08326d350fb2be43d208d88f32f7a
3
+ size 37830656
onnx/embed_tokens.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8bf0b4ab46771d3a553b9033087ce7ed4fd44d698bc261b68049e4d264578d40
3
+ size 430
onnx/embed_tokens.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acb8c75db211c85d7f46e468546ddb83024980d986de642cead93eda75cb595b
3
+ size 67125248
onnx/embed_tokens_fp16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:205a64e0dd69dbe3a76addc817a839f2d24ffc97d58c78cf12c6cd09eee3a7d5
3
+ size 632
onnx/embed_tokens_fp16.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8351ef124676206587b890cdb3c3ec9a121c094e595d3d1c39f9056d070f8397
3
+ size 33562624
onnx/vision_encoder.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20d2a8cc3ee7fc4ef81b4e2eab3df6240afff4e2b5804fe20f491d56c59641b1
3
+ size 25003
onnx/vision_encoder.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f1866b0b229ef93c4c45c175662f5c8e3971cdfc304f547d4bce8fabc6e1efc
3
+ size 331776
onnx/vision_encoder_fp16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8e310de97411105f2b34c7678143238c0f46feecd10f8b47cc2bb30187190b8
3
+ size 24052
onnx/vision_encoder_fp16.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d022b963fd92dc9979fc510ff7aa26b7bedaa80bb2c9f4100f77b5ce1ca56f67
3
+ size 165376
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:daab2354f8a74e70d70b4d1f804939b68a8c9624dd06cb7858e52dd8970e9726
3
- size 33384567
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b472c7c2d1a2ac607daa53c794575e59504c06b39579a0f2399bb95cea6a4949
3
+ size 20323105
tokenizer_config.json CHANGED
@@ -20,5 +20,6 @@
20
  "spaces_between_special_tokens": false,
21
  "tokenizer_class": "GemmaTokenizer",
22
  "unk_token": "<unk>",
23
- "use_default_system_prompt": false
24
- }
 
 
20
  "spaces_between_special_tokens": false,
21
  "tokenizer_class": "GemmaTokenizer",
22
  "unk_token": "<unk>",
23
+ "use_default_system_prompt": false,
24
+ "chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[0]['content'] is string -%}\n {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n {%- else -%}\n {%- set first_user_prefix = messages[0]['content'][0]['text'] + '\n\n' -%}\n {%- endif -%}\n {%- set loop_messages = messages[1:] -%}\n{%- else -%}\n {%- set first_user_prefix = \"\" -%}\n {%- set loop_messages = messages -%}\n{%- endif -%}\n{%- for message in loop_messages -%}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) -%}\n {{ raise_exception(\"Conversation roles must alternate user/assistant/user/assistant/...\") }}\n {%- endif -%}\n {%- if (message['role'] == 'assistant') -%}\n {%- set role = \"model\" -%}\n {%- else -%}\n {%- set role = message['role'] -%}\n {%- endif -%}\n {{ '<start_of_turn>' + role + '\n' + (first_user_prefix if loop.first else \"\") }}\n {%- if message['content'] is string -%}\n {{ message['content'] | trim }}\n {%- elif message['content'] is iterable -%}\n {%- for item in message['content'] -%}\n {%- if item['type'] == 'image' -%}\n {{ '<start_of_image>' }}\n {%- elif item['type'] == 'text' -%}\n {{ item['text'] | trim }}\n {%- endif -%}\n {%- endfor -%}\n {%- else -%}\n {{ raise_exception(\"Invalid content type\") }}\n {%- endif -%}\n {{ '<end_of_turn>\n' }}\n{%- endfor -%}\n{%- if add_generation_prompt -%}\n {{'<start_of_turn>model\n'}}\n{%- endif -%}\n"
25
+ }