Upload folder using huggingface_hub
Browse files- .DS_Store +0 -0
- .gitattributes +6 -0
- config.json +9 -13
- generation_config.json +1 -0
- onnx/model.onnx +2 -2
- onnx/model.onnx_data +2 -2
- onnx/model_bnb4.onnx +2 -2
- onnx/model_fp16.onnx +2 -2
- onnx/model_fp16.onnx_data +2 -2
- onnx/model_int8.onnx +2 -2
- onnx/model_q4.onnx +2 -2
- onnx/model_q4f16.onnx +2 -2
- onnx/model_quantized.onnx +2 -2
- onnx/model_uint8.onnx +2 -2
- special_tokens_map.json +1 -1
- tokenizer_config.json +15 -3
.DS_Store
CHANGED
|
Binary files a/.DS_Store and b/.DS_Store differ
|
|
|
.gitattributes
CHANGED
|
@@ -34,5 +34,11 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
onnx/model.onnx_data filter=lfs diff=lfs merge=lfs -text
|
|
|
|
| 37 |
onnx/model_fp16.onnx_data filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 38 |
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
|
|
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
onnx/model.onnx_data filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
onnx/model_bnb4.onnx_data filter=lfs diff=lfs merge=lfs -text
|
| 38 |
onnx/model_fp16.onnx_data filter=lfs diff=lfs merge=lfs -text
|
| 39 |
+
onnx/model_int8.onnx_data filter=lfs diff=lfs merge=lfs -text
|
| 40 |
+
onnx/model_q4.onnx_data filter=lfs diff=lfs merge=lfs -text
|
| 41 |
+
onnx/model_q4f16.onnx_data filter=lfs diff=lfs merge=lfs -text
|
| 42 |
+
onnx/model_quantized.onnx_data filter=lfs diff=lfs merge=lfs -text
|
| 43 |
+
onnx/model_uint8.onnx_data filter=lfs diff=lfs merge=lfs -text
|
| 44 |
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
config.json
CHANGED
|
@@ -1,42 +1,38 @@
|
|
| 1 |
{
|
| 2 |
"_attn_implementation_autoset": true,
|
| 3 |
-
"_name_or_path": "
|
| 4 |
"architectures": [
|
| 5 |
"LlamaForCausalLM"
|
| 6 |
],
|
| 7 |
"attention_bias": false,
|
| 8 |
"attention_dropout": 0.0,
|
| 9 |
"bos_token_id": 128000,
|
| 10 |
-
"eos_token_id":
|
| 11 |
-
|
| 12 |
-
128008,
|
| 13 |
-
128009
|
| 14 |
-
],
|
| 15 |
-
"head_dim": 64,
|
| 16 |
"hidden_act": "silu",
|
| 17 |
-
"hidden_size":
|
| 18 |
"initializer_range": 0.02,
|
| 19 |
-
"intermediate_size":
|
| 20 |
"max_position_embeddings": 131072,
|
| 21 |
"mlp_bias": false,
|
| 22 |
"model_type": "llama",
|
| 23 |
"num_attention_heads": 32,
|
| 24 |
-
"num_hidden_layers":
|
| 25 |
"num_key_value_heads": 8,
|
| 26 |
"pad_token_id": 128004,
|
| 27 |
"pretraining_tp": 1,
|
| 28 |
"rms_norm_eps": 1e-05,
|
| 29 |
"rope_scaling": {
|
| 30 |
-
"factor":
|
| 31 |
"high_freq_factor": 4.0,
|
| 32 |
"low_freq_factor": 1.0,
|
| 33 |
"original_max_position_embeddings": 8192,
|
| 34 |
"rope_type": "llama3"
|
| 35 |
},
|
| 36 |
"rope_theta": 500000.0,
|
| 37 |
-
"tie_word_embeddings":
|
| 38 |
"transformers_version": "4.46.3",
|
| 39 |
-
"
|
| 40 |
"use_cache": false,
|
| 41 |
"vocab_size": 128256
|
| 42 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"_attn_implementation_autoset": true,
|
| 3 |
+
"_name_or_path": "NousResearch/DeepHermes-3-Llama-3-8B-Preview",
|
| 4 |
"architectures": [
|
| 5 |
"LlamaForCausalLM"
|
| 6 |
],
|
| 7 |
"attention_bias": false,
|
| 8 |
"attention_dropout": 0.0,
|
| 9 |
"bos_token_id": 128000,
|
| 10 |
+
"eos_token_id": 128009,
|
| 11 |
+
"head_dim": 128,
|
|
|
|
|
|
|
|
|
|
|
|
|
| 12 |
"hidden_act": "silu",
|
| 13 |
+
"hidden_size": 4096,
|
| 14 |
"initializer_range": 0.02,
|
| 15 |
+
"intermediate_size": 14336,
|
| 16 |
"max_position_embeddings": 131072,
|
| 17 |
"mlp_bias": false,
|
| 18 |
"model_type": "llama",
|
| 19 |
"num_attention_heads": 32,
|
| 20 |
+
"num_hidden_layers": 32,
|
| 21 |
"num_key_value_heads": 8,
|
| 22 |
"pad_token_id": 128004,
|
| 23 |
"pretraining_tp": 1,
|
| 24 |
"rms_norm_eps": 1e-05,
|
| 25 |
"rope_scaling": {
|
| 26 |
+
"factor": 8.0,
|
| 27 |
"high_freq_factor": 4.0,
|
| 28 |
"low_freq_factor": 1.0,
|
| 29 |
"original_max_position_embeddings": 8192,
|
| 30 |
"rope_type": "llama3"
|
| 31 |
},
|
| 32 |
"rope_theta": 500000.0,
|
| 33 |
+
"tie_word_embeddings": false,
|
| 34 |
"transformers_version": "4.46.3",
|
| 35 |
+
"unsloth_fixed": true,
|
| 36 |
"use_cache": false,
|
| 37 |
"vocab_size": 128256
|
| 38 |
}
|
generation_config.json
CHANGED
|
@@ -1,4 +1,5 @@
|
|
| 1 |
{
|
|
|
|
| 2 |
"bos_token_id": 128000,
|
| 3 |
"do_sample": true,
|
| 4 |
"eos_token_id": [
|
|
|
|
| 1 |
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
"bos_token_id": 128000,
|
| 4 |
"do_sample": true,
|
| 5 |
"eos_token_id": [
|
onnx/model.onnx
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bdd315c3543dc1c2c7fea02fee19dcc5988b3bf455365892132e0f1ed01f5b3f
|
| 3 |
+
size 858913
|
onnx/model.onnx_data
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2fe497e51b001cf53241abf12fd69a9561cb75e6290a7ef41b94d5519499e832
|
| 3 |
+
size 32121044992
|
onnx/model_bnb4.onnx
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c078b3be1198c2ef021344f5ce42438cf55beaa2315d635b20c457a4b90c6684
|
| 3 |
+
size 911812
|
onnx/model_fp16.onnx
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:76fcb21a89369fc32d3795df146930abb06c163443824f4f1ec2d90bb69b03d7
|
| 3 |
+
size 860108
|
onnx/model_fp16.onnx_data
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:02911f28a0354c44876f3329ecf587d5ca16e65698bb2869937d8544d652aeaa
|
| 3 |
+
size 16060522496
|
onnx/model_int8.onnx
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5b1bb76d6a54817b3f2520fcd98dc4e95dcbc9bfc0ce7c57dca96c16e4c2b2dd
|
| 3 |
+
size 1105763
|
onnx/model_q4.onnx
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3827b6b586a90af38f1f11a34192dbc80f08b3e8e30deeb68253b7dd1c4153fb
|
| 3 |
+
size 908949
|
onnx/model_q4f16.onnx
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6c7183ef2e2436a57314ca1603bbd407219c03ac137b8eef75ff92c4163f9631
|
| 3 |
+
size 910416
|
onnx/model_quantized.onnx
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:92209e407f1d3bf3bebc152743293fb7a3a4ec00fe2e514bbb378cc94796dac9
|
| 3 |
+
size 1107250
|
onnx/model_uint8.onnx
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0c9ed7e84bdb801d0e4a6606067eac6ae14dd76948b408ca9f49bc7475599665
|
| 3 |
+
size 1106175
|
special_tokens_map.json
CHANGED
|
@@ -14,7 +14,7 @@
|
|
| 14 |
"single_word": false
|
| 15 |
},
|
| 16 |
"pad_token": {
|
| 17 |
-
"content": "<|
|
| 18 |
"lstrip": false,
|
| 19 |
"normalized": false,
|
| 20 |
"rstrip": false,
|
|
|
|
| 14 |
"single_word": false
|
| 15 |
},
|
| 16 |
"pad_token": {
|
| 17 |
+
"content": "<|end_of_text|>",
|
| 18 |
"lstrip": false,
|
| 19 |
"normalized": false,
|
| 20 |
"rstrip": false,
|
tokenizer_config.json
CHANGED
|
@@ -1,4 +1,5 @@
|
|
| 1 |
{
|
|
|
|
| 2 |
"added_tokens_decoder": {
|
| 3 |
"128000": {
|
| 4 |
"content": "<|begin_of_text|>",
|
|
@@ -2050,15 +2051,26 @@
|
|
| 2050 |
}
|
| 2051 |
},
|
| 2052 |
"bos_token": "<|begin_of_text|>",
|
| 2053 |
-
"chat_template":
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2054 |
"clean_up_tokenization_spaces": true,
|
| 2055 |
"eos_token": "<|eot_id|>",
|
|
|
|
| 2056 |
"model_input_names": [
|
| 2057 |
"input_ids",
|
| 2058 |
"attention_mask"
|
| 2059 |
],
|
| 2060 |
"model_max_length": 131072,
|
| 2061 |
-
"pad_token": "<|
|
| 2062 |
"padding_side": "left",
|
| 2063 |
-
"tokenizer_class": "PreTrainedTokenizerFast"
|
|
|
|
| 2064 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"add_bos_token": true,
|
| 3 |
"added_tokens_decoder": {
|
| 4 |
"128000": {
|
| 5 |
"content": "<|begin_of_text|>",
|
|
|
|
| 2051 |
}
|
| 2052 |
},
|
| 2053 |
"bos_token": "<|begin_of_text|>",
|
| 2054 |
+
"chat_template": [
|
| 2055 |
+
{
|
| 2056 |
+
"name": "default",
|
| 2057 |
+
"template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}"
|
| 2058 |
+
},
|
| 2059 |
+
{
|
| 2060 |
+
"name": "tool_use",
|
| 2061 |
+
"template": "{%- macro json_to_python_type(json_spec) %}\n{%- set basic_type_map = {\n \"string\": \"str\",\n \"number\": \"float\",\n \"integer\": \"int\",\n \"boolean\": \"bool\"\n} %}\n\n{%- if basic_type_map[json_spec.type] is defined %}\n {{- basic_type_map[json_spec.type] }}\n{%- elif json_spec.type == \"array\" %}\n {{- \"list[\" + json_to_python_type(json_spec|items) + \"]\"}}\n{%- elif json_spec.type == \"object\" %}\n {%- if json_spec.additionalProperties is defined %}\n {{- \"dict[str, \" + json_to_python_type(json_spec.additionalProperties) + ']'}}\n {%- else %}\n {{- \"dict\" }}\n {%- endif %}\n{%- elif json_spec.type is iterable %}\n {{- \"Union[\" }}\n {%- for t in json_spec.type %}\n {{- json_to_python_type({\"type\": t}) }}\n {%- if not loop.last %}\n {{- \",\" }} \n {%- endif %}\n {%- endfor %}\n {{- \"]\" }}\n{%- else %}\n {{- \"Any\" }}\n{%- endif %}\n{%- endmacro %}\n\n\n{{- bos_token }}\n{{- '<|im_start|>system\n' }}\n{{- \"You are a function calling AI model. You are provided with function signatures within <tools></tools> XML tags. You may call one or more functions to assist with the user query. Don't make assumptions about what values to plug into functions. Here are the available tools: <tools> \" }}\n{%- for tool in tools %}\n {%- if tool.function is defined %}\n {%- set tool = tool.function %}\n {%- endif %}\n {{- '{\"type\": \"function\", \"function\": ' }}\n {{- '{\"name\": \"' + tool.name + '\", ' }}\n {{- '\"description\": \"' + tool.name + '(' }}\n {%- for param_name, param_fields in tool.parameters.properties|items %}\n {{- param_name + \": \" + json_to_python_type(param_fields) }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- if tool.return is defined %}\n {{- \" -> \" + json_to_python_type(tool.return) }}\n {%- endif %}\n {{- \" - \" + tool.description + \"\n\n\" }}\n {%- for param_name, param_fields in tool.parameters.properties|items %}\n {%- if loop.first %}\n {{- \" Args:\n\" }}\n {%- endif %}\n {{- \" \" + param_name + \"(\" + json_to_python_type(param_fields) + \"): \" + param_fields.description|trim }}\n {%- endfor %}\n {%- if tool.return is defined and tool.return.description is defined %}\n {{- \"\n Returns:\n \" + tool.return.description }}\n {%- endif %}\n {{- '\"' }}\n {{- ', \"parameters\": ' }}\n {%- if tool.parameters.properties | length == 0 %}\n {{- \"{}\" }}\n {%- else %}\n {{- tool.parameters|tojson }}\n {%- endif %}\n {{- \"}\" }}\n {%- if not loop.last %}\n {{- \"\n\" }}\n {%- endif %}\n{%- endfor %}\n{{- \" </tools>\" }}\n{{- 'Use the following pydantic model json schema for each tool call you will make: {\"properties\": {\"name\": {\"title\": \"Name\", \"type\": \"string\"}, \"arguments\": {\"title\": \"Arguments\", \"type\": \"object\"}}, \"required\": [\"name\", \"arguments\"], \"title\": \"FunctionCall\", \"type\": \"object\"}}\n' }}\n{{- \"For each function call return a json object with function name and arguments within <tool_call></tool_call> XML tags as follows:\n\" }}\n{{- \"<tool_call>\n\" }}\n{{- '{\"name\": <function-name>, \"arguments\": <args-dict>}\n' }}\n{{- '</tool_call><|im_end|>\n' }}\n{%- for message in messages %}\n {%- if message.role == \"user\" or message.role == \"system\" or (message.role == \"assistant\" and message.tool_calls is not defined) %}\n {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- for tool_call in message.tool_calls %}\n {{- '\n<tool_call>\n' }} {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '{' }}\n {{- '\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\"' }}\n {{- ', '}}\n {%- if tool_call.arguments is defined %}\n {{- '\"arguments\": ' }}\n {%- if tool_call.arguments is string %}\n {{- tool_call.arguments }}\n {%- else %}\n {{- tool_call.arguments|tojson }}\n {%- endif %}\n {%- endif %}\n {{- '}' }}\n {{- '\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if loop.previtem and loop.previtem.role != \"tool\" %}\n {{- '<|im_start|>tool\n' }}\n {%- endif %}\n {{- '<tool_response>\n' }}\n {{- message.content }}\n {%- if not loop.last %}\n {{- '\n</tool_response>\n' }}\n {%- else %}\n {{- '\n</tool_response>' }}\n {%- endif %}\n {%- if not loop.last and loop.nextitem.role != \"tool\" %}\n {{- '<|im_end|>' }}\n {%- elif loop.last %}\n {{- '<|im_end|>' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\n' }}\n{%- endif %}\n"
|
| 2062 |
+
}
|
| 2063 |
+
],
|
| 2064 |
"clean_up_tokenization_spaces": true,
|
| 2065 |
"eos_token": "<|eot_id|>",
|
| 2066 |
+
"extra_special_tokens": {},
|
| 2067 |
"model_input_names": [
|
| 2068 |
"input_ids",
|
| 2069 |
"attention_mask"
|
| 2070 |
],
|
| 2071 |
"model_max_length": 131072,
|
| 2072 |
+
"pad_token": "<|end_of_text|>",
|
| 2073 |
"padding_side": "left",
|
| 2074 |
+
"tokenizer_class": "PreTrainedTokenizerFast",
|
| 2075 |
+
"unk_token": null
|
| 2076 |
}
|