elribonazo commited on
Commit
d1e57c0
·
verified ·
1 Parent(s): 5b53d38

Upload folder using huggingface_hub

Browse files
.DS_Store CHANGED
Binary files a/.DS_Store and b/.DS_Store differ
 
.gitattributes CHANGED
@@ -34,5 +34,11 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  onnx/model.onnx_data filter=lfs diff=lfs merge=lfs -text
 
37
  onnx/model_fp16.onnx_data filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
38
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  onnx/model.onnx_data filter=lfs diff=lfs merge=lfs -text
37
+ onnx/model_bnb4.onnx_data filter=lfs diff=lfs merge=lfs -text
38
  onnx/model_fp16.onnx_data filter=lfs diff=lfs merge=lfs -text
39
+ onnx/model_int8.onnx_data filter=lfs diff=lfs merge=lfs -text
40
+ onnx/model_q4.onnx_data filter=lfs diff=lfs merge=lfs -text
41
+ onnx/model_q4f16.onnx_data filter=lfs diff=lfs merge=lfs -text
42
+ onnx/model_quantized.onnx_data filter=lfs diff=lfs merge=lfs -text
43
+ onnx/model_uint8.onnx_data filter=lfs diff=lfs merge=lfs -text
44
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
config.json CHANGED
@@ -1,42 +1,38 @@
1
  {
2
  "_attn_implementation_autoset": true,
3
- "_name_or_path": "nguyenthanhthuan/Llama_3.2_1B_Intruct_Tool_Calling_V2",
4
  "architectures": [
5
  "LlamaForCausalLM"
6
  ],
7
  "attention_bias": false,
8
  "attention_dropout": 0.0,
9
  "bos_token_id": 128000,
10
- "eos_token_id": [
11
- 128001,
12
- 128008,
13
- 128009
14
- ],
15
- "head_dim": 64,
16
  "hidden_act": "silu",
17
- "hidden_size": 2048,
18
  "initializer_range": 0.02,
19
- "intermediate_size": 8192,
20
  "max_position_embeddings": 131072,
21
  "mlp_bias": false,
22
  "model_type": "llama",
23
  "num_attention_heads": 32,
24
- "num_hidden_layers": 16,
25
  "num_key_value_heads": 8,
26
  "pad_token_id": 128004,
27
  "pretraining_tp": 1,
28
  "rms_norm_eps": 1e-05,
29
  "rope_scaling": {
30
- "factor": 32.0,
31
  "high_freq_factor": 4.0,
32
  "low_freq_factor": 1.0,
33
  "original_max_position_embeddings": 8192,
34
  "rope_type": "llama3"
35
  },
36
  "rope_theta": 500000.0,
37
- "tie_word_embeddings": true,
38
  "transformers_version": "4.46.3",
39
- "unsloth_version": "2024.10.7",
40
  "use_cache": false,
41
  "vocab_size": 128256
42
  }
 
1
  {
2
  "_attn_implementation_autoset": true,
3
+ "_name_or_path": "NousResearch/DeepHermes-3-Llama-3-8B-Preview",
4
  "architectures": [
5
  "LlamaForCausalLM"
6
  ],
7
  "attention_bias": false,
8
  "attention_dropout": 0.0,
9
  "bos_token_id": 128000,
10
+ "eos_token_id": 128009,
11
+ "head_dim": 128,
 
 
 
 
12
  "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
  "initializer_range": 0.02,
15
+ "intermediate_size": 14336,
16
  "max_position_embeddings": 131072,
17
  "mlp_bias": false,
18
  "model_type": "llama",
19
  "num_attention_heads": 32,
20
+ "num_hidden_layers": 32,
21
  "num_key_value_heads": 8,
22
  "pad_token_id": 128004,
23
  "pretraining_tp": 1,
24
  "rms_norm_eps": 1e-05,
25
  "rope_scaling": {
26
+ "factor": 8.0,
27
  "high_freq_factor": 4.0,
28
  "low_freq_factor": 1.0,
29
  "original_max_position_embeddings": 8192,
30
  "rope_type": "llama3"
31
  },
32
  "rope_theta": 500000.0,
33
+ "tie_word_embeddings": false,
34
  "transformers_version": "4.46.3",
35
+ "unsloth_fixed": true,
36
  "use_cache": false,
37
  "vocab_size": 128256
38
  }
generation_config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "bos_token_id": 128000,
3
  "do_sample": true,
4
  "eos_token_id": [
 
1
  {
2
+ "_from_model_config": true,
3
  "bos_token_id": 128000,
4
  "do_sample": true,
5
  "eos_token_id": [
onnx/model.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4d5f9e5b9f911961b179a4688efbb50b616ecd66342b0faf9da0b78a0e691e6d
3
- size 435319
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdd315c3543dc1c2c7fea02fee19dcc5988b3bf455365892132e0f1ed01f5b3f
3
+ size 858913
onnx/model.onnx_data CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c4f269b9750ade29e17b6daec21c5d76bf2f0c1206b9e8e1090a2ad2d3b1ef95
3
- size 4943257600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fe497e51b001cf53241abf12fd69a9561cb75e6290a7ef41b94d5519499e832
3
+ size 32121044992
onnx/model_bnb4.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:22b933289dbe1f9af7c3547f17c68fc157e5862960006d9e7c27522b6cdb4cea
3
- size 1598743831
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c078b3be1198c2ef021344f5ce42438cf55beaa2315d635b20c457a4b90c6684
3
+ size 911812
onnx/model_fp16.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:957b23c342e124cd7ffff7de2275b5b30a27f54d8cfee8f6a426ae1663d64e68
3
- size 436000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76fcb21a89369fc32d3795df146930abb06c163443824f4f1ec2d90bb69b03d7
3
+ size 860108
onnx/model_fp16.onnx_data CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c9733aab730ec38ed1ed2b8e85ba96773627c212e41c4faec2771f3163968ad1
3
- size 2471628800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02911f28a0354c44876f3329ecf587d5ca16e65698bb2869937d8544d652aeaa
3
+ size 16060522496
onnx/model_int8.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:827ec32473381c80c4a147332bf3bf7b9b6682ba15dd6da0f20f12457ca28071
3
- size 1236565479
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b1bb76d6a54817b3f2520fcd98dc4e95dcbc9bfc0ce7c57dca96c16e4c2b2dd
3
+ size 1105763
onnx/model_q4.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f54166d51772c5e300692fd1893ca72be7d91b92fa951c15d11c042992757680
3
- size 1659560375
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3827b6b586a90af38f1f11a34192dbc80f08b3e8e30deeb68253b7dd1c4153fb
3
+ size 908949
onnx/model_q4f16.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c6d37c47b5284b00dac46b21d04ea204f336424925e7f1e5fd3536140f815390
3
- size 1073271141
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c7183ef2e2436a57314ca1603bbd407219c03ac137b8eef75ff92c4163f9631
3
+ size 910416
onnx/model_quantized.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:827ec32473381c80c4a147332bf3bf7b9b6682ba15dd6da0f20f12457ca28071
3
- size 1236565479
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92209e407f1d3bf3bebc152743293fb7a3a4ec00fe2e514bbb378cc94796dac9
3
+ size 1107250
onnx/model_uint8.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:17af458ac90f9069de8f87cf5c38620c954f24fa73ef7eff32bd159427947a9d
3
- size 1236565541
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c9ed7e84bdb801d0e4a6606067eac6ae14dd76948b408ca9f49bc7475599665
3
+ size 1106175
special_tokens_map.json CHANGED
@@ -14,7 +14,7 @@
14
  "single_word": false
15
  },
16
  "pad_token": {
17
- "content": "<|finetune_right_pad_id|>",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
 
14
  "single_word": false
15
  },
16
  "pad_token": {
17
+ "content": "<|end_of_text|>",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
tokenizer_config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "added_tokens_decoder": {
3
  "128000": {
4
  "content": "<|begin_of_text|>",
@@ -2050,15 +2051,26 @@
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
2053
- "chat_template": "\n{%- if messages[0]['role'] == 'system' -%}\n<|start_header_id|>system<|end_header_id|>\n\n{{ messages[0]['content'] }}\n{%- endif -%}\n{%- for message in messages -%}\n {%- if message['role'] == 'user' -%}\n <|start_header_id|>user<|end_header_id|>\n {{ message['content'] }}\n {%- elif message['role'] in ['assistant'] -%}\n <|start_header_id|>assistant<|end_header_id|>\n {{ message['content'] }}<|eot_id|>\n {%- elif message['role'] in ['tool_call'] -%}\n <|start_header_id|>tool_call<|end_header_id|>\n {{ message['content'] }}<|eot_id|>\n {%- elif message['role'] == 'tool_response' -%}\n <|start_header_id|>tool_respone<|end_header_id|>\n {{ message['content'] }}\n {%- endif -%}\n{%- endfor -%}\n{%- if add_generation_prompt -%} <|start_header_id|>assistant<|end_header_id|>\n{%- endif -%}\n",
 
 
 
 
 
 
 
 
 
2054
  "clean_up_tokenization_spaces": true,
2055
  "eos_token": "<|eot_id|>",
 
2056
  "model_input_names": [
2057
  "input_ids",
2058
  "attention_mask"
2059
  ],
2060
  "model_max_length": 131072,
2061
- "pad_token": "<|finetune_right_pad_id|>",
2062
  "padding_side": "left",
2063
- "tokenizer_class": "PreTrainedTokenizerFast"
 
2064
  }
 
1
  {
2
+ "add_bos_token": true,
3
  "added_tokens_decoder": {
4
  "128000": {
5
  "content": "<|begin_of_text|>",
 
2051
  }
2052
  },
2053
  "bos_token": "<|begin_of_text|>",
2054
+ "chat_template": [
2055
+ {
2056
+ "name": "default",
2057
+ "template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}"
2058
+ },
2059
+ {
2060
+ "name": "tool_use",
2061
+ "template": "{%- macro json_to_python_type(json_spec) %}\n{%- set basic_type_map = {\n \"string\": \"str\",\n \"number\": \"float\",\n \"integer\": \"int\",\n \"boolean\": \"bool\"\n} %}\n\n{%- if basic_type_map[json_spec.type] is defined %}\n {{- basic_type_map[json_spec.type] }}\n{%- elif json_spec.type == \"array\" %}\n {{- \"list[\" + json_to_python_type(json_spec|items) + \"]\"}}\n{%- elif json_spec.type == \"object\" %}\n {%- if json_spec.additionalProperties is defined %}\n {{- \"dict[str, \" + json_to_python_type(json_spec.additionalProperties) + ']'}}\n {%- else %}\n {{- \"dict\" }}\n {%- endif %}\n{%- elif json_spec.type is iterable %}\n {{- \"Union[\" }}\n {%- for t in json_spec.type %}\n {{- json_to_python_type({\"type\": t}) }}\n {%- if not loop.last %}\n {{- \",\" }} \n {%- endif %}\n {%- endfor %}\n {{- \"]\" }}\n{%- else %}\n {{- \"Any\" }}\n{%- endif %}\n{%- endmacro %}\n\n\n{{- bos_token }}\n{{- '<|im_start|>system\n' }}\n{{- \"You are a function calling AI model. You are provided with function signatures within <tools></tools> XML tags. You may call one or more functions to assist with the user query. Don't make assumptions about what values to plug into functions. Here are the available tools: <tools> \" }}\n{%- for tool in tools %}\n {%- if tool.function is defined %}\n {%- set tool = tool.function %}\n {%- endif %}\n {{- '{\"type\": \"function\", \"function\": ' }}\n {{- '{\"name\": \"' + tool.name + '\", ' }}\n {{- '\"description\": \"' + tool.name + '(' }}\n {%- for param_name, param_fields in tool.parameters.properties|items %}\n {{- param_name + \": \" + json_to_python_type(param_fields) }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- if tool.return is defined %}\n {{- \" -> \" + json_to_python_type(tool.return) }}\n {%- endif %}\n {{- \" - \" + tool.description + \"\n\n\" }}\n {%- for param_name, param_fields in tool.parameters.properties|items %}\n {%- if loop.first %}\n {{- \" Args:\n\" }}\n {%- endif %}\n {{- \" \" + param_name + \"(\" + json_to_python_type(param_fields) + \"): \" + param_fields.description|trim }}\n {%- endfor %}\n {%- if tool.return is defined and tool.return.description is defined %}\n {{- \"\n Returns:\n \" + tool.return.description }}\n {%- endif %}\n {{- '\"' }}\n {{- ', \"parameters\": ' }}\n {%- if tool.parameters.properties | length == 0 %}\n {{- \"{}\" }}\n {%- else %}\n {{- tool.parameters|tojson }}\n {%- endif %}\n {{- \"}\" }}\n {%- if not loop.last %}\n {{- \"\n\" }}\n {%- endif %}\n{%- endfor %}\n{{- \" </tools>\" }}\n{{- 'Use the following pydantic model json schema for each tool call you will make: {\"properties\": {\"name\": {\"title\": \"Name\", \"type\": \"string\"}, \"arguments\": {\"title\": \"Arguments\", \"type\": \"object\"}}, \"required\": [\"name\", \"arguments\"], \"title\": \"FunctionCall\", \"type\": \"object\"}}\n' }}\n{{- \"For each function call return a json object with function name and arguments within <tool_call></tool_call> XML tags as follows:\n\" }}\n{{- \"<tool_call>\n\" }}\n{{- '{\"name\": <function-name>, \"arguments\": <args-dict>}\n' }}\n{{- '</tool_call><|im_end|>\n' }}\n{%- for message in messages %}\n {%- if message.role == \"user\" or message.role == \"system\" or (message.role == \"assistant\" and message.tool_calls is not defined) %}\n {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- for tool_call in message.tool_calls %}\n {{- '\n<tool_call>\n' }} {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '{' }}\n {{- '\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\"' }}\n {{- ', '}}\n {%- if tool_call.arguments is defined %}\n {{- '\"arguments\": ' }}\n {%- if tool_call.arguments is string %}\n {{- tool_call.arguments }}\n {%- else %}\n {{- tool_call.arguments|tojson }}\n {%- endif %}\n {%- endif %}\n {{- '}' }}\n {{- '\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if loop.previtem and loop.previtem.role != \"tool\" %}\n {{- '<|im_start|>tool\n' }}\n {%- endif %}\n {{- '<tool_response>\n' }}\n {{- message.content }}\n {%- if not loop.last %}\n {{- '\n</tool_response>\n' }}\n {%- else %}\n {{- '\n</tool_response>' }}\n {%- endif %}\n {%- if not loop.last and loop.nextitem.role != \"tool\" %}\n {{- '<|im_end|>' }}\n {%- elif loop.last %}\n {{- '<|im_end|>' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\n' }}\n{%- endif %}\n"
2062
+ }
2063
+ ],
2064
  "clean_up_tokenization_spaces": true,
2065
  "eos_token": "<|eot_id|>",
2066
+ "extra_special_tokens": {},
2067
  "model_input_names": [
2068
  "input_ids",
2069
  "attention_mask"
2070
  ],
2071
  "model_max_length": 131072,
2072
+ "pad_token": "<|end_of_text|>",
2073
  "padding_side": "left",
2074
+ "tokenizer_class": "PreTrainedTokenizerFast",
2075
+ "unk_token": null
2076
  }