elribonazo commited on
Commit
e1c8b22
·
verified ·
1 Parent(s): f39d702

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_attn_implementation_autoset": true,
3
- "_name_or_path": "nguyenthanhthuan/Llama_3.2_1B_Intruct_Tool_Calling",
4
  "architectures": [
5
  "LlamaForCausalLM"
6
  ],
 
1
  {
2
  "_attn_implementation_autoset": true,
3
+ "_name_or_path": "nguyenthanhthuan/Llama_3.2_1B_Intruct_Tool_Calling_V2",
4
  "architectures": [
5
  "LlamaForCausalLM"
6
  ],
onnx/model.onnx_data CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:923b8df813e2e261b5c44aa0c883111628657a2f58816174ebb37cdc5dab8878
3
  size 4943257600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4f269b9750ade29e17b6daec21c5d76bf2f0c1206b9e8e1090a2ad2d3b1ef95
3
  size 4943257600
onnx/model_bnb4.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e4b48a54e5d610d90e305ce6c5be96571aed3e65d76e651093703bf72c15151c
3
- size 938490
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:973e5064cbd1cffa31cd30c9a1371cc648af2831391fdbf7768cd9419e7d1c00
3
+ size 1598757277
onnx/model_fp16.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a11061369a411450806a507cd93cd765200b4b6cc313384a55c28755fa1e190d
3
- size 903217
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a2f337289d0805d9807d78c4194f2dbc1d3dfb4d38e1447f8134d76284b3711
3
+ size 457645
onnx/model_fp16.onnx_data CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0fe6e4fcfbedbdb5b9466da557bc8e02d0fa1d5407ff30e96c8dcf818eef3bc
3
- size 16060522496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9733aab730ec38ed1ed2b8e85ba96773627c212e41c4faec2771f3163968ad1
3
+ size 2471628800
onnx/model_int8.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:30baf2344e2b73f544d962a305cbb48c791f8485c879870cc0d15f160f5a2941
3
- size 1132497
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a392606574450b062054554547f0b8d9c8b77b1678417107e37d007e31acc5b3
3
+ size 1236578925
onnx/model_q4.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3ab92135858a6c4909f7692c800a7204ed3fdd05f8a5dd4b552e3b9aba4ccd5e
3
- size 935627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5aa75b94996669549ce465cbe97688b269e438d4986e665695a7e98a89927a74
3
+ size 1659573821
onnx/model_q4f16.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0c610b112b9ae7d09194d888e7525b5c788f03f257f8ddd50e3b753760b6c18f
3
- size 953581
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f678c74e7eab4dac069eb38d0575cef3571205c12531fee6cd3d30341735a54
3
+ size 1073292786
onnx/model_quantized.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a362851c5bde1c40dfb8ac41b7e0429e2a5c72fbc3be86d77899e438206aeb5f
3
- size 1133984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a392606574450b062054554547f0b8d9c8b77b1678417107e37d007e31acc5b3
3
+ size 1236578925
onnx/model_uint8.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a73570bf3ac871dfbcb0e8006d46154cf174476393113d0b6a2539b43a097eae
3
- size 1132905
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:607c40015dceb3d5970af129484bce2f8dba78f3ad10881ea4586d20a850cfc6
3
+ size 1236578987
tokenizer_config.json CHANGED
@@ -2050,7 +2050,7 @@
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
2053
- "chat_template": "\n{%- if messages[0]['role'] == 'system' -%}\n<|start_header_id|>system<|end_header_id|>\n\n{{ messages[0]['content'] }}\n{%- endif -%}\n{%- for message in messages -%}\n {%- if message['role'] == 'user' -%}\n <|start_header_id|>user<|end_header_id|>\n {{ message['content'] }}\n {%- elif message['role'] in ['tool', 'assistant'] -%}\n <|start_header_id|>assistant<|end_header_id|>\n {{ message['content'] }}<|eot_id|>\n {%- elif message['role'] == 'tool_response' -%}\n <|start_header_id|>assistant<|end_header_id|>\n {{ message['content'] }}\n {%- endif -%}\n{%- endfor -%}\n{%- if add_generation_prompt -%} <|start_header_id|>assistant<|end_header_id|>\n{%- endif -%}\n",
2054
  "clean_up_tokenization_spaces": true,
2055
  "eos_token": "<|eot_id|>",
2056
  "model_input_names": [
 
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
2053
+ "chat_template": "\n{%- if messages[0]['role'] == 'system' -%}\n<|start_header_id|>system<|end_header_id|>\n\n{{ messages[0]['content'] }}\n{%- endif -%}\n{%- for message in messages -%}\n {%- if message['role'] == 'user' -%}\n <|start_header_id|>user<|end_header_id|>\n {{ message['content'] }}\n {%- elif message['role'] in ['assistant'] -%}\n <|start_header_id|>assistant<|end_header_id|>\n {{ message['content'] }}<|eot_id|>\n {%- elif message['role'] in ['tool_call'] -%}\n <|start_header_id|>tool_call<|end_header_id|>\n {{ message['content'] }}<|eot_id|>\n {%- elif message['role'] == 'tool_response' -%}\n <|start_header_id|>tool_respone<|end_header_id|>\n {{ message['content'] }}\n {%- endif -%}\n{%- endfor -%}\n{%- if add_generation_prompt -%} <|start_header_id|>assistant<|end_header_id|>\n{%- endif -%}\n",
2054
  "clean_up_tokenization_spaces": true,
2055
  "eos_token": "<|eot_id|>",
2056
  "model_input_names": [