shed219 commited on
Commit
6fd90c7
·
verified ·
1 Parent(s): 1c546fd

Upload folder using huggingface_hub

Browse files
added_tokens.json CHANGED
@@ -2,10 +2,10 @@
2
  "</think>": 151668,
3
  "</tool_call>": 151658,
4
  "</tool_response>": 151666,
5
- "<<|im_start|>>": 151669,
6
  "<think>": 151667,
7
  "<tool_call>": 151657,
8
  "<tool_response>": 151665,
 
9
  "<|box_end|>": 151649,
10
  "<|box_start|>": 151648,
11
  "<|endoftext|>": 151643,
 
2
  "</think>": 151668,
3
  "</tool_call>": 151658,
4
  "</tool_response>": 151666,
 
5
  "<think>": 151667,
6
  "<tool_call>": 151657,
7
  "<tool_response>": 151665,
8
+ "<|PAD_TOKEN|>": 151669,
9
  "<|box_end|>": 151649,
10
  "<|box_start|>": 151648,
11
  "<|endoftext|>": 151643,
chat_template.jinja CHANGED
@@ -1 +1,97 @@
1
- {% for message in messages %}{% if message['role'] == 'user' %}{{ '<<|im_start|>>user\n' + message['content'] + '<<|im_end|>>\n' }}{% elif message['role'] == 'assistant' %}{{ '<<|im_start|>>assistant\n' + message['content'] + '<<|im_end|>>\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<<|im_start|>>assistant\n' }}{% endif %}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
18
+ {%- for forward_message in messages %}
19
+ {%- set index = (messages|length - 1) - loop.index0 %}
20
+ {%- set message = messages[index] %}
21
+ {%- set tool_start = '<tool_response>' %}
22
+ {%- set tool_start_length = tool_start|length %}
23
+ {%- set start_of_message = message.content[:tool_start_length] %}
24
+ {%- set tool_end = '</tool_response>' %}
25
+ {%- set tool_end_length = tool_end|length %}
26
+ {%- set start_pos = (message.content|length) - tool_end_length %}
27
+ {%- if start_pos < 0 %}
28
+ {%- set start_pos = 0 %}
29
+ {%- endif %}
30
+ {%- set end_of_message = message.content[start_pos:] %}
31
+ {%- if ns.multi_step_tool and message.role == "user" and not(start_of_message == tool_start and end_of_message == tool_end) %}
32
+ {%- set ns.multi_step_tool = false %}
33
+ {%- set ns.last_query_index = index %}
34
+ {%- endif %}
35
+ {%- endfor %}
36
+ {%- for message in messages %}
37
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
38
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
39
+ {%- elif message.role == "assistant" %}
40
+ {%- set content = message.content %}
41
+ {%- set reasoning_content = '' %}
42
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
43
+ {%- set reasoning_content = message.reasoning_content %}
44
+ {%- else %}
45
+ {%- if '</think>' in message.content %}
46
+ {%- set content = (message.content.split('</think>')|last).lstrip('\n') %}
47
+ {%- set reasoning_content = (message.content.split('</think>')|first).rstrip('\n') %}
48
+ {%- set reasoning_content = (reasoning_content.split('<think>')|last).lstrip('\n') %}
49
+ {%- endif %}
50
+ {%- endif %}
51
+ {%- if loop.index0 > ns.last_query_index %}
52
+ {%- if loop.last or (not loop.last and reasoning_content) %}
53
+ {{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
54
+ {%- else %}
55
+ {{- '<|im_start|>' + message.role + '\n' + content }}
56
+ {%- endif %}
57
+ {%- else %}
58
+ {{- '<|im_start|>' + message.role + '\n' + content }}
59
+ {%- endif %}
60
+ {%- if message.tool_calls %}
61
+ {%- for tool_call in message.tool_calls %}
62
+ {%- if (loop.first and content) or (not loop.first) %}
63
+ {{- '\n' }}
64
+ {%- endif %}
65
+ {%- if tool_call.function %}
66
+ {%- set tool_call = tool_call.function %}
67
+ {%- endif %}
68
+ {{- '<tool_call>\n{"name": "' }}
69
+ {{- tool_call.name }}
70
+ {{- '", "arguments": ' }}
71
+ {%- if tool_call.arguments is string %}
72
+ {{- tool_call.arguments }}
73
+ {%- else %}
74
+ {{- tool_call.arguments | tojson }}
75
+ {%- endif %}
76
+ {{- '}\n</tool_call>' }}
77
+ {%- endfor %}
78
+ {%- endif %}
79
+ {{- '<|im_end|>\n' }}
80
+ {%- elif message.role == "tool" %}
81
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
82
+ {{- '<|im_start|>user' }}
83
+ {%- endif %}
84
+ {{- '\n<tool_response>\n' }}
85
+ {{- message.content }}
86
+ {{- '\n</tool_response>' }}
87
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
88
+ {{- '<|im_end|>\n' }}
89
+ {%- endif %}
90
+ {%- endif %}
91
+ {%- endfor %}
92
+ {%- if add_generation_prompt %}
93
+ {{- '<|im_start|>assistant\n' }}
94
+ {%- if enable_thinking is defined and enable_thinking is false %}
95
+ {{- '<think>\n\n</think>\n\n' }}
96
+ {%- endif %}
97
+ {%- endif %}
config.json ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen3ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "dtype": "float16",
8
+ "eos_token_id": 151645,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 2560,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 9728,
14
+ "layer_types": [
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention"
51
+ ],
52
+ "max_position_embeddings": 40960,
53
+ "max_window_layers": 36,
54
+ "model_type": "qwen3",
55
+ "num_attention_heads": 32,
56
+ "num_hidden_layers": 36,
57
+ "num_key_value_heads": 8,
58
+ "pad_token_id": 151669,
59
+ "quantization_config": {
60
+ "bnb_4bit_compute_dtype": "float16",
61
+ "bnb_4bit_quant_type": "nf4",
62
+ "bnb_4bit_use_double_quant": true,
63
+ "llm_int8_enable_fp32_cpu_offload": false,
64
+ "llm_int8_has_fp16_weight": false,
65
+ "llm_int8_skip_modules": null,
66
+ "llm_int8_threshold": 6.0,
67
+ "load_in_4bit": true,
68
+ "load_in_8bit": false,
69
+ "quant_method": "bitsandbytes"
70
+ },
71
+ "rms_norm_eps": 1e-06,
72
+ "rope_scaling": null,
73
+ "rope_theta": 1000000,
74
+ "sliding_window": null,
75
+ "tie_word_embeddings": true,
76
+ "transformers_version": "4.57.6",
77
+ "unsloth_fixed": true,
78
+ "unsloth_version": "2026.2.1",
79
+ "use_cache": true,
80
+ "use_sliding_window": false,
81
+ "vocab_size": 151936
82
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "eos_token_id": 151645,
4
+ "max_length": 40960,
5
+ "pad_token_id": 151669,
6
+ "transformers_version": "4.57.6"
7
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d48d36b25f5ce2187d8b2e89cd04840f3f9e28deebd615d8d40875f71f15e17f
3
+ size 2653133499
special_tokens_map.json CHANGED
@@ -14,22 +14,15 @@
14
  "<|image_pad|>",
15
  "<|video_pad|>"
16
  ],
17
- "bos_token": {
18
- "content": "<<|im_start|>>",
19
- "lstrip": false,
20
- "normalized": false,
21
- "rstrip": false,
22
- "single_word": false
23
- },
24
  "eos_token": {
25
- "content": "<|endoftext|>",
26
  "lstrip": false,
27
  "normalized": false,
28
  "rstrip": false,
29
  "single_word": false
30
  },
31
  "pad_token": {
32
- "content": "<|vision_pad|>",
33
  "lstrip": false,
34
  "normalized": false,
35
  "rstrip": false,
 
14
  "<|image_pad|>",
15
  "<|video_pad|>"
16
  ],
 
 
 
 
 
 
 
17
  "eos_token": {
18
+ "content": "<|im_end|>",
19
  "lstrip": false,
20
  "normalized": false,
21
  "rstrip": false,
22
  "single_word": false
23
  },
24
  "pad_token": {
25
+ "content": "<|PAD_TOKEN|>",
26
  "lstrip": false,
27
  "normalized": false,
28
  "rstrip": false,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:57a156d3ba9c5a6938632dba2d9a635cbf34ab18ac3233e32644deefca019db5
3
- size 11422845
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10ba4ba91270b1a50e5cd8e51023bccc66fc4ac4909dd7ae7ab29433411c9bb9
3
+ size 11422844
tokenizer_config.json CHANGED
@@ -211,7 +211,7 @@
211
  "special": false
212
  },
213
  "151669": {
214
- "content": "<<|im_start|>>",
215
  "lstrip": false,
216
  "normalized": false,
217
  "rstrip": false,
@@ -234,13 +234,13 @@
234
  "<|image_pad|>",
235
  "<|video_pad|>"
236
  ],
237
- "bos_token": "<<|im_start|>>",
238
  "clean_up_tokenization_spaces": false,
239
- "eos_token": "<|endoftext|>",
240
  "errors": "replace",
241
  "extra_special_tokens": {},
242
- "model_max_length": 32768,
243
- "pad_token": "<|vision_pad|>",
244
  "padding_side": "left",
245
  "split_special_tokens": false,
246
  "tokenizer_class": "Qwen2Tokenizer",
 
211
  "special": false
212
  },
213
  "151669": {
214
+ "content": "<|PAD_TOKEN|>",
215
  "lstrip": false,
216
  "normalized": false,
217
  "rstrip": false,
 
234
  "<|image_pad|>",
235
  "<|video_pad|>"
236
  ],
237
+ "bos_token": null,
238
  "clean_up_tokenization_spaces": false,
239
+ "eos_token": "<|im_end|>",
240
  "errors": "replace",
241
  "extra_special_tokens": {},
242
+ "model_max_length": 40960,
243
+ "pad_token": "<|PAD_TOKEN|>",
244
  "padding_side": "left",
245
  "split_special_tokens": false,
246
  "tokenizer_class": "Qwen2Tokenizer",