helcig commited on
Commit
8dd3a63
·
verified ·
1 Parent(s): f45185b

Add files using upload-large-folder tool

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Qwen3-8B/ll_4bit/tokenizer.json filter=lfs diff=lfs merge=lfs -text
Qwen3-8B/ll_4bit/README.md ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ # Quantized Model Checkpoint
2
+
3
+ **Base model:** unknown
4
+
5
+ **Average bitwidth:** unknown
6
+
7
+ See `quantization_config.txt` for full configuration details.
Qwen3-8B/ll_4bit/chat_template.jinja ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
18
+ {%- for message in messages[::-1] %}
19
+ {%- set index = (messages|length - 1) - loop.index0 %}
20
+ {%- if ns.multi_step_tool and message.role == "user" and message.content is string and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}
21
+ {%- set ns.multi_step_tool = false %}
22
+ {%- set ns.last_query_index = index %}
23
+ {%- endif %}
24
+ {%- endfor %}
25
+ {%- for message in messages %}
26
+ {%- if message.content is string %}
27
+ {%- set content = message.content %}
28
+ {%- else %}
29
+ {%- set content = '' %}
30
+ {%- endif %}
31
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
32
+ {{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
33
+ {%- elif message.role == "assistant" %}
34
+ {%- set reasoning_content = '' %}
35
+ {%- if message.reasoning_content is string %}
36
+ {%- set reasoning_content = message.reasoning_content %}
37
+ {%- else %}
38
+ {%- if '</think>' in content %}
39
+ {%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
40
+ {%- set content = content.split('</think>')[-1].lstrip('\n') %}
41
+ {%- endif %}
42
+ {%- endif %}
43
+ {%- if loop.index0 > ns.last_query_index %}
44
+ {%- if loop.last or (not loop.last and reasoning_content) %}
45
+ {{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
46
+ {%- else %}
47
+ {{- '<|im_start|>' + message.role + '\n' + content }}
48
+ {%- endif %}
49
+ {%- else %}
50
+ {{- '<|im_start|>' + message.role + '\n' + content }}
51
+ {%- endif %}
52
+ {%- if message.tool_calls %}
53
+ {%- for tool_call in message.tool_calls %}
54
+ {%- if (loop.first and content) or (not loop.first) %}
55
+ {{- '\n' }}
56
+ {%- endif %}
57
+ {%- if tool_call.function %}
58
+ {%- set tool_call = tool_call.function %}
59
+ {%- endif %}
60
+ {{- '<tool_call>\n{"name": "' }}
61
+ {{- tool_call.name }}
62
+ {{- '", "arguments": ' }}
63
+ {%- if tool_call.arguments is string %}
64
+ {{- tool_call.arguments }}
65
+ {%- else %}
66
+ {{- tool_call.arguments | tojson }}
67
+ {%- endif %}
68
+ {{- '}\n</tool_call>' }}
69
+ {%- endfor %}
70
+ {%- endif %}
71
+ {{- '<|im_end|>\n' }}
72
+ {%- elif message.role == "tool" %}
73
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
74
+ {{- '<|im_start|>user' }}
75
+ {%- endif %}
76
+ {{- '\n<tool_response>\n' }}
77
+ {{- content }}
78
+ {{- '\n</tool_response>' }}
79
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
80
+ {{- '<|im_end|>\n' }}
81
+ {%- endif %}
82
+ {%- endif %}
83
+ {%- endfor %}
84
+ {%- if add_generation_prompt %}
85
+ {{- '<|im_start|>assistant\n' }}
86
+ {%- if enable_thinking is defined and enable_thinking is false %}
87
+ {{- '<think>\n\n</think>\n\n' }}
88
+ {%- endif %}
89
+ {%- endif %}
Qwen3-8B/ll_4bit/config.json ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen3ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "dtype": "float16",
9
+ "eos_token_id": 151645,
10
+ "head_dim": 128,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 4096,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 12288,
15
+ "layer_types": [
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention"
52
+ ],
53
+ "max_position_embeddings": 40960,
54
+ "max_window_layers": 36,
55
+ "model_type": "qwen3",
56
+ "num_attention_heads": 32,
57
+ "num_hidden_layers": 36,
58
+ "num_key_value_heads": 8,
59
+ "pad_token_id": null,
60
+ "rms_norm_eps": 1e-06,
61
+ "rope_parameters": {
62
+ "rope_theta": 1000000,
63
+ "rope_type": "default"
64
+ },
65
+ "sliding_window": null,
66
+ "tie_word_embeddings": false,
67
+ "transformers_version": "5.3.0",
68
+ "use_cache": false,
69
+ "use_sliding_window": false,
70
+ "vocab_size": 151936
71
+ }
Qwen3-8B/ll_4bit/generation_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "temperature": 0.6,
10
+ "top_k": 20,
11
+ "top_p": 0.95,
12
+ "transformers_version": "5.3.0"
13
+ }
Qwen3-8B/ll_4bit/humming_online_quant_config.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "quant_method": "gptq",
3
+ "bits": 4,
4
+ "group_size": 128
5
+ }
Qwen3-8B/ll_4bit/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a198d1c3eae2ad7b97fececf22213c03d0c53d9f9a0809ca3d8d62e9d5249950
3
+ size 16381516808
Qwen3-8B/ll_4bit/quantization_config.txt ADDED
@@ -0,0 +1,252 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model.layers.23.self_attn.k_proj: 4
2
+ model.layers.22.self_attn.o_proj: 4
3
+ model.layers.4.self_attn.o_proj: 4
4
+ model.layers.1.mlp.down_proj: 4
5
+ model.layers.11.mlp.up_proj: 4
6
+ model.layers.5.self_attn.k_proj: 4
7
+ model.layers.14.self_attn.q_proj: 4
8
+ model.layers.17.mlp.down_proj: 4
9
+ model.layers.33.mlp.gate_proj: 4
10
+ model.layers.5.self_attn.v_proj: 4
11
+ model.layers.23.self_attn.v_proj: 4
12
+ model.layers.27.mlp.gate_proj: 4
13
+ model.layers.2.mlp.up_proj: 4
14
+ model.layers.15.self_attn.q_proj: 4
15
+ model.layers.23.self_attn.o_proj: 4
16
+ model.layers.21.mlp.up_proj: 4
17
+ model.layers.22.self_attn.k_proj: 4
18
+ model.layers.4.mlp.down_proj: 4
19
+ model.layers.4.self_attn.k_proj: 4
20
+ model.layers.5.self_attn.o_proj: 4
21
+ model.layers.31.mlp.up_proj: 4
22
+ model.layers.4.self_attn.v_proj: 4
23
+ model.layers.22.self_attn.v_proj: 4
24
+ model.layers.22.mlp.gate_proj: 4
25
+ model.layers.8.mlp.gate_proj: 4
26
+ model.layers.12.mlp.down_proj: 4
27
+ model.layers.7.mlp.gate_proj: 4
28
+ model.layers.35.self_attn.o_proj: 4
29
+ model.layers.21.self_attn.v_proj: 4
30
+ model.layers.34.self_attn.k_proj: 4
31
+ model.layers.7.self_attn.v_proj: 4
32
+ model.layers.7.self_attn.k_proj: 4
33
+ model.layers.6.self_attn.o_proj: 4
34
+ model.layers.34.self_attn.v_proj: 4
35
+ model.layers.20.self_attn.o_proj: 4
36
+ model.layers.13.mlp.up_proj: 4
37
+ model.layers.21.self_attn.k_proj: 4
38
+ model.layers.35.mlp.down_proj: 4
39
+ model.layers.28.self_attn.q_proj: 4
40
+ model.layers.11.mlp.gate_proj: 4
41
+ model.layers.16.self_attn.q_proj: 4
42
+ model.layers.21.mlp.down_proj: 4
43
+ model.layers.35.self_attn.k_proj: 4
44
+ model.layers.33.mlp.up_proj: 4
45
+ model.layers.20.self_attn.v_proj: 4
46
+ model.layers.34.self_attn.o_proj: 4
47
+ model.layers.6.self_attn.v_proj: 4
48
+ model.layers.28.mlp.gate_proj: 4
49
+ model.layers.2.mlp.gate_proj: 4
50
+ model.layers.18.mlp.down_proj: 4
51
+ model.layers.17.self_attn.q_proj: 4
52
+ model.layers.0.mlp.up_proj: 4
53
+ model.layers.24.mlp.down_proj: 4
54
+ model.layers.23.mlp.up_proj: 4
55
+ model.layers.7.self_attn.o_proj: 4
56
+ model.layers.6.self_attn.k_proj: 4
57
+ model.layers.29.self_attn.q_proj: 4
58
+ model.layers.20.self_attn.k_proj: 4
59
+ model.layers.30.mlp.down_proj: 4
60
+ model.layers.14.mlp.gate_proj: 4
61
+ model.layers.21.self_attn.o_proj: 4
62
+ model.layers.35.self_attn.v_proj: 4
63
+ model.layers.6.mlp.up_proj: 4
64
+ model.layers.9.self_attn.q_proj: 4
65
+ model.layers.0.self_attn.k_proj: 4
66
+ model.layers.9.mlp.up_proj: 4
67
+ model.layers.1.self_attn.o_proj: 4
68
+ model.layers.33.self_attn.v_proj: 4
69
+ model.layers.27.self_attn.o_proj: 4
70
+ model.layers.26.self_attn.k_proj: 4
71
+ model.layers.19.self_attn.o_proj: 4
72
+ model.layers.0.mlp.gate_proj: 4
73
+ model.layers.11.self_attn.q_proj: 4
74
+ model.layers.18.self_attn.k_proj: 4
75
+ model.layers.25.mlp.up_proj: 4
76
+ model.layers.26.mlp.down_proj: 4
77
+ model.layers.18.self_attn.v_proj: 4
78
+ model.layers.35.mlp.up_proj: 4
79
+ model.layers.32.self_attn.o_proj: 4
80
+ model.layers.26.self_attn.v_proj: 4
81
+ model.layers.33.self_attn.k_proj: 4
82
+ model.layers.32.mlp.down_proj: 4
83
+ model.layers.16.mlp.gate_proj: 4
84
+ model.layers.0.self_attn.v_proj: 4
85
+ model.layers.15.mlp.up_proj: 4
86
+ model.layers.19.self_attn.k_proj: 4
87
+ model.layers.10.self_attn.q_proj: 4
88
+ model.layers.5.mlp.gate_proj: 4
89
+ model.layers.18.self_attn.o_proj: 4
90
+ model.layers.0.self_attn.o_proj: 4
91
+ model.layers.1.self_attn.k_proj: 4
92
+ model.layers.8.self_attn.q_proj: 4
93
+ model.layers.27.self_attn.k_proj: 4
94
+ model.layers.26.self_attn.o_proj: 4
95
+ model.layers.32.self_attn.v_proj: 4
96
+ model.layers.9.mlp.down_proj: 4
97
+ model.layers.32.self_attn.k_proj: 4
98
+ model.layers.13.mlp.gate_proj: 4
99
+ model.layers.27.self_attn.v_proj: 4
100
+ model.layers.33.self_attn.o_proj: 4
101
+ model.layers.1.self_attn.v_proj: 4
102
+ model.layers.23.mlp.down_proj: 4
103
+ model.layers.19.self_attn.v_proj: 4
104
+ model.layers.2.self_attn.v_proj: 4
105
+ model.layers.6.mlp.down_proj: 4
106
+ model.layers.31.self_attn.k_proj: 4
107
+ model.layers.24.self_attn.v_proj: 4
108
+ model.layers.30.self_attn.o_proj: 4
109
+ model.layers.24.self_attn.k_proj: 4
110
+ model.layers.4.mlp.up_proj: 4
111
+ model.layers.25.self_attn.o_proj: 4
112
+ model.layers.31.self_attn.v_proj: 4
113
+ model.layers.20.mlp.gate_proj: 4
114
+ model.layers.3.self_attn.o_proj: 4
115
+ model.layers.2.self_attn.k_proj: 4
116
+ model.layers.10.mlp.down_proj: 4
117
+ model.layers.34.mlp.gate_proj: 4
118
+ model.layers.13.self_attn.q_proj: 4
119
+ model.layers.28.mlp.up_proj: 4
120
+ model.layers.27.mlp.up_proj: 4
121
+ model.layers.3.mlp.down_proj: 4
122
+ model.layers.3.self_attn.v_proj: 4
123
+ model.layers.19.mlp.gate_proj: 4
124
+ model.layers.31.self_attn.o_proj: 4
125
+ model.layers.25.self_attn.v_proj: 4
126
+ model.layers.30.self_attn.k_proj: 4
127
+ model.layers.29.mlp.down_proj: 4
128
+ model.layers.18.mlp.up_proj: 4
129
+ model.layers.15.mlp.down_proj: 4
130
+ model.layers.17.mlp.up_proj: 4
131
+ model.layers.31.mlp.gate_proj: 4
132
+ model.layers.12.self_attn.q_proj: 4
133
+ model.layers.30.self_attn.v_proj: 4
134
+ model.layers.24.self_attn.o_proj: 4
135
+ model.layers.25.self_attn.k_proj: 4
136
+ model.layers.25.mlp.gate_proj: 4
137
+ model.layers.3.self_attn.k_proj: 4
138
+ model.layers.2.self_attn.o_proj: 4
139
+ model.layers.2.mlp.down_proj: 4
140
+ model.layers.16.self_attn.o_proj: 4
141
+ model.layers.18.mlp.gate_proj: 4
142
+ model.layers.17.self_attn.k_proj: 4
143
+ model.layers.6.self_attn.q_proj: 4
144
+ model.layers.28.mlp.down_proj: 4
145
+ model.layers.28.self_attn.o_proj: 4
146
+ model.layers.20.self_attn.q_proj: 4
147
+ model.layers.29.self_attn.k_proj: 4
148
+ model.layers.29.self_attn.v_proj: 4
149
+ model.layers.14.mlp.down_proj: 4
150
+ model.layers.35.self_attn.q_proj: 4
151
+ model.layers.30.mlp.gate_proj: 4
152
+ model.layers.24.mlp.gate_proj: 4
153
+ model.layers.14.mlp.up_proj: 4
154
+ model.layers.17.self_attn.v_proj: 4
155
+ model.layers.7.self_attn.q_proj: 4
156
+ model.layers.28.self_attn.k_proj: 4
157
+ model.layers.21.self_attn.q_proj: 4
158
+ model.layers.29.self_attn.o_proj: 4
159
+ model.layers.7.mlp.down_proj: 4
160
+ model.layers.16.self_attn.k_proj: 4
161
+ model.layers.17.self_attn.o_proj: 4
162
+ model.layers.34.mlp.up_proj: 4
163
+ model.layers.21.mlp.gate_proj: 4
164
+ model.layers.16.self_attn.v_proj: 4
165
+ model.layers.24.mlp.up_proj: 4
166
+ model.layers.34.self_attn.q_proj: 4
167
+ model.layers.8.mlp.up_proj: 4
168
+ model.layers.11.mlp.down_proj: 4
169
+ model.layers.7.mlp.up_proj: 4
170
+ model.layers.35.mlp.gate_proj: 4
171
+ model.layers.28.self_attn.v_proj: 4
172
+ model.layers.4.mlp.gate_proj: 4
173
+ model.layers.16.mlp.up_proj: 4
174
+ model.layers.15.self_attn.v_proj: 4
175
+ model.layers.19.mlp.up_proj: 4
176
+ model.layers.8.mlp.down_proj: 4
177
+ model.layers.12.mlp.gate_proj: 4
178
+ model.layers.15.self_attn.k_proj: 4
179
+ model.layers.14.self_attn.o_proj: 4
180
+ model.layers.22.self_attn.q_proj: 4
181
+ model.layers.22.mlp.down_proj: 4
182
+ model.layers.4.self_attn.q_proj: 4
183
+ model.layers.14.self_attn.v_proj: 4
184
+ model.layers.26.mlp.up_proj: 4
185
+ model.layers.29.mlp.up_proj: 4
186
+ model.layers.5.mlp.up_proj: 4
187
+ model.layers.1.mlp.gate_proj: 4
188
+ model.layers.27.mlp.down_proj: 4
189
+ model.layers.23.self_attn.q_proj: 4
190
+ model.layers.5.self_attn.q_proj: 4
191
+ model.layers.33.mlp.down_proj: 4
192
+ model.layers.17.mlp.gate_proj: 4
193
+ model.layers.15.self_attn.o_proj: 4
194
+ model.layers.14.self_attn.k_proj: 4
195
+ model.layers.12.self_attn.k_proj: 4
196
+ model.layers.13.self_attn.o_proj: 4
197
+ model.layers.29.mlp.gate_proj: 4
198
+ model.layers.25.self_attn.q_proj: 4
199
+ model.layers.30.mlp.up_proj: 4
200
+ model.layers.3.mlp.gate_proj: 4
201
+ model.layers.19.mlp.down_proj: 4
202
+ model.layers.3.self_attn.q_proj: 4
203
+ model.layers.20.mlp.up_proj: 4
204
+ model.layers.30.self_attn.q_proj: 4
205
+ model.layers.25.mlp.down_proj: 4
206
+ model.layers.12.self_attn.v_proj: 4
207
+ model.layers.3.mlp.up_proj: 4
208
+ model.layers.31.mlp.down_proj: 4
209
+ model.layers.15.mlp.gate_proj: 4
210
+ model.layers.24.self_attn.q_proj: 4
211
+ model.layers.2.self_attn.q_proj: 4
212
+ model.layers.6.mlp.gate_proj: 4
213
+ model.layers.12.self_attn.o_proj: 4
214
+ model.layers.13.self_attn.k_proj: 4
215
+ model.layers.13.self_attn.v_proj: 4
216
+ model.layers.34.mlp.down_proj: 4
217
+ model.layers.10.mlp.gate_proj: 4
218
+ model.layers.10.mlp.up_proj: 4
219
+ model.layers.20.mlp.down_proj: 4
220
+ model.layers.31.self_attn.q_proj: 4
221
+ model.layers.22.mlp.up_proj: 4
222
+ model.layers.32.self_attn.q_proj: 4
223
+ model.layers.8.self_attn.v_proj: 4
224
+ model.layers.5.mlp.down_proj: 4
225
+ model.layers.10.self_attn.v_proj: 4
226
+ model.layers.1.mlp.up_proj: 4
227
+ model.layers.11.self_attn.o_proj: 4
228
+ model.layers.10.self_attn.k_proj: 4
229
+ model.layers.19.self_attn.q_proj: 4
230
+ model.layers.23.mlp.gate_proj: 4
231
+ model.layers.8.self_attn.k_proj: 4
232
+ model.layers.32.mlp.up_proj: 4
233
+ model.layers.1.self_attn.q_proj: 4
234
+ model.layers.9.self_attn.o_proj: 4
235
+ model.layers.9.mlp.gate_proj: 4
236
+ model.layers.13.mlp.down_proj: 4
237
+ model.layers.27.self_attn.q_proj: 4
238
+ model.layers.0.mlp.down_proj: 4
239
+ model.layers.11.self_attn.v_proj: 4
240
+ model.layers.33.self_attn.q_proj: 4
241
+ model.layers.9.self_attn.v_proj: 4
242
+ model.layers.12.mlp.up_proj: 4
243
+ model.layers.8.self_attn.o_proj: 4
244
+ model.layers.0.self_attn.q_proj: 4
245
+ model.layers.9.self_attn.k_proj: 4
246
+ model.layers.26.self_attn.q_proj: 4
247
+ model.layers.16.mlp.down_proj: 4
248
+ model.layers.32.mlp.gate_proj: 4
249
+ model.layers.18.self_attn.q_proj: 4
250
+ model.layers.11.self_attn.k_proj: 4
251
+ model.layers.10.self_attn.o_proj: 4
252
+ model.layers.26.mlp.gate_proj: 4
Qwen3-8B/ll_4bit/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be75606093db2094d7cd20f3c2f385c212750648bd6ea4fb2bf507a6a4c55506
3
+ size 11422650
Qwen3-8B/ll_4bit/tokenizer_config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "backend": "tokenizers",
4
+ "bos_token": null,
5
+ "clean_up_tokenization_spaces": false,
6
+ "eos_token": "<|im_end|>",
7
+ "errors": "replace",
8
+ "extra_special_tokens": [
9
+ "<|im_start|>",
10
+ "<|im_end|>",
11
+ "<|object_ref_start|>",
12
+ "<|object_ref_end|>",
13
+ "<|box_start|>",
14
+ "<|box_end|>",
15
+ "<|quad_start|>",
16
+ "<|quad_end|>",
17
+ "<|vision_start|>",
18
+ "<|vision_end|>",
19
+ "<|vision_pad|>",
20
+ "<|image_pad|>",
21
+ "<|video_pad|>"
22
+ ],
23
+ "is_local": false,
24
+ "model_max_length": 131072,
25
+ "pad_token": "<|endoftext|>",
26
+ "split_special_tokens": false,
27
+ "tokenizer_class": "Qwen2Tokenizer",
28
+ "unk_token": null
29
+ }