lobbuko commited on
Commit
68658bd
·
verified ·
1 Parent(s): 40a7bf9

Upload GLM-4.6V-Flash-oQ4 via oMLX

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: mlx
3
+ tags:
4
+ - mlx
5
+ - oq
6
+ - quantized
7
+ ---
8
+
9
+ # GLM-4.6V-Flash-oQ4
10
+
11
+ This model was quantized using [oQ](https://github.com/jundot/omlx) mixed-precision quantization.
12
+
13
+ ## Quantization details
14
+
15
+ - **Model type**: glm4v
16
+ - **Bits**: 4
17
+ - **Group size**: 64
18
+ - **Format**: MLX safetensors
chat_template.jinja ADDED
@@ -0,0 +1,140 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [gMASK]<sop>
2
+ {%- if tools -%}
3
+ <|system|>
4
+ # Tools
5
+
6
+ You may call one or more functions to assist with the user query.
7
+
8
+ You are provided with function signatures within <tools></tools> XML tags:
9
+ <tools>
10
+ {% for tool in tools %}
11
+ {{ tool | tojson(ensure_ascii=False) }}
12
+ {% endfor %}
13
+ </tools>
14
+
15
+ For each function call, output the function name and arguments within the following XML format:
16
+ <tool_call>{function-name}
17
+ <arg_key>{arg-key-1}</arg_key>
18
+ <arg_value>{arg-value-1}</arg_value>
19
+ <arg_key>{arg-key-2}</arg_key>
20
+ <arg_value>{arg-value-2}</arg_value>
21
+ ...
22
+ </tool_call>{%- endif -%}
23
+ {%- macro visible_text(content) -%}
24
+ {%- if content is string -%}
25
+ {{- content }}
26
+ {%- elif content is iterable and content is not mapping -%}
27
+ {%- for item in content -%}
28
+ {%- if item is mapping and item.type == 'text' -%}
29
+ {{- item.text }}
30
+ {%- elif item is mapping and (item.type == 'image' or 'image' in item) -%}
31
+ <|begin_of_image|><|image|><|end_of_image|>
32
+ {%- elif item is mapping and (item.type == 'video' or 'video' in item) -%}
33
+ <|begin_of_video|><|video|><|end_of_video|>
34
+ {%- elif item is string -%}
35
+ {{- item }}
36
+ {%- endif -%}
37
+ {%- endfor -%}
38
+ {%- else -%}
39
+ {{- content }}
40
+ {%- endif -%}
41
+ {%- endmacro -%}
42
+ {%- set ns = namespace(last_user_index=-1) %}
43
+ {%- for m in messages %}
44
+ {%- if m.role == 'user' %}
45
+ {% set ns.last_user_index = loop.index0 -%}
46
+ {%- endif %}
47
+ {%- endfor %}
48
+ {% for m in messages %}
49
+ {%- if m.role == 'user' -%}<|user|>
50
+ {% if m.content is string %}
51
+ {{ m.content }}
52
+ {%- else %}
53
+ {%- for item in m.content %}
54
+ {% if item.type == 'video' or 'video' in item %}
55
+ <|begin_of_video|><|video|><|end_of_video|>{% elif item.type == 'image' or 'image' in item %}
56
+ <|begin_of_image|><|image|><|end_of_image|>{% elif item.type == 'text' %}
57
+ {{ item.text }}
58
+ {%- endif %}
59
+ {%- endfor %}
60
+ {%- endif %}
61
+ {{- '/nothink' if (enable_thinking is defined and not enable_thinking and not visible_text(m.content).endswith("/nothink")) else '' -}}
62
+ {%- elif m.role == 'assistant' -%}
63
+ <|assistant|>
64
+ {%- set reasoning_content = '' %}
65
+ {%- set content = visible_text(m.content) %}
66
+ {%- if m.reasoning_content is string %}
67
+ {%- set reasoning_content = m.reasoning_content %}
68
+ {%- else %}
69
+ {%- if '</think>' in content %}
70
+ {%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
71
+ {%- set content = content.split('</think>')[-1].lstrip('\n') %}
72
+ {%- endif %}
73
+ {%- endif %}
74
+ {%- if loop.index0 > ns.last_user_index and reasoning_content -%}
75
+ {{ '\n<think>' + reasoning_content.strip() + '</think>'}}
76
+ {%- else -%}
77
+ {{ '\n<think></think>' }}
78
+ {%- endif -%}
79
+ {%- if content.strip() -%}
80
+ {{ '\n' + content.strip() }}
81
+ {%- endif -%}
82
+ {% if m.tool_calls %}
83
+ {% for tc in m.tool_calls %}
84
+ {%- if tc.function %}
85
+ {%- set tc = tc.function %}
86
+ {%- endif %}
87
+ {{ '\n<tool_call>' + tc.name }}
88
+ {% set _args = tc.arguments %}
89
+ {% for k, v in _args.items() %}
90
+ <arg_key>{{ k }}</arg_key>
91
+ <arg_value>{{ v | tojson(ensure_ascii=False) if v is not string else v }}</arg_value>
92
+ {% endfor %}
93
+ </tool_call>{% endfor %}
94
+ {% endif %}
95
+ {%- elif m.role == 'tool' -%}
96
+ {%- if m.content is string -%}
97
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
98
+ {{- '<|observation|>' }}
99
+ {%- endif %}
100
+ {{- '\n<tool_response>\n' }}
101
+ {{- m.content }}
102
+ {{- '\n</tool_response>' }}
103
+ {% elif m.content is iterable and m.content is not mapping %}
104
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
105
+ {{- '<|observation|>' }}
106
+ {%- endif %}
107
+ {{- '\n<tool_response>\n' }}
108
+ {%- for tr in m.content -%}
109
+ {%- if tr is mapping and tr.type is defined -%}
110
+ {%- set t = tr.type | lower -%}
111
+ {%- if t == 'text' and tr.text is defined -%}
112
+ {{ tr.text }}
113
+ {%- elif t in ['image', 'image_url'] -%}
114
+ <|begin_of_image|><|image|><|end_of_image|>
115
+ {%- elif t in ['video', 'video_url'] -%}
116
+ <|begin_of_video|><|video|><|end_of_video|>
117
+ {%- else -%}
118
+ {{ tr | tojson(ensure_ascii=False) }}
119
+ {%- endif -%}
120
+ {%- else -%}
121
+ {{ tr.output if tr.output is defined else tr }}
122
+ {%- endif -%}
123
+ {%- endfor -%}
124
+ {{- '\n</tool_response>' }}
125
+ {%- else -%}
126
+ <|observation|>{% for tr in m.content %}
127
+
128
+ <tool_response>
129
+ {{ tr.output if tr.output is defined else tr }}
130
+ </tool_response>{% endfor -%}
131
+ {% endif -%}
132
+ {%- elif m.role == 'system' -%}
133
+ <|system|>
134
+ {{ visible_text(m.content) }}
135
+ {%- endif -%}
136
+ {%- endfor -%}
137
+ {%- if add_generation_prompt -%}
138
+ <|assistant|>
139
+ {{'<think></think>\n' if (enable_thinking is defined and not enable_thinking) else ''}}
140
+ {%- endif -%}
config.json ADDED
@@ -0,0 +1,722 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Glm4vForConditionalGeneration"
4
+ ],
5
+ "model_type": "glm4v",
6
+ "text_config": {
7
+ "attention_bias": true,
8
+ "attention_dropout": 0.0,
9
+ "dtype": "bfloat16",
10
+ "eos_token_id": [
11
+ 151329,
12
+ 151336,
13
+ 151338
14
+ ],
15
+ "hidden_act": "silu",
16
+ "hidden_size": 4096,
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 13696,
19
+ "max_position_embeddings": 131072,
20
+ "model_type": "glm4v_text",
21
+ "num_attention_heads": 32,
22
+ "num_hidden_layers": 40,
23
+ "num_key_value_heads": 2,
24
+ "pad_token_id": 151329,
25
+ "rms_norm_eps": 1e-05,
26
+ "rope_parameters": {
27
+ "mrope_section": [
28
+ 8,
29
+ 12,
30
+ 12
31
+ ],
32
+ "partial_rotary_factor": 0.5,
33
+ "rope_theta": 500000,
34
+ "rope_type": "default"
35
+ },
36
+ "use_cache": true,
37
+ "vocab_size": 151552
38
+ },
39
+ "tie_word_embeddings": false,
40
+ "transformers_version": "5.0.0rc0",
41
+ "image_start_token_id": 151339,
42
+ "image_end_token_id": 151340,
43
+ "video_start_token_id": 151341,
44
+ "video_end_token_id": 151342,
45
+ "image_token_id": 151363,
46
+ "video_token_id": 151364,
47
+ "vision_config": {
48
+ "attention_bias": false,
49
+ "attention_dropout": 0.0,
50
+ "depth": 24,
51
+ "hidden_act": "silu",
52
+ "hidden_dropout_prob": 0.0,
53
+ "hidden_size": 1536,
54
+ "image_size": 336,
55
+ "in_channels": 3,
56
+ "initializer_range": 0.02,
57
+ "intermediate_size": 13696,
58
+ "model_type": "glm4v_vision",
59
+ "num_heads": 12,
60
+ "out_hidden_size": 4096,
61
+ "patch_size": 14,
62
+ "rms_norm_eps": 1e-05,
63
+ "spatial_merge_size": 2,
64
+ "temporal_patch_size": 2
65
+ },
66
+ "eos_token_id": [
67
+ 151329,
68
+ 151336,
69
+ 151338,
70
+ 151348
71
+ ],
72
+ "quantization": {
73
+ "group_size": 64,
74
+ "bits": 4,
75
+ "mode": "affine",
76
+ "language_model.model.layers.9.self_attn.o_proj": {
77
+ "bits": 5,
78
+ "group_size": 64,
79
+ "mode": "affine"
80
+ },
81
+ "language_model.model.layers.8.self_attn.o_proj": {
82
+ "bits": 5,
83
+ "group_size": 64,
84
+ "mode": "affine"
85
+ },
86
+ "language_model.model.layers.8.mlp.down_proj": {
87
+ "bits": 5,
88
+ "group_size": 64,
89
+ "mode": "affine"
90
+ },
91
+ "language_model.model.layers.9.mlp.down_proj": {
92
+ "bits": 5,
93
+ "group_size": 64,
94
+ "mode": "affine"
95
+ },
96
+ "language_model.model.layers.7.self_attn.o_proj": {
97
+ "bits": 5,
98
+ "group_size": 64,
99
+ "mode": "affine"
100
+ },
101
+ "language_model.model.layers.6.mlp.down_proj": {
102
+ "bits": 5,
103
+ "group_size": 64,
104
+ "mode": "affine"
105
+ },
106
+ "language_model.model.layers.0.self_attn.v_proj": {
107
+ "bits": 6,
108
+ "group_size": 64,
109
+ "mode": "affine"
110
+ },
111
+ "language_model.model.layers.5.self_attn.o_proj": {
112
+ "bits": 5,
113
+ "group_size": 64,
114
+ "mode": "affine"
115
+ },
116
+ "language_model.model.layers.4.self_attn.q_proj": {
117
+ "bits": 5,
118
+ "group_size": 64,
119
+ "mode": "affine"
120
+ },
121
+ "language_model.model.layers.3.self_attn.v_proj": {
122
+ "bits": 6,
123
+ "group_size": 64,
124
+ "mode": "affine"
125
+ },
126
+ "language_model.model.layers.2.self_attn.v_proj": {
127
+ "bits": 6,
128
+ "group_size": 64,
129
+ "mode": "affine"
130
+ },
131
+ "language_model.model.layers.4.self_attn.k_proj": {
132
+ "bits": 5,
133
+ "group_size": 64,
134
+ "mode": "affine"
135
+ },
136
+ "language_model.model.layers.4.mlp.down_proj": {
137
+ "bits": 6,
138
+ "group_size": 64,
139
+ "mode": "affine"
140
+ },
141
+ "language_model.model.layers.0.self_attn.k_proj": {
142
+ "bits": 5,
143
+ "group_size": 64,
144
+ "mode": "affine"
145
+ },
146
+ "language_model.model.layers.3.self_attn.o_proj": {
147
+ "bits": 5,
148
+ "group_size": 64,
149
+ "mode": "affine"
150
+ },
151
+ "language_model.model.layers.7.mlp.down_proj": {
152
+ "bits": 5,
153
+ "group_size": 64,
154
+ "mode": "affine"
155
+ },
156
+ "language_model.model.layers.4.self_attn.v_proj": {
157
+ "bits": 6,
158
+ "group_size": 64,
159
+ "mode": "affine"
160
+ },
161
+ "language_model.model.layers.3.self_attn.q_proj": {
162
+ "bits": 5,
163
+ "group_size": 64,
164
+ "mode": "affine"
165
+ },
166
+ "language_model.model.layers.6.self_attn.o_proj": {
167
+ "bits": 5,
168
+ "group_size": 64,
169
+ "mode": "affine"
170
+ },
171
+ "language_model.model.layers.2.self_attn.o_proj": {
172
+ "bits": 5,
173
+ "group_size": 64,
174
+ "mode": "affine"
175
+ },
176
+ "language_model.model.layers.1.self_attn.v_proj": {
177
+ "bits": 6,
178
+ "group_size": 64,
179
+ "mode": "affine"
180
+ },
181
+ "language_model.model.layers.1.self_attn.q_proj": {
182
+ "bits": 5,
183
+ "group_size": 64,
184
+ "mode": "affine"
185
+ },
186
+ "language_model.model.layers.0.self_attn.q_proj": {
187
+ "bits": 5,
188
+ "group_size": 64,
189
+ "mode": "affine"
190
+ },
191
+ "language_model.model.layers.2.self_attn.q_proj": {
192
+ "bits": 5,
193
+ "group_size": 64,
194
+ "mode": "affine"
195
+ },
196
+ "language_model.model.layers.1.self_attn.k_proj": {
197
+ "bits": 5,
198
+ "group_size": 64,
199
+ "mode": "affine"
200
+ },
201
+ "language_model.model.layers.3.self_attn.k_proj": {
202
+ "bits": 5,
203
+ "group_size": 64,
204
+ "mode": "affine"
205
+ },
206
+ "language_model.model.layers.0.mlp.down_proj": {
207
+ "bits": 6,
208
+ "group_size": 64,
209
+ "mode": "affine"
210
+ },
211
+ "language_model.model.layers.5.mlp.down_proj": {
212
+ "bits": 5,
213
+ "group_size": 64,
214
+ "mode": "affine"
215
+ },
216
+ "language_model.model.layers.0.self_attn.o_proj": {
217
+ "bits": 5,
218
+ "group_size": 64,
219
+ "mode": "affine"
220
+ },
221
+ "language_model.model.layers.2.mlp.down_proj": {
222
+ "bits": 6,
223
+ "group_size": 64,
224
+ "mode": "affine"
225
+ },
226
+ "language_model.model.layers.1.mlp.down_proj": {
227
+ "bits": 6,
228
+ "group_size": 64,
229
+ "mode": "affine"
230
+ },
231
+ "language_model.model.layers.3.mlp.down_proj": {
232
+ "bits": 6,
233
+ "group_size": 64,
234
+ "mode": "affine"
235
+ },
236
+ "language_model.model.layers.2.self_attn.k_proj": {
237
+ "bits": 5,
238
+ "group_size": 64,
239
+ "mode": "affine"
240
+ },
241
+ "language_model.model.layers.4.self_attn.o_proj": {
242
+ "bits": 5,
243
+ "group_size": 64,
244
+ "mode": "affine"
245
+ },
246
+ "language_model.model.layers.1.self_attn.o_proj": {
247
+ "bits": 5,
248
+ "group_size": 64,
249
+ "mode": "affine"
250
+ },
251
+ "language_model.model.layers.21.self_attn.o_proj": {
252
+ "bits": 5,
253
+ "group_size": 64,
254
+ "mode": "affine"
255
+ },
256
+ "language_model.model.layers.22.mlp.down_proj": {
257
+ "bits": 5,
258
+ "group_size": 64,
259
+ "mode": "affine"
260
+ },
261
+ "language_model.model.layers.21.mlp.down_proj": {
262
+ "bits": 5,
263
+ "group_size": 64,
264
+ "mode": "affine"
265
+ },
266
+ "language_model.model.layers.20.mlp.down_proj": {
267
+ "bits": 5,
268
+ "group_size": 64,
269
+ "mode": "affine"
270
+ },
271
+ "language_model.model.layers.19.mlp.down_proj": {
272
+ "bits": 5,
273
+ "group_size": 64,
274
+ "mode": "affine"
275
+ },
276
+ "language_model.model.layers.18.self_attn.o_proj": {
277
+ "bits": 5,
278
+ "group_size": 64,
279
+ "mode": "affine"
280
+ },
281
+ "language_model.model.layers.18.mlp.down_proj": {
282
+ "bits": 5,
283
+ "group_size": 64,
284
+ "mode": "affine"
285
+ },
286
+ "language_model.model.layers.17.self_attn.o_proj": {
287
+ "bits": 5,
288
+ "group_size": 64,
289
+ "mode": "affine"
290
+ },
291
+ "language_model.model.layers.17.mlp.down_proj": {
292
+ "bits": 5,
293
+ "group_size": 64,
294
+ "mode": "affine"
295
+ },
296
+ "language_model.model.layers.11.self_attn.o_proj": {
297
+ "bits": 5,
298
+ "group_size": 64,
299
+ "mode": "affine"
300
+ },
301
+ "language_model.model.layers.10.mlp.down_proj": {
302
+ "bits": 5,
303
+ "group_size": 64,
304
+ "mode": "affine"
305
+ },
306
+ "language_model.model.layers.16.mlp.down_proj": {
307
+ "bits": 5,
308
+ "group_size": 64,
309
+ "mode": "affine"
310
+ },
311
+ "language_model.model.layers.15.mlp.down_proj": {
312
+ "bits": 5,
313
+ "group_size": 64,
314
+ "mode": "affine"
315
+ },
316
+ "language_model.model.layers.14.mlp.down_proj": {
317
+ "bits": 5,
318
+ "group_size": 64,
319
+ "mode": "affine"
320
+ },
321
+ "language_model.model.layers.14.self_attn.o_proj": {
322
+ "bits": 5,
323
+ "group_size": 64,
324
+ "mode": "affine"
325
+ },
326
+ "language_model.model.layers.10.self_attn.o_proj": {
327
+ "bits": 5,
328
+ "group_size": 64,
329
+ "mode": "affine"
330
+ },
331
+ "language_model.model.layers.13.self_attn.o_proj": {
332
+ "bits": 5,
333
+ "group_size": 64,
334
+ "mode": "affine"
335
+ },
336
+ "language_model.model.layers.20.self_attn.o_proj": {
337
+ "bits": 5,
338
+ "group_size": 64,
339
+ "mode": "affine"
340
+ },
341
+ "language_model.model.layers.19.self_attn.o_proj": {
342
+ "bits": 5,
343
+ "group_size": 64,
344
+ "mode": "affine"
345
+ },
346
+ "language_model.model.layers.13.mlp.down_proj": {
347
+ "bits": 5,
348
+ "group_size": 64,
349
+ "mode": "affine"
350
+ },
351
+ "language_model.model.layers.11.mlp.down_proj": {
352
+ "bits": 5,
353
+ "group_size": 64,
354
+ "mode": "affine"
355
+ },
356
+ "language_model.model.layers.12.self_attn.o_proj": {
357
+ "bits": 5,
358
+ "group_size": 64,
359
+ "mode": "affine"
360
+ },
361
+ "language_model.model.layers.16.self_attn.o_proj": {
362
+ "bits": 5,
363
+ "group_size": 64,
364
+ "mode": "affine"
365
+ },
366
+ "language_model.model.layers.15.self_attn.o_proj": {
367
+ "bits": 5,
368
+ "group_size": 64,
369
+ "mode": "affine"
370
+ },
371
+ "language_model.model.layers.22.self_attn.o_proj": {
372
+ "bits": 5,
373
+ "group_size": 64,
374
+ "mode": "affine"
375
+ },
376
+ "language_model.model.layers.35.self_attn.v_proj": {
377
+ "bits": 6,
378
+ "group_size": 64,
379
+ "mode": "affine"
380
+ },
381
+ "language_model.model.layers.35.self_attn.k_proj": {
382
+ "bits": 5,
383
+ "group_size": 64,
384
+ "mode": "affine"
385
+ },
386
+ "language_model.model.layers.34.self_attn.o_proj": {
387
+ "bits": 5,
388
+ "group_size": 64,
389
+ "mode": "affine"
390
+ },
391
+ "language_model.model.layers.33.self_attn.o_proj": {
392
+ "bits": 5,
393
+ "group_size": 64,
394
+ "mode": "affine"
395
+ }
396
+ },
397
+ "quantization_config": {
398
+ "group_size": 64,
399
+ "bits": 4,
400
+ "mode": "affine",
401
+ "language_model.model.layers.9.self_attn.o_proj": {
402
+ "bits": 5,
403
+ "group_size": 64,
404
+ "mode": "affine"
405
+ },
406
+ "language_model.model.layers.8.self_attn.o_proj": {
407
+ "bits": 5,
408
+ "group_size": 64,
409
+ "mode": "affine"
410
+ },
411
+ "language_model.model.layers.8.mlp.down_proj": {
412
+ "bits": 5,
413
+ "group_size": 64,
414
+ "mode": "affine"
415
+ },
416
+ "language_model.model.layers.9.mlp.down_proj": {
417
+ "bits": 5,
418
+ "group_size": 64,
419
+ "mode": "affine"
420
+ },
421
+ "language_model.model.layers.7.self_attn.o_proj": {
422
+ "bits": 5,
423
+ "group_size": 64,
424
+ "mode": "affine"
425
+ },
426
+ "language_model.model.layers.6.mlp.down_proj": {
427
+ "bits": 5,
428
+ "group_size": 64,
429
+ "mode": "affine"
430
+ },
431
+ "language_model.model.layers.0.self_attn.v_proj": {
432
+ "bits": 6,
433
+ "group_size": 64,
434
+ "mode": "affine"
435
+ },
436
+ "language_model.model.layers.5.self_attn.o_proj": {
437
+ "bits": 5,
438
+ "group_size": 64,
439
+ "mode": "affine"
440
+ },
441
+ "language_model.model.layers.4.self_attn.q_proj": {
442
+ "bits": 5,
443
+ "group_size": 64,
444
+ "mode": "affine"
445
+ },
446
+ "language_model.model.layers.3.self_attn.v_proj": {
447
+ "bits": 6,
448
+ "group_size": 64,
449
+ "mode": "affine"
450
+ },
451
+ "language_model.model.layers.2.self_attn.v_proj": {
452
+ "bits": 6,
453
+ "group_size": 64,
454
+ "mode": "affine"
455
+ },
456
+ "language_model.model.layers.4.self_attn.k_proj": {
457
+ "bits": 5,
458
+ "group_size": 64,
459
+ "mode": "affine"
460
+ },
461
+ "language_model.model.layers.4.mlp.down_proj": {
462
+ "bits": 6,
463
+ "group_size": 64,
464
+ "mode": "affine"
465
+ },
466
+ "language_model.model.layers.0.self_attn.k_proj": {
467
+ "bits": 5,
468
+ "group_size": 64,
469
+ "mode": "affine"
470
+ },
471
+ "language_model.model.layers.3.self_attn.o_proj": {
472
+ "bits": 5,
473
+ "group_size": 64,
474
+ "mode": "affine"
475
+ },
476
+ "language_model.model.layers.7.mlp.down_proj": {
477
+ "bits": 5,
478
+ "group_size": 64,
479
+ "mode": "affine"
480
+ },
481
+ "language_model.model.layers.4.self_attn.v_proj": {
482
+ "bits": 6,
483
+ "group_size": 64,
484
+ "mode": "affine"
485
+ },
486
+ "language_model.model.layers.3.self_attn.q_proj": {
487
+ "bits": 5,
488
+ "group_size": 64,
489
+ "mode": "affine"
490
+ },
491
+ "language_model.model.layers.6.self_attn.o_proj": {
492
+ "bits": 5,
493
+ "group_size": 64,
494
+ "mode": "affine"
495
+ },
496
+ "language_model.model.layers.2.self_attn.o_proj": {
497
+ "bits": 5,
498
+ "group_size": 64,
499
+ "mode": "affine"
500
+ },
501
+ "language_model.model.layers.1.self_attn.v_proj": {
502
+ "bits": 6,
503
+ "group_size": 64,
504
+ "mode": "affine"
505
+ },
506
+ "language_model.model.layers.1.self_attn.q_proj": {
507
+ "bits": 5,
508
+ "group_size": 64,
509
+ "mode": "affine"
510
+ },
511
+ "language_model.model.layers.0.self_attn.q_proj": {
512
+ "bits": 5,
513
+ "group_size": 64,
514
+ "mode": "affine"
515
+ },
516
+ "language_model.model.layers.2.self_attn.q_proj": {
517
+ "bits": 5,
518
+ "group_size": 64,
519
+ "mode": "affine"
520
+ },
521
+ "language_model.model.layers.1.self_attn.k_proj": {
522
+ "bits": 5,
523
+ "group_size": 64,
524
+ "mode": "affine"
525
+ },
526
+ "language_model.model.layers.3.self_attn.k_proj": {
527
+ "bits": 5,
528
+ "group_size": 64,
529
+ "mode": "affine"
530
+ },
531
+ "language_model.model.layers.0.mlp.down_proj": {
532
+ "bits": 6,
533
+ "group_size": 64,
534
+ "mode": "affine"
535
+ },
536
+ "language_model.model.layers.5.mlp.down_proj": {
537
+ "bits": 5,
538
+ "group_size": 64,
539
+ "mode": "affine"
540
+ },
541
+ "language_model.model.layers.0.self_attn.o_proj": {
542
+ "bits": 5,
543
+ "group_size": 64,
544
+ "mode": "affine"
545
+ },
546
+ "language_model.model.layers.2.mlp.down_proj": {
547
+ "bits": 6,
548
+ "group_size": 64,
549
+ "mode": "affine"
550
+ },
551
+ "language_model.model.layers.1.mlp.down_proj": {
552
+ "bits": 6,
553
+ "group_size": 64,
554
+ "mode": "affine"
555
+ },
556
+ "language_model.model.layers.3.mlp.down_proj": {
557
+ "bits": 6,
558
+ "group_size": 64,
559
+ "mode": "affine"
560
+ },
561
+ "language_model.model.layers.2.self_attn.k_proj": {
562
+ "bits": 5,
563
+ "group_size": 64,
564
+ "mode": "affine"
565
+ },
566
+ "language_model.model.layers.4.self_attn.o_proj": {
567
+ "bits": 5,
568
+ "group_size": 64,
569
+ "mode": "affine"
570
+ },
571
+ "language_model.model.layers.1.self_attn.o_proj": {
572
+ "bits": 5,
573
+ "group_size": 64,
574
+ "mode": "affine"
575
+ },
576
+ "language_model.model.layers.21.self_attn.o_proj": {
577
+ "bits": 5,
578
+ "group_size": 64,
579
+ "mode": "affine"
580
+ },
581
+ "language_model.model.layers.22.mlp.down_proj": {
582
+ "bits": 5,
583
+ "group_size": 64,
584
+ "mode": "affine"
585
+ },
586
+ "language_model.model.layers.21.mlp.down_proj": {
587
+ "bits": 5,
588
+ "group_size": 64,
589
+ "mode": "affine"
590
+ },
591
+ "language_model.model.layers.20.mlp.down_proj": {
592
+ "bits": 5,
593
+ "group_size": 64,
594
+ "mode": "affine"
595
+ },
596
+ "language_model.model.layers.19.mlp.down_proj": {
597
+ "bits": 5,
598
+ "group_size": 64,
599
+ "mode": "affine"
600
+ },
601
+ "language_model.model.layers.18.self_attn.o_proj": {
602
+ "bits": 5,
603
+ "group_size": 64,
604
+ "mode": "affine"
605
+ },
606
+ "language_model.model.layers.18.mlp.down_proj": {
607
+ "bits": 5,
608
+ "group_size": 64,
609
+ "mode": "affine"
610
+ },
611
+ "language_model.model.layers.17.self_attn.o_proj": {
612
+ "bits": 5,
613
+ "group_size": 64,
614
+ "mode": "affine"
615
+ },
616
+ "language_model.model.layers.17.mlp.down_proj": {
617
+ "bits": 5,
618
+ "group_size": 64,
619
+ "mode": "affine"
620
+ },
621
+ "language_model.model.layers.11.self_attn.o_proj": {
622
+ "bits": 5,
623
+ "group_size": 64,
624
+ "mode": "affine"
625
+ },
626
+ "language_model.model.layers.10.mlp.down_proj": {
627
+ "bits": 5,
628
+ "group_size": 64,
629
+ "mode": "affine"
630
+ },
631
+ "language_model.model.layers.16.mlp.down_proj": {
632
+ "bits": 5,
633
+ "group_size": 64,
634
+ "mode": "affine"
635
+ },
636
+ "language_model.model.layers.15.mlp.down_proj": {
637
+ "bits": 5,
638
+ "group_size": 64,
639
+ "mode": "affine"
640
+ },
641
+ "language_model.model.layers.14.mlp.down_proj": {
642
+ "bits": 5,
643
+ "group_size": 64,
644
+ "mode": "affine"
645
+ },
646
+ "language_model.model.layers.14.self_attn.o_proj": {
647
+ "bits": 5,
648
+ "group_size": 64,
649
+ "mode": "affine"
650
+ },
651
+ "language_model.model.layers.10.self_attn.o_proj": {
652
+ "bits": 5,
653
+ "group_size": 64,
654
+ "mode": "affine"
655
+ },
656
+ "language_model.model.layers.13.self_attn.o_proj": {
657
+ "bits": 5,
658
+ "group_size": 64,
659
+ "mode": "affine"
660
+ },
661
+ "language_model.model.layers.20.self_attn.o_proj": {
662
+ "bits": 5,
663
+ "group_size": 64,
664
+ "mode": "affine"
665
+ },
666
+ "language_model.model.layers.19.self_attn.o_proj": {
667
+ "bits": 5,
668
+ "group_size": 64,
669
+ "mode": "affine"
670
+ },
671
+ "language_model.model.layers.13.mlp.down_proj": {
672
+ "bits": 5,
673
+ "group_size": 64,
674
+ "mode": "affine"
675
+ },
676
+ "language_model.model.layers.11.mlp.down_proj": {
677
+ "bits": 5,
678
+ "group_size": 64,
679
+ "mode": "affine"
680
+ },
681
+ "language_model.model.layers.12.self_attn.o_proj": {
682
+ "bits": 5,
683
+ "group_size": 64,
684
+ "mode": "affine"
685
+ },
686
+ "language_model.model.layers.16.self_attn.o_proj": {
687
+ "bits": 5,
688
+ "group_size": 64,
689
+ "mode": "affine"
690
+ },
691
+ "language_model.model.layers.15.self_attn.o_proj": {
692
+ "bits": 5,
693
+ "group_size": 64,
694
+ "mode": "affine"
695
+ },
696
+ "language_model.model.layers.22.self_attn.o_proj": {
697
+ "bits": 5,
698
+ "group_size": 64,
699
+ "mode": "affine"
700
+ },
701
+ "language_model.model.layers.35.self_attn.v_proj": {
702
+ "bits": 6,
703
+ "group_size": 64,
704
+ "mode": "affine"
705
+ },
706
+ "language_model.model.layers.35.self_attn.k_proj": {
707
+ "bits": 5,
708
+ "group_size": 64,
709
+ "mode": "affine"
710
+ },
711
+ "language_model.model.layers.34.self_attn.o_proj": {
712
+ "bits": 5,
713
+ "group_size": 64,
714
+ "mode": "affine"
715
+ },
716
+ "language_model.model.layers.33.self_attn.o_proj": {
717
+ "bits": 5,
718
+ "group_size": 64,
719
+ "mode": "affine"
720
+ }
721
+ }
722
+ }
generation_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151329,
6
+ 151336,
7
+ 151338,
8
+ 151348
9
+ ],
10
+ "pad_token_id": 151329,
11
+ "top_p": 0.6,
12
+ "temperature": 0.8,
13
+ "top_k": 2,
14
+ "transformers_version": "4.57.1"
15
+ }
model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a706f9e53bd9f36572687872ce4a069ee81df1db6e651aea8389c254e63b5be
3
+ size 5105533104
model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40396c4c41cd44e0f094ba4cc5eea45607fd9fe7e4f15a3244981a1bd607bd71
3
+ size 2223092170
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
preprocessor_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "size": {"shortest_edge": 12544, "longest_edge": 9633792},
3
+ "do_rescale": true,
4
+ "patch_size": 14,
5
+ "temporal_patch_size": 2,
6
+ "merge_size": 2,
7
+ "image_mean": [0.48145466, 0.4578275, 0.40821073],
8
+ "image_std": [0.26862954, 0.26130258, 0.27577711],
9
+ "image_processor_type": "Glm46VImageProcessor",
10
+ "processor_class": "Glm46VProcessor"
11
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9340665016419c825c4bdabbcc9acc43b7ca2c68ce142724afa829abb1be5efd
3
+ size 19970699
tokenizer_config.json ADDED
@@ -0,0 +1,327 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "151329": {
4
+ "content": "<|endoftext|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "151330": {
12
+ "content": "[MASK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "151331": {
20
+ "content": "[gMASK]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "151332": {
28
+ "content": "[sMASK]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "151333": {
36
+ "content": "<sop>",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "151334": {
44
+ "content": "<eop>",
45
+ "lstrip": false,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "151335": {
52
+ "content": "<|system|>",
53
+ "lstrip": false,
54
+ "normalized": false,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ },
59
+ "151336": {
60
+ "content": "<|user|>",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": true
66
+ },
67
+ "151337": {
68
+ "content": "<|assistant|>",
69
+ "lstrip": false,
70
+ "normalized": false,
71
+ "rstrip": false,
72
+ "single_word": false,
73
+ "special": true
74
+ },
75
+ "151338": {
76
+ "content": "<|observation|>",
77
+ "lstrip": false,
78
+ "normalized": false,
79
+ "rstrip": false,
80
+ "single_word": false,
81
+ "special": true
82
+ },
83
+ "151339": {
84
+ "content": "<|begin_of_image|>",
85
+ "lstrip": false,
86
+ "normalized": false,
87
+ "rstrip": false,
88
+ "single_word": false,
89
+ "special": true
90
+ },
91
+ "151340": {
92
+ "content": "<|end_of_image|>",
93
+ "lstrip": false,
94
+ "normalized": false,
95
+ "rstrip": false,
96
+ "single_word": false,
97
+ "special": true
98
+ },
99
+ "151341": {
100
+ "content": "<|begin_of_video|>",
101
+ "lstrip": false,
102
+ "normalized": false,
103
+ "rstrip": false,
104
+ "single_word": false,
105
+ "special": true
106
+ },
107
+ "151342": {
108
+ "content": "<|end_of_video|>",
109
+ "lstrip": false,
110
+ "normalized": false,
111
+ "rstrip": false,
112
+ "single_word": false,
113
+ "special": true
114
+ },
115
+ "151343": {
116
+ "content": "<|begin_of_audio|>",
117
+ "lstrip": false,
118
+ "normalized": false,
119
+ "rstrip": false,
120
+ "single_word": false,
121
+ "special": true
122
+ },
123
+ "151344": {
124
+ "content": "<|end_of_audio|>",
125
+ "lstrip": false,
126
+ "normalized": false,
127
+ "rstrip": false,
128
+ "single_word": false,
129
+ "special": true
130
+ },
131
+ "151345": {
132
+ "content": "<|begin_of_transcription|>",
133
+ "lstrip": false,
134
+ "normalized": false,
135
+ "rstrip": false,
136
+ "single_word": false,
137
+ "special": true
138
+ },
139
+ "151346": {
140
+ "content": "<|end_of_transcription|>",
141
+ "lstrip": false,
142
+ "normalized": false,
143
+ "rstrip": false,
144
+ "single_word": false,
145
+ "special": true
146
+ },
147
+ "151347": {
148
+ "content": "<|code_prefix|>",
149
+ "lstrip": false,
150
+ "normalized": false,
151
+ "rstrip": false,
152
+ "single_word": false,
153
+ "special": true
154
+ },
155
+ "151348": {
156
+ "content": "<|code_middle|>",
157
+ "lstrip": false,
158
+ "normalized": false,
159
+ "rstrip": false,
160
+ "single_word": false,
161
+ "special": true
162
+ },
163
+ "151349": {
164
+ "content": "<|code_suffix|>",
165
+ "lstrip": false,
166
+ "normalized": false,
167
+ "rstrip": false,
168
+ "single_word": false,
169
+ "special": true
170
+ },
171
+ "151350": {
172
+ "content": "<think>",
173
+ "lstrip": false,
174
+ "normalized": false,
175
+ "rstrip": false,
176
+ "single_word": false,
177
+ "special": false
178
+ },
179
+ "151351": {
180
+ "content": "</think>",
181
+ "lstrip": false,
182
+ "normalized": false,
183
+ "rstrip": false,
184
+ "single_word": false,
185
+ "special": false
186
+ },
187
+ "151352": {
188
+ "content": "<tool_call>",
189
+ "lstrip": false,
190
+ "normalized": false,
191
+ "rstrip": false,
192
+ "single_word": false,
193
+ "special": false
194
+ },
195
+ "151353": {
196
+ "content": "</tool_call>",
197
+ "lstrip": false,
198
+ "normalized": false,
199
+ "rstrip": false,
200
+ "single_word": false,
201
+ "special": false
202
+ },
203
+ "151354": {
204
+ "content": "<tool_response>",
205
+ "lstrip": false,
206
+ "normalized": false,
207
+ "rstrip": false,
208
+ "single_word": false,
209
+ "special": false
210
+ },
211
+ "151355": {
212
+ "content": "</tool_response>",
213
+ "lstrip": false,
214
+ "normalized": false,
215
+ "rstrip": false,
216
+ "single_word": false,
217
+ "special": false
218
+ },
219
+ "151356": {
220
+ "content": "<arg_key>",
221
+ "lstrip": false,
222
+ "normalized": false,
223
+ "rstrip": false,
224
+ "single_word": false,
225
+ "special": false
226
+ },
227
+ "151357": {
228
+ "content": "</arg_key>",
229
+ "lstrip": false,
230
+ "normalized": false,
231
+ "rstrip": false,
232
+ "single_word": false,
233
+ "special": false
234
+ },
235
+ "151358": {
236
+ "content": "<arg_value>",
237
+ "lstrip": false,
238
+ "normalized": false,
239
+ "rstrip": false,
240
+ "single_word": false,
241
+ "special": false
242
+ },
243
+ "151359": {
244
+ "content": "</arg_value>",
245
+ "lstrip": false,
246
+ "normalized": false,
247
+ "rstrip": false,
248
+ "single_word": false,
249
+ "special": false
250
+ },
251
+ "151360": {
252
+ "content": "/nothink",
253
+ "lstrip": false,
254
+ "normalized": false,
255
+ "rstrip": false,
256
+ "single_word": false,
257
+ "special": true
258
+ },
259
+ "151361": {
260
+ "content": "<|begin_of_box|>",
261
+ "lstrip": false,
262
+ "normalized": false,
263
+ "rstrip": false,
264
+ "single_word": false,
265
+ "special": false
266
+ },
267
+ "151362": {
268
+ "content": "<|end_of_box|>",
269
+ "lstrip": false,
270
+ "normalized": false,
271
+ "rstrip": false,
272
+ "single_word": false,
273
+ "special": false
274
+ },
275
+ "151363": {
276
+ "content": "<|image|>",
277
+ "lstrip": false,
278
+ "normalized": false,
279
+ "rstrip": false,
280
+ "single_word": false,
281
+ "special": false
282
+ },
283
+ "151364": {
284
+ "content": "<|video|>",
285
+ "lstrip": false,
286
+ "normalized": false,
287
+ "rstrip": false,
288
+ "single_word": false,
289
+ "special": false
290
+ }
291
+ },
292
+ "additional_special_tokens": [
293
+ "<|endoftext|>",
294
+ "[MASK]",
295
+ "[gMASK]",
296
+ "[sMASK]",
297
+ "<sop>",
298
+ "<eop>",
299
+ "<|system|>",
300
+ "<|user|>",
301
+ "<|assistant|>",
302
+ "<|observation|>",
303
+ "<|begin_of_image|>",
304
+ "<|end_of_image|>",
305
+ "<|begin_of_video|>",
306
+ "<|end_of_video|>",
307
+ "<|begin_of_audio|>",
308
+ "<|end_of_audio|>",
309
+ "<|image|>",
310
+ "<|video|>",
311
+ "<|begin_of_transcription|>",
312
+ "<|end_of_transcription|>",
313
+ "<|code_prefix|>",
314
+ "<|code_middle|>",
315
+ "<|code_suffix|>",
316
+ "/nothink"
317
+ ],
318
+ "clean_up_tokenization_spaces": false,
319
+ "do_lower_case": false,
320
+ "eos_token": "<|endoftext|>",
321
+ "extra_special_tokens": {},
322
+ "model_max_length": 128000,
323
+ "pad_token": "<|endoftext|>",
324
+ "padding_side": "left",
325
+ "remove_space": false,
326
+ "tokenizer_class": "PreTrainedTokenizer"
327
+ }