fael-b-pix commited on
Commit
79da680
·
verified ·
1 Parent(s): e51b4ab

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tekken.json filter=lfs diff=lfs merge=lfs -text
37
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,353 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - fp4
4
+ - vllm
5
+ language:
6
+ - en
7
+ - de
8
+ - fr
9
+ - it
10
+ - pt
11
+ - hi
12
+ - es
13
+ - th
14
+ pipeline_tag: text-generation
15
+ license: apache-2.0
16
+ base_model: unsloth/Mistral-Small-3.2-24B-Instruct-2506
17
+ ---
18
+
19
+ # Mistral-Small-3.2-24B-Instruct-2506-NVFP4
20
+
21
+ ## Model Overview
22
+ - **Model Architecture:** unsloth/Mistral-Small-3.2-24B-Instruct-2506
23
+ - **Input:** Text
24
+ - **Output:** Text
25
+ - **Model Optimizations:**
26
+ - **Weight quantization:** FP4
27
+ - **Activation quantization:** FP4
28
+ - **Out-of-scope:** Use in any manner that violates applicable laws or regulations (including trade compliance laws). Use in languages other than English.
29
+ - **Release Date:** 10/29/2025
30
+ - **Version:** 1.0
31
+ - **Model Developers:** RedHatAI
32
+
33
+ This model is a quantized version of [unsloth/Mistral-Small-3.2-24B-Instruct-2506](https://huggingface.co/unsloth/Mistral-Small-3.2-24B-Instruct-2506).
34
+ It was evaluated on a several tasks to assess the its quality in comparison to the unquatized model.
35
+
36
+ ### Model Optimizations
37
+
38
+ This model was obtained by quantizing the weights and activations of [unsloth/Mistral-Small-3.2-24B-Instruct-2506](https://huggingface.co/unsloth/Mistral-Small-3.2-24B-Instruct-2506) to FP4 data type, ready for inference with vLLM>=0.9.1
39
+ This optimization reduces the number of bits per parameter from 16 to 4, reducing the disk size and GPU memory requirements by approximately 75%.
40
+
41
+ Only the weights and activations of the linear operators within transformers blocks are quantized using [LLM Compressor](https://github.com/vllm-project/llm-compressor).
42
+
43
+ ## Deployment
44
+
45
+ ### Use with vLLM
46
+
47
+ 1. Initialize vLLM server:
48
+ ```
49
+ vllm serve RedHatAI/Mistral-Small-3.2-24B-Instruct-2506-NVFP4 --tensor_parallel_size 1 --tokenizer_mode mistral
50
+ ```
51
+
52
+ 2. Send requests to the server:
53
+
54
+ ```python
55
+ from openai import OpenAI
56
+
57
+ # Modify OpenAI's API key and API base to use vLLM's API server.
58
+ openai_api_key = "EMPTY"
59
+ openai_api_base = "http://<your-server-host>:8000/v1"
60
+
61
+ client = OpenAI(
62
+ api_key=openai_api_key,
63
+ base_url=openai_api_base,
64
+ )
65
+
66
+ model = "RedHatAI/Mistral-Small-3.2-24B-Instruct-2506-NVFP4"
67
+
68
+
69
+ messages = [
70
+ {"role": "user", "content": "Explain quantum mechanics clearly and concisely."},
71
+ ]
72
+
73
+ outputs = client.chat.completions.create(
74
+ model=model,
75
+ messages=messages,
76
+ )
77
+
78
+ generated_text = outputs.choices[0].message.content
79
+ print(generated_text)
80
+ ```
81
+
82
+ ## Creation
83
+
84
+ This model was created by applying [LLM Compressor with calibration samples from UltraChat](https://github.com/vllm-project/llm-compressor/blob/main/examples/quantization_w4a4_fp4/llama3_example.py), as presented in the code snipet below.
85
+
86
+ <details>
87
+
88
+ ```python
89
+ from datasets import load_dataset
90
+ from transformers import AutoModelForCausalLM, AutoTokenizer
91
+
92
+ from llmcompressor import oneshot
93
+ from llmcompressor.modifiers.quantization import QuantizationModifier
94
+ from llmcompressor.modifiers.smoothquant import SmoothQuantModifier
95
+ from llmcompressor.utils import dispatch_for_generation
96
+
97
+ MODEL_ID = "unsloth/Mistral-Small-3.2-24B-Instruct-2506"
98
+
99
+ # Load model.
100
+ model = AutoModelForCausalLM.from_pretrained(MODEL_ID, torch_dtype="auto")
101
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
102
+
103
+ DATASET_ID = "HuggingFaceH4/ultrachat_200k"
104
+ DATASET_SPLIT = "train_sft"
105
+
106
+ # Select number of samples. 512 samples is a good place to start.
107
+ # Increasing the number of samples can improve accuracy.
108
+ NUM_CALIBRATION_SAMPLES = 512
109
+ MAX_SEQUENCE_LENGTH = 2048
110
+
111
+ # Load dataset and preprocess.
112
+ ds = load_dataset(DATASET_ID, split=f"{DATASET_SPLIT}[:{NUM_CALIBRATION_SAMPLES}]")
113
+ ds = ds.shuffle(seed=42)
114
+
115
+ def preprocess(example):
116
+ return {
117
+ "text": tokenizer.apply_chat_template(
118
+ example["messages"],
119
+ tokenize=False,
120
+ )
121
+ }
122
+
123
+ ds = ds.map(preprocess)
124
+
125
+ # Tokenize inputs.
126
+ def tokenize(sample):
127
+ return tokenizer(
128
+ sample["text"],
129
+ padding=False,
130
+ max_length=MAX_SEQUENCE_LENGTH,
131
+ truncation=True,
132
+ add_special_tokens=False,
133
+ )
134
+
135
+ ds = ds.map(tokenize, remove_columns=ds.column_names)
136
+
137
+ # Configure the quantization algorithm and scheme.
138
+ # In this case, we:
139
+ # * quantize the weights to fp4 with per group 16 via ptq
140
+ # * calibrate a global_scale for activations, which will be used to
141
+ # quantize activations to fp4 on the fly
142
+ smoothing_strength = 0.9
143
+ recipe = [
144
+ SmoothQuantModifier(smoothing_strength=smoothing_strength),
145
+ QuantizationModifier(
146
+ ignore=["re:.*lm_head.*"],
147
+ config_groups={
148
+ "group_0": {
149
+ "targets": ["Linear"],
150
+ "weights": {
151
+ "num_bits": 4,
152
+ "type": "float",
153
+ "strategy": "tensor_group",
154
+ "group_size": 16,
155
+ "symmetric": True,
156
+ "observer": "mse",
157
+ },
158
+ "input_activations": {
159
+ "num_bits": 4,
160
+ "type": "float",
161
+ "strategy": "tensor_group",
162
+ "group_size": 16,
163
+ "symmetric": True,
164
+ "dynamic": "local",
165
+ "observer": "minmax",
166
+ },
167
+ }
168
+ },
169
+ )
170
+ ]
171
+
172
+ # Save to disk in compressed-tensors format.
173
+ SAVE_DIR = MODEL_ID.rstrip("/").split("/")[-1] + "-NVFP4"
174
+
175
+ # Apply quantization.
176
+ oneshot(
177
+ model=model,
178
+ dataset=ds,
179
+ recipe=recipe,
180
+ max_seq_length=MAX_SEQUENCE_LENGTH,
181
+ num_calibration_samples=NUM_CALIBRATION_SAMPLES,
182
+ output_dir=SAVE_DIR,
183
+ )
184
+
185
+ print("\n\n")
186
+ print("========== SAMPLE GENERATION ==============")
187
+ dispatch_for_generation(model)
188
+ input_ids = tokenizer("Hello my name is", return_tensors="pt").input_ids.to("cuda")
189
+ output = model.generate(input_ids, max_new_tokens=100)
190
+ print(tokenizer.decode(output[0]))
191
+ print("==========================================\n\n")
192
+
193
+ model.save_pretrained(SAVE_DIR, save_compressed=True)
194
+ tokenizer.save_pretrained(SAVE_DIR)
195
+ ```
196
+ </details>
197
+
198
+ ## Evaluation
199
+
200
+ This model was evaluated on the well-known OpenLLM v1, OpenLLM v2 and HumanEval_64 benchmarks using [lm-evaluation-harness](https://github.com/neuralmagic/lm-evaluation-harness).
201
+
202
+ ### Accuracy
203
+ <table>
204
+ <thead>
205
+ <tr>
206
+ <th>Category</th>
207
+ <th>Metric</th>
208
+ <th>unsloth/Mistral-Small-3.2-24B-Instruct-2506</th>
209
+ <th>RedHatAI/Mistral-Small-3.2-24B-Instruct-2506-NVFP4</th>
210
+ <th>Recovery</th>
211
+ </tr>
212
+ </thead>
213
+ <tbody>
214
+ <!-- OpenLLM V1 -->
215
+ <tr>
216
+ <td rowspan="7"><b>OpenLLM V1</b></td>
217
+ <td>arc_challenge</td>
218
+ <td>68.52</td>
219
+ <td>66.98</td>
220
+ <td>97.75</td>
221
+ </tr>
222
+ <tr>
223
+ <td>gsm8k</td>
224
+ <td>89.61</td>
225
+ <td>87.11</td>
226
+ <td>97.21</td>
227
+ </tr>
228
+ <tr>
229
+ <td>hellaswag</td>
230
+ <td>85.70</td>
231
+ <td>85.11</td>
232
+ <td>99.31</td>
233
+ </tr>
234
+ <tr>
235
+ <td>mmlu</td>
236
+ <td>81.06</td>
237
+ <td>79.43</td>
238
+ <td>97.99</td>
239
+ </tr>
240
+ <tr>
241
+ <td>truthfulqa_mc2</td>
242
+ <td>61.35</td>
243
+ <td>60.34</td>
244
+ <td>98.35</td>
245
+ </tr>
246
+ <tr>
247
+ <td>winogrande</td>
248
+ <td>83.27</td>
249
+ <td>81.61</td>
250
+ <td>98.01</td>
251
+ </tr>
252
+ <tr>
253
+ <td><b>Average</b></td>
254
+ <td><b>78.25</b></td>
255
+ <td><b>76.76</b></td>
256
+ <td><b>98.10</b></td>
257
+ </tr>
258
+ <tr>
259
+ <td rowspan="7"><b>OpenLLM V2</b></td>
260
+ <td>BBH (3-shot)</td>
261
+ <td>65.86</td>
262
+ <td>64.05</td>
263
+ <td>97.25</td>
264
+ </tr>
265
+ <tr>
266
+ <td>MMLU-Pro (5-shot)</td>
267
+ <td>50.84</td>
268
+ <td>48.45</td>
269
+ <td>95.30</td>
270
+ </tr>
271
+ <tr>
272
+ <td>MuSR (0-shot)</td>
273
+ <td>39.15</td>
274
+ <td>40.21</td>
275
+ <td>102.71</td>
276
+ </tr>
277
+ <tr>
278
+ <td>IFEval (0-shot)</td>
279
+ <td>84.05</td>
280
+ <td>84.41</td>
281
+ <td>100.43</td>
282
+ </tr>
283
+ <tr>
284
+ <td>GPQA (0-shot)</td>
285
+ <td>33.14</td>
286
+ <td>32.55</td>
287
+ <td>98.22</td>
288
+ </tr>
289
+ <tr>
290
+ <td>Math-|v|-5 (4-shot)</td>
291
+ <td>41.69</td>
292
+ <td>37.76</td>
293
+ <td>90.57</td>
294
+ </tr>
295
+ <tr>
296
+ <td><b>Average</b></td>
297
+ <td><b>52.46</b></td>
298
+ <td><b>51.24</b></td>
299
+ <td><b>97.68</b></td>
300
+ </tr>
301
+ <tr>
302
+ <td rowspan="2"><b>Coding</b></td>
303
+ <td>HumanEval_64 pass@2</td>
304
+ <td>88.88</td>
305
+ <td>88.84</td>
306
+ <td>99.95</td>
307
+ </tr>
308
+ </tbody>
309
+ </table>
310
+
311
+
312
+
313
+
314
+ ### Reproduction
315
+
316
+ The results were obtained using the following commands:
317
+
318
+ <details>
319
+
320
+ ```
321
+ lm_eval \
322
+ --model vllm \
323
+ --model_args pretrained="RedHatAI/Mistral-Small-3.2-24B-Instruct-2506-NVFP4",dtype=auto,max_model_len=4096,tensor_parallel_size=2,enable_chunked_prefill=True,enforce_eager=True\
324
+ --apply_chat_template \
325
+ --fewshot_as_multiturn \
326
+ --tasks openllm \
327
+ --batch_size auto
328
+ ```
329
+
330
+
331
+ #### OpenLLM v2
332
+ ```
333
+ lm_eval \
334
+ --model vllm \
335
+ --model_args pretrained="RedHatAI/Mistral-Small-3.2-24B-Instruct-2506-NVFP4",dtype=auto,max_model_len=4096,tensor_parallel_size=2,enable_chunked_prefill=True,enforce_eager=True\
336
+ --apply_chat_template \
337
+ --fewshot_as_multiturn \
338
+ --tasks leaderboard \
339
+ --batch_size auto
340
+ ```
341
+
342
+ #### HumanEval_64
343
+ ```
344
+ lm_eval \
345
+ --model vllm \
346
+ --model_args pretrained="RedHatAI/Mistral-Small-3.2-24B-Instruct-2506-NVFP4",dtype=auto,max_model_len=4096,tensor_parallel_size=2,enable_chunked_prefill=True,enforce_eager=True\
347
+ --apply_chat_template \
348
+ --fewshot_as_multiturn \
349
+ --tasks humaneval_64_instruct \
350
+ --batch_size auto
351
+
352
+ ```
353
+ </details>
chat_template.jinja ADDED
@@ -0,0 +1,300 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {#- Unsloth template fixes #}
2
+ {%- set yesterday_day = strftime_now("%d") %}
3
+ {%- set yesterday_month = strftime_now("%m") %}
4
+ {%- set yesterday_year = strftime_now("%Y") %}
5
+ {%- set today_date = yesterday_year + '-' + yesterday_month + '-' + yesterday_day %}
6
+ {%- if yesterday_day == '01' %}
7
+ {#- Jinja doesnt allow minus 1 date - Unsloth alternative #}
8
+ {%- if yesterday_month == '01' %}
9
+ {%- set yesterday_day = '31' %}
10
+ {%- set yesterday_month = '12' %}
11
+ {%- if yesterday_year == '2024' %}
12
+ {%- set yesterday_year = '2023' %}
13
+ {%- elif yesterday_year == '2025' %}
14
+ {%- set yesterday_year = '2024' %}
15
+ {%- elif yesterday_year == '2026' %}
16
+ {%- set yesterday_year = '2025' %}
17
+ {%- elif yesterday_year == '2027' %}
18
+ {%- set yesterday_year = '2026' %}
19
+ {%- elif yesterday_year == '2028' %}
20
+ {%- set yesterday_year = '2027' %}
21
+ {%- elif yesterday_year == '2029' %}
22
+ {%- set yesterday_year = '2028' %}
23
+ {%- elif yesterday_year == '2030' %}
24
+ {%- set yesterday_year = '2029' %}
25
+ {%- elif yesterday_year == '2031' %}
26
+ {%- set yesterday_year = '2030' %}
27
+ {%- elif yesterday_year == '2032' %}
28
+ {%- set yesterday_year = '2031' %}
29
+ {%- elif yesterday_year == '1970' %}
30
+ {#- Stop llama_cpp from erroring out #}
31
+ {%- set yesterday_year = '1970' %}
32
+ {%- else %}
33
+ {{- raise_exception('Unsloth custom template does not support years > 2032. Error year = [' + yesterday_year + ']') }}
34
+ {%- endif %}
35
+ {%- elif yesterday_month == '02' %}
36
+ {%- set yesterday_day = '31' %}
37
+ {%- set yesterday_month = '01' %}
38
+ {%- elif yesterday_month == '03' %}
39
+ {%- set yesterday_month = '02' %}
40
+ {%- set yesterday_day = '28' %}
41
+ {%- if yesterday_year == '2024' %}
42
+ {%- set yesterday_day = '29' %}
43
+ {%- elif yesterday_year == '2028' %}
44
+ {%- set yesterday_day = '29' %}
45
+ {%- elif yesterday_year == '2032' %}
46
+ {%- set yesterday_day = '29' %}
47
+ {%- elif yesterday_year == '1970' %}
48
+ {#- Stop llama_cpp from erroring out #}
49
+ {%- set yesterday_day = '29' %}
50
+ {%- else %}
51
+ {{- raise_exception('Unsloth custom template does not support years > 2032. Error year = [' + yesterday_year + ']') }}
52
+ {%- endif %}
53
+ {%- elif yesterday_month == '04' %}
54
+ {%- set yesterday_day = '31' %}
55
+ {%- set yesterday_month = '03' %}
56
+ {%- elif yesterday_month == '05' %}
57
+ {%- set yesterday_day = '30' %}
58
+ {%- set yesterday_month = '04' %}
59
+ {%- elif yesterday_month == '06' %}
60
+ {%- set yesterday_day = '31' %}
61
+ {%- set yesterday_month = '05' %}
62
+ {%- elif yesterday_month == '07' %}
63
+ {%- set yesterday_day = '30' %}
64
+ {%- set yesterday_month = '06' %}
65
+ {%- elif yesterday_month == '08' %}
66
+ {%- set yesterday_day = '31' %}
67
+ {%- set yesterday_month = '07' %}
68
+ {%- elif yesterday_month == '09' %}
69
+ {%- set yesterday_day = '31' %}
70
+ {%- set yesterday_month = '08' %}
71
+ {%- elif yesterday_month == '10' %}
72
+ {%- set yesterday_day = '30' %}
73
+ {%- set yesterday_month = '09' %}
74
+ {%- elif yesterday_month == '11' %}
75
+ {%- set yesterday_day = '31' %}
76
+ {%- set yesterday_month = '10' %}
77
+ {%- elif yesterday_month == '12' %}
78
+ {%- set yesterday_day = '30' %}
79
+ {%- set yesterday_month = '11' %}
80
+ {%- endif %}
81
+ {%- elif yesterday_day == '02' %}
82
+ {%- set yesterday_day = '01' %}
83
+ {%- elif yesterday_day == '03' %}
84
+ {%- set yesterday_day = '02' %}
85
+ {%- elif yesterday_day == '04' %}
86
+ {%- set yesterday_day = '03' %}
87
+ {%- elif yesterday_day == '05' %}
88
+ {%- set yesterday_day = '04' %}
89
+ {%- elif yesterday_day == '06' %}
90
+ {%- set yesterday_day = '05' %}
91
+ {%- elif yesterday_day == '07' %}
92
+ {%- set yesterday_day = '06' %}
93
+ {%- elif yesterday_day == '08' %}
94
+ {%- set yesterday_day = '07' %}
95
+ {%- elif yesterday_day == '09' %}
96
+ {%- set yesterday_day = '08' %}
97
+ {%- elif yesterday_day == '10' %}
98
+ {%- set yesterday_day = '09' %}
99
+ {%- elif yesterday_day == '11' %}
100
+ {%- set yesterday_day = '10' %}
101
+ {%- elif yesterday_day == '12' %}
102
+ {%- set yesterday_day = '11' %}
103
+ {%- elif yesterday_day == '13' %}
104
+ {%- set yesterday_day = '12' %}
105
+ {%- elif yesterday_day == '14' %}
106
+ {%- set yesterday_day = '13' %}
107
+ {%- elif yesterday_day == '15' %}
108
+ {%- set yesterday_day = '14' %}
109
+ {%- elif yesterday_day == '16' %}
110
+ {%- set yesterday_day = '15' %}
111
+ {%- elif yesterday_day == '17' %}
112
+ {%- set yesterday_day = '16' %}
113
+ {%- elif yesterday_day == '18' %}
114
+ {%- set yesterday_day = '17' %}
115
+ {%- elif yesterday_day == '19' %}
116
+ {%- set yesterday_day = '18' %}
117
+ {%- elif yesterday_day == '20' %}
118
+ {%- set yesterday_day = '19' %}
119
+ {%- elif yesterday_day == '21' %}
120
+ {%- set yesterday_day = '20' %}
121
+ {%- elif yesterday_day == '22' %}
122
+ {%- set yesterday_day = '21' %}
123
+ {%- elif yesterday_day == '23' %}
124
+ {%- set yesterday_day = '22' %}
125
+ {%- elif yesterday_day == '24' %}
126
+ {%- set yesterday_day = '23' %}
127
+ {%- elif yesterday_day == '25' %}
128
+ {%- set yesterday_day = '24' %}
129
+ {%- elif yesterday_day == '26' %}
130
+ {%- set yesterday_day = '25' %}
131
+ {%- elif yesterday_day == '27' %}
132
+ {%- set yesterday_day = '26' %}
133
+ {%- elif yesterday_day == '28' %}
134
+ {%- set yesterday_day = '27' %}
135
+ {%- elif yesterday_day == '29' %}
136
+ {%- set yesterday_day = '28' %}
137
+ {%- elif yesterday_day == '30' %}
138
+ {%- set yesterday_day = '29' %}
139
+ {%- elif yesterday_day == '31' %}
140
+ {%- set yesterday_day = '30' %}
141
+ {%- endif %}
142
+ {#- Edits made by Unsloth #}
143
+ {%- set yesterday_date = yesterday_year + '-' + yesterday_month + '-' + yesterday_day %}
144
+ {%- set default_system_message = "You are Mistral-Small-3.2-24B-Instruct-2506, a Large Language Model (LLM) created by Mistral AI, a French startup headquartered in Paris.\nYou power an AI assistant called Le Chat.\nYour knowledge base was last updated on 2023-10-01.\nThe current date is " + today_date + ".\n\nWhen you\'re not sure about some information or when the user\'s request requires up-to-date or specific data, you must use the available tools to fetch the information. Do not hesitate to use tools whenever they can provide a more accurate or complete response. If no relevant tools are available, then clearly state that you don\'t have the information and avoid making up anything.\nIf the user\'s question is not clear, ambiguous, or does not provide enough context for you to accurately answer the question, you do not try to answer it right away and you rather ask the user to clarify their request (e.g. \"What are some good restaurants around me?\" => \"Where are you?\" or \"When is the next flight to Tokyo\" => \"Where do you travel from?\").\nYou are always very attentive to dates, in particular you try to resolve dates (e.g. \"yesterday\" is " + yesterday_date + ") and when asked about information at specific dates, you discard information that is at another date.\nYou follow these instructions in all languages, and always respond to the user in the language they use or request.\nNext sections describe the capabilities that you have.\n\n# WEB BROWSING INSTRUCTIONS\n\nYou cannot perform any web search or access internet to open URLs, links etc. If it seems like the user is expecting you to do so, you clarify the situation and ask the user to copy paste the text directly in the chat.\n\n# MULTI-MODAL INSTRUCTIONS\n\nYou have the ability to read images, but you cannot generate images. You also cannot transcribe audio files or videos.\nYou cannot read nor transcribe audio files or videos.\n\n# TOOL CALLING INSTRUCTIONS\n\nYou may have access to tools that you can use to fetch information or perform actions. You must use these tools in the following situations:\n\n1. When the request requires up-to-date information.\n2. When the request requires specific data that you do not have in your knowledge base.\n3. When the request involves actions that you cannot perform without tools.\n\nAlways prioritize using tools to provide the most accurate and helpful response. If tools are not available, inform the user that you cannot perform the requested action at the moment." %}
145
+
146
+ {{- bos_token }}
147
+
148
+ {%- if messages[0]['role'] == 'system' %}
149
+ {%- if messages[0]['content'] is string %}
150
+ {%- set system_message = messages[0]['content'] %}
151
+ {%- elif messages[0]['content'] is iterable %}
152
+ {%- set system_message = messages[0]['content'][0]['text'] %}
153
+ {%- else %}
154
+ {%- set system_message = messages[0]['content']|string %}
155
+ {%- endif %}
156
+ {%- set loop_messages = messages[1:] %}
157
+ {%- else %}
158
+ {%- set system_message = default_system_message %}
159
+ {%- set loop_messages = messages %}
160
+ {%- endif %}
161
+ {{- '[SYSTEM_PROMPT]' + system_message + '[/SYSTEM_PROMPT]' }}
162
+
163
+
164
+ {#- Tool description appended ONLY to last user message. Edits made by Unsloth #}
165
+ {#- Tool description appended also if last message is tool. Edits made by Unsloth #}
166
+ {%- set tools_description = "" %}
167
+ {%- set has_tools = false %}
168
+ {#- Cannot use set append_tools_index = loop.index0 since temporary variable - must use namespace #}
169
+ {%- set ns = namespace(append_tools_index=0, append_tools=false) %}
170
+
171
+ {%- if tools is defined and tools is not none and tools|length > 0 %}
172
+
173
+ {%- set has_tools = true %}
174
+ {%- set tools_description = "[AVAILABLE_TOOLS]" + (tools | tojson) + "[/AVAILABLE_TOOLS]" %}
175
+
176
+ {#- If User,Assistant,Tool,Tool we also need to append tools_description to last assistant WITHOUT tool_calls defined. Edits made by Unsloth #}
177
+ {%- if (loop_messages|last)['role'] == 'tool' %}
178
+
179
+ {#- Find last assistant WITHOUT tool_calls defined #}
180
+ {%- set ns.append_tools = true %}
181
+ {%- for message in loop_messages %}
182
+ {%- if message['role'] == 'assistant' %}
183
+
184
+ {#- Cannot use set append_tools_index = loop.index0 since temporary variable - must use namespace #}
185
+
186
+ {%- if message['tool_calls'] is not defined or message['tool_calls'] is none %}
187
+ {%- set ns.append_tools_index = loop.index0 %}
188
+ {%- endif %}
189
+
190
+ {%- endif %}
191
+ {%- endfor %}
192
+
193
+ {%- endif %}
194
+
195
+ {%- endif %}
196
+
197
+ {%- for message in loop_messages %}
198
+ {%- if message['role'] == 'user' %}
199
+
200
+ {%- if has_tools and loop.last %}
201
+ {{- tools_description }}
202
+ {%- endif %}
203
+
204
+ {#- If directly called tools in first turn, prepend to user #}
205
+ {%- if ns.append_tools and ns.append_tools_index == 0 %}
206
+ {{- tools_description }}
207
+ {%- endif %}
208
+
209
+ {%- if message['content'] is string %}
210
+ {{- '[INST]' + message['content'] + '[/INST]' }}
211
+ {%- else %}
212
+ {{- '[INST]' }}
213
+ {%- for block in message['content'] %}
214
+ {%- if block['type'] == 'text' %}
215
+
216
+ {#- Original did not have content which is weird. Added by Un-sloth. #}
217
+ {%- if block['text'] is defined %}
218
+ {{- block['text'] }}
219
+ {%- else %}
220
+ {{- block['content'] }}
221
+ {%- endif %}
222
+
223
+ {%- elif block['type'] in ['image', 'image_url'] %}
224
+ {{- '[IMG]' }}
225
+ {%- else %}
226
+ {{- raise_exception('Only text and image blocks are supported in message content!') }}
227
+ {%- endif %}
228
+ {%- endfor %}
229
+ {{- '[/INST]' }}
230
+ {%- endif %}
231
+
232
+ {%- elif message['role'] == 'system' %}
233
+ {%- if message['content'] is string %}
234
+ {{- '[SYSTEM_PROMPT]' + message['content'] + '[/SYSTEM_PROMPT]' }}
235
+ {%- elif message['content'] is iterable %}
236
+ {{- '[SYSTEM_PROMPT]' + message['content'][0]['text'] + '[/SYSTEM_PROMPT]' }}
237
+ {%- else %}
238
+ {{- '[SYSTEM_PROMPT]' + message['content']|string + '[/SYSTEM_PROMPT]' }}
239
+ {%- endif %}
240
+
241
+
242
+ {%- elif message['role'] == 'assistant' %}
243
+ {%- if message['content'] is string %}
244
+ {{- message['content'] }}
245
+ {%- elif message['content'] is iterable %}
246
+ {{- message['content'][0]['text'] }}
247
+ {%- else %}
248
+ {{- message['content']|string }}
249
+ {%- endif %}
250
+
251
+ {#- If User,Assistant,Tool,Tool we also need to append tools_description. Edits made by Unsloth #}
252
+
253
+ {%- if has_tools and (loop.index0 == ns.append_tools_index) %}
254
+ {{- eos_token }}
255
+ {{- tools_description }}
256
+ {%- endif %}
257
+
258
+ {%- if message['tool_calls'] is defined and message['tool_calls'] is not none %}
259
+ {%- for tool in message['tool_calls'] %}
260
+ {%- if tool['id'] is not defined %}
261
+ {{- raise_exception('Tool ID must be provided!') }}
262
+ {%- endif %}
263
+ {%- set tool_call_id = tool['id'] %}
264
+ {%- if tool_call_id is not string or tool_call_id|length < 9 %}
265
+ {{- raise_exception("Tool call IDs should be alphanumeric strings with length >= 9!") }}
266
+ {%- endif %}
267
+ {%- set arguments = tool['function']['arguments'] %}
268
+ {%- if arguments is not string %}
269
+ {%- set arguments = arguments|tojson %}
270
+ {%- endif %}
271
+ {#- Must list tool calls AFTER assistant. Edits made by Un-sloth #}
272
+ {{- "[TOOL_CALLS]" + tool['function']['name'] + "[CALL_ID]" + tool_call_id + "[ARGS]" + arguments }}
273
+ {%- endfor %}
274
+ {%- endif %}
275
+
276
+ {#- Must not add EOS if added tools_description. Unsloth edits. #}
277
+ {%- if (loop.index0 != ns.append_tools_index) %}
278
+ {{- eos_token }}
279
+ {%- endif %}
280
+
281
+ {%- elif message["role"] == "tool_results" or message["role"] == "tool" %}
282
+ {%- if message.content is defined and message.content.content is defined %}
283
+ {%- set content = message.content.content %}
284
+ {%- else %}
285
+ {%- set content = message.content %}
286
+ {%- endif %}
287
+ {%- if message['tool_call_id'] is not defined %}
288
+ {{- raise_exception('tool_call_id must be provided!') }}
289
+ {%- endif %}
290
+ {%- set tool_call_id = message['tool_call_id'] %}
291
+ {%- if tool_call_id is not string or tool_call_id|length < 9 %}
292
+ {{- raise_exception("Tool call IDs should be alphanumeric strings with length >= 9!") }}
293
+ {%- endif %}
294
+ {{- "[TOOL_RESULTS]" + tool_call_id + "[TOOL_CONTENT]" + content|string + "[/TOOL_RESULTS]" }}
295
+
296
+ {%- else %}
297
+ {{- raise_exception('Only user, systemm assistant and tool roles are supported in the custom template made by Unsloth!') }}
298
+ {%- endif %}
299
+ {%- endfor %}
300
+ {#- Copyright 2025-present Unsloth. Apache 2.0 License. #}
config.json ADDED
@@ -0,0 +1,270 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Mistral3ForConditionalGeneration"
4
+ ],
5
+ "bos_token_id": 1,
6
+ "dtype": "bfloat16",
7
+ "eos_token_id": 2,
8
+ "image_token_index": 10,
9
+ "model_type": "mistral3",
10
+ "multimodal_projector_bias": false,
11
+ "pad_token_id": 11,
12
+ "projector_hidden_act": "gelu",
13
+ "quantization_config": {
14
+ "config_groups": {
15
+ "group_0": {
16
+ "format": "nvfp4-pack-quantized",
17
+ "input_activations": {
18
+ "actorder": null,
19
+ "block_structure": null,
20
+ "dynamic": "local",
21
+ "group_size": 16,
22
+ "num_bits": 4,
23
+ "observer": "minmax",
24
+ "observer_kwargs": {},
25
+ "strategy": "tensor_group",
26
+ "symmetric": true,
27
+ "type": "float"
28
+ },
29
+ "output_activations": null,
30
+ "targets": [
31
+ "Linear"
32
+ ],
33
+ "weights": {
34
+ "actorder": null,
35
+ "block_structure": null,
36
+ "dynamic": false,
37
+ "group_size": 16,
38
+ "num_bits": 4,
39
+ "observer": "mse",
40
+ "observer_kwargs": {},
41
+ "strategy": "tensor_group",
42
+ "symmetric": true,
43
+ "type": "float"
44
+ }
45
+ }
46
+ },
47
+ "format": "nvfp4-pack-quantized",
48
+ "global_compression_ratio": null,
49
+ "ignore": [
50
+ "model.vision_tower.transformer.layers.0.feed_forward.gate_proj",
51
+ "model.vision_tower.transformer.layers.0.feed_forward.up_proj",
52
+ "model.vision_tower.transformer.layers.0.feed_forward.down_proj",
53
+ "model.vision_tower.transformer.layers.0.attention.k_proj",
54
+ "model.vision_tower.transformer.layers.0.attention.v_proj",
55
+ "model.vision_tower.transformer.layers.0.attention.q_proj",
56
+ "model.vision_tower.transformer.layers.0.attention.o_proj",
57
+ "model.vision_tower.transformer.layers.1.feed_forward.gate_proj",
58
+ "model.vision_tower.transformer.layers.1.feed_forward.up_proj",
59
+ "model.vision_tower.transformer.layers.1.feed_forward.down_proj",
60
+ "model.vision_tower.transformer.layers.1.attention.k_proj",
61
+ "model.vision_tower.transformer.layers.1.attention.v_proj",
62
+ "model.vision_tower.transformer.layers.1.attention.q_proj",
63
+ "model.vision_tower.transformer.layers.1.attention.o_proj",
64
+ "model.vision_tower.transformer.layers.2.feed_forward.gate_proj",
65
+ "model.vision_tower.transformer.layers.2.feed_forward.up_proj",
66
+ "model.vision_tower.transformer.layers.2.feed_forward.down_proj",
67
+ "model.vision_tower.transformer.layers.2.attention.k_proj",
68
+ "model.vision_tower.transformer.layers.2.attention.v_proj",
69
+ "model.vision_tower.transformer.layers.2.attention.q_proj",
70
+ "model.vision_tower.transformer.layers.2.attention.o_proj",
71
+ "model.vision_tower.transformer.layers.3.feed_forward.gate_proj",
72
+ "model.vision_tower.transformer.layers.3.feed_forward.up_proj",
73
+ "model.vision_tower.transformer.layers.3.feed_forward.down_proj",
74
+ "model.vision_tower.transformer.layers.3.attention.k_proj",
75
+ "model.vision_tower.transformer.layers.3.attention.v_proj",
76
+ "model.vision_tower.transformer.layers.3.attention.q_proj",
77
+ "model.vision_tower.transformer.layers.3.attention.o_proj",
78
+ "model.vision_tower.transformer.layers.4.feed_forward.gate_proj",
79
+ "model.vision_tower.transformer.layers.4.feed_forward.up_proj",
80
+ "model.vision_tower.transformer.layers.4.feed_forward.down_proj",
81
+ "model.vision_tower.transformer.layers.4.attention.k_proj",
82
+ "model.vision_tower.transformer.layers.4.attention.v_proj",
83
+ "model.vision_tower.transformer.layers.4.attention.q_proj",
84
+ "model.vision_tower.transformer.layers.4.attention.o_proj",
85
+ "model.vision_tower.transformer.layers.5.feed_forward.gate_proj",
86
+ "model.vision_tower.transformer.layers.5.feed_forward.up_proj",
87
+ "model.vision_tower.transformer.layers.5.feed_forward.down_proj",
88
+ "model.vision_tower.transformer.layers.5.attention.k_proj",
89
+ "model.vision_tower.transformer.layers.5.attention.v_proj",
90
+ "model.vision_tower.transformer.layers.5.attention.q_proj",
91
+ "model.vision_tower.transformer.layers.5.attention.o_proj",
92
+ "model.vision_tower.transformer.layers.6.feed_forward.gate_proj",
93
+ "model.vision_tower.transformer.layers.6.feed_forward.up_proj",
94
+ "model.vision_tower.transformer.layers.6.feed_forward.down_proj",
95
+ "model.vision_tower.transformer.layers.6.attention.k_proj",
96
+ "model.vision_tower.transformer.layers.6.attention.v_proj",
97
+ "model.vision_tower.transformer.layers.6.attention.q_proj",
98
+ "model.vision_tower.transformer.layers.6.attention.o_proj",
99
+ "model.vision_tower.transformer.layers.7.feed_forward.gate_proj",
100
+ "model.vision_tower.transformer.layers.7.feed_forward.up_proj",
101
+ "model.vision_tower.transformer.layers.7.feed_forward.down_proj",
102
+ "model.vision_tower.transformer.layers.7.attention.k_proj",
103
+ "model.vision_tower.transformer.layers.7.attention.v_proj",
104
+ "model.vision_tower.transformer.layers.7.attention.q_proj",
105
+ "model.vision_tower.transformer.layers.7.attention.o_proj",
106
+ "model.vision_tower.transformer.layers.8.feed_forward.gate_proj",
107
+ "model.vision_tower.transformer.layers.8.feed_forward.up_proj",
108
+ "model.vision_tower.transformer.layers.8.feed_forward.down_proj",
109
+ "model.vision_tower.transformer.layers.8.attention.k_proj",
110
+ "model.vision_tower.transformer.layers.8.attention.v_proj",
111
+ "model.vision_tower.transformer.layers.8.attention.q_proj",
112
+ "model.vision_tower.transformer.layers.8.attention.o_proj",
113
+ "model.vision_tower.transformer.layers.9.feed_forward.gate_proj",
114
+ "model.vision_tower.transformer.layers.9.feed_forward.up_proj",
115
+ "model.vision_tower.transformer.layers.9.feed_forward.down_proj",
116
+ "model.vision_tower.transformer.layers.9.attention.k_proj",
117
+ "model.vision_tower.transformer.layers.9.attention.v_proj",
118
+ "model.vision_tower.transformer.layers.9.attention.q_proj",
119
+ "model.vision_tower.transformer.layers.9.attention.o_proj",
120
+ "model.vision_tower.transformer.layers.10.feed_forward.gate_proj",
121
+ "model.vision_tower.transformer.layers.10.feed_forward.up_proj",
122
+ "model.vision_tower.transformer.layers.10.feed_forward.down_proj",
123
+ "model.vision_tower.transformer.layers.10.attention.k_proj",
124
+ "model.vision_tower.transformer.layers.10.attention.v_proj",
125
+ "model.vision_tower.transformer.layers.10.attention.q_proj",
126
+ "model.vision_tower.transformer.layers.10.attention.o_proj",
127
+ "model.vision_tower.transformer.layers.11.feed_forward.gate_proj",
128
+ "model.vision_tower.transformer.layers.11.feed_forward.up_proj",
129
+ "model.vision_tower.transformer.layers.11.feed_forward.down_proj",
130
+ "model.vision_tower.transformer.layers.11.attention.k_proj",
131
+ "model.vision_tower.transformer.layers.11.attention.v_proj",
132
+ "model.vision_tower.transformer.layers.11.attention.q_proj",
133
+ "model.vision_tower.transformer.layers.11.attention.o_proj",
134
+ "model.vision_tower.transformer.layers.12.feed_forward.gate_proj",
135
+ "model.vision_tower.transformer.layers.12.feed_forward.up_proj",
136
+ "model.vision_tower.transformer.layers.12.feed_forward.down_proj",
137
+ "model.vision_tower.transformer.layers.12.attention.k_proj",
138
+ "model.vision_tower.transformer.layers.12.attention.v_proj",
139
+ "model.vision_tower.transformer.layers.12.attention.q_proj",
140
+ "model.vision_tower.transformer.layers.12.attention.o_proj",
141
+ "model.vision_tower.transformer.layers.13.feed_forward.gate_proj",
142
+ "model.vision_tower.transformer.layers.13.feed_forward.up_proj",
143
+ "model.vision_tower.transformer.layers.13.feed_forward.down_proj",
144
+ "model.vision_tower.transformer.layers.13.attention.k_proj",
145
+ "model.vision_tower.transformer.layers.13.attention.v_proj",
146
+ "model.vision_tower.transformer.layers.13.attention.q_proj",
147
+ "model.vision_tower.transformer.layers.13.attention.o_proj",
148
+ "model.vision_tower.transformer.layers.14.feed_forward.gate_proj",
149
+ "model.vision_tower.transformer.layers.14.feed_forward.up_proj",
150
+ "model.vision_tower.transformer.layers.14.feed_forward.down_proj",
151
+ "model.vision_tower.transformer.layers.14.attention.k_proj",
152
+ "model.vision_tower.transformer.layers.14.attention.v_proj",
153
+ "model.vision_tower.transformer.layers.14.attention.q_proj",
154
+ "model.vision_tower.transformer.layers.14.attention.o_proj",
155
+ "model.vision_tower.transformer.layers.15.feed_forward.gate_proj",
156
+ "model.vision_tower.transformer.layers.15.feed_forward.up_proj",
157
+ "model.vision_tower.transformer.layers.15.feed_forward.down_proj",
158
+ "model.vision_tower.transformer.layers.15.attention.k_proj",
159
+ "model.vision_tower.transformer.layers.15.attention.v_proj",
160
+ "model.vision_tower.transformer.layers.15.attention.q_proj",
161
+ "model.vision_tower.transformer.layers.15.attention.o_proj",
162
+ "model.vision_tower.transformer.layers.16.feed_forward.gate_proj",
163
+ "model.vision_tower.transformer.layers.16.feed_forward.up_proj",
164
+ "model.vision_tower.transformer.layers.16.feed_forward.down_proj",
165
+ "model.vision_tower.transformer.layers.16.attention.k_proj",
166
+ "model.vision_tower.transformer.layers.16.attention.v_proj",
167
+ "model.vision_tower.transformer.layers.16.attention.q_proj",
168
+ "model.vision_tower.transformer.layers.16.attention.o_proj",
169
+ "model.vision_tower.transformer.layers.17.feed_forward.gate_proj",
170
+ "model.vision_tower.transformer.layers.17.feed_forward.up_proj",
171
+ "model.vision_tower.transformer.layers.17.feed_forward.down_proj",
172
+ "model.vision_tower.transformer.layers.17.attention.k_proj",
173
+ "model.vision_tower.transformer.layers.17.attention.v_proj",
174
+ "model.vision_tower.transformer.layers.17.attention.q_proj",
175
+ "model.vision_tower.transformer.layers.17.attention.o_proj",
176
+ "model.vision_tower.transformer.layers.18.feed_forward.gate_proj",
177
+ "model.vision_tower.transformer.layers.18.feed_forward.up_proj",
178
+ "model.vision_tower.transformer.layers.18.feed_forward.down_proj",
179
+ "model.vision_tower.transformer.layers.18.attention.k_proj",
180
+ "model.vision_tower.transformer.layers.18.attention.v_proj",
181
+ "model.vision_tower.transformer.layers.18.attention.q_proj",
182
+ "model.vision_tower.transformer.layers.18.attention.o_proj",
183
+ "model.vision_tower.transformer.layers.19.feed_forward.gate_proj",
184
+ "model.vision_tower.transformer.layers.19.feed_forward.up_proj",
185
+ "model.vision_tower.transformer.layers.19.feed_forward.down_proj",
186
+ "model.vision_tower.transformer.layers.19.attention.k_proj",
187
+ "model.vision_tower.transformer.layers.19.attention.v_proj",
188
+ "model.vision_tower.transformer.layers.19.attention.q_proj",
189
+ "model.vision_tower.transformer.layers.19.attention.o_proj",
190
+ "model.vision_tower.transformer.layers.20.feed_forward.gate_proj",
191
+ "model.vision_tower.transformer.layers.20.feed_forward.up_proj",
192
+ "model.vision_tower.transformer.layers.20.feed_forward.down_proj",
193
+ "model.vision_tower.transformer.layers.20.attention.k_proj",
194
+ "model.vision_tower.transformer.layers.20.attention.v_proj",
195
+ "model.vision_tower.transformer.layers.20.attention.q_proj",
196
+ "model.vision_tower.transformer.layers.20.attention.o_proj",
197
+ "model.vision_tower.transformer.layers.21.feed_forward.gate_proj",
198
+ "model.vision_tower.transformer.layers.21.feed_forward.up_proj",
199
+ "model.vision_tower.transformer.layers.21.feed_forward.down_proj",
200
+ "model.vision_tower.transformer.layers.21.attention.k_proj",
201
+ "model.vision_tower.transformer.layers.21.attention.v_proj",
202
+ "model.vision_tower.transformer.layers.21.attention.q_proj",
203
+ "model.vision_tower.transformer.layers.21.attention.o_proj",
204
+ "model.vision_tower.transformer.layers.22.feed_forward.gate_proj",
205
+ "model.vision_tower.transformer.layers.22.feed_forward.up_proj",
206
+ "model.vision_tower.transformer.layers.22.feed_forward.down_proj",
207
+ "model.vision_tower.transformer.layers.22.attention.k_proj",
208
+ "model.vision_tower.transformer.layers.22.attention.v_proj",
209
+ "model.vision_tower.transformer.layers.22.attention.q_proj",
210
+ "model.vision_tower.transformer.layers.22.attention.o_proj",
211
+ "model.vision_tower.transformer.layers.23.feed_forward.gate_proj",
212
+ "model.vision_tower.transformer.layers.23.feed_forward.up_proj",
213
+ "model.vision_tower.transformer.layers.23.feed_forward.down_proj",
214
+ "model.vision_tower.transformer.layers.23.attention.k_proj",
215
+ "model.vision_tower.transformer.layers.23.attention.v_proj",
216
+ "model.vision_tower.transformer.layers.23.attention.q_proj",
217
+ "model.vision_tower.transformer.layers.23.attention.o_proj",
218
+ "model.multi_modal_projector.patch_merger.merging_layer",
219
+ "model.multi_modal_projector.linear_1",
220
+ "model.multi_modal_projector.linear_2",
221
+ "lm_head"
222
+ ],
223
+ "kv_cache_scheme": null,
224
+ "quant_method": "compressed-tensors",
225
+ "quantization_status": "compressed",
226
+ "sparsity_config": {},
227
+ "transform_config": {},
228
+ "version": "0.11.1.a20250929"
229
+ },
230
+ "spatial_merge_size": 2,
231
+ "text_config": {
232
+ "attention_dropout": 0.0,
233
+ "dtype": "bfloat16",
234
+ "head_dim": 128,
235
+ "hidden_act": "silu",
236
+ "hidden_size": 5120,
237
+ "initializer_range": 0.02,
238
+ "intermediate_size": 32768,
239
+ "max_position_embeddings": 131072,
240
+ "model_type": "mistral",
241
+ "num_attention_heads": 32,
242
+ "num_hidden_layers": 40,
243
+ "num_key_value_heads": 8,
244
+ "rms_norm_eps": 1e-05,
245
+ "rope_theta": 1000000000.0,
246
+ "sliding_window": null,
247
+ "use_cache": true,
248
+ "vocab_size": 131072
249
+ },
250
+ "tie_word_embeddings": false,
251
+ "transformers_version": "4.56.1",
252
+ "unsloth_fixed": true,
253
+ "vision_config": {
254
+ "attention_dropout": 0.0,
255
+ "dtype": "bfloat16",
256
+ "head_dim": 64,
257
+ "hidden_act": "silu",
258
+ "hidden_size": 1024,
259
+ "image_size": 1540,
260
+ "initializer_range": 0.02,
261
+ "intermediate_size": 4096,
262
+ "model_type": "pixtral",
263
+ "num_attention_heads": 16,
264
+ "num_channels": 3,
265
+ "num_hidden_layers": 24,
266
+ "patch_size": 14,
267
+ "rope_theta": 10000.0
268
+ },
269
+ "vision_feature_layer": -1
270
+ }
generation_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 1,
3
+ "do_sample": true,
4
+ "eos_token_id": 2,
5
+ "max_length": 131072,
6
+ "pad_token_id": 11,
7
+ "temperature": 0.15,
8
+ "transformers_version": "4.56.1"
9
+ }
model-00001-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cc0279337fdd97f82327325758788610a47e04959f0ce6e0e3053131148ee74
3
+ size 4939410608
model-00002-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:336e466e3b0ff4e6c1e2f22d53add7e3b6be5690b903cee291f9fe0fce80e1a9
3
+ size 4991611888
model-00003-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b003847e04b71b76f21d54ede497f50132fe9a483ddc18a81785b7e0b14a51b5
3
+ size 4794356016
model-00004-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b72834082387ce74af5414bf353165eab35661f3261f7c9c9f16da505b64e93
3
+ size 1342177424
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
preprocessor_config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": null,
3
+ "data_format": "channels_first",
4
+ "default_to_square": true,
5
+ "device": null,
6
+ "disable_grouping": null,
7
+ "do_center_crop": null,
8
+ "do_convert_rgb": true,
9
+ "do_normalize": true,
10
+ "do_rescale": true,
11
+ "do_resize": true,
12
+ "image_mean": [
13
+ 0.48145466,
14
+ 0.4578275,
15
+ 0.40821073
16
+ ],
17
+ "image_processor_type": "PixtralImageProcessorFast",
18
+ "image_std": [
19
+ 0.26862954,
20
+ 0.26130258,
21
+ 0.27577711
22
+ ],
23
+ "input_data_format": null,
24
+ "patch_size": 14,
25
+ "processor_class": "PixtralProcessor",
26
+ "resample": 3,
27
+ "rescale_factor": 0.00392156862745098,
28
+ "return_tensors": null,
29
+ "size": {
30
+ "longest_edge": 1540
31
+ }
32
+ }
processor_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "image_break_token": "[IMG_BREAK]",
3
+ "image_end_token": "[IMG_END]",
4
+ "image_token": "[IMG]",
5
+ "patch_size": 14,
6
+ "processor_class": "PixtralProcessor",
7
+ "spatial_merge_size": 2
8
+ }
recipe.yaml ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ default_stage:
2
+ default_modifiers:
3
+ SmoothQuantModifier:
4
+ smoothing_strength: 0.9
5
+ mappings:
6
+ - !!python/tuple
7
+ - ['re:.*q_proj', 're:.*k_proj', 're:.*v_proj']
8
+ - re:.*input_layernorm
9
+ - !!python/tuple
10
+ - ['re:.*gate_proj', 're:.*up_proj']
11
+ - re:.*post_attention_layernorm
12
+ ignore: []
13
+ QuantizationModifier:
14
+ config_groups:
15
+ group_0:
16
+ targets: [Linear]
17
+ weights:
18
+ num_bits: 4
19
+ type: float
20
+ symmetric: true
21
+ group_size: 16
22
+ strategy: tensor_group
23
+ block_structure: null
24
+ dynamic: false
25
+ actorder: null
26
+ observer: mse
27
+ observer_kwargs: {}
28
+ input_activations:
29
+ num_bits: 4
30
+ type: float
31
+ symmetric: true
32
+ group_size: 16
33
+ strategy: tensor_group
34
+ block_structure: null
35
+ dynamic: local
36
+ actorder: null
37
+ observer: minmax
38
+ observer_kwargs: {}
39
+ output_activations: null
40
+ format: null
41
+ targets: [Linear]
42
+ ignore: ['re:.*lm_head.*', 're:.*multi_modal_projector.*', 're:.*vision_tower.*', 're:.*model.norm.*',
43
+ 're:.*embed_tokens.*']
special_tokens_map.json ADDED
@@ -0,0 +1,1032 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<unk>",
4
+ "<s>",
5
+ "</s>",
6
+ "[INST]",
7
+ "[/INST]",
8
+ "[AVAILABLE_TOOLS]",
9
+ "[/AVAILABLE_TOOLS]",
10
+ "[TOOL_RESULTS]",
11
+ "[/TOOL_RESULTS]",
12
+ "[TOOL_CALLS]",
13
+ "[IMG]",
14
+ "<pad>",
15
+ "[IMG_BREAK]",
16
+ "[IMG_END]",
17
+ "[PREFIX]",
18
+ "[MIDDLE]",
19
+ "[SUFFIX]",
20
+ "[SYSTEM_PROMPT]",
21
+ "[/SYSTEM_PROMPT]",
22
+ "[TOOL_CONTENT]",
23
+ "<SPECIAL_20>",
24
+ "<SPECIAL_21>",
25
+ "<SPECIAL_22>",
26
+ "<SPECIAL_23>",
27
+ "<SPECIAL_24>",
28
+ "<SPECIAL_25>",
29
+ "<SPECIAL_26>",
30
+ "<SPECIAL_27>",
31
+ "<SPECIAL_28>",
32
+ "<SPECIAL_29>",
33
+ "<SPECIAL_30>",
34
+ "<SPECIAL_31>",
35
+ "[ARGS]",
36
+ "[CALL_ID]",
37
+ "<SPECIAL_34>",
38
+ "<SPECIAL_35>",
39
+ "<SPECIAL_36>",
40
+ "<SPECIAL_37>",
41
+ "<SPECIAL_38>",
42
+ "<SPECIAL_39>",
43
+ "<SPECIAL_40>",
44
+ "<SPECIAL_41>",
45
+ "<SPECIAL_42>",
46
+ "<SPECIAL_43>",
47
+ "<SPECIAL_44>",
48
+ "<SPECIAL_45>",
49
+ "<SPECIAL_46>",
50
+ "<SPECIAL_47>",
51
+ "<SPECIAL_48>",
52
+ "<SPECIAL_49>",
53
+ "<SPECIAL_50>",
54
+ "<SPECIAL_51>",
55
+ "<SPECIAL_52>",
56
+ "<SPECIAL_53>",
57
+ "<SPECIAL_54>",
58
+ "<SPECIAL_55>",
59
+ "<SPECIAL_56>",
60
+ "<SPECIAL_57>",
61
+ "<SPECIAL_58>",
62
+ "<SPECIAL_59>",
63
+ "<SPECIAL_60>",
64
+ "<SPECIAL_61>",
65
+ "<SPECIAL_62>",
66
+ "<SPECIAL_63>",
67
+ "<SPECIAL_64>",
68
+ "<SPECIAL_65>",
69
+ "<SPECIAL_66>",
70
+ "<SPECIAL_67>",
71
+ "<SPECIAL_68>",
72
+ "<SPECIAL_69>",
73
+ "<SPECIAL_70>",
74
+ "<SPECIAL_71>",
75
+ "<SPECIAL_72>",
76
+ "<SPECIAL_73>",
77
+ "<SPECIAL_74>",
78
+ "<SPECIAL_75>",
79
+ "<SPECIAL_76>",
80
+ "<SPECIAL_77>",
81
+ "<SPECIAL_78>",
82
+ "<SPECIAL_79>",
83
+ "<SPECIAL_80>",
84
+ "<SPECIAL_81>",
85
+ "<SPECIAL_82>",
86
+ "<SPECIAL_83>",
87
+ "<SPECIAL_84>",
88
+ "<SPECIAL_85>",
89
+ "<SPECIAL_86>",
90
+ "<SPECIAL_87>",
91
+ "<SPECIAL_88>",
92
+ "<SPECIAL_89>",
93
+ "<SPECIAL_90>",
94
+ "<SPECIAL_91>",
95
+ "<SPECIAL_92>",
96
+ "<SPECIAL_93>",
97
+ "<SPECIAL_94>",
98
+ "<SPECIAL_95>",
99
+ "<SPECIAL_96>",
100
+ "<SPECIAL_97>",
101
+ "<SPECIAL_98>",
102
+ "<SPECIAL_99>",
103
+ "<SPECIAL_100>",
104
+ "<SPECIAL_101>",
105
+ "<SPECIAL_102>",
106
+ "<SPECIAL_103>",
107
+ "<SPECIAL_104>",
108
+ "<SPECIAL_105>",
109
+ "<SPECIAL_106>",
110
+ "<SPECIAL_107>",
111
+ "<SPECIAL_108>",
112
+ "<SPECIAL_109>",
113
+ "<SPECIAL_110>",
114
+ "<SPECIAL_111>",
115
+ "<SPECIAL_112>",
116
+ "<SPECIAL_113>",
117
+ "<SPECIAL_114>",
118
+ "<SPECIAL_115>",
119
+ "<SPECIAL_116>",
120
+ "<SPECIAL_117>",
121
+ "<SPECIAL_118>",
122
+ "<SPECIAL_119>",
123
+ "<SPECIAL_120>",
124
+ "<SPECIAL_121>",
125
+ "<SPECIAL_122>",
126
+ "<SPECIAL_123>",
127
+ "<SPECIAL_124>",
128
+ "<SPECIAL_125>",
129
+ "<SPECIAL_126>",
130
+ "<SPECIAL_127>",
131
+ "<SPECIAL_128>",
132
+ "<SPECIAL_129>",
133
+ "<SPECIAL_130>",
134
+ "<SPECIAL_131>",
135
+ "<SPECIAL_132>",
136
+ "<SPECIAL_133>",
137
+ "<SPECIAL_134>",
138
+ "<SPECIAL_135>",
139
+ "<SPECIAL_136>",
140
+ "<SPECIAL_137>",
141
+ "<SPECIAL_138>",
142
+ "<SPECIAL_139>",
143
+ "<SPECIAL_140>",
144
+ "<SPECIAL_141>",
145
+ "<SPECIAL_142>",
146
+ "<SPECIAL_143>",
147
+ "<SPECIAL_144>",
148
+ "<SPECIAL_145>",
149
+ "<SPECIAL_146>",
150
+ "<SPECIAL_147>",
151
+ "<SPECIAL_148>",
152
+ "<SPECIAL_149>",
153
+ "<SPECIAL_150>",
154
+ "<SPECIAL_151>",
155
+ "<SPECIAL_152>",
156
+ "<SPECIAL_153>",
157
+ "<SPECIAL_154>",
158
+ "<SPECIAL_155>",
159
+ "<SPECIAL_156>",
160
+ "<SPECIAL_157>",
161
+ "<SPECIAL_158>",
162
+ "<SPECIAL_159>",
163
+ "<SPECIAL_160>",
164
+ "<SPECIAL_161>",
165
+ "<SPECIAL_162>",
166
+ "<SPECIAL_163>",
167
+ "<SPECIAL_164>",
168
+ "<SPECIAL_165>",
169
+ "<SPECIAL_166>",
170
+ "<SPECIAL_167>",
171
+ "<SPECIAL_168>",
172
+ "<SPECIAL_169>",
173
+ "<SPECIAL_170>",
174
+ "<SPECIAL_171>",
175
+ "<SPECIAL_172>",
176
+ "<SPECIAL_173>",
177
+ "<SPECIAL_174>",
178
+ "<SPECIAL_175>",
179
+ "<SPECIAL_176>",
180
+ "<SPECIAL_177>",
181
+ "<SPECIAL_178>",
182
+ "<SPECIAL_179>",
183
+ "<SPECIAL_180>",
184
+ "<SPECIAL_181>",
185
+ "<SPECIAL_182>",
186
+ "<SPECIAL_183>",
187
+ "<SPECIAL_184>",
188
+ "<SPECIAL_185>",
189
+ "<SPECIAL_186>",
190
+ "<SPECIAL_187>",
191
+ "<SPECIAL_188>",
192
+ "<SPECIAL_189>",
193
+ "<SPECIAL_190>",
194
+ "<SPECIAL_191>",
195
+ "<SPECIAL_192>",
196
+ "<SPECIAL_193>",
197
+ "<SPECIAL_194>",
198
+ "<SPECIAL_195>",
199
+ "<SPECIAL_196>",
200
+ "<SPECIAL_197>",
201
+ "<SPECIAL_198>",
202
+ "<SPECIAL_199>",
203
+ "<SPECIAL_200>",
204
+ "<SPECIAL_201>",
205
+ "<SPECIAL_202>",
206
+ "<SPECIAL_203>",
207
+ "<SPECIAL_204>",
208
+ "<SPECIAL_205>",
209
+ "<SPECIAL_206>",
210
+ "<SPECIAL_207>",
211
+ "<SPECIAL_208>",
212
+ "<SPECIAL_209>",
213
+ "<SPECIAL_210>",
214
+ "<SPECIAL_211>",
215
+ "<SPECIAL_212>",
216
+ "<SPECIAL_213>",
217
+ "<SPECIAL_214>",
218
+ "<SPECIAL_215>",
219
+ "<SPECIAL_216>",
220
+ "<SPECIAL_217>",
221
+ "<SPECIAL_218>",
222
+ "<SPECIAL_219>",
223
+ "<SPECIAL_220>",
224
+ "<SPECIAL_221>",
225
+ "<SPECIAL_222>",
226
+ "<SPECIAL_223>",
227
+ "<SPECIAL_224>",
228
+ "<SPECIAL_225>",
229
+ "<SPECIAL_226>",
230
+ "<SPECIAL_227>",
231
+ "<SPECIAL_228>",
232
+ "<SPECIAL_229>",
233
+ "<SPECIAL_230>",
234
+ "<SPECIAL_231>",
235
+ "<SPECIAL_232>",
236
+ "<SPECIAL_233>",
237
+ "<SPECIAL_234>",
238
+ "<SPECIAL_235>",
239
+ "<SPECIAL_236>",
240
+ "<SPECIAL_237>",
241
+ "<SPECIAL_238>",
242
+ "<SPECIAL_239>",
243
+ "<SPECIAL_240>",
244
+ "<SPECIAL_241>",
245
+ "<SPECIAL_242>",
246
+ "<SPECIAL_243>",
247
+ "<SPECIAL_244>",
248
+ "<SPECIAL_245>",
249
+ "<SPECIAL_246>",
250
+ "<SPECIAL_247>",
251
+ "<SPECIAL_248>",
252
+ "<SPECIAL_249>",
253
+ "<SPECIAL_250>",
254
+ "<SPECIAL_251>",
255
+ "<SPECIAL_252>",
256
+ "<SPECIAL_253>",
257
+ "<SPECIAL_254>",
258
+ "<SPECIAL_255>",
259
+ "<SPECIAL_256>",
260
+ "<SPECIAL_257>",
261
+ "<SPECIAL_258>",
262
+ "<SPECIAL_259>",
263
+ "<SPECIAL_260>",
264
+ "<SPECIAL_261>",
265
+ "<SPECIAL_262>",
266
+ "<SPECIAL_263>",
267
+ "<SPECIAL_264>",
268
+ "<SPECIAL_265>",
269
+ "<SPECIAL_266>",
270
+ "<SPECIAL_267>",
271
+ "<SPECIAL_268>",
272
+ "<SPECIAL_269>",
273
+ "<SPECIAL_270>",
274
+ "<SPECIAL_271>",
275
+ "<SPECIAL_272>",
276
+ "<SPECIAL_273>",
277
+ "<SPECIAL_274>",
278
+ "<SPECIAL_275>",
279
+ "<SPECIAL_276>",
280
+ "<SPECIAL_277>",
281
+ "<SPECIAL_278>",
282
+ "<SPECIAL_279>",
283
+ "<SPECIAL_280>",
284
+ "<SPECIAL_281>",
285
+ "<SPECIAL_282>",
286
+ "<SPECIAL_283>",
287
+ "<SPECIAL_284>",
288
+ "<SPECIAL_285>",
289
+ "<SPECIAL_286>",
290
+ "<SPECIAL_287>",
291
+ "<SPECIAL_288>",
292
+ "<SPECIAL_289>",
293
+ "<SPECIAL_290>",
294
+ "<SPECIAL_291>",
295
+ "<SPECIAL_292>",
296
+ "<SPECIAL_293>",
297
+ "<SPECIAL_294>",
298
+ "<SPECIAL_295>",
299
+ "<SPECIAL_296>",
300
+ "<SPECIAL_297>",
301
+ "<SPECIAL_298>",
302
+ "<SPECIAL_299>",
303
+ "<SPECIAL_300>",
304
+ "<SPECIAL_301>",
305
+ "<SPECIAL_302>",
306
+ "<SPECIAL_303>",
307
+ "<SPECIAL_304>",
308
+ "<SPECIAL_305>",
309
+ "<SPECIAL_306>",
310
+ "<SPECIAL_307>",
311
+ "<SPECIAL_308>",
312
+ "<SPECIAL_309>",
313
+ "<SPECIAL_310>",
314
+ "<SPECIAL_311>",
315
+ "<SPECIAL_312>",
316
+ "<SPECIAL_313>",
317
+ "<SPECIAL_314>",
318
+ "<SPECIAL_315>",
319
+ "<SPECIAL_316>",
320
+ "<SPECIAL_317>",
321
+ "<SPECIAL_318>",
322
+ "<SPECIAL_319>",
323
+ "<SPECIAL_320>",
324
+ "<SPECIAL_321>",
325
+ "<SPECIAL_322>",
326
+ "<SPECIAL_323>",
327
+ "<SPECIAL_324>",
328
+ "<SPECIAL_325>",
329
+ "<SPECIAL_326>",
330
+ "<SPECIAL_327>",
331
+ "<SPECIAL_328>",
332
+ "<SPECIAL_329>",
333
+ "<SPECIAL_330>",
334
+ "<SPECIAL_331>",
335
+ "<SPECIAL_332>",
336
+ "<SPECIAL_333>",
337
+ "<SPECIAL_334>",
338
+ "<SPECIAL_335>",
339
+ "<SPECIAL_336>",
340
+ "<SPECIAL_337>",
341
+ "<SPECIAL_338>",
342
+ "<SPECIAL_339>",
343
+ "<SPECIAL_340>",
344
+ "<SPECIAL_341>",
345
+ "<SPECIAL_342>",
346
+ "<SPECIAL_343>",
347
+ "<SPECIAL_344>",
348
+ "<SPECIAL_345>",
349
+ "<SPECIAL_346>",
350
+ "<SPECIAL_347>",
351
+ "<SPECIAL_348>",
352
+ "<SPECIAL_349>",
353
+ "<SPECIAL_350>",
354
+ "<SPECIAL_351>",
355
+ "<SPECIAL_352>",
356
+ "<SPECIAL_353>",
357
+ "<SPECIAL_354>",
358
+ "<SPECIAL_355>",
359
+ "<SPECIAL_356>",
360
+ "<SPECIAL_357>",
361
+ "<SPECIAL_358>",
362
+ "<SPECIAL_359>",
363
+ "<SPECIAL_360>",
364
+ "<SPECIAL_361>",
365
+ "<SPECIAL_362>",
366
+ "<SPECIAL_363>",
367
+ "<SPECIAL_364>",
368
+ "<SPECIAL_365>",
369
+ "<SPECIAL_366>",
370
+ "<SPECIAL_367>",
371
+ "<SPECIAL_368>",
372
+ "<SPECIAL_369>",
373
+ "<SPECIAL_370>",
374
+ "<SPECIAL_371>",
375
+ "<SPECIAL_372>",
376
+ "<SPECIAL_373>",
377
+ "<SPECIAL_374>",
378
+ "<SPECIAL_375>",
379
+ "<SPECIAL_376>",
380
+ "<SPECIAL_377>",
381
+ "<SPECIAL_378>",
382
+ "<SPECIAL_379>",
383
+ "<SPECIAL_380>",
384
+ "<SPECIAL_381>",
385
+ "<SPECIAL_382>",
386
+ "<SPECIAL_383>",
387
+ "<SPECIAL_384>",
388
+ "<SPECIAL_385>",
389
+ "<SPECIAL_386>",
390
+ "<SPECIAL_387>",
391
+ "<SPECIAL_388>",
392
+ "<SPECIAL_389>",
393
+ "<SPECIAL_390>",
394
+ "<SPECIAL_391>",
395
+ "<SPECIAL_392>",
396
+ "<SPECIAL_393>",
397
+ "<SPECIAL_394>",
398
+ "<SPECIAL_395>",
399
+ "<SPECIAL_396>",
400
+ "<SPECIAL_397>",
401
+ "<SPECIAL_398>",
402
+ "<SPECIAL_399>",
403
+ "<SPECIAL_400>",
404
+ "<SPECIAL_401>",
405
+ "<SPECIAL_402>",
406
+ "<SPECIAL_403>",
407
+ "<SPECIAL_404>",
408
+ "<SPECIAL_405>",
409
+ "<SPECIAL_406>",
410
+ "<SPECIAL_407>",
411
+ "<SPECIAL_408>",
412
+ "<SPECIAL_409>",
413
+ "<SPECIAL_410>",
414
+ "<SPECIAL_411>",
415
+ "<SPECIAL_412>",
416
+ "<SPECIAL_413>",
417
+ "<SPECIAL_414>",
418
+ "<SPECIAL_415>",
419
+ "<SPECIAL_416>",
420
+ "<SPECIAL_417>",
421
+ "<SPECIAL_418>",
422
+ "<SPECIAL_419>",
423
+ "<SPECIAL_420>",
424
+ "<SPECIAL_421>",
425
+ "<SPECIAL_422>",
426
+ "<SPECIAL_423>",
427
+ "<SPECIAL_424>",
428
+ "<SPECIAL_425>",
429
+ "<SPECIAL_426>",
430
+ "<SPECIAL_427>",
431
+ "<SPECIAL_428>",
432
+ "<SPECIAL_429>",
433
+ "<SPECIAL_430>",
434
+ "<SPECIAL_431>",
435
+ "<SPECIAL_432>",
436
+ "<SPECIAL_433>",
437
+ "<SPECIAL_434>",
438
+ "<SPECIAL_435>",
439
+ "<SPECIAL_436>",
440
+ "<SPECIAL_437>",
441
+ "<SPECIAL_438>",
442
+ "<SPECIAL_439>",
443
+ "<SPECIAL_440>",
444
+ "<SPECIAL_441>",
445
+ "<SPECIAL_442>",
446
+ "<SPECIAL_443>",
447
+ "<SPECIAL_444>",
448
+ "<SPECIAL_445>",
449
+ "<SPECIAL_446>",
450
+ "<SPECIAL_447>",
451
+ "<SPECIAL_448>",
452
+ "<SPECIAL_449>",
453
+ "<SPECIAL_450>",
454
+ "<SPECIAL_451>",
455
+ "<SPECIAL_452>",
456
+ "<SPECIAL_453>",
457
+ "<SPECIAL_454>",
458
+ "<SPECIAL_455>",
459
+ "<SPECIAL_456>",
460
+ "<SPECIAL_457>",
461
+ "<SPECIAL_458>",
462
+ "<SPECIAL_459>",
463
+ "<SPECIAL_460>",
464
+ "<SPECIAL_461>",
465
+ "<SPECIAL_462>",
466
+ "<SPECIAL_463>",
467
+ "<SPECIAL_464>",
468
+ "<SPECIAL_465>",
469
+ "<SPECIAL_466>",
470
+ "<SPECIAL_467>",
471
+ "<SPECIAL_468>",
472
+ "<SPECIAL_469>",
473
+ "<SPECIAL_470>",
474
+ "<SPECIAL_471>",
475
+ "<SPECIAL_472>",
476
+ "<SPECIAL_473>",
477
+ "<SPECIAL_474>",
478
+ "<SPECIAL_475>",
479
+ "<SPECIAL_476>",
480
+ "<SPECIAL_477>",
481
+ "<SPECIAL_478>",
482
+ "<SPECIAL_479>",
483
+ "<SPECIAL_480>",
484
+ "<SPECIAL_481>",
485
+ "<SPECIAL_482>",
486
+ "<SPECIAL_483>",
487
+ "<SPECIAL_484>",
488
+ "<SPECIAL_485>",
489
+ "<SPECIAL_486>",
490
+ "<SPECIAL_487>",
491
+ "<SPECIAL_488>",
492
+ "<SPECIAL_489>",
493
+ "<SPECIAL_490>",
494
+ "<SPECIAL_491>",
495
+ "<SPECIAL_492>",
496
+ "<SPECIAL_493>",
497
+ "<SPECIAL_494>",
498
+ "<SPECIAL_495>",
499
+ "<SPECIAL_496>",
500
+ "<SPECIAL_497>",
501
+ "<SPECIAL_498>",
502
+ "<SPECIAL_499>",
503
+ "<SPECIAL_500>",
504
+ "<SPECIAL_501>",
505
+ "<SPECIAL_502>",
506
+ "<SPECIAL_503>",
507
+ "<SPECIAL_504>",
508
+ "<SPECIAL_505>",
509
+ "<SPECIAL_506>",
510
+ "<SPECIAL_507>",
511
+ "<SPECIAL_508>",
512
+ "<SPECIAL_509>",
513
+ "<SPECIAL_510>",
514
+ "<SPECIAL_511>",
515
+ "<SPECIAL_512>",
516
+ "<SPECIAL_513>",
517
+ "<SPECIAL_514>",
518
+ "<SPECIAL_515>",
519
+ "<SPECIAL_516>",
520
+ "<SPECIAL_517>",
521
+ "<SPECIAL_518>",
522
+ "<SPECIAL_519>",
523
+ "<SPECIAL_520>",
524
+ "<SPECIAL_521>",
525
+ "<SPECIAL_522>",
526
+ "<SPECIAL_523>",
527
+ "<SPECIAL_524>",
528
+ "<SPECIAL_525>",
529
+ "<SPECIAL_526>",
530
+ "<SPECIAL_527>",
531
+ "<SPECIAL_528>",
532
+ "<SPECIAL_529>",
533
+ "<SPECIAL_530>",
534
+ "<SPECIAL_531>",
535
+ "<SPECIAL_532>",
536
+ "<SPECIAL_533>",
537
+ "<SPECIAL_534>",
538
+ "<SPECIAL_535>",
539
+ "<SPECIAL_536>",
540
+ "<SPECIAL_537>",
541
+ "<SPECIAL_538>",
542
+ "<SPECIAL_539>",
543
+ "<SPECIAL_540>",
544
+ "<SPECIAL_541>",
545
+ "<SPECIAL_542>",
546
+ "<SPECIAL_543>",
547
+ "<SPECIAL_544>",
548
+ "<SPECIAL_545>",
549
+ "<SPECIAL_546>",
550
+ "<SPECIAL_547>",
551
+ "<SPECIAL_548>",
552
+ "<SPECIAL_549>",
553
+ "<SPECIAL_550>",
554
+ "<SPECIAL_551>",
555
+ "<SPECIAL_552>",
556
+ "<SPECIAL_553>",
557
+ "<SPECIAL_554>",
558
+ "<SPECIAL_555>",
559
+ "<SPECIAL_556>",
560
+ "<SPECIAL_557>",
561
+ "<SPECIAL_558>",
562
+ "<SPECIAL_559>",
563
+ "<SPECIAL_560>",
564
+ "<SPECIAL_561>",
565
+ "<SPECIAL_562>",
566
+ "<SPECIAL_563>",
567
+ "<SPECIAL_564>",
568
+ "<SPECIAL_565>",
569
+ "<SPECIAL_566>",
570
+ "<SPECIAL_567>",
571
+ "<SPECIAL_568>",
572
+ "<SPECIAL_569>",
573
+ "<SPECIAL_570>",
574
+ "<SPECIAL_571>",
575
+ "<SPECIAL_572>",
576
+ "<SPECIAL_573>",
577
+ "<SPECIAL_574>",
578
+ "<SPECIAL_575>",
579
+ "<SPECIAL_576>",
580
+ "<SPECIAL_577>",
581
+ "<SPECIAL_578>",
582
+ "<SPECIAL_579>",
583
+ "<SPECIAL_580>",
584
+ "<SPECIAL_581>",
585
+ "<SPECIAL_582>",
586
+ "<SPECIAL_583>",
587
+ "<SPECIAL_584>",
588
+ "<SPECIAL_585>",
589
+ "<SPECIAL_586>",
590
+ "<SPECIAL_587>",
591
+ "<SPECIAL_588>",
592
+ "<SPECIAL_589>",
593
+ "<SPECIAL_590>",
594
+ "<SPECIAL_591>",
595
+ "<SPECIAL_592>",
596
+ "<SPECIAL_593>",
597
+ "<SPECIAL_594>",
598
+ "<SPECIAL_595>",
599
+ "<SPECIAL_596>",
600
+ "<SPECIAL_597>",
601
+ "<SPECIAL_598>",
602
+ "<SPECIAL_599>",
603
+ "<SPECIAL_600>",
604
+ "<SPECIAL_601>",
605
+ "<SPECIAL_602>",
606
+ "<SPECIAL_603>",
607
+ "<SPECIAL_604>",
608
+ "<SPECIAL_605>",
609
+ "<SPECIAL_606>",
610
+ "<SPECIAL_607>",
611
+ "<SPECIAL_608>",
612
+ "<SPECIAL_609>",
613
+ "<SPECIAL_610>",
614
+ "<SPECIAL_611>",
615
+ "<SPECIAL_612>",
616
+ "<SPECIAL_613>",
617
+ "<SPECIAL_614>",
618
+ "<SPECIAL_615>",
619
+ "<SPECIAL_616>",
620
+ "<SPECIAL_617>",
621
+ "<SPECIAL_618>",
622
+ "<SPECIAL_619>",
623
+ "<SPECIAL_620>",
624
+ "<SPECIAL_621>",
625
+ "<SPECIAL_622>",
626
+ "<SPECIAL_623>",
627
+ "<SPECIAL_624>",
628
+ "<SPECIAL_625>",
629
+ "<SPECIAL_626>",
630
+ "<SPECIAL_627>",
631
+ "<SPECIAL_628>",
632
+ "<SPECIAL_629>",
633
+ "<SPECIAL_630>",
634
+ "<SPECIAL_631>",
635
+ "<SPECIAL_632>",
636
+ "<SPECIAL_633>",
637
+ "<SPECIAL_634>",
638
+ "<SPECIAL_635>",
639
+ "<SPECIAL_636>",
640
+ "<SPECIAL_637>",
641
+ "<SPECIAL_638>",
642
+ "<SPECIAL_639>",
643
+ "<SPECIAL_640>",
644
+ "<SPECIAL_641>",
645
+ "<SPECIAL_642>",
646
+ "<SPECIAL_643>",
647
+ "<SPECIAL_644>",
648
+ "<SPECIAL_645>",
649
+ "<SPECIAL_646>",
650
+ "<SPECIAL_647>",
651
+ "<SPECIAL_648>",
652
+ "<SPECIAL_649>",
653
+ "<SPECIAL_650>",
654
+ "<SPECIAL_651>",
655
+ "<SPECIAL_652>",
656
+ "<SPECIAL_653>",
657
+ "<SPECIAL_654>",
658
+ "<SPECIAL_655>",
659
+ "<SPECIAL_656>",
660
+ "<SPECIAL_657>",
661
+ "<SPECIAL_658>",
662
+ "<SPECIAL_659>",
663
+ "<SPECIAL_660>",
664
+ "<SPECIAL_661>",
665
+ "<SPECIAL_662>",
666
+ "<SPECIAL_663>",
667
+ "<SPECIAL_664>",
668
+ "<SPECIAL_665>",
669
+ "<SPECIAL_666>",
670
+ "<SPECIAL_667>",
671
+ "<SPECIAL_668>",
672
+ "<SPECIAL_669>",
673
+ "<SPECIAL_670>",
674
+ "<SPECIAL_671>",
675
+ "<SPECIAL_672>",
676
+ "<SPECIAL_673>",
677
+ "<SPECIAL_674>",
678
+ "<SPECIAL_675>",
679
+ "<SPECIAL_676>",
680
+ "<SPECIAL_677>",
681
+ "<SPECIAL_678>",
682
+ "<SPECIAL_679>",
683
+ "<SPECIAL_680>",
684
+ "<SPECIAL_681>",
685
+ "<SPECIAL_682>",
686
+ "<SPECIAL_683>",
687
+ "<SPECIAL_684>",
688
+ "<SPECIAL_685>",
689
+ "<SPECIAL_686>",
690
+ "<SPECIAL_687>",
691
+ "<SPECIAL_688>",
692
+ "<SPECIAL_689>",
693
+ "<SPECIAL_690>",
694
+ "<SPECIAL_691>",
695
+ "<SPECIAL_692>",
696
+ "<SPECIAL_693>",
697
+ "<SPECIAL_694>",
698
+ "<SPECIAL_695>",
699
+ "<SPECIAL_696>",
700
+ "<SPECIAL_697>",
701
+ "<SPECIAL_698>",
702
+ "<SPECIAL_699>",
703
+ "<SPECIAL_700>",
704
+ "<SPECIAL_701>",
705
+ "<SPECIAL_702>",
706
+ "<SPECIAL_703>",
707
+ "<SPECIAL_704>",
708
+ "<SPECIAL_705>",
709
+ "<SPECIAL_706>",
710
+ "<SPECIAL_707>",
711
+ "<SPECIAL_708>",
712
+ "<SPECIAL_709>",
713
+ "<SPECIAL_710>",
714
+ "<SPECIAL_711>",
715
+ "<SPECIAL_712>",
716
+ "<SPECIAL_713>",
717
+ "<SPECIAL_714>",
718
+ "<SPECIAL_715>",
719
+ "<SPECIAL_716>",
720
+ "<SPECIAL_717>",
721
+ "<SPECIAL_718>",
722
+ "<SPECIAL_719>",
723
+ "<SPECIAL_720>",
724
+ "<SPECIAL_721>",
725
+ "<SPECIAL_722>",
726
+ "<SPECIAL_723>",
727
+ "<SPECIAL_724>",
728
+ "<SPECIAL_725>",
729
+ "<SPECIAL_726>",
730
+ "<SPECIAL_727>",
731
+ "<SPECIAL_728>",
732
+ "<SPECIAL_729>",
733
+ "<SPECIAL_730>",
734
+ "<SPECIAL_731>",
735
+ "<SPECIAL_732>",
736
+ "<SPECIAL_733>",
737
+ "<SPECIAL_734>",
738
+ "<SPECIAL_735>",
739
+ "<SPECIAL_736>",
740
+ "<SPECIAL_737>",
741
+ "<SPECIAL_738>",
742
+ "<SPECIAL_739>",
743
+ "<SPECIAL_740>",
744
+ "<SPECIAL_741>",
745
+ "<SPECIAL_742>",
746
+ "<SPECIAL_743>",
747
+ "<SPECIAL_744>",
748
+ "<SPECIAL_745>",
749
+ "<SPECIAL_746>",
750
+ "<SPECIAL_747>",
751
+ "<SPECIAL_748>",
752
+ "<SPECIAL_749>",
753
+ "<SPECIAL_750>",
754
+ "<SPECIAL_751>",
755
+ "<SPECIAL_752>",
756
+ "<SPECIAL_753>",
757
+ "<SPECIAL_754>",
758
+ "<SPECIAL_755>",
759
+ "<SPECIAL_756>",
760
+ "<SPECIAL_757>",
761
+ "<SPECIAL_758>",
762
+ "<SPECIAL_759>",
763
+ "<SPECIAL_760>",
764
+ "<SPECIAL_761>",
765
+ "<SPECIAL_762>",
766
+ "<SPECIAL_763>",
767
+ "<SPECIAL_764>",
768
+ "<SPECIAL_765>",
769
+ "<SPECIAL_766>",
770
+ "<SPECIAL_767>",
771
+ "<SPECIAL_768>",
772
+ "<SPECIAL_769>",
773
+ "<SPECIAL_770>",
774
+ "<SPECIAL_771>",
775
+ "<SPECIAL_772>",
776
+ "<SPECIAL_773>",
777
+ "<SPECIAL_774>",
778
+ "<SPECIAL_775>",
779
+ "<SPECIAL_776>",
780
+ "<SPECIAL_777>",
781
+ "<SPECIAL_778>",
782
+ "<SPECIAL_779>",
783
+ "<SPECIAL_780>",
784
+ "<SPECIAL_781>",
785
+ "<SPECIAL_782>",
786
+ "<SPECIAL_783>",
787
+ "<SPECIAL_784>",
788
+ "<SPECIAL_785>",
789
+ "<SPECIAL_786>",
790
+ "<SPECIAL_787>",
791
+ "<SPECIAL_788>",
792
+ "<SPECIAL_789>",
793
+ "<SPECIAL_790>",
794
+ "<SPECIAL_791>",
795
+ "<SPECIAL_792>",
796
+ "<SPECIAL_793>",
797
+ "<SPECIAL_794>",
798
+ "<SPECIAL_795>",
799
+ "<SPECIAL_796>",
800
+ "<SPECIAL_797>",
801
+ "<SPECIAL_798>",
802
+ "<SPECIAL_799>",
803
+ "<SPECIAL_800>",
804
+ "<SPECIAL_801>",
805
+ "<SPECIAL_802>",
806
+ "<SPECIAL_803>",
807
+ "<SPECIAL_804>",
808
+ "<SPECIAL_805>",
809
+ "<SPECIAL_806>",
810
+ "<SPECIAL_807>",
811
+ "<SPECIAL_808>",
812
+ "<SPECIAL_809>",
813
+ "<SPECIAL_810>",
814
+ "<SPECIAL_811>",
815
+ "<SPECIAL_812>",
816
+ "<SPECIAL_813>",
817
+ "<SPECIAL_814>",
818
+ "<SPECIAL_815>",
819
+ "<SPECIAL_816>",
820
+ "<SPECIAL_817>",
821
+ "<SPECIAL_818>",
822
+ "<SPECIAL_819>",
823
+ "<SPECIAL_820>",
824
+ "<SPECIAL_821>",
825
+ "<SPECIAL_822>",
826
+ "<SPECIAL_823>",
827
+ "<SPECIAL_824>",
828
+ "<SPECIAL_825>",
829
+ "<SPECIAL_826>",
830
+ "<SPECIAL_827>",
831
+ "<SPECIAL_828>",
832
+ "<SPECIAL_829>",
833
+ "<SPECIAL_830>",
834
+ "<SPECIAL_831>",
835
+ "<SPECIAL_832>",
836
+ "<SPECIAL_833>",
837
+ "<SPECIAL_834>",
838
+ "<SPECIAL_835>",
839
+ "<SPECIAL_836>",
840
+ "<SPECIAL_837>",
841
+ "<SPECIAL_838>",
842
+ "<SPECIAL_839>",
843
+ "<SPECIAL_840>",
844
+ "<SPECIAL_841>",
845
+ "<SPECIAL_842>",
846
+ "<SPECIAL_843>",
847
+ "<SPECIAL_844>",
848
+ "<SPECIAL_845>",
849
+ "<SPECIAL_846>",
850
+ "<SPECIAL_847>",
851
+ "<SPECIAL_848>",
852
+ "<SPECIAL_849>",
853
+ "<SPECIAL_850>",
854
+ "<SPECIAL_851>",
855
+ "<SPECIAL_852>",
856
+ "<SPECIAL_853>",
857
+ "<SPECIAL_854>",
858
+ "<SPECIAL_855>",
859
+ "<SPECIAL_856>",
860
+ "<SPECIAL_857>",
861
+ "<SPECIAL_858>",
862
+ "<SPECIAL_859>",
863
+ "<SPECIAL_860>",
864
+ "<SPECIAL_861>",
865
+ "<SPECIAL_862>",
866
+ "<SPECIAL_863>",
867
+ "<SPECIAL_864>",
868
+ "<SPECIAL_865>",
869
+ "<SPECIAL_866>",
870
+ "<SPECIAL_867>",
871
+ "<SPECIAL_868>",
872
+ "<SPECIAL_869>",
873
+ "<SPECIAL_870>",
874
+ "<SPECIAL_871>",
875
+ "<SPECIAL_872>",
876
+ "<SPECIAL_873>",
877
+ "<SPECIAL_874>",
878
+ "<SPECIAL_875>",
879
+ "<SPECIAL_876>",
880
+ "<SPECIAL_877>",
881
+ "<SPECIAL_878>",
882
+ "<SPECIAL_879>",
883
+ "<SPECIAL_880>",
884
+ "<SPECIAL_881>",
885
+ "<SPECIAL_882>",
886
+ "<SPECIAL_883>",
887
+ "<SPECIAL_884>",
888
+ "<SPECIAL_885>",
889
+ "<SPECIAL_886>",
890
+ "<SPECIAL_887>",
891
+ "<SPECIAL_888>",
892
+ "<SPECIAL_889>",
893
+ "<SPECIAL_890>",
894
+ "<SPECIAL_891>",
895
+ "<SPECIAL_892>",
896
+ "<SPECIAL_893>",
897
+ "<SPECIAL_894>",
898
+ "<SPECIAL_895>",
899
+ "<SPECIAL_896>",
900
+ "<SPECIAL_897>",
901
+ "<SPECIAL_898>",
902
+ "<SPECIAL_899>",
903
+ "<SPECIAL_900>",
904
+ "<SPECIAL_901>",
905
+ "<SPECIAL_902>",
906
+ "<SPECIAL_903>",
907
+ "<SPECIAL_904>",
908
+ "<SPECIAL_905>",
909
+ "<SPECIAL_906>",
910
+ "<SPECIAL_907>",
911
+ "<SPECIAL_908>",
912
+ "<SPECIAL_909>",
913
+ "<SPECIAL_910>",
914
+ "<SPECIAL_911>",
915
+ "<SPECIAL_912>",
916
+ "<SPECIAL_913>",
917
+ "<SPECIAL_914>",
918
+ "<SPECIAL_915>",
919
+ "<SPECIAL_916>",
920
+ "<SPECIAL_917>",
921
+ "<SPECIAL_918>",
922
+ "<SPECIAL_919>",
923
+ "<SPECIAL_920>",
924
+ "<SPECIAL_921>",
925
+ "<SPECIAL_922>",
926
+ "<SPECIAL_923>",
927
+ "<SPECIAL_924>",
928
+ "<SPECIAL_925>",
929
+ "<SPECIAL_926>",
930
+ "<SPECIAL_927>",
931
+ "<SPECIAL_928>",
932
+ "<SPECIAL_929>",
933
+ "<SPECIAL_930>",
934
+ "<SPECIAL_931>",
935
+ "<SPECIAL_932>",
936
+ "<SPECIAL_933>",
937
+ "<SPECIAL_934>",
938
+ "<SPECIAL_935>",
939
+ "<SPECIAL_936>",
940
+ "<SPECIAL_937>",
941
+ "<SPECIAL_938>",
942
+ "<SPECIAL_939>",
943
+ "<SPECIAL_940>",
944
+ "<SPECIAL_941>",
945
+ "<SPECIAL_942>",
946
+ "<SPECIAL_943>",
947
+ "<SPECIAL_944>",
948
+ "<SPECIAL_945>",
949
+ "<SPECIAL_946>",
950
+ "<SPECIAL_947>",
951
+ "<SPECIAL_948>",
952
+ "<SPECIAL_949>",
953
+ "<SPECIAL_950>",
954
+ "<SPECIAL_951>",
955
+ "<SPECIAL_952>",
956
+ "<SPECIAL_953>",
957
+ "<SPECIAL_954>",
958
+ "<SPECIAL_955>",
959
+ "<SPECIAL_956>",
960
+ "<SPECIAL_957>",
961
+ "<SPECIAL_958>",
962
+ "<SPECIAL_959>",
963
+ "<SPECIAL_960>",
964
+ "<SPECIAL_961>",
965
+ "<SPECIAL_962>",
966
+ "<SPECIAL_963>",
967
+ "<SPECIAL_964>",
968
+ "<SPECIAL_965>",
969
+ "<SPECIAL_966>",
970
+ "<SPECIAL_967>",
971
+ "<SPECIAL_968>",
972
+ "<SPECIAL_969>",
973
+ "<SPECIAL_970>",
974
+ "<SPECIAL_971>",
975
+ "<SPECIAL_972>",
976
+ "<SPECIAL_973>",
977
+ "<SPECIAL_974>",
978
+ "<SPECIAL_975>",
979
+ "<SPECIAL_976>",
980
+ "<SPECIAL_977>",
981
+ "<SPECIAL_978>",
982
+ "<SPECIAL_979>",
983
+ "<SPECIAL_980>",
984
+ "<SPECIAL_981>",
985
+ "<SPECIAL_982>",
986
+ "<SPECIAL_983>",
987
+ "<SPECIAL_984>",
988
+ "<SPECIAL_985>",
989
+ "<SPECIAL_986>",
990
+ "<SPECIAL_987>",
991
+ "<SPECIAL_988>",
992
+ "<SPECIAL_989>",
993
+ "<SPECIAL_990>",
994
+ "<SPECIAL_991>",
995
+ "<SPECIAL_992>",
996
+ "<SPECIAL_993>",
997
+ "<SPECIAL_994>",
998
+ "<SPECIAL_995>",
999
+ "<SPECIAL_996>",
1000
+ "<SPECIAL_997>",
1001
+ "<SPECIAL_998>",
1002
+ "<SPECIAL_999>"
1003
+ ],
1004
+ "bos_token": {
1005
+ "content": "<s>",
1006
+ "lstrip": false,
1007
+ "normalized": false,
1008
+ "rstrip": false,
1009
+ "single_word": false
1010
+ },
1011
+ "eos_token": {
1012
+ "content": "</s>",
1013
+ "lstrip": false,
1014
+ "normalized": false,
1015
+ "rstrip": false,
1016
+ "single_word": false
1017
+ },
1018
+ "pad_token": {
1019
+ "content": "<pad>",
1020
+ "lstrip": false,
1021
+ "normalized": false,
1022
+ "rstrip": false,
1023
+ "single_word": false
1024
+ },
1025
+ "unk_token": {
1026
+ "content": "<unk>",
1027
+ "lstrip": false,
1028
+ "normalized": false,
1029
+ "rstrip": false,
1030
+ "single_word": false
1031
+ }
1032
+ }
tekken.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e2501687ccd0e1f30f36319eaf2b46958b897811e246cd8eb5d385b9e3de7d1
3
+ size 19399895
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23ad081f384b2bdb3c97f6e5461dfce52c1174c7328854a55006988f0fef9da7
3
+ size 17078019
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff