elribonazo commited on
Commit
dbef287
·
verified ·
1 Parent(s): e1c8b22

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,29 +1,24 @@
1
  {
2
  "_attn_implementation_autoset": true,
3
- "_name_or_path": "nguyenthanhthuan/Llama_3.2_1B_Intruct_Tool_Calling_V2",
4
  "architectures": [
5
  "LlamaForCausalLM"
6
  ],
7
  "attention_bias": false,
8
  "attention_dropout": 0.0,
9
  "bos_token_id": 128000,
10
- "eos_token_id": [
11
- 128001,
12
- 128008,
13
- 128009
14
- ],
15
- "head_dim": 64,
16
  "hidden_act": "silu",
17
- "hidden_size": 2048,
18
  "initializer_range": 0.02,
19
  "intermediate_size": 8192,
20
  "max_position_embeddings": 131072,
21
  "mlp_bias": false,
22
  "model_type": "llama",
23
- "num_attention_heads": 32,
24
- "num_hidden_layers": 16,
25
  "num_key_value_heads": 8,
26
- "pad_token_id": 128004,
27
  "pretraining_tp": 1,
28
  "rms_norm_eps": 1e-05,
29
  "rope_scaling": {
@@ -36,7 +31,6 @@
36
  "rope_theta": 500000.0,
37
  "tie_word_embeddings": true,
38
  "transformers_version": "4.46.3",
39
- "unsloth_version": "2024.10.7",
40
  "use_cache": true,
41
  "vocab_size": 128256
42
  }
 
1
  {
2
  "_attn_implementation_autoset": true,
3
+ "_name_or_path": "NousResearch/Hermes-3-Llama-3.2-3B",
4
  "architectures": [
5
  "LlamaForCausalLM"
6
  ],
7
  "attention_bias": false,
8
  "attention_dropout": 0.0,
9
  "bos_token_id": 128000,
10
+ "eos_token_id": 128039,
11
+ "head_dim": 128,
 
 
 
 
12
  "hidden_act": "silu",
13
+ "hidden_size": 3072,
14
  "initializer_range": 0.02,
15
  "intermediate_size": 8192,
16
  "max_position_embeddings": 131072,
17
  "mlp_bias": false,
18
  "model_type": "llama",
19
+ "num_attention_heads": 24,
20
+ "num_hidden_layers": 28,
21
  "num_key_value_heads": 8,
 
22
  "pretraining_tp": 1,
23
  "rms_norm_eps": 1e-05,
24
  "rope_scaling": {
 
31
  "rope_theta": 500000.0,
32
  "tie_word_embeddings": true,
33
  "transformers_version": "4.46.3",
 
34
  "use_cache": true,
35
  "vocab_size": 128256
36
  }
generation_config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "bos_token_id": 128000,
3
  "do_sample": true,
4
  "eos_token_id": [
@@ -6,8 +7,6 @@
6
  128008,
7
  128009
8
  ],
9
- "max_length": 131072,
10
- "pad_token_id": 128004,
11
  "temperature": 0.6,
12
  "top_p": 0.9,
13
  "transformers_version": "4.46.3"
 
1
  {
2
+ "_from_model_config": true,
3
  "bos_token_id": 128000,
4
  "do_sample": true,
5
  "eos_token_id": [
 
7
  128008,
8
  128009
9
  ],
 
 
10
  "temperature": 0.6,
11
  "top_p": 0.9,
12
  "transformers_version": "4.46.3"
onnx/model.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4be0a1c8367f0f08fa30f806a6703e403d88a8f864dbbb24adf920799a132dcf
3
- size 448765
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5d41a9ea95192f86edd67138737ce7ad0fc0d86853a6d5facce1f0ed10041c6
3
+ size 776375
onnx/model.onnx_data CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c4f269b9750ade29e17b6daec21c5d76bf2f0c1206b9e8e1090a2ad2d3b1ef95
3
- size 4943257600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e32108d9e31e081988686862455045e14bb58e868ec76e552a9edcd8367996b
3
+ size 12850999296
onnx/model_bnb4.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:973e5064cbd1cffa31cd30c9a1371cc648af2831391fdbf7768cd9419e7d1c00
3
- size 1598757277
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3028fefc768fbea550846e7e26e7a9d2f07801158f0e7c43d4a5591bb63b6347
3
+ size 822344
onnx/model_bnb4.onnx_data CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fd39711b03ac0b3460bbb30b0f0cec6e8ef82acf58f88bb2110c851fb4c8ecd8
3
- size 6323781632
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ae8301ff4d0d8ca41d75596769753f408add40db29558f0c4f87024b0f75d3e
3
+ size 3162157056
onnx/model_fp16.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a2f337289d0805d9807d78c4194f2dbc1d3dfb4d38e1447f8134d76284b3711
3
- size 457645
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:602f3d0ed7296cfc7c3ea4e1a0e3e3ba1eb0b8cbada03e1519aecdaaa1f324d3
3
+ size 791765
onnx/model_fp16.onnx_data CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c9733aab730ec38ed1ed2b8e85ba96773627c212e41c4faec2771f3163968ad1
3
- size 2471628800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5b07b6bdd0967f15120ecb05ecb75ce1b7ef7bb9f39b66f0c315688340810e9
3
+ size 6425499648
onnx/model_int8.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a392606574450b062054554547f0b8d9c8b77b1678417107e37d007e31acc5b3
3
- size 1236578925
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1a0ce71408e06e3d2395ad382b45a961508a99fcb0e4d6f4af910c0e1ef975d
3
+ size 991648
onnx/model_int8.onnx_data CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:89ee296e92fdd6fb0cdcc3d13f43a24d7a1a28c37d546a7c98392c013b4e186b
3
- size 8031059968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:657f76d3174d98bdec5b911858a5cade2863a345eb391bd97b79dd6e8deb2e84
3
+ size 3213275136
onnx/model_q4.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5aa75b94996669549ce465cbe97688b269e438d4986e665695a7e98a89927a74
3
- size 1659573821
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80b807bdfa953bb48484bdcd13fd917f5ef9fbdb4b511c3037c18aac71edd5f8
3
+ size 819792
onnx/model_q4.onnx_data CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:460a97e29c395098e40630ae977e9a8f1158e1b474c9209e4ad0e46334afa2a4
3
- size 6792822784
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96e6b616d56af52e3014c1bf52b870b5fd5fb32f6df66614eb1eb4a0401eb949
3
+ size 3338317824
onnx/model_q4f16.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2f678c74e7eab4dac069eb38d0575cef3571205c12531fee6cd3d30341735a54
3
- size 1073292786
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9653d2bb65a19201c6b3aa760cc26cd748179d641d95b08bde9817876ac204d7
3
+ size 835365
onnx/model_q4f16.onnx_data CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:73694682378ff20d4d65c680ee98b5415a0f74d96d3437db644b702ff05833c9
3
- size 5272576000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55cb782dce22fcccbd96d2dfa2905246f79412edfc99ebd9d774ac442e0c4f29
3
+ size 2373801984
onnx/model_quantized.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a392606574450b062054554547f0b8d9c8b77b1678417107e37d007e31acc5b3
3
- size 1236578925
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cc3735938054c404db34b6281455df06be1eef9574a3abf8093a3402dd2cd73
3
+ size 992946
onnx/model_quantized.onnx_data CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:89ee296e92fdd6fb0cdcc3d13f43a24d7a1a28c37d546a7c98392c013b4e186b
3
- size 8031059968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:657f76d3174d98bdec5b911858a5cade2863a345eb391bd97b79dd6e8deb2e84
3
+ size 3213275136
onnx/model_uint8.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:607c40015dceb3d5970af129484bce2f8dba78f3ad10881ea4586d20a850cfc6
3
- size 1236578987
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:811d5dd6980f7fef1f903a9ba8eff4459a10e2608fc3f36824bea6debf1d91f5
3
+ size 991999
onnx/model_uint8.onnx_data CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d881b50aa33ca4d9b531031029fc3b40c677033299addbf58f5ee1c468c85e43
3
- size 8031059968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e7b640fee0a83fb4cc57cd0a3c6908fc934e336d1f5f8b5fe6dea70da7a0c15
3
+ size 3213275136
special_tokens_map.json CHANGED
@@ -7,14 +7,14 @@
7
  "single_word": false
8
  },
9
  "eos_token": {
10
- "content": "<|eot_id|>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
  "pad_token": {
17
- "content": "<|finetune_right_pad_id|>",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
 
7
  "single_word": false
8
  },
9
  "eos_token": {
10
+ "content": "<|im_end|>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
  "pad_token": {
17
+ "content": "<|end_of_text|>",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
3
- size 17209920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f908f9b84390fd12c6d0c356765257846c53f60bf472ff4996a440a1e230373
3
+ size 17209403
tokenizer_config.json CHANGED
@@ -17,20 +17,20 @@
17
  "special": true
18
  },
19
  "128002": {
20
- "content": "<|reserved_special_token_0|>",
21
  "lstrip": false,
22
  "normalized": false,
23
  "rstrip": false,
24
  "single_word": false,
25
- "special": true
26
  },
27
  "128003": {
28
- "content": "<|reserved_special_token_1|>",
29
  "lstrip": false,
30
  "normalized": false,
31
  "rstrip": false,
32
  "single_word": false,
33
- "special": true
34
  },
35
  "128004": {
36
  "content": "<|finetune_right_pad_id|>",
@@ -89,231 +89,231 @@
89
  "special": true
90
  },
91
  "128011": {
92
- "content": "<|reserved_special_token_3|>",
93
  "lstrip": false,
94
  "normalized": false,
95
  "rstrip": false,
96
  "single_word": false,
97
- "special": true
98
  },
99
  "128012": {
100
- "content": "<|reserved_special_token_4|>",
101
  "lstrip": false,
102
  "normalized": false,
103
  "rstrip": false,
104
  "single_word": false,
105
- "special": true
106
  },
107
  "128013": {
108
- "content": "<|reserved_special_token_5|>",
109
  "lstrip": false,
110
  "normalized": false,
111
  "rstrip": false,
112
  "single_word": false,
113
- "special": true
114
  },
115
  "128014": {
116
- "content": "<|reserved_special_token_6|>",
117
  "lstrip": false,
118
  "normalized": false,
119
  "rstrip": false,
120
  "single_word": false,
121
- "special": true
122
  },
123
  "128015": {
124
- "content": "<|reserved_special_token_7|>",
125
  "lstrip": false,
126
  "normalized": false,
127
  "rstrip": false,
128
  "single_word": false,
129
- "special": true
130
  },
131
  "128016": {
132
- "content": "<|reserved_special_token_8|>",
133
  "lstrip": false,
134
  "normalized": false,
135
  "rstrip": false,
136
  "single_word": false,
137
- "special": true
138
  },
139
  "128017": {
140
- "content": "<|reserved_special_token_9|>",
141
  "lstrip": false,
142
  "normalized": false,
143
  "rstrip": false,
144
  "single_word": false,
145
- "special": true
146
  },
147
  "128018": {
148
- "content": "<|reserved_special_token_10|>",
149
  "lstrip": false,
150
  "normalized": false,
151
  "rstrip": false,
152
  "single_word": false,
153
- "special": true
154
  },
155
  "128019": {
156
- "content": "<|reserved_special_token_11|>",
157
  "lstrip": false,
158
  "normalized": false,
159
  "rstrip": false,
160
  "single_word": false,
161
- "special": true
162
  },
163
  "128020": {
164
- "content": "<|reserved_special_token_12|>",
165
  "lstrip": false,
166
  "normalized": false,
167
  "rstrip": false,
168
  "single_word": false,
169
- "special": true
170
  },
171
  "128021": {
172
- "content": "<|reserved_special_token_13|>",
173
  "lstrip": false,
174
  "normalized": false,
175
  "rstrip": false,
176
  "single_word": false,
177
- "special": true
178
  },
179
  "128022": {
180
- "content": "<|reserved_special_token_14|>",
181
  "lstrip": false,
182
  "normalized": false,
183
  "rstrip": false,
184
  "single_word": false,
185
- "special": true
186
  },
187
  "128023": {
188
- "content": "<|reserved_special_token_15|>",
189
  "lstrip": false,
190
  "normalized": false,
191
  "rstrip": false,
192
  "single_word": false,
193
- "special": true
194
  },
195
  "128024": {
196
- "content": "<|reserved_special_token_16|>",
197
  "lstrip": false,
198
  "normalized": false,
199
  "rstrip": false,
200
  "single_word": false,
201
- "special": true
202
  },
203
  "128025": {
204
- "content": "<|reserved_special_token_17|>",
205
  "lstrip": false,
206
  "normalized": false,
207
  "rstrip": false,
208
  "single_word": false,
209
- "special": true
210
  },
211
  "128026": {
212
- "content": "<|reserved_special_token_18|>",
213
  "lstrip": false,
214
  "normalized": false,
215
  "rstrip": false,
216
  "single_word": false,
217
- "special": true
218
  },
219
  "128027": {
220
- "content": "<|reserved_special_token_19|>",
221
  "lstrip": false,
222
  "normalized": false,
223
  "rstrip": false,
224
  "single_word": false,
225
- "special": true
226
  },
227
  "128028": {
228
- "content": "<|reserved_special_token_20|>",
229
  "lstrip": false,
230
  "normalized": false,
231
  "rstrip": false,
232
  "single_word": false,
233
- "special": true
234
  },
235
  "128029": {
236
- "content": "<|reserved_special_token_21|>",
237
  "lstrip": false,
238
  "normalized": false,
239
  "rstrip": false,
240
  "single_word": false,
241
- "special": true
242
  },
243
  "128030": {
244
- "content": "<|reserved_special_token_22|>",
245
  "lstrip": false,
246
  "normalized": false,
247
  "rstrip": false,
248
  "single_word": false,
249
- "special": true
250
  },
251
  "128031": {
252
- "content": "<|reserved_special_token_23|>",
253
  "lstrip": false,
254
  "normalized": false,
255
  "rstrip": false,
256
  "single_word": false,
257
- "special": true
258
  },
259
  "128032": {
260
- "content": "<|reserved_special_token_24|>",
261
  "lstrip": false,
262
  "normalized": false,
263
  "rstrip": false,
264
  "single_word": false,
265
- "special": true
266
  },
267
  "128033": {
268
- "content": "<|reserved_special_token_25|>",
269
  "lstrip": false,
270
  "normalized": false,
271
  "rstrip": false,
272
  "single_word": false,
273
- "special": true
274
  },
275
  "128034": {
276
- "content": "<|reserved_special_token_26|>",
277
  "lstrip": false,
278
  "normalized": false,
279
  "rstrip": false,
280
  "single_word": false,
281
- "special": true
282
  },
283
  "128035": {
284
- "content": "<|reserved_special_token_27|>",
285
  "lstrip": false,
286
  "normalized": false,
287
  "rstrip": false,
288
  "single_word": false,
289
- "special": true
290
  },
291
  "128036": {
292
- "content": "<|reserved_special_token_28|>",
293
  "lstrip": false,
294
  "normalized": false,
295
  "rstrip": false,
296
  "single_word": false,
297
- "special": true
298
  },
299
  "128037": {
300
- "content": "<|reserved_special_token_29|>",
301
  "lstrip": false,
302
  "normalized": false,
303
  "rstrip": false,
304
  "single_word": false,
305
- "special": true
306
  },
307
  "128038": {
308
- "content": "<|reserved_special_token_30|>",
309
  "lstrip": false,
310
  "normalized": false,
311
  "rstrip": false,
312
  "single_word": false,
313
- "special": true
314
  },
315
  "128039": {
316
- "content": "<|reserved_special_token_31|>",
317
  "lstrip": false,
318
  "normalized": false,
319
  "rstrip": false,
@@ -321,12 +321,12 @@
321
  "special": true
322
  },
323
  "128040": {
324
- "content": "<|reserved_special_token_32|>",
325
  "lstrip": false,
326
  "normalized": false,
327
  "rstrip": false,
328
  "single_word": false,
329
- "special": true
330
  },
331
  "128041": {
332
  "content": "<|reserved_special_token_33|>",
@@ -2050,15 +2050,14 @@
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
2053
- "chat_template": "\n{%- if messages[0]['role'] == 'system' -%}\n<|start_header_id|>system<|end_header_id|>\n\n{{ messages[0]['content'] }}\n{%- endif -%}\n{%- for message in messages -%}\n {%- if message['role'] == 'user' -%}\n <|start_header_id|>user<|end_header_id|>\n {{ message['content'] }}\n {%- elif message['role'] in ['assistant'] -%}\n <|start_header_id|>assistant<|end_header_id|>\n {{ message['content'] }}<|eot_id|>\n {%- elif message['role'] in ['tool_call'] -%}\n <|start_header_id|>tool_call<|end_header_id|>\n {{ message['content'] }}<|eot_id|>\n {%- elif message['role'] == 'tool_response' -%}\n <|start_header_id|>tool_respone<|end_header_id|>\n {{ message['content'] }}\n {%- endif -%}\n{%- endfor -%}\n{%- if add_generation_prompt -%} <|start_header_id|>assistant<|end_header_id|>\n{%- endif -%}\n",
2054
  "clean_up_tokenization_spaces": true,
2055
- "eos_token": "<|eot_id|>",
2056
  "model_input_names": [
2057
  "input_ids",
2058
  "attention_mask"
2059
  ],
2060
  "model_max_length": 131072,
2061
- "pad_token": "<|finetune_right_pad_id|>",
2062
- "padding_side": "left",
2063
  "tokenizer_class": "PreTrainedTokenizerFast"
2064
  }
 
17
  "special": true
18
  },
19
  "128002": {
20
+ "content": "<tool_call>",
21
  "lstrip": false,
22
  "normalized": false,
23
  "rstrip": false,
24
  "single_word": false,
25
+ "special": false
26
  },
27
  "128003": {
28
+ "content": "<tool_response>",
29
  "lstrip": false,
30
  "normalized": false,
31
  "rstrip": false,
32
  "single_word": false,
33
+ "special": false
34
  },
35
  "128004": {
36
  "content": "<|finetune_right_pad_id|>",
 
89
  "special": true
90
  },
91
  "128011": {
92
+ "content": "<tools>",
93
  "lstrip": false,
94
  "normalized": false,
95
  "rstrip": false,
96
  "single_word": false,
97
+ "special": false
98
  },
99
  "128012": {
100
+ "content": "</tools>",
101
  "lstrip": false,
102
  "normalized": false,
103
  "rstrip": false,
104
  "single_word": false,
105
+ "special": false
106
  },
107
  "128013": {
108
+ "content": "</tool_call>",
109
  "lstrip": false,
110
  "normalized": false,
111
  "rstrip": false,
112
  "single_word": false,
113
+ "special": false
114
  },
115
  "128014": {
116
+ "content": "</tool_response>",
117
  "lstrip": false,
118
  "normalized": false,
119
  "rstrip": false,
120
  "single_word": false,
121
+ "special": false
122
  },
123
  "128015": {
124
+ "content": "<schema>",
125
  "lstrip": false,
126
  "normalized": false,
127
  "rstrip": false,
128
  "single_word": false,
129
+ "special": false
130
  },
131
  "128016": {
132
+ "content": "</schema>",
133
  "lstrip": false,
134
  "normalized": false,
135
  "rstrip": false,
136
  "single_word": false,
137
+ "special": false
138
  },
139
  "128017": {
140
+ "content": "<scratch_pad>",
141
  "lstrip": false,
142
  "normalized": false,
143
  "rstrip": false,
144
  "single_word": false,
145
+ "special": false
146
  },
147
  "128018": {
148
+ "content": "</scratch_pad>",
149
  "lstrip": false,
150
  "normalized": false,
151
  "rstrip": false,
152
  "single_word": false,
153
+ "special": false
154
  },
155
  "128019": {
156
+ "content": "<SCRATCHPAD>",
157
  "lstrip": false,
158
  "normalized": false,
159
  "rstrip": false,
160
  "single_word": false,
161
+ "special": false
162
  },
163
  "128020": {
164
+ "content": "</SCRATCHPAD>",
165
  "lstrip": false,
166
  "normalized": false,
167
  "rstrip": false,
168
  "single_word": false,
169
+ "special": false
170
  },
171
  "128021": {
172
+ "content": "<REASONING>",
173
  "lstrip": false,
174
  "normalized": false,
175
  "rstrip": false,
176
  "single_word": false,
177
+ "special": false
178
  },
179
  "128022": {
180
+ "content": "</REASONING>",
181
  "lstrip": false,
182
  "normalized": false,
183
  "rstrip": false,
184
  "single_word": false,
185
+ "special": false
186
  },
187
  "128023": {
188
+ "content": "<INNER_MONOLOGUE>",
189
  "lstrip": false,
190
  "normalized": false,
191
  "rstrip": false,
192
  "single_word": false,
193
+ "special": false
194
  },
195
  "128024": {
196
+ "content": "</INNER_MONOLOGUE>",
197
  "lstrip": false,
198
  "normalized": false,
199
  "rstrip": false,
200
  "single_word": false,
201
+ "special": false
202
  },
203
  "128025": {
204
+ "content": "<PLAN>",
205
  "lstrip": false,
206
  "normalized": false,
207
  "rstrip": false,
208
  "single_word": false,
209
+ "special": false
210
  },
211
  "128026": {
212
+ "content": "</PLAN>",
213
  "lstrip": false,
214
  "normalized": false,
215
  "rstrip": false,
216
  "single_word": false,
217
+ "special": false
218
  },
219
  "128027": {
220
+ "content": "<EXECUTION>",
221
  "lstrip": false,
222
  "normalized": false,
223
  "rstrip": false,
224
  "single_word": false,
225
+ "special": false
226
  },
227
  "128028": {
228
+ "content": "</EXECUTION>",
229
  "lstrip": false,
230
  "normalized": false,
231
  "rstrip": false,
232
  "single_word": false,
233
+ "special": false
234
  },
235
  "128029": {
236
+ "content": "<REFLECTION>",
237
  "lstrip": false,
238
  "normalized": false,
239
  "rstrip": false,
240
  "single_word": false,
241
+ "special": false
242
  },
243
  "128030": {
244
+ "content": "</REFLECTION>",
245
  "lstrip": false,
246
  "normalized": false,
247
  "rstrip": false,
248
  "single_word": false,
249
+ "special": false
250
  },
251
  "128031": {
252
+ "content": "<THINKING>",
253
  "lstrip": false,
254
  "normalized": false,
255
  "rstrip": false,
256
  "single_word": false,
257
+ "special": false
258
  },
259
  "128032": {
260
+ "content": "</THINKING>",
261
  "lstrip": false,
262
  "normalized": false,
263
  "rstrip": false,
264
  "single_word": false,
265
+ "special": false
266
  },
267
  "128033": {
268
+ "content": "<SOLUTION>",
269
  "lstrip": false,
270
  "normalized": false,
271
  "rstrip": false,
272
  "single_word": false,
273
+ "special": false
274
  },
275
  "128034": {
276
+ "content": "</SOLUTION>",
277
  "lstrip": false,
278
  "normalized": false,
279
  "rstrip": false,
280
  "single_word": false,
281
+ "special": false
282
  },
283
  "128035": {
284
+ "content": "<EXPLANATION>",
285
  "lstrip": false,
286
  "normalized": false,
287
  "rstrip": false,
288
  "single_word": false,
289
+ "special": false
290
  },
291
  "128036": {
292
+ "content": "</EXPLANATION>",
293
  "lstrip": false,
294
  "normalized": false,
295
  "rstrip": false,
296
  "single_word": false,
297
+ "special": false
298
  },
299
  "128037": {
300
+ "content": "<UNIT_TEST>",
301
  "lstrip": false,
302
  "normalized": false,
303
  "rstrip": false,
304
  "single_word": false,
305
+ "special": false
306
  },
307
  "128038": {
308
+ "content": "</UNIT_TEST>",
309
  "lstrip": false,
310
  "normalized": false,
311
  "rstrip": false,
312
  "single_word": false,
313
+ "special": false
314
  },
315
  "128039": {
316
+ "content": "<|im_end|>",
317
  "lstrip": false,
318
  "normalized": false,
319
  "rstrip": false,
 
321
  "special": true
322
  },
323
  "128040": {
324
+ "content": "<|im_start|>",
325
  "lstrip": false,
326
  "normalized": false,
327
  "rstrip": false,
328
  "single_word": false,
329
+ "special": false
330
  },
331
  "128041": {
332
  "content": "<|reserved_special_token_33|>",
 
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
2053
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
2054
  "clean_up_tokenization_spaces": true,
2055
+ "eos_token": "<|im_end|>",
2056
  "model_input_names": [
2057
  "input_ids",
2058
  "attention_mask"
2059
  ],
2060
  "model_max_length": 131072,
2061
+ "pad_token": "<|end_of_text|>",
 
2062
  "tokenizer_class": "PreTrainedTokenizerFast"
2063
  }