Fix: add chat_template required for OpenAI-compatible chat/completions
Browse files- tokenizer_config.json +2 -1
tokenizer_config.json
CHANGED
|
@@ -234,5 +234,6 @@
|
|
| 234 |
"pad_token": "<|endoftext|>",
|
| 235 |
"split_special_tokens": false,
|
| 236 |
"tokenizer_class": "Qwen2Tokenizer",
|
| 237 |
-
"unk_token": null
|
|
|
|
| 238 |
}
|
|
|
|
| 234 |
"pad_token": "<|endoftext|>",
|
| 235 |
"split_special_tokens": false,
|
| 236 |
"tokenizer_class": "Qwen2Tokenizer",
|
| 237 |
+
"unk_token": null,
|
| 238 |
+
"chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% set loop_messages = messages[1:] %}{% else %}{% set system_message = '' %}{% set loop_messages = messages %}{% endif %}{% if system_message != '' %}<|im_start|>system\n{{ system_message }}<|im_end|>\n{% endif %}{% for message in loop_messages %}<|im_start|>{{ message['role'] }}\n{{ message['content'] }}<|im_end|>\n{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant\n{% endif %}"
|
| 239 |
}
|