Upload tokenizer
Browse files- chat_template.jinja +2 -2
- special_tokens_map.json +1 -1
- tokenizer_config.json +2 -2
chat_template.jinja
CHANGED
|
@@ -3,7 +3,7 @@
|
|
| 3 |
{%- if messages[0]['role'] == 'system' %}
|
| 4 |
{{- messages[0]['content'] }}
|
| 5 |
{%- else %}
|
| 6 |
-
{{- 'You are a helpful assistant.' }}
|
| 7 |
{%- endif %}
|
| 8 |
{{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
|
| 9 |
{%- for tool in tools %}
|
|
@@ -15,7 +15,7 @@
|
|
| 15 |
{%- if messages[0]['role'] == 'system' %}
|
| 16 |
{{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
|
| 17 |
{%- else %}
|
| 18 |
-
{{- '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}
|
| 19 |
{%- endif %}
|
| 20 |
{%- endif %}
|
| 21 |
{%- for message in messages %}
|
|
|
|
| 3 |
{%- if messages[0]['role'] == 'system' %}
|
| 4 |
{{- messages[0]['content'] }}
|
| 5 |
{%- else %}
|
| 6 |
+
{{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}
|
| 7 |
{%- endif %}
|
| 8 |
{{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
|
| 9 |
{%- for tool in tools %}
|
|
|
|
| 15 |
{%- if messages[0]['role'] == 'system' %}
|
| 16 |
{{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
|
| 17 |
{%- else %}
|
| 18 |
+
{{- '<|im_start|>system\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\n' }}
|
| 19 |
{%- endif %}
|
| 20 |
{%- endif %}
|
| 21 |
{%- for message in messages %}
|
special_tokens_map.json
CHANGED
|
@@ -15,7 +15,7 @@
|
|
| 15 |
"<|video_pad|>"
|
| 16 |
],
|
| 17 |
"eos_token": {
|
| 18 |
-
"content": "<|
|
| 19 |
"lstrip": false,
|
| 20 |
"normalized": false,
|
| 21 |
"rstrip": false,
|
|
|
|
| 15 |
"<|video_pad|>"
|
| 16 |
],
|
| 17 |
"eos_token": {
|
| 18 |
+
"content": "<|im_end|>",
|
| 19 |
"lstrip": false,
|
| 20 |
"normalized": false,
|
| 21 |
"rstrip": false,
|
tokenizer_config.json
CHANGED
|
@@ -196,10 +196,10 @@
|
|
| 196 |
],
|
| 197 |
"bos_token": null,
|
| 198 |
"clean_up_tokenization_spaces": false,
|
| 199 |
-
"eos_token": "<|
|
| 200 |
"errors": "replace",
|
| 201 |
"extra_special_tokens": {},
|
| 202 |
-
"model_max_length":
|
| 203 |
"pad_token": "<|endoftext|>",
|
| 204 |
"split_special_tokens": false,
|
| 205 |
"tokenizer_class": "Qwen2Tokenizer",
|
|
|
|
| 196 |
],
|
| 197 |
"bos_token": null,
|
| 198 |
"clean_up_tokenization_spaces": false,
|
| 199 |
+
"eos_token": "<|im_end|>",
|
| 200 |
"errors": "replace",
|
| 201 |
"extra_special_tokens": {},
|
| 202 |
+
"model_max_length": 131072,
|
| 203 |
"pad_token": "<|endoftext|>",
|
| 204 |
"split_special_tokens": false,
|
| 205 |
"tokenizer_class": "Qwen2Tokenizer",
|