{ "add_prefix_space": false, "added_tokens_decoder": { "0": { "content": "<|endoftext|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "32768": { "content": "<|im_start|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "32769": { "content": "<|im_end|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "32770": { "content": "", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": false }, "32771": { "content": "", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": false } }, "additional_special_tokens": [ "<|im_start|>", "<|im_end|>" ], "bos_token": null, "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null, "chat_template": "{%- if messages[0].role == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0].content + '<|im_end|>\\n' }}\n{%- endif %}\n{%- for message in messages %}\n {%- if message.content is string %}\n {%- set content = message.content %}\n {%- else %}\n {%- set content = '' %}\n {%- endif %}\n {%- if message.role == \"user\" %}\n {{- '<|im_start|>user\\n' + content + '<|im_end|>\\n' }}\n {%- elif message.role == \"assistant\" %}\n {%- set has_think = false %}\n {%- set think_content = '' %}\n {%- set main_content = content %}\n \n {%- if '' in content %}\n {%- set think_content = content.split('')[0].rstrip('\\n').split('')[-1].lstrip('\\n') %}\n {%- set main_content = content.split('')[-1].lstrip('\\n') %}\n {%- set has_think = true %}\n {%- endif %}\n \n {{- '<|im_start|>assistant\\n' }}\n {%- if has_think %}\n {{- '\\n' + think_content + '\\n\\n\\n' + main_content }}\n {%- else %}\n {{- content }}\n {%- endif %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}", "clean_up_tokenization_spaces": false, "errors": "replace", "model_max_length": 32768, "split_special_tokens": false, "tokenizer_class": "PreTrainedTokenizerFast", "add_bos_token": false }