{ "tokenizer_class": "GPT2TokenizerFast", "tokenizer_file": "tokenizer.json", "model_max_length": 1024, "add_prefix_space": true, "clean_up_tokenization_spaces": false, "bos_token": "<|begin_of_text|>", "eos_token": "<|eot_id|>", "eot_token": "<|eot_id|>", "pad_token": "<|pad|>", "unk_token": "<|unk|>", "additional_special_tokens": [ "<|start_header_id|>", "<|end_header_id|>", "<|eot_id|>" ], "added_tokens_decoder": { "0": { "content": "<|pad|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "1": { "content": "<|unk|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "2": { "content": "<|begin_of_text|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "3": { "content": "<|end_of_text|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "4": { "content": "<|start_header_id|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "5": { "content": "<|end_header_id|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "6": { "content": "<|eot_id|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true } }, "chat_template": "{% for message in messages %}{% if loop.first %}<|begin_of_text|>{% endif %}<|start_header_id|>{{ message['role'] }}<|end_header_id|>\n\n{{ message['content'] }}<|eot_id|>{% endfor %}{% if add_generation_prompt %}<|start_header_id|>assistant<|end_header_id|>\n\n{% endif %}" }