brandon-tiny-10m-instruct / tokenizer_config.json
xaskasdf's picture
Upload folder using huggingface_hub
d0e32f5 verified
raw
history blame contribute delete
453 Bytes
{
"model_type": "sentencepiece",
"vocab_size": 8192,
"bos_token": "<|bos|>",
"eos_token": "<|eos|>",
"pad_token": "<|pad|>",
"chat_template": "{% for message in messages %}<|im_start|>{{ message.role }}\n{{ message.content }}<|im_end|>\n{% endfor %}<|im_start|>assistant\n",
"special_tokens": {
"im_start": "<|im_start|>",
"im_end": "<|im_end|>",
"bos": "<|bos|>",
"eos": "<|eos|>",
"pad": "<|pad|>"
}
}