{
"tokenizer_class": "PreTrainedTokenizerFast",
"model_max_length": 512,
"bos_token": "",
"eos_token": "",
"unk_token": "",
"pad_token": "",
"clean_up_tokenization_spaces": false,
"note": "This model uses a tiktoken-based tokenizer (cl100k_base remapped to 8192 tokens). The tokenizer.model file is for reference but requires custom loading."
}