{ "tokenizer_class": "PreTrainedTokenizerFast", "model_max_length": 512, "bos_token": "", "eos_token": "", "unk_token": "", "pad_token": "", "clean_up_tokenization_spaces": false, "note": "This model uses a tiktoken-based tokenizer (cl100k_base remapped to 8192 tokens). The tokenizer.model file is for reference but requires custom loading." }