{ "add_prefix_space": false, "added_tokens_decoder": { "50256": { "content": "<|endoftext|>", "lstrip": false, "normalized": true, "rstrip": false, "single_word": false, "special": true }, "50257": { "content": "<|bos|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "50258": { "content": "<|pad|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "50259": { "content": "<|system|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "50260": { "content": "<|user|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "50261": { "content": "<|assistant|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "50262": { "content": "<|tool|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "50263": { "content": "<|observation|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "50264": { "content": "<|document|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "50265": { "content": "<|retrieval|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "50266": { "content": "<|scratchpad|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "50267": { "content": "<|thought|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "50268": { "content": "<|code|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "50269": { "content": "<|sep|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "50270": { "content": "<|endofchunk|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true } }, "additional_special_tokens": [ "<|system|>", "<|user|>", "<|assistant|>", "<|tool|>", "<|observation|>", "<|document|>", "<|retrieval|>", "<|scratchpad|>", "<|thought|>", "<|code|>", "<|sep|>", "<|endofchunk|>" ], "bos_token": "<|bos|>", "clean_up_tokenization_spaces": false, "eos_token": "", "extra_special_tokens": {}, "model_max_length": 1024, "pad_token": "<|pad|>", "tokenizer_class": "GPT2Tokenizer", "unk_token": "<|endoftext|>" }