{ "tokenizer_class": "PreTrainedTokenizerFast", "vocab_size": 32000, "model_max_length": 2048, "padding_side": "left", "truncation_side": "right", "special_tokens": { "": 0, "": 1, "": 2, "": 3 }, "bos_token": "", "eos_token": "", "pad_token": "", "unk_token": "" }