{ "tokenizer_class": "WordTokenizer", "model_max_length": 128, "bos_token": "", "eos_token": "", "unk_token": "", "pad_token": "", "vocab_size": 69, "name_or_path": "./" }