{
"tokenizer_class": "PreTrainedTokenizerFast",
"auto_map": {
"AutoTokenizer": [
"transformers",
"PreTrainedTokenizerFast"
]
},
"bos_token": "",
"eos_token": "",
"unk_token": "",
"pad_token": "",
"model_max_length": 512,
"special_tokens_map_file": null,
"name_or_path": "mELECTRA",
"tokenizer_type": "SentencePiece"
}