test1 / tokenizer_config.json
gszabo's picture
Upload tokenizer
dcd7533
raw
history blame contribute delete
243 Bytes
{
"name_or_path": "gszabo/test1",
"special_tokens_map_file": "/root/.cache/huggingface/hub/models--gszabo--test1/snapshots/6f8df0182582a5d3ae2d140e30c94b8bf36bc3f4/special_tokens_map.json",
"tokenizer_class": "PreTrainedTokenizerFast"
}