KoVicuna-5.8b-v1 / tokenizer_config.json
rippertnt's picture
Upload 11 files
88145bc
raw
history blame contribute delete
352 Bytes
{
"eos_token": "<|endoftext|>",
"model_max_length": 1000000000000000019884624838656,
"pad_token": "<|endoftext|>",
"special_tokens_map_file": "/home/circulus/.cache/huggingface/hub/models--EleutherAI--polyglot-ko-5.8b/snapshots/7d4b97cde25b41489579b0919d381161cec0300f/special_tokens_map.json",
"tokenizer_class": "PreTrainedTokenizerFast"
}