tokenizerchess / tokenizer_config.json
ItsMaxNorm's picture
Upload folder using huggingface_hub
5c25055 verified
{
"tokenizer_type": "BPE",
"vocab_size": 388,
"pattern": "'(?i:[sdmt]|ll|ve|re)|[^\\r\\n\\p{L}\\p{N}]?+\\p{L}+|\\p{N}{1,3}| ?[^\\s\\p{L}\\p{N}]++[\\r\\n]*|\\s*[\\r\\n]|\\s+(?!\\S)|\\s+",
"special_tokens": {},
"training_config": {
"vocab_size": 256,
"dataset_fraction": "train[0:1000]",
"moves_key": "moves_custom",
"separator": " "
}
}