| { | |
| "version": "1.0", | |
| "model_type": "GPT-2", | |
| "tokenizer_class": "GPT2Tokenizer", | |
| "vocab_file": "vocab.json", | |
| "merges_file": "merges.txt", | |
| "special_tokens_map": { | |
| "pad_token": "<PAD>", | |
| "unk_token": "<UNK>", | |
| "cls_token": "<CLS>", | |
| "sep_token": "<SEP>", | |
| "mask_token": "<MASK>" | |
| } | |
| } | |