Upload tokenizer
Browse files- special_tokens_map.json +7 -1
special_tokens_map.json
CHANGED
|
@@ -32,7 +32,13 @@
|
|
| 32 |
"rstrip": false,
|
| 33 |
"single_word": false
|
| 34 |
},
|
| 35 |
-
"pad_token":
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 36 |
"unk_token": {
|
| 37 |
"content": "<|endoftext|>",
|
| 38 |
"lstrip": false,
|
|
|
|
| 32 |
"rstrip": false,
|
| 33 |
"single_word": false
|
| 34 |
},
|
| 35 |
+
"pad_token": {
|
| 36 |
+
"content": "<|endoftext|>",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false
|
| 41 |
+
},
|
| 42 |
"unk_token": {
|
| 43 |
"content": "<|endoftext|>",
|
| 44 |
"lstrip": false,
|