Update tokenizer.json
Browse files- tokenizer.json +2 -2
tokenizer.json
CHANGED
|
@@ -5,7 +5,7 @@
|
|
| 5 |
"added_tokens": [
|
| 6 |
{
|
| 7 |
"id": 0,
|
| 8 |
-
"content": "<|
|
| 9 |
"single_word": false,
|
| 10 |
"lstrip": false,
|
| 11 |
"rstrip": false,
|
|
@@ -41,7 +41,7 @@
|
|
| 41 |
"fuse_unk": false,
|
| 42 |
"byte_fallback": false,
|
| 43 |
"vocab": {
|
| 44 |
-
"<|
|
| 45 |
"!": 1,
|
| 46 |
"\"": 2,
|
| 47 |
"#": 3,
|
|
|
|
| 5 |
"added_tokens": [
|
| 6 |
{
|
| 7 |
"id": 0,
|
| 8 |
+
"content": "<|endofmolecule|>",
|
| 9 |
"single_word": false,
|
| 10 |
"lstrip": false,
|
| 11 |
"rstrip": false,
|
|
|
|
| 41 |
"fuse_unk": false,
|
| 42 |
"byte_fallback": false,
|
| 43 |
"vocab": {
|
| 44 |
+
"<|endofmolecule|>": 0,
|
| 45 |
"!": 1,
|
| 46 |
"\"": 2,
|
| 47 |
"#": 3,
|