Final Rescue: Full vocab restoration with Unigram scores and Metaspace fix
Browse files- tokenizer.json +2 -2
tokenizer.json
CHANGED
|
@@ -43,7 +43,7 @@
|
|
| 43 |
"normalizer": null,
|
| 44 |
"pre_tokenizer": {
|
| 45 |
"type": "Metaspace",
|
| 46 |
-
"replacement": "
|
| 47 |
"prepend_scheme": "always",
|
| 48 |
"split": true
|
| 49 |
},
|
|
@@ -112,7 +112,7 @@
|
|
| 112 |
},
|
| 113 |
"decoder": {
|
| 114 |
"type": "Metaspace",
|
| 115 |
-
"replacement": "
|
| 116 |
"prepend_scheme": "always",
|
| 117 |
"split": true
|
| 118 |
},
|
|
|
|
| 43 |
"normalizer": null,
|
| 44 |
"pre_tokenizer": {
|
| 45 |
"type": "Metaspace",
|
| 46 |
+
"replacement": "▁",
|
| 47 |
"prepend_scheme": "always",
|
| 48 |
"split": true
|
| 49 |
},
|
|
|
|
| 112 |
},
|
| 113 |
"decoder": {
|
| 114 |
"type": "Metaspace",
|
| 115 |
+
"replacement": "▁",
|
| 116 |
"prepend_scheme": "always",
|
| 117 |
"split": true
|
| 118 |
},
|