Update README.md
Browse files
README.md
CHANGED
|
@@ -3,3 +3,13 @@ license: other
|
|
| 3 |
license_name: microsoft
|
| 4 |
license_link: https://huggingface.co/microsoft/Orca-2-13b/blob/main/LICENSE
|
| 5 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3 |
license_name: microsoft
|
| 4 |
license_link: https://huggingface.co/microsoft/Orca-2-13b/blob/main/LICENSE
|
| 5 |
---
|
| 6 |
+
|
| 7 |
+
Lost of good models use Orca for their merges, however vanilla Orca has vocabulary size of 32003,
|
| 8 |
+
3 last tokens are ChatML tokens and a PAD token. This causes errors during a merge with models with standard 32000 vocabulary size.
|
| 9 |
+
|
| 10 |
+
I've removed those tokens from volabulary and resized model embeddings to mach 32000 standard size. So this model is ready to be used
|
| 11 |
+
as a merge component in mergekit. It may not work on its own with ChatML template anymore.
|
| 12 |
+
|
| 13 |
+
```python
|
| 14 |
+
model.resize_token_embeddings(32000)
|
| 15 |
+
```
|