Update README.md
Browse files
README.md
CHANGED
|
@@ -10,7 +10,7 @@ tags:
|
|
| 10 |
- 6-bit
|
| 11 |
---
|
| 12 |
|
| 13 |
-
##
|
| 14 |
|
| 15 |
Original model: https://huggingface.co/inclusionAI/Ring-mini-2.0
|
| 16 |
|
|
@@ -161,7 +161,7 @@ Evals for the model are available here: https://huggingface.co/spaces/steampunq
|
|
| 161 |
## Download the file from below:
|
| 162 |
| Link | Type | Size/e9 B | Notes |
|
| 163 |
|------|------|-----------|-------|
|
| 164 |
-
| [Ring-mini-2.0.Q6_K_H.gguf](https://huggingface.co/steampunque/Ring-mini-2.0-
|
| 165 |
|
| 166 |
A discussion thread about the hybrid layer quant approach can be found here on the llama.cpp git repository:
|
| 167 |
|
|
|
|
| 10 |
- 6-bit
|
| 11 |
---
|
| 12 |
|
| 13 |
+
## Mixed Precision GGUF layer quantization of Ring-mini-2.0 by inclusionAI
|
| 14 |
|
| 15 |
Original model: https://huggingface.co/inclusionAI/Ring-mini-2.0
|
| 16 |
|
|
|
|
| 161 |
## Download the file from below:
|
| 162 |
| Link | Type | Size/e9 B | Notes |
|
| 163 |
|------|------|-----------|-------|
|
| 164 |
+
| [Ring-mini-2.0.Q6_K_H.gguf](https://huggingface.co/steampunque/Ring-mini-2.0-MP-GGUF/resolve/main/Ring-mini-2.0.Q6_K_H.gguf) | Q6_K_H | 13.2e9 B | ~Q6_K size |
|
| 165 |
|
| 166 |
A discussion thread about the hybrid layer quant approach can be found here on the llama.cpp git repository:
|
| 167 |
|