steampunque commited on
Commit
4592246
·
verified ·
1 Parent(s): c767975

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -10,7 +10,7 @@ tags:
10
  - 6-bit
11
  ---
12
 
13
- ## Llama.cpp hybrid layer quantization of Ring-mini-2.0 by inclusionAI
14
 
15
  Original model: https://huggingface.co/inclusionAI/Ring-mini-2.0
16
 
@@ -161,7 +161,7 @@ Evals for the model are available here: https://huggingface.co/spaces/steampunq
161
  ## Download the file from below:
162
  | Link | Type | Size/e9 B | Notes |
163
  |------|------|-----------|-------|
164
- | [Ring-mini-2.0.Q6_K_H.gguf](https://huggingface.co/steampunque/Ring-mini-2.0-Hybrid-GGUF/resolve/main/Ring-mini-2.0.Q6_K_H.gguf) | Q6_K_H | 13.2e9 B | ~Q6_K size |
165
 
166
  A discussion thread about the hybrid layer quant approach can be found here on the llama.cpp git repository:
167
 
 
10
  - 6-bit
11
  ---
12
 
13
+ ## Mixed Precision GGUF layer quantization of Ring-mini-2.0 by inclusionAI
14
 
15
  Original model: https://huggingface.co/inclusionAI/Ring-mini-2.0
16
 
 
161
  ## Download the file from below:
162
  | Link | Type | Size/e9 B | Notes |
163
  |------|------|-----------|-------|
164
+ | [Ring-mini-2.0.Q6_K_H.gguf](https://huggingface.co/steampunque/Ring-mini-2.0-MP-GGUF/resolve/main/Ring-mini-2.0.Q6_K_H.gguf) | Q6_K_H | 13.2e9 B | ~Q6_K size |
165
 
166
  A discussion thread about the hybrid layer quant approach can be found here on the llama.cpp git repository:
167