Update README.md
Browse files
README.md
CHANGED
|
@@ -17,6 +17,13 @@ Working GGUF of [Qwen/Qwen3-Reranker-8B](https://huggingface.co/Qwen/Qwen3-Reran
|
|
| 17 |
|
| 18 |
> **Other sizes:** [0.6B](https://huggingface.co/Voodisss/Qwen3-Reranker-0.6B-GGUF-llama_cpp) 路 [4B](https://huggingface.co/Voodisss/Qwen3-Reranker-4B-GGUF-llama_cpp) 路 [8B (this)](https://huggingface.co/Voodisss/Qwen3-Reranker-8B-GGUF-llama_cpp)
|
| 19 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 20 |
## Does it work?
|
| 21 |
|
| 22 |
Yes. Most community GGUFs of Qwen3-Reranker produce garbage scores (`4.5e-23`) because they're missing reranker-specific tensors. See [llama.cpp #16407](https://github.com/ggml-org/llama.cpp/issues/16407). This one works:
|
|
|
|
| 17 |
|
| 18 |
> **Other sizes:** [0.6B](https://huggingface.co/Voodisss/Qwen3-Reranker-0.6B-GGUF-llama_cpp) 路 [4B](https://huggingface.co/Voodisss/Qwen3-Reranker-4B-GGUF-llama_cpp) 路 [8B (this)](https://huggingface.co/Voodisss/Qwen3-Reranker-8B-GGUF-llama_cpp)
|
| 19 |
|
| 20 |
+
## Available files
|
| 21 |
+
|
| 22 |
+
| File | Quant | Size | Description |
|
| 23 |
+
| --------------------------- | ----- | -------- | -------------------------------------------------- |
|
| 24 |
+
| `Qwen3-Reranker-8B-F16.gguf` | F16 | 14.10 GB | Full precision, no quality loss |
|
| 25 |
+
| `Qwen3-Reranker-8B-Q8_0.gguf` | Q8_0 | 7.49 GB | 8-bit quantized, half the size |
|
| 26 |
+
|
| 27 |
## Does it work?
|
| 28 |
|
| 29 |
Yes. Most community GGUFs of Qwen3-Reranker produce garbage scores (`4.5e-23`) because they're missing reranker-specific tensors. See [llama.cpp #16407](https://github.com/ggml-org/llama.cpp/issues/16407). This one works:
|