Update README.md
Browse files
README.md
CHANGED
|
@@ -15,7 +15,7 @@ revision: 602838d # Aug 19 2025
|
|
| 15 |
# Qwen3-Reranker-0.6B-GGUF
|
| 16 |
|
| 17 |
**🚨 REQUIRED Llama.cpp build:** https://github.com/ngxson/llama.cpp/tree/xsn/qwen3_embd_rerank
|
| 18 |
-
**This unmerged fix branch is mandatory** to run Qwen3 reranking models. Other HF GGUF quantizations of the reranker typically fail in mainline `llama.cpp` because they were not produced with this build. **This quantization was produced with the above build and works.**
|
| 19 |
|
| 20 |
## Purpose
|
| 21 |
Multilingual **text-reranking** model in **GGUF** for efficient CPU/GPU inference with *llama.cpp*-compatible back-ends.
|
|
|
|
| 15 |
# Qwen3-Reranker-0.6B-GGUF
|
| 16 |
|
| 17 |
**🚨 REQUIRED Llama.cpp build:** https://github.com/ngxson/llama.cpp/tree/xsn/qwen3_embd_rerank
|
| 18 |
+
**This unmerged fix branch is mandatory** to run Qwen3 reranking models. Other HF GGUF quantizations of the 0.6B reranker typically fail in mainline `llama.cpp` because they were not produced with this build. **This quantization was produced with the above build and works.**
|
| 19 |
|
| 20 |
## Purpose
|
| 21 |
Multilingual **text-reranking** model in **GGUF** for efficient CPU/GPU inference with *llama.cpp*-compatible back-ends.
|