Zen Embedding 8B GGUF
High-performance text embedding model based on Qwen3-Embedding-8B, optimized for efficient inference.
Downloads
| Source | URL |
|---|---|
| HuggingFace | hf download zenlm/zen-embedding-8B-GGUF |
| Direct | https://download.hanzo.ai/llm-models/zen-embedding-8B-Q4_K_M.gguf |
Features
- 100+ language support
- #1 on MTEB multilingual leaderboard
- Optimized for semantic search and retrieval
- GGUF format for efficient CPU/GPU inference
- Q4_K_M quantization (4.68 GB)
Usage
Works with llama.cpp and compatible inference engines.
License
Apache 2.0 (inherited from Qwen3-Embedding)
- Downloads last month
- 75
Hardware compatibility
Log In
to add your hardware
4-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support