Qwen3.5-9B β GGUF (Q6_K)
π· Model Overview
This repository contains a GGUF Q6_K conversion of:
- Base Model: Qwen3.5-9B
- Developer: Qwen
- Format: GGUF (optimized for llama.cpp)
- Precision: Q6_K
This model is designed for high-quality local inference.
π¦ Files
| File | Description |
|---|---|
Qwen3.5-9B_Q6_K.gguf |
Q6_K GGUF model |
βοΈ Technical Details
| Parameter | Value |
|---|---|
| Architecture | Qwen3.5-9B |
| Format | GGUF |
| Precision | Q6_K |
| Runtime | llama.cpp |
| Use Case | High-quality inference |
β‘ Why GGUF?
GGUF enables:
- Efficient CPU inference via llama.cpp
- Single-file model distribution
- Fast loading using memory mapping
- Cross-platform compatibility
β οΈ License & Usage
This is a converted derivative model.
You must comply with the original license for Qwen series models
Important:
- β Not an official Qwen release
- β No additional rights granted
- β Original model ownership remains with Qwen
- β οΈ Use responsibly under original license terms
π Quick Start (llama.cpp)
./llama-cli -m Qwen3.5-9B-Q6_K.gguf -p "Explain AI simply"
- Downloads last month
- 12
Hardware compatibility
Log In to add your hardware
6-bit