Qwen3-Coder-0.6B โ GGUF (q6_k)
This repository contains a GGUF-quantized version of Qwen3-Coder-0.6B, optimized for local inference using llama.cpp-compatible runtimes.
Model details
- Base model: Qwen3-Coder-0.6B
- Quantization: q6_k
- Format: GGUF
- Use case: Code generation, code completion, lightweight coding tasks
Compatibility
Tested with:
- llama.cpp
- LM Studio
- text-generation-webui
- koboldcpp
Usage (llama.cpp)
./llama-cli -m qwen3-coder-0.6b-q6_k.gguf -p "Write a Python function to reverse a list"
- Downloads last month
- 36
Hardware compatibility
Log In
to view the estimation
6-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support