Qwen3-Coder-0.6B โ€“ GGUF (q6_k)

This repository contains a GGUF-quantized version of Qwen3-Coder-0.6B, optimized for local inference using llama.cpp-compatible runtimes.

Model details

  • Base model: Qwen3-Coder-0.6B
  • Quantization: q6_k
  • Format: GGUF
  • Use case: Code generation, code completion, lightweight coding tasks

Compatibility

Tested with:

  • llama.cpp
  • LM Studio
  • text-generation-webui
  • koboldcpp

Usage (llama.cpp)

./llama-cli -m qwen3-coder-0.6b-q6_k.gguf -p "Write a Python function to reverse a list"
Downloads last month
36
GGUF
Model size
0.8B params
Architecture
qwen3
Hardware compatibility
Log In to view the estimation

6-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support