Source: Qwen/Qwen2.5-Coder-7B-Instruct-GGUF
This repo bundles the Q4_KM GGUF variant along with necessary configuration and tokenizer files. No weight merging. Compatible with llama.cpp ecosystem.
Chat template
4-bit