Qwen 3.5
Collection
10 items • Updated
The following quantization formats are available in this repository:
You can run these models locally using llama.cpp or compatible interfaces like LM Studio, Ollama, or text-generation-webui.
Example using llama.cpp in the terminal:
./main -m Qwen3.5-4B-Python-Coder-Q4_K_M.gguf -n 512 --color -i -cml -p "<|im_start|>user\nWrite a Python script to scrape a website.<|im_end|>\n<|im_start|>assistant\n"
3-bit
4-bit
5-bit
6-bit
8-bit
Base model
Qwen/Qwen3.5-4B-Base