This repository hosts quantized versions of the QWQ-32B reasoning model.
Format: GGUFConverter: llama.cpp ba7654380a3c7c1b5ae154bea19134a3a9417a1eQuantizer: LM-Kit.NET 2025.3.3
For more detailed information on the base model, please visit the following link
Chat template
4-bit