NousCoder-14B GGUF

GGUF quantizations of NousResearch/NousCoder-14B for use with llama.cpp, Ollama, LM Studio, and other GGUF-compatible inference engines.

Credits

All credit goes to NousResearch for training and releasing the original NousCoder-14B model. This repo only provides quantized GGUF versions for easier local inference.

Available Quants

Filename Quant Size Description
nouscoder-14b-q4_k_m.gguf Q4_K_M 8.4 GB Good balance of quality and size
nouscoder-14b-q3_k_m.gguf Q3_K_M 6.9 GB Lower memory, good quality
nouscoder-14b-q3_k_s.gguf Q3_K_S 6.2 GB Smallest, for constrained environments

Usage

llama.cpp

./llama-cli -m nouscoder-14b-q4_k_m.gguf -p "def fibonacci(n):"

Ollama

ollama run hf.co/bigatuna/NousCoder-14B-GGUF:Q4_K_M

Original Model

Downloads last month
132
GGUF
Model size
15B params
Architecture
qwen3
Hardware compatibility
Log In to view the estimation

3-bit

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for bigatuna/NousCoder-14B-GGUF

Finetuned
Qwen/Qwen3-14B
Quantized
(25)
this model