Brick Complexity Extractor (Q4_K_M GGUF)

Q4_K_M quantized GGUF of regolo/brick-complexity-extractor

Regolo.ai | Original Model | Dataset | Brick SR1 on GitHub

License: CC BY-NC 4.0 Base Model


Model Details

Property Value
Quantization Q4_K_M
File brick-complexity-extractor-Q4_K_M.gguf
Size 494 MB
Bits per weight 5.5
Original model regolo/brick-complexity-extractor
Base model Qwen/Qwen3.5-0.8B
Output classes 3 (easy, medium, hard)
License CC BY-NC 4.0

4-bit k-quant mixed precision. Best quality/size ratio. Ideal for edge and resource-constrained deployments.

This is a full merged model (base Qwen3.5-0.8B + LoRA adapter merged and quantized), so no separate adapter loading is needed.

All Available Quantizations

Model Quant Size BPW
BF16-GGUF BF16 1.5 GB 16.0
Q8_0-GGUF Q8_0 775 MB 8.0
Q4_K_M-GGUF Q4_K_M 494 MB 5.5

Usage with llama.cpp

# Download
huggingface-cli download regolo/brick-complexity-extractor-Q4_K_M-GGUF \
    brick-complexity-extractor-Q4_K_M.gguf --local-dir ./models

# Run inference
./llama-cli -m ./models/brick-complexity-extractor-Q4_K_M.gguf \
    -p "<|im_start|>system
You are a query difficulty classifier for an LLM routing system.
Classify each query as easy, medium, or hard based on the cognitive depth and domain expertise required to answer correctly.
Respond with ONLY one word: easy, medium, or hard.<|im_end|>
<|im_start|>user
Classify: What is the capital of France?<|im_end|>
<|im_start|>assistant
" \
    -n 5 --temp 0

Usage with Ollama

cat > Modelfile <<EOF
FROM ./brick-complexity-extractor-Q4_K_M.gguf

SYSTEM \"\"\"You are a query difficulty classifier for an LLM routing system.
Classify each query as easy, medium, or hard based on the cognitive depth and domain expertise required to answer correctly.
Respond with ONLY one word: easy, medium, or hard.\"\"\"

TEMPLATE \"\"\"<|im_start|>system
{{ .System }}<|im_end|>
<|im_start|>user
Classify: {{ .Prompt }}<|im_end|>
<|im_start|>assistant
\"\"\"

PARAMETER temperature 0
PARAMETER num_predict 5
EOF

ollama create brick-complexity -f Modelfile
ollama run brick-complexity "Design a distributed consensus algorithm"
# Output: hard

Usage with vLLM

from vllm import LLM, SamplingParams

llm = LLM(model="regolo/brick-complexity-extractor-Q4_K_M-GGUF")
sampling_params = SamplingParams(temperature=0, max_tokens=5)

prompt = \"\"\"<|im_start|>system
You are a query difficulty classifier for an LLM routing system.
Classify each query as easy, medium, or hard.
Respond with ONLY one word: easy, medium, or hard.<|im_end|>
<|im_start|>user
Classify: Explain the rendering equation from radiometric first principles<|im_end|>
<|im_start|>assistant
\"\"\"

output = llm.generate([prompt], sampling_params)
print(output[0].outputs[0].text.strip())
# Output: hard

Note on GGUF Inference

The GGUF model uses generative text output (generates "easy", "medium", or "hard") rather than logit-based classification used by the original LoRA adapter. For production deployments requiring maximum accuracy, consider using the original LoRA adapter with the PEFT library.

About

Regolo.ai is the EU-sovereign LLM inference platform built on Seeweb infrastructure. Brick is our open-source semantic routing system that intelligently distributes queries across model pools, optimizing for cost, latency, and quality.

Website | Docs | GitHub | Discord

Downloads last month
50
GGUF
Model size
0.8B params
Architecture
qwen35
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for regolo/brick-complexity-extractor-Q4_K_M-GGUF

Quantized
(3)
this model

Collection including regolo/brick-complexity-extractor-Q4_K_M-GGUF

Evaluation results