Carnice
Collection
Carnice is a project to make a model preform well in a specific agentic harness (hermes-agent • 6 items • Updated • 4
GGUF builds of Carnice-9b, a Hermes-Agent-specialized model built from Qwen/Qwen3.5-9B and trained specifically for the Hermes-Agent harness.
This repo contains three quantized variants:
Carnice-9b-Q4_K_M.ggufCarnice-9b-Q6_K.ggufCarnice-9b-Q8_0.gguf| File | Quant | Size | Recommended use |
|---|---|---|---|
Carnice-9b-Q4_K_M.gguf |
4-bit | 5.3 GB | fastest local testing |
Carnice-9b-Q6_K.gguf |
6-bit | 6.9 GB | best quality/size balance |
Carnice-9b-Q8_0.gguf |
8-bit | 8.9 GB | highest quality GGUF option |
Merged source model:
Carnice-9b was trained specifically around Hermes-Agent behavior rather than generic chat polish. The training mixture emphasized:
llama-cli -m Carnice-9b-Q6_K.gguf -p "Reply with exactly READY." -n 16
These are GGUF exports of the merged standalone Carnice model, not PEFT adapters.
4-bit
6-bit
8-bit