prism-coder:8b โ€” AAC Tool Router (8B)

Fine-tuned from Qwen3-8B for deterministic tool routing in the Prism AAC system.

BFCL accuracy: 95% on 100-case ร— 3 seeds routing benchmark (v30 corpus).

What it does

Routes user messages to one of 6 tools or plain text. Intermediate quality tier in the desktop cascade (14B โ†’ 32B โ†’ cloud). 8B is retained as an iOS/edge tier for 8GB-RAM devices where 14B does not fit.

Tool Trigger
session_load_context Load/fetch context for project X
session_save_ledger Note / jot down / log / remember
session_save_handoff Handoff to next agent / pass on
session_compact_ledger Compact/archive/trim the ledger
session_search_memory What did we discuss / recall session
knowledge_search What do I know / stored notes
(plain text) AAC phrases, math, facts, translation, time

Deployment

ollama run dcostenco/prism-coder:8b-v30

Files

File Size Format
prism-coder-8b-v30-q4km.gguf 4.7 GB Q4_K_M GGUF (v30, recommended)
prism-aac-8b-q4km.gguf 5.0 GB Q4_K_M GGUF (legacy v29)

Training

  • Base: Qwen3-8B
  • Method: MLX LoRA fine-tuning (mlx_lm.lora)
  • Dataset: v36_1b7 routing corpus (6-tool system prompt)
  • Hardware: Apple Silicon (M-series)
  • Eval: BFCL 100-case benchmark โ†’ 95%

Cascade position

Desktop cascade: 14B โ†’ 32B โ†’ cloud Claude
iOS cascade: 8B (primary offline tier for 8GB-RAM devices)

Downloads last month
209
GGUF
Model size
8B params
Architecture
qwen3
Hardware compatibility
Log In to add your hardware

We're not able to determine the quantization variants.

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for dcostenco/prism-coder-8b

Finetuned
Qwen/Qwen3-8B
Quantized
(280)
this model