Ryan Fetterman
Set base_model to fdtn-ai/Foundation-Sec-8B-Instruct with quantized relation
72b2abc
---
license: apache-2.0
library_name: llama.cpp
pipeline_tag: text-generation
tags:
- gguf
- quantized
language:
- en
base_model: fdtn-ai/Foundation-Sec-8B-Instruct
base_model_relation: quantized
---
# Foundation-Sec-8B-Instruct — GGUF (Q4_K_M)
Public GGUF quantization (`Q4_K_M`) for local inference (llama.cpp / LM Studio / Ollama).
- File: `Foundation-Sec-8B-Instruct-Q4_K_M.gguf`
- Pull via Git LFS.