GHOSTAI-Spooky / README.md
ghostai1's picture
Update README.md
640cc41 verified
---
license: mit
language:
- en
tags:
- gguf
- llama.cpp
- mistral
- instruct
- lora
- horror
- roleplay
pipeline_tag: text-generation
---
<!--
GHOSTAI • HORROR GGUF (7B)
Professional Hugging Face README for a quantized-only GGUF release.
-->
<p align="center">
<img src="https://capsule-render.vercel.app/api?type=waving&color=0:0b0b0f,50:2b0a2a,100:0b0b0f&height=160&section=header&text=GHOSTAI%20—%20HORROR%20GGUF%20(7B)&fontSize=44&fontColor=EAEAEA&animation=twinkling" />
</p>
<p align="center">
<img alt="Format" src="https://img.shields.io/badge/Format-GGUF-8A2BE2?style=for-the-badge">
<img alt="Runtime" src="https://img.shields.io/badge/Runtime-llama.cpp-5B2C83?style=for-the-badge">
<img alt="Model Size" src="https://img.shields.io/badge/Model-7B-3A0CA3?style=for-the-badge">
<img alt="Theme" src="https://img.shields.io/badge/Theme-Horror-8B0000?style=for-the-badge">
<a href="https://www.linkedin.com/in/ccengineering/">
<img alt="LinkedIn" src="https://img.shields.io/badge/LinkedIn-ccengineering-0A66C2?style=for-the-badge&logo=linkedin&logoColor=white">
</a>
</p>
<p align="center">
<strong>GHOSTAI — HORROR GGUF (7B)</strong><br/>
A focused, horror-themed 7B model released exclusively in quantized GGUF format for the <strong>llama.cpp</strong> ecosystem.<br/>
<sub>Quantized-only release. No FP16 weights included.</sub>
</p>
---
## Overview
**GHOSTAI** is a compact, atmosphere-driven horror model designed for narrative generation, roleplay, and dark storytelling.
It prioritizes tone, pacing, and vivid imagery over generic assistant behavior.
This repository provides **multiple GGUF quantizations**, allowing you to choose the best balance of quality, speed, and memory usage for your hardware.
The model runs:
- Fully on **CPU**
- With optional **GPU offload** (CUDA / Metal / Vulkan builds of llama.cpp)
Quantization choice is independent of whether you use CPU or GPU.
---
## Files
| File | Quant | Approx size | Rough RAM needed (4k ctx) |
|---|---:|---:|---:|
| `ghostai-horror-7b.Q8_0.gguf` | Q8_0 | ~7.2 GB | ~10–11 GB |
| `ghostai-horror-7b.Q6_K.gguf` | Q6_K | ~5.5 GB | ~8–9 GB |
| `ghostai-horror-7b.Q5_K_M.gguf` | Q5_K_M | ~4.8 GB | ~7–8 GB |
| `ghostai-horror-7b.Q5_K_S.gguf` | Q5_K_S | ~4.7 GB | ~7–8 GB |
| `ghostai-horror-7b.Q4_K_M.gguf` | Q4_K_M | ~4.1 GB | ~6–7 GB |
| `ghostai-horror-7b.Q4_K_S.gguf` | Q4_K_S | ~3.9 GB | ~6–7 GB |
| `ghostai-horror-7b.Q3_K_M.gguf` | Q3_K_M | ~3.3 GB | ~5–6 GB |
| `ghostai-horror-7b.Q3_K_S.gguf` | Q3_K_S | ~3.0 GB | ~5–6 GB |
| `ghostai-horror-7b.Q2_K.gguf` | Q2_K | ~2.5 GB | ~4–5 GB |
| `ghostai-horror-7b.TQ1_0.gguf` | TQ1_0 | ~1.6 GB | ~3–4 GB |
Notes:
- “Rough RAM needed” assumes **~4k context** and typical llama.cpp overhead.
- For **8k context**, plan **+1–2 GB** extra.
- GPU offload can shift some load to VRAM, but you still need system RAM.
---
## Recommended Downloads
- Best default: **`Q4_K_M`**
- More quality (more RAM): **`Q5_K_M`**, **`Q6_K`**, **`Q8_0`**
- Low RAM: **`Q3_K_S`**, **`Q2_K`**
- Ultra-small / experimental: **`TQ1_0`** (expect noticeable quality loss)
---
## Quickstart (llama.cpp)
### 1) Run on CPU
```bash
./llama-cli \
-m ghostai-horror-7b.Q4_K_M.gguf \
-c 4096 \
-t 8 \
-p "You are GHOSTAI. Speak like a calm horror narrator. Keep it tight and vivid."