Update README.md
Browse files
README.md
CHANGED
|
@@ -2,63 +2,74 @@
|
|
| 2 |
license: mit
|
| 3 |
---
|
| 4 |
|
| 5 |
-
<!--
|
| 6 |
-
GHOSTAI • HORROR GGUF RELEASE README
|
| 7 |
-
Drop this into README.md at the root of your Hugging Face repo.
|
| 8 |
-
-->
|
| 9 |
|
| 10 |
<p align="center">
|
| 11 |
-
<img src="https://capsule-render.vercel.app/api?type=waving&color=0:0b0b0f,50:2b0a2a,100:0b0b0f&height=160§ion=header&text=GHOSTAI%20%E2%80%94%20HORROR%20GGUF&fontSize=
|
| 12 |
</p>
|
| 13 |
|
| 14 |
<p align="center">
|
| 15 |
<img alt="GGUF" src="https://img.shields.io/badge/GGUF-llama.cpp-8A2BE2?style=for-the-badge">
|
| 16 |
-
<img alt="
|
| 17 |
-
<img alt="Quant" src="https://img.shields.io/badge/Quant-Q4__K__M%20%7C%20F16-3A0CA3?style=for-the-badge">
|
| 18 |
<img alt="Theme" src="https://img.shields.io/badge/Theme-Horror-8B0000?style=for-the-badge">
|
|
|
|
| 19 |
</p>
|
| 20 |
|
| 21 |
<p align="center">
|
| 22 |
-
<b>GHOSTAI</b> is a horror-
|
| 23 |
-
<
|
| 24 |
-
Pick your haunt: <b>F16</b> for max fidelity or <b>Q4_K_M</b> for the best everyday balance.
|
| 25 |
</p>
|
| 26 |
|
| 27 |
---
|
| 28 |
|
| 29 |
## 🩸 What’s inside
|
| 30 |
|
| 31 |
-
|
| 32 |
|
| 33 |
-
### 🎃
|
| 34 |
|
| 35 |
-
|
|
| 36 |
-
|---|---|---:|---|
|
| 37 |
-
|
|
| 38 |
-
|
|
| 39 |
-
|
|
| 40 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 41 |
|
| 42 |
-
|
|
|
|
|
|
|
|
|
|
| 43 |
|
| 44 |
---
|
| 45 |
|
| 46 |
-
##
|
| 47 |
|
| 48 |
-
- **
|
| 49 |
-
- **
|
| 50 |
-
- **
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 51 |
|
| 52 |
---
|
| 53 |
|
| 54 |
## ⚰️ Quickstart (llama.cpp)
|
| 55 |
|
| 56 |
-
###
|
| 57 |
|
| 58 |
```bash
|
| 59 |
./llama-cli \
|
| 60 |
-
-m
|
| 61 |
-ngl 99 \
|
| 62 |
-c 4096 \
|
| 63 |
-
-p "You are GHOSTAI. Speak like a calm
|
|
|
|
| 64 |
|
|
|
|
| 2 |
license: mit
|
| 3 |
---
|
| 4 |
|
| 5 |
+
<!-- GHOSTAI • HORROR GGUF (7B) — README -->
|
|
|
|
|
|
|
|
|
|
| 6 |
|
| 7 |
<p align="center">
|
| 8 |
+
<img src="https://capsule-render.vercel.app/api?type=waving&color=0:0b0b0f,50:2b0a2a,100:0b0b0f&height=160§ion=header&text=GHOSTAI%20%E2%80%94%20HORROR%20GGUF%20(7B)&fontSize=42&fontColor=EAEAEA&animation=twinkling" />
|
| 9 |
</p>
|
| 10 |
|
| 11 |
<p align="center">
|
| 12 |
<img alt="GGUF" src="https://img.shields.io/badge/GGUF-llama.cpp-8A2BE2?style=for-the-badge">
|
| 13 |
+
<img alt="7B" src="https://img.shields.io/badge/Size-7B-5B2C83?style=for-the-badge">
|
|
|
|
| 14 |
<img alt="Theme" src="https://img.shields.io/badge/Theme-Horror-8B0000?style=for-the-badge">
|
| 15 |
+
<img alt="Quant" src="https://img.shields.io/badge/Quants-Q8__0%20%7C%20Q6__K%20%7C%20Q5__K__M%20%7C%20Q4__K__M-3A0CA3?style=for-the-badge">
|
| 16 |
</p>
|
| 17 |
|
| 18 |
<p align="center">
|
| 19 |
+
<b>GHOSTAI</b> is a <b>horror-themed</b> <b>7B</b> GGUF release for the <b>llama.cpp</b> ecosystem.<br/>
|
| 20 |
+
This repo contains <b>quantized GGUFs only</b> (no FP16).
|
|
|
|
| 21 |
</p>
|
| 22 |
|
| 23 |
---
|
| 24 |
|
| 25 |
## 🩸 What’s inside
|
| 26 |
|
| 27 |
+
Quantized GGUF files (7B) ready for llama.cpp-compatible runtimes.
|
| 28 |
|
| 29 |
+
### 🎃 Files in this release
|
| 30 |
|
| 31 |
+
| File | Quant | Approx size | Rough RAM needed (4k ctx) |
|
| 32 |
+
|---|---:|---:|---:|
|
| 33 |
+
| `ghostai-horror-7b.Q8_0.gguf` | Q8_0 | ~7.2 GB | ~10–11 GB |
|
| 34 |
+
| `ghostai-horror-7b.Q6_K.gguf` | Q6_K | ~5.5 GB | ~8–9 GB |
|
| 35 |
+
| `ghostai-horror-7b.Q5_K_M.gguf` | Q5_K_M | ~4.8 GB | ~7–8 GB |
|
| 36 |
+
| `ghostai-horror-7b.Q5_K_S.gguf` | Q5_K_S | ~4.7 GB | ~7–8 GB |
|
| 37 |
+
| `ghostai-horror-7b.Q4_K_M.gguf` | Q4_K_M | ~4.1 GB | ~6–7 GB |
|
| 38 |
+
| `ghostai-horror-7b.Q4_K_S.gguf` | Q4_K_S | ~3.9 GB | ~6–7 GB |
|
| 39 |
+
| `ghostai-horror-7b.Q3_K_M.gguf` | Q3_K_M | ~3.3 GB | ~5–6 GB |
|
| 40 |
+
| `ghostai-horror-7b.Q3_K_S.gguf` | Q3_K_S | ~3.0 GB | ~5–6 GB |
|
| 41 |
+
| `ghostai-horror-7b.Q2_K.gguf` | Q2_K | ~2.5 GB | ~4–5 GB |
|
| 42 |
+
| `ghostai-horror-7b.TQ1_0.gguf` | TQ1_0 | ~1.6 GB | ~3–4 GB |
|
| 43 |
|
| 44 |
+
**RAM notes (rough):**
|
| 45 |
+
- “Rough RAM needed” assumes **~4k context** and typical llama.cpp overhead.
|
| 46 |
+
- If you run **8k context**, add roughly **+1–2 GB**.
|
| 47 |
+
- GPU offload doesn’t remove the need for RAM; it shifts some weight/KV usage to VRAM depending on settings.
|
| 48 |
|
| 49 |
---
|
| 50 |
|
| 51 |
+
## 🧟 Which quant should I use?
|
| 52 |
|
| 53 |
+
- **Best default:** `Q4_K_M`
|
| 54 |
+
- **Higher quality:** `Q5_K_M` or `Q6_K`
|
| 55 |
+
- **If you have plenty of RAM:** `Q8_0`
|
| 56 |
+
- **Low RAM:** `Q3_K_S` / `Q2_K`
|
| 57 |
+
- **Tiny / experimental:** `TQ1_0` (expect quality loss)
|
| 58 |
+
|
| 59 |
+
These formats are **not “CPU vs GPU.”**
|
| 60 |
+
You can run any quant on CPU-only or with GPU offload.
|
| 61 |
|
| 62 |
---
|
| 63 |
|
| 64 |
## ⚰️ Quickstart (llama.cpp)
|
| 65 |
|
| 66 |
+
### GPU offload (CUDA build)
|
| 67 |
|
| 68 |
```bash
|
| 69 |
./llama-cli \
|
| 70 |
+
-m ghostai-horror-7b.Q4_K_M.gguf \
|
| 71 |
-ngl 99 \
|
| 72 |
-c 4096 \
|
| 73 |
+
-p "You are GHOSTAI. Speak like a calm horror narrator. Keep it tight and vivid."
|
| 74 |
+
|
| 75 |
|