|
|
--- |
|
|
license: apache-2.0 |
|
|
tags: |
|
|
- gguf |
|
|
- llama.cpp |
|
|
- quantized |
|
|
- deepseek |
|
|
- stheno |
|
|
--- |
|
|
|
|
|
# DeepSeek Sunfall Merged - GGUF Quantized Models |
|
|
|
|
|
This repository contains multiple **quantized GGUF variants** of the merged DeepSeek + Sunfall model, compatible with `llama.cpp`. |
|
|
|
|
|
## 🧠 Available Quantized Formats |
|
|
|
|
|
| Format | File Name | Description | |
|
|
|-------------|--------------------------------------------------|---------------------------------| |
|
|
| Q3_K_M | `deepseek_sunfall_merged_Model.Q3_K_M.gguf` | Smallest size, fastest inference | |
|
|
| Q4_K_M | `deepseek_sunfall_merged_Model.Q4_K_M.gguf` | Balanced speed & performance | |
|
|
| Q5_K_M | `deepseek_sunfall_merged_Model.Q5_K_M.gguf` | Better quality, slower | |
|
|
| Q6_K | `deepseek_sunfall_merged_Model.Q6_K.gguf` | Near full precision | |
|
|
| Q8_0 | `deepseek_sunfall_merged_Model.Q8_0.gguf` | Almost no compression loss | |
|
|
|
|
|
## 🔧 Usage (Python) |
|
|
|
|
|
Install `llama-cpp-python`: |
|
|
|
|
|
```bash |
|
|
pip install llama-cpp-python |
|
|
from llama_cpp import Llama |
|
|
|
|
|
model = Llama(model_path="deepseek_sunfall_merged_Model.Q4_K_M.gguf") # or Q3_K_M, etc. |
|
|
output = model("Tell me a story about stars.") |
|
|
print(output)``` |