Baskar2005 commited on
Commit
786c071
·
verified ·
1 Parent(s): eefb091

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +3 -26
README.md CHANGED
@@ -4,33 +4,10 @@ tags:
4
  - gguf
5
  - llama.cpp
6
  - quantized
7
- - deepseek
8
- - stheno
9
  ---
10
 
11
- # DeepSeek Sunfall Merged - GGUF Quantized Models
12
 
13
- This repository contains multiple **quantized GGUF variants** of the merged DeepSeek + Sunfall model, compatible with `llama.cpp`.
14
 
15
- ## 🧠 Available Quantized Formats
16
-
17
- | Format | File Name | Description |
18
- |-------------|--------------------------------------------------|---------------------------------|
19
- | Q3_K_M | `deepseek_sunfall_merged_Model.Q3_K_M.gguf` | Smallest size, fastest inference |
20
- | Q4_K_M | `deepseek_sunfall_merged_Model.Q4_K_M.gguf` | Balanced speed & performance |
21
- | Q5_K_M | `deepseek_sunfall_merged_Model.Q5_K_M.gguf` | Better quality, slower |
22
- | Q6_K | `deepseek_sunfall_merged_Model.Q6_K.gguf` | Near full precision |
23
- | Q8_0 | `deepseek_sunfall_merged_Model.Q8_0.gguf` | Almost no compression loss |
24
-
25
- ## 🔧 Usage (Python)
26
-
27
- Install `llama-cpp-python`:
28
-
29
- ```bash
30
- pip install llama-cpp-python
31
- from llama_cpp import Llama
32
-
33
- model = Llama(model_path="deepseek_sunfall_merged_Model.Q4_K_M.gguf") # or Q3_K_M, etc.
34
- output = model("Tell me a story about stars.")
35
- print(output)
36
- ```
 
4
  - gguf
5
  - llama.cpp
6
  - quantized
 
 
7
  ---
8
 
9
+ # deepseek_sunfall_merged_Model.Q3_K_M
10
 
11
+ This repository contains the `deepseek_sunfall_merged_Model.Q3_K_M.gguf` model in GGUF format, quantized with llama.cpp.
12
 
13
+ ```