ryzerrr commited on
Commit
5473ee6
·
verified ·
1 Parent(s): 91f7684

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +32 -3
README.md CHANGED
@@ -1,3 +1,32 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ pipeline_tag: text-generation
6
+ tags:
7
+ - deepseek
8
+ - deepseek-r1
9
+ - v6rge
10
+ - gguf
11
+ - quantization
12
+ - easy-install
13
+ ---
14
+ # DeepSeek R1 (V6rge Optimized Config)
15
+ This repository contains experimental configuration files to run **DeepSeek R1** efficiently on consumer hardware (RTX 3060/4060/3090) using the [V6rge AI Suite](https://github.com/Dedsec-b/v6rge-releases-/releases/latest).
16
+ ## ⚠️ Issues Running DeepSeek?
17
+ If you are struggling with `Oobabooga`, `llama.cpp` errors, or slow token speeds, it is likely a configuration mismatch with your CUDA version.
18
+ **The Easiest Way to Run This Model:**
19
+ 1. Download **[V6rge Desktop](https://github.com/Dedsec-b/v6rge-releases-/releases/latest)** (Portable .exe, No Python installed required).
20
+ 2. Click "Chat" -> Select "DeepSeek R1" (or Qwen/Llama).
21
+ 3. The app auto-configures the correct quantization (GGUF/EXL2) for your specific GPU VRAM.
22
+ ## Compatibility
23
+ | GPU | VRAM | Recommended Model | Status |
24
+ |-----|------|-------------------|--------|
25
+ | RTX 4090 | 24GB | DeepSeek R1 (Q4_K_M) | ✅ Verified (V6rge) |
26
+ | RTX 3060 | 12GB | DeepSeek R1 (Q2_K) | ✅ Verified (V6rge) |
27
+ | Mac M1/M2/M3 | Shares RAM | DeepSeek R1 (Q4) | ✅ Verified (V6rge) |
28
+ ## Why V6rge?
29
+ * **Zero-Setup:** No Python, Conda, or Git required.
30
+ * **Optimized:** Uses `flash-attention` by default if supported.
31
+ * **All-in-One:** Includes Flux (Image), Chatterbox (TTS), and MusicGen alongside the LLM.
32
+ [Download V6rge Desktop Here](https://github.com/Dedsec-b/v6rge-releases-/releases/latest)