|
|
--- |
|
|
license: apache-2.0 |
|
|
language: |
|
|
- en |
|
|
pipeline_tag: text-generation |
|
|
tags: |
|
|
- deepseek |
|
|
- deepseek-r1 |
|
|
- v6rge |
|
|
- gguf |
|
|
- quantization |
|
|
- easy-install |
|
|
--- |
|
|
# DeepSeek R1 (V6rge Optimized Config) |
|
|
This repository contains experimental configuration files to run **DeepSeek R1** efficiently on consumer hardware (RTX 3060/4060/3090) using the [V6rge AI Suite](https://github.com/Dedsec-b/v6rge-releases-/releases/latest). |
|
|
## ⚠️ Issues Running DeepSeek? |
|
|
If you are struggling with `Oobabooga`, `llama.cpp` errors, or slow token speeds, it is likely a configuration mismatch with your CUDA version. |
|
|
**The Easiest Way to Run This Model:** |
|
|
1. Download **[V6rge Desktop](https://github.com/Dedsec-b/v6rge-releases-/releases/latest)** (Portable .exe, No Python installed required). |
|
|
2. Click "Chat" -> Select "DeepSeek R1" (or Qwen/Llama). |
|
|
3. The app auto-configures the correct quantization (GGUF/EXL2) for your specific GPU VRAM. |
|
|
## Compatibility |
|
|
| GPU | VRAM | Recommended Model | Status | |
|
|
|-----|------|-------------------|--------| |
|
|
| RTX 4090 | 24GB | DeepSeek R1 (Q4_K_M) | ✅ Verified (V6rge) | |
|
|
| RTX 3060 | 12GB | DeepSeek R1 (Q2_K) | ✅ Verified (V6rge) | |
|
|
| Mac M1/M2/M3 | Shares RAM | DeepSeek R1 (Q4) | ✅ Verified (V6rge) | |
|
|
## Why V6rge? |
|
|
* **Zero-Setup:** No Python, Conda, or Git required. |
|
|
* **Optimized:** Uses `flash-attention` by default if supported. |
|
|
* **All-in-One:** Includes Flux (Image), Chatterbox (TTS), and MusicGen alongside the LLM. |
|
|
[Download V6rge Desktop Here](https://github.com/Dedsec-b/v6rge-releases-/releases/latest) |
|
|
|