fmasterpro27 commited on
Commit
50a910d
·
verified ·
1 Parent(s): 7e79404

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +67 -1
README.md CHANGED
@@ -3,5 +3,71 @@ license: mit
3
  library_name: transformers
4
  tags:
5
  - mlx
 
6
  base_model: deepseek-ai/DeepSeek-R1
7
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  library_name: transformers
4
  tags:
5
  - mlx
6
+ - open4bits
7
  base_model: deepseek-ai/DeepSeek-R1
8
+ pipeline_tag: text-generation
9
+ ---
10
+
11
+ # Open4bits / DeepSeek-R1-MLX-2Bit
12
+
13
+ This repository provides the **DeepSeek-R1 model quantized to 2-bit in MLX format**, published by Open4bits to enable highly efficient local inference with minimal memory usage and broad hardware compatibility.
14
+
15
+ The underlying DeepSeek-R1 model and architecture are **developed and owned by DeepSeek AI**. This repository contains only a 2-bit quantized MLX conversion of the original model weights.
16
+
17
+ The model is designed for lightweight, high-performance text generation and instruction-following tasks, making it well suited for resource-constrained and local deployments.
18
+
19
+ ---
20
+
21
+ ## Model Overview
22
+
23
+ DeepSeek-R1 is a transformer-based large language model developed for strong general language understanding and generation.
24
+ This release provides a **2-bit quantized checkpoint in MLX format**, enabling efficient inference on CPUs and supported accelerators with reduced memory footprint.
25
+
26
+ Open4bits has started supporting **MLX models** to broaden compatibility with emerging quantization formats and efficient runtimes.
27
+
28
+ ---
29
+
30
+ ## Model Details
31
+
32
+ * **Base Model:** DeepSeek-R1
33
+ * **Quantization:** 2-bit
34
+ * **Format:** MLX
35
+ * **Task:** Text generation, instruction following
36
+ * **Weight tying:** Preserved
37
+ * **Compatibility:** MLX-enabled inference engines and efficient runtimes
38
+
39
+ This quantized release is designed to balance strong generation performance with low resource requirements.
40
+
41
+ ---
42
+
43
+ ## Intended Use
44
+
45
+ This model is intended for:
46
+
47
+ * Local text generation and conversational applications
48
+ * CPU-based or low-resource deployments
49
+ * Research, prototyping, and experimentation
50
+ * Self-hosted or offline AI systems
51
+
52
+ ---
53
+
54
+ ## Limitations
55
+
56
+ * Reduced performance compared to full-precision variants
57
+ * Output quality depends on prompt design and inference settings
58
+ * Not specifically tuned for highly specialized or domain-specific tasks
59
+
60
+ ---
61
+
62
+ ## License
63
+
64
+ This model follows the **MIT** as defined by the base model creators.
65
+ Users must comply with the licensing conditions of the base DeepSeek-R1 model.
66
+
67
+ ---
68
+
69
+ ## Support
70
+
71
+ If you find this model useful, please consider supporting the project.
72
+ Your support helps Open4bits continue releasing and maintaining high-quality, efficient open models for the community.
73
+