figodeng commited on
Commit
d92e9f3
·
verified ·
1 Parent(s): ec48d34

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -0
README.md ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # DeepSeek-R1-Distill-Llama-70B-3bit
2
+ The Model DeepSeek-R1-Distill-Llama-70B-3bit was converted to MLX format from deepseek-ai/DeepSeek-R1-Distill-Llama-70B using mlx-lm version 0.21.1.
3
+
4
+ # Use with mlx
5
+
6
+ ```
7
+ pip install mlx-lm
8
+ ```
9
+
10
+ ```
11
+ from mlx_lm import load, generate
12
+
13
+ model, tokenizer = load("figodeng/DeepSeek-R1-Distill-Llama-70B-3bit")
14
+
15
+ prompt = "hello"
16
+
17
+ if tokenizer.chat_template is not None:
18
+ messages = [{"role": "user", "content": prompt}]
19
+ prompt = tokenizer.apply_chat_template(
20
+ messages, add_generation_prompt=True
21
+ )
22
+
23
+ response = generate(model, tokenizer, prompt=prompt, verbose=True)
24
+
25
+ ```