bibproj commited on
Commit
8ef85e9
·
verified ·
1 Parent(s): 04b5b48

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +39 -1
README.md CHANGED
@@ -9,4 +9,42 @@ license_name: modified-mit
9
  license_link: https://github.com/MiniMax-AI/MiniMax-M2.1/blob/main/LICENSE
10
  base_model:
11
  - MiniMaxAI/MiniMax-M2.1
12
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  license_link: https://github.com/MiniMax-AI/MiniMax-M2.1/blob/main/LICENSE
10
  base_model:
11
  - MiniMaxAI/MiniMax-M2.1
12
+ ---
13
+
14
+ # mlx-community/MiniMax-M2.1-8bit-gs32
15
+
16
+ This model [mlx-community/MiniMax-M2.1-8bit-gs32](https://huggingface.co/mlx-community/MiniMax-M2.1-8bit-gs32) was
17
+ converted to MLX format from [MiniMaxAI/MiniMax-M2.1](https://huggingface.co/MiniMaxAI/MiniMax-M2.1)
18
+ using mlx-lm version **0.29.1**.
19
+
20
+
21
+ ## Recipe:
22
+ * 8-bit
23
+ * group-size 32
24
+ * 9 bits per weight (bpw)
25
+
26
+ You can find more similar MLX model quants for a single Apple Mac Studio M3 Ultra with 512 GB at https://huggingface.co/bibproj
27
+
28
+ ---
29
+
30
+ ## Use with mlx
31
+
32
+ ```bash
33
+ pip install mlx-lm
34
+ ```
35
+
36
+ ```python
37
+ from mlx_lm import load, generate
38
+
39
+ model, tokenizer = load("mlx-community/MiniMax-M2.1-8bit-gs32")
40
+
41
+ prompt = "hello"
42
+
43
+ if tokenizer.chat_template is not None:
44
+ messages = [{"role": "user", "content": prompt}]
45
+ prompt = tokenizer.apply_chat_template(
46
+ messages, add_generation_prompt=True
47
+ )
48
+
49
+ response = generate(model, tokenizer, prompt=prompt, verbose=True)
50
+ ```