kernelpool commited on
Commit
37cf26d
·
verified ·
1 Parent(s): 1b773a2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +36 -0
README.md CHANGED
@@ -8,3 +8,39 @@ base_model: MiniMaxAI/MiniMax-M2.5
8
  tags:
9
  - mlx
10
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  tags:
9
  - mlx
10
  ---
11
+
12
+ # catalystsec/MiniMax-M2.5-3bit-DWQ
13
+
14
+ This model was quantized to 3-bit using DWQ with mlx-lm version **0.30.7**.
15
+
16
+ | Parameter | Value |
17
+ |---------------------------|--------------------------------|
18
+ | DWQ learning rate | 3e-7 |
19
+ | Batch size | 1 |
20
+ | Dataset | `allenai/tulu-3-sft-mixture` |
21
+ | Initial validation loss | 0.183 |
22
+ | Final validation loss | 0.110 |
23
+ | Relative KL reduction | ≈40 % |
24
+ | Tokens processed | ≈1.11 M |
25
+
26
+ ## Use with mlx
27
+
28
+ ```bash
29
+ pip install mlx-lm
30
+ ```
31
+
32
+ ```python
33
+ from mlx_lm import load, generate
34
+
35
+ model, tokenizer = load("catalystsec/MiniMax-M2.5-3bit-DWQ")
36
+ prompt = "hello"
37
+
38
+ if tokenizer.chat_template is not None:
39
+ prompt = tokenizer.apply_chat_template(
40
+ [{"role": "user", "content": prompt}],
41
+ add_generation_prompt=True,
42
+ )
43
+
44
+ response = generate(model, tokenizer, prompt=prompt, verbose=True)
45
+ print(response)
46
+ ```