shieldstackllc commited on
Commit
c2e0389
·
verified ·
1 Parent(s): fa88915

Add vMLX model card

Browse files
Files changed (1) hide show
  1. README.md +68 -0
README.md ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: mit
5
+ pipeline_tag: text-generation
6
+ tags:
7
+ - mlx
8
+ - mixture-of-experts
9
+ - moe
10
+ - pruning
11
+ - reap
12
+ - minimax
13
+ - 4bit
14
+ - quantized
15
+ - apple-silicon
16
+ library_name: mlx
17
+ base_model: Akicou/MiniMax-M2-5-REAP-29
18
+ ---
19
+
20
+ <p align="center">
21
+ <a href="https://vmlx.net">
22
+ <img src="vmlx-logo.png" alt="vMLX" width="120">
23
+ </a>
24
+ </p>
25
+
26
+ # MiniMax-M2.5 REAP-29 — MLX 4-bit
27
+
28
+ MLX 4-bit quantized version of [Akicou/MiniMax-M2-5-REAP-29](https://huggingface.co/Akicou/MiniMax-M2-5-REAP-29) for efficient local inference on Apple Silicon.
29
+
30
+ - **Quantization**: 4-bit (group size 64, affine mode; router gates at 8-bit)
31
+ - **Architecture**: MiniMax M2.5 MoE — 62 layers, 180 experts (REAP-pruned from 256), 8 active per token
32
+ - **Context**: 196K tokens
33
+ - **Size**: ~85 GB
34
+ - **Pruning**: 29% of experts removed via [REAP](https://github.com/CerebrasResearch/reap) (Router Expert Activation Pruning)
35
+
36
+ ## Usage
37
+
38
+ ```python
39
+ from mlx_lm import load, generate
40
+
41
+ model, tokenizer = load("shieldstackllc/MiniMax-M2.5-REAP-29-mlx-4bit")
42
+ response = generate(model, tokenizer, prompt="Hello!", verbose=True)
43
+ ```
44
+
45
+ Or with [vMLX](https://vmlx.net) for native macOS inference.
46
+
47
+ ## About
48
+
49
+ MiniMax-M2.5 is a large Mixture-of-Experts language model by MiniMax AI. This variant was pruned to 29% fewer experts by [Akicou](https://huggingface.co/Akicou) using REAP (Router Expert Activation Pruning), reducing model size and memory footprint while maintaining strong performance. MLX quantization by [vMLX](https://vmlx.net).
50
+
51
+ ## Also Available
52
+
53
+ - [MiniMax-M2.5-REAP-39 MLX 4-bit](https://huggingface.co/shieldstackllc/MiniMax-M2-5-REAP-39-mlx-4bit) (~73 GB) — 39% pruned variant
54
+ - [MiniMax-M2.5-REAP-39 MLX 8-bit](https://huggingface.co/shieldstackllc/MiniMax-M2-5-REAP-39-mlx-8bit) (~138 GB) — 39% pruned variant
55
+
56
+ ## Made for vMLX
57
+
58
+ This model was converted and optimized for [vMLX](https://vmlx.net) — a free, open source macOS native MLX inference engine for Apple Silicon. Download vMLX to run this model locally with zero configuration.
59
+
60
+ ## Credits
61
+
62
+ - **Base model**: [MiniMaxAI/MiniMax-M2.5](https://huggingface.co/MiniMaxAI/MiniMax-M2.5) by MiniMax AI
63
+ - **REAP pruning**: [Akicou/MiniMax-M2-5-REAP-29](https://huggingface.co/Akicou/MiniMax-M2-5-REAP-29) by Akicou
64
+ - **MLX conversion**: [vMLX](https://vmlx.net) — Run AI locally on Mac. No compromises.
65
+
66
+ ## Contact
67
+
68
+ For questions, issues, or collaboration: **admin@vmlx.net**