| | --- |
| | language: |
| | - en |
| | license: mit |
| | pipeline_tag: text-generation |
| | tags: |
| | - mlx |
| | - mixture-of-experts |
| | - moe |
| | - pruning |
| | - reap |
| | - minimax |
| | - 4bit |
| | - quantized |
| | - apple-silicon |
| | library_name: mlx |
| | base_model: Akicou/MiniMax-M2-5-REAP-39 |
| | --- |
| | |
| | # MiniMax-M2.5 REAP-39 — MLX 4-bit |
| |
|
| | MLX 4-bit quantized version of [Akicou/MiniMax-M2-5-REAP-39](https://huggingface.co/Akicou/MiniMax-M2-5-REAP-39) for efficient local inference on Apple Silicon. |
| |
|
| | - **Quantization**: 4-bit (4.5 bits per weight, group size 64, affine mode) |
| | - **Architecture**: MiniMax M2.5 MoE — 62 layers, 154 experts (REAP-pruned from 256), 8 active per token |
| | - **Context**: 196K tokens |
| | - **Size**: ~73 GB |
| | - **Pruning**: 39% of experts removed via [REAP](https://github.com/CerebrasResearch/reap) (Router Expert Activation Pruning) |
| |
|
| | ## Usage |
| |
|
| | ```python |
| | from mlx_lm import load, generate |
| | |
| | model, tokenizer = load("shieldstackllc/MiniMax-M2-5-REAP-39-mlx-4bit") |
| | response = generate(model, tokenizer, prompt="Hello!", verbose=True) |
| | ``` |
| |
|
| | Or with [vMLX](https://vmlx.net) for native macOS inference. |
| |
|
| | ## About |
| |
|
| | MiniMax-M2.5 is a large Mixture-of-Experts language model by MiniMax AI. This variant was pruned to 39% fewer experts by [Akicou](https://huggingface.co/Akicou) using REAP (Router Expert Activation Pruning), reducing model size and memory footprint while maintaining strong performance. MLX quantization by [vMLX](https://vmlx.net). |
| |
|
| | ## Also Available |
| |
|
| | - [MiniMax-M2.5-REAP-39 MLX 8-bit](https://huggingface.co/shieldstackllc/MiniMax-M2-5-REAP-39-mlx-8bit) (~131 GB) |
| |
|
| | ## Made for vMLX |
| |
|
| | This model was converted and optimized for [vMLX](https://vmlx.net) — a free, open source macOS native MLX inference engine for Apple Silicon. Download vMLX to run this model locally with zero configuration. |
| |
|
| | ## Credits |
| |
|
| | - **Base model**: [MiniMaxAI/MiniMax-M2.5](https://huggingface.co/MiniMaxAI/MiniMax-M2.5) by MiniMax AI |
| | - **REAP pruning**: [Akicou/MiniMax-M2-5-REAP-39](https://huggingface.co/Akicou/MiniMax-M2-5-REAP-39) by Akicou |
| | - **MLX conversion**: [vMLX](https://vmlx.net) — Run AI locally on Mac. No compromises. |
| |
|
| | ## Contact |
| |
|
| | For questions, issues, or collaboration: **admin@vmlx.net** |
| |
|