Text Generation
MLX
Safetensors
minimax_m2
minimax
m2.7
Mixture of Experts
quantized
rotorquant
kv-cache-quantization
conversational
custom_code
5-bit
Instructions to use majentik/MiniMax-M2.7-RotorQuant-MLX-5bit with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- MLX
How to use majentik/MiniMax-M2.7-RotorQuant-MLX-5bit with MLX:
# Make sure mlx-lm is installed # pip install --upgrade mlx-lm # Generate text with mlx-lm from mlx_lm import load, generate model, tokenizer = load("majentik/MiniMax-M2.7-RotorQuant-MLX-5bit") prompt = "Write a story about Einstein" messages = [{"role": "user", "content": prompt}] prompt = tokenizer.apply_chat_template( messages, add_generation_prompt=True ) text = generate(model, tokenizer, prompt=prompt, verbose=True) - Notebooks
- Google Colab
- Kaggle
- Local Apps
- LM Studio
- Pi new
How to use majentik/MiniMax-M2.7-RotorQuant-MLX-5bit with Pi:
Start the MLX server
# Install MLX LM: uv tool install mlx-lm # Start a local OpenAI-compatible server: mlx_lm.server --model "majentik/MiniMax-M2.7-RotorQuant-MLX-5bit"
Configure the model in Pi
# Install Pi: npm install -g @mariozechner/pi-coding-agent # Add to ~/.pi/agent/models.json: { "providers": { "mlx-lm": { "baseUrl": "http://localhost:8080/v1", "api": "openai-completions", "apiKey": "none", "models": [ { "id": "majentik/MiniMax-M2.7-RotorQuant-MLX-5bit" } ] } } }Run Pi
# Start Pi in your project directory: pi
- Hermes Agent new
How to use majentik/MiniMax-M2.7-RotorQuant-MLX-5bit with Hermes Agent:
Start the MLX server
# Install MLX LM: uv tool install mlx-lm # Start a local OpenAI-compatible server: mlx_lm.server --model "majentik/MiniMax-M2.7-RotorQuant-MLX-5bit"
Configure Hermes
# Install Hermes: curl -fsSL https://hermes-agent.nousresearch.com/install.sh | bash hermes setup # Point Hermes at the local server: hermes config set model.provider custom hermes config set model.base_url http://127.0.0.1:8080/v1 hermes config set model.default majentik/MiniMax-M2.7-RotorQuant-MLX-5bit
Run Hermes
hermes
- MLX LM
How to use majentik/MiniMax-M2.7-RotorQuant-MLX-5bit with MLX LM:
Generate or start a chat session
# Install MLX LM uv tool install mlx-lm # Interactive chat REPL mlx_lm.chat --model "majentik/MiniMax-M2.7-RotorQuant-MLX-5bit"
Run an OpenAI-compatible server
# Install MLX LM uv tool install mlx-lm # Start the server mlx_lm.server --model "majentik/MiniMax-M2.7-RotorQuant-MLX-5bit" # Calling the OpenAI-compatible server with curl curl -X POST "http://localhost:8000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "majentik/MiniMax-M2.7-RotorQuant-MLX-5bit", "messages": [ {"role": "user", "content": "Hello"} ] }'
docs: Tier 2 polish — variant matrix + quant trade-off
Browse files
README.md
CHANGED
|
@@ -6,15 +6,13 @@ license: other
|
|
| 6 |
license_name: minimax-model-license
|
| 7 |
license_link: https://huggingface.co/MiniMaxAI/MiniMax-M2.7/blob/main/LICENSE
|
| 8 |
tags:
|
| 9 |
-
- minimax
|
| 10 |
-
- m2.7
|
| 11 |
-
- moe
|
| 12 |
-
- quantized
|
| 13 |
-
- rotorquant
|
| 14 |
-
- kv-cache-quantization
|
| 15 |
-
- mlx
|
| 16 |
-
language:
|
| 17 |
-
- en
|
| 18 |
---
|
| 19 |
|
| 20 |
# MiniMax-M2.7-RotorQuant-MLX-5bit
|
|
@@ -91,3 +89,35 @@ print(response)
|
|
| 91 |
- [majentik/MiniMax-M2.7-TurboQuant-MLX-5bit](https://huggingface.co/majentik/MiniMax-M2.7-TurboQuant-MLX-5bit) -- TurboQuant MLX 5-bit
|
| 92 |
- [majentik/MiniMax-M2.7-RotorQuant-MLX-8bit](https://huggingface.co/majentik/MiniMax-M2.7-RotorQuant-MLX-8bit) -- MLX 8-bit
|
| 93 |
- [majentik/MiniMax-M2.7-RotorQuant-MLX-4bit](https://huggingface.co/majentik/MiniMax-M2.7-RotorQuant-MLX-4bit) -- MLX 4-bit
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
license_name: minimax-model-license
|
| 7 |
license_link: https://huggingface.co/MiniMaxAI/MiniMax-M2.7/blob/main/LICENSE
|
| 8 |
tags:
|
| 9 |
+
- minimax
|
| 10 |
+
- m2.7
|
| 11 |
+
- moe
|
| 12 |
+
- quantized
|
| 13 |
+
- rotorquant
|
| 14 |
+
- kv-cache-quantization
|
| 15 |
+
- mlx
|
|
|
|
|
|
|
| 16 |
---
|
| 17 |
|
| 18 |
# MiniMax-M2.7-RotorQuant-MLX-5bit
|
|
|
|
| 89 |
- [majentik/MiniMax-M2.7-TurboQuant-MLX-5bit](https://huggingface.co/majentik/MiniMax-M2.7-TurboQuant-MLX-5bit) -- TurboQuant MLX 5-bit
|
| 90 |
- [majentik/MiniMax-M2.7-RotorQuant-MLX-8bit](https://huggingface.co/majentik/MiniMax-M2.7-RotorQuant-MLX-8bit) -- MLX 8-bit
|
| 91 |
- [majentik/MiniMax-M2.7-RotorQuant-MLX-4bit](https://huggingface.co/majentik/MiniMax-M2.7-RotorQuant-MLX-4bit) -- MLX 4-bit
|
| 92 |
+
|
| 93 |
+
## Quant trade-off (MLX lane)
|
| 94 |
+
|
| 95 |
+
| Bits | Approx size | Use case | Recommendation |
|
| 96 |
+
|---|---|---|---|
|
| 97 |
+
| 2-bit | ~119 GB | Aggressive quantization | Very low-RAM Macs |
|
| 98 |
+
| 3-bit | ~164 GB | Lossy but small | Low-RAM Macs |
|
| 99 |
+
| 4-bit | ~192 GB | Balanced default | Recommended for most Macs |
|
| 100 |
+
| **5-bit** | ~228 GB | Higher fidelity | **Quality-sensitive** |
|
| 101 |
+
| 6-bit | ~274 GB | Approaching FP16 quality | High-fidelity |
|
| 102 |
+
| 8-bit | ~347 GB | Near-lossless reference | Fidelity-critical work |
|
| 103 |
+
|
| 104 |
+
(Current variant — **5bit** — is bolded.)
|
| 105 |
+
|
| 106 |
+
## Variants in this family
|
| 107 |
+
|
| 108 |
+
(Showing 12 sibling variants under `majentik/minimax-m2.7-*`. The current variant — `RotorQuant-MLX-5bit` — is **bolded**.)
|
| 109 |
+
|
| 110 |
+
| Variant | Runtime | Approx size | Use case |
|
| 111 |
+
|---|---|---|---|
|
| 112 |
+
| [RotorQuant](https://huggingface.co/majentik/minimax-m2.7-rotorquant) | runtime modifier | n/a | KV-cache root (weight-agnostic) |
|
| 113 |
+
| [RotorQuant-MLX-2bit](https://huggingface.co/majentik/minimax-m2.7-rotorquant-mlx-2bit) | mlx-lm | ~885 MB | Apple Silicon, smallest |
|
| 114 |
+
| [RotorQuant-MLX-3bit](https://huggingface.co/majentik/minimax-m2.7-rotorquant-mlx-3bit) | mlx-lm | ~1.2 GB | Apple Silicon, small |
|
| 115 |
+
| [RotorQuant-MLX-4bit](https://huggingface.co/majentik/minimax-m2.7-rotorquant-mlx-4bit) | mlx-lm | ~1.7 GB | Apple Silicon balanced |
|
| 116 |
+
| **RotorQuant-MLX-5bit** | mlx-lm | ~2.1 GB | Apple Silicon, higher fidelity |
|
| 117 |
+
| [RotorQuant-MLX-8bit](https://huggingface.co/majentik/minimax-m2.7-rotorquant-mlx-8bit) | mlx-lm | ~3.2 GB | Apple Silicon reference |
|
| 118 |
+
| [TurboQuant](https://huggingface.co/majentik/minimax-m2.7-turboquant) | runtime modifier | n/a | KV-cache root (weight-agnostic) |
|
| 119 |
+
| [TurboQuant-MLX-2bit](https://huggingface.co/majentik/minimax-m2.7-turboquant-mlx-2bit) | mlx-lm | ~885 MB | Apple Silicon, smallest |
|
| 120 |
+
| [TurboQuant-MLX-3bit](https://huggingface.co/majentik/minimax-m2.7-turboquant-mlx-3bit) | mlx-lm | ~1.2 GB | Apple Silicon, small |
|
| 121 |
+
| [TurboQuant-MLX-4bit](https://huggingface.co/majentik/minimax-m2.7-turboquant-mlx-4bit) | mlx-lm | ~1.7 GB | Apple Silicon balanced |
|
| 122 |
+
| [TurboQuant-MLX-5bit](https://huggingface.co/majentik/minimax-m2.7-turboquant-mlx-5bit) | mlx-lm | ~2.1 GB | Apple Silicon, higher fidelity |
|
| 123 |
+
| [TurboQuant-MLX-8bit](https://huggingface.co/majentik/minimax-m2.7-turboquant-mlx-8bit) | mlx-lm | ~3.2 GB | Apple Silicon reference |
|