File size: 1,363 Bytes
1afac07
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
---
license: apache-2.0
base_model: MiniMaxAI/MiniMax-M2.7
base_model_relation: quantized
tags:
  - auto-round
  - int4
  - w4a16
  - quantization
  - moe
library_name: transformers
---

# MiniMax-M2.7 INT4 AutoRound

4-bit quantized version of [MiniMaxAI/MiniMax-M2.7](https://huggingface.co/MiniMaxAI/MiniMax-M2.7) using [Intel AutoRound](https://github.com/intel/auto-round).

## Quantization Config

| Setting | Value |
|---|---|
| Scheme | W4A16 (INT4 weights, FP16 activations) |
| Group size | 128 |
| Ignored layers | MoE `gate` layers (kept at full precision) |
| Method | RTN (iters=0) |

## Usage

### vLLM

```bash
vllm serve Lasimeri/MiniMax-M2.7-int4-AutoRound \
  --trust-remote-code \
  --tensor-parallel-size 8 \
  --enable-auto-tool-choice \
  --tool-call-parser minimax_m2 \
  --reasoning-parser minimax_m2_append_think
```

### SGLang

```bash
python -m sglang.launch_server \
  --model-path Lasimeri/MiniMax-M2.7-int4-AutoRound \
  --trust-remote-code \
  --tp 8 \
  --reasoning-parser minimax-append-think \
  --tool-call-parser minimax-m2
```

## Quantization Hardware

Quantized on a single-node rig:

| Component | Spec |
|---|---|
| CPU | AMD EPYC 7742 (64C / 128T) |
| RAM | 251 GB DDR4 |
| GPUs | 8× RTX 3080 (20 GB modded) |

Peak resource usage during quantization: ~25.6 GB RAM, ~5 GB VRAM on GPU 0, ~1.3 GB on each remaining GPU.