Update README.md
Browse files
README.md
CHANGED
|
@@ -89,7 +89,7 @@ from llmcompressor.modeling.minimax_m2 import ( # noqa: F401
|
|
| 89 |
from llmcompressor.modifiers.quantization import QuantizationModifier
|
| 90 |
|
| 91 |
# Load the model
|
| 92 |
-
model_id = "
|
| 93 |
config = AutoConfig.from_pretrained(model_id, trust_remote_code=True)
|
| 94 |
model = AutoModelForCausalLM.from_pretrained(
|
| 95 |
model_id, dtype=torch.bfloat16, config=config,trust_remote_code=True
|
|
@@ -253,7 +253,7 @@ The model was evaluated on the ifeval, mmlu_pro and gsm8k_platinum using [lm-ev
|
|
| 253 |
|
| 254 |
### Accuracy
|
| 255 |
|
| 256 |
-
| Benchmark |
|
| 257 |
|-----------|------------------------------------------|-------------------------------------------|--------------|
|
| 258 |
| GSM8k Platinum (0-shot) | 95.15 | 93.91 | 98.70 |
|
| 259 |
| IfEval (0-shot) | 92.05 | 89.89 | 97.66 |
|
|
|
|
| 89 |
from llmcompressor.modifiers.quantization import QuantizationModifier
|
| 90 |
|
| 91 |
# Load the model
|
| 92 |
+
model_id = "RedHatAI/MiniMax-M2.5-BF16"
|
| 93 |
config = AutoConfig.from_pretrained(model_id, trust_remote_code=True)
|
| 94 |
model = AutoModelForCausalLM.from_pretrained(
|
| 95 |
model_id, dtype=torch.bfloat16, config=config,trust_remote_code=True
|
|
|
|
| 253 |
|
| 254 |
### Accuracy
|
| 255 |
|
| 256 |
+
| Benchmark | RedHatAI/MiniMax-M2.5-BF16 | RedHatAI/MiniMax-M2.5-NVFP4 | Recovery (%) |
|
| 257 |
|-----------|------------------------------------------|-------------------------------------------|--------------|
|
| 258 |
| GSM8k Platinum (0-shot) | 95.15 | 93.91 | 98.70 |
|
| 259 |
| IfEval (0-shot) | 92.05 | 89.89 | 97.66 |
|