alexmarques commited on
Commit
2efb9ef
·
verified ·
1 Parent(s): bdee684

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -89,7 +89,7 @@ from llmcompressor.modeling.minimax_m2 import ( # noqa: F401
89
  from llmcompressor.modifiers.quantization import QuantizationModifier
90
 
91
  # Load the model
92
- model_id = "inference-optimization/MiniMax-M2.5-BF16"
93
  config = AutoConfig.from_pretrained(model_id, trust_remote_code=True)
94
  model = AutoModelForCausalLM.from_pretrained(
95
  model_id, dtype=torch.bfloat16, config=config,trust_remote_code=True
@@ -253,7 +253,7 @@ The model was evaluated on the ifeval, mmlu_pro and gsm8k_platinum using [lm-ev
253
 
254
  ### Accuracy
255
 
256
- | Benchmark | inference-optimization/MiniMax-M2.5-BF16 | inference-optimization/MiniMax-M2.5-NVFP4 | Recovery (%) |
257
  |-----------|------------------------------------------|-------------------------------------------|--------------|
258
  | GSM8k Platinum (0-shot) | 95.15 | 93.91 | 98.70 |
259
  | IfEval (0-shot) | 92.05 | 89.89 | 97.66 |
 
89
  from llmcompressor.modifiers.quantization import QuantizationModifier
90
 
91
  # Load the model
92
+ model_id = "RedHatAI/MiniMax-M2.5-BF16"
93
  config = AutoConfig.from_pretrained(model_id, trust_remote_code=True)
94
  model = AutoModelForCausalLM.from_pretrained(
95
  model_id, dtype=torch.bfloat16, config=config,trust_remote_code=True
 
253
 
254
  ### Accuracy
255
 
256
+ | Benchmark | RedHatAI/MiniMax-M2.5-BF16 | RedHatAI/MiniMax-M2.5-NVFP4 | Recovery (%) |
257
  |-----------|------------------------------------------|-------------------------------------------|--------------|
258
  | GSM8k Platinum (0-shot) | 95.15 | 93.91 | 98.70 |
259
  | IfEval (0-shot) | 92.05 | 89.89 | 97.66 |