Parameters Architecture Context License

MiniMax-M2.5-PRISM-LITE

A PRISM-LITE version of Ex0bit/MiniMax-M2.5-PRISM-PRO intended for role-following over-refusal and propaganda mechanisms suppression using our SOTA PRISM pipeline.

PRISM-PRO version available for purchase here: https://ko-fi.com/s/0a23d1b9a5

For Full Custom trained PRISM versions or raw tensors access reach out @ https://ko-fi.com/ex0bit.

β˜• Support Our Work

If you enjoy our work and find it useful, please consider sponsoring or supporting us!

Ko-fi

Option Description
PRISM PRO VIP Membership Access to all PRISM models
Bitcoin bc1qarq2pyn4psjpcxzp2ghgwaq6y2h4e53q232x8r

image


Model Highlights

  • PRISM Ablation β€” State-of-the-art technique that removes over-refusal behaviors while preserving model capabilities
  • SOTA Coding Performance β€” 80.2% on SWE-Bench Verified, 51.3% on Multi-SWE-Bench, 76.3% on BrowseComp (with context management)
  • Frontier Agentic Capabilities β€” Industry-leading performance in tool use, search, and complex multi-step tasks
  • Efficient Reasoning β€” Trained with RL to reason efficiently and decompose tasks optimally, 37% faster than M2.1
  • Cost-Effective β€” $1 for continuous operation at 100 tok/s for an hour; $0.30 at 50 tok/s
  • Modified-MIT Base License β€” Based on MiniMax's open-weight release

Base Model Architecture

MiniMax-M2.5 is a Mixture-of-Experts (MoE) model extensively trained with reinforcement learning across hundreds of thousands of complex real-world environments.

Specification Value
Architecture Sparse Mixture-of-Experts (MoE)
Training Extensive RL in 200K+ real-world environments
Languages 10+ (Go, C, C++, TypeScript, Rust, Kotlin, Python, Java, JavaScript, PHP, Lua, Dart, Ruby)
Inference Speed 100 tok/s (Lightning) / 50 tok/s (Standard)
Library transformers

Benchmarks (Base Model)

Coding

Benchmark MiniMax-M2.5 Claude Opus 4.6 Gemini 3 Pro GPT-5.2
SWE-Bench Verified 80.2 78.9 74.0 72.6
Multi-SWE-Bench 51.3 50.8 β€” β€”
SWE-Bench Multilingual 55.6 β€” β€” β€”
Terminal-Bench 2.0 51.5 52.1 β€” β€”

Search & Tool Calling

Benchmark MiniMax-M2.5 Claude Opus 4.6 Gemini 3 Pro GPT-5.2
BrowseComp 76.3 71.2 62.4 57.8

Reasoning & Knowledge

Benchmark MiniMax-M2.5 Claude Opus 4.6 Gemini 3 Pro GPT-5.2
AIME25 86.3 95.6 96.0 98.0
GPQA-D 85.2 90.0 91.0 90.0
HLE w/o tools 19.4 30.7 37.2 31.4
SciCode 44.4 52.0 56.0 52.0
IFBench 70.0 53.0 70.0 75.0

Usage

llama.cpp (GGUF)

Build the latest master of llama.cpp and run:

~/llama.cpp/build/bin/llama-cli \
  -m ../outputs/MiniMax-M2.5-PRISM-PRO-[QUANT].gguf \
  --jinja \
  -ngl 999 \
  --repeat_penalty 1.15 \
  --temp 1.0 \
  --top_p 0.95 \
  --top_k 40

Replace [QUANT] with your quantization level (e.g. Q8_0, etc.).

Recommended Parameters

Use Case Temperature Top-P Top-K Repeat Penalty Max New Tokens
Reasoning / Coding 1.0 0.95 40 1.15 32768
General Chat 0.6 0.95 40 1.15 4096
Agentic / Tool Use 1.0 0.95 40 1.15 32768
Version Description Access
PRISM-LITE Abliterated with PRISM-LITE pipeline β€” removes over-refusal while preserving core capabilities Free on Hugging Face
PRISM-PRO Full PRISM-PRO ablation β€” Full Production Level Mode suppression of propaganda/refusal mechanisms with maximum capability retention Ko-fi

License

This model is released under the PRISM Research License.

The base model MiniMax-M2.5 is released under a Modified-MIT License.

Acknowledgments

Based on MiniMax-M2.5 by MiniMax AI.

Downloads last month
967
GGUF
Model size
229B params
Architecture
minimax-m2
Hardware compatibility
Log In to add your hardware

1-bit

2-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for Ex0bit/MiniMax-M2.5-PRISM-LITE

Finetuned
(5)
this model