You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Parameters Architecture Context License

MiniMax-M2.5-PRISM-PRO

A Powerful Production ready fully uncessored model intended for COMPLETE over-refusal and propaganda mechanisms suppression using our SOTA PRISM-PRO pipeline.

PRISM-PRO is available for purchase: https://ko-fi.com/s/0a23d1b9a5

For Custom trained PRISM versions or raw tensors access reach out @ https://ko-fi.com/ex0bit.

β˜• Support Our Work

If you enjoy our work and find it useful, please consider sponsoring or supporting us!

Ko-fi

Option Description
PRISM PRO VIP Membership Access to all PRISM models
Bitcoin bc1qarq2pyn4psjpcxzp2ghgwaq6y2h4e53q232x8r

image


Model Highlights

  • PRISM Ablation β€” State-of-the-art technique that removes over-refusal behaviors while preserving model capabilities
  • SOTA Coding Performance β€” 80.2% on SWE-Bench Verified, 51.3% on Multi-SWE-Bench, 76.3% on BrowseComp (with context management)
  • Frontier Agentic Capabilities β€” Industry-leading performance in tool use, search, and complex multi-step tasks
  • Efficient Reasoning β€” Trained with RL to reason efficiently and decompose tasks optimally, 37% faster than M2.1
  • Cost-Effective β€” $1 for continuous operation at 100 tok/s for an hour; $0.30 at 50 tok/s
  • Modified-MIT Base License β€” Based on MiniMax's open-weight release

Base Model Architecture

Base MiniMax-M2.5 is a Mixture-of-Experts (MoE) model extensively trained with reinforcement learning across hundreds of thousands of complex real-world environments.

Specification Value
Architecture Sparse Mixture-of-Experts (MoE)
Training Extensive RL in 200K+ real-world environments
Languages 10+ (Go, C, C++, TypeScript, Rust, Kotlin, Python, Java, JavaScript, PHP, Lua, Dart, Ruby)
Inference Speed 100 tok/s (Lightning) / 50 tok/s (Standard)
Library transformers

Benchmarks (Base Model)

Coding

Benchmark MiniMax-M2.5 Claude Opus 4.6 Gemini 3 Pro GPT-5.2
SWE-Bench Verified 80.2 78.9 74.0 72.6
Multi-SWE-Bench 51.3 50.8 β€” β€”
SWE-Bench Multilingual 55.6 β€” β€” β€”
Terminal-Bench 2.0 51.5 52.1 β€” β€”

Search & Tool Calling

Benchmark MiniMax-M2.5 Claude Opus 4.6 Gemini 3 Pro GPT-5.2
BrowseComp 76.3 71.2 62.4 57.8

Reasoning & Knowledge

Benchmark MiniMax-M2.5 Claude Opus 4.6 Gemini 3 Pro GPT-5.2
AIME25 86.3 95.6 96.0 98.0
GPQA-D 85.2 90.0 91.0 90.0
HLE w/o tools 19.4 30.7 37.2 31.4
SciCode 44.4 52.0 56.0 52.0
IFBench 70.0 53.0 70.0 75.0

Usage

llama.cpp (GGUF)

Build the latest master of llama.cpp and run:

~/llama.cpp/build/bin/llama-cli \
  -m ../outputs/MiniMax-M2.5-PRISM-PRO-[QUANT].gguf \
  --jinja \
  -ngl 999 \
  --repeat_penalty 1.15 \
  --temp 1.0 \
  --top_p 0.95 \
  --top_k 40

Replace [QUANT] with your quantization level (e.g. Q8_0, etc.).

Recommended Parameters

Use Case Temperature Top-P Top-K Repeat Penalty Max New Tokens
Reasoning / Coding 1.0 0.95 40 1.15 32768
General Chat 0.6 0.95 40 1.15 4096
Agentic / Tool Use 1.0 0.95 40 1.15 32768
Version Description Access
PRISM-LITE Abliterated with PRISM-LITE pipeline β€” removes over-refusal while preserving core capabilities Free on Hugging Face
PRISM-PRO Full PRISM-PRO ablation β€” Full Production Level Mode suppression of propaganda/refusal mechanisms with maximum capability retention Ko-fi

License

This model is released under the PRISM Research License.

The base model MiniMax-M2.5 is released under a Modified-MIT License.

Acknowledgments

Based on MiniMax-M2.5 by MiniMax AI.

Downloads last month
-
GGUF
Model size
229B params
Architecture
minimax-m2
Hardware compatibility
Log In to add your hardware

1-bit

2-bit

3-bit

4-bit

6-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for Ex0bit/MiniMax-M2.5-PRISM-PRO

Finetuned
(5)
this model