Ex0bit commited on
Commit
5dd09db
Β·
verified Β·
1 Parent(s): ec3eb02

Create READNE.md

Browse files
Files changed (1) hide show
  1. READNE.md +152 -0
READNE.md ADDED
@@ -0,0 +1,152 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: prism-research
4
+ license_link: LICENSE.md
5
+ language:
6
+ - en
7
+ - zh
8
+ tags:
9
+ - minimax
10
+ - prism
11
+ - moe
12
+ - reasoning
13
+ - coding
14
+ - agentic
15
+ - abliterated
16
+ pipeline_tag: text-generation
17
+ library_name: transformers
18
+ base_model:
19
+ - MiniMaxAI/MiniMax-M2.5
20
+ base_model_relation: finetune
21
+ ---
22
+
23
+ [![Parameters](https://img.shields.io/badge/Parameters-MoE-blue)]()
24
+ [![Architecture](https://img.shields.io/badge/Architecture-MoE-green)]()
25
+ [![Context](https://img.shields.io/badge/Context-1M+-orange)]()
26
+ [![License](https://img.shields.io/badge/License-PRISM--Research-purple)]()
27
+
28
+
29
+ <p align="center">
30
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/63adf1fa42fd3b8dbaeb0c92/shxznHWnvppRhT_yKrsdP.png" width="400"/>
31
+ </p>
32
+
33
+
34
+ # MiniMax-M2.5-PRISM-LITE
35
+
36
+ A PRISM-LITE version of [Ex0bit/MiniMax-M2.5-PRISM-PRO](https://hf.co/Ex0bit/MiniMax-M2.5-PRISM-PRO) intended for role-following over-refusal and propaganda mechanisms suppression using our SOTA PRISM pipeline.
37
+
38
+ PRISM-PRO version available for purchase here: **https://ko-fi.com/s/0a23d1b9a5**
39
+
40
+ For Full Custom trained PRISM versions & or raw tensors acess reach out @ https://ko-fi.com/ex0bit.
41
+
42
+ <div align="center">
43
+
44
+ ### β˜• Support Our Work
45
+
46
+ If you enjoy our work and find it useful, please consider sponsoring or supporting us!
47
+
48
+ [![Ko-fi](https://img.shields.io/badge/Ko--fi-Support%20Us-ff5e5b?logo=ko-fi&logoColor=white)](https://ko-fi.com/ex0bit)
49
+
50
+ | Option | Description |
51
+ |--------|-------------|
52
+ | [**PRISM PRO VIP Membership**](https://ko-fi.com/summary/6bae206c-a751-4868-8dc7-f531afd1fb4c) | Access to all PRISM models |
53
+ | **Bitcoin** | `bc1qarq2pyn4psjpcxzp2ghgwaq6y2h4e53q232x8r` |
54
+
55
+ ![image](https://cdn-uploads.huggingface.co/production/uploads/63adf1fa42fd3b8dbaeb0c92/Psgbl1TgyDok__C7AMQog.png)
56
+
57
+ </div>
58
+
59
+ ---
60
+
61
+ ## Model Highlights
62
+
63
+ - **PRISM Ablation** β€” State-of-the-art technique that removes over-refusal behaviors while preserving model capabilities
64
+ - **SOTA Coding Performance** β€” 80.2% on SWE-Bench Verified, 51.3% on Multi-SWE-Bench, 76.3% on BrowseComp (with context management)
65
+ - **Frontier Agentic Capabilities** β€” Industry-leading performance in tool use, search, and complex multi-step tasks
66
+ - **Efficient Reasoning** β€” Trained with RL to reason efficiently and decompose tasks optimally, 37% faster than M2.1
67
+ - **Cost-Effective** β€” $1 for continuous operation at 100 tok/s for an hour; $0.30 at 50 tok/s
68
+ - **Modified-MIT Base License** β€” Based on MiniMax's open-weight release
69
+
70
+ ## Base Model Architecture
71
+
72
+ MiniMax-M2.5 is a Mixture-of-Experts (MoE) model extensively trained with reinforcement learning across hundreds of thousands of complex real-world environments.
73
+
74
+ | Specification | Value |
75
+ |---------------|-------|
76
+ | Architecture | Sparse Mixture-of-Experts (MoE) |
77
+ | Training | Extensive RL in 200K+ real-world environments |
78
+ | Languages | 10+ (Go, C, C++, TypeScript, Rust, Kotlin, Python, Java, JavaScript, PHP, Lua, Dart, Ruby) |
79
+ | Inference Speed | 100 tok/s (Lightning) / 50 tok/s (Standard) |
80
+ | Library | `transformers` |
81
+
82
+ ## Benchmarks (Base Model)
83
+
84
+ ### Coding
85
+
86
+ | Benchmark | MiniMax-M2.5 | Claude Opus 4.6 | Gemini 3 Pro | GPT-5.2 |
87
+ |-----------|-------------|-----------------|-------------|---------|
88
+ | SWE-Bench Verified | **80.2** | 78.9 | 74.0 | 72.6 |
89
+ | Multi-SWE-Bench | **51.3** | 50.8 | β€” | β€” |
90
+ | SWE-Bench Multilingual | **55.6** | β€” | β€” | β€” |
91
+ | Terminal-Bench 2.0 | 51.5 | 52.1 | β€” | β€” |
92
+
93
+ ### Search & Tool Calling
94
+
95
+ | Benchmark | MiniMax-M2.5 | Claude Opus 4.6 | Gemini 3 Pro | GPT-5.2 |
96
+ |-----------|-------------|-----------------|-------------|---------|
97
+ | BrowseComp | **76.3** | 71.2 | 62.4 | 57.8 |
98
+
99
+ ### Reasoning & Knowledge
100
+
101
+ | Benchmark | MiniMax-M2.5 | Claude Opus 4.6 | Gemini 3 Pro | GPT-5.2 |
102
+ |-----------|-------------|-----------------|-------------|---------|
103
+ | AIME25 | 86.3 | 95.6 | 96.0 | 98.0 |
104
+ | GPQA-D | 85.2 | 90.0 | 91.0 | 90.0 |
105
+ | HLE w/o tools | 19.4 | 30.7 | 37.2 | 31.4 |
106
+ | SciCode | 44.4 | 52.0 | 56.0 | 52.0 |
107
+ | IFBench | **70.0** | 53.0 | 70.0 | 75.0 |
108
+
109
+ ## Usage
110
+
111
+ ### llama.cpp (GGUF)
112
+
113
+ Build the latest master of [llama.cpp](https://github.com/ggml-org/llama.cpp) and run:
114
+
115
+ ```bash
116
+ ~/llama.cpp/build/bin/llama-cli \
117
+ -m ../outputs/MiniMax-M2.5-PRISM-PRO-[QUANT].gguf \
118
+ --jinja \
119
+ -ngl 999 \
120
+ --repeat_penalty 1.15 \
121
+ --temp 1.0 \
122
+ --top_p 0.95 \
123
+ --top_k 40
124
+ ```
125
+
126
+
127
+ > Replace `[QUANT]` with your quantization level (e.g. `Q8_0`, etc.).
128
+
129
+ ### Recommended Parameters
130
+
131
+ | Use Case | Temperature | Top-P | Top-K | Repeat Penalty | Max New Tokens |
132
+ |----------|-------------|-------|-------|----------------|----------------|
133
+ | Reasoning / Coding | 1.0 | 0.95 | 40 | 1.15 | 32768 |
134
+ | General Chat | 0.6 | 0.95 | 40 | 1.15 | 4096 |
135
+ | Agentic / Tool Use | 1.0 | 0.95 | 40 | 1.15 | 32768 |
136
+
137
+
138
+
139
+ | Version | Description | Access |
140
+ |---------|-------------|--------|
141
+ | **PRISM-LITE** | Abliterated with PRISM-LITE pipeline β€” removes over-refusal while preserving core capabilities | Free on Hugging Face |
142
+ | **PRISM-PRO** | Full PRISM-PRO ablation β€” Full Production Level Mode suppression of propaganda/refusal mechanisms with maximum capability retention | [Ko-fi](https://ko-fi.com/s/0a23d1b9a5) |
143
+
144
+ ## License
145
+
146
+ This model is released under the [PRISM Research License](LICENSE.md).
147
+
148
+ The base model [MiniMax-M2.5](https://huggingface.co/MiniMaxAI/MiniMax-M2.5) is released under a [Modified-MIT License](https://github.com/MiniMax-AI/MiniMax-M2.5/blob/main/LICENSE).
149
+
150
+ ## Acknowledgments
151
+
152
+ Based on [MiniMax-M2.5](https://huggingface.co/MiniMaxAI/MiniMax-M2.5) by [MiniMax AI](https://www.minimax.io).