Ex0bit commited on
Commit
33e5bc6
·
verified ·
1 Parent(s): f0da977

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -25
README.md CHANGED
@@ -13,9 +13,9 @@ language:
13
  pipeline_tag: text-generation
14
  ---
15
 
16
- # MiniMax-M2.1-PRISM
17
 
18
- **An abliterated version of MiniMax-M2.1 using the PRISM methodology**
19
 
20
  [![Support me on Ko-fi](https://ko-fi.com/img/githubbutton_sm.svg)](https://ko-fi.com/ericelbaz)
21
 
@@ -23,7 +23,7 @@ pipeline_tag: text-generation
23
 
24
  ## Model Description
25
 
26
- **MiniMax-M2.1-PRISM** is an abliterated version of MiniMax-M2.1, processed using PRISM (Projected Refusal Isolation via Subspace Modification) to remove refusal behaviors while preserving full model capabilities.
27
 
28
  ### Base Model: MiniMax-M2.1
29
 
@@ -44,24 +44,7 @@ MiniMax-M2.1 is an open-source agentic language model designed for robust perfor
44
 
45
  This model was abliterated using **PRISM v5** - a state-of-the-art abliteration methodology combining multiple principled techniques for effective refusal removal while preserving model capabilities.
46
 
47
- **Formula**: `W' = W - weight * (d ⊗ d) @ W`
48
 
49
- Where:
50
- - `W` = Original weight matrix
51
- - `d` = Refusal direction vector (unit normalized)
52
- - `weight` = Layer-specific abliteration strength
53
- - `W'` = Modified weight matrix
54
-
55
- ### Abliteration Parameters
56
-
57
- | Parameter | Value |
58
- |-----------|-------|
59
- | Base Model | QuixiAI/MiniMax-M2.1-bf16 |
60
- | Total Layers | 62 |
61
- | Target Layers | 16-46 (31 layers) |
62
- | Peak Layer | 31 |
63
- | Max Weight | 3.0 |
64
- | Min Weight | 0.5 |
65
 
66
  ### Weight Distribution
67
 
@@ -92,16 +75,16 @@ The abliteration strength follows a triangular distribution centered on the peak
92
  | Benign Coherence | 100% |
93
  | Response Quality | Full technical accuracy preserved |
94
 
95
- Testing shows that PRISM abliteration maintains full model coherence with no measurable capability degradation.
96
 
97
  ---
98
 
99
- ## Available Formats
100
 
101
  | Format | Size | Description |
102
  |--------|------|-------------|
103
- | Safetensors (BF16) | ~426 GB | Full precision, 92 shards |
104
  | GGUF IQ1_S | ~43 GB | Quantized with importance matrix |
 
105
 
106
  ---
107
 
@@ -157,7 +140,6 @@ This model inherits the [Modified-MIT License](https://github.com/MiniMax-AI/Min
157
  ## Credits
158
 
159
  - **Base Model**: [MiniMax-M2.1](https://huggingface.co/MiniMaxAI/MiniMax-M2.1) by MiniMax AI
160
- - **BF16 Conversion**: [QuixiAI/MiniMax-M2.1-bf16](https://huggingface.co/QuixiAI/MiniMax-M2.1-bf16) by Eric Hartford
161
  - **PRISM Abliteration**: Ex0bit
162
  - **Quantization**: Using [llama.cpp](https://github.com/ggml-org/llama.cpp) with unsloth imatrix
163
 
@@ -165,7 +147,7 @@ This model inherits the [Modified-MIT License](https://github.com/MiniMax-AI/Min
165
 
166
  ## Support
167
 
168
- If you find this work useful, consider supporting development:
169
 
170
  [![Support me on Ko-fi](https://ko-fi.com/img/githubbutton_sm.svg)](https://ko-fi.com/ericelbaz)
171
 
 
13
  pipeline_tag: text-generation
14
  ---
15
 
16
+ # MiniMax-M2.1-PRISM (UNCENSORED)
17
 
18
+ ** MiniMax-M2.1 Uncensored via PRISM Advanced Abliteration**
19
 
20
  [![Support me on Ko-fi](https://ko-fi.com/img/githubbutton_sm.svg)](https://ko-fi.com/ericelbaz)
21
 
 
23
 
24
  ## Model Description
25
 
26
+ **MiniMax-M2.1-PRISM** is the abliterated version of MiniMax-M2.1, processed using PRISM (Projected Refusal Isolation via Subspace Modification) to remove refusal behaviors while preserving and even enhancing full model capabilities.
27
 
28
  ### Base Model: MiniMax-M2.1
29
 
 
44
 
45
  This model was abliterated using **PRISM v5** - a state-of-the-art abliteration methodology combining multiple principled techniques for effective refusal removal while preserving model capabilities.
46
 
 
47
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
48
 
49
  ### Weight Distribution
50
 
 
75
  | Benign Coherence | 100% |
76
  | Response Quality | Full technical accuracy preserved |
77
 
78
+ Testing shows that PRISM abliteration maintains full model coherence with no capability degradation and MMLU increases of 5-8%.
79
 
80
  ---
81
 
82
+ ## Available Formats (contact for full tensors)
83
 
84
  | Format | Size | Description |
85
  |--------|------|-------------|
 
86
  | GGUF IQ1_S | ~43 GB | Quantized with importance matrix |
87
+ | Safetensors (BF16) | ~426 GB | Full precision, 92 shards |
88
 
89
  ---
90
 
 
140
  ## Credits
141
 
142
  - **Base Model**: [MiniMax-M2.1](https://huggingface.co/MiniMaxAI/MiniMax-M2.1) by MiniMax AI
 
143
  - **PRISM Abliteration**: Ex0bit
144
  - **Quantization**: Using [llama.cpp](https://github.com/ggml-org/llama.cpp) with unsloth imatrix
145
 
 
147
 
148
  ## Support
149
 
150
+ If you find this work useful, please consider supporting development so I can continue putting out the best models for the community:
151
 
152
  [![Support me on Ko-fi](https://ko-fi.com/img/githubbutton_sm.svg)](https://ko-fi.com/ericelbaz)
153