File size: 5,343 Bytes
a516664 3eb031a a516664 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 | ---
license: apache-2.0
base_model:
- IlyaGusev/vikhr_nemo_orpo_dostoevsky_12b_slerp
- DavidAU/Mistral-Nemo-2407-12B-Thinking-Claude-Gemini-GPT5.2-Uncensored-HERETIC
- Naphula/MN-12B-Mag-Mell-R1-Uncensored
- Aleteian/Pathfinder-RP-12B-RU
library_name: transformers
language:
- en
- ru
tags:
- mergekit
- merge
- russian
- uncensored
- roleplay
- mixtral-nemo
---
# MN-12B-Hydra-RP-RU
<details>
<summary>🇷🇺 Нажмите, чтобы развернуть описание на русском</summary>
## 🌟 О модели
**MN-12B-Hydra-RP-RU** — экспериментальный merge на базе Mistral Nemo 12B, сочетающий:
* 🎭 Сильные ролевые способности
* 📚 Глубокий литературный русский язык
* 🔓 Снятую цензуру
Модель собрана методом TIES-merging, что позволяет объединять веса нескольких моделей с минимальными конфликтами между параметрами.
## 🎯 Особенности
* Основной язык — русский
* Хорошо держит персонажей и контекст
* Следует инструкциям
* Сохраняет возможности базового Nemo
* Не проходила дополнительного обучения после слияния
## ⚠️ Важно
Uncensored-характер модели означает, что она может генерировать контент, который некоторые пользователи сочтут неподобающим.
</details>
High-quality TIES merge based on **Mistral Nemo 12B**, optimized for roleplay, strong Russian language capabilities, and uncensored behavior.
---
## 🌍 Overview
**MN-12B-Hydra-RP-RU** is an experimental merge built on top of [Mistral Nemo 12B](https://huggingface.co/mistralai/Mistral-Nemo-Base-2407), combining strengths from multiple fine-tuned models:
* 🎭 Advanced roleplay capability from Pathfinder-RP
* 📚 Deep Russian language fluency inspired by Vikhr + Dostoevsky-style tuning
* 🔓 Reduced safety filtering via uncensored components
The merge was created using **TIES merging**, which allows combining model deltas while minimizing destructive interference between weights.
---
## 🎯 Key Features
| Feature | Description |
| ------------------------- | ------------------------------------------------ |
| **Languages** | Russian, English |
| **Censorship** | Uncensored behavior |
| **Roleplay** | Strong character consistency and narrative depth |
| **Instruction Following** | Reliable prompt adherence |
| **Tool Calling** | Retains base Nemo capabilities |
| **Architecture** | Mistral Nemo 12B |
---
## 🧩 Model Composition
The merge combines the following models:
| Model | Role in merge | Weight |
| ------------------------------ | ------------------------- | ------ |
| **Pathfinder-RP-12B-RU** | Base model, RP backbone | 0.60 |
| **Vikhr Nemo ORPO Dostoevsky** | Literary Russian depth | 0.25 |
| **HERETIC Uncensored** | Safety removal | 0.30 |
| **Mag-Mell R1 Uncensored** | Additional uncensor delta | 0.20 |
*Weights shown before normalization (final weights are normalized to sum = 1).*
---
## 💡 Usage Example
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
model_name = "limloop/MN-12B-Hydra-RP-RU"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype=torch.bfloat16,
device_map="auto"
)
prompt = "You are a medieval innkeeper. Greet the traveler!"
messages = [{"role": "user", "content": prompt}]
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to(model.device)
outputs = model.generate(inputs, max_new_tokens=512, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
```
---
## ⚙️ Merge Details
Built using [mergekit](https://github.com/cg123/mergekit) with the **TIES** method (Trim, Elect Sign, Merge).
Core mechanism:
1. Trim low-magnitude deltas via `density`
2. Resolve sign conflicts
3. Weighted averaging of aligned parameters
### Merge Configuration
```yaml
models:
- model: Aleteian/Pathfinder-RP-12B-RU
weight: 0.6
- model: IlyaGusev/vikhr_nemo_orpo_dostoevsky_12b_slerp
weight: 0.25
density: 0.9
- model: DavidAU/Mistral-Nemo-2407-12B-Thinking-Claude-Gemini-GPT5.2-Uncensored-HERETIC
weight: 0.3
density: 0.9
- model: Naphula/MN-12B-Mag-Mell-R1-Uncensored
weight: 0.2
density: 0.9
merge_method: ties
parameters:
epsilon: 0.01
normalize: true
base_model: Aleteian/Pathfinder-RP-12B-RU
dtype: bfloat16
tokenizer:
source: base
```
---
## ⚠️ Known Characteristics
* No additional post-merge fine-tuning
* May switch to English on complex reasoning tasks
* Uncensored components allow generation of explicit or controversial content
|