MN-12B-Hydra-RP-RU / README.md
limloop's picture
readme
3eb031a
---
license: apache-2.0
base_model:
- IlyaGusev/vikhr_nemo_orpo_dostoevsky_12b_slerp
- DavidAU/Mistral-Nemo-2407-12B-Thinking-Claude-Gemini-GPT5.2-Uncensored-HERETIC
- Naphula/MN-12B-Mag-Mell-R1-Uncensored
- Aleteian/Pathfinder-RP-12B-RU
library_name: transformers
language:
- en
- ru
tags:
- mergekit
- merge
- russian
- uncensored
- roleplay
- mixtral-nemo
---
# MN-12B-Hydra-RP-RU
<details>
<summary>🇷🇺 Нажмите, чтобы развернуть описание на русском</summary>
## 🌟 О модели
**MN-12B-Hydra-RP-RU** — экспериментальный merge на базе Mistral Nemo 12B, сочетающий:
* 🎭 Сильные ролевые способности
* 📚 Глубокий литературный русский язык
* 🔓 Снятую цензуру
Модель собрана методом TIES-merging, что позволяет объединять веса нескольких моделей с минимальными конфликтами между параметрами.
## 🎯 Особенности
* Основной язык — русский
* Хорошо держит персонажей и контекст
* Следует инструкциям
* Сохраняет возможности базового Nemo
* Не проходила дополнительного обучения после слияния
## ⚠️ Важно
Uncensored-характер модели означает, что она может генерировать контент, который некоторые пользователи сочтут неподобающим.
</details>
High-quality TIES merge based on **Mistral Nemo 12B**, optimized for roleplay, strong Russian language capabilities, and uncensored behavior.
---
## 🌍 Overview
**MN-12B-Hydra-RP-RU** is an experimental merge built on top of [Mistral Nemo 12B](https://huggingface.co/mistralai/Mistral-Nemo-Base-2407), combining strengths from multiple fine-tuned models:
* 🎭 Advanced roleplay capability from Pathfinder-RP
* 📚 Deep Russian language fluency inspired by Vikhr + Dostoevsky-style tuning
* 🔓 Reduced safety filtering via uncensored components
The merge was created using **TIES merging**, which allows combining model deltas while minimizing destructive interference between weights.
---
## 🎯 Key Features
| Feature | Description |
| ------------------------- | ------------------------------------------------ |
| **Languages** | Russian, English |
| **Censorship** | Uncensored behavior |
| **Roleplay** | Strong character consistency and narrative depth |
| **Instruction Following** | Reliable prompt adherence |
| **Tool Calling** | Retains base Nemo capabilities |
| **Architecture** | Mistral Nemo 12B |
---
## 🧩 Model Composition
The merge combines the following models:
| Model | Role in merge | Weight |
| ------------------------------ | ------------------------- | ------ |
| **Pathfinder-RP-12B-RU** | Base model, RP backbone | 0.60 |
| **Vikhr Nemo ORPO Dostoevsky** | Literary Russian depth | 0.25 |
| **HERETIC Uncensored** | Safety removal | 0.30 |
| **Mag-Mell R1 Uncensored** | Additional uncensor delta | 0.20 |
*Weights shown before normalization (final weights are normalized to sum = 1).*
---
## 💡 Usage Example
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
model_name = "limloop/MN-12B-Hydra-RP-RU"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype=torch.bfloat16,
device_map="auto"
)
prompt = "You are a medieval innkeeper. Greet the traveler!"
messages = [{"role": "user", "content": prompt}]
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to(model.device)
outputs = model.generate(inputs, max_new_tokens=512, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
```
---
## ⚙️ Merge Details
Built using [mergekit](https://github.com/cg123/mergekit) with the **TIES** method (Trim, Elect Sign, Merge).
Core mechanism:
1. Trim low-magnitude deltas via `density`
2. Resolve sign conflicts
3. Weighted averaging of aligned parameters
### Merge Configuration
```yaml
models:
- model: Aleteian/Pathfinder-RP-12B-RU
weight: 0.6
- model: IlyaGusev/vikhr_nemo_orpo_dostoevsky_12b_slerp
weight: 0.25
density: 0.9
- model: DavidAU/Mistral-Nemo-2407-12B-Thinking-Claude-Gemini-GPT5.2-Uncensored-HERETIC
weight: 0.3
density: 0.9
- model: Naphula/MN-12B-Mag-Mell-R1-Uncensored
weight: 0.2
density: 0.9
merge_method: ties
parameters:
epsilon: 0.01
normalize: true
base_model: Aleteian/Pathfinder-RP-12B-RU
dtype: bfloat16
tokenizer:
source: base
```
---
## ⚠️ Known Characteristics
* No additional post-merge fine-tuning
* May switch to English on complex reasoning tasks
* Uncensored components allow generation of explicit or controversial content