MN-12B-Faun-RP-RU / README.md
limloop's picture
init
bdfc761
---
license: apache-2.0
base_model:
- ReadyArt/Forgotten-Safeword-12B-v4.0
- limloop/MN-12B-Hydra-RP-RU
- SicariusSicariiStuff/Impish_Bloodmoon_12B
library_name: transformers
language:
- en
- ru
tags:
- mergekit
- merge
- russian
- uncensored
- roleplay
- mixtral-nemo
---
# MN-12B-Faun-RP-RU
<details>
<summary>🇷🇺 Нажмите, чтобы развернуть описание на русском</summary>
## 🌟 О модели
**MN-12B-Faun-RP-RU** — улучшенный merge на базе Mistral Nemo 12B, развивающий идеи Hydra и ориентированный на:
* 🎭 Более стабильный и выразительный roleplay
* 📚 Улучшенный русский язык
* 🧠 Расширенный словарный запас, включая сложные и NSFW-темы
* 🔓 Практически отсутствующую цензуру
*Модель собрана методом TIES-merging и не проходила дополнительного обучения после слияния.*
## 🎯 Особенности
* Основной фокус — русский язык
* Лучше удерживает персонажей и стиль диалога
* Более богатая и вариативная генерация
* Улучшенная стабильность на длинных контекстах (проверено до ~8192 токенов)
* Следует инструкциям, но может добавлять дисклеймеры на чувствительные запросы
## ⚠️ Важно
Модель сохраняет uncensored-характер, однако в некоторых случаях может добавлять предупреждения о неподходящем контенте при прямых запросах.
При этом генерация **не блокируется** и продолжается после дисклеймера.
</details>
High-quality TIES merge based on **Mistral Nemo 12B**, focused on improved Russian fluency, stronger roleplay, richer vocabulary, and stable long-context performance.
---
## 🌍 Overview
**MN-12B-Faun-RP-RU** is an evolution of the Hydra-style merge, designed to push further in roleplay quality, language richness, and generation stability.
Key improvements include:
* 📚 Better Russian
* 🎭 More consistent and immersive roleplay behavior
* 🧠 Expanded vocabulary, including expressive and NSFW domains
* 🔁 More stable handling of long conversations (tested up to ~8k tokens)
The model may occasionally produce **safety disclaimers** when prompted directly for sensitive content, but generation continues normally afterward.
*Built using TIES merging, which minimizes destructive interference between merged model weights.*
---
## 🎯 Key Features
| Feature | Description |
| ------------------------- | ----------------------------------------------- |
| **Languages** | Russian, English |
| **Censorship** | Mostly uncensored (with occasional disclaimers) |
| **Roleplay** | Improved consistency and immersion |
| **Instruction Following** | Strong |
| **Vocabulary** | Expanded, including NSFW domains |
| **Context Length** | Stable up to ~8192 tokens |
| **Architecture** | Mistral Nemo 12B |
---
## 🧩 Model Composition
The merge combines the following models:
| Model | Role in merge | Weight |
| ------------------------------- | ------------------- | ------ |
| **MN-12B-Hydra-RP-RU** | Base / foundation | 0.60 |
| **Impish_Bloodmoon_12B** | RP + style boost | 0.25 |
| **Forgotten-Safeword-12B-v4.0** | Uncensored behavior | 0.10 |
*Weights shown before normalization.*
---
## 💡 Usage Example
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
model_name = "limloop/MN-12B-Faun-RP-RU"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype=torch.bfloat16,
device_map="auto"
)
prompt = "You are a mysterious forest faun speaking in poetic Russian."
messages = [{"role": "user", "content": prompt}]
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to(model.device)
outputs = model.generate(inputs, max_new_tokens=512, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
```
---
## ⚙️ Merge Details
Built using [mergekit](https://github.com/arcee-ai/mergekit) with the **TIES** method (Trim, Elect Sign, Merge).
Core mechanism:
1. Trim low-magnitude deltas via `density`
2. Resolve sign conflicts
3. Weighted averaging of aligned parameters
### Merge Configuration
```yaml
models:
- model: limloop/MN-12B-Hydra-RP-RU
parameters:
weight: 0.6
- model: SicariusSicariiStuff/Impish_Bloodmoon_12B
parameters:
weight: 0.25
density: 0.9
- model: ReadyArt/Forgotten-Safeword-12B-v4.0
parameters:
weight: 0.1
density: 0.6
merge_method: ties
parameters:
epsilon: 0.01
normalize: true
base_model: limloop/MN-12B-Hydra-RP-RU
dtype: bfloat16
tokenizer:
source: "base"
```
---
## ⚠️ Known Characteristics
* No post-merge fine-tuning
* May produce safety disclaimers before sensitive outputs
* Occasionally switches to English in complex reasoning
* Stronger stylistic bias in roleplay compared to Hydra