limloop commited on
Commit
7a11c68
·
1 Parent(s): fba807b
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ *.gguf filter=lfs diff=lfs merge=lfs -text
MN-12B-Hydra-RP-RU.Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75be70b4ff0f409ac77148a05edf411475070950cccd94e28d9187f655c5153c
3
+ size 6083103008
MN-12B-Hydra-RP-RU.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5170aee1ad06c0b4005b0cc796eca5e2f405d68a8a80631b8cac9cdb70056a3d
3
+ size 7477218912
MN-12B-Hydra-RP-RU.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1882344acc121596bd2593c0c4985919c5a85f572e0b33dfcc09268f397c54d4
3
+ size 8727647072
MN-12B-Hydra-RP-RU.Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4769fc86191245f7960379e91b8fb99458c6e41973f569470cf6d4226ce12f3
3
+ size 10056227008
MN-12B-Hydra-RP-RU.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acc6db200c547e90c2bfe2e877f0c110f058e28785450d08f19c60a5d9ba67d8
3
+ size 13022391296
README.md CHANGED
@@ -1,3 +1,156 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: limloop/MN-12B-Hydra-RP-RU
4
+ language:
5
+ - en
6
+ - ru
7
+ tags:
8
+ - GGUF
9
+ - russian
10
+ - uncensored
11
+ - roleplay
12
+ - mixtral-nemo
13
+ ---
14
+
15
+ # MN-12B-Hydra-RP-RU
16
+ [Original model](https://huggingface.co/limloop/MN-12B-Hydra-RP-RU)
17
+
18
+ <details>
19
+ <summary>🇷🇺 Нажмите, чтобы развернуть описание на русском</summary>
20
+
21
+ ## 🌟 О модели
22
+
23
+ **MN-12B-Hydra-RP-RU** — экспериментальный merge на базе Mistral Nemo 12B, сочетающий:
24
+
25
+ * 🎭 Сильные ролевые способности
26
+ * 📚 Глубокий литературный русский язык
27
+ * 🔓 Снятую цензуру
28
+
29
+ Модель собрана методом TIES-merging, что позволяет объединять веса нескольких моделей с минимальными конфликтами между параметрами.
30
+
31
+ ## 🎯 Особенности
32
+
33
+ * Основной язык — русский
34
+ * Хорошо держит персонажей и контекст
35
+ * Следует инструкциям
36
+ * Сохраняет возможности базового Nemo
37
+ * Не проходила дополнительного обучения после слияния
38
+
39
+ ## ⚠️ Важно
40
+
41
+ Uncensored-характер модели означает, что она может генерировать контент, который некоторые пользователи сочтут неподобающим.
42
+
43
+ </details>
44
+
45
+ High-quality TIES merge based on **Mistral Nemo 12B**, optimized for roleplay, strong Russian language capabilities, and uncensored behavior.
46
+
47
+ ---
48
+
49
+ ## 🌍 Overview
50
+
51
+ **MN-12B-Hydra-RP-RU** is an experimental merge built on top of [Mistral Nemo 12B](https://huggingface.co/mistralai/Mistral-Nemo-Base-2407), combining strengths from multiple fine-tuned models:
52
+
53
+ * 🎭 Advanced roleplay capability from Pathfinder-RP
54
+ * 📚 Deep Russian language fluency inspired by Vikhr + Dostoevsky-style tuning
55
+ * 🔓 Reduced safety filtering via uncensored components
56
+
57
+ The merge was created using **TIES merging**, which allows combining model deltas while minimizing destructive interference between weights.
58
+
59
+ ---
60
+
61
+ ## 🎯 Key Features
62
+
63
+ | Feature | Description |
64
+ | ------------------------- | ------------------------------------------------ |
65
+ | **Languages** | Russian, English |
66
+ | **Censorship** | Uncensored behavior |
67
+ | **Roleplay** | Strong character consistency and narrative depth |
68
+ | **Instruction Following** | Reliable prompt adherence |
69
+ | **Tool Calling** | Retains base Nemo capabilities |
70
+ | **Architecture** | Mistral Nemo 12B |
71
+
72
+ ---
73
+
74
+ ## 🧩 Model Composition
75
+
76
+ The merge combines the following models:
77
+
78
+ | Model | Role in merge | Weight |
79
+ | ------------------------------ | ------------------------- | ------ |
80
+ | **Pathfinder-RP-12B-RU** | Base model, RP backbone | 0.60 |
81
+ | **Vikhr Nemo ORPO Dostoevsky** | Literary Russian depth | 0.25 |
82
+ | **HERETIC Uncensored** | Safety removal | 0.30 |
83
+ | **Mag-Mell R1 Uncensored** | Additional uncensor delta | 0.20 |
84
+
85
+ *Weights shown before normalization (final weights are normalized to sum = 1).*
86
+
87
+ ---
88
+
89
+ ## 💡 Usage Example
90
+
91
+ ```python
92
+ from transformers import AutoTokenizer, AutoModelForCausalLM
93
+ import torch
94
+
95
+ model_name = "limloop/MN-12B-Hydra-RP-RU"
96
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
97
+ model = AutoModelForCausalLM.from_pretrained(
98
+ model_name,
99
+ torch_dtype=torch.bfloat16,
100
+ device_map="auto"
101
+ )
102
+
103
+ prompt = "You are a medieval innkeeper. Greet the traveler!"
104
+ messages = [{"role": "user", "content": prompt}]
105
+ inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to(model.device)
106
+
107
+ outputs = model.generate(inputs, max_new_tokens=512, temperature=0.7)
108
+ response = tokenizer.decode(outputs[0], skip_special_tokens=True)
109
+ print(response)
110
+ ```
111
+
112
+ ---
113
+
114
+ ## ⚙️ Merge Details
115
+
116
+ Built using [mergekit](https://github.com/cg123/mergekit) with the **TIES** method (Trim, Elect Sign, Merge).
117
+
118
+ Core mechanism:
119
+
120
+ 1. Trim low-magnitude deltas via `density`
121
+ 2. Resolve sign conflicts
122
+ 3. Weighted averaging of aligned parameters
123
+
124
+ ### Merge Configuration
125
+
126
+ ```yaml
127
+ models:
128
+ - model: Aleteian/Pathfinder-RP-12B-RU
129
+ weight: 0.6
130
+ - model: IlyaGusev/vikhr_nemo_orpo_dostoevsky_12b_slerp
131
+ weight: 0.25
132
+ density: 0.9
133
+ - model: DavidAU/Mistral-Nemo-2407-12B-Thinking-Claude-Gemini-GPT5.2-Uncensored-HERETIC
134
+ weight: 0.3
135
+ density: 0.9
136
+ - model: Naphula/MN-12B-Mag-Mell-R1-Uncensored
137
+ weight: 0.2
138
+ density: 0.9
139
+
140
+ merge_method: ties
141
+ parameters:
142
+ epsilon: 0.01
143
+ normalize: true
144
+ base_model: Aleteian/Pathfinder-RP-12B-RU
145
+ dtype: bfloat16
146
+ tokenizer:
147
+ source: base
148
+ ```
149
+
150
+ ---
151
+
152
+ ## ⚠️ Known Characteristics
153
+
154
+ * No additional post-merge fine-tuning
155
+ * May switch to English on complex reasoning tasks
156
+ * Uncensored components allow generation of explicit or controversial content