|
|
--- |
|
|
base_model: |
|
|
- Delta-Vector/Francois-PE-V2-Huali-12B |
|
|
- DoppelReflEx/MN-12B-Mimicore-GreenSnake |
|
|
- yamatazen/EtherealAurora-12B-v2 |
|
|
library_name: transformers |
|
|
tags: |
|
|
- mergekit |
|
|
- merge |
|
|
license: cc-by-nc-4.0 |
|
|
--- |
|
|
|
|
|
# What is this? |
|
|
|
|
|
Next-gen version of Mimicore. Balance of roleplaying performance, intelligent and model size. I like this model, nearly 80% of my 24B MiniusLight v2.1. |
|
|
|
|
|
## GGUF |
|
|
|
|
|
[Normal](https://huggingface.co/mradermacher/LilithCore-v1-12B-GGUF) - [IMatrix](https://huggingface.co/mradermacher/LilithCore-v1-12B-i1-GGUF) |
|
|
|
|
|
## Template: Although **Mistral Tekken** one is smarter, I recommend to use *ChatML* format for roleplaying. |
|
|
|
|
|
If you don't really care about the model intelligent, ChatML is better in some cases with more creative. Mistral Tekken for smarter model, sometimes give it a try is not too bad. |
|
|
|
|
|
## Configuration |
|
|
|
|
|
```yaml |
|
|
models: |
|
|
- model: Delta-Vector/Francois-PE-V2-Huali-12B |
|
|
parameters: |
|
|
density: 0.9 |
|
|
weight: 1 |
|
|
- model: DoppelReflEx/MN-12B-Mimicore-GreenSnake |
|
|
parameters: |
|
|
density: 0.6 |
|
|
weight: 0.8 |
|
|
- model: yamatazen/EtherealAurora-12B-v2 |
|
|
parameters: |
|
|
density: 0.8 |
|
|
weight: 0.6 |
|
|
merge_method: dare_ties |
|
|
base_model: Delta-Vector/Francois-PE-V2-Huali-12B |
|
|
tokenizer_source: base |
|
|
parameters: |
|
|
rescale: true |
|
|
dtype: bfloat16 |
|
|
``` |