|
|
| --- |
| |
| base_model: |
| - v000000/NM-12B-Lyris-dev-3 |
| - ArliAI/Mistral-Nemo-12B-ArliAI-RPMax-v1.1 |
| library_name: transformers |
| tags: |
| - mergekit |
| - merge |
|
|
|
|
| --- |
| |
| [](https://hf.co/QuantFactory) |
|
|
|
|
| # QuantFactory/Insanity-GGUF |
| This is quantized version of [Nohobby/Insanity](https://huggingface.co/Nohobby/Insanity) created using llama.cpp |
|
|
| # Original Model Card |
|
|
| # insanity |
|
|
| This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). |
|
|
| ## Merge Details |
| ### Merge Method |
|
|
| This model was merged using the della_linear merge method using [ArliAI/Mistral-Nemo-12B-ArliAI-RPMax-v1.1](https://huggingface.co/ArliAI/Mistral-Nemo-12B-ArliAI-RPMax-v1.1) as a base. |
| |
| ### Configuration |
| |
| The following YAML configuration was used to produce this model: |
| |
| ```yaml |
| models: |
| - model: natong19/Mistral-Nemo-Instruct-2407-abliterated |
| - model: Fizzarolli/MN-12b-Sunrose |
| parameters: |
| density: 0.5 |
| weight: [0.495, 0.165, 0.165, 0.495, 0.495, 0.165, 0.165, 0.495] |
| - model: nbeerbower/mistral-nemo-gutenberg-12B-v4 |
| parameters: |
| density: [0.35, 0.65, 0.5, 0.65, 0.35] |
| weight: [-0.01891, 0.01554, -0.01325, 0.01791, -0.01458] |
| merge_method: dare_ties |
| base_model: natong19/Mistral-Nemo-Instruct-2407-abliterated |
| parameters: |
| normalize: false |
| int8_mask: true |
| dtype: bfloat16 |
| name: uncen |
| |
| --- |
| models: |
| - model: unsloth/Mistral-Nemo-Instruct-2407 |
| - model: NeverSleep/Lumimaid-v0.2-12B |
| parameters: |
| density: 0.5 |
| weight: [0.139, 0.208, 0.139, 0.208, 0.139] |
| - model: nbeerbower/mistral-nemo-cc-12B |
| parameters: |
| density: [0.65, 0.35, 0.5, 0.35, 0.65] |
| weight: [0.01823, -0.01647, 0.01422, -0.01975, 0.01128] |
| - model: nbeerbower/mistral-nemo-bophades-12B |
| parameters: |
| density: [0.35, 0.65, 0.5, 0.65, 0.35] |
| weight: [-0.01891, 0.01554, -0.01325, 0.01791, -0.01458] |
| merge_method: della |
| base_model: unsloth/Mistral-Nemo-Instruct-2407 |
| parameters: |
| epsilon: 0.04 |
| lambda: 1.05 |
| normalize: false |
| int8_mask: true |
| dtype: bfloat16 |
| name: conv |
| --- |
| models: |
| - model: unsloth/Mistral-Nemo-Base-2407 |
| - model: elinas/Chronos-Gold-12B-1.0 |
| parameters: |
| density: 0.9 |
| gamma: 0.01 |
| weight: [0.139, 0.208, 0.208, 0.139, 0.139] |
| - model: shuttleai/shuttle-2.5-mini |
| parameters: |
| density: 0.9 |
| gamma: 0.01 |
| weight: [0.208, 0.139, 0.139, 0.139, 0.208] |
| - model: Epiculous/Violet_Twilight-v0.2 |
| parameters: |
| density: 0.9 |
| gamma: 0.01 |
| weight: [0.139, 0.139, 0.208, 0.208, 0.139] |
| merge_method: breadcrumbs_ties |
| base_model: unsloth/Mistral-Nemo-Base-2407 |
| parameters: |
| normalize: false |
| int8_mask: true |
| dtype: bfloat16 |
| name: chatml |
| --- |
| models: |
| - model: ArliAI/Mistral-Nemo-12B-ArliAI-RPMax-v1.1 |
| parameters: |
| weight: [0.2, 0.3, 0.2, 0.3, 0.2] |
| density: [0.45, 0.55, 0.45, 0.55, 0.45] |
| - model: chatml |
| parameters: |
| weight: [0.01768, -0.01675, 0.01285, -0.01696, 0.01421] |
| density: [0.6, 0.4, 0.5, 0.4, 0.6] |
| - model: uncen |
| parameters: |
| density: [0.6, 0.4, 0.5, 0.4, 0.6] |
| weight: [0.01768, -0.01675, 0.01285, -0.01696, 0.01421] |
| - model: conv |
| parameters: |
| weight: [0.208, 0.139, 0.139, 0.139, 0.208] |
| density: [0.7] |
| - model: v000000/NM-12B-Lyris-dev-3 |
| parameters: |
| weight: [0.33] |
| density: [0.45, 0.55, 0.45, 0.55, 0.45] |
| merge_method: della_linear |
| base_model: ArliAI/Mistral-Nemo-12B-ArliAI-RPMax-v1.1 |
| parameters: |
| epsilon: 0.04 |
| lambda: 1.05 |
| int8_mask: true |
| rescale: true |
| normalize: false |
| dtype: bfloat16 |
| tokenizer_source: base |
| ``` |
| |
|
|