Update README.md
Browse files
README.md
CHANGED
|
@@ -3,19 +3,19 @@ license: other
|
|
| 3 |
language:
|
| 4 |
- en
|
| 5 |
---
|
| 6 |
-
The Llama 2 sequel to my original experiment with gradient merges using [the following script](https://github.com/Gryphe/BlockMerge_Gradient). Its three models ([Hermes](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-13b), [Chronos](https://huggingface.co/elinas/chronos-13b-v2) and [Airoboros](https://huggingface.co/jondurbin/airoboros-l2-13b-gpt4-2.0)) are almost evenly divided over the layer structure this time. Airoboros was the "wildcard model" due to its superior ability to understand complex instructions.
|
| 7 |
|
| 8 |
## Model details
|
| 9 |
|
| 10 |
-
As
|
| 11 |
|
| 12 |
-
Illustrated below are the gradients used for this specific L2 recipe
|
| 13 |
|
| 14 |
-
[](MythoLogic-L2.png)
|
| 15 |
|
| 16 |
## Prompt Format
|
| 17 |
|
| 18 |
-
This model primarily uses (and
|
| 19 |
```
|
| 20 |
### Instruction:
|
| 21 |
Your instruction or question here.
|
|
|
|
| 3 |
language:
|
| 4 |
- en
|
| 5 |
---
|
| 6 |
+
The Llama 2 sequel to my [original experiment](https://huggingface.co/Gryphe/MythoLogic-13b) with gradient merges using [the following script](https://github.com/Gryphe/BlockMerge_Gradient). Its three models ([Hermes](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-13b), [Chronos](https://huggingface.co/elinas/chronos-13b-v2) and [Airoboros](https://huggingface.co/jondurbin/airoboros-l2-13b-gpt4-2.0)) are almost evenly divided over the layer structure this time. Airoboros was the "wildcard model" due to its superior ability to understand complex instructions.
|
| 7 |
|
| 8 |
## Model details
|
| 9 |
|
| 10 |
+
As before, the main objective was to create an all-round model with improved roleplaying capabilities. MythoLogic-L2 differs from its predecessor in that it focuses primarily on the understanding of instructions and personalities of complex character cards.
|
| 11 |
|
| 12 |
+
Illustrated below are the gradients used for this specific L2 recipe, with the top of the image representing layer 0 and the bottom layer 40.
|
| 13 |
|
| 14 |
+

|
| 15 |
|
| 16 |
## Prompt Format
|
| 17 |
|
| 18 |
+
This model primarily uses (and was heavily tested with) Alpaca formatting, so for optimal model performance, use:
|
| 19 |
```
|
| 20 |
### Instruction:
|
| 21 |
Your instruction or question here.
|