Language Models are Super Mario: Absorbing Abilities from Homologous Models as a Free Lunch
Paper • 2311.03099 • Published • 33
(This is made by me, I'm slowly figuring out how to quant them)
The result of this merge is in my opinion a more vibrant and less generic sonnet inspired prose, it's able to be gentle and harsh where asked. I've personally been trying to get a more spice while also compensating for the Magnum-v2.5 having the issue on my end that it simply won't stop yapping.
Both Mistral and ChatML should work though I had better results with ChatML: ChatML Example:
"""<|im_start|>user
Hi there!<|im_end|>
<|im_start|>assistant
Nice to meet you!<|im_end|>
<|im_start|>user
Can I ask a question?<|im_end|>
<|im_start|>assistant
"""
This model was merged using the DARE TIES merge method using Sao10K/MN-12B-Lyra-v3 as a base.
The following models were included in the merge:
Special thanks to the SillyTilly and myself for helping me find the energy to finish this.
Base model
mistralai/Mistral-Nemo-Base-2407