Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time
Paper • 2203.05482 • Published • 8
llm.create_chat_completion(
messages = "No input example has been defined for this model task."
)This is a merge of pre-trained language models created using mergekit.
This model was merged using the linear merge method.
The following models were included in the merge:
The following YAML configuration was used to produce this model:
merge_method: linear
models:
- model: new1+jeiku/Theory_of_Mind_128_StableLM
parameters:
weight: 1
- model: new1+jeiku/Everything_v3_128_StableLM
parameters:
weight: 1
- model: new1+jeiku/Gnosis_StableLM
parameters:
weight: 1
dtype: float16
2-bit
4-bit
6-bit
16-bit
# !pip install llama-cpp-python from llama_cpp import Llama llm = Llama.from_pretrained( repo_id="jeiku/General_Purpose_3B_GGUF", filename="", )