--- license: apache-2.0 language: - en library_name: transformers --- ## Model Composition - `NurtureAI/neural-chat-7b-v3-16k`: Weight - 30% - `xDAN-AI/xDAN-L1-Chat-RL-v1`: Weight - 30% - `rwitz/go-bruins-v2`: Weight - 30% - `segmed/MedMistral-7B-v0.1`: Weight - 10% ### Code Snippet for Model Merging The following Python code demonstrates how to create this mixed model using the LM-Cocktail approach: ```python from LM_Cocktail import mix_models_by_layers model = mix_models_by_layers( model_names_or_paths=[ "NurtureAI/neural-chat-7b-v3-16k", "xDAN-AI/xDAN-L1-Chat-RL-v1", "rwitz/go-bruins-v2", "segmed/MedMistral-7B-v0.1" ], model_type='decoder', weights=[0.3, 0.3, 0.3, 0.1], output_path='./mixed_llm' ) ``` --- license: apache-2.0 ---