gimlet / README.md
antiven0m's picture
Update README.md
04a797c verified
---
license: apache-2.0
language:
- en
library_name: transformers
---
## Model Composition
- `NurtureAI/neural-chat-7b-v3-16k`: Weight - 30%
- `xDAN-AI/xDAN-L1-Chat-RL-v1`: Weight - 30%
- `rwitz/go-bruins-v2`: Weight - 30%
- `segmed/MedMistral-7B-v0.1`: Weight - 10%
### Code Snippet for Model Merging
The following Python code demonstrates how to create this mixed model using the LM-Cocktail approach:
```python
from LM_Cocktail import mix_models_by_layers
model = mix_models_by_layers(
model_names_or_paths=[
"NurtureAI/neural-chat-7b-v3-16k",
"xDAN-AI/xDAN-L1-Chat-RL-v1",
"rwitz/go-bruins-v2",
"segmed/MedMistral-7B-v0.1"
],
model_type='decoder',
weights=[0.3, 0.3, 0.3, 0.1],
output_path='./mixed_llm'
)
```
---
license: apache-2.0
---