|
|
---
|
|
|
license: apache-2.0
|
|
|
base_model:
|
|
|
- Qwen/Qwen2.5-7B
|
|
|
pipeline_tag: text-generation
|
|
|
language:
|
|
|
- zho
|
|
|
- eng
|
|
|
- fra
|
|
|
- spa
|
|
|
- por
|
|
|
- deu
|
|
|
- ita
|
|
|
- rus
|
|
|
- jpn
|
|
|
- kor
|
|
|
- vie
|
|
|
- tha
|
|
|
- ara
|
|
|
library_name: transformers
|
|
|
tags:
|
|
|
- text-generation-inference
|
|
|
---
|
|
|
|
|
|
## Model Description
|
|
|
|
|
|
Optimized Layer Merging (OLM)
|
|
|
Is a transformer optimization framework implementing automated layer recombination.
|
|
|
|
|
|
Olm create Frankenstein's monster out of language models by cherry-picking the best performing layers across different models to create a superior hybrid.
|
|
|
The core mechanism:
|
|
|
|
|
|
- Takes multiple language models as input
|
|
|
- Uses a base model as the foundation
|
|
|
- Iteratively replaces individual layers, evaluating performance on specified datasets
|
|
|
- Keeps the best performing layer at each position based on metrics like perplexity, exact match, and a custom "quality" score
|
|
|
- Builds a fusion model layer-by-layer while maintaining or improving performance
|
|
|
|
|
|
https://github.com/jeffmeloy/olm |