Phi-4-RRStock / README.md
FINGU-AI's picture
Update README.md
d2a5483 verified
---
library_name: transformers
license: mit
---
# Phi-4 SLERP Merge Model
## Model Description
This is a merged language model created using the **Spherical Linear Interpolation (SLERP) merge method**, allowing for a smooth blend of features from both parent models across different layers. The merge optimizes reasoning, general knowledge, and task-specific performance by strategically interpolating attention and MLP components.
---
## Merge Details
**Merge Method:**
The model was merged using **SLERP (Spherical Linear Interpolation)** rather than a traditional linear merge, ensuring a well-balanced combination of both source models while maintaining coherent weight transitions.
**Base Model:**
- **bunnycore/Phi-4-RR-Shoup** (used as the primary base)
---
## Models Merged
The following models were included in this merge:
1. **bunnycore/Phi-4-RR-Shoup** (Primary base)
2. **bunnycore/Phi-4-Model-Stock-v4**
---
## Configuration
The following YAML configuration was used to produce this merged model:
```yaml
slices:
- sources:
- model: bunnycore/Phi-4-RR-Shoup
layer_range:
- 0
- 32
- model: bunnycore/Phi-4-Model-Stock-v4
layer_range:
- 0
- 32
merge_method: slerp
base_model: bunnycore/Phi-4-RR-Shoup
parameters:
t:
- filter: self_attn
value:
- 0
- 0.5
- 0.3
- 0.7
- 1
- filter: mlp
value:
- 1
- 0.5
- 0.7
- 0.3
- 0
- value: 0.5
dtype: bfloat16