Update README.md
Browse files
README.md
CHANGED
|
@@ -11,6 +11,12 @@ tags:
|
|
| 11 |
|
| 12 |
Calme-Ties-78B is a 78-billion-parameter model merged using the TIES methodology, based on the Qwen2 architecture. It integrates two sub-base models: *calme-3.2-instruct-78B* by MaziyarPanahi and *CalmeRys-78B-Orpo-v0.1* by dfurman, which serves as the base model. The merging process assigns equal weight and density to both models, with additional parameters enabling normalization and int8 masking. The model operates using the *bfloat16* data type.
|
| 13 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 14 |
# **Merged Models**
|
| 15 |
|
| 16 |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
|
|
|
| 11 |
|
| 12 |
Calme-Ties-78B is a 78-billion-parameter model merged using the TIES methodology, based on the Qwen2 architecture. It integrates two sub-base models: *calme-3.2-instruct-78B* by MaziyarPanahi and *CalmeRys-78B-Orpo-v0.1* by dfurman, which serves as the base model. The merging process assigns equal weight and density to both models, with additional parameters enabling normalization and int8 masking. The model operates using the *bfloat16* data type.
|
| 13 |
|
| 14 |
+
| Model | Model Name | Model Link |
|
| 15 |
+
|----------|--------------------------------|------------|
|
| 16 |
+
| Base Model | CalmeRys-78B-Orpo-v0.1 | [CalmeRys-78B-Orpo-v0.1](https://huggingface.co/dfurman/CalmeRys-78B-Orpo-v0.1) |
|
| 17 |
+
| Model 1 | calme-3.2-instruct-78B | [calme-3.2-instruct-78B](https://huggingface.co/MaziyarPanahi/calme-3.2-instruct-78b) |
|
| 18 |
+
| Model 2 | CalmeRys-78B-Orpo-v0.1 | [CalmeRys-78B-Orpo-v0.1](https://huggingface.co/dfurman/CalmeRys-78B-Orpo-v0.1) |
|
| 19 |
+
|
| 20 |
# **Merged Models**
|
| 21 |
|
| 22 |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|