Improve model card metadata and description
Browse filesHi! I'm Niels from the Hugging Face community science team. I noticed this model repository is associated with the paper "Mix Data or Merge Models? Balancing the Helpfulness, Honesty, and Harmlessness of Large Language Model via Model Merging".
This PR updates the model card to include:
- `library_name: transformers` and `pipeline_tag: text-generation` metadata for better discoverability.
- Relevant tags such as `model-merging` and `RESM`.
- A description of the model and its purpose (achieving balanced Helpfulness, Honesty, and Harmlessness alignment via the RESM method).
- A proper BibTeX citation for the associated paper.
Please review and merge if this looks good!
README.md
CHANGED
|
@@ -1,13 +1,39 @@
|
|
| 1 |
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
base_model:
|
| 4 |
- meta-llama/Meta-Llama-3-8B-Instruct
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
## Citation
|
| 7 |
-
|
|
|
|
|
|
|
|
|
|
| 8 |
@article{yang2025mix,
|
| 9 |
title={Mix Data or Merge Models? Balancing the Helpfulness, Honesty, and Harmlessness of Large Language Model via Model Merging},
|
| 10 |
author={Yang, Jinluan and Jin, Dingnan and Tang, Anke and Shen, Li and Zhu, Didi and Chen, Zhengyu and Wang, Daixin and Cui, Qing and Zhang, Zhiqiang and Zhou, Jun and others},
|
| 11 |
journal={arXiv preprint arXiv:2502.06876},
|
| 12 |
year={2025}
|
| 13 |
-
}
|
|
|
|
|
|
| 1 |
---
|
|
|
|
| 2 |
base_model:
|
| 3 |
- meta-llama/Meta-Llama-3-8B-Instruct
|
| 4 |
+
license: apache-2.0
|
| 5 |
+
library_name: transformers
|
| 6 |
+
pipeline_tag: text-generation
|
| 7 |
+
tags:
|
| 8 |
+
- model-merging
|
| 9 |
+
- RESM
|
| 10 |
+
- alignment
|
| 11 |
---
|
| 12 |
+
|
| 13 |
+
# RESM Llama-3-8B-Instruct
|
| 14 |
+
|
| 15 |
+
This model is a 3H-aligned (Helpfulness, Honesty, and Harmlessness) Large Language Model (LLM) developed as part of the research paper [Mix Data or Merge Models? Balancing the Helpfulness, Honesty, and Harmlessness of Large Language Model via Model Merging](https://huggingface.co/papers/2502.06876).
|
| 16 |
+
|
| 17 |
+
## Model Description
|
| 18 |
+
|
| 19 |
+
This model utilizes a novel merging method called **RESM** (**R**eweighting **E**nhanced task **S**ingular **M**erging). RESM addresses challenges in preference alignment such as noise accumulation and layer sparsity adaptation through outlier weighting and sparsity-aware rank selection strategies.
|
| 20 |
+
|
| 21 |
+
The goal of this model is to achieve a balanced alignment across the three pillars of responsible AI:
|
| 22 |
+
* **Helpfulness**
|
| 23 |
+
* **Honesty**
|
| 24 |
+
* **Harmlessness**
|
| 25 |
+
|
| 26 |
+
The paper systematically reveals the collaborative and conflict relationships among these dimensions and demonstrates that parameter-level merging using RESM outperforms standard data mixture methods.
|
| 27 |
+
|
| 28 |
## Citation
|
| 29 |
+
|
| 30 |
+
If you find this work useful, please cite the following paper:
|
| 31 |
+
|
| 32 |
+
```bibtex
|
| 33 |
@article{yang2025mix,
|
| 34 |
title={Mix Data or Merge Models? Balancing the Helpfulness, Honesty, and Harmlessness of Large Language Model via Model Merging},
|
| 35 |
author={Yang, Jinluan and Jin, Dingnan and Tang, Anke and Shen, Li and Zhu, Didi and Chen, Zhengyu and Wang, Daixin and Cui, Qing and Zhang, Zhiqiang and Zhou, Jun and others},
|
| 36 |
journal={arXiv preprint arXiv:2502.06876},
|
| 37 |
year={2025}
|
| 38 |
+
}
|
| 39 |
+
```
|