nielsr HF Staff commited on
Commit
f8d235f
·
verified ·
1 Parent(s): 00293f9

Improve model card metadata and description

Browse files

Hi! I'm Niels from the Hugging Face community science team. I noticed this model repository is associated with the paper "Mix Data or Merge Models? Balancing the Helpfulness, Honesty, and Harmlessness of Large Language Model via Model Merging".

This PR updates the model card to include:
- `library_name: transformers` and `pipeline_tag: text-generation` metadata for better discoverability.
- Relevant tags such as `model-merging` and `RESM`.
- A description of the model and its purpose (achieving balanced Helpfulness, Honesty, and Harmlessness alignment via the RESM method).
- A proper BibTeX citation for the associated paper.

Please review and merge if this looks good!

Files changed (1) hide show
  1. README.md +29 -3
README.md CHANGED
@@ -1,13 +1,39 @@
1
  ---
2
- license: apache-2.0
3
  base_model:
4
  - meta-llama/Meta-Llama-3-8B-Instruct
 
 
 
 
 
 
 
5
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  ## Citation
7
- ```
 
 
 
8
  @article{yang2025mix,
9
  title={Mix Data or Merge Models? Balancing the Helpfulness, Honesty, and Harmlessness of Large Language Model via Model Merging},
10
  author={Yang, Jinluan and Jin, Dingnan and Tang, Anke and Shen, Li and Zhu, Didi and Chen, Zhengyu and Wang, Daixin and Cui, Qing and Zhang, Zhiqiang and Zhou, Jun and others},
11
  journal={arXiv preprint arXiv:2502.06876},
12
  year={2025}
13
- }
 
 
1
  ---
 
2
  base_model:
3
  - meta-llama/Meta-Llama-3-8B-Instruct
4
+ license: apache-2.0
5
+ library_name: transformers
6
+ pipeline_tag: text-generation
7
+ tags:
8
+ - model-merging
9
+ - RESM
10
+ - alignment
11
  ---
12
+
13
+ # RESM Llama-3-8B-Instruct
14
+
15
+ This model is a 3H-aligned (Helpfulness, Honesty, and Harmlessness) Large Language Model (LLM) developed as part of the research paper [Mix Data or Merge Models? Balancing the Helpfulness, Honesty, and Harmlessness of Large Language Model via Model Merging](https://huggingface.co/papers/2502.06876).
16
+
17
+ ## Model Description
18
+
19
+ This model utilizes a novel merging method called **RESM** (**R**eweighting **E**nhanced task **S**ingular **M**erging). RESM addresses challenges in preference alignment such as noise accumulation and layer sparsity adaptation through outlier weighting and sparsity-aware rank selection strategies.
20
+
21
+ The goal of this model is to achieve a balanced alignment across the three pillars of responsible AI:
22
+ * **Helpfulness**
23
+ * **Honesty**
24
+ * **Harmlessness**
25
+
26
+ The paper systematically reveals the collaborative and conflict relationships among these dimensions and demonstrates that parameter-level merging using RESM outperforms standard data mixture methods.
27
+
28
  ## Citation
29
+
30
+ If you find this work useful, please cite the following paper:
31
+
32
+ ```bibtex
33
  @article{yang2025mix,
34
  title={Mix Data or Merge Models? Balancing the Helpfulness, Honesty, and Harmlessness of Large Language Model via Model Merging},
35
  author={Yang, Jinluan and Jin, Dingnan and Tang, Anke and Shen, Li and Zhu, Didi and Chen, Zhengyu and Wang, Daixin and Cui, Qing and Zhang, Zhiqiang and Zhou, Jun and others},
36
  journal={arXiv preprint arXiv:2502.06876},
37
  year={2025}
38
+ }
39
+ ```