|
|
--- |
|
|
license: apache-2.0 |
|
|
datasets: |
|
|
- songff/UltraPrompt |
|
|
language: |
|
|
- en |
|
|
base_model: |
|
|
- meta-llama/Llama-3.2-3B-Instruct |
|
|
library_name: transformers |
|
|
--- |
|
|
|
|
|
# P-Aligner |
|
|
|
|
|
|
|
|
## Quick Start |
|
|
|
|
|
```python |
|
|
from vllm import LLM, SamplingParams |
|
|
from transformers import AutoTokenizer |
|
|
|
|
|
raw_instruction = "What is the capital of France?" |
|
|
model_path = "P-Aligner" |
|
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True) |
|
|
model = LLM( |
|
|
model=model_path, |
|
|
gpu_memory_utilization=0.9, |
|
|
enable_prefix_caching=True, |
|
|
dtype="bfloat16", |
|
|
) |
|
|
|
|
|
outputs = model.generate( |
|
|
[raw_instruction], |
|
|
sampling_params=SamplingParams( |
|
|
temperature=0.0, |
|
|
max_tokens=2048, |
|
|
), |
|
|
) |
|
|
better_instruction = tokenizer.parse_output( |
|
|
outputs[0].outputs[0].text, |
|
|
raw_instruction, |
|
|
) |
|
|
|
|
|
print(better_instruction) |
|
|
``` |
|
|
|
|
|
If you find this work useful, please consider citing: |
|
|
``` |
|
|
@misc{song2025paligner, |
|
|
title={P-Aligner: Enabling Pre-Alignment of Language Models via Principled Instruction Synthesis}, |
|
|
author={Song, Feifan and Gao, Bofei and Song, Yifan and Liu, Yi and Xiong, Weimin and Song, Yuyang and Liu, Tianyu and Wang, Guoyin and Wang, Houfeng}, |
|
|
year={2025}, |
|
|
eprint={2508.04626}, |
|
|
archivePrefix={arXiv}, |
|
|
primaryClass={cs.CL} |
|
|
} |
|
|
``` |