|
|
--- |
|
|
license: cc-by-4.0 |
|
|
task_categories: |
|
|
- image-text-to-text |
|
|
configs: |
|
|
- config_name: default |
|
|
data_files: |
|
|
- split: HCMAS_train |
|
|
path: version_v4/HCMAS-train.json |
|
|
- split: HCMAS_test |
|
|
path: version_v4/HCMAS-test.json |
|
|
- split: HCSHR_train |
|
|
path: version_v4/HCSHR-train.json |
|
|
- split: HCSHR_test |
|
|
path: version_v4/HCSHR-test.json |
|
|
--- |
|
|
|
|
|
# Aligning VLM Assistants with Personalized Situated Cognition (ACL 2025 main) |
|
|
|
|
|
[](https://github.com/liyongqi2002/PCogAlign) |
|
|
[](https://huggingface.co/datasets/YongqiLi/PCogAlignBench) |
|
|
[](https://arxiv.org/abs/2506.00930) |
|
|
|
|
|
This repository contains the constructed benchmark in our ACL 2025 main paper **"Aligning VLM Assistants with Personalized Situated Cognition"**. |
|
|
|
|
|
> ⚠️ This project is for academic research only and not intended for commercial use. |
|
|
|
|
|
## Abstract |
|
|
|
|
|
Vision-language models (VLMs) aligned with general human objectives, such as being harmless and hallucination-free, have become valuable assistants of humans in managing visual tasks. |
|
|
However, people with diversified backgrounds have different cognition even in the same situation. Consequently, they may have personalized expectations for VLM assistants. |
|
|
This highlights the urgent need to align VLM assistants with personalized situated cognition for real-world assistance. |
|
|
To study this problem, we first simplify it by characterizing individuals based on the sociological concept of Role-Set. Then, we propose to evaluate the individuals' actions to examine whether the personalized alignment is achieved. |
|
|
Further, we construct a benchmark named PCogAlignBench, which includes 18k instances and 20 individuals with different Role-Sets. |
|
|
Finally, we present a framework called PCogAlign, which constructs a cognition-aware and action-based reward model for personalized alignment. |
|
|
Experimental results and human evaluations demonstrate the reliability of the PCogAlignBench and the effectiveness of our proposed PCogAlign. |
|
|
|
|
|
## 🙌 Acknowledgments |
|
|
|
|
|
All datasets and models used are obtained through legal and ethical means. For detailed ethical considerations, please refer to our paper's Ethics Statement section. |
|
|
|
|
|
## 📬 Contact |
|
|
|
|
|
For any questions or feedback, feel free to reach out to us at [liyongqi@whu.edu.cn]. |
|
|
|
|
|
--- |
|
|
|
|
|
✨ Thank you for your interest in PCogAlign! Stay tuned for more updates. |