|
|
--- |
|
|
license: apache-2.0 |
|
|
datasets: |
|
|
- helehan/topic-overwrite |
|
|
language: |
|
|
- en |
|
|
--- |
|
|
|
|
|
# Model Card for Model ID |
|
|
|
|
|
[GitHub](https://github.com/topic-overwrite/topic-level-overwrite/tree/main) | [Paper](https://arxiv.org/abs/2411.17265) |
|
|
|
|
|
## Model Details |
|
|
|
|
|
The model, trained using the RLHF/RLAIF methods proposed in the [TPO paper](https://arxiv.org/abs/2411.17265) by llava, has enhanced trustworthiness and reduced hallucinations. |
|
|
|
|
|
## Model Description |
|
|
|
|
|
- **Trained from model:** [llava-v1.5-7B](https://huggingface.co/liuhaotian/llava-v1.5-7b) |
|
|
- **Trained on data:** [TPO-Dataset](https://huggingface.co/datasets/helehan/topic-overwrite) |
|
|
|
|
|
## Usage |
|
|
|
|
|
Please look at [GitHub](https://github.com/topic-overwrite/topic-level-overwrite/tree/main) for more details about usage. |
|
|
|
|
|
## Citation |
|
|
|
|
|
```bibtex |
|
|
@article{he2024topic, |
|
|
title={A Topic-level Self-Correctional Approach to Mitigate Hallucinations in MLLMs}, |
|
|
author={He, Lehan and Chen, Zeren and Shi, Zhelun and Yu, Tianyu and Shao, Jing and Sheng, Lu}, |
|
|
journal={arXiv preprint arXiv:2411.17265}, |
|
|
year={2024} |
|
|
} |
|
|
``` |