File size: 1,155 Bytes
1f66f17
e729204
 
 
 
 
1f66f17
 
 
 
e729204
1f66f17
1a75acf
 
 
 
e729204
1f66f17
e729204
6a3f294
e729204
1f66f17
6a3f294
 
 
 
e729204
1f66f17
e729204
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
---
license: apache-2.0
datasets:
- helehan/topic-overwrite
language:
- en
---

# Model Card for Model ID

[GitHub](https://github.com/topic-overwrite/topic-level-overwrite/tree/main) | [Paper](https://arxiv.org/abs/2411.17265)

## Model Details

The model, trained using the RLHF/RLAIF methods proposed in the [TPO paper](https://arxiv.org/abs/2411.17265) by llava, has enhanced trustworthiness and reduced hallucinations.

## Model Description

- **Trained from model:** [llava-v1.5-7B](https://huggingface.co/liuhaotian/llava-v1.5-7b)
- **Lora Config:** [llava-v1.5-7B-lora](https://huggingface.co/liuhaotian/llava-v1.5-7b-lora)
- **Trained on data:** [TPO-Dataset](https://huggingface.co/datasets/helehan/topic-overwrite)

## Usage

Please look at [GitHub](https://github.com/topic-overwrite/topic-level-overwrite/tree/main) for more details about usage.

## Citation

```bibtex
@article{he2024topic,
  title={A Topic-level Self-Correctional Approach to Mitigate Hallucinations in MLLMs},
  author={He, Lehan and Chen, Zeren and Shi, Zhelun and Yu, Tianyu and Shao, Jing and Sheng, Lu},
  journal={arXiv preprint arXiv:2411.17265},
  year={2024}
}
```