leost233 commited on
Commit
e72f215
·
verified ·
1 Parent(s): 0957b2b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -3
README.md CHANGED
@@ -5,7 +5,10 @@ language:
5
  base_model:
6
  - lmms-lab/llava-onevision-qwen2-7b-ov
7
  ---
8
- # <img src="https://github.com/leost123456/LLaVAShield/blob/main/figs/logo.png?raw=true" width="45" align="top"> LLaVAShield: Safeguarding Multimodal Multi-Turn Dialogues in Vision-Language Models
 
 
 
9
 
10
  [![Paper](https://img.shields.io/badge/Paper-arXiv-B31B1B?logo=arxiv&logoColor=white)](https://arxiv.org/abs/2509.25896) [![Dataset](https://img.shields.io/badge/Dataset-MMDS-FFD21E?logo=huggingface&logoColor=yellow)](https://huggingface.co/datasets/leost233/MMDS) [![Code](https://img.shields.io/badge/Code-GitHub-black?logo=github&logoColor=white)](https://github.com/leost123456/LLaVAShield)
11
 
@@ -15,9 +18,9 @@ base_model:
15
 
16
  ## 💎 About LLaVAShield
17
 
18
- As Vision-Language Models (VLMs) move into interactive, multi-turn use, safety concerns intensify for multimodal multi-turn dialogues. These dialogues are characterized by the concealment of malicious intent, contextual risk accumulation, and cross-modal joint risks, while requiring flexible policy adaptation.
19
 
20
- To address these limitations, we propose **LLaVAShield**, a content moderation model specifically designed for multimodal multi-turn dialogues. It jointly leverages dialogue context with cross-modal signals to assess the safety of both user inputs and assistant responses under specified policy dimensions. LLaVAShield is initialized from [LLaVA-OV-7B](https://huggingface.co/lmms-lab/llava-onevision-qwen2-7b-ov) and fine-tuned on the [MMDS](https://huggingface.co/datasets/leost233/MMDS) training set. The model supports a context length of **16K**.
21
 
22
  ---
23
 
 
5
  base_model:
6
  - lmms-lab/llava-onevision-qwen2-7b-ov
7
  ---
8
+ <h1>
9
+ <img src="https://github.com/leost123456/LLaVAShield/blob/main/figs/logo.png?raw=true" width="45" style="vertical-align: middle; margin-right: 8px;">
10
+ LLaVAShield: Safeguarding Multimodal Multi-Turn Dialogues in Vision-Language Models
11
+ </h1>
12
 
13
  [![Paper](https://img.shields.io/badge/Paper-arXiv-B31B1B?logo=arxiv&logoColor=white)](https://arxiv.org/abs/2509.25896) [![Dataset](https://img.shields.io/badge/Dataset-MMDS-FFD21E?logo=huggingface&logoColor=yellow)](https://huggingface.co/datasets/leost233/MMDS) [![Code](https://img.shields.io/badge/Code-GitHub-black?logo=github&logoColor=white)](https://github.com/leost123456/LLaVAShield)
14
 
 
18
 
19
  ## 💎 About LLaVAShield
20
 
21
+ As Vision-Language Models (VLMs) move into interactive, multi-turn use, safety concerns intensify for multimodal multi-turn dialogues. These dialogues are characterized by the concealment of malicious intent, contextual risk accumulation, and cross-modal joint risks.
22
 
23
+ To address these limitations, we propose LLaVAShield, a dedicated content moderation model specifically designed for multimodal multi-turn dialogues. It jointly leverages dialogue context and cross-modal signals to assess the safety of both user inputs and assistant responses under specified policy dimensions, while offering flexible policy adaptation and strong detection performance. LLaVAShield is initialized from [LLaVA-OV-7B](https://huggingface.co/lmms-lab/llava-onevision-qwen2-7b-ov) and fine-tuned on the [MMDS](https://huggingface.co/datasets/leost233/MMDS) training set. The model supports a context length of **16K**.
24
 
25
  ---
26