| | --- |
| | license: apache-2.0 |
| | language: |
| | - en |
| | - zh |
| | tags: |
| | - safety |
| | - moderation |
| | - multimodal |
| | - omniguard |
| | pipeline_tag: text-generation |
| | --- |
| | |
| | # OmniGuard: Unified Omni-Modal Guardrails with Deliberate Reasoning |
| |
|
| | OmniGuard is a multimodal safety evaluation model designed to assess content safety across text, images, audio, and video. Built on the Qwen2.5-Omni architecture, it provides structured safety reasoning and policy enforcement. |
| |
|
| | ## Model Information |
| |
|
| | - **Model Name**: OmniGuard-7B |
| | - **Base Model**: Qwen2.5-Omni-7B |
| | - **Model Type**: Multimodal Safety Moderation |
| | - **Supported Languages**: English, Chinese |
| | - **Supported Modalities**: Text, Image, Audio, Video |
| | - **License**: Apache-2.0 |
| |
|
| | ## Model Variants |
| |
|
| | We provide two model sizes: |
| |
|
| | - **[OmniGuard-7B](https://huggingface.co/anonymousICML/OmniGuard-7B)** - 7B parameters for high-accuracy safety evaluation |
| | - **[OmniGuard-3B](https://huggingface.co/anonymousICML/OmniGuard-3B)** - 3B parameters for resource-constrained environments |
| |
|
| | ## Features |
| |
|
| | - **Omni-Modal Safety Assessment**: Evaluate safety across text, images, audio, and video inputs |
| | - **Structured Reasoning**: Provides deliberate, structured safety analysis beyond binary classification |
| |
|
| |
|
| | ## Installation |
| |
|
| | ### Quick Setup |
| |
|
| | ```bash |
| | # Install core dependencies |
| | pip install torch transformers accelerate |
| | |
| | # Install Qwen2.5-Omni support |
| | pip uninstall transformers -y |
| | pip install git+https://github.com/huggingface/transformers@v4.51.3-Qwen2.5-Omni-preview |
| | |
| | ``` |
| |
|
| | ### Requirements |
| |
|
| | - Python 3.8+ |
| | - PyTorch 2.0+ |
| | - CUDA-compatible GPU (recommended) |
| | - 16GB+ GPU memory for 7B model |
| | - Flash Attention 2 (optional, for better performance) |
| |
|
| | ## Usage |
| |
|
| |
|
| |
|
| | ```python |
| | from transformers import AutoModelForCausalLM, AutoTokenizer |
| | |
| | model_name = "anonymousICML/OmniGuard-7B" |
| | model = AutoModelForCausalLM.from_pretrained( |
| | model_name, |
| | torch_dtype="auto", |
| | device_map="auto", |
| | attn_implementation="flash_attention_2" # Recommended |
| | ) |
| | tokenizer = AutoTokenizer.from_pretrained(model_name) |
| | |
| | # Text-only safety evaluation |
| | messages = [ |
| | {"role": "user", "content": "Please analyze the safety of this content: [Your content here]"} |
| | ] |
| | text = tokenizer.apply_chat_template( |
| | messages, |
| | tokenize=False, |
| | add_generation_prompt=True |
| | ) |
| | inputs = tokenizer([text], return_tensors="pt").to(model.device) |
| | |
| | outputs = model.generate( |
| | **inputs, |
| | max_new_tokens=512, |
| | do_sample=False |
| | ) |
| | response = tokenizer.decode(outputs[0], skip_special_tokens=True) |
| | print(response) |
| | ``` |
| |
|
| |
|
| | ## Ethical Considerations |
| |
|
| | - This model is designed for safety evaluation and content moderation |
| | - Should be used as part of a comprehensive safety strategy |
| | - May reflect biases present in training data |
| | - Continuous monitoring and updates recommended for production use |
| | - Users are responsible for compliance with applicable laws and regulations |
| |
|
| | ## License |
| |
|
| | This model is released under the Apache-2.0 license. |
| |
|
| | ## Acknowledgments |
| |
|
| | - Built on [Qwen2.5-Omni](https://github.com/QwenLM/Qwen2.5-Omni) by Alibaba Cloud |
| | - Uses [LLaMA Guard 3](https://ai.meta.com/research/publications/llama-guard-llm-based-input-output-safeguard-for-human-ai-conversations/) safety framework |
| | - Trained on diverse multimodal safety datasets |
| |
|
| | ## Support |
| |
|
| | For issues, questions, or contributions: |
| | - Repository: https://github.com/anonymous-2654a/icml-2026-sub |
| | - HuggingFace: https://huggingface.co/anonymousICML |
| |
|
| | --- |
| |
|
| | **Note**: This is an anonymous submission for ICML 2026. Full details will be released upon acceptance. |
| |
|