Ali7880's picture
Upload 5 files
31e7458 verified
---
license: apache-2.0
base_model: Falconsai/nsfw_image_detection
tags:
- image-classification
- content-moderation
- violence-detection
- nsfw-detection
- multi-task-learning
---
# Multi-Head Content Moderator
A multi-task image moderation model with **two classification heads**:
- **NSFW Detection**: Detects explicit/adult content (preserved from Falconsai)
- **Violence Detection**: Detects violent content (newly trained)
## Architecture
- Base: ViT (Vision Transformer) from Falconsai/nsfw_image_detection
- Head 1: NSFW classifier (frozen, pretrained)
- Head 2: Violence classifier (trained on violence dataset)
## Categories
### NSFW Head
- nsfw
- safe
### Violence Head
- safe
- violence
## Performance (Violence Detection)
- Accuracy: 0.9075
- F1 Score: 0.9076
## Usage
```python
import torch
from transformers import AutoImageProcessor
# Load
checkpoint = torch.load('multihead_model.pt')
processor = AutoImageProcessor.from_pretrained('path/to/model')
# Create model class (see notebook for full class definition)
# model = MultiHeadContentModerator(...)
# model.load_state_dict(checkpoint['model_state_dict'])
# Inference
inputs = processor(images=image, return_tensors='pt')
with torch.no_grad():
# Get both predictions
outputs = model(inputs['pixel_values'], task='both')
nsfw_pred = outputs['nsfw'].argmax(-1)
violence_pred = outputs['violence'].argmax(-1)
```