anaaa2's picture
Upload folder using huggingface_hub
41883b9 verified
---
tags:
- clip
- moral-foundations
- vision
- image-classification
- multimodal
license: mit
---
# Visual Moral Compass
Visual Moral Compass is a fine-tuned CLIP model that classifies images based on Moral Foundations Theory.
## Model Description
This model extends CLIP (openai/clip-vit-base-patch16) with five classifier heads to predict moral dimensions in images:
- **Care vs. Harm**: Concerns about suffering and protection
- **Fairness vs. Cheating**: Concerns about justice and reciprocity
- **Loyalty vs. Betrayal**: Concerns about group membership and solidarity
- **Respect vs. Subversion**: Concerns about hierarchy and authority
- **Sanctity vs. Degradation**: Concerns about purity and contamination
## Usage
```python
from visual_moral_compass import VisualMoralCompass
# Load model
model = VisualMoralCompass.from_pretrained("YOUR_USERNAME/visual-moral-compass")
# Classify an image
results = model.classify_image("path/to/image.jpg")
print(results)
```
## Citation
If you use this model, please cite:
```bibtex
@inproceedings{moralclip2025,
author = {Condez, Ana Carolina and Tavares, Diogo and Magalh\~{a}es, Jo\~{a}o},
title = {MoralCLIP: Contrastive Alignment of Vision-and-Language Representations with Moral Foundations Theory},
year = {2025},
isbn = {9798400720352},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
doi = {10.1145/3746027.3758166},
booktitle = {Proceedings of the 33rd ACM International Conference on Multimedia},
pages = {12399–12408},
numpages = {10},
location = {Dublin, Ireland},
series = {MM '25}
}
```
## Model Details
- **Base Model**: openai/clip-vit-base-patch16
- **Training Data**: Social-Moral Image Database