DevicaAI Toxicity Detection — XLM-RoBERTa-Large

Multilingual toxicity classifier fine-tuned on XLM-RoBERTa-Large (559M params) for Indian language content moderation.

Supported Languages

Hindi, English, Bengali, Tamil, Telugu, Marathi, Kannada, Malayalam, Gujarati, Punjabi, Hinglish (code-mixed)

Performance

Metric Score
Accuracy ~93-95%
F1 (weighted) ~93-95%
F1 (toxic) ~94%

Usage

from transformers import pipeline
classifier = pipeline("text-classification", model="tsmaitry/devica-toxicity-xlmr-large")
classifier("तुम बहुत अच्छे हो")   # → non-toxic
classifier("Saala kutta kamina")   # → toxic
classifier("நீ முட்டாள்")          # → toxic
Downloads last month
51
Safetensors
Model size
0.6B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for tsmaitry/devica-toxicity-xlmr-large

Finetuned
(927)
this model

Dataset used to train tsmaitry/devica-toxicity-xlmr-large