Text Classification
Transformers
Safetensors
xlm-roberta
File size: 3,010 Bytes
45d6b74
 
b94d83f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
94331a8
 
 
 
 
 
 
6297ac2
45d6b74
 
94331a8
 
 
 
 
 
023e4e8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
94331a8
926ad04
 
 
 
 
 
 
 
 
 
 
 
 
 
 
94331a8
dff4c4f
94331a8
cde9d07
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
---
library_name: transformers
language:
- en
- fr
- it
- es
- ru
- uk
- tt
- ar
- hi
- ja
- zh
- he
- am
- de
license: openrail++
datasets:
- textdetox/multilingual_toxicity_dataset
metrics:
- f1
base_model:
- FacebookAI/xlm-roberta-large
pipeline_tag: text-classification
---

## Multilingual Toxicity Classifier for 15 Languages (2025)

This is an instance of [xlm-roberta-large](https://huggingface.co/FacebookAI/xlm-roberta-large) that was fine-tuned on binary toxicity classification task based on our updated (2025) dataset [textdetox/multilingual_toxicity_dataset](https://huggingface.co/datasets/textdetox/multilingual_toxicity_dataset).

Now, the models covers 15 languages from various language families:

| Language  | Code | F1 Score |
|-----------|------|---------|
| English   | en   | 0.9225  |
| Russian   | ru   | 0.9525  |
| Ukrainian | uk   | 0.96    |
| German    | de   | 0.7325  |
| Spanish   | es   | 0.7125  |
| Arabic    | ar   | 0.6625  |
| Amharic   | am   | 0.5575  |
| Hindi     | hi   | 0.9725  |
| Chinese   | zh   | 0.9175  |
| Italian   | it   | 0.5864  |
| French    | fr   | 0.9235  |
| Hinglish  | hin  | 0.61    |
| Hebrew    | he   | 0.8775  |
| Japanese  | ja   | 0.8773  |
| Tatar     | tt   | 0.5744  |

## How to use

```python
import torch
from transformers import AutoTokenizer, AutoModelForSequenceClassification

tokenizer = AutoTokenizer.from_pretrained('textdetox/xlmr-large-toxicity-classifier-v2')
model = AutoModelForSequenceClassification.from_pretrained('textdetox/xlmr-large-toxicity-classifier-v2')

batch = tokenizer.encode("You are amazing!", return_tensors="pt")

output = model(batch)
# idx 0 for neutral, idx 1 for toxic
```

## Citation
The model is prepared for [TextDetox 2025 Shared Task](https://pan.webis.de/clef25/pan25-web/text-detoxification.html) evaluation.

```
@inproceedings{dementieva2025overview,
  title={Overview of the Multilingual Text Detoxification Task at PAN 2025},
  author={Dementieva, Daryna and
      Protasov, Vitaly and
      Babakov, Nikolay and
      Rizwan, Naquee and
      Alimova, Ilseyar and
      Brune, Caroline and
      Konovalov, Vasily and
      Muti, Arianna and
      Liebeskind, Chaya and
      Litvak, Marina and
      Nozza, Debora, and
      Shah Khan, Shehryaar and
      Takeshita, Sotaro and
      Vanetik, Natalia and
      Ayele, Abinew Ali and
      Schneider, Frolian and
      Wang, Xintog and
      Yimam, Seid Muhie and
      Elnagar, Ashraf and
      Mukherjee, Animesh and
      Panchenko, Alexander},
    booktitle={Working Notes of CLEF 2025 -- Conference and Labs of the Evaluation Forum},
    editor={Guglielmo Faggioli and Nicola Ferro and Paolo Rosso and Damiano Spina},
    month =                    sep,
    publisher =                {CEUR-WS.org},
    series =                   {CEUR Workshop Proceedings},
    site =                     {Vienna, Austria},
    url =                      {https://ceur-ws.org/Vol-4038/paper_278.pdf},
    year =                     2025
}
```