Text Classification
Safetensors
Galician
bert
hate speech
palomapiot commited on
Commit
e81f4a2
·
verified ·
1 Parent(s): 5e99f9d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +66 -3
README.md CHANGED
@@ -1,3 +1,66 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - irlab-udc/MetaHate-mBERT-GL-es
5
+ language:
6
+ - gl
7
+ metrics:
8
+ - accuracy
9
+ - precision
10
+ - recall
11
+ - f1
12
+ pipeline_tag: text-classification
13
+ tags:
14
+ - hate speech
15
+ ---
16
+ # MetaHate-mBERT-GL-es
17
+
18
+ ## Model Description
19
+
20
+ This is a fine-tuned mBERT model specifically designed to detect hate speech in text in Galician (Spanish variety). The model is based on the `bert-base-multilingual-cased` architecture and has been fine-tuned on a custom dataset for the task of binary text classification, where the labels are `no hate` and `hate`.
21
+
22
+ ## Intended Uses & Limitations
23
+
24
+ ### Intended Uses
25
+ - **Hate Speech Detection**: This model is intended for detecting hate speech in social media comments, forums, and other text data sources.
26
+ - **Content Moderation**: Can be used by platforms to automatically flag potentially harmful content.
27
+
28
+ ### Limitations
29
+ - **Biases**: The model may carry biases present in the training data.
30
+ - **False Positives/Negatives**: It's not perfect and may misclassify some instances.
31
+ - **Domain Specificity**: Performance may vary across different domains.
32
+
33
+ ## Citation
34
+
35
+ If you use this model, please cite the following reference:
36
+
37
+ ```bibtex
38
+ @misc{piot2025bridginggapshatespeech,
39
+ title={Bridging Gaps in Hate Speech Detection: Meta-Collections and Benchmarks for Low-Resource Iberian Languages},
40
+ author={Paloma Piot and José Ramom Pichel Campos and Javier Parapar},
41
+ year={2025},
42
+ eprint={2510.11167},
43
+ archivePrefix={arXiv},
44
+ primaryClass={cs.CL},
45
+ url={https://arxiv.org/abs/2510.11167},
46
+ }
47
+ ```
48
+
49
+ ## Acknowledgements
50
+ The authors thank the funding from the Horizon Europe research and innovation programme under the Marie Skłodowska-Curie Grant Agreement No. 101073351. The authors also thank the financial support supplied by the Consellería de Cultura, Educación, Formación Profesional e Universidades (accreditation 2019-2022 ED431G/01, ED431B 2022/33) and the European Regional Development Fund, which acknowledges the CITIC Research Center in ICT of the University of A Coruña as a Research Center of the Galician University System and the project PID2022-137061OB-C21 (Ministerio de Ciencia e Innovación, Agencia Estatal de Investigación, Proyectos de Generación de Conocimiento; supported by the European Regional Development Fund). The authors also thank the funding of project PLEC2021-007662 (MCIN/AEI/10.13039/501100011033, Ministerio de Ciencia e Innovación, Agencia Estatal de Investigación, Plan de Recuperación, Transformación y Resiliencia, Unión Europea-Next Generation EU).
51
+
52
+ ## Usage
53
+
54
+ ### Inference
55
+
56
+ To use this model, you can load it via the `transformers` library:
57
+
58
+ ```python
59
+ from transformers import pipeline
60
+
61
+ # Load the model
62
+ classifier = pipeline("text-classification", model="irlab-udc/MetaHate-mBERT-GL-es")
63
+
64
+ # Test the model
65
+ result = classifier("Your input text here")
66
+ print(result) # Should print the labels "no hate" or "hate"