PatrickSchwabl commited on
Commit
e9b1bfd
·
verified ·
1 Parent(s): 8179fc0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +58 -1
README.md CHANGED
@@ -12,4 +12,61 @@ tags:
12
  - social science
13
  - communication
14
  - hatespeech
15
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
  - social science
13
  - communication
14
  - hatespeech
15
+ ---
16
+ This model card provides details about the DistilBERT-based classifier designed to detect fear speech (FS) in German language Telegram posts. The classifier was developed to study the prevalence and dynamics of FS in the communication of radical and extremist actors on Telegram.
17
+ Model Details
18
+ Model Description
19
+
20
+ The distilbert_fearspeech_classifier is a fine-tuned DistilBERT model aimed at identifying and classifying fear speech in social media posts, particularly those from radical and extremist groups. It has been trained on a dataset of annotated Telegram posts from far-right, COVID-19 protest, and conspiracy-focused actors.
21
+
22
+ Developed by: Simon Greipl, Julian Hohner, Heidi Schulze, Patrick Schwabl, Diana Rieger
23
+ Model type: DistilBERT for text classification
24
+ Language(s) (NLP): German
25
+ License: Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International Public License
26
+
27
+ Model Sources
28
+
29
+ Paper: "You are doomed!" Crisis-specific and Dynamic Use of Fear Speech in Protest and Extremist Radical Social Movements
30
+
31
+
32
+ Uses
33
+
34
+ The model is used directly to classify Telegram posts into fear speech (FS) and non-fear speech (no FS) categories. This is particularly useful for researchers studying online radicalization and the dynamics of fear speech in social media.
35
+
36
+ Downstream Use
37
+
38
+ The model can be fine-tuned for specific tasks related to hate speech detection, communication studies, and social media analysis.
39
+ Out-of-Scope Use
40
+
41
+ This model should not be used for labeling and penalizing individuals or groups without context or understanding of the nuances in their communication. Misuse could lead to unjust outcomes.
42
+ Bias, Risks, and Limitations
43
+
44
+ The model was trained on data from specific Telegram channels and groups known for their extremist content. Inherent biases in the data may affect the model's predictions. Misuse of the model could lead to labeling and penalizing individuals or groups without proper context.
45
+ Recommendations
46
+
47
+ Users should be aware of the risks, biases, and limitations of the model. Further research and contextual understanding are recommended before using the model for critical decision-making.
48
+ How to Get Started with the Model
49
+
50
+ Use the following code to get started with the model:
51
+
52
+ ```
53
+ from transformers import AutoTokenizer, AutoModelForSequenceClassification
54
+ import torch
55
+
56
+ # Load the model and tokenizer
57
+ tokenizer = AutoTokenizer.from_pretrained("PatrickSchwabl/distilbert_fearspeech_classifier")
58
+ model = AutoModelForSequenceClassification.from_pretrained("PatrickSchwabl/distilbert_fearspeech_classifier")
59
+
60
+ # Tokenize input text
61
+ inputs = tokenizer("Your text here", return_tensors="pt")
62
+
63
+ # Get model predictions
64
+ outputs = model(**inputs)
65
+ predictions = torch.nn.functional.softmax(outputs.logits, dim=-1)
66
+
67
+ # Print predictions
68
+ print(predictions)
69
+
70
+ ```
71
+
72
+