PatrickSchwabl commited on
Commit
8179fc0
·
verified ·
1 Parent(s): 3b047b1

Create README.md

Browse files

Model Details

Model Name: distilbert_fearspeech_classifier

Model Version: 1.0

Paper: "You are doomed!" Crisis-specific and Dynamic Use of Fear Speech in Protest and Extremist Radical Social Movements

Authors: Simon Greipl, Julian Hohner, Heidi Schulze, Patrick Schwabl, Diana Rieger

Institution: Ludwig-Maximilians-University of Munich, Germany

Contact: simon.greipl@ifkw.lmu.de

Date: 2024-04-26
Overview

This model is a DistilBERT-based classifier trained to detect fear speech (FS) in German language Telegram posts. The classifier was developed to study the prevalence and dynamics of FS in the communication of radical and extremist actors on Telegram.
Intended Use

Primary Use Case: Detection and analysis of fear speech in social media posts, particularly those from radical and extremist groups.

Primary Users: Researchers in social media analysis, hate speech detection, and radicalization studies.

Languages: German

Domains: Social media, Telegram, radical and extremist communication
Model Description

Model Architecture: DistilBERT

Training Data:

The model was fine-tuned on a dataset of manually annotated Telegram posts from radical and extremist groups.
The dataset consists of posts related to six crisis-specific topics: COVID-19, Conspiracy Narratives, Russian Invasion of Ukraine (RioU), Energy Crisis, Inflation, and Migration.
The training dataset includes 2,460 posts, equally split between fear speech (FS) and non-fear speech (no FS).

Performance:

Validation Set Macro F1 Score: 0.82
Test Set Macro F1 Score: 0.79
Detailed performance metrics for the classifier are provided in the paper.

Evaluation

Metrics:

Precision, Recall, F1-Score for both FS and no FS classes.

Validation and Test Set Performance:

Validation Set Precision: 0.82
Validation Set Recall: 0.82
Validation Set F1-Score: 0.82
Test Set Precision: 0.79
Test Set Recall: 0.79
Test Set F1-Score: 0.79

Ethical Considerations

Bias and Fairness: The model was trained on data from specific Telegram channels and groups known for their extremist content. While efforts were made to ensure balanced representation, inherent biases in the data may affect the model's predictions.
Misuse Potential: The classifier is intended for research purposes. Misuse of the model to label and penalize individuals or groups without context or understanding of the nuances in their communication could lead to unjust outcomes.

Training Data

Source:

Data collected from German language Telegram channels and groups associated with far-right, COVID-19 protest, and conspiracy-focused actors.

Annotation Process:

A nominal six-scale annotation scheme was used to capture and distinguish the varying degrees of FS manifestations.
The scheme was compressed into a dichotomous scale for training, where 0 indicates no FS and 1 indicates FS.

How to Use

The trained classifier can be accessed and used via the Hugging Face model hub:

Model URL: distilbert_fearspeech_classifier

Example Usage:

```
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch

# Load the model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("PatrickSchwabl/distilbert_fearspeech_classifier")
model = AutoModelForSequenceClassification.from_pretrained("PatrickSchwabl/distilbert_fearspeech_classifier")

# Tokenize input text
inputs = tokenizer("Your text here", return_tensors="pt")

# Get model predictions
outputs = model(**inputs)
predictions = torch.nn.functional.softmax(outputs.logits, dim=-1)

# Print predictions
print(predictions)


```

Files changed (1) hide show
  1. README.md +15 -0
README.md ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - de
4
+ metrics:
5
+ - f1
6
+ - precision
7
+ - recall
8
+ pipeline_tag: text-classification
9
+ tags:
10
+ - fearspeech
11
+ - classification
12
+ - social science
13
+ - communication
14
+ - hatespeech
15
+ ---