File size: 2,508 Bytes
3624161
 
 
 
 
 
 
 
 
 
 
8077c9a
 
 
5263183
 
 
 
 
3624161
 
 
 
 
8077c9a
 
5263183
 
3624161
28e5440
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
---
dataset_info:
  features:
  - name: text
    dtype: string
  - name: label
    dtype: int64
  splits:
  - name: train
    num_bytes: 473209.3747297414
    num_examples: 4000
  - name: validation
    num_bytes: 16726.983668160137
    num_examples: 150
  - name: test
    num_bytes: 11572.123015873016
    num_examples: 100
  download_size: 340291
  dataset_size: 501508.4814137746
configs:
- config_name: default
  data_files:
  - split: train
    path: data/train-*
  - split: validation
    path: data/validation-*
  - split: test
    path: data/test-*
---
## Dataset Description

This dataset is designed for fine-tuning language models, particularly the [Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct) model, for the task of hate speech detection in social media text (tweets). It focuses on both **implicit** and **explicit** forms of hate speech, aiming to improve the performance of smaller language models in this challenging task.

The dataset is a combination of two existing datasets:

*   **Hate Speech Examples:**  Examples of implicit hate speech are sourced from the [SALT-NLP/ImplicitHate](https://huggingface.co/datasets/SALT-NLP/ImplicitHate) dataset. This dataset contains tweets annotated as containing implicit hate speech, categorized into types like grievance, incitement, inferiority, irony, stereotyping, and threatening.
*   **Non-Hate Speech Examples:** Examples of non-hate speech are sourced from the [TweetEval](https://huggingface.co/datasets/tweet_eval) dataset, specifically the `hate` configuration. This configuration provides tweets labeled as 'non-hate'.

By combining these two sources, we create a dataset suitable for binary classification of tweets into "hate speech" and "not hate speech" categories.

## Dataset Splits

The dataset is divided into the following splits:

*   **`train`**:  Contains 2500 examples for training the model.
*   **`validation`**: Contains 150 examples for evaluating and tuning the model during training.
*   **`test`**: Contains 100 examples for final evaluation of the trained model's performance.

These splits are designed to be relatively balanced in terms of class distribution (hate vs. not hate) to ensure fair evaluation.

## Dataset Fields

Each example in the dataset consists of the following fields:

*   **`text`**:  (`string`) The text content of the tweet.
*   **`label`**: (`int`) The label for the tweet, with the following mapping:
    *   `0`:  Not Hate Speech
    *   `1`:  Hate Speech