|
|
--- |
|
|
language: |
|
|
- en |
|
|
dataset_info: |
|
|
features: |
|
|
- name: text |
|
|
dtype: string |
|
|
- name: label |
|
|
dtype: int64 |
|
|
splits: |
|
|
- name: train |
|
|
num_bytes: 473209.3747297414 |
|
|
num_examples: 4000 |
|
|
- name: validation |
|
|
num_bytes: 16726.983668160137 |
|
|
num_examples: 150 |
|
|
- name: test |
|
|
num_bytes: 11572.123015873016 |
|
|
num_examples: 100 |
|
|
download_size: 340291 |
|
|
dataset_size: 501508.4814137746 |
|
|
configs: |
|
|
- config_name: default |
|
|
data_files: |
|
|
- split: train |
|
|
path: data/train-* |
|
|
- split: validation |
|
|
path: data/validation-* |
|
|
- split: test |
|
|
path: data/test-* |
|
|
--- |
|
|
## Dataset Description |
|
|
|
|
|
This dataset is designed for fine-tuning language models, particularly the [Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct) model, for the task of hate speech detection in social media text (tweets). It focuses on both **implicit** and **explicit** forms of hate speech, aiming to improve the performance of smaller language models in this challenging task. |
|
|
|
|
|
The dataset is a combination of two existing datasets: |
|
|
|
|
|
* **Hate Speech Examples:** Examples of implicit hate speech are sourced from the [SALT-NLP/ImplicitHate](https://huggingface.co/datasets/SALT-NLP/ImplicitHate) dataset. This dataset contains tweets annotated as containing implicit hate speech, categorized into types like grievance, incitement, inferiority, irony, stereotyping, and threatening. |
|
|
* **Non-Hate Speech Examples:** Examples of non-hate speech are sourced from the [TweetEval](https://huggingface.co/datasets/tweet_eval) dataset, specifically the `hate` configuration. This configuration provides tweets labeled as 'non-hate'. |
|
|
|
|
|
By combining these two sources, we create a dataset suitable for binary classification of tweets into "hate speech" and "not hate speech" categories. |
|
|
|
|
|
## Dataset Splits |
|
|
|
|
|
The dataset is divided into the following splits: |
|
|
|
|
|
* **`train`**: Contains 2500 examples for training the model. |
|
|
* **`validation`**: Contains 150 examples for evaluating and tuning the model during training. |
|
|
* **`test`**: Contains 100 examples for final evaluation of the trained model's performance. |
|
|
|
|
|
These splits are designed to be relatively balanced in terms of class distribution (hate vs. not hate) to ensure fair evaluation. |
|
|
|
|
|
## Dataset Fields |
|
|
|
|
|
Each example in the dataset consists of the following fields: |
|
|
|
|
|
* **`text`**: (`string`) The text content of the tweet. |
|
|
* **`label`**: (`int`) The label for the tweet, with the following mapping: |
|
|
* `0`: Not Hate Speech |
|
|
* `1`: Hate Speech |
|
|
|