Copyright 2026 Harikrishna Srinivasan
DistilBERT for Hate Speech Classifier (LoRA)
Summary
This model is a LoRA fine-tuned DistilBERT base uncased model for binary hate speech classification (Hate / Not Hate). It is optimized for efficient fine-tuning using Low-Rank Adaptation (LoRA) via the Hugging Face PEFT library.
Details
Description
- Developed by: Harikrishna Srinivasan
- Model type: Fine-Tuned (LoRA) DistilBERT base uncased finetuned SST-2
- Task: Binary text classification
- Language(s): English
- License: Apache 2.0
- Finetuned from:
distilbert/distilbert-base-uncased-finetuned-sst-2-english
This model uses Low-Rank Adaptation (LoRA) to fine-tune only a small subset of parameters, enabling efficient training while preserving the lightweight and fast inference capabilities of DistilBERT.
Sources
- Base Model: https://huggingface.co/distilbert/distilbert-base-uncased-finetuned-sst-2-english
- Training Framework: Hugging Face Transformers + PEFT
- Repository: Not publicly linked (local / academic project)
Uses
Direct Use
This model can be used directly for:
- Hate speech detection in English text
- Moderation pipelines
- Dataset auditing
- Research on implicit hate and biased language
- Pre-filtering content for human moderation
Dataset Citation
@misc{srinivasan2026hatespeech,
author = {Harikrishna Srinivasan},
title = {Hate-Speech Dataset},
year = {2026},
publisher = {Hugging Face Datasets},
url = {https://huggingface.co/datasets/Harikrishna-Srinivasan/Hate-Speech}
}
Example:
from transformers import AutoTokenizer, AutoModelForSequenceClassification
MODEL_NAME = "Harikrishna-Srinivasan/Hate-Speech-DistilBERT"
model = PeftModel.from_pretrained(MODEL_NAME)
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, use_fast=True)
- Downloads last month
- -