Copyright 2026 Harikrishna Srinivasan

DistilBERT for Hate Speech Classifier (LoRA)

Summary

This model is a LoRA fine-tuned DistilBERT base uncased model for binary hate speech classification (Hate / Not Hate). It is optimized for efficient fine-tuning using Low-Rank Adaptation (LoRA) via the Hugging Face PEFT library.


Details

Description

  • Developed by: Harikrishna Srinivasan
  • Model type: Fine-Tuned (LoRA) DistilBERT base uncased finetuned SST-2
  • Task: Binary text classification
  • Language(s): English
  • License: Apache 2.0
  • Finetuned from: distilbert/distilbert-base-uncased-finetuned-sst-2-english

This model uses Low-Rank Adaptation (LoRA) to fine-tune only a small subset of parameters, enabling efficient training while preserving the lightweight and fast inference capabilities of DistilBERT.

Sources


Uses

Direct Use

This model can be used directly for:

  • Hate speech detection in English text
  • Moderation pipelines
  • Dataset auditing
  • Research on implicit hate and biased language
  • Pre-filtering content for human moderation

Dataset Citation

@misc{srinivasan2026hatespeech,
  author       = {Harikrishna Srinivasan},
  title        = {Hate-Speech Dataset},
  year         = {2026},
  publisher    = {Hugging Face Datasets},
  url          = {https://huggingface.co/datasets/Harikrishna-Srinivasan/Hate-Speech}
}

Example:

from transformers import AutoTokenizer, AutoModelForSequenceClassification

MODEL_NAME = "Harikrishna-Srinivasan/Hate-Speech-DistilBERT"

model = PeftModel.from_pretrained(MODEL_NAME)
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, use_fast=True)
Downloads last month
-
Safetensors
Model size
67M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Harikrishna-Srinivasan/Hate-Speech-DistilBERT

Dataset used to train Harikrishna-Srinivasan/Hate-Speech-DistilBERT