DistilBERT NLU Intent Classification

Fine-tuned DistilBERT model for intent classification in Natural Language Understanding (NLU) systems.

Model Details

  • Base Model: distilbert-base-uncased
  • Task: Intent Classification (Sequence Classification)
  • Number of Labels: 8
  • Framework: PyTorch + Transformers

Supported Intents

ID Intent Description
0 BILLING_ISSUE Problems with bills or charges
1 CANCEL_SUBSCRIPTION Cancel service requests
2 CHECK_BALANCE Balance inquiry
3 GOODBYE Farewell messages
4 GREETING Hello/welcome messages
5 MODIFY_PROFILE Update account details
6 ROAMING_ACTIVATION Enable roaming
7 ROAMING_DEACTIVATION Disable roaming

Usage

from transformers import pipeline

classifier = pipeline("text-classification", model="sidde/distilbert-nlu-intent-classification")

# Single prediction
result = classifier("I want to check my balance")
print(result)
# [{"label": "CHECK_BALANCE", "score": 0.98}]

Training Details

  • Dataset: 772 examples (custom intent dataset)
  • Train/Eval Split: 80/20 with stratification
  • Epochs: 10
  • Batch Size: 16
  • Learning Rate: 2e-5
  • Hardware: NVIDIA L4 GPU on OpenShift AI

Deployment

This model is deployed on OpenShift AI using KServe.

License

Apache 2.0

Downloads last month
16
Safetensors
Model size
67M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support