File size: 1,604 Bytes
e54703b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
---
language: en
license: apache-2.0
tags:
  - text-classification
  - intent-detection
  - distilbert
  - nlu
datasets:
  - custom
metrics:
  - accuracy
  - f1
pipeline_tag: text-classification
---

# DistilBERT NLU Intent Classification

Fine-tuned DistilBERT model for intent classification in Natural Language Understanding (NLU) systems.

## Model Details

- **Base Model:** distilbert-base-uncased
- **Task:** Intent Classification (Sequence Classification)
- **Number of Labels:** 8
- **Framework:** PyTorch + Transformers

## Supported Intents

| ID | Intent | Description |
|----|--------|-------------|
| 0 | BILLING_ISSUE | Problems with bills or charges |
| 1 | CANCEL_SUBSCRIPTION | Cancel service requests |
| 2 | CHECK_BALANCE | Balance inquiry |
| 3 | GOODBYE | Farewell messages |
| 4 | GREETING | Hello/welcome messages |
| 5 | MODIFY_PROFILE | Update account details |
| 6 | ROAMING_ACTIVATION | Enable roaming |
| 7 | ROAMING_DEACTIVATION | Disable roaming |

## Usage

    from transformers import pipeline

    classifier = pipeline("text-classification", model="sidde/distilbert-nlu-intent-classification")

    # Single prediction
    result = classifier("I want to check my balance")
    print(result)
    # [{"label": "CHECK_BALANCE", "score": 0.98}]

## Training Details

- **Dataset:** 772 examples (custom intent dataset)
- **Train/Eval Split:** 80/20 with stratification
- **Epochs:** 10
- **Batch Size:** 16
- **Learning Rate:** 2e-5
- **Hardware:** NVIDIA L4 GPU on OpenShift AI

## Deployment

This model is deployed on OpenShift AI using KServe.

## License

Apache 2.0