File size: 2,753 Bytes
179ea10
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
---
license: apache-2.0
base_model: HuggingFaceTB/SmolLM-360M-Instruct
tags:
- smollm
- customer-support
- fine-tuned
- lora
- instruct
language:
- en
---

# SmolLM-360M-CustomerSupport-Instruct

This is a fine-tuned version of [HuggingFaceTB/SmolLM-360M-Instruct](https://huggingface.co/HuggingFaceTB/SmolLM-360M-Instruct) trained on customer support conversations.

## Model Details

- **Base Model**: SmolLM-360M-Instruct (360M parameters)
- **Fine-tuning Method**: LoRA (Low-Rank Adaptation)
- **Dataset**: Bitext Customer Support Dataset
- **Training Samples**: 2,000 customer support conversations
- **Use Case**: Customer service chatbot, support ticket handling

## Training Details

- **Epochs**: 3
- **Learning Rate**: 1e-4
- **Batch Size**: 8
- **Max Sequence Length**: 512
- **Optimizer**: AdamW with cosine learning rate schedule
- **Hardware**: Single GPU (Colab/Kaggle)

## Usage

```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "sweatSmile/SmolLM-360M-CustomerSupport-Instruct"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

# Format your prompt
prompt = "<|im_start|>user\nHow do I reset my password?<|im_end|>\n<|im_start|>assistant\n"

inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=150, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)

print(response)
```

## Example Outputs

**Input**: "My order hasn't arrived yet, what should I do?"

**Output**: "I apologize for the delay with your order. Let me help you track it. Could you please provide your order number? Once I have that, I can check the current status and estimated delivery date."

## Limitations

- Trained on English customer support conversations only
- Best for customer service domains (banking, e-commerce, general support)
- May not perform well on highly technical or domain-specific queries outside training data
- Small model size (360M) - not as capable as larger models but much faster

## Intended Use

- Customer support chatbots
- Automated ticket response systems
- FAQ assistance
- Initial customer inquiry handling

## Training Data

The model was fine-tuned on the [Bitext Customer Support Dataset](https://huggingface.co/datasets/bitext/Bitext-customer-support-llm-chatbot-training-dataset), which contains diverse customer service scenarios.

## Citation

```bibtex
@misc{smollm-customer-support-2025,
  author = {sweatSmile},
  title = {SmolLM-360M Fine-tuned for Customer Support},
  year = {2025},
  publisher = {HuggingFace},
  url = {https://huggingface.co/sweatSmile/SmolLM-360M-CustomerSupport-Instruct}
}
```

## License

Apache 2.0 (same as base model)