sweatSmile's picture
Update README.md
179ea10 verified
---
license: apache-2.0
base_model: HuggingFaceTB/SmolLM-360M-Instruct
tags:
- smollm
- customer-support
- fine-tuned
- lora
- instruct
language:
- en
---
# SmolLM-360M-CustomerSupport-Instruct
This is a fine-tuned version of [HuggingFaceTB/SmolLM-360M-Instruct](https://huggingface.co/HuggingFaceTB/SmolLM-360M-Instruct) trained on customer support conversations.
## Model Details
- **Base Model**: SmolLM-360M-Instruct (360M parameters)
- **Fine-tuning Method**: LoRA (Low-Rank Adaptation)
- **Dataset**: Bitext Customer Support Dataset
- **Training Samples**: 2,000 customer support conversations
- **Use Case**: Customer service chatbot, support ticket handling
## Training Details
- **Epochs**: 3
- **Learning Rate**: 1e-4
- **Batch Size**: 8
- **Max Sequence Length**: 512
- **Optimizer**: AdamW with cosine learning rate schedule
- **Hardware**: Single GPU (Colab/Kaggle)
## Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "sweatSmile/SmolLM-360M-CustomerSupport-Instruct"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)
# Format your prompt
prompt = "<|im_start|>user\nHow do I reset my password?<|im_end|>\n<|im_start|>assistant\n"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=150, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
```
## Example Outputs
**Input**: "My order hasn't arrived yet, what should I do?"
**Output**: "I apologize for the delay with your order. Let me help you track it. Could you please provide your order number? Once I have that, I can check the current status and estimated delivery date."
## Limitations
- Trained on English customer support conversations only
- Best for customer service domains (banking, e-commerce, general support)
- May not perform well on highly technical or domain-specific queries outside training data
- Small model size (360M) - not as capable as larger models but much faster
## Intended Use
- Customer support chatbots
- Automated ticket response systems
- FAQ assistance
- Initial customer inquiry handling
## Training Data
The model was fine-tuned on the [Bitext Customer Support Dataset](https://huggingface.co/datasets/bitext/Bitext-customer-support-llm-chatbot-training-dataset), which contains diverse customer service scenarios.
## Citation
```bibtex
@misc{smollm-customer-support-2025,
author = {sweatSmile},
title = {SmolLM-360M Fine-tuned for Customer Support},
year = {2025},
publisher = {HuggingFace},
url = {https://huggingface.co/sweatSmile/SmolLM-360M-CustomerSupport-Instruct}
}
```
## License
Apache 2.0 (same as base model)