File size: 3,654 Bytes
16a7680
 
96fff49
 
 
 
 
 
 
16a7680
 
96fff49
16a7680
96fff49
16a7680
96fff49
16a7680
96fff49
16a7680
 
 
 
 
96fff49
16a7680
96fff49
 
 
 
 
 
16a7680
96fff49
16a7680
96fff49
 
 
 
 
 
 
16a7680
96fff49
16a7680
96fff49
 
16a7680
96fff49
16a7680
96fff49
16a7680
 
 
96fff49
16a7680
96fff49
 
 
 
 
16a7680
96fff49
16a7680
96fff49
16a7680
96fff49
 
 
 
16a7680
 
 
96fff49
16a7680
96fff49
 
 
 
16a7680
96fff49
16a7680
96fff49
16a7680
96fff49
 
 
 
16a7680
 
 
96fff49
 
 
16a7680
96fff49
16a7680
 
 
96fff49
 
16a7680
96fff49
16a7680
96fff49
 
16a7680
96fff49
16a7680
96fff49
 
16a7680
96fff49
c7c6ea6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
---
library_name: transformers
tags:
- smollm2
- automotive
- question-answering
- instruction-tuning
- domain-adaptation
- workshop-assistant
---

# Model Card for SmolLM2-135M-Technician-QA

This model is a domain-adapted version of **HuggingFaceTB/SmolLM2-135M**, fine-tuned to answer questions related to automotive service, technician workflows, diagnostics, and spare part replacement scenarios.

It is optimized for lightweight deployment in workshop assistants, service center copilots, and edge devices.

---

## Model Details

### Model Description

SmolLM2-135M-Technician-QA is a compact instruction-following language model fine-tuned on a curated dataset of technician question-answer pairs covering:

- Customer vehicle issues  
- Technical diagnostics  
- Work order lifecycle  
- Periodic service procedures  
- Spare part replacement decisions  
- On-site breakdown support  

The model is designed for real-world automotive service environments where fast and efficient inference is required.

- **Developed by:** Shailesh H  
- **Funded by:** Self / Research & Development  
- **Shared by:** Shailesh H  
- **Model type:** Causal Language Model (Instruction-tuned)  
- **Language(s) (NLP):** English  
- **License:** Apache-2.0  
- **Finetuned from model:** HuggingFaceTB/SmolLM2-135M  

### Model Sources

- **Repository:** https://huggingface.co/<your-username>/SmolLM2-135M-Technician-QA  
- **Base Model:** https://huggingface.co/HuggingFaceTB/SmolLM2-135M  

---

## Uses

### Direct Use

This model can be used for:

- Automotive technician assistants  
- Workshop chatbot systems  
- Service advisor support  
- Troubleshooting guidance  
- Training simulators for technicians  

### Downstream Use

The model can be integrated into:

- RAG systems with service manuals  
- Mobile workshop applications  
- Edge diagnostic tools  
- Voice-based service assistants  

### Out-of-Scope Use

This model should NOT be used for:

- Safety-critical vehicle control  
- Legal or compliance decisions  
- Autonomous driving systems  
- Financial or medical advice  

---

## Bias, Risks, and Limitations

- Trained on synthetic domain data → may not cover all vehicle models
- Limited general world knowledge due to small model size
- May generate plausible but incorrect repair steps
- English-only responses

### Recommendations

- Always verify outputs with OEM service manuals
- Use as an assistive tool, not a final authority
- Combine with RAG for production deployment

---

## How to Get Started with the Model

```python
from transformers import AutoTokenizer, AutoModelForCausalLM

model_id = "<your-username>/SmolLM2-135M-Technician-QA"

tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)

prompt = "Customer says the car battery drains overnight. What should you check?"

inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=120)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
---
## Evaluation
## Testing Data, Factors & Metrics
## Testing Data

Held-out automotive technician QA samples from the same domain.

## Factors

Customer complaint handling

Diagnostic reasoning

Spare part replacement logic

Service workflow understanding

## Metrics

Perplexity

Instruction-following accuracy

Manual domain evaluation

## Results

Strong performance on workshop troubleshooting queries

Accurate step-by-step diagnostic suggestions

Fast inference on CPU

## Summary

The fine-tuned model shows clear domain adaptation compared to the base SmolLM2 model, especially for automotive service workflows.