File size: 3,891 Bytes
cf15f7c
7006620
ac71cd7
2e28c73
 
ac71cd7
 
2e28c73
 
7006620
 
 
 
 
 
ac71cd7
 
 
2e28c73
cf15f7c
7006620
 
cf15f7c
ac71cd7
7006620
 
 
f932031
cf15f7c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ac71cd7
cf15f7c
 
 
 
ac71cd7
cf15f7c
 
 
 
 
ac71cd7
cf15f7c
2e28c73
cf15f7c
 
 
2e28c73
cf15f7c
 
 
 
 
c442de1
cf15f7c
 
7006620
cf15f7c
7006620
cf15f7c
7006620
2e28c73
7006620
 
 
cf15f7c
7006620
cf15f7c
 
 
 
7006620
cf15f7c
 
 
 
 
7006620
cf15f7c
 
 
 
7006620
cf15f7c
d542c29
7006620
 
 
87fab78
7006620
 
cf15f7c
2e28c73
cf15f7c
7006620
cf15f7c
7006620
cf15f7c
7006620
 
 
 
 
87fab78
cf15f7c
87fab78
 
 
 
cf15f7c
 
 
87fab78
cf15f7c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
87fab78
cf15f7c
 
 
 
 
 
 
 
 
 
 
 
87fab78
cf15f7c
 
 
 
 
 
 
 
 
 
7006620
 
 
 
2e28c73
cf15f7c
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
---
base_model: google/gemma-3-1b-it
tags:
- text-generation
- finetune
- transformers
- unsloth
- gemma3
- wall-e
- lightweight
- mobile-friendly
- local-ai
- multilingual
- coding-assistant
- text-summarization
license: apache-2.0
language:
- en
- fa
- de
library_name: transformers
pipeline_tag: text-generation
---

[![Open in HF Space](https://img.shields.io/badge/🤗%20Try%20Live%20Demo-FFD21E?style=for-the-badge&logo=huggingface)](https://huggingface.co/spaces/sinamsv0/WALL-E-DEMO)
[![GitHub](https://img.shields.io/badge/⭐%20GitHub-181717?style=for-the-badge&logo=github)](https://github.com/unknownmsv/WALL-E)
[![License](https://img.shields.io/badge/License-Apache%202.0-blue.svg?style=for-the-badge)](LICENSE)

# 🤖 WALL•E — Lightweight Local AI Assistant (1B)

**WALL•E** is a fine-tuned, lightweight language model based on **Gemma 3 1B**, designed for **local, privacy-preserving AI usage**.  
It focuses on *practical tasks*, *fast responses*, and *real-world utility* rather than model size.

---

## 🎯 Why WALL•E?

Most modern AI models are either:
- Too large to run locally, or
- Too generic for everyday tasks

**WALL•E** is built to fill that gap.

✅ Runs entirely locally  
✅ No API keys or cloud services  
✅ Designed for low-resource environments  
✅ Open-source and transparent  

---

## ✨ Key Capabilities

### 🌐 Multilingual Support
- **English** – primary interaction language
- **فارسی (Persian)** – natural and fluent responses
- **Deutsch (German)** – conversational support

### 🛠 Practical Task Focus
- 📝 Text summarization (articles, notes, reports)
- 💻 Coding help (Python, JavaScript, Bash, shell)
- 🖥 Linux command explanations & troubleshooting
- 📚 Short factual answers and guidance

The model is optimized to handle **short and minimal prompts** naturally (e.g. *"Hi"*, *"Explain ls -la"*), avoiding over-generation.

---

## ⚙️ Technical Overview

| Component        | Details |
|------------------|--------|
| Base Model       | Google Gemma 3 1B |
| Fine-tuning      | Supervised Fine-Tuning (SFT) |
| Framework        | Unsloth |
| Context Length   | 3200 tokens |
| Precision        | BF16 |
| License          | Apache 2.0 |

---

## 🚀 Quick Start (Transformers)

```python
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline

model_id = "sinamsv0/WALL-E"

tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
    model_id,
    device_map="auto"
)

pipe = pipeline(
    "text-generation",
    model=model,
    tokenizer=tokenizer
)

response = pipe(
    "Summarize this text: Artificial intelligence is...",
    max_new_tokens=120
)

print(response[0]["generated_text"])
```



## 🧪 Training Summary


Method: Supervised Fine-Tuning (SFT)

Data: Custom multilingual datasets with safety-focused filtering

Hardware: Single consumer GPU

Goal: Improve instruction-following, multilingual responses, and short-prompt behavior





## 🛡 Safety & Limitations

- ✅ Trained with safety-aware data
- ✅ Avoids harmful or unethical requests
- ⚠️ Limited reasoning depth due to 1B parameter size
- ⚠️ Not intended for complex multi-step reasoning or creative writing



## 🌍 Ideal Use Cases


Local coding assistant

Study and document summarization

Privacy-focused users

Lightweight edge deployments

Research and experimentation with small LLMs





## 🤝 Community & Links


GitHub: https://github.com/unknownmsv/WALL-E

Hugging Face Model: https://huggingface.co/sinamsv0/WALL-E

Hugging Face Space: https://huggingface.co/spaces/sinamsv0/WALL-E-DEMO





## 🔮 Roadmap (Planned)


UI tools for local use

Optional voice interface

Extended language support

Performance benchmarking on edge devices






Small model, focused design.
WALL•E proves that useful AI doesn’t have to be huge.