sinamsv0 commited on
Commit
2e28c73
·
verified ·
1 Parent(s): b7246d0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -8
README.md CHANGED
@@ -1,21 +1,48 @@
1
  ---
2
  base_model: unsloth/gemma-3-1b-it
3
  tags:
4
- - text-generation-inference
 
5
  - transformers
6
  - unsloth
7
- - gemma3_text
 
8
  license: apache-2.0
9
  language:
10
  - en
 
 
11
  ---
12
 
13
- # Uploaded finetuned model
14
 
15
- - **Developed by:** sinamsv0
16
- - **License:** apache-2.0
17
- - **Finetuned from model :** unsloth/gemma-3-1b-it
18
 
19
- This gemma3_text model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
 
 
 
 
20
 
21
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  base_model: unsloth/gemma-3-1b-it
3
  tags:
4
+ - text-generation
5
+ - finetune
6
  - transformers
7
  - unsloth
8
+ - gemma3
9
+ - wall-e
10
  license: apache-2.0
11
  language:
12
  - en
13
+ - fa
14
+ - de
15
  ---
16
 
17
+ # WALL•E — Finetuned Gemma 3 Model
18
 
19
+ **WALL•E** is a lightweight, multilingual AI model finetuned by **sinamsv0**
20
+ based on **unsloth/gemma-3-1b-it**.
21
+ هدف اصلی این مدل ارائه پاسخ‌های دقیق، امن و سازگار برای مکالمه‌های عمومی است.
22
 
23
+ ### 🔧 Features
24
+ - **Multilingual ability (EN / فارسی / Deutsch)**
25
+ - **Fast inference** thanks to Unsloth optimizations
26
+ - **Safety-aligned** for general-purpose assistants
27
+ - **Lightweight** and suitable for local/edge deployment
28
 
29
+ ### 🧠 Training
30
+ This model was finetuned using:
31
+ - **Unsloth** (for accelerated training)
32
+ - **HuggingFace TRL**
33
+ - **Custom safety & multi-language dataset**
34
+
35
+ ### 📦 Base Model
36
+ - **unsloth/gemma-3-1b-it**
37
+ Apache-2.0 licensed.
38
+
39
+ ### 📌 Usage Example
40
+ ```python
41
+ from transformers import AutoModelForCausalLM, AutoTokenizer
42
+
43
+ model = AutoModelForCausalLM.from_pretrained("sinamsv0/WALL-E")
44
+ tokenizer = AutoTokenizer.from_pretrained("sinamsv0/WALL-E")
45
+
46
+ inputs = tokenizer("Hello WALL•E!", return_tensors="pt")
47
+ outputs = model.generate(**inputs, max_new_tokens=100)
48
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))