File size: 2,878 Bytes
182a8e1
 
 
 
 
 
5690616
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
---
license: apache-2.0
language:
- pt
base_model:
- google/gemma-3-270m
---

# 🐢 DogeAI-v1.5-Coder

DogeAI-v1.5-Coder is a **small, experimental code-focused language model** fine-tuned from **Gemma 3 (270M parameters)**.

This model was created as a learning and experimentation project, focusing on **code generation and completion** with limited resources. It is **not intended to compete with large-scale coding models**, but rather to explore how far a compact model can go when domain-focused.

---

## πŸ” Model Details

- **Base model:** Gemma 3 – 270M  
- **Fine-tuning type:** Supervised fine-tuning (SFT)  
- **Primary domain:** Programming / code-related text  
- **Languages:** Mixed (depends on dataset; mainly scripting-style code)  
- **Parameters:** ~270 million  
- **Context length:** Limited (inherits base model constraints)

---

## 🎯 Intended Use

DogeAI-v1.5-Coder is best suited for:

- Simple code completion
- Small scripting examples
- Educational purposes (learning how fine-tuning works)
- Research on **small language models**
- Benchmarking and experimentation

It performs best when:
- Prompts are short and explicit
- The task is narrow and well-defined
- Expectations are aligned with its size

---

## ⚠️ Limitations

This model has **clear and expected limitations**:

- Weak long-range reasoning
- Inconsistent performance on complex programming tasks
- Limited generalization outside the training distribution
- Not reliable for production or critical systems

These limitations are a direct consequence of its **small scale and experimental nature**.

---

## πŸ§ͺ Training Notes

- The model was fine-tuned on a custom dataset focused on code-related text.
- No reinforcement learning or advanced alignment techniques were used.
- The goal was experimentation and learning, not optimization for benchmarks.

---

## πŸ“š Why This Model Exists

DogeAI-v1.5-Coder exists as a **learning artifact**.

It represents:
- Early experimentation with fine-tuning
- Exploration of low-parameter models
- A step in understanding data quality, formatting, and model behavior

Small models are valuable tools for understanding how language models actually work.

---

## 🚫 What This Model Is NOT

- ❌ A replacement for large coding assistants
- ❌ A reasoning-focused model
- ❌ Production-ready
- ❌ Instruction-following at a high level

---

## πŸ“œ License

This model follows the same license as its base model (Gemma).  
Please ensure compliance with the original license when using or redistributing.

---

## πŸ™Œ Acknowledgements

- Google Gemma team for the base model
- The open-source ML community

---

## 🧠 Final Note

DogeAI-v1.5-Coder is small, imperfect, and honest.  
Its value lies in experimentation, not performance.

Sometimes, understanding the limits teaches more than chasing scale.

MADE BY AXIONLAB