AxionLab-official commited on
Commit
5690616
Β·
verified Β·
1 Parent(s): 182a8e1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +103 -1
README.md CHANGED
@@ -4,4 +4,106 @@ language:
4
  - pt
5
  base_model:
6
  - google/gemma-3-270m
7
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  - pt
5
  base_model:
6
  - google/gemma-3-270m
7
+ ---
8
+
9
+ # 🐢 DogeAI-v1.5-Coder
10
+
11
+ DogeAI-v1.5-Coder is a **small, experimental code-focused language model** fine-tuned from **Gemma 3 (270M parameters)**.
12
+
13
+ This model was created as a learning and experimentation project, focusing on **code generation and completion** with limited resources. It is **not intended to compete with large-scale coding models**, but rather to explore how far a compact model can go when domain-focused.
14
+
15
+ ---
16
+
17
+ ## πŸ” Model Details
18
+
19
+ - **Base model:** Gemma 3 – 270M
20
+ - **Fine-tuning type:** Supervised fine-tuning (SFT)
21
+ - **Primary domain:** Programming / code-related text
22
+ - **Languages:** Mixed (depends on dataset; mainly scripting-style code)
23
+ - **Parameters:** ~270 million
24
+ - **Context length:** Limited (inherits base model constraints)
25
+
26
+ ---
27
+
28
+ ## 🎯 Intended Use
29
+
30
+ DogeAI-v1.5-Coder is best suited for:
31
+
32
+ - Simple code completion
33
+ - Small scripting examples
34
+ - Educational purposes (learning how fine-tuning works)
35
+ - Research on **small language models**
36
+ - Benchmarking and experimentation
37
+
38
+ It performs best when:
39
+ - Prompts are short and explicit
40
+ - The task is narrow and well-defined
41
+ - Expectations are aligned with its size
42
+
43
+ ---
44
+
45
+ ## ⚠️ Limitations
46
+
47
+ This model has **clear and expected limitations**:
48
+
49
+ - Weak long-range reasoning
50
+ - Inconsistent performance on complex programming tasks
51
+ - Limited generalization outside the training distribution
52
+ - Not reliable for production or critical systems
53
+
54
+ These limitations are a direct consequence of its **small scale and experimental nature**.
55
+
56
+ ---
57
+
58
+ ## πŸ§ͺ Training Notes
59
+
60
+ - The model was fine-tuned on a custom dataset focused on code-related text.
61
+ - No reinforcement learning or advanced alignment techniques were used.
62
+ - The goal was experimentation and learning, not optimization for benchmarks.
63
+
64
+ ---
65
+
66
+ ## πŸ“š Why This Model Exists
67
+
68
+ DogeAI-v1.5-Coder exists as a **learning artifact**.
69
+
70
+ It represents:
71
+ - Early experimentation with fine-tuning
72
+ - Exploration of low-parameter models
73
+ - A step in understanding data quality, formatting, and model behavior
74
+
75
+ Small models are valuable tools for understanding how language models actually work.
76
+
77
+ ---
78
+
79
+ ## 🚫 What This Model Is NOT
80
+
81
+ - ❌ A replacement for large coding assistants
82
+ - ❌ A reasoning-focused model
83
+ - ❌ Production-ready
84
+ - ❌ Instruction-following at a high level
85
+
86
+ ---
87
+
88
+ ## πŸ“œ License
89
+
90
+ This model follows the same license as its base model (Gemma).
91
+ Please ensure compliance with the original license when using or redistributing.
92
+
93
+ ---
94
+
95
+ ## πŸ™Œ Acknowledgements
96
+
97
+ - Google Gemma team for the base model
98
+ - The open-source ML community
99
+
100
+ ---
101
+
102
+ ## 🧠 Final Note
103
+
104
+ DogeAI-v1.5-Coder is small, imperfect, and honest.
105
+ Its value lies in experimentation, not performance.
106
+
107
+ Sometimes, understanding the limits teaches more than chasing scale.
108
+
109
+ MADE BY AXIONLAB