|
|
--- |
|
|
license: apache-2.0 |
|
|
language: |
|
|
- pt |
|
|
base_model: |
|
|
- google/gemma-3-270m |
|
|
--- |
|
|
|
|
|
# πΆ DogeAI-v1.5-Coder |
|
|
|
|
|
DogeAI-v1.5-Coder is a **small, experimental code-focused language model** fine-tuned from **Gemma 3 (270M parameters)**. |
|
|
|
|
|
This model was created as a learning and experimentation project, focusing on **code generation and completion** with limited resources. It is **not intended to compete with large-scale coding models**, but rather to explore how far a compact model can go when domain-focused. |
|
|
|
|
|
--- |
|
|
|
|
|
## π Model Details |
|
|
|
|
|
- **Base model:** Gemma 3 β 270M |
|
|
- **Fine-tuning type:** Supervised fine-tuning (SFT) |
|
|
- **Primary domain:** Programming / code-related text |
|
|
- **Languages:** Mixed (depends on dataset; mainly scripting-style code) |
|
|
- **Parameters:** ~270 million |
|
|
- **Context length:** Limited (inherits base model constraints) |
|
|
|
|
|
--- |
|
|
|
|
|
## π― Intended Use |
|
|
|
|
|
DogeAI-v1.5-Coder is best suited for: |
|
|
|
|
|
- Simple code completion |
|
|
- Small scripting examples |
|
|
- Educational purposes (learning how fine-tuning works) |
|
|
- Research on **small language models** |
|
|
- Benchmarking and experimentation |
|
|
|
|
|
It performs best when: |
|
|
- Prompts are short and explicit |
|
|
- The task is narrow and well-defined |
|
|
- Expectations are aligned with its size |
|
|
|
|
|
--- |
|
|
|
|
|
## β οΈ Limitations |
|
|
|
|
|
This model has **clear and expected limitations**: |
|
|
|
|
|
- Weak long-range reasoning |
|
|
- Inconsistent performance on complex programming tasks |
|
|
- Limited generalization outside the training distribution |
|
|
- Not reliable for production or critical systems |
|
|
|
|
|
These limitations are a direct consequence of its **small scale and experimental nature**. |
|
|
|
|
|
--- |
|
|
|
|
|
## π§ͺ Training Notes |
|
|
|
|
|
- The model was fine-tuned on a custom dataset focused on code-related text. |
|
|
- No reinforcement learning or advanced alignment techniques were used. |
|
|
- The goal was experimentation and learning, not optimization for benchmarks. |
|
|
|
|
|
--- |
|
|
|
|
|
## π Why This Model Exists |
|
|
|
|
|
DogeAI-v1.5-Coder exists as a **learning artifact**. |
|
|
|
|
|
It represents: |
|
|
- Early experimentation with fine-tuning |
|
|
- Exploration of low-parameter models |
|
|
- A step in understanding data quality, formatting, and model behavior |
|
|
|
|
|
Small models are valuable tools for understanding how language models actually work. |
|
|
|
|
|
--- |
|
|
|
|
|
## π« What This Model Is NOT |
|
|
|
|
|
- β A replacement for large coding assistants |
|
|
- β A reasoning-focused model |
|
|
- β Production-ready |
|
|
- β Instruction-following at a high level |
|
|
|
|
|
--- |
|
|
|
|
|
## π License |
|
|
|
|
|
This model follows the same license as its base model (Gemma). |
|
|
Please ensure compliance with the original license when using or redistributing. |
|
|
|
|
|
--- |
|
|
|
|
|
## π Acknowledgements |
|
|
|
|
|
- Google Gemma team for the base model |
|
|
- The open-source ML community |
|
|
|
|
|
--- |
|
|
|
|
|
## π§ Final Note |
|
|
|
|
|
DogeAI-v1.5-Coder is small, imperfect, and honest. |
|
|
Its value lies in experimentation, not performance. |
|
|
|
|
|
Sometimes, understanding the limits teaches more than chasing scale. |
|
|
|
|
|
MADE BY AXIONLAB |