File size: 2,878 Bytes
182a8e1 5690616 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 |
---
license: apache-2.0
language:
- pt
base_model:
- google/gemma-3-270m
---
# πΆ DogeAI-v1.5-Coder
DogeAI-v1.5-Coder is a **small, experimental code-focused language model** fine-tuned from **Gemma 3 (270M parameters)**.
This model was created as a learning and experimentation project, focusing on **code generation and completion** with limited resources. It is **not intended to compete with large-scale coding models**, but rather to explore how far a compact model can go when domain-focused.
---
## π Model Details
- **Base model:** Gemma 3 β 270M
- **Fine-tuning type:** Supervised fine-tuning (SFT)
- **Primary domain:** Programming / code-related text
- **Languages:** Mixed (depends on dataset; mainly scripting-style code)
- **Parameters:** ~270 million
- **Context length:** Limited (inherits base model constraints)
---
## π― Intended Use
DogeAI-v1.5-Coder is best suited for:
- Simple code completion
- Small scripting examples
- Educational purposes (learning how fine-tuning works)
- Research on **small language models**
- Benchmarking and experimentation
It performs best when:
- Prompts are short and explicit
- The task is narrow and well-defined
- Expectations are aligned with its size
---
## β οΈ Limitations
This model has **clear and expected limitations**:
- Weak long-range reasoning
- Inconsistent performance on complex programming tasks
- Limited generalization outside the training distribution
- Not reliable for production or critical systems
These limitations are a direct consequence of its **small scale and experimental nature**.
---
## π§ͺ Training Notes
- The model was fine-tuned on a custom dataset focused on code-related text.
- No reinforcement learning or advanced alignment techniques were used.
- The goal was experimentation and learning, not optimization for benchmarks.
---
## π Why This Model Exists
DogeAI-v1.5-Coder exists as a **learning artifact**.
It represents:
- Early experimentation with fine-tuning
- Exploration of low-parameter models
- A step in understanding data quality, formatting, and model behavior
Small models are valuable tools for understanding how language models actually work.
---
## π« What This Model Is NOT
- β A replacement for large coding assistants
- β A reasoning-focused model
- β Production-ready
- β Instruction-following at a high level
---
## π License
This model follows the same license as its base model (Gemma).
Please ensure compliance with the original license when using or redistributing.
---
## π Acknowledgements
- Google Gemma team for the base model
- The open-source ML community
---
## π§ Final Note
DogeAI-v1.5-Coder is small, imperfect, and honest.
Its value lies in experimentation, not performance.
Sometimes, understanding the limits teaches more than chasing scale.
MADE BY AXIONLAB |