File size: 4,128 Bytes
40c3edb a9fdb63 d644072 a9fdb63 40c3edb a9fdb63 40c3edb 165dad9 40c3edb 165dad9 40c3edb 165dad9 40c3edb 165dad9 40c3edb 165dad9 40c3edb 165dad9 40c3edb 165dad9 40c3edb 165dad9 40c3edb 165dad9 40c3edb 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 a9fdb63 165dad9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 |
---
base_model: google/gemma-3-270m-it
library_name: peft
pipeline_tag: text-generation
tags:
- base_model:adapter:google/gemma-3-270m-it
- lora
- transformers
---
🧠 MODEL CARD — DogeAI-v1.0-instruct
Model Details
Model Description
DogeAI-v1.0-instruct is an early-stage instruction-following language model fine-tuned for conversational use and experimentation. This version is intended as a proof of concept (v1) and focuses on language generation rather than reliable logical reasoning.
Developed by: Arthur(loboGOAT)
Funded by: Independent / Community-driven
Shared by: Arthur(loboGOAT)
Model type: Small Instruction-Tuned Language Model
Language(s): Portuguese (primary), multilingual tendencies inherited from base model
License: Apache 2.0 (or the same license as the base model, if different)
Finetuned from model: Gemma-3-270M-it
Model Sources
Repository: loboGOAT/DogeAI-v1.0-instruct
Paper: Not available
Demo: Not available
Uses
Direct Use
Conversational experiments
Text generation and rewriting
Prompt testing and evaluation
Educational use to study limitations of small LLMs
Downstream Use (Optional)
Further fine-tuning
Research on alignment, reasoning, and instruction-following
Benchmarking small models
Out-of-Scope Use
Tasks requiring reliable logical reasoning
Mathematical proof or formal logic
Decision-making systems
Safety-critical or automated validation tasks
Recommendations
This model should not be relied upon for reasoning-intensive tasks.
Users are encouraged to treat DogeAI-v1.0-instruct as an experimental model and expect occasional logical inconsistencies, multilingual drift, or overgeneration.
Future versions aim to address these limitations through:
cleaner datasets
improved stopping criteria
alternative base models
How to Get Started with the Model
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("loboGOAT/DogeAI-v1.0-instruct")
model = AutoModelForCausalLM.from_pretrained("loboGOAT/DogeAI-v1.0-instruct")
inputs = tokenizer("Olá! Vamos conversar?", return_tensors="pt")
outputs = model.generate(
**inputs,
max_new_tokens=128,
temperature=0.65,
top_p=0.95
)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Training Details
Training Data
The model was fine-tuned on a custom instruction-style dataset, primarily in Portuguese, designed to encourage conversational responses.
The dataset does not focus on formal logic or structured reasoning.
Training Procedure
Preprocessing
Instruction–response formatting
Text normalization
No explicit chain-of-thought supervision
Training Hyperparameters
Training regime: Supervised fine-tuning (SFT)
PEFT: Yes (LoRA-based fine-tuning)
Evaluation
Testing Data
Manual testing and prompt-based evaluation.
Factors
Logical consistency
Instruction-following
Language fluency
Metrics
No automated benchmarks were used for this version.
Results
Strong conversational fluency for model size
Inconsistent logical reasoning
Occasional overgeneration beyond intended response
Summary
Model Examination
DogeAI-v1.0-instruct demonstrates the strengths and limitations of small instruction-tuned language models.
While capable of natural conversation, it lacks robust reasoning abilities, which will be a focus of future iterations.
Environmental Impact
Hardware Type: Consumer GPU / Local Machine
Hours used: Low
Cloud Provider: None
Compute Region: Local
Carbon Emitted: Negligible
Technical Specifications
Model Architecture and Objective
Decoder-only Transformer
Next-token prediction
Instruction-following objective
Compute Infrastructure
Local training environment.
Hardware
Consumer-grade GPU / CPU
Software
Transformers
PEFT 0.18.0
PyTorch
Citation
BibTeX:
@misc{dogeai_v1_2025,
title={DogeAI-v1.0-instruct},
author={Arthur},
year={2025},
note={Early experimental instruction-tuned language model}
}
APA:
Arthur (2025). DogeAI-v1.0-instruct: An experimental instruction-tuned language model.
Model Card Authors
Arthur
Model Card Contact
(your Hugging Face profile or GitHub) |