File size: 4,128 Bytes
40c3edb
a9fdb63
d644072
a9fdb63
40c3edb
a9fdb63
 
 
40c3edb
 
165dad9
 
40c3edb
165dad9
 
40c3edb
165dad9
 
 
 
 
 
 
40c3edb
165dad9
40c3edb
165dad9
 
 
40c3edb
165dad9
 
40c3edb
165dad9
40c3edb
165dad9
40c3edb
165dad9
40c3edb
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
 
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
 
a9fdb63
165dad9
 
a9fdb63
165dad9
 
 
 
 
 
 
a9fdb63
165dad9
a9fdb63
165dad9
 
a9fdb63
165dad9
 
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
 
a9fdb63
165dad9
 
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
 
a9fdb63
165dad9
 
a9fdb63
165dad9
a9fdb63
165dad9
 
 
 
 
a9fdb63
165dad9
 
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
 
 
 
 
 
a9fdb63
 
165dad9
 
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
a9fdb63
165dad9
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
---
base_model: google/gemma-3-270m-it
library_name: peft
pipeline_tag: text-generation
tags:
- base_model:adapter:google/gemma-3-270m-it
- lora
- transformers
---

🧠 MODEL CARD — DogeAI-v1.0-instruct
Model Details

Model Description
DogeAI-v1.0-instruct is an early-stage instruction-following language model fine-tuned for conversational use and experimentation. This version is intended as a proof of concept (v1) and focuses on language generation rather than reliable logical reasoning.

Developed by: Arthur(loboGOAT)
Funded by: Independent / Community-driven
Shared by: Arthur(loboGOAT)
Model type: Small Instruction-Tuned Language Model
Language(s): Portuguese (primary), multilingual tendencies inherited from base model
License: Apache 2.0 (or the same license as the base model, if different)
Finetuned from model: Gemma-3-270M-it

Model Sources

Repository: loboGOAT/DogeAI-v1.0-instruct
Paper: Not available
Demo: Not available

Uses
Direct Use

Conversational experiments

Text generation and rewriting

Prompt testing and evaluation

Educational use to study limitations of small LLMs

Downstream Use (Optional)

Further fine-tuning

Research on alignment, reasoning, and instruction-following

Benchmarking small models

Out-of-Scope Use

Tasks requiring reliable logical reasoning

Mathematical proof or formal logic

Decision-making systems

Safety-critical or automated validation tasks

Recommendations

This model should not be relied upon for reasoning-intensive tasks.
Users are encouraged to treat DogeAI-v1.0-instruct as an experimental model and expect occasional logical inconsistencies, multilingual drift, or overgeneration.

Future versions aim to address these limitations through:

cleaner datasets

improved stopping criteria

alternative base models

How to Get Started with the Model
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("loboGOAT/DogeAI-v1.0-instruct")
model = AutoModelForCausalLM.from_pretrained("loboGOAT/DogeAI-v1.0-instruct")

inputs = tokenizer("Olá! Vamos conversar?", return_tensors="pt")
outputs = model.generate(
    **inputs,
    max_new_tokens=128,
    temperature=0.65,
    top_p=0.95
)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Training Details
Training Data

The model was fine-tuned on a custom instruction-style dataset, primarily in Portuguese, designed to encourage conversational responses.
The dataset does not focus on formal logic or structured reasoning.

Training Procedure

Preprocessing

Instruction–response formatting

Text normalization

No explicit chain-of-thought supervision

Training Hyperparameters

Training regime: Supervised fine-tuning (SFT)
PEFT: Yes (LoRA-based fine-tuning)

Evaluation
Testing Data

Manual testing and prompt-based evaluation.

Factors

Logical consistency

Instruction-following

Language fluency

Metrics

No automated benchmarks were used for this version.

Results

Strong conversational fluency for model size

Inconsistent logical reasoning

Occasional overgeneration beyond intended response

Summary
Model Examination

DogeAI-v1.0-instruct demonstrates the strengths and limitations of small instruction-tuned language models.
While capable of natural conversation, it lacks robust reasoning abilities, which will be a focus of future iterations.

Environmental Impact

Hardware Type: Consumer GPU / Local Machine
Hours used: Low
Cloud Provider: None
Compute Region: Local
Carbon Emitted: Negligible

Technical Specifications
Model Architecture and Objective

Decoder-only Transformer

Next-token prediction

Instruction-following objective

Compute Infrastructure

Local training environment.

Hardware

Consumer-grade GPU / CPU

Software

Transformers

PEFT 0.18.0

PyTorch

Citation

BibTeX:

@misc{dogeai_v1_2025,
  title={DogeAI-v1.0-instruct},
  author={Arthur},
  year={2025},
  note={Early experimental instruction-tuned language model}
}


APA:
Arthur (2025). DogeAI-v1.0-instruct: An experimental instruction-tuned language model.

Model Card Authors

Arthur

Model Card Contact

(your Hugging Face profile or GitHub)