Text Generation
Portuguese
File size: 3,882 Bytes
b8a046a
 
 
 
29249c7
 
 
 
 
b8a046a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
---
license: apache-2.0
language:
- pt
datasets:
- AxionLab-Co/ThinkSet-PTBR
metrics:
- accuracy: 16.9%
pipeline_tag: text-generation
---

**๐Ÿง  MiniAxion1.5-3M**

**Emergent reasoning in a 2.7M parameter model.
A tiny Portuguese-first language model that learns how to think before it learns how to be correct.**

**๐Ÿš€ Overview**

MiniAxion1.5-3M is an ultra-compact (~2.7M parameters) GPT-style language model designed to investigate reasoning emergence at extreme small scale.

Unlike typical small models optimized for fluency, MiniAxion is explicitly trained to produce:

Structured reasoning traces
Step-by-step thinking (<THINK><STEP>)
Deterministic answer formatting

It operates primarily in Portuguese, making it a rare example of a non-English reasoning-first nano model.

**โšก Why This Model Is Interesting**

Most models follow this trajectory:

Language โ†’ Knowledge โ†’ Reasoning

MiniAxion flips part of that:

Structure โ†’ Reasoning format โ†’ (still learning correctness)

**๐Ÿ’ก Key insight:**

The model demonstrates that reasoning structure can emerge independently of reasoning accuracy.

**๐Ÿงช Evaluation**
Task Performance
Task	Accuracy
Addition	10%
Subtraction	10%
Multiplication	0%
Even/Odd	100%
Comparison	5%
Sequence Completion	0%
Word Problems (Addition)	10%
Word Problems (Subtraction)	0%
Word Problems (Multiplication)	10%
True/False	100%
Chat/Greetings	100%

**๐Ÿง  Reasoning Behavior Metrics**
Metric	Score
Thinking Rate	100%
Step Format	100%
Answer Completion	100%

โœ” The model always thinks
โœ” The model always structures reasoning
โœ” The model always produces an answer

**๐Ÿ“Š Interpretation**

MiniAxion exhibits a clear dissociation:

โœ… What it learned
Reasoning format
Step-by-step decomposition
Logical task patterns (parity, boolean)
โŒ What it did NOT learn
Arithmetic correctness
Numerical reasoning
Multi-step computation

**๐Ÿ”ฌ Core Finding**

Reasoning โ‰  Correctness

MiniAxion shows that:

Models can internalize thinking patterns
Without actually learning how to solve problems

This makes it a strong candidate for studying:

Emergent reasoning
Tiny Recursive Models (TRMs)
Reasoning distillation

**๐Ÿ—๏ธ Architecture**
Type: GPT-style Transformer
Parameters: ~2.7M
Objective: Next-token prediction
Language: Portuguese (primary)
Specialization: Structured reasoning traces

**๐Ÿง  Training Strategy**

The model was trained with a reasoning-first approach:

Portuguese language grounding
Structured reasoning data (<THINK><STEP>)
Emphasis on:
Deterministic formats
Multi-step thinking
Explicit reasoning tokens

๐Ÿšซ No RLHF
๐Ÿšซ No instruction tuning at scale
๐Ÿšซ No large model distillation (yet)

โš ๏ธ Limitations
1. Arithmetic Collapse

Near-random performance in:

Addition

Subtraction

Multiplication

โ†’ Indicates lack of numerical representation learning

Strong dependence on:

Prompt format

Token patterns

Seen reasoning templates

**๐Ÿ”ฎ Future Work**

This model is just the beginning.

๐Ÿ“ˆ Scaling

5M / 10M / 20M versions

Track emergence of correctness

๐Ÿงช Distillation

Inject reasoning from larger models

Improve accuracy without scaling params

๐Ÿ” Self-Play / Synthetic Data

Generate reasoning loops

Reinforce correct chains

๐Ÿงฉ Hybrid Reasoning

Combine symbolic + neural learning

Fix arithmetic weakness

๐Ÿงพ Example Output

<THINK>
<STEP> Identifico os nรบmeros
<STEP> Tento somar os valores
<STEP> Ajusto o resultado
</THINK>
<ANSWER> 74 </ANSWER>

โœ” Perfect reasoning structure
โŒ Incorrect answer

**๐Ÿ’ก Takeaway**

MiniAxion1.5-3M proves something important:

Even a 2.7M model can learn to simulate thinking before it learns to actually think correctly.

**๐Ÿค Use Cases**

Research on emergent reasoning

Tiny model experimentation (CPU-friendly)

Educational demos of:

Chain-of-Thought

Reasoning failure modes

Base model for:

Distillation

NRM experiments