Text Generation
Portuguese
MiniAxion1.5-3M / README.md
AxionLab-official's picture
Update README.md
29249c7 verified
---
license: apache-2.0
language:
- pt
datasets:
- AxionLab-Co/ThinkSet-PTBR
metrics:
- accuracy: 16.9%
pipeline_tag: text-generation
---
**๐Ÿง  MiniAxion1.5-3M**
**Emergent reasoning in a 2.7M parameter model.
A tiny Portuguese-first language model that learns how to think before it learns how to be correct.**
**๐Ÿš€ Overview**
MiniAxion1.5-3M is an ultra-compact (~2.7M parameters) GPT-style language model designed to investigate reasoning emergence at extreme small scale.
Unlike typical small models optimized for fluency, MiniAxion is explicitly trained to produce:
Structured reasoning traces
Step-by-step thinking (<THINK><STEP>)
Deterministic answer formatting
It operates primarily in Portuguese, making it a rare example of a non-English reasoning-first nano model.
**โšก Why This Model Is Interesting**
Most models follow this trajectory:
Language โ†’ Knowledge โ†’ Reasoning
MiniAxion flips part of that:
Structure โ†’ Reasoning format โ†’ (still learning correctness)
**๐Ÿ’ก Key insight:**
The model demonstrates that reasoning structure can emerge independently of reasoning accuracy.
**๐Ÿงช Evaluation**
Task Performance
Task Accuracy
Addition 10%
Subtraction 10%
Multiplication 0%
Even/Odd 100%
Comparison 5%
Sequence Completion 0%
Word Problems (Addition) 10%
Word Problems (Subtraction) 0%
Word Problems (Multiplication) 10%
True/False 100%
Chat/Greetings 100%
**๐Ÿง  Reasoning Behavior Metrics**
Metric Score
Thinking Rate 100%
Step Format 100%
Answer Completion 100%
โœ” The model always thinks
โœ” The model always structures reasoning
โœ” The model always produces an answer
**๐Ÿ“Š Interpretation**
MiniAxion exhibits a clear dissociation:
โœ… What it learned
Reasoning format
Step-by-step decomposition
Logical task patterns (parity, boolean)
โŒ What it did NOT learn
Arithmetic correctness
Numerical reasoning
Multi-step computation
**๐Ÿ”ฌ Core Finding**
Reasoning โ‰  Correctness
MiniAxion shows that:
Models can internalize thinking patterns
Without actually learning how to solve problems
This makes it a strong candidate for studying:
Emergent reasoning
Tiny Recursive Models (TRMs)
Reasoning distillation
**๐Ÿ—๏ธ Architecture**
Type: GPT-style Transformer
Parameters: ~2.7M
Objective: Next-token prediction
Language: Portuguese (primary)
Specialization: Structured reasoning traces
**๐Ÿง  Training Strategy**
The model was trained with a reasoning-first approach:
Portuguese language grounding
Structured reasoning data (<THINK><STEP>)
Emphasis on:
Deterministic formats
Multi-step thinking
Explicit reasoning tokens
๐Ÿšซ No RLHF
๐Ÿšซ No instruction tuning at scale
๐Ÿšซ No large model distillation (yet)
โš ๏ธ Limitations
1. Arithmetic Collapse
Near-random performance in:
Addition
Subtraction
Multiplication
โ†’ Indicates lack of numerical representation learning
Strong dependence on:
Prompt format
Token patterns
Seen reasoning templates
**๐Ÿ”ฎ Future Work**
This model is just the beginning.
๐Ÿ“ˆ Scaling
5M / 10M / 20M versions
Track emergence of correctness
๐Ÿงช Distillation
Inject reasoning from larger models
Improve accuracy without scaling params
๐Ÿ” Self-Play / Synthetic Data
Generate reasoning loops
Reinforce correct chains
๐Ÿงฉ Hybrid Reasoning
Combine symbolic + neural learning
Fix arithmetic weakness
๐Ÿงพ Example Output
<THINK>
<STEP> Identifico os nรบmeros
<STEP> Tento somar os valores
<STEP> Ajusto o resultado
</THINK>
<ANSWER> 74 </ANSWER>
โœ” Perfect reasoning structure
โŒ Incorrect answer
**๐Ÿ’ก Takeaway**
MiniAxion1.5-3M proves something important:
Even a 2.7M model can learn to simulate thinking before it learns to actually think correctly.
**๐Ÿค Use Cases**
Research on emergent reasoning
Tiny model experimentation (CPU-friendly)
Educational demos of:
Chain-of-Thought
Reasoning failure modes
Base model for:
Distillation
NRM experiments