File size: 4,253 Bytes
8bed955
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
317cb78
8bed955
 
 
 
 
 
 
 
 
 
 
 
e219fe7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8bed955
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
317cb78
8bed955
 
 
 
 
 
317cb78
8bed955
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
---
license: mit
language:
- en
metrics:

pipeline_tag: text-generation
tags:
- nrm
- nano
- reasoning
- thinking
- sub-1m
- lowparams
- custom_code
---

# ๐Ÿง  MiniAxion1-0.9M

**MiniAxion1-0.9M** is a Nano Reasoning Model (NRM) with ~920K parameters designed to explore the emergence of structured reasoning in extremely small neural networks.

Despite its minimal size, the model demonstrates strong consistency in reasoning format and step-based thinking using explicit `<THINK>` and `<STEP>` tokens.

---

## ๐Ÿš€ Overview

* **Model Type:** Nano Reasoning Model (NRM)
* **Parameters:** ~920,833
* **Architecture:** Transformer (6 layers: 2 entry + 2 shared + 2 exit)
* **d_model:** 256
* **Heads:** 8
* **FFN size:** 512
* **LoRA Rank:** 16
* **Vocabulary Size:** 2048
* **Training Time:** ~80 minutes (CPU)

---

## ๐Ÿง  Key Capabilities

### โœ… Structured Reasoning

The model reliably produces structured reasoning traces:

```
<THINK>
<STEP> ...
<STEP> ...
</THINK>
<ANS>...</ANS>
```

* 100% usage of reasoning tokens
* Consistent multi-step formatting
* Stable output structure across tasks

---

### โšก Ultra-Lightweight

* Runs efficiently on CPU
* Designed for experimentation and rapid iteration
* Suitable for embedded or game-like environments

---

### ๐Ÿงช Research-Oriented Design

MiniAxion1 is not intended to compete with large-scale models. Instead, it is built to:

* Study reasoning emergence in small models
* Explore structure vs correctness trade-offs
* Enable fast iteration cycles for AI research

---

## ๐Ÿ“Š Evaluation Results

| Task                    | Accuracy |
| ----------------------- | -------- |
| Arithmetic              | 3.3%     |
| Two-Step Arithmetic     | 10.0%    |
| Even/Odd                | 100.0%   |
| Comparison              | 5.0%     |
| Pattern Completion      | 0.0%     |
| Word Problems           | 0.0%     |
| Sorting                 | 0.0%     |
| Chain-of-Thought Format | 100.0%   |

**Average Accuracy:** 16.9%

---

## ๐Ÿ” Observations

* The model learns reasoning *structure* before reasoning *correctness*
* Chain-of-thought formatting is highly reliable
* Arithmetic and symbolic reasoning remain limited at this scale
* Evidence of partial decoupling between reasoning steps and final answers

---

## โš ๏ธ Limitations

* Weak performance on arithmetic and multi-step reasoning tasks
* Susceptible to incorrect intermediate reasoning steps
* Limited generalization beyond trained patterns
* Not suitable for production use in critical systems
* Due to 920k parameters, low results on evaluation is expected

---

## ๐ŸŽฏ Intended Use Cases

* ๐Ÿงช AI research and experimentation
* ๐ŸŽฎ Game AI / NPC reasoning simulation
* ๐Ÿ“š Educational demonstrations of reasoning structure
* โš™๏ธ Lightweight reasoning prototypes

---

### Quick start

```python

import torch
from model import NRMModel
from tokenizer import Tokenizer

# load
model = NRMModel.from_config("config.json")
model.load_state_dict(torch.load("model.pt"))
model.eval()

tokenizer = Tokenizer.load("tokenizer.json")

def generate(prompt):
    tokens = tokenizer.encode(prompt)
    output = model.generate(tokens)
    return tokenizer.decode(output)

print(generate("<INST>What is 2 + 2?</INST>"))
```

## ๐Ÿง  Philosophy

MiniAxion1 explores a key question:

> *Can structured reasoning emerge in extremely small models?*

This model provides early evidence that:

* Reasoning format can be learned efficiently
* Structure and correctness are separable capabilities
* Useful behavior can emerge even at sub-1M scale

---

## ๐Ÿ”ฎ Future Directions

* Improved dataset alignment for arithmetic reasoning
* Scaling parameters (1M โ†’ 10M range)
* Better coupling between reasoning and answers
* Task-specific specialization (e.g., math-only variants)
* distillation knowledge on bigger models

---

## ๐Ÿค Acknowledgments

This model was developed as part of ongoing experimentation in nano-scale reasoning systems.
the main question was: "How low could a model think(or mimic it)?

---

## ๐Ÿ“Ž Model

๐Ÿ‘‰ https://huggingface.co/AxionLab-Co/MiniAxion1-0.9M

---

## ๐Ÿงช Disclaimer

This is an experimental research model. Outputs may be incorrect even when reasoning appears structured or convincing.