File size: 4,253 Bytes
8bed955 317cb78 8bed955 e219fe7 8bed955 317cb78 8bed955 317cb78 8bed955 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 | ---
license: mit
language:
- en
metrics:
pipeline_tag: text-generation
tags:
- nrm
- nano
- reasoning
- thinking
- sub-1m
- lowparams
- custom_code
---
# ๐ง MiniAxion1-0.9M
**MiniAxion1-0.9M** is a Nano Reasoning Model (NRM) with ~920K parameters designed to explore the emergence of structured reasoning in extremely small neural networks.
Despite its minimal size, the model demonstrates strong consistency in reasoning format and step-based thinking using explicit `<THINK>` and `<STEP>` tokens.
---
## ๐ Overview
* **Model Type:** Nano Reasoning Model (NRM)
* **Parameters:** ~920,833
* **Architecture:** Transformer (6 layers: 2 entry + 2 shared + 2 exit)
* **d_model:** 256
* **Heads:** 8
* **FFN size:** 512
* **LoRA Rank:** 16
* **Vocabulary Size:** 2048
* **Training Time:** ~80 minutes (CPU)
---
## ๐ง Key Capabilities
### โ
Structured Reasoning
The model reliably produces structured reasoning traces:
```
<THINK>
<STEP> ...
<STEP> ...
</THINK>
<ANS>...</ANS>
```
* 100% usage of reasoning tokens
* Consistent multi-step formatting
* Stable output structure across tasks
---
### โก Ultra-Lightweight
* Runs efficiently on CPU
* Designed for experimentation and rapid iteration
* Suitable for embedded or game-like environments
---
### ๐งช Research-Oriented Design
MiniAxion1 is not intended to compete with large-scale models. Instead, it is built to:
* Study reasoning emergence in small models
* Explore structure vs correctness trade-offs
* Enable fast iteration cycles for AI research
---
## ๐ Evaluation Results
| Task | Accuracy |
| ----------------------- | -------- |
| Arithmetic | 3.3% |
| Two-Step Arithmetic | 10.0% |
| Even/Odd | 100.0% |
| Comparison | 5.0% |
| Pattern Completion | 0.0% |
| Word Problems | 0.0% |
| Sorting | 0.0% |
| Chain-of-Thought Format | 100.0% |
**Average Accuracy:** 16.9%
---
## ๐ Observations
* The model learns reasoning *structure* before reasoning *correctness*
* Chain-of-thought formatting is highly reliable
* Arithmetic and symbolic reasoning remain limited at this scale
* Evidence of partial decoupling between reasoning steps and final answers
---
## โ ๏ธ Limitations
* Weak performance on arithmetic and multi-step reasoning tasks
* Susceptible to incorrect intermediate reasoning steps
* Limited generalization beyond trained patterns
* Not suitable for production use in critical systems
* Due to 920k parameters, low results on evaluation is expected
---
## ๐ฏ Intended Use Cases
* ๐งช AI research and experimentation
* ๐ฎ Game AI / NPC reasoning simulation
* ๐ Educational demonstrations of reasoning structure
* โ๏ธ Lightweight reasoning prototypes
---
### Quick start
```python
import torch
from model import NRMModel
from tokenizer import Tokenizer
# load
model = NRMModel.from_config("config.json")
model.load_state_dict(torch.load("model.pt"))
model.eval()
tokenizer = Tokenizer.load("tokenizer.json")
def generate(prompt):
tokens = tokenizer.encode(prompt)
output = model.generate(tokens)
return tokenizer.decode(output)
print(generate("<INST>What is 2 + 2?</INST>"))
```
## ๐ง Philosophy
MiniAxion1 explores a key question:
> *Can structured reasoning emerge in extremely small models?*
This model provides early evidence that:
* Reasoning format can be learned efficiently
* Structure and correctness are separable capabilities
* Useful behavior can emerge even at sub-1M scale
---
## ๐ฎ Future Directions
* Improved dataset alignment for arithmetic reasoning
* Scaling parameters (1M โ 10M range)
* Better coupling between reasoning and answers
* Task-specific specialization (e.g., math-only variants)
* distillation knowledge on bigger models
---
## ๐ค Acknowledgments
This model was developed as part of ongoing experimentation in nano-scale reasoning systems.
the main question was: "How low could a model think(or mimic it)?
---
## ๐ Model
๐ https://huggingface.co/AxionLab-Co/MiniAxion1-0.9M
---
## ๐งช Disclaimer
This is an experimental research model. Outputs may be incorrect even when reasoning appears structured or convincing.
|