KennedyOfficaly's picture
Upload 2 files
66e3bbc verified
# AXL — Architecture eXperimental Lab
**27 CPU-Optimized Code Generation Models** by [Koinic](https://koinic.ai)
All models are trained from scratch on consumer hardware (AMD Ryzen 5 5600G, 16GB RAM). No GPU required.
## Models
### Lion-Optimized (Recommended)
| Model | Params | PPL | GGUF (F16) | GGUF (Q4_K_M) |
|-------|--------|-----|-----------|---------------|
| [AXL-Code-1B-Lion](AXL-Code-1B-Lion) | 318M | **1.90** | 606 MB | 188 MB |
| [AXL-Reasoning-Lion](AXL-Reasoning-Lion) | 70M | **1.03** | 134 MB | 44 MB |
| [AXL-Refactor-Lion](AXL-Refactor-Lion) | 19.1M | 1.02 | 37 MB | 12 MB |
| [AXL-TestGen-Lion](AXL-TestGen-Lion) | 15.2M | 1.02 | 30 MB | 18 MB |
| [AXL-Chat-Lion](AXL-Chat-Lion) | 9.9M | 1.03 | 19 MB | 7 MB |
| [AXL-Micro-Lion](AXL-Micro-Lion) | 12.8M | 1.04 | 25 MB | 15 MB |
| [AXL-Secure-Lion](AXL-Secure-Lion) | 11.7M | 1.03 | 23 MB | 8 MB |
| [AXL-Docs-Lion](AXL-Docs-Lion) | 9.9M | 1.01 | 19 MB | 7 MB |
| [AXL-Comment-Lion](AXL-Comment-Lion) | 7.2M | 1.02 | 14 MB | 5 MB |
### SGD Models
| Model | Params | PPL | Focus |
|-------|--------|-----|-------|
| [AXL-Micro-600K](AXL-Micro-600K) | 600K | 63.08 | Demo |
| [AXL-Micro-8M](AXL-Micro-8M) | 12.8M | 3.13 | Code gen |
| [AXL-Coder-15M](AXL-Coder-15M) | 26.0M | 5.97 | Agentic |
| [AXL-Debugger-8M](AXL-Debugger-8M) | 14.1M | 6.60 | Bug fixing |
| [AXL-Fixer-12M](AXL-Fixer-12M) | 20.9M | 5.90 | Debug |
| [AXL-Reasoning-70M](AXL-Reasoning-70M) | 70M | 1.93 | CoT |
| [AXL-300M](AXL-300M) | 322M | 5.98 | Flagship |
| [AXL-Chat-10M](AXL-Chat-10M) | 9.9M | 1.02 | Dialogue |
| [AXL-TestGen-15M](AXL-TestGen-15M) | 15.2M | 1.01 | Test gen |
| [AXL-Refactor-20M](AXL-Refactor-20M) | 19.1M | 1.01 | Refactoring |
| [AXL-Docs-8M](AXL-Docs-8M) | 9.9M | 1.03 | Docstrings |
| [AXL-Comment-5M](AXL-Comment-5M) | 7.2M | 1.01 | Comments |
| [AXL-Secure-10M](AXL-Secure-10M) | 11.7M | 1.01 | Security |
### Specialized Models
| Model | Params | PPL | Focus |
|-------|--------|-----|-------|
| [AXL-Code-1B](AXL-Code-1B) | 318M | 31.22 | Code gen (SGD) |
| [AXL-Chat-Pro](AXL-Chat-Pro) | 12.8M | 3.42 | Advanced chat |
| [AXL-Translate](AXL-Translate) | 15.2M | 1.86 | Code translation |
| [AXL-Vision-0.8M](AXL-Vision-0.8M) | 1M | — | Vision encoder |
| [AXL-Vision-v2](AXL-Vision-v2) | 4.1M | — | UI vision |
## Quick Start
### Python API Server (Full Quality - Recommended)
```bash
pip install -e .
python AXL/API/serve_model.py --model AXL-Micro-Lion/ --port 8880
# OpenAI-compatible endpoint:
curl http://localhost:8880/v1/completions \
-H "Content-Type: application/json" \
-d '{"prompt": "def fibonacci(n):", "max_tokens": 100}'
# Works with Continue.dev, LlamaIndex, LangChain, Cursor
```
### With Ollama (Degraded Quality)
> **Warning:** GGUF files for Ollama use only the fine-scale encoder (1/3 of the AXL architecture). The reported PPL values apply to the full multi-scale model. Use the Python API above for full quality.
```bash
cd AXL-Micro-Lion
ollama create axl-micro-lion -f Modelfile
ollama run axl-micro-lion "def fibonacci(n):"
```
### With Python (Direct Inference)
```python
import torch
from multiscale_transformer.model.config import load_config, ModelConfig
from multiscale_transformer.model.model import MultiScaleTransformer
from multiscale_transformer.training.tokenizer import ByteTokenizer
config = load_config("AXL-Micro-Lion/config.json")
model = MultiScaleTransformer(config)
ckpt = torch.load("AXL-Micro-Lion/axl_micro_lion.pt", map_location="cpu")
model.load_state_dict(ckpt["model_state_dict"])
model.eval()
tokenizer = ByteTokenizer()
ids = torch.tensor([tokenizer.encode("def hello():")], dtype=torch.long)
out = model.generate(ids, max_new_tokens=50, temperature=0.8, top_k=40)
print(tokenizer.decode(out[0].tolist()))
```
## Architecture
AXL processes token sequences at three parallel resolution scales:
- **Fine (1x)**: All tokens. Attention cost: O(N^2 d)
- **Medium (2x)**: Grouped in pairs. Cost: O(N^2 d/4)
- **Coarse (4x)**: Grouped in quadruplets. Cost: O(N^2 d/16)
Cross-scale attention connects all scale pairs. Adaptive gating fusion combines representations.
**Lion optimizer**: Sign-based momentum, 20x faster convergence than SGD, 50% less memory than AdamW.
**Byte-level tokenizer**: 258 vocab (256 bytes + BOS + EOS). No vocabulary training. Works with any programming language.
## Training Cost
| Model | Time | Cost (USD) |
|-------|------|-----------|
| AXL-Comment-Lion | 2 min | $0.0004 |
| AXL-Code-1B-Lion | 20 min | $0.004 |
| All 9 Lion models | 49 min | $0.010 |
Based on AMD Ryzen 5 5600G (100W system, $0.12/kWh).
## Papers
- [Research Paper (PDF)](paper_axl.pdf)
- [Research Paper (LaTeX)](paper_axl.tex)
## Code
Full training code: [GitHub](https://github.com/koinic/axl)