| # AXL — Architecture eXperimental Lab
|
|
|
| **27 CPU-Optimized Code Generation Models** by [Koinic](https://koinic.ai)
|
|
|
| All models are trained from scratch on consumer hardware (AMD Ryzen 5 5600G, 16GB RAM). No GPU required.
|
|
|
| ## Models
|
|
|
| ### Lion-Optimized (Recommended)
|
|
|
| | Model | Params | PPL | GGUF (F16) | GGUF (Q4_K_M) |
|
| |-------|--------|-----|-----------|---------------|
|
| | [AXL-Code-1B-Lion](AXL-Code-1B-Lion) | 318M | **1.90** | 606 MB | 188 MB |
|
| | [AXL-Reasoning-Lion](AXL-Reasoning-Lion) | 70M | **1.03** | 134 MB | 44 MB |
|
| | [AXL-Refactor-Lion](AXL-Refactor-Lion) | 19.1M | 1.02 | 37 MB | 12 MB |
|
| | [AXL-TestGen-Lion](AXL-TestGen-Lion) | 15.2M | 1.02 | 30 MB | 18 MB |
|
| | [AXL-Chat-Lion](AXL-Chat-Lion) | 9.9M | 1.03 | 19 MB | 7 MB |
|
| | [AXL-Micro-Lion](AXL-Micro-Lion) | 12.8M | 1.04 | 25 MB | 15 MB |
|
| | [AXL-Secure-Lion](AXL-Secure-Lion) | 11.7M | 1.03 | 23 MB | 8 MB |
|
| | [AXL-Docs-Lion](AXL-Docs-Lion) | 9.9M | 1.01 | 19 MB | 7 MB |
|
| | [AXL-Comment-Lion](AXL-Comment-Lion) | 7.2M | 1.02 | 14 MB | 5 MB |
|
|
|
| ### SGD Models
|
|
|
| | Model | Params | PPL | Focus |
|
| |-------|--------|-----|-------|
|
| | [AXL-Micro-600K](AXL-Micro-600K) | 600K | 63.08 | Demo |
|
| | [AXL-Micro-8M](AXL-Micro-8M) | 12.8M | 3.13 | Code gen |
|
| | [AXL-Coder-15M](AXL-Coder-15M) | 26.0M | 5.97 | Agentic |
|
| | [AXL-Debugger-8M](AXL-Debugger-8M) | 14.1M | 6.60 | Bug fixing |
|
| | [AXL-Fixer-12M](AXL-Fixer-12M) | 20.9M | 5.90 | Debug |
|
| | [AXL-Reasoning-70M](AXL-Reasoning-70M) | 70M | 1.93 | CoT |
|
| | [AXL-300M](AXL-300M) | 322M | 5.98 | Flagship |
|
| | [AXL-Chat-10M](AXL-Chat-10M) | 9.9M | 1.02 | Dialogue |
|
| | [AXL-TestGen-15M](AXL-TestGen-15M) | 15.2M | 1.01 | Test gen |
|
| | [AXL-Refactor-20M](AXL-Refactor-20M) | 19.1M | 1.01 | Refactoring |
|
| | [AXL-Docs-8M](AXL-Docs-8M) | 9.9M | 1.03 | Docstrings |
|
| | [AXL-Comment-5M](AXL-Comment-5M) | 7.2M | 1.01 | Comments |
|
| | [AXL-Secure-10M](AXL-Secure-10M) | 11.7M | 1.01 | Security |
|
|
|
| ### Specialized Models
|
|
|
| | Model | Params | PPL | Focus |
|
| |-------|--------|-----|-------|
|
| | [AXL-Code-1B](AXL-Code-1B) | 318M | 31.22 | Code gen (SGD) |
|
| | [AXL-Chat-Pro](AXL-Chat-Pro) | 12.8M | 3.42 | Advanced chat |
|
| | [AXL-Translate](AXL-Translate) | 15.2M | 1.86 | Code translation |
|
| | [AXL-Vision-0.8M](AXL-Vision-0.8M) | 1M | — | Vision encoder |
|
| | [AXL-Vision-v2](AXL-Vision-v2) | 4.1M | — | UI vision |
|
|
|
| ## Quick Start |
|
|
| ### Python API Server (Full Quality - Recommended) |
| ```bash |
| pip install -e . |
| python AXL/API/serve_model.py --model AXL-Micro-Lion/ --port 8880 |
| |
| # OpenAI-compatible endpoint: |
| curl http://localhost:8880/v1/completions \ |
| -H "Content-Type: application/json" \ |
| -d '{"prompt": "def fibonacci(n):", "max_tokens": 100}' |
| |
| # Works with Continue.dev, LlamaIndex, LangChain, Cursor |
| ``` |
|
|
| ### With Ollama (Degraded Quality) |
| > **Warning:** GGUF files for Ollama use only the fine-scale encoder (1/3 of the AXL architecture). The reported PPL values apply to the full multi-scale model. Use the Python API above for full quality. |
|
|
| ```bash |
| cd AXL-Micro-Lion |
| ollama create axl-micro-lion -f Modelfile |
| ollama run axl-micro-lion "def fibonacci(n):" |
| ``` |
|
|
| ### With Python (Direct Inference) |
| ```python
|
| import torch
|
| from multiscale_transformer.model.config import load_config, ModelConfig
|
| from multiscale_transformer.model.model import MultiScaleTransformer
|
| from multiscale_transformer.training.tokenizer import ByteTokenizer
|
|
|
| config = load_config("AXL-Micro-Lion/config.json")
|
| model = MultiScaleTransformer(config)
|
| ckpt = torch.load("AXL-Micro-Lion/axl_micro_lion.pt", map_location="cpu")
|
| model.load_state_dict(ckpt["model_state_dict"])
|
| model.eval()
|
|
|
| tokenizer = ByteTokenizer()
|
| ids = torch.tensor([tokenizer.encode("def hello():")], dtype=torch.long)
|
| out = model.generate(ids, max_new_tokens=50, temperature=0.8, top_k=40)
|
| print(tokenizer.decode(out[0].tolist()))
|
| ```
|
|
|
| ## Architecture
|
|
|
| AXL processes token sequences at three parallel resolution scales:
|
| - **Fine (1x)**: All tokens. Attention cost: O(N^2 d)
|
| - **Medium (2x)**: Grouped in pairs. Cost: O(N^2 d/4)
|
| - **Coarse (4x)**: Grouped in quadruplets. Cost: O(N^2 d/16)
|
|
|
| Cross-scale attention connects all scale pairs. Adaptive gating fusion combines representations.
|
|
|
| **Lion optimizer**: Sign-based momentum, 20x faster convergence than SGD, 50% less memory than AdamW.
|
|
|
| **Byte-level tokenizer**: 258 vocab (256 bytes + BOS + EOS). No vocabulary training. Works with any programming language.
|
|
|
| ## Training Cost
|
|
|
| | Model | Time | Cost (USD) |
|
| |-------|------|-----------|
|
| | AXL-Comment-Lion | 2 min | $0.0004 |
|
| | AXL-Code-1B-Lion | 20 min | $0.004 |
|
| | All 9 Lion models | 49 min | $0.010 |
|
|
|
| Based on AMD Ryzen 5 5600G (100W system, $0.12/kWh).
|
|
|
| ## Papers
|
|
|
| - [Research Paper (PDF)](paper_axl.pdf)
|
| - [Research Paper (LaTeX)](paper_axl.tex)
|
|
|
| ## Code
|
|
|
| Full training code: [GitHub](https://github.com/koinic/axl)
|
|
|