| | --- |
| | license: apache-2.0 |
| | datasets: |
| | - HuggingFaceFW/fineweb-edu |
| | - microsoft/orca-math-word-problems-200k |
| | - sahil2801/CodeAlpaca-20k |
| | - ttbui/alpaca_data_with_html_output |
| | - yahma/alpaca-cleaned |
| | language: |
| | - en |
| | tags: |
| | - small |
| | - tiny |
| | - llm |
| | - finetuned |
| | - instruct |
| | - code |
| | - coding |
| | - math |
| | - cpu |
| | - fast |
| | --- |
| | |
| | **After the success of Apex 1.5 Coder, I've built something entirely new: *Axiom 1 Coder*. It's 350M, but trained with 120k Orca-Math samples and FineWeb-Edu.** |
| |
|
| | This model is based on Apex 1.6 Instruct, my newest and best model for chat and facts without coding. |
| |
|
| | Stay tuned - weights and code coming march 2026! |