Axon / README.md
AIencoder's picture
Update README.md
1ecbdcf verified
metadata
title: Axon v26
emoji: 
colorFrom: indigo
colorTo: purple
sdk: docker
pinned: true
license: mit
short_description: Free AI Coding Assistant - 5 Models - 25 Tools - 100% Local

⚡ Axon v26

The Ultimate Free AI Coding Assistant

5 Models • 25 Tools • 100% Local • Zero API Keys

Built with llama.cpp Models Tools License AVX2 Optimized


Try Axon Now →


🚀 What is Axon?

Axon is a free, privacy-first AI coding assistant that runs entirely locally using llama.cpp. No API keys, no rate limits, no data collection - just powerful AI coding tools at your fingertips.

Built from the ground up after spending 2 days building llama-cpp-python wheels that didn't exist. Now you don't have to.


🤖 5 Powerful Models

Choose the right model for your task - from instant responses to complex reasoning.

Model Size Speed Best For
🧠 DeepSeek V2 Lite ~9GB ⭐⭐ Complex logic, MoE architecture
⚖️ Qwen2.5 Coder 7B ~4.5GB ⭐⭐⭐ Balanced quality & speed
🚀 Qwen2.5 Coder 3B ~2GB ⭐⭐⭐⭐ Fast & highly capable
💨 Qwen2.5 Coder 1.5B ~1GB ⭐⭐⭐⭐⭐ Quick tasks
🔬 Qwen2.5 Coder 0.5B ~0.3GB Instant responses

Models download automatically on first use. Storage persists between sessions.


🛠️ 25 Tools

Core Tools

Tool Description
💬 Chat Conversational coding help with streaming responses
Generate Create code from natural language descriptions
🔍 Explain Understand any code (Brief / Normal / Detailed modes)
🔧 Debug Find and fix bugs with error context
📋 Review Code quality, security & performance review

Advanced Tools

Tool Description
🔐 Security Scan Find vulnerabilities (SQL injection, XSS, etc.)
📊 Complexity Big O analysis for time & space
🔄 Convert Translate between 22+ programming languages
🧪 Test Generate comprehensive unit tests
📝 Document Add docstrings, comments & inline docs
🚀 Optimize Performance improvements & refactoring
🔀 Diff Compare two code snippets
📐 Pseudocode Convert code to pseudocode/flowcharts
🎓 Interview Generate coding challenges & solutions

Builders

Tool Description
🗄️ SQL Builder Natural language → SQL queries
🐚 Shell Builder Natural language → Bash/PowerShell commands
Cron Builder Create cron schedule expressions
🎯 Regex Builder Pattern creation with explanations
🔗 API Builder Generate REST endpoint boilerplate

Data Tools

Tool Description
📦 Mock Data Generate realistic test data (JSON, CSV, etc.)
🔄 Format Converter Convert between JSON/YAML/XML/CSV/TOML

NEW in v26

Tool Description
🎨 Refactor Restructure code for better design patterns
📊 Benchmark Generate performance benchmark code
🔗 Dependency Analyzer Analyze imports & dependencies
📋 Changelog Generate changelogs from code diffs
💡 Suggest AI-powered improvement suggestions

🎤 Voice Input

Speak your code requests using Whisper-powered speech-to-text. Just click the microphone and talk naturally.


🌙 Dark Mode

Toggle between light and dark themes. Your preference is saved automatically.


💾 Export

Save your chat history and generated code for later reference.


⚡ Performance

Axon uses AVX2-optimized llama-cpp-python wheels for 2-3x faster inference compared to basic builds.

Build Type Tokens/sec (3B) Compatibility
Basic ~10-15 All x86_64
AVX2 (Axon) ~30-40 Intel Haswell+ / AMD Zen+ (2013+)

🔒 Privacy First

Feature Axon Cloud Alternatives
100% Local Processing
No API Keys Required
No Rate Limits
No Data Collection
Works Offline
Free Forever

Your code never leaves your browser. Period.


💻 Supported Languages

Python • JavaScript • TypeScript • Go • Rust • Java • C++ • C# • C • PHP • Ruby • Swift • Kotlin • Scala • R • Julia • Perl • HTML/CSS • SQL • Bash • PowerShell • Lua


🛞 Pre-built Wheels

Tired of building llama-cpp-python from source? Use our AVX2-optimized wheels:

# Python 3.10
pip install https://huggingface.co/datasets/AIencoder/llama-cpp-wheels/resolve/main/llama_cpp_python-0.3.16-cp310-cp310-manylinux_2_31_x86_64.whl

# Python 3.11
pip install https://huggingface.co/datasets/AIencoder/llama-cpp-wheels/resolve/main/llama_cpp_python-0.3.16-cp311-cp311-manylinux_2_31_x86_64.whl

Features:

  • AVX2 + FMA + F16C enabled
  • 2-3x faster than basic builds
  • Works on Intel Haswell+ (2013+) and AMD Zen+ (2018+)

Browse all wheels →


🏗️ Tech Stack

Component Technology
Inference llama.cpp via llama-cpp-python
Wheels AIencoder/llama-cpp-wheels (AVX2 optimized)
UI Gradio
Speech faster-whisper
Models GGUF format from HuggingFace
Hosting HuggingFace Spaces (Docker)

🚀 Self-Hosting

Want to run Axon on your own machine?

# Clone the space
git clone https://huggingface.co/spaces/AIencoder/Axon
cd Axon

# Build and run
docker build -t axon .
docker run -p 7860:7860 -v axon_data:/data axon

Then open http://localhost:7860


📊 Changelog

v26 (Current) - The FINAL Version

  • ✨ Added 6 new tools (25 total)
  • 🎨 Redesigned UI with better UX
  • ⚡ AVX2-optimized wheels for 2-3x speed boost
  • 🔧 Gradio 6.0 compatibility fixes
  • 📦 Optimized storage usage

v6

  • 🚀 Initial public release
  • 🤖 8 models (reduced to 5 due to storage)
  • 🛠️ 19 tools
  • 🎤 Whisper voice input

🙏 Credits


📄 License

MIT License - Use it, modify it, share it!


⭐ Support

If Axon helps you code faster, consider:

  • ⭐ Starring the Space
  • 🐛 Reporting issues
  • 💡 Suggesting features
  • 📢 Sharing with friends

Built with ❤️ and mass caffeine by AIencoder

No sleep was had in the making of those wheels.