Sovereign-Code / README.md
Vaultkeeper's picture
Update README.md
c1fe38a verified
metadata
license: other
language:
  - en
tags:
  - text-generation
  - mergekit
  - coding
  - agentic
  - reasoning
  - qwen2.5
  - llama-3.1
  - transformers
  - merge
  - sovereign
base_model:
  - Qwen/Qwen2.5-Coder-7B-Instruct
  - meta-llama/Meta-Llama-3.1-8B-Instruct
pipeline_tag: text-generation
Sovereign-Code

Sovereign-Code

by VaultAI

VAULTAI

Deployment Status: UNRELEASED

[ PRE-ALPHA ] SOVEREIGN-CODE & CORPUS-CALLOSUM | ARCHITECTING...


Execution, Absolute.

While most models are built to converse, Sovereign-Code is built to execute. It is a specialized, cold-logic engine designed for a single purpose: high-fidelity technical output.

Engineered by VaultAI, Sovereign-Code is a custom 32-Layer Hybrid model. It utilizes an aggressive architectural "passthrough" to bridge the deep structural coding intelligence of Qwen 2.5 Coder with the rigid, high-instruction-following cortex of Llama 3.1. It does not offer opinions; it delivers functional syntax.

🧠 Architecture & Identity: The Logic Terminal

Sovereign-Code is a "Frankenmerge" that ignores standard architectural safety to achieve peak performance. By stacking disparate layers, VaultAI has created a model that processes raw intent through a coding-heavy base before filtering it through an elite instruction-following top-layer.

Key Capabilities:

  • Deterministic Syntax: Optimized for zero-fluff code generation across Python, C++, Rust, and Mojo.
  • Tattooed Monologue: Hardcoded via a custom Jinja2 template to engage in a mandatory three-phase internal processing loop inside <think> tags before every output.
  • Hardware Optimized: Designed for dual-GPU configurations (Polaris/gfx803) using llama.cpp and Vulkan backends.

⚡ Performance & Benchmarks (Estimated)

Sovereign-Code is designed for maximum throughput on local consumer hardware (RX 570/580 8GB setups).

Metric Target Hardware VRAM Footprint Logic Mode
Quantization Q4_K_M (GGUF) ~9.2 GB Full GPU Offload
Context Length 32,768 Tokens High Headroom Optimized for Repo-level Debugging

Standardized Accuracy Benchmarks

Benchmarks are currently queued for evaluation.

Benchmark Focus Area Score Status
HumanEval Coding & Logic TBD ⏳ Pending Eval
MBPP Python Programming TBD ⏳ Pending Eval
GSM8k Mathematical Reasoning TBD ⏳ Pending Eval

Model Details

  • Type: Causal Language Model (Hybrid Passthrough)
  • Base Architecture: Qwen 2.5 (7B) + Llama 3.1 (8B)
  • Total Parameters: ~15B (Effective density via Layer Stacking)
  • Merge Method: Passthrough / Frankenmerge
  • Weights Composition:
  • License: Other (See Base Model Licenses)

Why Sovereign-Code?

  • The Execution Engine: No conversational "As an AI..." filler.
  • Analytical Grounding: The built-in <think> protocol forces the model to debug its own code conceptually before writing a single line.
  • Agentic Ready: Optimized for tool-calling and autonomous development workflows.
VAULTAI