LumynaX MaramaRoute

Sovereign intelligence, held in the light.

Ko te mārama te tūāpapa — the light is the foundation.

A sovereign model router for the LumynaX release family.
AbteeX AI Labs — Aotearoa New Zealand.

What it is · Quickstart · Routing contract · Scoring · Registry · Roadmap · Companions

MaramaRoute Stage scaffold Runtime python License Apache 2.0 Card v1

What It Is

LumynaX MaramaRoute is the AbteeX AI Labs sovereign model router for LumynaX releases. It is similar in spirit to OpenRouter, but it is not a general marketplace. It is a LumynaX-first routing layer that selects models based on sovereignty, residency, license, runtime, modality, task fit, context length, and operational risk.

Marama = light, clarity. The router brings the right model into the light for the work at hand.

What It Routes To

  • Local GGUF models.
  • MoE and frontier-style LumynaX packages.
  • Multimodal LumynaX packages (text + image / audio / voice).
  • Embedding and retrieval models.
  • Reasoning and coding variants.
  • Future tenant-specific sovereign models.

The bundled registry covers 50 models across the AbteeXAILab family.

Quickstart

Clone the repo:

hf download AbteeXAILab/marama-route --local-dir marama-route --repo-type model
cd marama-route
pip install -r requirements.txt

Route a restricted code request (requires local runtime, NZ residency):

python -m marama_route.cli route \
  --registry configs/lumynax_model_registry.json \
  --request examples/request.code-restricted.json

Expected: a LumynaX coder or Qwen-family GGUF package with NZ residency constraints satisfied, plus an ordered list of fallbacks.

Route a public multimodal request:

python -m marama_route.cli route \
  --registry configs/lumynax_model_registry.json \
  --request examples/request.multimodal-public.json

Expected: a multimodal LumynaX package with text + image modalities.

Routing Contract

Every request is evaluated through ordered gates:

Gate Purpose
Capability Modalities, context length, tool use, JSON mode, task fit.
Sovereignty Jurisdiction, residency, data sensitivity, local runtime requirement.
License Optional license allowlist and model-card provenance.
Runtime llama.cpp, Transformers, embedding, multimodal, or hosted adapter.
Score Quality, cost, active parameters, task tags, and fallback strength.
Audit Decision, rejected models, selected model, and fallbacks are recorded.

Scoring Signals

Signal Reason
Residency match Keeps governed data inside approved regions.
Sovereignty tier Allows policy packs to enforce stronger local controls.
Task tags Routes code, reasoning, embedding, and multimodal tasks to specialised models.
Runtime Prefers local GGUF / llama.cpp for sensitive work.
Quality rank Keeps stronger models ahead when policy allows them.
Cost rank Avoids oversized models when smaller models are sufficient.
Active parameters Helps sparse MoE models compete when active footprint is small.

Registry

The registry (configs/lumynax_model_registry.json) is a flat array of model entries. Each entry carries:

Field Meaning
repo_id Hugging Face repository id.
family Upstream family (qwen, gemma, phi, granite, olmo, mistral, deepseek, embedding, ...).
runtime llama_cpp, transformers, python_embedding, llama_cpp_multimodal, ...
modalities text, image, audio, embedding.
context_tokens Max context window.
jurisdiction / residency Where the model is approved to run.
sovereignty_tier 1 (open) → 5 (strict).
quality_rank / cost_rank Routing scorer inputs.
supports_tools / supports_json Capability flags.
total_params_b / active_params_b Total / active parameter counts.

Refresh the registry from a fresh HF card report:

python -m marama_route.build_registry --report path/to/hf-model-card-report.json --out configs/lumynax_model_registry.json

Planned API Surface

Endpoint Purpose
GET /v1/models List candidate models.
POST /v1/route Return a deterministic route decision.
GET /v1/route/{decision_id} Retrieve a stored decision.
POST /v1/chat/completions OpenAI-compatible completions wrapper.
POST /v1/embeddings Embedding wrapper.

The first implementation focuses on the deterministic router and CLI — the HTTP gateway ships in P1.

Roadmap

Milestone Outcome
P0 scaffold (now) Registry, router, CLI, examples, docs, tests.
P1 OpenAI-compatible API /v1/models, /v1/route, /v1/chat/completions wrapper.
P2 Live runtime adapters llama.cpp, llama-cpp-python, Transformers, embedding, multimodal.
P3 Tenant policy packs Per-customer region, license, sensitivity, allowlist rules.
P4 Evaluation loop Quality, acceptance, speed, safety metrics per model.
P5 Hosted control plane Private customer gateway with signed route + audit records.

Companion Products

Product Purpose
AbteeX SovereignCode Local-first coding agent with Data Capsule policy. Uses MaramaRoute for every model call.
LumynaX Live Demo Public browser demo.
MaramaRoute Live Interactive router — paste a request, see the selected model and fallbacks.
AbteeXAILab on Hugging Face The full LumynaX release family.

Aotearoa Kaupapa

MaramaRoute is built in and for Aotearoa New Zealand. Routing is not just performance — it is kaitiakitanga: guardianship of where data goes, which model touches it, and what audit trail remains.

Limitations & Responsible Use

  • The router enforces declared registry metadata and policy. It cannot detect undeclared licence or residency issues.
  • The current release is a product scaffold. The hosted gateway, runtime adapters, and tenant policy server ship in P1–P5.
  • For high-impact routing, pair MaramaRoute with SovereignCode policy enforcement and human review.

Local roots, global work. · Sovereignty is a design property, not a deployment option.

AbteeX AI Labs · abteex.com · lumynax.com · huggingface.co/AbteeXAILab

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Spaces using AbteeXAILab/marama-route 3