ATLAS-1: Neural World Model

ATLAS-1 is a LoRA adapter for Qwen3.5-35B-A3B (MoE: 35B total params, 3B active) that predicts execution outcomes before tasks run — a neural world model for AI orchestration.

What it does

Given a task description and system state, ATLAS-1 predicts:

  • Success probability (0-1)
  • Estimated duration (milliseconds)
  • Token usage and cost (USD)
  • Recommended agent and model
  • Risk factors (specific, actionable warnings)
  • Prediction confidence (0-1)

Example

Input (system state):

{
  "task": "Refactor the authentication service to use dependency injection",
  "taskType": "code",
  "complexity": "complex",
  "selectedAgent": "economist",
  "selectedModel": "deepseek-chat",
  "sessionLength": 45,
  "systemLoad": 0.82,
  "recentAgentSuccess": 0.6
}

Output (prediction):

{
  "willSucceed": 0.31,
  "estimatedDuration": 185000,
  "estimatedTokens": 22000,
  "estimatedCost": 0.0088,
  "recommendedAgent": "nova",
  "recommendedModel": "claude-sonnet-4-20250514",
  "riskFactors": [
    "Agent-task mismatch: economist for code refactoring",
    "High system load may cause timeouts",
    "Long session context adds noise"
  ],
  "confidence": 0.88
}

Training Details

  • Base model: Qwen3.5-35B-A3B (hybrid linear attention + Mamba SSM + 256 MoE experts)
  • Method: LoRA (r=32, alpha=64, targets: q_proj + v_proj)
  • Training data: 10,500 examples (500 real execution outcomes + 10,000 distilled from Grok 4.1 Fast Reasoning)
  • Epochs: 3 (1,770 steps)
  • Best validation loss: 0.1952
  • Training time: 22.0 hours on NVIDIA H100 80GB
  • Precision: BF16 with BitsAndBytes 4-bit quantization (NF4)
  • Optimizer: AdamW 8-bit, LR=2e-5 with cosine decay

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
import json

base_model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen3.5-35B-A3B", trust_remote_code=True)
model = PeftModel.from_pretrained(base_model, "hyperspaceai/atlas-1-lora")
tokenizer = AutoTokenizer.from_pretrained("hyperspaceai/atlas-1-lora")

system = "You are ATLAS-1, a neural world model for the Thor AI orchestration system. Given a structured description of a task and system state, predict the execution outcome. Respond with valid JSON only."
state = json.dumps({"task": "Fix the login bug", "taskType": "debug", "complexity": "moderate", "selectedAgent": "nova", "selectedModel": "grok-4-1-fast-reasoning", "sessionLength": 5, "systemLoad": 0.3, "recentAgentSuccess": 0.85})

prompt = f"<|im_start|>system\n{system}<|im_end|>\n<|im_start|>user\n{state}<|im_end|>\n<|im_start|>assistant\n"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=256)
prediction = json.loads(tokenizer.decode(outputs[0], skip_special_tokens=True).split("assistant\n")[-1])

Data Generation

Training data was generated using knowledge distillation from Grok 4.1 Fast Reasoning, which reasons about agent-task fit, model-task fit, system load effects, and realistic cost/duration estimates. Real execution outcomes from the Thor production system provide ground truth grounding.

Part of the Thor Ecosystem

ATLAS-1 runs as a pre-execution prediction step in the Thor AI orchestration pipeline, enabling proactive risk detection and resource optimization before tasks execute.

Developed by: HyperspaceAI

Downloads last month
15
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for hyperspaceai/atlas-1-lora

Adapter
(13)
this model