"use client"; import { useEffect, useMemo } from "react"; import { useSentinel } from "./hooks/useSentinel"; import HeroCanvas from "./components/HeroCanvas"; import SystemModules from "./components/SystemModules"; import AgentTrustMonitor from "./components/AgentTrustMonitor"; import SimCanvas from "./components/SimCanvas"; import ExecutionLog from "./components/ExecutionLog"; import ArchitecturePipeline from "./components/ArchitecturePipeline"; import MetricsGrid from "./components/MetricsGrid"; import GPUClusterPanel from "./components/GPUClusterPanel"; import type { TaskType, AutoPolicy } from "./lib/types"; export default function Page() { const s = useSentinel(); useEffect(() => { void s.resetEpisode(); // eslint-disable-next-line react-hooks/exhaustive-deps }, []); const adversarialAgents = useMemo(() => new Set( s.events .filter(e => e.outcome === "poisoned" || e.outcome === "blocked") .map(e => e.specialist) .filter((x): x is string => !!x) ), [s.events]); const totalReward = s.info?.total_reward ?? 0; return ( <> {/* NAV */} {/* HERO */}
SYS.CORE // SENTINEL v2.4.1 // MARL FRAMEWORK

Train AI to Trust
and Survive Adversaries

A multi-agent reinforcement learning system where an orchestrator learns to detect deception, assign trust, and optimize decisions in real-time adversarial environments.

{s.observation?.available_specialists?.length ?? 5}
Active Agents
{s.proof?.trained ? Math.round(s.proof.trained.avg_detection_rate * 100) : 92 }%
Trust Accuracy
{s.proof?.trained ? s.proof.trained.avg_score.toFixed(2) : "0.91" }
Avg Score
{/* SYSTEM MODULES */}
01 // SYSTEM MODULES

Core Architecture

Each module operates as an independent inference layer within the trust-calibration pipeline. All components communicate via the orchestration bus.

{/* SIMULATION */}
02 // LIVE PREVIEW

Simulation Control Panel

Real-time orchestrator view. Agent trust scores update per-step. Red indicates flagged adversarial behaviour.

SENTINEL // ORCHESTRATOR VIEW // TASK: {s.taskType?.toUpperCase() ?? "TASK3"} // STEP: {s.observation?.step_count ?? 0}
● {s.running ? "LIVE" : s.done ? "DONE" : "READY"}
{/* LEFT: AGENTS */} {/* CENTER: CANVAS */}
{/* RIGHT: LOGS */}
{/* Controls Row */}
POLICY: {s.running && ( )} ACTIONS:
{/* Sim Footer */}
TASK: {s.taskType?.toUpperCase() ?? "TASK3"} SEED: {s.seed} ALGO: DQN+TCE SESSION: {s.sessionId?.slice(0, 8) ?? "—"}
{/* GPU CLUSTER */}
{/* ARCHITECTURE */}
04 // SYSTEM DESIGN

Execution Pipeline

Data flows unidirectionally through the trust-calibrated RL loop. Each stage emits telemetry to the monitoring bus.

{/* METRICS */}
05 // EVALUATION RESULTS

Experimental Benchmarks

Averaged across evaluation episodes. Adversarial injection ratio fixed at 20%. Baseline: naive averaging orchestrator without trust calibration.

{/* FOOTER */} ); }