ADAPT-Chase's picture
Add files using upload-large-folder tool
2021f39 verified

Nova Autonomous LLM System

This directory scaffolds a next‑gen autonomous LLM stack suitable for container deployment using supervisord. It includes a gateway, a basic router, and a tools MCP server, and is designed to integrate with vLLM and external public endpoints.

Key components

  • plane/api/gateway.py: OpenAI‑compatible proxy and route registry
  • plane/orchestration/router.py: Task routing stub for multi‑backend selection
  • plane/tools/mcp_server.py: Minimal MCP‑style tool registry and executor
  • config/external_routes.yaml: External/public URL candidates (Cloudflare and local)
  • requirements-nova.txt: Minimal runtime dependencies

Start locally

  • vLLM: bash elizabeth/deployment_configs/serve_vllm_no_steering.sh
  • Supervisord: supervisord -c supervisord.conf then supervisorctl status

Environment

  • UPSTREAM_OPENAI_BASE: upstream OpenAI‑compatible base URL (default http://127.0.0.1:8000)
  • EXTERNAL_ROUTES_FILE: path to routes YAML
  • PORT: per‑service port overrides