Spaces:
Running
Running
File size: 1,561 Bytes
8b3905d | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 | # Core services
DATABASE_URL=postgresql+asyncpg://sentinel:sentinel@localhost:5432/sentinelai
REDIS_URL=redis://localhost:6379/0
SKIP_DB=1
# AsyncPG connect timeout (seconds) when PostgreSQL is unreachable — avoids long hangs
DB_CONNECT_TIMEOUT_SEC=10
CORS_ORIGINS=http://localhost:3000
# Collectors
COLLECTOR_WATCH_DIR=./demo_logs
COLLECTOR_POLL_SEC=1.0
COLLECTOR_MISSING_RETRY_SEC=10
COLLECTOR_HOSTNAME=edge-01
COLLECTOR_FILE_PATHS=
COLLECT_AUTH_LOG=0
AUTH_LOG_PATH=/var/log/auth.log
ENABLE_MOCK_CLOUD_POLL=1
# LangGraph compile dry-run runs in the background after startup (optional skip / timeout)
# SKIP_LANGGRAPH_WARMUP=1
# LANGGRAPH_WARMUP_TIMEOUT_SEC=120
AUTO_AI_ON_INCIDENT=1
AUTO_AI_MIN_SEC=75
# Local LLM (Ollama)
OLLAMA_HOST=http://localhost:11434
OLLAMA_MODEL=llama3
# OpenAI-compatible inference (vLLM, Fireworks, OpenAI, etc.)
# Must be the INFERENCE base URL — not http://localhost:8000 (that is this app’s API).
# Fireworks example: https://api.fireworks.ai/inference/v1
VLLM_BASE_URL=
VLLM_API_KEY=
# Fireworks example: accounts/fireworks/models/deepseek-v4-pro
SENTINEL_LLM_MODEL=llama3
OPENAI_BASE_URL=
OPENAI_API_KEY=
LLM_TEMPERATURE=0.2
LLM_MAX_TOKENS=4096
# Optional Fireworks-style sampling: LLM_TOP_P=1 LLM_TOP_K=40
# Optional threat intel
ABUSEIPDB_API_KEY=
VIRUSTOTAL_API_KEY=
OTX_API_KEY=
# Alerting
SLACK_WEBHOOK_URL=
DISCORD_WEBHOOK_URL=
TEAMS_WEBHOOK_URL=
GENERIC_ALERT_WEBHOOK=
# Frontend (Next.js) — also set when running npm run dev
NEXT_PUBLIC_API_URL=http://localhost:8000
CHROMA_PERSIST_DIR=./chroma_data
|