""" ctx_config.py -- Central configuration loader for the Alive Skill System. Loads from (in priority order): 1. ~/.claude/skill-system-config.json (user's deployed config, highest priority) 2. /config.json (repo default config) Usage: from ctx_config import cfg wiki_dir = cfg.wiki_dir max_skills = cfg.max_skills """ import json import os import sys import tempfile from importlib import resources from pathlib import Path from typing import Any _SCRIPT_DIR = Path(__file__).parent _DEFAULT_CONFIG = _SCRIPT_DIR / "config.json" _USER_CONFIG = Path(os.path.expanduser("~/.claude/skill-system-config.json")) def _read_default_config() -> dict[str, Any]: """Read repo defaults from source checkout or packaged wheel data.""" try: if _DEFAULT_CONFIG.exists(): return json.loads(_DEFAULT_CONFIG.read_text(encoding="utf-8")) if _DEFAULT_CONFIG != _SCRIPT_DIR / "config.json": return {} packaged = resources.files("ctx").joinpath("config.json") if packaged.is_file(): return json.loads(packaged.read_text(encoding="utf-8")) except Exception as exc: print(f"Warning: failed to load default config: {exc}", file=sys.stderr) return {} def _load_raw() -> dict[str, Any]: """Load and merge default + user config.""" raw: dict[str, Any] = _read_default_config() if _USER_CONFIG.exists(): try: user = json.loads(_USER_CONFIG.read_text(encoding="utf-8")) # Deep merge: user values override defaults _deep_merge(raw, user) except Exception as exc: print(f"Warning: failed to load user config: {exc}", file=sys.stderr) return raw def _deep_merge(base: dict, override: dict) -> None: """Merge override into base in-place (recursive for nested dicts).""" for k, v in override.items(): if k in base and isinstance(base[k], dict) and isinstance(v, dict): _deep_merge(base[k], v) else: base[k] = v def _expand(value: str) -> str: """Expand ~ and env vars in path strings.""" return os.path.expandvars(os.path.expanduser(value)) def _default_stack_profile_tmp() -> Path: return Path(tempfile.gettempdir()) / "skill-stack-profile.json" class Config: """Typed access to configuration values.""" def __init__(self, raw: dict[str, Any]) -> None: self._raw = raw paths = raw.get("paths", {}) resolver = raw.get("resolver", {}) monitor = raw.get("context_monitor", {}) tracker = raw.get("usage_tracker", {}) transformer = raw.get("skill_transformer", {}) router = raw.get("skill_router", {}) intake = raw.get("intake", {}) intake_emb = intake.get("embedding", {}) if isinstance(intake, dict) else {} harness = raw.get("harness", {}) if isinstance(raw.get("harness"), dict) else {} bsitter = raw.get("babysitter", {}) # ── Paths ────────────────────────────────────────────────────────── self.claude_dir = Path(_expand(paths.get("claude_dir", "~/.claude"))) self.wiki_dir = Path(_expand(paths.get("wiki_dir", "~/.claude/skill-wiki"))) self.skills_dir = Path(_expand(paths.get("skills_dir", "~/.claude/skills"))) self.agents_dir = Path(_expand(paths.get("agents_dir", "~/.claude/agents"))) self.skill_manifest = Path(_expand(paths.get("skill_manifest", "~/.claude/skill-manifest.json"))) self.intent_log = Path(_expand(paths.get("intent_log", "~/.claude/intent-log.jsonl"))) self.pending_skills = Path(_expand(paths.get("pending_skills", "~/.claude/pending-skills.json"))) self.skill_registry = Path(_expand(paths.get("skill_registry", "~/.claude/skill-registry.json"))) stack_profile_tmp = paths.get("stack_profile_tmp") self.stack_profile_tmp = ( Path(_expand(stack_profile_tmp)) if stack_profile_tmp else _default_stack_profile_tmp() ) self.catalog = Path(_expand(paths.get("catalog", "~/.claude/skill-wiki/catalog.md"))) # ── Resolver ─────────────────────────────────────────────────────── self.max_skills: int = resolver.get("max_skills", 15) self.recommendation_top_k: int = int(resolver.get("recommendation_top_k", 5)) self.recommendation_min_normalized_score: float = float( resolver.get("recommendation_min_normalized_score", 0.30) ) self.intent_boost_per_signal: int = resolver.get("intent_boost_per_signal", 5) self.intent_boost_max: int = resolver.get("intent_boost_max", 15) self.staleness_penalty: int = resolver.get("staleness_penalty", -8) self.meta_skills: list[str] = resolver.get("meta_skills", ["skill-router", "file-reading"]) if not (1 <= self.recommendation_top_k <= 5): raise ValueError( f"resolver.recommendation_top_k must be in [1, 5] " f"(got {self.recommendation_top_k})" ) if not (0.0 <= self.recommendation_min_normalized_score <= 1.0): raise ValueError( "resolver.recommendation_min_normalized_score must be in [0, 1] " f"(got {self.recommendation_min_normalized_score})" ) # ── Harness Catalog ──────────────────────────────────────────────── self.harness_recommendation_min_normalized_score: float = float( harness.get("recommendation_min_normalized_score", 0.85) ) self.harness_recommendation_min_fit_score: float = float( harness.get( "recommendation_min_fit_score", self.harness_recommendation_min_normalized_score, ) ) raw_reliability_weights = harness.get("reliability_weights", {}) if not isinstance(raw_reliability_weights, dict): raw_reliability_weights = {} default_reliability_weights = { "context": 0.34, "constraints": 0.33, "convergence": 0.33, } self.harness_reliability_weights: dict[str, float] = {} for dimension, default in default_reliability_weights.items(): value = float(raw_reliability_weights.get(dimension, default)) if value < 0: raise ValueError( f"harness.reliability_weights.{dimension} must be >= 0 " f"(got {value})" ) self.harness_reliability_weights[dimension] = value reliability_weight_total = sum(self.harness_reliability_weights.values()) if reliability_weight_total <= 0: raise ValueError("harness.reliability_weights must sum to > 0") self.harness_reliability_weights = { dimension: value / reliability_weight_total for dimension, value in self.harness_reliability_weights.items() } if not (0.0 <= self.harness_recommendation_min_normalized_score <= 1.0): raise ValueError( "harness.recommendation_min_normalized_score must be in [0, 1] " f"(got {self.harness_recommendation_min_normalized_score})" ) if not (0.0 <= self.harness_recommendation_min_fit_score <= 1.0): raise ValueError( "harness.recommendation_min_fit_score must be in [0, 1] " f"(got {self.harness_recommendation_min_fit_score})" ) # ── Context Monitor ──────────────────────────────────────────────── self.unmatched_signal_threshold: int = monitor.get("unmatched_signal_threshold", 3) self.manifest_stale_minutes: int = monitor.get("manifest_stale_minutes", 60) # ── Usage Tracker ────────────────────────────────────────────────── self.stale_threshold_sessions: int = tracker.get("stale_threshold_sessions", 30) self.keep_log_days: int = tracker.get("keep_log_days", 5) # ── Skill Transformer ────────────────────────────────────────────── raw_line_threshold = transformer.get("line_threshold", 180) if ( isinstance(raw_line_threshold, bool) or not isinstance(raw_line_threshold, int) ): raise ValueError( "skill_transformer.line_threshold must be an integer >= 1 " f"(got {raw_line_threshold!r})" ) self.line_threshold = raw_line_threshold if self.line_threshold < 1: raise ValueError( "skill_transformer.line_threshold must be an integer >= 1 " f"(got {self.line_threshold})" ) self.max_stage_lines: int = transformer.get("max_stage_lines", 40) self.stage_count: int = transformer.get("stage_count", 5) # ── Skill Router ─────────────────────────────────────────────────── self.manifest_stale_router_minutes: int = router.get("manifest_stale_minutes", 60) self.manifest_max_age_hours: int = router.get("manifest_max_age_hours", 24) # ── Extra Skill Dirs ─────────────────────────────────────────────── self.extra_skill_dirs: list[Path] = [ Path(_expand(d)) for d in raw.get("extra_skill_dirs", []) ] # ── Tag Taxonomy ────────────────────────────────────────────────── self.all_tags: list[str] = raw.get("tags", [ "python", "javascript", "typescript", "rust", "go", "java", "ruby", "swift", "kotlin", "react", "vue", "angular", "nextjs", "fastapi", "django", "express", "flask", "docker", "kubernetes", "terraform", "ci-cd", "aws", "gcp", "azure", "sql", "nosql", "redis", "kafka", "spark", "dbt", "airflow", "llm", "agents", "mcp", "langchain", "embeddings", "fine-tuning", "rag", "testing", "linting", "typing", "security", "performance", "documentation", "api-spec", "markdown", "diagrams", "comparison", "decision", "pattern", "troubleshooting", "marketplace", "registry", "versioning", "compatibility", ]) # ── Intake Gate ──────────────────────────────────────────────────── # Phase 2 similarity/structure gate for skill_add / agent_add. # When disabled, callers should skip the gate entirely — the # thresholds here only apply when ``intake_enabled`` is True. self.intake_enabled: bool = bool(intake.get("enabled", True)) self.intake_dup_threshold: float = float(intake.get("dup_threshold", 0.93)) self.intake_near_dup_threshold: float = float( intake.get("near_dup_threshold", 0.80) ) self.intake_min_neighbors: int = int(intake.get("min_neighbors", 0)) self.intake_min_neighbor_score: float = float( intake.get("min_neighbor_score", 0.30) ) self.intake_min_body_chars: int = int(intake.get("min_body_chars", 120)) self.intake_cache_root: Path = Path( _expand(intake.get("cache_root", "~/.claude/skills/_embeddings")) ) self.intake_backend: str = str( intake_emb.get("backend", "sentence-transformers") ) # ``None``-valued keys flow through unchanged so downstream # factories can distinguish "use backend default" (None) from # "forced empty string" (never). model = intake_emb.get("model") self.intake_model: str | None = model if isinstance(model, str) else None base_url = intake_emb.get("base_url") self.intake_base_url: str | None = ( base_url if isinstance(base_url, str) else None ) self.intake_allow_remote: bool = bool(intake_emb.get("allow_remote", False)) # ── Babysitter ───────────────────────────────────────────────────── self.babysitter_plugin_root: str = bsitter.get("plugin_root", "") self.babysitter_runs_dir: str = bsitter.get("runs_dir", ".a5c/runs") self.babysitter_sdk_version: str = bsitter.get("sdk_version", "latest") # ── Graph Edge Weights ────────────────────────────────────────────── # wiki_graphify builds base edges from semantic/tag/slug-token # weights plus source/direct evidence, then applies additive # explainability boosts. The semantic/tag/token weights must sum # to 1.0 so the primary blend remains comparable across configs. graph = raw.get("graph", {}) if isinstance(raw.get("graph"), dict) else {} ew = graph.get("edge_weights", {}) if isinstance(graph.get("edge_weights"), dict) else {} self.graph_edge_weight_semantic: float = float(ew.get("semantic", 0.70)) self.graph_edge_weight_tags: float = float(ew.get("tags", 0.15)) self.graph_edge_weight_tokens: float = float(ew.get("slug_tokens", 0.15)) self.graph_edge_min_weight: float = float(graph.get("min_edge_weight", 0.0)) sem = graph.get("semantic", {}) if isinstance(graph.get("semantic"), dict) else {} self.graph_semantic_top_k: int = int(sem.get("top_k", 20)) self.graph_semantic_build_floor: float = float(sem.get("build_floor", 0.50)) self.graph_semantic_min_cosine: float = float(sem.get("min_cosine", 0.80)) self.graph_semantic_batch_size: int = int(sem.get("batch_size", 128)) self.graph_semantic_cache_dir: Path = Path(_expand( sem.get("cache_dir", "~/.claude/skill-wiki/.embedding-cache/graph") )) # Strict (0, 1) open interval on both thresholds. 0 would # include every pair (N^2 explosion); 1 would only match # exact duplicates (floating-point drift means even identical # texts rarely hit exactly 1.0). And build_floor must be # <= min_cosine — otherwise min_cosine would accept edges # the graph never materialised, producing silent gaps. if not (0.0 < self.graph_semantic_build_floor < 1.0): raise ValueError( f"graph.semantic.build_floor must be strictly in (0, 1) " f"(got {self.graph_semantic_build_floor})" ) if not (0.0 < self.graph_semantic_min_cosine < 1.0): raise ValueError( f"graph.semantic.min_cosine must be strictly in (0, 1) " f"(got {self.graph_semantic_min_cosine})" ) if self.graph_semantic_build_floor > self.graph_semantic_min_cosine: raise ValueError( f"graph.semantic.build_floor ({self.graph_semantic_build_floor}) " f"must be <= min_cosine ({self.graph_semantic_min_cosine}); " "otherwise query-time filtering would ask for edges that " "were never materialised into the graph." ) te = graph.get("tag_edges", {}) if isinstance(graph.get("tag_edges"), dict) else {} self.graph_dense_tag_threshold: int = int(te.get("dense_tag_threshold", 500)) self.graph_shared_tag_saturation: int = int(te.get("shared_tag_saturation", 5)) toe = graph.get("token_edges", {}) if isinstance(graph.get("token_edges"), dict) else {} self.graph_dense_token_threshold: int = int(toe.get("dense_token_threshold", 500)) self.graph_shared_token_saturation: int = int(toe.get("shared_token_saturation", 3)) se = graph.get("source_edges", {}) if isinstance(graph.get("source_edges"), dict) else {} self.graph_dense_source_threshold: int = int(se.get("dense_source_threshold", 50)) boosts = graph.get("edge_boosts", {}) if isinstance(graph.get("edge_boosts"), dict) else {} self.graph_edge_boost_direct_link: float = float(boosts.get("direct_link", 0.10)) self.graph_edge_boost_source_overlap: float = float(boosts.get("source_overlap", 0.05)) self.graph_edge_boost_adamic_adar: float = float(boosts.get("adamic_adar", 0.04)) self.graph_edge_boost_type_affinity: float = float(boosts.get("type_affinity", 0.03)) self.graph_edge_boost_usage: float = float(boosts.get("usage", 0.02)) self.graph_edge_boost_quality: float = float(boosts.get("quality", 0.02)) # Validate the blend weights sum to 1.0 (±1e-6 tolerance). A # misconfigured user config is better caught here than 10 min # into a regraphify pass when scores come out wrong. weight_sum = ( self.graph_edge_weight_semantic + self.graph_edge_weight_tags + self.graph_edge_weight_tokens ) if abs(weight_sum - 1.0) > 1e-6: raise ValueError( f"graph.edge_weights must sum to 1.0 " f"(got {weight_sum:.4f}: semantic={self.graph_edge_weight_semantic}, " f"tags={self.graph_edge_weight_tags}, " f"slug_tokens={self.graph_edge_weight_tokens})" ) for name, val in ( ("semantic", self.graph_edge_weight_semantic), ("tags", self.graph_edge_weight_tags), ("slug_tokens", self.graph_edge_weight_tokens), ): if val < 0.0: raise ValueError( f"graph.edge_weights.{name} must be >= 0 (got {val})" ) if not (0.0 <= self.graph_edge_min_weight <= 1.0): raise ValueError( "graph.min_edge_weight must be in [0, 1] " f"(got {self.graph_edge_min_weight})" ) if self.graph_dense_source_threshold < 1: raise ValueError( "graph.source_edges.dense_source_threshold must be >= 1 " f"(got {self.graph_dense_source_threshold})" ) for name, val in ( ("direct_link", self.graph_edge_boost_direct_link), ("source_overlap", self.graph_edge_boost_source_overlap), ("adamic_adar", self.graph_edge_boost_adamic_adar), ("type_affinity", self.graph_edge_boost_type_affinity), ("usage", self.graph_edge_boost_usage), ("quality", self.graph_edge_boost_quality), ): if val < 0.0: raise ValueError( f"graph.edge_boosts.{name} must be >= 0 (got {val})" ) def get(self, key: str, default: Any = None) -> Any: """Raw key access (dot-separated: 'paths.wiki_dir').""" parts = key.split(".") node: Any = self._raw for p in parts: if isinstance(node, dict) and p in node: node = node[p] else: return default return node def all_skill_dirs(self) -> list[Path]: """Return all skill directories (primary + extra).""" dirs = [self.skills_dir, self.agents_dir] + self.extra_skill_dirs return [d for d in dirs if d.exists()] def build_intake_config(self) -> Any: """Construct an ``intake_gate.IntakeConfig`` from these settings. Lazy-imported so ``ctx_config`` stays free of the numpy / embedding dependency graph when callers don't need the intake gate. """ from intake_gate import IntakeConfig # noqa: PLC0415 return IntakeConfig( dup_threshold=self.intake_dup_threshold, near_dup_threshold=self.intake_near_dup_threshold, min_neighbors=self.intake_min_neighbors, min_neighbor_score=self.intake_min_neighbor_score, min_body_chars=self.intake_min_body_chars, ) def build_intake_embedder(self) -> Any: """Construct the configured embedding backend. Lazy-imported for the same reason as ``build_intake_config``. Callers pay the heavy-model cost only when they ask for it. """ from embedding_backend import get_embedder # noqa: PLC0415 return get_embedder( backend=self.intake_backend, model=self.intake_model, base_url=self.intake_base_url, allow_remote=self.intake_allow_remote, ) # Singleton instance — import this cfg = Config(_load_raw()) def reload() -> None: """Reload config from disk (useful if config changed during session).""" global cfg cfg = Config(_load_raw())