ctx / src /change_detector.py
Stevesolun's picture
Sync ctx 4f0c3de
21ff762 verified
"""
change_detector.py -- Detect file changes between current state and last snapshot.
Used by ``backup_mirror.py`` to implement ``snapshot-if-changed``: we
take a new snapshot only when at least one tracked file has actually
changed (by SHA-256 content hash), not on every hook fire.
The public entry point is :func:`detect_changes`. It is a pure function
— no filesystem writes, no globals — so it can be reused by the
PostToolUse hook, the CLI, and a future daemon trigger.
"""
from __future__ import annotations
import hashlib
import json
import os
import stat as _stat
from dataclasses import asdict, dataclass
from pathlib import Path
from typing import Iterable
from backup_config import BackupConfig
# ── Report types ────────────────────────────────────────────────────────────
@dataclass(frozen=True)
class ChangeReport:
"""Diff between current tracked state and a previous snapshot."""
# Files present now but not in the last snapshot.
new: tuple[str, ...]
# Files present in both but with a different content hash.
changed: tuple[str, ...]
# Files in the last snapshot but not present now.
removed: tuple[str, ...]
# Count of files whose hash is identical to the last snapshot.
unchanged: int
# Snapshot ID we compared against, or None if no baseline existed.
baseline_snapshot: str | None
@property
def has_changes(self) -> bool:
"""True if a fresh snapshot would capture new information."""
return bool(self.new or self.changed or self.removed)
@property
def total_current(self) -> int:
"""Total files currently tracked (new + changed + unchanged)."""
return len(self.new) + len(self.changed) + self.unchanged
def to_dict(self) -> dict:
out = asdict(self)
out["has_changes"] = self.has_changes
out["total_current"] = self.total_current
return out
# ── Hashing (duplicated from backup_mirror to keep this module import-cheap) ─
def _sha256_file(path: Path) -> str | None:
"""Content hash, or None if the file is a symlink / unreadable.
Skipped files won't trigger a change-detected event; they also
aren't in any snapshot's manifest since ``backup_mirror`` never
hashes them. Treating both sides identically avoids false positives.
"""
try:
if path.is_symlink():
return None
h = hashlib.sha256()
with open(path, "rb") as fh:
while True:
chunk = fh.read(1 << 20)
if not chunk:
break
h.update(chunk)
return h.hexdigest()
except OSError:
return None
# ── Iteration (mirrors backup_mirror's discovery rules) ─────────────────────
def _iter_top_files(cfg: BackupConfig, claude_home: Path) -> Iterable[tuple[str, Path]]:
for name in cfg.top_files:
src = claude_home / name
if src.is_file() and not src.is_symlink():
yield (name, src)
def _iter_tree_files(cfg: BackupConfig, claude_home: Path) -> Iterable[tuple[str, Path]]:
for tree in cfg.trees:
root = claude_home / tree.src
if not root.is_dir():
continue
for dirpath, _dirnames, filenames in os.walk(root, followlinks=False):
for name in filenames:
src = Path(dirpath) / name
# Skip files that will be dropped during capture.
try:
st = os.lstat(src)
except OSError:
continue
if _stat.S_ISLNK(st.st_mode):
continue
if st.st_size > cfg.max_file_bytes:
continue
rel = src.relative_to(root)
dest_rel = (Path(tree.dest) / rel).as_posix()
yield (dest_rel, src)
def _iter_memory_files(cfg: BackupConfig, claude_home: Path) -> Iterable[tuple[str, Path]]:
if not cfg.memory_glob:
return
projects = claude_home / "projects"
if not projects.is_dir():
return
for slug_dir in projects.iterdir():
if not slug_dir.is_dir():
continue
memory_dir = slug_dir / "memory"
if not memory_dir.is_dir():
continue
for dirpath, _dirnames, filenames in os.walk(memory_dir, followlinks=False):
for name in filenames:
src = Path(dirpath) / name
try:
st = os.lstat(src)
except OSError:
continue
if _stat.S_ISLNK(st.st_mode):
continue
if st.st_size > cfg.max_file_bytes:
continue
rel = src.relative_to(memory_dir)
dest_rel = (Path("memory") / slug_dir.name / rel).as_posix()
yield (dest_rel, src)
def _current_state(cfg: BackupConfig, claude_home: Path) -> dict[str, str]:
"""Hash every currently-tracked file. Returns {dest_rel: sha256}."""
state: dict[str, str] = {}
for dest, src in _iter_top_files(cfg, claude_home):
digest = _sha256_file(src)
if digest is not None:
state[dest] = digest
for dest, src in _iter_tree_files(cfg, claude_home):
digest = _sha256_file(src)
if digest is not None:
state[dest] = digest
for dest, src in _iter_memory_files(cfg, claude_home):
digest = _sha256_file(src)
if digest is not None:
state[dest] = digest
return state
# ── Baseline loading ────────────────────────────────────────────────────────
def _load_snapshot_hashes(snap_path: Path) -> dict[str, str]:
"""Read manifest.json and return {dest: sha256}, skipping skipped entries."""
manifest_path = snap_path / "manifest.json"
if not manifest_path.is_file():
return {}
try:
manifest = json.loads(manifest_path.read_text(encoding="utf-8"))
except (OSError, json.JSONDecodeError):
return {}
entries = manifest.get("entries") or []
out: dict[str, str] = {}
for e in entries:
dest = e.get("dest")
digest = e.get("sha256")
if isinstance(dest, str) and isinstance(digest, str):
out[dest] = digest
return out
def _snapshot_id(snap_path: Path) -> str:
manifest_path = snap_path / "manifest.json"
try:
manifest = json.loads(manifest_path.read_text(encoding="utf-8"))
return str(manifest.get("snapshot_id") or snap_path.name)
except (OSError, json.JSONDecodeError):
return snap_path.name
# ── Public API ──────────────────────────────────────────────────────────────
def detect_changes(
cfg: BackupConfig,
claude_home: Path,
last_snapshot: Path | None,
) -> ChangeReport:
"""Compare current tracked files against a previous snapshot.
Parameters
----------
cfg
Config describing what to track (top_files, trees, memory_glob,
max_file_bytes, excludes).
claude_home
Root of the live ~/.claude tree to inspect.
last_snapshot
Directory of the previous snapshot, or ``None`` to treat every
current file as new.
Returns
-------
ChangeReport
Diff between current state and the snapshot's manifest. Files
exceeding ``max_file_bytes`` or that are symlinks are ignored on
both sides so they never produce phantom changes.
"""
current = _current_state(cfg, claude_home)
if last_snapshot is None or not last_snapshot.is_dir():
return ChangeReport(
new=tuple(sorted(current.keys())),
changed=(),
removed=(),
unchanged=0,
baseline_snapshot=None,
)
baseline = _load_snapshot_hashes(last_snapshot)
baseline_id = _snapshot_id(last_snapshot)
new: list[str] = []
changed: list[str] = []
unchanged = 0
for dest, digest in current.items():
if dest not in baseline:
new.append(dest)
elif baseline[dest] != digest:
changed.append(dest)
else:
unchanged += 1
removed = [d for d in baseline if d not in current]
return ChangeReport(
new=tuple(sorted(new)),
changed=tuple(sorted(changed)),
removed=tuple(sorted(removed)),
unchanged=unchanged,
baseline_snapshot=baseline_id,
)