import argparse import glob import os import re import sys from typing import List, Sequence, Tuple import torch from torch.utils.data import DataLoader sys.path.append(os.path.join(os.path.dirname(__file__), "src")) from gliomasam3_moe.data.brats_dataset import SegMambaNPZDataset, split_npz_paths from gliomasam3_moe.models.gliomasam3_moe import GliomaSAM3_MoE from train import evaluate_test, load_config def _find_latest_ckpt(ckpt_dir: str) -> str: pattern = os.path.join(ckpt_dir, "ckpt_step*.pt") matches = [] for path in glob.glob(pattern): m = re.search(r"ckpt_step(\d+)\.pt$", path) if m: matches.append((int(m.group(1)), path)) if not matches: raise FileNotFoundError(f"No checkpoints found under {ckpt_dir}.") matches.sort(key=lambda x: x[0]) return matches[-1][1] def _select_train_subset( data_dir: str, train_rate: float, val_rate: float, test_rate: float, seed: int, ) -> Tuple[Sequence[str], int, int]: train_paths, _, test_paths = split_npz_paths( data_dir, train_rate=train_rate, val_rate=val_rate, test_rate=test_rate, seed=seed ) test_n = len(test_paths) if test_n == 0: raise ValueError("Test split size is 0; cannot match train subset size.") subset_n = min(len(train_paths), test_n) rng = torch.Generator().manual_seed(seed) perm = torch.randperm(len(train_paths), generator=rng).tolist() subset_paths = [train_paths[i] for i in perm[:subset_n]] return subset_paths, test_n, len(train_paths) def main(): parser = argparse.ArgumentParser() parser.add_argument("--config", type=str, default="configs/train.yaml") parser.add_argument("--checkpoint", type=str, default=None, help="Path to ckpt_step*.pt (default: latest in ckpt_dir).") parser.add_argument("--max_cases", type=int, default=0, help="Optional cap on subset size.") args = parser.parse_args() cfg = load_config(args.config) data_dir = cfg.data.root_dir if not os.path.isdir(data_dir): raise FileNotFoundError(f"data.root_dir does not exist: {data_dir}") if getattr(cfg.data, "format", "nifti") != "segmamba_npz": raise ValueError("Only segmamba_npz format is supported for this evaluation script.") ckpt_path = args.checkpoint or _find_latest_ckpt(cfg.train.ckpt_dir) if not os.path.isfile(ckpt_path): raise FileNotFoundError(f"Checkpoint not found: {ckpt_path}") subset_paths, test_n, train_n = _select_train_subset( data_dir, train_rate=getattr(cfg.data, "train_rate", 0.7), val_rate=getattr(cfg.data, "val_rate", 0.1), test_rate=getattr(cfg.data, "test_rate", 0.2), seed=cfg.seed, ) if args.max_cases and args.max_cases > 0: subset_paths = subset_paths[: min(len(subset_paths), args.max_cases)] device = torch.device(cfg.device if torch.cuda.is_available() else "cpu") model = GliomaSAM3_MoE(**cfg.model.__dict__).to(device) ckpt = torch.load(ckpt_path, map_location=device) model.load_state_dict(ckpt["model"], strict=True) ensure_npy = bool(getattr(cfg.data, "segmamba_unpack", True)) dataset = SegMambaNPZDataset( data_dir=data_dir, npz_paths=subset_paths, test=False, ensure_npy=ensure_npy, map_et_to_4=True, ) loader = DataLoader( dataset, batch_size=1, shuffle=False, num_workers=max(0, int(cfg.train.num_workers)), ) metrics = evaluate_test(model, loader, cfg, device) print(f"[TRAIN-SUBSET] ckpt={ckpt_path}") print(f"[TRAIN-SUBSET] total_train={train_n} test_count={test_n} subset={len(subset_paths)}") print( f"[TRAIN-SUBSET] dice[WT,TC,ET]={metrics['dice']} " f"hd95[WT,TC,ET]={metrics['hd95']}" ) if __name__ == "__main__": main()