ChipYTY's picture
Add files using upload-large-folder tool
fe8202e verified
import argparse
import glob
import os
import re
import sys
import numpy as np
import torch
import SimpleITK as sitk
# Prefer pip-installed MONAI over the local monai/ folder.
os.environ.setdefault("MONAI_SKIP_SUBMODULES", "1")
_repo_root = os.path.abspath(os.path.dirname(__file__))
if "" in sys.path:
sys.path.remove("")
if _repo_root in sys.path:
sys.path.remove(_repo_root)
import monai # noqa: E402
sys.path.insert(0, _repo_root)
from monai.inferers import SlidingWindowInferer
from monai.utils import set_determinism
from light_training.dataloading.dataset import MedicalDataset, get_train_val_test_loader_from_train
from light_training.evaluation.metric import dice
from light_training.prediction import Predictor
from light_training.trainer import Trainer
set_determinism(123)
def _parse_csv_ints(s: str, n: int):
parts = [p.strip() for p in str(s).split(",") if p.strip()]
if len(parts) != n:
raise ValueError(f"expect {n} integers like '128,128,128', got: {s}")
return [int(x) for x in parts]
def _parse_csv_floats(s: str, n: int):
parts = [p.strip() for p in str(s).split(",") if p.strip()]
if len(parts) != n:
raise ValueError(f"expect {n} floats like '1,1,1', got: {s}")
return [float(x) for x in parts]
def _find_ckpt_from_logdir(logdir: str, prefer: str = "best") -> str:
model_dir = os.path.join(logdir, "model")
if not os.path.isdir(model_dir):
raise FileNotFoundError(f"model dir not found: {model_dir}")
best = sorted(glob.glob(os.path.join(model_dir, "best_model_*.pt")))
final = sorted(glob.glob(os.path.join(model_dir, "final_model_*.pt")))
tmp = sorted(glob.glob(os.path.join(model_dir, "tmp_model_ep*.pt")))
any_pt = sorted(glob.glob(os.path.join(model_dir, "*.pt")))
def pick_by_score(paths):
# filenames like best_model_0.9038.pt / final_model_0.9038.pt
scored = []
for p in paths:
m = re.search(r"_(\d+\\.?\\d*)\\.pt$", os.path.basename(p))
if m is None:
continue
try:
scored.append((float(m.group(1)), p))
except ValueError:
continue
if scored:
scored.sort(key=lambda x: x[0], reverse=True)
return scored[0][1]
return None
if prefer == "best":
picked = pick_by_score(best) or (best[-1] if best else None)
if picked:
return picked
if prefer in {"best", "final"}:
picked = pick_by_score(final) or (final[-1] if final else None)
if picked:
return picked
if prefer in {"best", "final", "latest"}:
if tmp:
tmp.sort(key=lambda p: os.path.getmtime(p), reverse=True)
return tmp[0]
if any_pt:
any_pt.sort(key=lambda p: os.path.getmtime(p), reverse=True)
return any_pt[0]
raise FileNotFoundError(f"no checkpoint found under: {model_dir}")
class BraTSTrainer(Trainer):
def __init__(
self,
ckpt_path: str,
save_path: str,
patch_size,
sw_batch_size: int = 2,
overlap: float = 0.5,
mirror_axes=(0, 1, 2),
raw_spacing=(1.0, 1.0, 1.0),
device="cuda:0",
print_dice: bool = False,
):
super().__init__(
env_type="pytorch",
max_epochs=1,
batch_size=1,
device=device,
val_every=1,
num_gpus=1,
logdir="",
master_port=17751,
training_script=__file__,
)
self.patch_size = patch_size
self.augmentation = False
self.print_dice = print_dice
self.save_path = save_path
self.raw_spacing = raw_spacing
from model_segmamba.segmamba import SegMamba
self.model = SegMamba(
in_chans=4,
out_chans=4,
depths=[2, 2, 2, 2],
feat_size=[48, 96, 192, 384],
)
self.load_state_dict(ckpt_path, strict=True)
self.model.eval()
window_infer = SlidingWindowInferer(
roi_size=patch_size,
sw_batch_size=sw_batch_size,
overlap=overlap,
progress=True,
mode="gaussian",
)
self.predictor = Predictor(
window_infer=window_infer,
mirror_axes=list(mirror_axes) if mirror_axes is not None else None,
)
os.makedirs(self.save_path, exist_ok=True)
def convert_labels(self, labels):
## TC, WT and ET
result = [(labels == 1) | (labels == 3), (labels == 1) | (labels == 3) | (labels == 2), labels == 3]
return torch.cat(result, dim=1).float()
def get_input(self, batch):
image = batch["data"]
label = batch["seg"]
properties = batch["properties"]
label = self.convert_labels(label)
return image, label, properties
def validation_step(self, batch):
image, label, properties = self.get_input(batch)
# The preprocessed datasets used in many setups (including /data/yty/brats23_processed)
# do NOT contain cropping/resample metadata (shape_before_cropping, bbox_used_for_cropping, ...),
# so we directly save predictions in the same (D,H,W) space as the inputs.
#
# We save as a TRUE 4D NIfTI (t,z,y,x) with t=3 (TC/WT/ET) so that
# `sitk.GetArrayFromImage` returns shape (3, D, H, W), matching `5_compute_metrics.py`.
logits = self.predictor.maybe_mirror_and_predict(image, self.model, device=self.device) # (1,4,D,H,W) on CPU
pred_lbl = logits.argmax(dim=1) # (1,D,H,W)
pred_3c = self.convert_labels(pred_lbl[:, None])[0].cpu().numpy().astype(np.uint8) # (3,D,H,W)
if self.print_dice:
gt_3c = label[0].cpu().numpy()
dices = [dice(pred_3c[i], gt_3c[i]) for i in range(3)]
print(dices)
case_name = properties.get("name", "")
if isinstance(case_name, (list, tuple)) and len(case_name) > 0:
case_name = case_name[0]
out_path = os.path.join(self.save_path, f"{case_name}.nii.gz")
pred_itk = sitk.GetImageFromArray(pred_3c, isVector=False)
pred_itk.SetSpacing((float(self.raw_spacing[0]), float(self.raw_spacing[1]), float(self.raw_spacing[2]), 1.0))
sitk.WriteImage(pred_itk, out_path)
print(f"saved: {out_path}")
return 0
def convert_labels_dim0(self, labels):
## TC, WT and ET
result = [(labels == 1) | (labels == 3), (labels == 1) | (labels == 3) | (labels == 2), labels == 3]
return torch.cat(result, dim=0).float()
def main():
parser = argparse.ArgumentParser(description="SegMamba inference/prediction for BraTS2023.")
parser.add_argument("--data_dir", type=str, default="./data/fullres/train", help="Preprocessed data directory (contains *.npz).")
parser.add_argument("--split", type=str, default="test", choices=["train", "val", "test", "all"])
parser.add_argument("--train_rate", type=float, default=0.7)
parser.add_argument("--val_rate", type=float, default=0.1)
parser.add_argument("--test_rate", type=float, default=0.2)
parser.add_argument("--seed", type=int, default=42)
parser.add_argument("--ckpt", type=str, default="", help="Checkpoint path (*.pt). If empty, will search under --logdir/model.")
parser.add_argument("--logdir", type=str, default="./logs/segmamba", help="Training logdir to locate checkpoints when --ckpt is empty.")
parser.add_argument("--ckpt_prefer", type=str, default="best", choices=["best", "final", "latest"])
parser.add_argument("--save_dir", type=str, default="./prediction_results/segmamba", help="Directory to save prediction nii.gz.")
parser.add_argument("--device", type=str, default="cuda:0")
parser.add_argument("--patch_size", type=str, default="128,128,128")
parser.add_argument("--sw_batch_size", type=int, default=2)
parser.add_argument("--overlap", type=float, default=0.5)
parser.add_argument("--raw_spacing", type=str, default="1,1,1", help="Spacing used when saving NIfTI, e.g. '1,1,1'.")
parser.add_argument("--no_mirror", action="store_true", help="Disable mirror TTA.")
parser.add_argument("--print_dice", action="store_true", help="Print dice against preprocessed seg (if available).")
args = parser.parse_args()
patch_size = _parse_csv_ints(args.patch_size, 3)
raw_spacing = _parse_csv_floats(args.raw_spacing, 3)
ckpt_path = args.ckpt.strip()
if ckpt_path == "":
ckpt_path = _find_ckpt_from_logdir(args.logdir, prefer=args.ckpt_prefer)
if not os.path.isfile(ckpt_path):
raise FileNotFoundError(f"checkpoint not found: {ckpt_path}")
print(f"Using checkpoint: {ckpt_path}")
trainer = BraTSTrainer(
ckpt_path=ckpt_path,
save_path=args.save_dir,
patch_size=patch_size,
sw_batch_size=args.sw_batch_size,
overlap=args.overlap,
mirror_axes=None if args.no_mirror else (0, 1, 2),
raw_spacing=raw_spacing,
device=args.device,
print_dice=args.print_dice,
)
if args.split == "all":
all_paths = sorted(glob.glob(os.path.join(args.data_dir, "*.npz")))
ds = MedicalDataset(all_paths, test=False)
else:
train_ds, val_ds, test_ds = get_train_val_test_loader_from_train(
args.data_dir,
train_rate=args.train_rate,
val_rate=args.val_rate,
test_rate=args.test_rate,
seed=args.seed,
)
ds = {"train": train_ds, "val": val_ds, "test": test_ds}[args.split]
trainer.validation_single_gpu(ds)
if __name__ == "__main__":
main()