MEDHEB's picture
init commit
4ab0193
import os
import re
import json
import random
from pathlib import Path
import numpy as np
from PIL import Image
DATASET_A_INPUT_JSON = r"./Kvasir-SEG/kavsir_bboxes.json"
DATASET_A_IMAGE_DIR = r"./Kvasir-SEG/images"
DATASET_A_SPLIT_CSV = r"./split.csv"
DATASET_A_IMAGE_SUFFIX = ".jpg"
DATASET_B_DOWNLOAD_ROOT = r"./Gastronintestinal"
DATASET_B_DOWNLOAD_SIZE = 512
DATASET_B_NPZ_PATH = None
DATASET_B_NPZ_EXTRACT_ROOT = r"./Gastronintestinal"
DATASET_B_SPLITS = ["train", "test", "val"]
OUTPUT_IMAGE_DIR = r"./Gastronintestinal/images"
RANDOM_SEED = 42
IMAGE_EXTS = {".jpg", ".jpeg", ".png", ".bmp", ".tif", ".tiff"}
random.seed(RANDOM_SEED)
def ensure_dir(path):
if path:
os.makedirs(path, exist_ok=True)
def safe_int(x):
return int(float(x))
def to_uint8(arr):
arr = np.asarray(arr)
if arr.dtype == np.uint8:
return arr
arr = arr.astype(np.float32)
if arr.size > 0 and arr.min() >= 0 and arr.max() <= 1.0:
arr = arr * 255.0
arr = np.clip(arr, 0, 255).astype(np.uint8)
return arr
def save_image_from_array(arr, save_path):
arr = to_uint8(arr)
if arr.ndim == 2:
img = Image.fromarray(arr, mode="L")
elif arr.ndim == 3:
if arr.shape[2] == 1:
img = Image.fromarray(arr[:, :, 0], mode="L")
elif arr.shape[2] == 3:
img = Image.fromarray(arr, mode="RGB")
elif arr.shape[2] == 4:
img = Image.fromarray(arr, mode="RGBA")
else:
raise ValueError(f"Unsupported image shape: {arr.shape}")
else:
raise ValueError(f"Unsupported image shape: {arr.shape}")
img.save(save_path)
def crop_and_save(bbox_rows, output_image_dir):
ensure_dir(output_image_dir)
total = len(bbox_rows)
saved_count = 0
print(f"[CROP] Total rows: {total}")
for idx, row in enumerate(bbox_rows):
split = str(row["split"]).strip()
image_name = str(row["image_name"]).strip()
image_path = str(row["image_path"]).strip()
x_min = safe_int(row["x_min"])
y_min = safe_int(row["y_min"])
x_max = safe_int(row["x_max"])
y_max = safe_int(row["y_max"])
if not os.path.exists(image_path):
print(f"[WARN] Image not found, skip: {image_path}")
continue
try:
img = Image.open(image_path).convert("RGB")
except Exception as e:
print(f"[WARN] Failed to open image: {image_path}, error: {e}")
continue
w, h = img.size
x_min = max(0, min(x_min, w - 1))
y_min = max(0, min(y_min, h - 1))
x_max = max(0, min(x_max, w))
y_max = max(0, min(y_max, h))
if x_max <= x_min or y_max <= y_min:
print(f"[WARN] Invalid box, skip: {image_name}, box=({x_min},{y_min},{x_max},{y_max})")
continue
stem = os.path.splitext(image_name)[0]
new_image_name = f"{split}_{stem}.png"
crop_image_name = f"{split}_{stem}_polyp.png"
new_image_path = os.path.join(output_image_dir, new_image_name)
crop_image_path = os.path.join(output_image_dir, crop_image_name)
try:
img.save(new_image_path, format="PNG")
except Exception as e:
print(f"[WARN] Failed to save original: {new_image_path}, error: {e}")
continue
crop = img.crop((x_min, y_min, x_max, y_max))
try:
crop.save(crop_image_path, format="PNG")
except Exception as e:
print(f"[WARN] Failed to save crop: {crop_image_path}, error: {e}")
continue
saved_count += 1
if (idx + 1) % 200 == 0 or (idx + 1) == total:
print(f"[CROP] Processed {idx + 1}/{total}")
print(f"[CROP] Saved {saved_count} samples")
return saved_count
def load_split_csv(csv_path):
split_map = {}
with open(csv_path, "r", encoding="utf-8") as f:
header = f.readline()
for line in f:
line = line.strip()
if not line:
continue
parts = line.split(",", 1)
if len(parts) != 2:
continue
split_name = parts[0].strip()
image_filename = parts[1].strip()
stem = os.path.splitext(image_filename)[0]
split_map[stem] = split_name
return split_map
def bbox_to_int(bbox):
return (
int(float(bbox["xmin"])),
int(float(bbox["ymin"])),
int(float(bbox["xmax"])),
int(float(bbox["ymax"])),
)
def is_valid_bbox(bbox):
required_keys = {"xmin", "ymin", "xmax", "ymax"}
if not isinstance(bbox, dict):
return False
return required_keys.issubset(set(bbox.keys()))
def bbox_area(bbox):
try:
x_min, y_min, x_max, y_max = bbox_to_int(bbox)
w = max(0, x_max - x_min)
h = max(0, y_max - y_min)
return w * h
except Exception:
return -1
def select_single_bbox(bboxes):
valid_bboxes = [b for b in bboxes if is_valid_bbox(b)]
if not valid_bboxes:
return None
return max(valid_bboxes, key=bbox_area)
def pipeline_dataset_a():
print("\n" + "=" * 60)
print("Dataset A: Kvasir-SEG Pipeline")
print("=" * 60)
print(f"Loading JSON: {DATASET_A_INPUT_JSON}")
with open(DATASET_A_INPUT_JSON, "r", encoding="utf-8") as f:
data = json.load(f)
print(f"Loading split CSV: {DATASET_A_SPLIT_CSV}")
split_map = load_split_csv(DATASET_A_SPLIT_CSV)
print(f"Split CSV entries: {len(split_map)}")
samples = []
missing_images = []
no_split_info = []
for image_id, info in data.items():
image_name = f"{image_id}{DATASET_A_IMAGE_SUFFIX}"
image_path = os.path.join(DATASET_A_IMAGE_DIR, image_name)
if not os.path.isfile(image_path):
missing_images.append(image_name)
continue
split_name = split_map.get(image_id)
if split_name is None:
no_split_info.append(image_id)
continue
bboxes = info.get("bbox", [])
if not isinstance(bboxes, list) or len(bboxes) == 0:
continue
selected_bbox = select_single_bbox(bboxes)
if selected_bbox is None:
continue
samples.append({
"image_id": image_id,
"image_name": image_name,
"image_path": image_path,
"split": split_name,
"bbox": selected_bbox,
})
print(f"Valid images: {len(samples)}")
print(f"Missing images: {len(missing_images)}")
print(f"No split info: {len(no_split_info)}")
if len(samples) == 0:
print("[ERROR] Dataset A: No valid samples found.")
return 0
bbox_rows = []
for sample in samples:
bb = sample["bbox"]
x_min, y_min, x_max, y_max = bbox_to_int(bb)
bbox_rows.append({
"split": sample["split"],
"image_name": sample["image_name"],
"image_path": sample["image_path"],
"x_min": x_min,
"y_min": y_min,
"x_max": x_max,
"y_max": y_max,
})
return crop_and_save(bbox_rows, OUTPUT_IMAGE_DIR)
def is_binary_mask(arr):
unique_vals = np.unique(arr)
return set(unique_vals.tolist()).issubset({0, 1})
def key_to_subdir(key):
parts = key.split("_")
if len(parts) >= 2 and parts[0] in {"train", "val", "valid", "validation", "test"}:
split = parts[0]
rest = "_".join(parts[1:])
return Path(split) / rest
return Path(key)
def extract_npz(npz_path, output_root):
npz_name = Path(npz_path).stem
npz_out_root = Path(output_root) / npz_name
ensure_dir(npz_out_root)
print(f"[NPZ] Loading: {npz_path}")
data = np.load(npz_path, allow_pickle=True)
print("[NPZ] Keys:")
for key in data.files:
arr = data[key]
print(f" - {key}: shape={arr.shape}, dtype={arr.dtype}")
for key in data.files:
arr = np.asarray(data[key])
subdir = key_to_subdir(key)
target_dir = npz_out_root / subdir
ensure_dir(target_dir)
print(f"[NPZ] Exporting key: {key}")
if is_binary_mask(arr):
print(f" [binary mask -> npy] unique={np.unique(arr)[:10].tolist()}")
_save_npz_batch_as_npy(arr, target_dir, key)
continue
if arr.ndim == 2:
save_image_from_array(arr, target_dir / "0.png")
continue
if arr.ndim == 3 and arr.shape[-1] in (1, 3, 4):
save_image_from_array(arr, target_dir / "0.png")
continue
if arr.ndim == 3:
for i in range(arr.shape[0]):
if is_binary_mask(arr[i]):
np.save(target_dir / f"{i:05d}.npy", arr[i])
else:
save_image_from_array(arr[i], target_dir / f"{i:05d}.png")
continue
if arr.ndim == 4 and arr.shape[-1] in (1, 3, 4):
for i in range(arr.shape[0]):
save_image_from_array(arr[i], target_dir / f"{i:05d}.png")
continue
np.save(target_dir / f"{key}.npy", arr)
return str(npz_out_root)
def _save_npz_batch_as_npy(arr, target_dir, key):
if arr.ndim == 2:
np.save(target_dir / "0.npy", arr)
return
if arr.ndim == 3 and arr.shape[-1] in (1, 3, 4):
np.save(target_dir / "0.npy", arr)
return
if arr.ndim >= 3:
for i in range(arr.shape[0]):
np.save(target_dir / f"{i:05d}.npy", arr[i])
return
np.save(target_dir / f"{key}.npy", arr)
def get_file_stem_to_path(folder, valid_exts=None):
mapping = {}
folder = Path(folder)
if not folder.exists():
print(f"[WARN] Folder does not exist: {folder}")
return mapping
for p in folder.iterdir():
if not p.is_file():
continue
if valid_exts is not None and p.suffix.lower() not in valid_exts:
continue
mapping[p.stem] = str(p)
return mapping
def mask_to_bbox(mask):
if mask.ndim != 2:
raise ValueError(f"Mask should be 2D, got shape={mask.shape}")
ys, xs = np.where(mask > 0)
if len(xs) == 0 or len(ys) == 0:
return None
return int(xs.min()), int(ys.min()), int(xs.max()), int(ys.max())
def mask_dir_to_bbox_rows(extracted_root, splits):
all_rows = []
for split in splits:
image_dir = os.path.join(extracted_root, split, "images")
label_dir = os.path.join(extracted_root, split, "label")
image_map = get_file_stem_to_path(image_dir, valid_exts=IMAGE_EXTS)
mask_map = get_file_stem_to_path(label_dir, valid_exts={".npy"})
print(f"[MASK->BBOX] Split: {split}")
print(f" Image dir: {image_dir} ({len(image_map)} files)")
print(f" Label dir: {label_dir} ({len(mask_map)} files)")
common_names = sorted(set(image_map.keys()) & set(mask_map.keys()))
print(f" Matched pairs: {len(common_names)}")
skipped = 0
for name in common_names:
image_path = image_map[name]
mask_path = mask_map[name]
try:
mask = np.load(mask_path)
if mask.ndim == 3 and mask.shape[-1] == 1:
mask = np.squeeze(mask, axis=-1)
bbox = mask_to_bbox(mask)
if bbox is None:
skipped += 1
continue
x_min, y_min, x_max, y_max = bbox
all_rows.append({
"split": split,
"image_name": os.path.basename(image_path),
"image_path": image_path,
"x_min": x_min,
"y_min": y_min,
"x_max": x_max,
"y_max": y_max,
})
except Exception as e:
skipped += 1
print(f" [ERROR] {name}: {e}")
print(f" Valid: {len(all_rows)}, Skipped: {skipped}")
return all_rows
def find_npz_file(download_root):
for root, dirs, files in os.walk(download_root):
for f in files:
if f.endswith(".npz"):
return os.path.join(root, f)
return None
def pipeline_dataset_b():
print("\n" + "=" * 60)
print("Dataset B: PolypGenMSBench Pipeline")
print("=" * 60)
print("[DOWNLOAD] Downloading PolypGenMSBench...")
try:
from medsegbench import PolypGenMSBench
for split_name in ["train", "test"]:
print(f" Downloading split: {split_name}")
_ = PolypGenMSBench(
root=DATASET_B_DOWNLOAD_ROOT,
split=split_name,
download=True,
size=DATASET_B_DOWNLOAD_SIZE,
)
print("[DOWNLOAD] Done.")
except ImportError:
print("[WARN] medsegbench not installed. Skipping download.")
print(" Please install: pip install medsegbench")
except Exception as e:
print(f"[WARN] Download error: {e}")
npz_path = DATASET_B_NPZ_PATH
if npz_path is None:
print("[NPZ] Searching for NPZ file...")
npz_path = find_npz_file(DATASET_B_DOWNLOAD_ROOT)
if npz_path is None:
print("[ERROR] No NPZ file found. Dataset B pipeline aborted.")
return 0
print(f"[NPZ] Found: {npz_path}")
extracted_root = extract_npz(npz_path, DATASET_B_NPZ_EXTRACT_ROOT)
bbox_rows = mask_dir_to_bbox_rows(extracted_root, DATASET_B_SPLITS)
if len(bbox_rows) == 0:
print("[ERROR] Dataset B: No valid bbox rows generated.")
return 0
return crop_and_save(bbox_rows, OUTPUT_IMAGE_DIR)
def main():
print("=" * 60)
print("Integrated Pipeline: Multi-Dataset Processing")
print("=" * 60)
print(f"Output image dir: {OUTPUT_IMAGE_DIR}")
ensure_dir(OUTPUT_IMAGE_DIR)
count_a = pipeline_dataset_a()
count_b = pipeline_dataset_b()
print("\n" + "=" * 60)
print("Summary")
print("=" * 60)
print(f"Dataset A (Kvasir-SEG): {count_a} samples")
print(f"Dataset B (PolypGenMSBench): {count_b} samples")
print(f"Total: {count_a + count_b} samples")
print(f"All images saved to: {OUTPUT_IMAGE_DIR}")
print("Done.")
if __name__ == "__main__":
main()