WildRoad / script /process_single_split.py
guanxiaofei's picture
Added scripts to reproduce the processed dataset from raw source data.
23a6ff8 verified
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
独立单流程处理脚本:处理指定目录(如 train/val/test)下的图片和路网文件
不依赖 json 分割文件,直接把处理好的 A 和 AB 结果归置到对应 _patches 文件夹。
同时,保证 mask 也能正确无误地提取并保存下来。
"""
import os
import re
import sys
import time
import shutil
import argparse
import subprocess
from typing import Dict, List, Tuple, Optional
IMG_EXTS = (".png", ".jpg", ".jpeg", ".tif", ".tiff", ".bmp")
def is_image(path: str) -> bool:
ext = os.path.splitext(path)[1].lower()
return ext in IMG_EXTS
def find_pairs(split_dir: str) -> List[Tuple[str, str, str]]:
files = [f for f in os.listdir(split_dir) if os.path.isfile(os.path.join(split_dir, f))]
images_by_id: Dict[str, str] = {}
graphs_by_id: Dict[str, str] = {}
pat = re.compile(r"^(data\d+)")
for fname in files:
m = pat.match(fname)
if not m:
continue
data_id = m.group(1)
full = os.path.join(split_dir, fname)
if is_image(full):
images_by_id.setdefault(data_id, full)
elif fname.lower().endswith(".pickle"):
prev = graphs_by_id.get(data_id)
# 优先保留带有 _gt_graph 的后缀
if prev is None or fname.endswith("_gt_graph.pickle"):
graphs_by_id[data_id] = full
pairs: List[Tuple[str, str, str]] = []
for data_id, img in images_by_id.items():
g = graphs_by_id.get(data_id)
if g:
pairs.append((data_id, img, g))
def id_key(t: Tuple[str, str, str]) -> int:
m = re.search(r"data(\d+)", t[0])
return int(m.group(1)) if m else 0
pairs.sort(key=id_key)
return pairs
def run_cmd(cmd: List[str], cwd: Optional[str] = None) -> int:
print("[RUN]", " ".join(cmd))
try:
res = subprocess.run(cmd, cwd=cwd, check=False)
return res.returncode
except Exception as e:
print(f"[ERR] Failed to run: {' '.join(cmd)} | {e}")
return 1
def collect_pairs_to_output(
processed_root: str,
output_dir: str,
only_A: bool = False
) -> int:
os.makedirs(output_dir, exist_ok=True)
pat = re.compile(r"^data(\d+)_tiled_candidates$")
folder_map: Dict[int, str] = {}
for name in os.listdir(processed_root):
full = os.path.join(processed_root, name)
if not os.path.isdir(full):
continue
m = pat.match(name)
if m:
idx = int(m.group(1))
folder_map[idx] = full
global_idx = 0
# 遍历所有被处理过的 data{i}
for data_idx in sorted(folder_map.keys()):
data_dir = folder_map[data_idx]
# 要收集的子文件夹:只收集A 或者 收集A和B
dirs_to_check = ["A"] if only_A else ["A", "B"]
for d in dirs_to_check:
sub_dir = os.path.join(data_dir, d)
if not os.path.isdir(sub_dir):
continue
# 寻找 graph.pickle 作为锚点,提取 prefix
for fname in os.listdir(sub_dir):
if fname.endswith("_graph.pickle"):
prefix = fname[:-len("_graph.pickle")]
rgb_name = prefix + "_rgb.png"
graph_path = os.path.join(sub_dir, fname)
rgb_path = os.path.join(sub_dir, rgb_name)
if os.path.exists(rgb_path):
# 目标文件命名
dst_rgb = os.path.join(output_dir, f"data_{global_idx}.png")
dst_graph = os.path.join(output_dir, f"gt_graph_{global_idx}.pickle")
shutil.copy2(rgb_path, dst_rgb)
shutil.copy2(graph_path, dst_graph)
global_idx += 1
return global_idx
def main() -> int:
parser = argparse.ArgumentParser(description="按序处理单分片目录(比如 test 或 val),不需要 split json 文件")
parser.add_argument("split_dir", help="需要处理的目录路径 (例如 test)")
parser.add_argument("--patch_size", type=int, default=1024)
parser.add_argument("--overlaps", type=int, nargs="*", default=[256, 384])
parser.add_argument("--edge_width", type=int, default=6)
parser.add_argument("--inner_offset", type=int, default=5)
parser.add_argument("--min_edge_ratio", type=float, default=0.05)
parser.add_argument("--a_density_min", type=float, default=0.001)
parser.add_argument("--b_density_min", type=float, default=0.001)
parser.add_argument("--wl_iterations", type=int, default=3)
parser.add_argument("--sim_threshold", type=float, default=0.7)
parser.add_argument("--seed", type=int, default=2025)
parser.add_argument("--workers", type=int, default=4, help="用于并行裁剪的线程数")
args = parser.parse_args()
split_dir = os.path.abspath(args.split_dir)
split_name = os.path.basename(split_dir)
project_root = os.path.dirname(split_dir)
script_dir = os.path.dirname(os.path.abspath(__file__))
# 定义输出路径
processed_dir = os.path.join(project_root, f"{split_name}_processed")
patches_dir = os.path.join(project_root, f"{split_name}_patches")
if not os.path.isdir(split_dir):
print(f"[ERR] 找不到指定的目录: {split_dir}")
return 1
os.makedirs(processed_dir, exist_ok=True)
os.makedirs(patches_dir, exist_ok=True)
pairs = find_pairs(split_dir)
if not pairs:
print(f"[WARN] 在目录中未找到任何 (图像, 路网) 文件对: {split_dir}")
return 0
print(f"找到 {len(pairs)} 对数据在目录 {split_dir} 中,准备进行处理。")
# 强制使用并行的tiler,保障速度与xy的正确映射
tiler = os.path.join(script_dir, "tile_and_crop_patches_parallel.py")
selector = os.path.join(script_dir, "select_by_wl_similarity.py")
for data_id, img_path, graph_path in pairs:
out_dir = os.path.join(processed_dir, f"{data_id}_tiled_candidates")
os.makedirs(out_dir, exist_ok=True)
print(f"\n--- 正在处理 {data_id} ---")
# 第一步:多线程切分和初步过滤
tile_cmd = [
sys.executable, tiler,
img_path, graph_path,
"--output", out_dir,
"--patch_size", str(args.patch_size),
"--overlaps", *[str(m) for m in args.overlaps],
"--edge_width", str(args.edge_width),
"--inner_offset", str(args.inner_offset),
"--min_edge_ratio", str(args.min_edge_ratio),
"--workers", str(args.workers)
]
t0 = time.time()
if run_cmd(tile_cmd, cwd=script_dir) != 0:
print(f"[ERR] 对 {data_id} 执行并行裁剪失败")
continue
print(f"裁剪耗时 {time.time() - t0:.2f}s")
# 第二步:基于 WL 相似度的精选
sel_cmd = [
sys.executable, selector,
out_dir,
"--a_density_min", str(args.a_density_min),
"--b_density_min", str(args.b_density_min),
"--wl_iterations", str(args.wl_iterations),
"--sim_threshold", str(args.sim_threshold),
"--seed", str(args.seed),
"--debug_max", "0"
]
t1 = time.time()
if run_cmd(sel_cmd, cwd=script_dir) != 0:
print(f"[ERR] 对 {data_id} 进行 WL 相似度筛选失败")
continue
print(f"筛选耗时 {time.time() - t1:.2f}s")
# 第三步:收集结果整理到最终的补丁文件夹里(同时复制 RGB,MASK 和 GRAPH)
print(f"\n--- 正在将处理结果收集归类为 {split_name}_A 和 {split_name}_AB ---")
out_A = os.path.join(patches_dir, f"{split_name}_A")
count_A = collect_pairs_to_output(processed_dir, out_A, only_A=True)
print(f"已收集纯 A 划分数据: {count_A} 对 -> {out_A}")
out_AB = os.path.join(patches_dir, f"{split_name}_AB")
count_AB = collect_pairs_to_output(processed_dir, out_AB, only_A=False)
print(f"已收集包含重叠块(A+B)划分数据: {count_AB} 对 -> {out_AB}")
print("\n✓ 全部处理完成!")
return 0
if __name__ == "__main__":
raise SystemExit(main())