File size: 8,655 Bytes
23a6ff8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""

独立单流程处理脚本:处理指定目录(如 train/val/test)下的图片和路网文件

不依赖 json 分割文件,直接把处理好的 A 和 AB 结果归置到对应 _patches 文件夹。

同时,保证 mask 也能正确无误地提取并保存下来。

"""

import os
import re
import sys
import time
import shutil
import argparse
import subprocess
from typing import Dict, List, Tuple, Optional

IMG_EXTS = (".png", ".jpg", ".jpeg", ".tif", ".tiff", ".bmp")

def is_image(path: str) -> bool:
    ext = os.path.splitext(path)[1].lower()
    return ext in IMG_EXTS

def find_pairs(split_dir: str) -> List[Tuple[str, str, str]]:
    files = [f for f in os.listdir(split_dir) if os.path.isfile(os.path.join(split_dir, f))]
    images_by_id: Dict[str, str] = {}
    graphs_by_id: Dict[str, str] = {}

    pat = re.compile(r"^(data\d+)")

    for fname in files:
        m = pat.match(fname)
        if not m:
            continue
        data_id = m.group(1)
        full = os.path.join(split_dir, fname)
        if is_image(full):
            images_by_id.setdefault(data_id, full)
        elif fname.lower().endswith(".pickle"):
            prev = graphs_by_id.get(data_id)
            # 优先保留带有 _gt_graph 的后缀
            if prev is None or fname.endswith("_gt_graph.pickle"):
                graphs_by_id[data_id] = full

    pairs: List[Tuple[str, str, str]] = []
    for data_id, img in images_by_id.items():
        g = graphs_by_id.get(data_id)
        if g:
            pairs.append((data_id, img, g))

    def id_key(t: Tuple[str, str, str]) -> int:
        m = re.search(r"data(\d+)", t[0])
        return int(m.group(1)) if m else 0
    pairs.sort(key=id_key)
    return pairs

def run_cmd(cmd: List[str], cwd: Optional[str] = None) -> int:
    print("[RUN]", " ".join(cmd))
    try:
        res = subprocess.run(cmd, cwd=cwd, check=False)
        return res.returncode
    except Exception as e:
        print(f"[ERR] Failed to run: {' '.join(cmd)} | {e}")
        return 1

def collect_pairs_to_output(

    processed_root: str,

    output_dir: str,

    only_A: bool = False

) -> int:
    os.makedirs(output_dir, exist_ok=True)
    pat = re.compile(r"^data(\d+)_tiled_candidates$")
    folder_map: Dict[int, str] = {}

    for name in os.listdir(processed_root):
        full = os.path.join(processed_root, name)
        if not os.path.isdir(full):
            continue
        m = pat.match(name)
        if m:
            idx = int(m.group(1))
            folder_map[idx] = full

    global_idx = 0
    # 遍历所有被处理过的 data{i}
    for data_idx in sorted(folder_map.keys()):
        data_dir = folder_map[data_idx]
        
        # 要收集的子文件夹:只收集A 或者 收集A和B
        dirs_to_check = ["A"] if only_A else ["A", "B"]
        
        for d in dirs_to_check:
            sub_dir = os.path.join(data_dir, d)
            if not os.path.isdir(sub_dir):
                continue
            # 寻找 graph.pickle 作为锚点,提取 prefix
            for fname in os.listdir(sub_dir):
                if fname.endswith("_graph.pickle"):
                    prefix = fname[:-len("_graph.pickle")]
                    rgb_name = prefix + "_rgb.png"
                    
                    graph_path = os.path.join(sub_dir, fname)
                    rgb_path = os.path.join(sub_dir, rgb_name)
                    
                    if os.path.exists(rgb_path):
                        # 目标文件命名
                        dst_rgb = os.path.join(output_dir, f"data_{global_idx}.png")
                        dst_graph = os.path.join(output_dir, f"gt_graph_{global_idx}.pickle")
                        
                        shutil.copy2(rgb_path, dst_rgb)
                        shutil.copy2(graph_path, dst_graph)
                        global_idx += 1
    return global_idx

def main() -> int:
    parser = argparse.ArgumentParser(description="按序处理单分片目录(比如 test 或 val),不需要 split json 文件")
    parser.add_argument("split_dir", help="需要处理的目录路径 (例如 test)")
    parser.add_argument("--patch_size", type=int, default=1024)
    parser.add_argument("--overlaps", type=int, nargs="*", default=[256, 384])
    parser.add_argument("--edge_width", type=int, default=6)
    parser.add_argument("--inner_offset", type=int, default=5)
    parser.add_argument("--min_edge_ratio", type=float, default=0.05)
    parser.add_argument("--a_density_min", type=float, default=0.001)
    parser.add_argument("--b_density_min", type=float, default=0.001)
    parser.add_argument("--wl_iterations", type=int, default=3)
    parser.add_argument("--sim_threshold", type=float, default=0.7)
    parser.add_argument("--seed", type=int, default=2025)
    parser.add_argument("--workers", type=int, default=4, help="用于并行裁剪的线程数")
    
    args = parser.parse_args()

    split_dir = os.path.abspath(args.split_dir)
    split_name = os.path.basename(split_dir)
    project_root = os.path.dirname(split_dir)
    
    script_dir = os.path.dirname(os.path.abspath(__file__))
    
    # 定义输出路径
    processed_dir = os.path.join(project_root, f"{split_name}_processed")
    patches_dir = os.path.join(project_root, f"{split_name}_patches")
    
    if not os.path.isdir(split_dir):
        print(f"[ERR] 找不到指定的目录: {split_dir}")
        return 1
        
    os.makedirs(processed_dir, exist_ok=True)
    os.makedirs(patches_dir, exist_ok=True)

    pairs = find_pairs(split_dir)
    if not pairs:
        print(f"[WARN] 在目录中未找到任何 (图像, 路网) 文件对: {split_dir}")
        return 0
        
    print(f"找到 {len(pairs)} 对数据在目录 {split_dir} 中,准备进行处理。")
    
    # 强制使用并行的tiler,保障速度与xy的正确映射
    tiler = os.path.join(script_dir, "tile_and_crop_patches_parallel.py")
    selector = os.path.join(script_dir, "select_by_wl_similarity.py")

    for data_id, img_path, graph_path in pairs:
        out_dir = os.path.join(processed_dir, f"{data_id}_tiled_candidates")
        os.makedirs(out_dir, exist_ok=True)
        
        print(f"\n--- 正在处理 {data_id} ---")
        
        # 第一步:多线程切分和初步过滤
        tile_cmd = [
            sys.executable, tiler,
            img_path, graph_path,
            "--output", out_dir,
            "--patch_size", str(args.patch_size),
            "--overlaps", *[str(m) for m in args.overlaps],
            "--edge_width", str(args.edge_width),
            "--inner_offset", str(args.inner_offset),
            "--min_edge_ratio", str(args.min_edge_ratio),
            "--workers", str(args.workers)
        ]
        
        t0 = time.time()
        if run_cmd(tile_cmd, cwd=script_dir) != 0:
            print(f"[ERR] 对 {data_id} 执行并行裁剪失败")
            continue
        print(f"裁剪耗时 {time.time() - t0:.2f}s")
        
        # 第二步:基于 WL 相似度的精选
        sel_cmd = [
            sys.executable, selector,
            out_dir,
            "--a_density_min", str(args.a_density_min),
            "--b_density_min", str(args.b_density_min),
            "--wl_iterations", str(args.wl_iterations),
            "--sim_threshold", str(args.sim_threshold),
            "--seed", str(args.seed),
            "--debug_max", "0"
        ]
        
        t1 = time.time()
        if run_cmd(sel_cmd, cwd=script_dir) != 0:
            print(f"[ERR] 对 {data_id} 进行 WL 相似度筛选失败")
            continue
        print(f"筛选耗时 {time.time() - t1:.2f}s")

    # 第三步:收集结果整理到最终的补丁文件夹里(同时复制 RGB,MASK 和 GRAPH)
    print(f"\n--- 正在将处理结果收集归类为 {split_name}_A 和 {split_name}_AB ---")
    
    out_A = os.path.join(patches_dir, f"{split_name}_A")
    count_A = collect_pairs_to_output(processed_dir, out_A, only_A=True)
    print(f"已收集纯 A 划分数据: {count_A} 对 -> {out_A}")
    
    out_AB = os.path.join(patches_dir, f"{split_name}_AB")
    count_AB = collect_pairs_to_output(processed_dir, out_AB, only_A=False)
    print(f"已收集包含重叠块(A+B)划分数据: {count_AB} 对 -> {out_AB}")
    
    print("\n✓ 全部处理完成!")
    return 0

if __name__ == "__main__":
    raise SystemExit(main())