|
|
import os |
|
|
import random |
|
|
import shutil |
|
|
from pathlib import Path |
|
|
import json |
|
|
|
|
|
import numpy as np |
|
|
|
|
|
SEED = 42 |
|
|
random.seed(SEED) |
|
|
np.random.seed(SEED) |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
root_indir = Path("./raw-datasets/DigitizePID_Dataset") |
|
|
imgs_indir = root_indir / "image_2" |
|
|
imgs_in = os.listdir(imgs_indir) |
|
|
|
|
|
root_outdir = Path("./processed-datasets/DigitizePID_Dataset") |
|
|
|
|
|
|
|
|
for split in ("train", "val"): |
|
|
(root_outdir / split).mkdir(parents=True, exist_ok=True) |
|
|
|
|
|
|
|
|
imgs_in = os.listdir(imgs_indir) |
|
|
random.shuffle(imgs_in) |
|
|
|
|
|
n = len(imgs_in) |
|
|
train_end = int(0.8 * n) |
|
|
|
|
|
splits = ( |
|
|
("train", imgs_in[:train_end]), |
|
|
("val", imgs_in[train_end:]), |
|
|
) |
|
|
|
|
|
for split, files in splits: |
|
|
metadata_lines = [] |
|
|
for img_fname in files: |
|
|
idx = int(Path(img_fname).stem) |
|
|
shutil.copy(imgs_indir / img_fname, root_outdir / split / img_fname) |
|
|
|
|
|
symbols = np.load( |
|
|
root_indir / str(idx) / f"{idx}_symbols.npy", allow_pickle=True |
|
|
) |
|
|
|
|
|
metadata_lines.append({ |
|
|
"file_name": img_fname, |
|
|
"symbols": { |
|
|
"bbox": [[int(n) for n in symbol[1]] for symbol in symbols], |
|
|
"labels": [int(symbol[2]) for symbol in symbols], |
|
|
}, |
|
|
}) |
|
|
|
|
|
with open(root_outdir / split / "metadata.jsonl", "w") as f: |
|
|
for line in metadata_lines: |
|
|
f.write(json.dumps(line) + "\n") |
|
|
|