|
|
import json |
|
|
from .basemodel import TextDetBase, TextDetBaseDNN |
|
|
import os.path as osp |
|
|
from tqdm import tqdm |
|
|
import numpy as np |
|
|
import cv2 |
|
|
import torch |
|
|
from pathlib import Path |
|
|
import einops |
|
|
|
|
|
from utils.io_utils import find_all_imgs, NumpyEncoder |
|
|
from utils.imgproc_utils import letterbox, xyxy2yolo, get_yololabel_strings, square_pad_resize |
|
|
|
|
|
from modules.textdetector.yolov5.yolov5_utils import non_max_suppression |
|
|
from modules.textdetector.db_utils import SegDetectorRepresenter |
|
|
from utils.textblock import TextBlock, group_output, mit_merge_textlines |
|
|
from modules.textdetector.ctd.textmask import refine_mask, refine_undetected_mask, REFINEMASK_INPAINT, REFINEMASK_ANNOTATION |
|
|
from pathlib import Path |
|
|
from typing import Union, List, Tuple, Callable |
|
|
|
|
|
CTD_MODEL_PATH = r'data/models/comictextdetector.pt' |
|
|
|
|
|
def det_rearrange_forward( |
|
|
img: np.ndarray, |
|
|
dbnet_batch_forward: Callable[[np.ndarray, str], Tuple[np.ndarray, np.ndarray]], |
|
|
tgt_size: int = 1280, |
|
|
max_batch_size: int = 4, |
|
|
device='cuda', |
|
|
crop_as_square=False, verbose=False): |
|
|
''' |
|
|
Rearrange image to square batches before feeding into network if following conditions are satisfied: \n |
|
|
1. Extreme aspect ratio |
|
|
2. Is too tall or wide for detect size (tgt_size) |
|
|
|
|
|
Returns: |
|
|
DBNet output, mask or None, None if rearrangement is not required |
|
|
''' |
|
|
|
|
|
def _unrearrange(patch_lst: List[np.ndarray], transpose: bool, channel=1, pad_num=0): |
|
|
_psize = _h = patch_lst[0].shape[-1] |
|
|
_step = int(ph_step * _psize / patch_size) |
|
|
_pw = int(_psize / pw_num) |
|
|
_h = int(_pw / w * h) |
|
|
tgtmap = np.zeros((channel, _h, _pw), dtype=np.float32) |
|
|
num_patches = len(patch_lst) * pw_num - pad_num |
|
|
for ii, p in enumerate(patch_lst): |
|
|
if transpose: |
|
|
p = einops.rearrange(p, 'c h w -> c w h') |
|
|
for jj in range(pw_num): |
|
|
pidx = ii * pw_num + jj |
|
|
rel_t = rel_step_list[pidx] |
|
|
t = int(round(rel_t * _h)) |
|
|
b = min(t + _psize, _h) |
|
|
l = jj * _pw |
|
|
r = l + _pw |
|
|
tgtmap[..., t: b, :] += p[..., : b - t, l: r] |
|
|
if pidx > 0: |
|
|
interleave = _psize - _step |
|
|
tgtmap[..., t: t+interleave, :] /= 2. |
|
|
|
|
|
if pidx >= num_patches - 1: |
|
|
break |
|
|
|
|
|
if transpose: |
|
|
tgtmap = einops.rearrange(tgtmap, 'c h w -> c w h') |
|
|
return tgtmap[None, ...] |
|
|
|
|
|
def _patch2batches(patch_lst: List[np.ndarray], p_num: int, transpose: bool): |
|
|
if transpose: |
|
|
patch_lst = einops.rearrange(patch_lst, '(p_num pw_num) ph pw c -> p_num (pw_num pw) ph c', p_num=p_num) |
|
|
else: |
|
|
patch_lst = einops.rearrange(patch_lst, '(p_num pw_num) ph pw c -> p_num ph (pw_num pw) c', p_num=p_num) |
|
|
|
|
|
batches = [[]] |
|
|
for ii, patch in enumerate(patch_lst): |
|
|
|
|
|
if len(batches[-1]) >= max_batch_size: |
|
|
batches.append([]) |
|
|
p, down_scale_ratio, pad_h, pad_w = square_pad_resize(patch, tgt_size=tgt_size) |
|
|
|
|
|
assert pad_h == pad_w |
|
|
pad_size = pad_h |
|
|
batches[-1].append(p) |
|
|
if verbose: |
|
|
cv2.imwrite(f'result/rearrange_{ii}.png', p[..., ::-1]) |
|
|
return batches, down_scale_ratio, pad_size |
|
|
|
|
|
h, w = img.shape[:2] |
|
|
transpose = False |
|
|
if h < w: |
|
|
transpose = True |
|
|
h, w = img.shape[1], img.shape[0] |
|
|
|
|
|
asp_ratio = h / w |
|
|
down_scale_ratio = h / tgt_size |
|
|
|
|
|
|
|
|
require_rearrange = down_scale_ratio > 2.5 and asp_ratio > 3 |
|
|
if not require_rearrange: |
|
|
return None, None |
|
|
|
|
|
if verbose: |
|
|
print(f'Input image will be rearranged to square batches before fed into network.\ |
|
|
\n Rearranged batches will be saved to result/rearrange_%d.png') |
|
|
|
|
|
if transpose: |
|
|
img = einops.rearrange(img, 'h w c -> w h c') |
|
|
|
|
|
if crop_as_square: |
|
|
pw_num = 1 |
|
|
else: |
|
|
pw_num = max(int(np.floor(2 * tgt_size / w)), 2) |
|
|
patch_size = ph = pw_num * w |
|
|
|
|
|
ph_num = int(np.ceil(h / ph)) |
|
|
ph_step = int((h - ph) / (ph_num - 1)) if ph_num > 1 else 0 |
|
|
rel_step_list = [] |
|
|
patch_list = [] |
|
|
for ii in range(ph_num): |
|
|
t = ii * ph_step |
|
|
b = t + ph |
|
|
rel_step_list.append(t / h) |
|
|
patch_list.append(img[t: b]) |
|
|
|
|
|
p_num = int(np.ceil(ph_num / pw_num)) |
|
|
pad_num = p_num * pw_num - ph_num |
|
|
for ii in range(pad_num): |
|
|
patch_list.append(np.zeros_like(patch_list[0])) |
|
|
|
|
|
batches, down_scale_ratio, pad_size = _patch2batches(patch_list, p_num, transpose) |
|
|
|
|
|
db_lst, mask_lst = [], [] |
|
|
for batch in batches: |
|
|
batch = np.array(batch) |
|
|
db, mask = dbnet_batch_forward(batch, device=device) |
|
|
|
|
|
for ii, (d, m) in enumerate(zip(db, mask)): |
|
|
if pad_size > 0: |
|
|
paddb = int(db.shape[-1] / tgt_size * pad_size) |
|
|
padmsk = int(mask.shape[-1] / tgt_size * pad_size) |
|
|
d = d[..., :-paddb, :-paddb] |
|
|
m = m[..., :-padmsk, :-padmsk] |
|
|
db_lst.append(d) |
|
|
mask_lst.append(m) |
|
|
if verbose: |
|
|
cv2.imwrite(f'result/rearrange_db_{ii}.png', (d[0] * 255).astype(np.uint8)) |
|
|
cv2.imwrite(f'result/rearrange_thr_{ii}.png', (d[1] * 255).astype(np.uint8)) |
|
|
|
|
|
db = _unrearrange(db_lst, transpose, channel=2, pad_num=pad_num) |
|
|
mask = _unrearrange(mask_lst, transpose, channel=1, pad_num=pad_num) |
|
|
return db, mask |
|
|
|
|
|
def model2annotations(model_path, img_dir_list, save_dir, save_json=False): |
|
|
if isinstance(img_dir_list, str): |
|
|
img_dir_list = [img_dir_list] |
|
|
cuda = torch.cuda.is_available() |
|
|
device = 'cuda' if cuda else 'cpu' |
|
|
model = TextDetector(model_path=model_path, detect_size=1024, device=device, act='leaky') |
|
|
imglist = [] |
|
|
for img_dir in img_dir_list: |
|
|
imglist += find_all_imgs(img_dir, abs_path=True) |
|
|
for img_path in tqdm(imglist): |
|
|
imgname = osp.basename(img_path) |
|
|
img = cv2.imread(img_path) |
|
|
im_h, im_w = img.shape[:2] |
|
|
imname = imgname.replace(Path(imgname).suffix, '') |
|
|
maskname = 'mask-'+imname+'.png' |
|
|
poly_save_path = osp.join(save_dir, 'line-' + imname + '.txt') |
|
|
mask, mask_refined, blk_list = model(img, refine_mode=REFINEMASK_ANNOTATION, keep_undetected_mask=True) |
|
|
polys = [] |
|
|
blk_xyxy = [] |
|
|
blk_dict_list = [] |
|
|
for blk in blk_list: |
|
|
polys += blk.lines |
|
|
blk_xyxy.append(blk.xyxy) |
|
|
blk_dict_list.append(blk.to_dict()) |
|
|
blk_xyxy = xyxy2yolo(blk_xyxy, im_w, im_h) |
|
|
if blk_xyxy is not None: |
|
|
cls_list = [1] * len(blk_xyxy) |
|
|
yolo_label = get_yololabel_strings(cls_list, blk_xyxy) |
|
|
else: |
|
|
yolo_label = '' |
|
|
with open(osp.join(save_dir, imname+'.txt'), 'w', encoding='utf8') as f: |
|
|
f.write(yolo_label) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
if len(polys) != 0: |
|
|
if isinstance(polys, list): |
|
|
polys = np.array(polys) |
|
|
polys = polys.reshape(-1, 8) |
|
|
np.savetxt(poly_save_path, polys, fmt='%d') |
|
|
if save_json: |
|
|
with open(osp.join(save_dir, imname+'.json'), 'w', encoding='utf8') as f: |
|
|
f.write(json.dumps(blk_dict_list, ensure_ascii=False, cls=NumpyEncoder)) |
|
|
cv2.imwrite(osp.join(save_dir, imgname), img) |
|
|
cv2.imwrite(osp.join(save_dir, maskname), mask_refined) |
|
|
|
|
|
def preprocess_img(img, detect_size=(1024, 1024), device='cpu', bgr2rgb=True, half=False, to_tensor=True): |
|
|
if isinstance(detect_size, int): |
|
|
detect_size = (detect_size, detect_size) |
|
|
|
|
|
if bgr2rgb: |
|
|
img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB) |
|
|
img_in, ratio, (dw, dh) = letterbox(img, new_shape=detect_size, auto=False, stride=64) |
|
|
if to_tensor: |
|
|
img_in = img_in.transpose((2, 0, 1))[::-1] |
|
|
img_in = np.array([np.ascontiguousarray(img_in)]).astype(np.float32) / 255 |
|
|
if to_tensor: |
|
|
img_in = torch.from_numpy(img_in).to(device) |
|
|
if half: |
|
|
img_in = img_in.half() |
|
|
return img_in, ratio, int(dw), int(dh) |
|
|
|
|
|
def postprocess_mask(img: Union[torch.Tensor, np.ndarray], thresh=None): |
|
|
|
|
|
if isinstance(img, torch.Tensor): |
|
|
img = img.squeeze_() |
|
|
if img.device != 'cpu': |
|
|
img = img.detach().cpu() |
|
|
img = img.numpy() |
|
|
else: |
|
|
img = img.squeeze() |
|
|
if thresh is not None: |
|
|
img = img > thresh |
|
|
img = img * 255 |
|
|
|
|
|
|
|
|
return img.astype(np.uint8) |
|
|
|
|
|
def postprocess_yolo(det, conf_thresh, nms_thresh, resize_ratio, sort_func=None): |
|
|
det = non_max_suppression(det, conf_thresh, nms_thresh)[0] |
|
|
|
|
|
if det.device != 'cpu': |
|
|
det = det.detach_().cpu().numpy() |
|
|
det[..., [0, 2]] = det[..., [0, 2]] * resize_ratio[0] |
|
|
det[..., [1, 3]] = det[..., [1, 3]] * resize_ratio[1] |
|
|
if sort_func is not None: |
|
|
det = sort_func(det) |
|
|
|
|
|
blines = det[..., 0:4].astype(np.int32) |
|
|
confs = np.round(det[..., 4], 3) |
|
|
cls = det[..., 5].astype(np.int32) |
|
|
return blines, cls, confs |
|
|
|
|
|
class TextDetector: |
|
|
lang_list = ['eng', 'ja', 'unknown'] |
|
|
langcls2idx = {'eng': 0, 'ja': 1, 'unknown': 2} |
|
|
|
|
|
def __init__(self, model_path, detect_size=1024, device='cpu', half=False, nms_thresh=0.35, conf_thresh=0.4, det_rearrange_max_batches=4): |
|
|
super(TextDetector, self).__init__() |
|
|
|
|
|
self.net: Union[TextDetBase, TextDetBaseDNN] = None |
|
|
self.backend: str = None |
|
|
|
|
|
self.detect_size = detect_size |
|
|
self.device = device |
|
|
self.half = half |
|
|
self.conf_thresh = conf_thresh |
|
|
self.nms_thresh = nms_thresh |
|
|
self.seg_rep = SegDetectorRepresenter(thresh=0.3) |
|
|
|
|
|
self.backend = 'torch' |
|
|
self.load_model(model_path) |
|
|
|
|
|
self.det_rearrange_max_batches = det_rearrange_max_batches |
|
|
|
|
|
def load_model(self, model_path: str): |
|
|
if Path(model_path).suffix == '.onnx': |
|
|
self.net = TextDetBaseDNN(1024, model_path) |
|
|
self.backend = 'opencv' |
|
|
else: |
|
|
self.net = TextDetBase(model_path, device=self.device, act='leaky', half=self.half) |
|
|
self.backend = 'torch' |
|
|
|
|
|
def set_device(self, device: str): |
|
|
if self.device == device: |
|
|
return |
|
|
model_path = CTD_MODEL_PATH+'.onnx' if device == 'cpu' else CTD_MODEL_PATH |
|
|
if not osp.exists(model_path): |
|
|
raise FileNotFoundError(f'CTD model not found: {model_path}') |
|
|
self.load_model(model_path) |
|
|
|
|
|
def det_batch_forward_ctd(self, batch: np.ndarray, device: str) -> Tuple[np.ndarray, np.ndarray]: |
|
|
|
|
|
if isinstance(self.net, TextDetBase): |
|
|
batch = einops.rearrange(batch.astype(np.float32) / 255., 'n h w c -> n c h w') |
|
|
batch = torch.from_numpy(batch).to(device) |
|
|
_, mask, lines = self.net(batch) |
|
|
mask = mask.cpu().numpy() |
|
|
lines = lines.cpu().numpy() |
|
|
elif isinstance(self.net, TextDetBaseDNN): |
|
|
mask_lst, line_lst = [], [] |
|
|
for b in batch: |
|
|
_, mask, lines = self.net(b) |
|
|
if mask.shape[1] == 2: |
|
|
tmp = mask |
|
|
mask = lines |
|
|
lines = tmp |
|
|
mask_lst.append(mask) |
|
|
line_lst.append(lines) |
|
|
lines, mask = np.concatenate(line_lst, 0), np.concatenate(mask_lst, 0) |
|
|
else: |
|
|
raise NotImplementedError |
|
|
return lines, mask |
|
|
|
|
|
@torch.no_grad() |
|
|
def __call__(self, img, refine_mode=REFINEMASK_INPAINT, keep_undetected_mask=False) -> Tuple[np.ndarray, np.ndarray, List[TextBlock]]: |
|
|
|
|
|
detect_size = self.detect_size if not self.backend == 'opencv' else 1024 |
|
|
im_h, im_w = img.shape[:2] |
|
|
lines_map, mask = det_rearrange_forward(img, self.det_batch_forward_ctd, detect_size, self.det_rearrange_max_batches, self.device) |
|
|
blks = [] |
|
|
resize_ratio = [1, 1] |
|
|
if lines_map is None: |
|
|
img_in, ratio, dw, dh = preprocess_img(img, bgr2rgb=False, detect_size=detect_size, device=self.device, half=self.half, to_tensor=self.backend=='torch') |
|
|
blks, mask, lines_map = self.net(img_in) |
|
|
if self.backend == 'opencv': |
|
|
if mask.shape[1] == 2: |
|
|
tmp = mask |
|
|
mask = lines_map |
|
|
lines_map = tmp |
|
|
mask = mask.squeeze() |
|
|
resize_ratio = (im_w / (detect_size - dw), im_h / (detect_size - dh)) |
|
|
blks = postprocess_yolo(blks, self.conf_thresh, self.nms_thresh, resize_ratio) |
|
|
mask = mask[..., :mask.shape[0]-dh, :mask.shape[1]-dw] |
|
|
lines_map = lines_map[..., :lines_map.shape[2]-dh, :lines_map.shape[3]-dw] |
|
|
|
|
|
mask = postprocess_mask(mask) |
|
|
lines, scores = self.seg_rep(None, lines_map, height=im_h, width=im_w) |
|
|
box_thresh = 0.6 |
|
|
idx = np.where(scores[0] > box_thresh) |
|
|
lines, scores = lines[0][idx], scores[0][idx] |
|
|
|
|
|
|
|
|
mask = cv2.resize(mask, (im_w, im_h), interpolation=cv2.INTER_LINEAR) |
|
|
if lines.size == 0: |
|
|
lines = [] |
|
|
else: |
|
|
lines = lines.astype(np.int64) |
|
|
blk_list = group_output([], lines, im_w, im_h, mask, canvas=img) |
|
|
|
|
|
|
|
|
mask_refined = refine_mask(img, mask, blk_list, refine_mode=refine_mode) |
|
|
if keep_undetected_mask: |
|
|
mask_refined = refine_undetected_mask(img, mask, mask_refined, blk_list, refine_mode=refine_mode) |
|
|
|
|
|
return mask, mask_refined, blk_list |