| import numpy as np |
| import cv2 |
| import torch |
|
|
| import os |
| from modules import devices |
| from annotator.annotator_path import models_path |
|
|
| import mmcv |
| from mmdet.apis import inference_detector, init_detector |
| from mmpose.apis import inference_top_down_pose_model |
| from mmpose.apis import init_pose_model, process_mmdet_results, vis_pose_result |
|
|
|
|
| def preprocessing(image, device): |
| |
| scale = 640 / max(image.shape[:2]) |
| image = cv2.resize(image, dsize=None, fx=scale, fy=scale) |
| raw_image = image.astype(np.uint8) |
|
|
| |
| image = image.astype(np.float32) |
| image -= np.array( |
| [ |
| float(104.008), |
| float(116.669), |
| float(122.675), |
| ] |
| ) |
|
|
| |
| image = torch.from_numpy(image.transpose(2, 0, 1)).float().unsqueeze(0) |
| image = image.to(device) |
|
|
| return image, raw_image |
|
|
|
|
| def imshow_keypoints(img, |
| pose_result, |
| skeleton=None, |
| kpt_score_thr=0.1, |
| pose_kpt_color=None, |
| pose_link_color=None, |
| radius=4, |
| thickness=1): |
| """Draw keypoints and links on an image. |
| Args: |
| img (ndarry): The image to draw poses on. |
| pose_result (list[kpts]): The poses to draw. Each element kpts is |
| a set of K keypoints as an Kx3 numpy.ndarray, where each |
| keypoint is represented as x, y, score. |
| kpt_score_thr (float, optional): Minimum score of keypoints |
| to be shown. Default: 0.3. |
| pose_kpt_color (np.array[Nx3]`): Color of N keypoints. If None, |
| the keypoint will not be drawn. |
| pose_link_color (np.array[Mx3]): Color of M links. If None, the |
| links will not be drawn. |
| thickness (int): Thickness of lines. |
| """ |
|
|
| img_h, img_w, _ = img.shape |
| img = np.zeros(img.shape) |
|
|
| for idx, kpts in enumerate(pose_result): |
| if idx > 1: |
| continue |
| kpts = kpts['keypoints'] |
| |
| kpts = np.array(kpts, copy=False) |
|
|
| |
| if pose_kpt_color is not None: |
| assert len(pose_kpt_color) == len(kpts) |
|
|
| for kid, kpt in enumerate(kpts): |
| x_coord, y_coord, kpt_score = int(kpt[0]), int(kpt[1]), kpt[2] |
|
|
| if kpt_score < kpt_score_thr or pose_kpt_color[kid] is None: |
| |
| continue |
|
|
| color = tuple(int(c) for c in pose_kpt_color[kid]) |
| cv2.circle(img, (int(x_coord), int(y_coord)), |
| radius, color, -1) |
|
|
| |
| if skeleton is not None and pose_link_color is not None: |
| assert len(pose_link_color) == len(skeleton) |
|
|
| for sk_id, sk in enumerate(skeleton): |
| pos1 = (int(kpts[sk[0], 0]), int(kpts[sk[0], 1])) |
| pos2 = (int(kpts[sk[1], 0]), int(kpts[sk[1], 1])) |
|
|
| if (pos1[0] <= 0 or pos1[0] >= img_w or pos1[1] <= 0 or pos1[1] >= img_h or pos2[0] <= 0 |
| or pos2[0] >= img_w or pos2[1] <= 0 or pos2[1] >= img_h or kpts[sk[0], 2] < kpt_score_thr |
| or kpts[sk[1], 2] < kpt_score_thr or pose_link_color[sk_id] is None): |
| |
| continue |
| color = tuple(int(c) for c in pose_link_color[sk_id]) |
| cv2.line(img, pos1, pos2, color, thickness=thickness) |
|
|
| return img |
|
|
|
|
| human_det, pose_model = None, None |
| det_model_path = "https://download.openmmlab.com/mmdetection/v2.0/faster_rcnn/faster_rcnn_r50_fpn_1x_coco/faster_rcnn_r50_fpn_1x_coco_20200130-047c8118.pth" |
| pose_model_path = "https://download.openmmlab.com/mmpose/top_down/hrnet/hrnet_w48_coco_256x192-b9e0b3ab_20200708.pth" |
|
|
| modeldir = os.path.join(models_path, "keypose") |
| old_modeldir = os.path.dirname(os.path.realpath(__file__)) |
|
|
| det_config = 'faster_rcnn_r50_fpn_coco.py' |
| pose_config = 'hrnet_w48_coco_256x192.py' |
|
|
| det_checkpoint = 'faster_rcnn_r50_fpn_1x_coco_20200130-047c8118.pth' |
| pose_checkpoint = 'hrnet_w48_coco_256x192-b9e0b3ab_20200708.pth' |
| det_cat_id = 1 |
| bbox_thr = 0.2 |
|
|
| skeleton = [ |
| [15, 13], [13, 11], [16, 14], [14, 12], [11, 12], [5, 11], [6, 12], [5, 6], [5, 7], [6, 8], |
| [7, 9], [8, 10], |
| [1, 2], [0, 1], [0, 2], [1, 3], [2, 4], [3, 5], [4, 6] |
| ] |
|
|
| pose_kpt_color = [ |
| [51, 153, 255], [51, 153, 255], [51, 153, 255], [51, 153, 255], [51, 153, 255], |
| [0, 255, 0], |
| [255, 128, 0], [0, 255, 0], [255, 128, 0], [0, 255, 0], [255, 128, 0], [0, 255, 0], |
| [255, 128, 0], |
| [0, 255, 0], [255, 128, 0], [0, 255, 0], [255, 128, 0] |
| ] |
|
|
| pose_link_color = [ |
| [0, 255, 0], [0, 255, 0], [255, 128, 0], [255, 128, 0], |
| [51, 153, 255], [51, 153, 255], [51, 153, 255], [51, 153, 255], [0, 255, 0], |
| [255, 128, 0], |
| [0, 255, 0], [255, 128, 0], [51, 153, 255], [51, 153, 255], [51, 153, 255], |
| [51, 153, 255], |
| [51, 153, 255], [51, 153, 255], [51, 153, 255] |
| ] |
|
|
| def find_download_model(checkpoint, remote_path): |
| modelpath = os.path.join(modeldir, checkpoint) |
| old_modelpath = os.path.join(old_modeldir, checkpoint) |
| |
| if os.path.exists(old_modelpath): |
| modelpath = old_modelpath |
| elif not os.path.exists(modelpath): |
| from scripts.utils import load_file_from_url |
| load_file_from_url(remote_path, model_dir=modeldir) |
| |
| return modelpath |
|
|
| def apply_keypose(input_image): |
| global human_det, pose_model |
| if netNetwork is None: |
| det_model_local = find_download_model(det_checkpoint, det_model_path) |
| hrnet_model_local = find_download_model(pose_checkpoint, pose_model_path) |
| det_config_mmcv = mmcv.Config.fromfile(det_config) |
| pose_config_mmcv = mmcv.Config.fromfile(pose_config) |
| human_det = init_detector(det_config_mmcv, det_model_local, device=devices.get_device_for("controlnet")) |
| pose_model = init_pose_model(pose_config_mmcv, hrnet_model_local, device=devices.get_device_for("controlnet")) |
|
|
| assert input_image.ndim == 3 |
| input_image = input_image.copy() |
| with torch.no_grad(): |
| image = torch.from_numpy(input_image).float().to(devices.get_device_for("controlnet")) |
| image = image / 255.0 |
| mmdet_results = inference_detector(human_det, image) |
| |
| |
| person_results = process_mmdet_results(mmdet_results, det_cat_id) |
| |
| return_heatmap = False |
| dataset = pose_model.cfg.data['test']['type'] |
| |
| |
| output_layer_names = None |
| pose_results, _ = inference_top_down_pose_model( |
| pose_model, |
| image, |
| person_results, |
| bbox_thr=bbox_thr, |
| format='xyxy', |
| dataset=dataset, |
| dataset_info=None, |
| return_heatmap=return_heatmap, |
| outputs=output_layer_names |
| ) |
| |
| im_keypose_out = imshow_keypoints( |
| image, |
| pose_results, |
| skeleton=skeleton, |
| pose_kpt_color=pose_kpt_color, |
| pose_link_color=pose_link_color, |
| radius=2, |
| thickness=2 |
| ) |
| im_keypose_out = im_keypose_out.astype(np.uint8) |
|
|
| |
| |
| |
| return im_keypose_out |
|
|
|
|
| def unload_hed_model(): |
| global netNetwork |
| if netNetwork is not None: |
| netNetwork.cpu() |
|
|