| | |
| | import math |
| | import cv2 |
| | import os |
| | import math |
| | import glob |
| |
|
| | from utils_from_Depth2HHA_python.rgbd_util import * |
| | from utils_from_Depth2HHA_python.getCameraParam import * |
| |
|
| | def config_setup(): |
| | config = {} |
| | config["home_param"] = "<scene>/" |
| | return config |
| |
|
| | ''' |
| | must use 'COLOR_BGR2GRAY' here, or you will get a different gray-value with what MATLAB gets. |
| | ''' |
| | def getImage(root='demo'): |
| | D = cv2.imread(os.path.join(root, '0.png'), cv2.COLOR_BGR2GRAY)/10000 |
| | RD = cv2.imread(os.path.join(root, '0_raw.png'), cv2.COLOR_BGR2GRAY)/10000 |
| | return D, RD |
| |
|
| |
|
| | ''' |
| | C: Camera matrix |
| | D: Depth image, the unit of each element in it is "meter" |
| | RD: Raw depth image, the unit of each element in it is "meter" |
| | ''' |
| | def getHHA(C, D, RD): |
| | missingMask = (RD == 0); |
| | pc, N, yDir, h, pcRot, NRot = processDepthImage(D * 100, missingMask, C) |
| |
|
| | tmp = np.multiply(N, yDir) |
| | acosValue = np.minimum(1,np.maximum(-1,np.sum(tmp, axis=2))) |
| | angle = np.array([math.degrees(math.acos(x)) for x in acosValue.flatten()]) |
| | angle = np.reshape(angle, h.shape) |
| |
|
| | ''' |
| | Must convert nan to 180 as the MATLAB program actually does. |
| | Or we will get a HHA image whose border region is different |
| | with that of MATLAB program's output. |
| | ''' |
| | angle[np.isnan(angle)] = 180 |
| |
|
| |
|
| | pc[:,:,2] = np.maximum(pc[:,:,2], 100) |
| | I = np.zeros(pc.shape) |
| |
|
| | |
| | I[:,:,2] = 31000/pc[:,:,2] |
| | I[:,:,1] = h |
| | I[:,:,0] = (angle + 128-90) |
| |
|
| | |
| |
|
| | ''' |
| | np.uint8 seems to use 'floor', but in matlab, it seems to use 'round'. |
| | So I convert it to integer myself. |
| | ''' |
| | I = np.rint(I) |
| |
|
| | |
| | I[I>255] = 255 |
| | HHA = I.astype(np.uint8) |
| | return HHA |
| |
|
| | def main(): |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | config = config_setup() |
| | |
| | depth_paths = sorted(glob.glob(config["scene_path"] + "depth/*.png")) |
| |
|
| | for i, depth_path in enumerate(depth_paths): |
| | depth = cv2.imread(depth_path, -1) |
| | |
| | depth = depth / 1000 |
| | H_ori, W_ori = (depth.shape[0], depth.shape[1]) |
| | camera_matrix = np.array([[max(H_ori, W_ori), 0, W_ori/2], [0, max(H_ori, W_ori), H_ori/2], [0, 0, 1]]) |
| | H, W = (int(depth.shape[0]/4), int(depth.shape[1]/4)) |
| | depth_resize = cv2.resize(depth, (W, H), interpolation=cv2.INTER_NEAREST) |
| | hha = getHHA(camera_matrix, depth_resize, depth_resize) |
| | cv2.imwrite(config["scene_path"]+f'HHA/{i:03d}_equi_hha.png', cv2.resize(hha, (W_ori, H_ori), interpolation=cv2.INTER_NEAREST)) |
| |
|
| |
|
| | if __name__ == "__main__": |
| | main() |
| | |
| | ''' multi-peocessing example ''' |
| | ''' |
| | from multiprocessing import Pool |
| | |
| | def generate_hha(i): |
| | # generate hha for the i-th image |
| | return |
| | |
| | processNum = 16 |
| | pool = Pool(processNum) |
| | |
| | for i in range(img_num): |
| | print(i) |
| | pool.apply_async(generate_hha, args=(i,)) |
| | pool.close() |
| | pool.join() |
| | ''' |
| |
|