File size: 2,237 Bytes
c94c8c9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
import os
import json
from joblib import Parallel, delayed, parallel_backend
from glob import glob
from tqdm import tqdm
import numpy as np
import argparse


def load_semantic_anno(semantic_txt):
    semantic_color = []
    obj_name_list = []
    color_2_name = {}
    color_2_id = {}
    with open(semantic_txt) as f:
        lines = f.readlines()[1:]
        for line in lines:
            obj_id = int(line.split(',')[0])
            color_str = line.split(',')[1]
            if len(color_str) != 6:
                color_str = '0' * (6 - len(color_str)) + color_str
            r = int(color_str[0:2], 16)
            g = int(color_str[2:4], 16)
            b = int(color_str[4:6], 16)
            obj_name = line.split(',')[2][1:-1]
            obj_name_list.append(obj_name)
            rgb_value = np.array([r, g, b], dtype=np.uint8).reshape(1, 3)
            semantic_color.append(rgb_value)
            color_2_name[(r, g, b)] = obj_name
            color_2_id[(r, g, b)] = obj_id
    return np.concatenate(semantic_color, axis=0), obj_name_list, color_2_name, color_2_id


def scene_proc(scene_input):
    scene_name = scene_input.split('/')[-1]
    scene_uid = scene_name.split('-')[1]
    sem_dir = scene_input + '/' + scene_uid + '.semantic'
    print(scene_name)

    # load obj semantics anno
    semantic_anno_color, obj_name_list, color_2_name, color_2_id = load_semantic_anno(sem_dir+'.txt')

    tgt_id2obj_id = {}
    # obj assignment and export
    semantic_anno_set = set(list(zip(*(semantic_anno_color.T))))
    for _i, sem in enumerate(tqdm(semantic_anno_set)):
        obj_name = color_2_name[(sem[0], sem[1], sem[2])]
        obj_id = color_2_id[(sem[0], sem[1], sem[2])]
        tgt_id2obj_id[_i+1] = (obj_id, obj_name)
    json.dump(tgt_id2obj_id, open(os.path.join(scene_input, 'tgt_id2obj_id.json'), 'w'), indent=4)


if __name__ == "__main__":
    parser = argparse.ArgumentParser()
    parser.add_argument('--data_root', type=str, default='./hm3d-train-annots', help='data root for hm-semantics data')
    args = parser.parse_args()
    scene_list = glob(args.data_root + '/*')
    with parallel_backend('multiprocessing', n_jobs=1):
        Parallel()(delayed(scene_proc)(scene) for scene in scene_list)