|
|
import os |
|
|
import collections |
|
|
import json |
|
|
import random |
|
|
from copy import deepcopy |
|
|
import jsonlines |
|
|
from tqdm import tqdm |
|
|
import numpy as np |
|
|
import torch |
|
|
|
|
|
from .default import DATASET_REGISTRY |
|
|
from ..data_utils import build_rotate_mat |
|
|
from data.data_utils import (VICUNA_ACTION_TOKENS) |
|
|
from .scannet_base import ScanNetBase |
|
|
from scipy.spatial.transform import Rotation as R |
|
|
|
|
|
ONESTEPNAVI_ACTION_SPACE = { |
|
|
'move_forward': 0, |
|
|
'turn_left': 1, |
|
|
'move_backward': 2, |
|
|
'turn_right': 3, |
|
|
'turn_left_forward': 4, |
|
|
'turn_left_backward': 5, |
|
|
'turn_right_backward': 6, |
|
|
'turn_right_forward': 7, |
|
|
} |
|
|
|
|
|
ONESTEPNAVI_ACTION_SPACE_TOKENIZE = { |
|
|
k: v for k, v in zip(list(ONESTEPNAVI_ACTION_SPACE.values()), list(VICUNA_ACTION_TOKENS.keys())[:len(ONESTEPNAVI_ACTION_SPACE)]) |
|
|
} |
|
|
|
|
|
NAVI_ACTION_POOL = [ |
|
|
"What action should I take next step?", |
|
|
] |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
@DATASET_REGISTRY.register() |
|
|
class ScanNetOneStepNavi(ScanNetBase): |
|
|
def __init__(self, cfg, split): |
|
|
super().__init__(cfg, split) |
|
|
self.dataset_cfg = cfg.data.next_step_navigation.args |
|
|
|
|
|
self.num_points = self.dataset_cfg.get('num_points', 1024) |
|
|
self.max_obj_len = self.dataset_cfg.get('max_obj_len', 60) |
|
|
self.pc_type = self.dataset_cfg.get('pc_type', 'gt') |
|
|
self.action_type = self.dataset_cfg.get('action_type', 'four_direction') |
|
|
self.modality_type = self.dataset_cfg.get('modality_type', 'multimodal') |
|
|
|
|
|
assert self.pc_type in ['gt', 'pred'] |
|
|
assert self.split in ['train', 'val', 'test'] |
|
|
if self.split == 'train': |
|
|
self.pc_type = 'gt' |
|
|
if self.split == 'test': |
|
|
self.split = 'val' |
|
|
|
|
|
self.action_mapping = { |
|
|
"four_direction": {0:0, 1:1, 2:2, 3:3, 4:0}, |
|
|
"eight_direction": {0:0, 2:1, 4:2, 6:3, 8:0, 1:4, 3:5, 5:6, 7:7}, |
|
|
} |
|
|
|
|
|
self.scan_ids = self._load_split(cfg, self.split) |
|
|
anno_file_path = os.path.join(cfg.data.msnn_base, 'msnn_scannet.json') |
|
|
with open(anno_file_path, 'r') as f: |
|
|
anno_info_all = json.load(f) |
|
|
|
|
|
print(f"Loading ScanNet ScanNetOneStepNavi {split}-set language") |
|
|
self.data, self.scan_ids = self._load_lang(anno_info_all, self.scan_ids) |
|
|
if cfg.debug.flag: |
|
|
self.data = self.data[:cfg.debug.debug_size] |
|
|
print(f"Finish loading ScanNetOneStepNavi {split}-set language") |
|
|
|
|
|
|
|
|
print(f"Loading ScanNet ScanNetOneStepNavi {split}-set scans") |
|
|
self.scan_data = self._load_scannet(self.scan_ids, self.pc_type, self.pc_type == 'gt') |
|
|
print(f"Finish loading ScanNet ScanNetOneStepNavi {split}-set data") |
|
|
|
|
|
def _load_lang(self, anno_info_all, select_scan_ids): |
|
|
output_list = [] |
|
|
scan_ids = [] |
|
|
for scan_id, samples_one_scene in anno_info_all.items(): |
|
|
if scan_id not in select_scan_ids: |
|
|
continue |
|
|
scan_ids.append(scan_id) |
|
|
for one_sample in samples_one_scene.values(): |
|
|
one_sample['insts'] = [int(x) for x in one_sample['insts']] |
|
|
output_list.append(one_sample) |
|
|
scan_ids = list(set(scan_ids)) |
|
|
return output_list, scan_ids |
|
|
|
|
|
def __len__(self): |
|
|
return len(self.data) |
|
|
|
|
|
|
|
|
def preprocess_pcd(self, obj_pcds, return_anchor = False, rot_aug = True, situation = None): |
|
|
|
|
|
rot_matrix = build_rotate_mat(self.split, rot_aug=rot_aug) |
|
|
|
|
|
|
|
|
obj_fts = [] |
|
|
obj_locs = [] |
|
|
for i, obj_pcd in enumerate(obj_pcds): |
|
|
if rot_matrix is not None: |
|
|
obj_pcd[:, :3] = np.matmul(obj_pcd[:, :3], rot_matrix.transpose()) |
|
|
|
|
|
obj_center = obj_pcd[:, :3].mean(0) |
|
|
obj_size = obj_pcd[:, :3].max(0) - obj_pcd[:, :3].min(0) |
|
|
obj_locs.append(np.concatenate([obj_center, obj_size], 0)) |
|
|
if return_anchor and i == 0: |
|
|
|
|
|
anchor_loc = obj_pcd[:, :3].min(0) + np.random.rand(3) * obj_size |
|
|
|
|
|
|
|
|
pcd_idxs = np.random.choice(len(obj_pcd), size=self.num_points, |
|
|
replace=len(obj_pcd) < self.num_points) |
|
|
obj_pcd = obj_pcd[pcd_idxs] |
|
|
|
|
|
|
|
|
obj_pcd[:, :3] = obj_pcd[:, :3] - obj_pcd[:, :3].mean(0) |
|
|
max_dist = np.sqrt((obj_pcd[:, :3]**2).sum(1)).max() |
|
|
if max_dist < 1e-6: |
|
|
max_dist = 1 |
|
|
obj_pcd[:, :3] = obj_pcd[:, :3] / max_dist |
|
|
obj_fts.append(obj_pcd) |
|
|
|
|
|
|
|
|
obj_fts = torch.from_numpy(np.stack(obj_fts, 0)) |
|
|
obj_locs = torch.from_numpy(np.array(obj_locs)) |
|
|
if return_anchor: |
|
|
anchor_loc = torch.from_numpy(anchor_loc) |
|
|
else: |
|
|
anchor_loc = torch.zeros(3).float() |
|
|
|
|
|
output_dict = { |
|
|
'obj_fts': obj_fts, |
|
|
'obj_locs': obj_locs, |
|
|
'anchor_loc': anchor_loc, |
|
|
} |
|
|
|
|
|
if situation is not None: |
|
|
if rot_matrix is None: |
|
|
output_dict["situation"] = situation |
|
|
else: |
|
|
pos, ori = situation |
|
|
pos = np.array(pos) |
|
|
ori = np.array(ori) |
|
|
pos_new = pos.reshape(1, 3) @ rot_matrix.transpose() |
|
|
pos_new = pos_new.reshape(-1) |
|
|
ori_new = R.from_quat(ori).as_matrix() |
|
|
ori_new = rot_matrix @ ori_new |
|
|
ori_new = R.from_matrix(ori_new).as_quat() |
|
|
ori_new = ori_new.reshape(-1) |
|
|
output_dict["situation"] = (pos_new, ori_new) |
|
|
return output_dict |
|
|
|
|
|
|
|
|
def _get_scene_encoder_input(self, obj_pcds, scan_insts, situation = None): |
|
|
|
|
|
if len(obj_pcds) <= self.max_obj_len: |
|
|
|
|
|
selected_obj_pcds = list(obj_pcds.values()) |
|
|
else: |
|
|
|
|
|
selected_obj_pcds = [] |
|
|
|
|
|
|
|
|
for i in scan_insts: |
|
|
if i in obj_pcds: |
|
|
selected_obj_pcds.append(obj_pcds[i]) |
|
|
|
|
|
num_selected_objs = len(selected_obj_pcds) |
|
|
if num_selected_objs >= self.max_obj_len: |
|
|
random.shuffle(selected_obj_pcds) |
|
|
selected_obj_pcds = selected_obj_pcds[:self.max_obj_len] |
|
|
else: |
|
|
|
|
|
remained_obj_idx = [i for i in obj_pcds.keys() if i not in scan_insts] |
|
|
random.shuffle(remained_obj_idx) |
|
|
for i in remained_obj_idx[: self.max_obj_len - num_selected_objs]: |
|
|
selected_obj_pcds.append(obj_pcds[i]) |
|
|
|
|
|
assert len(selected_obj_pcds) == self.max_obj_len |
|
|
output_dict = self.preprocess_pcd(selected_obj_pcds, return_anchor = False, rot_aug = True, situation = situation) |
|
|
return output_dict |
|
|
|
|
|
def __getitem__(self, index): |
|
|
one_sample = self.data[index] |
|
|
|
|
|
if self.modality_type == 'multimodal': |
|
|
situation = one_sample['situation_multimodal'] |
|
|
else: |
|
|
situation = one_sample['situation_text'] |
|
|
interaction = one_sample['interaction'] |
|
|
anchor_loc = one_sample['location'] |
|
|
anchor_orientation = one_sample['orientation'] |
|
|
question = random.choice(NAVI_ACTION_POOL) |
|
|
question = interaction + " " + question |
|
|
|
|
|
|
|
|
scan_id = one_sample['scan_id'] |
|
|
obj_pcds = self.scan_data[scan_id]['obj_pcds'] |
|
|
obj_pcds = {int(k): obj_pcds[k] for k in range(len(obj_pcds))} |
|
|
|
|
|
action_token_list = [] |
|
|
action_gt_code = one_sample['action'][self.action_type][0] |
|
|
action_gt_code = self.action_mapping[self.action_type][action_gt_code] |
|
|
action_gt = ONESTEPNAVI_ACTION_SPACE_TOKENIZE[action_gt_code] |
|
|
action_token_list.append(action_gt) |
|
|
|
|
|
action_text_list = [] |
|
|
action_text = one_sample['action'][self.action_type][1] |
|
|
action_text_list.append(action_text) |
|
|
|
|
|
|
|
|
output_dict = self._get_scene_encoder_input(obj_pcds, one_sample['insts'], situation = (anchor_loc, anchor_orientation)) |
|
|
obj_fts = output_dict['obj_fts'] |
|
|
obj_locs = output_dict['obj_locs'] |
|
|
anchor_loc, anchor_orientation = output_dict["situation"] |
|
|
|
|
|
data_dict = { |
|
|
"situation": situation, |
|
|
"situation_pos": np.array(anchor_loc), |
|
|
"situation_rot": np.array(anchor_orientation), |
|
|
"question": question, |
|
|
"action_token_list": action_token_list, |
|
|
"action_text_list": action_text_list, |
|
|
"obj_fts": obj_fts, |
|
|
"obj_locs": obj_locs, |
|
|
"scan_id": scan_id, |
|
|
} |
|
|
return data_dict |