Dorothydu's picture
Add files using upload-large-folder tool
e4d50d8 verified
import os
import glob
import random
from PIL import Image
import numpy as np
import trimesh
from lib.data.core import Field
from lib.common import random_crop_occ
class IndexField(Field):
''' Basic index field.'''
# def load(self, model_path, idx, category):
def load(self, model_path, idx, start_idx=0, dataset_folder=None, **kwargs):
''' Loads the index field.
Args:
model_path (str): path to model
idx (int): ID of data point
start_idx (int): id of sequence start
dataset_folder (str): dataset folder
'''
return idx
def check_complete(self, files):
''' Check if field is complete.
Args:
files: files
'''
return True
class PointsSubseqField(Field):
''' Points subsequence field class.
Args:
folder_name (str): points folder name
transform (transform): transform
seq_len (int): length of sequence
all_steps (bool): whether to return all time steps
fixed_time_step (int): if and which fixed time step to use
unpackbits (bool): whether to unpack bits
scale_type (str, optional): Specifies the type of transformation to apply to the point cloud:
``'cr'`` | ``'oflow'``. ``'cr'``: transform the point cloud to align with the output,
``'oflow'``: scale the point cloud w.r.t. the first point cloud of the sequence
spatial_completion (bool): whether to remove some points for 4D spatial completion experiment
'''
def __init__(self, folder_name, transform=None, seq_len=17,
all_steps=False, fixed_time_step=None, unpackbits=False,
scale_type=None, spatial_completion=False, **kwargs):
self.folder_name = folder_name
self.transform = transform
self.seq_len = seq_len
self.all_steps = all_steps
self.sample_padding = 0.1
self.fixed_time_step = fixed_time_step
self.unpackbits = unpackbits
self.scale_type = scale_type
self.spatial_completion = spatial_completion
if scale_type is not None:
assert scale_type in ['oflow', 'cr']
def get_loc_scale(self, mesh):
''' Returns location and scale of mesh.
Args:
mesh (trimesh): mesh
'''
bbox = mesh.bounding_box.bounds
# Compute location and scale with padding of 0.1
loc = (bbox[0] + bbox[1]) / 2
scale = (bbox[1] - bbox[0]).max() / (1 - self.sample_padding)
return loc, scale
def normalize_mesh(self, mesh, loc, scale):
''' Normalize mesh.
Args:
mesh (trimesh): mesh
loc (tuple): location for normalization
scale (float): scale for normalization
'''
# Transform input mesh
mesh.apply_translation(-loc)
mesh.apply_scale(1 / scale)
return mesh
def load_files(self, model_path, start_idx):
''' Loads the model files.
Args:
model_path (str): path to model
start_idx (int): id of sequence start
'''
folder = os.path.join(model_path, self.folder_name)
files = glob.glob(os.path.join(folder, '*.npz'))
files.sort()
files = files[start_idx:start_idx+self.seq_len]
return files
def load_all_steps(self, files, loc0, scale0, loc_global, scale_global, dataset_folder):
''' Loads data for all steps.
Args:
files (list): list of files
points_dict (dict): points dictionary for first step of sequence
loc0 (tuple): location of first time step mesh
scale0 (float): scale of first time step mesh
'''
p_list = []
o_list = []
t_list = []
for i, f in enumerate(files):
points_dict = np.load(f)
# Load points
points = points_dict['points']
if (points.dtype == np.float16):
# break symmetry (nec. for some version)
points = points.astype(np.float32)
points += 1e-4 * np.random.randn(*points.shape)
occupancies = points_dict['occupancies']
if self.unpackbits:
occupancies = np.unpackbits(occupancies)[:points.shape[0]]
occupancies = occupancies.astype(np.float32)
loc = points_dict['loc'].astype(np.float32)
scale = points_dict['scale'].astype(np.float32)
model_id, _, frame_id = f.split('/')[-3:]
# Remove some points for 4D spatial completion experiment
if self.spatial_completion:
data_folder = os.path.join(dataset_folder, 'test', 'D-FAUST', model_id)
mask_folder = os.path.join(dataset_folder, 'spatial_mask', model_id)
if not os.path.exists(mask_folder):
os.makedirs(mask_folder)
mask_file = os.path.join(mask_folder, frame_id.replace('.npz', '.npy'))
if os.path.exists(mask_file):
mask = np.load(mask_file)
else:
pcl = np.load(os.path.join(data_folder, 'pcl_seq', frame_id))['points']
mask, _, _ = random_crop_occ(points, pcl)
np.save(mask_file, mask)
points = points[mask, :]
occupancies = occupancies[mask]
if self.scale_type is not None:
# Transform to loc0, scale0
if self.scale_type == 'oflow':
points = (loc + scale * points - loc0) / scale0
# Align the testing data of the original D-FAUST with the output of our model
if self.scale_type == 'cr':
trans = np.load(os.path.join(dataset_folder, 'smpl_params', model_id, frame_id))['trans']
loc -= trans
points = (loc + scale * points - loc_global) / scale_global
points = points.astype(np.float32)
time = np.array(i / (self.seq_len - 1), dtype=np.float32)
p_list.append(points)
o_list.append(occupancies)
t_list.append(time)
if not self.spatial_completion:
data = {
None: np.stack(p_list),
'occ': np.stack(o_list),
'time': np.stack(t_list),
}
else:
data = {
None: p_list,
'occ': o_list,
'time': np.stack(t_list),
}
return data
def load_single_step(self, files, points_dict, loc0, scale0):
''' Loads data for a single step.
Args:
files (list): list of files
points_dict (dict): points dictionary for first step of sequence
loc0 (tuple): location of first time step mesh
scale0 (float): scale of first time step mesh
'''
if self.fixed_time_step is None:
# Random time step
time_step = np.random.choice(self.seq_len)
else:
time_step = int(self.fixed_time_step)
if time_step != 0:
points_dict = np.load(files[time_step])
# Load points
points = points_dict['points'].astype(np.float32)
occupancies = points_dict['occupancies']
if self.unpackbits:
occupancies = np.unpackbits(occupancies)[:points.shape[0]]
occupancies = occupancies.astype(np.float32)
if self.scale_type == 'oflow':
loc = points_dict['loc'].astype(np.float32)
scale = points_dict['scale'].astype(np.float32)
# Transform to loc0, scale0
points = (loc + scale * points - loc0) / scale0
if self.seq_len > 1:
time = np.array(
time_step / (self.seq_len - 1), dtype=np.float32)
else:
time = np.array([1], dtype=np.float32)
data = {
None: points,
'occ': occupancies,
'time': time,
}
return data
def load(self, model_path, idx, c_idx=None, start_idx=0, dataset_folder=None, **kwargs):
''' Loads the points subsequence field.
Args:
model_path (str): path to model
idx (int): ID of data point
start_idx (int): id of sequence start
dataset_folder (str): dataset folder
'''
files = self.load_files(model_path, start_idx)
# Load loc and scale from t_0, we use the global loc and scale calculated from the whole training set
points_dict = np.load(files[0])
loc0 = points_dict['loc'].astype(np.float32)
scale0 = points_dict['scale'].astype(np.float32)
loc_global = np.array([-0.005493, -0.1888, 0.07587]).astype(np.float32)
scale_global = 2.338
if self.all_steps:
data = self.load_all_steps(files, loc0, scale0, loc_global, scale_global, dataset_folder)
else:
data = self.load_single_step(files, points_dict, loc0, scale0)
if self.transform is not None:
data = self.transform(data)
return data
class PointCloudSubseqField(Field):
''' Point cloud subsequence field class.
Args:
folder_name (str): points folder name
transform (transform): transform
seq_len (int): length of sequence
only_end_points (bool): whether to only return end points
scale_type (str, optional): Specifies the type of transformation to apply to the input point cloud:
``'cr'`` | ``'oflow'``. ``'cr'``: transform the point cloud the original scale and location of SMPL model,
``'oflow'``: scale the point cloud w.r.t. the first point cloud of the sequence
'''
def __init__(self, folder_name, transform=None, seq_len=17,
only_end_points=False, scale_type=None, eval_mode=False):
self.folder_name = folder_name
self.transform = transform
self.seq_len = seq_len
self.only_end_points = only_end_points
self.scale_type = scale_type
self.eval_mode = eval_mode
if scale_type is not None:
assert scale_type in ['oflow', 'cr']
def return_loc_scale(self, mesh):
''' Returns location and scale of mesh.
Args:
mesh (trimesh): mesh
'''
bbox = mesh.bounding_box.bounds
# Compute location and scale
loc = (bbox[0] + bbox[1]) / 2
scale = (bbox[1] - bbox[0]).max() / (1 - 0)
return loc, scale
def apply_normalization(self, mesh, loc, scale):
''' Normalizes the mesh.
Args:
mesh (trimesh): mesh
loc (tuple): location for normalization
scale (float): scale for normalization
'''
mesh.apply_translation(-loc)
mesh.apply_scale(1/scale)
return mesh
def load_files(self, model_path, start_idx):
''' Loads the model files.
Args:
model_path (str): path to model
start_idx (int): id of sequence start
'''
folder = os.path.join(model_path, self.folder_name)
files = glob.glob(os.path.join(folder, '*.npz'))
files.sort()
files = files[start_idx:start_idx+self.seq_len]
if self.only_end_points:
files = [files[0], files[-1]]
return files
def load_single_file(self, file_path):
''' Loads a single file.
Args:
file_path (str): file path
'''
pointcloud_dict = np.load(file_path)
points = pointcloud_dict['points'].astype(np.float32)
loc = pointcloud_dict['loc'].astype(np.float32)
scale = pointcloud_dict['scale'].astype(np.float32)
return points, loc, scale
def get_time_values(self):
''' Returns the time values.
'''
if self.seq_len > 1:
time = \
np.array([i/(self.seq_len - 1) for i in range(self.seq_len)],
dtype=np.float32)
else:
time = np.array([1]).astype(np.float32)
return time
def load(self, model_path, idx, c_idx=None, start_idx=0, dataset_folder=None, **kwargs):
''' Loads the point cloud sequence field.
Args:
model_path (str): path to model
idx (int): ID of data point
c_idx (int): index of category
start_idx (int): id of sequence start
dataset_folder (str): dataset folder
'''
pc_seq = []
# Get file paths
files = self.load_files(model_path, start_idx)
# Load first pcl file
_, loc0, scale0 = self.load_single_file(files[0])
loc_global = np.array([-0.005493, -0.1888, 0.07587]).astype(np.float32)
scale_global = 2.338
for f in files:
points, loc, scale = self.load_single_file(f)
if self.scale_type is not None:
# Transform mesh to loc0 / scale0
if self.scale_type == 'oflow':
points = (loc + scale * points - loc0) / scale0
# Transform to original scale and location of SMPL model
if self.scale_type == 'cr':
points = loc + scale * points
model_id, _, frame_id = f.split('/')[-3:]
trans = np.load(os.path.join(dataset_folder, 'smpl_params', model_id, frame_id))['trans']
points = points - trans
# Only for evaluation, align the output with the testing data in D-FAUST
if self.eval_mode:
points = (points - loc_global) / scale_global
pc_seq.append(points)
data = {
None: np.stack(pc_seq),
'time': self.get_time_values(),
}
if self.transform is not None:
data = self.transform(data)
return data