| | import os |
| | from torch.utils import data |
| | import numpy as np |
| |
|
| |
|
| |
|
| | class HumansDataset(data.Dataset): |
| | ''' 3D Shapes dataset class. |
| | ''' |
| |
|
| | def __init__(self, dataset_folder, fields, mode, split=None, category='D-FAUST', |
| | length_sequence=17, n_files_per_sequence=-1, offset_sequence=0, |
| | ex_folder_name='pcl_seq', specific_model=None): |
| | ''' Initialization of the the 3D shape dataset. |
| | |
| | Args: |
| | dataset_folder (str): dataset folder |
| | fields (dict): dictionary of fields |
| | category (str): category of data |
| | split (str): which split is used |
| | ''' |
| |
|
| | |
| | self.dataset_folder = dataset_folder |
| | self.fields = fields |
| | self.mode = mode |
| | self.length_sequence = length_sequence |
| | self.n_files_per_sequence = n_files_per_sequence |
| | self.offset_sequence = offset_sequence |
| | self.ex_folder_name = ex_folder_name |
| |
|
| | if mode == 'train': |
| | with open(os.path.join(self.dataset_folder, 'train_human_ids.lst'), 'r') as f: |
| | self.hid = f.read().split('\n') |
| |
|
| | if specific_model is not None: |
| | self.models = [{'category': category, 'model': specific_model['seq'], |
| | 'start_idx': specific_model['start_idx']}] |
| | else: |
| | |
| | self.models = [] |
| | subpath = os.path.join(self.dataset_folder, 'test', category) |
| | if split is not None and os.path.exists( |
| | os.path.join(subpath, split + '.lst')): |
| | split_file = os.path.join(subpath, split + '.lst') |
| | with open(split_file, 'r') as f: |
| | models_c = f.read().split('\n') |
| | else: |
| | models_c = [f for f in os.listdir(subpath) if |
| | os.path.isdir(os.path.join(subpath, f))] |
| | models_c = list(filter(lambda x: len(x) > 0, models_c)) |
| | models_len = self.get_models_seq_len(subpath, models_c) |
| | models_c, start_idx = self.subdivide_into_sequences( |
| | models_c, models_len) |
| | self.models += [ |
| | {'category': category, 'model': m, 'start_idx': start_idx[i]} |
| | for i, m in enumerate(models_c) |
| | ] |
| |
|
| |
|
| | def __len__(self): |
| | if self.mode == 'train': |
| | return 2048 * 16 |
| | else: |
| | return len(self.models) |
| |
|
| |
|
| | def __getitem__(self, idx): |
| | '''identity enchange is conducted only in training mode |
| | ''' |
| | if self.mode == 'train': |
| | id1, id2 = np.random.choice(self.hid, size=2, replace=False) |
| | m1, m2 = np.random.choice(self.models, size=2) |
| | data1 = self.get_data_dict(m1, idx, id1, id2) |
| | data2 = self.get_data_dict(m2, idx, id2, id1) |
| | return [data1, data2] |
| | else: |
| | m = self.models[idx] |
| | data = self.get_data_dict(m, idx) |
| | return data |
| |
|
| |
|
| | def get_data_dict(self, motion, idx=0, id1=None, id2=None): |
| | category = motion['category'] |
| | model = motion['model'] |
| | start_idx = motion['start_idx'] |
| |
|
| | data = {} |
| |
|
| | for field_name, field in self.fields.items(): |
| | if self.mode == 'train': |
| | if field_name in ['points_ex', 'points_t_ex', 'points_iou_ex']: |
| | model_path = os.path.join(self.dataset_folder, 'train', category, id2, model) |
| | else: |
| | model_path = os.path.join(self.dataset_folder, 'train', category, id1, model) |
| | else: |
| | model_path = os.path.join(self.dataset_folder, 'test', category, model) |
| |
|
| | field_data = field.load(model_path, idx, start_idx=start_idx, dataset_folder=self.dataset_folder) |
| |
|
| | if isinstance(field_data, dict): |
| | for k, v in field_data.items(): |
| | if k is None: |
| | data[field_name] = v |
| | else: |
| | data['%s.%s' % (field_name, k)] = v |
| | else: |
| | data[field_name] = field_data |
| |
|
| | return data |
| |
|
| |
|
| | def get_model_dict(self, idx): |
| | return self.models[idx] |
| |
|
| |
|
| | def get_models_seq_len(self, subpath, models): |
| | ''' Returns the sequence length of a specific model. |
| | |
| | This is a little "hacky" as we assume the existence of the folder |
| | self.ex_folder_name. However, in our case this is always given. |
| | |
| | Args: |
| | subpath (str): subpath of model category |
| | models (list): list of model names |
| | ''' |
| | ex_folder_name = self.ex_folder_name |
| | models_seq_len = [ |
| | len(os.listdir(os.path.join(subpath, m, ex_folder_name))) |
| | for m in models] |
| |
|
| | return models_seq_len |
| |
|
| |
|
| | def subdivide_into_sequences(self, models, models_len): |
| | ''' Subdivides model sequence into smaller sequences. |
| | |
| | Args: |
| | models (list): list of model names |
| | models_len (list): list of lengths of model sequences |
| | ''' |
| | length_sequence = self.length_sequence |
| | n_files_per_sequence = self.n_files_per_sequence |
| | offset_sequence = self.offset_sequence |
| |
|
| | |
| | models_len = [l - offset_sequence for l in models_len] |
| |
|
| | |
| | if n_files_per_sequence > 0: |
| | models_len = [min(n_files_per_sequence, l) for l in models_len] |
| |
|
| | models_out = [] |
| | start_idx = [] |
| | for idx, model in enumerate(models): |
| | for n in range(0, models_len[idx] - length_sequence + 1): |
| | models_out.append(model) |
| | start_idx.append(n + offset_sequence) |
| |
|
| | return models_out, start_idx |
| |
|