|
|
import os |
|
|
import logging |
|
|
from torch.utils import data |
|
|
import numpy as np |
|
|
import yaml |
|
|
|
|
|
|
|
|
logger = logging.getLogger(__name__) |
|
|
|
|
|
|
|
|
|
|
|
class Field(object): |
|
|
''' Data fields class. |
|
|
''' |
|
|
|
|
|
def load(self, data_path, idx, category): |
|
|
''' Loads a data point. |
|
|
|
|
|
Args: |
|
|
data_path (str): path to data file |
|
|
idx (int): index of data point |
|
|
category (int): index of category |
|
|
''' |
|
|
raise NotImplementedError |
|
|
|
|
|
def check_complete(self, files): |
|
|
''' Checks if set is complete. |
|
|
|
|
|
Args: |
|
|
files: files |
|
|
''' |
|
|
raise NotImplementedError |
|
|
|
|
|
|
|
|
class Shapes3dDataset(data.Dataset): |
|
|
''' 3D Shapes dataset class. |
|
|
''' |
|
|
|
|
|
def __init__(self, dataset_folder, fields, split=None, |
|
|
categories=None, no_except=True, transform=None): |
|
|
''' Initialization of the the 3D shape dataset. |
|
|
|
|
|
Args: |
|
|
dataset_folder (str): dataset folder |
|
|
fields (dict): dictionary of fields |
|
|
split (str): which split is used |
|
|
categories (list): list of categories to use |
|
|
no_except (bool): no exception |
|
|
transform (callable): transformation applied to data points |
|
|
''' |
|
|
|
|
|
self.dataset_folder = dataset_folder |
|
|
self.fields = fields |
|
|
self.no_except = no_except |
|
|
self.transform = transform |
|
|
|
|
|
|
|
|
if categories is None: |
|
|
categories = os.listdir(dataset_folder) |
|
|
categories = [c for c in categories |
|
|
if os.path.isdir(os.path.join(dataset_folder, c))] |
|
|
|
|
|
|
|
|
metadata_file = os.path.join(dataset_folder, 'metadata.yaml') |
|
|
|
|
|
if os.path.exists(metadata_file): |
|
|
with open(metadata_file, 'r') as f: |
|
|
self.metadata = yaml.load(f) |
|
|
else: |
|
|
self.metadata = { |
|
|
c: {'id': c, 'name': 'n/a'} for c in categories |
|
|
} |
|
|
|
|
|
|
|
|
for c_idx, c in enumerate(categories): |
|
|
self.metadata[c]['idx'] = c_idx |
|
|
|
|
|
|
|
|
self.models = [] |
|
|
for c_idx, c in enumerate(categories): |
|
|
subpath = os.path.join(dataset_folder, c) |
|
|
if not os.path.isdir(subpath): |
|
|
logger.warning('Category %s does not exist in dataset.' % c) |
|
|
|
|
|
split_file = os.path.join(subpath, split + '.lst') |
|
|
with open(split_file, 'r') as f: |
|
|
models_c = f.read().split('\n') |
|
|
|
|
|
self.models += [ |
|
|
{'category': c, 'model': m} |
|
|
for m in models_c |
|
|
] |
|
|
|
|
|
def __len__(self): |
|
|
''' Returns the length of the dataset. |
|
|
''' |
|
|
return len(self.models) |
|
|
|
|
|
def __getitem__(self, idx): |
|
|
''' Returns an item of the dataset. |
|
|
|
|
|
Args: |
|
|
idx (int): ID of data point |
|
|
''' |
|
|
category = self.models[idx]['category'] |
|
|
model = self.models[idx]['model'] |
|
|
c_idx = self.metadata[category]['idx'] |
|
|
|
|
|
model_path = os.path.join(self.dataset_folder, category, model) |
|
|
data = {} |
|
|
|
|
|
for field_name, field in self.fields.items(): |
|
|
try: |
|
|
field_data = field.load(model_path, idx, c_idx) |
|
|
except Exception: |
|
|
if self.no_except: |
|
|
logger.warn( |
|
|
'Error occured when loading field %s of model %s' |
|
|
% (field_name, model) |
|
|
) |
|
|
return None |
|
|
else: |
|
|
raise |
|
|
|
|
|
if isinstance(field_data, dict): |
|
|
for k, v in field_data.items(): |
|
|
if k is None: |
|
|
data[field_name] = v |
|
|
else: |
|
|
data['%s.%s' % (field_name, k)] = v |
|
|
else: |
|
|
data[field_name] = field_data |
|
|
|
|
|
if self.transform is not None: |
|
|
data = self.transform(data) |
|
|
|
|
|
return data |
|
|
|
|
|
def get_model_dict(self, idx): |
|
|
return self.models[idx] |
|
|
|
|
|
def test_model_complete(self, category, model): |
|
|
''' Tests if model is complete. |
|
|
|
|
|
Args: |
|
|
model (str): modelname |
|
|
''' |
|
|
model_path = os.path.join(self.dataset_folder, category, model) |
|
|
files = os.listdir(model_path) |
|
|
for field_name, field in self.fields.items(): |
|
|
if not field.check_complete(files): |
|
|
logger.warn('Field "%s" is incomplete: %s' |
|
|
% (field_name, model_path)) |
|
|
return False |
|
|
|
|
|
return True |
|
|
|
|
|
|
|
|
def collate_remove_none(batch): |
|
|
''' Collater that puts each data field into a tensor with outer dimension |
|
|
batch size. |
|
|
|
|
|
Args: |
|
|
batch: batch |
|
|
''' |
|
|
|
|
|
batch = list(filter(lambda x: x is not None, batch)) |
|
|
return data.dataloader.default_collate(batch) |
|
|
|
|
|
|
|
|
def worker_init_fn(worker_id): |
|
|
''' Worker init function to ensure true randomness. |
|
|
''' |
|
|
random_data = os.urandom(4) |
|
|
base_seed = int.from_bytes(random_data, byteorder="big") |
|
|
np.random.seed(base_seed + worker_id) |
|
|
|