|
|
import math |
|
|
import bisect |
|
|
|
|
|
import imgaug |
|
|
import numpy as np |
|
|
|
|
|
import torch |
|
|
import torch.distributed as dist |
|
|
from torch.utils.data import Sampler, ConcatDataset, BatchSampler |
|
|
|
|
|
from concern.config import Configurable, State |
|
|
|
|
|
|
|
|
def default_worker_init_fn(worker_id): |
|
|
np.random.seed(worker_id) |
|
|
imgaug.seed(worker_id) |
|
|
|
|
|
|
|
|
class DataLoader(Configurable, torch.utils.data.DataLoader): |
|
|
dataset = State() |
|
|
batch_size = State(default=256) |
|
|
num_workers = State(default=10) |
|
|
is_train = State(default=True) |
|
|
collect_fn = State(default=None) |
|
|
drop_last = State(default=True) |
|
|
shuffle = State() |
|
|
|
|
|
def __init__(self, **kwargs): |
|
|
self.load_all(**kwargs) |
|
|
if self.collect_fn is None: |
|
|
self.collect_fn = torch.utils.data.dataloader.default_collate |
|
|
cmd = kwargs.get('cmd', {}) |
|
|
self.is_train = cmd['is_train'] |
|
|
if 'batch_size' in cmd: |
|
|
self.batch_size = cmd['batch_size'] |
|
|
if self.shuffle is None: |
|
|
self.shuffle = self.is_train |
|
|
self.num_workers = cmd.get('num_workers', self.num_workers) |
|
|
|
|
|
if cmd.get('distributed'): |
|
|
sampler = DistributedSampler( |
|
|
self.dataset, shuffle=self.shuffle, |
|
|
num_replicas=cmd['num_gpus']) |
|
|
batch_sampler = BatchSampler( |
|
|
sampler, self.batch_size//cmd['num_gpus'], False) |
|
|
torch.utils.data.DataLoader.__init__( |
|
|
self, self.dataset, batch_sampler=batch_sampler, |
|
|
num_workers=self.num_workers, pin_memory=False, |
|
|
drop_last=self.drop_last, collate_fn=self.collect_fn, |
|
|
worker_init_fn=default_worker_init_fn) |
|
|
else: |
|
|
torch.utils.data.DataLoader.__init__( |
|
|
self, self.dataset, |
|
|
batch_size=self.batch_size, num_workers=self.num_workers, |
|
|
drop_last=self.drop_last, shuffle=self.shuffle, |
|
|
pin_memory=True, collate_fn=self.collect_fn, |
|
|
worker_init_fn=default_worker_init_fn) |
|
|
self.collect_fn = str(self.collect_fn) |
|
|
|
|
|
|
|
|
class SuccessiveRandomSampler(Sampler): |
|
|
'''Random Sampler that yields sorted data in successive ranges. |
|
|
Args: |
|
|
dataset: Dataset used for sampling. |
|
|
''' |
|
|
def __init__(self, dataset): |
|
|
self.dataset = dataset |
|
|
self.epoch = 0 |
|
|
|
|
|
def __iter__(self): |
|
|
if self.shuffle: |
|
|
|
|
|
g = torch.Generator() |
|
|
g.manual_seed(self.epoch) |
|
|
indices = torch.randperm(len(self.dataset)).tolist() |
|
|
else: |
|
|
indices = torch.arange(len(self.dataset)).tolist() |
|
|
|
|
|
|
|
|
indices += indices[: (self.total_size - len(indices))] |
|
|
assert len(indices) == self.total_size |
|
|
|
|
|
|
|
|
offset = self.num_samples * self.rank |
|
|
indices = indices[offset: offset + self.num_samples] |
|
|
assert len(indices) == self.num_samples |
|
|
|
|
|
return iter(indices) |
|
|
|
|
|
def __len__(self): |
|
|
return len(self.dataset) |
|
|
|
|
|
def set_epoch(self, epoch): |
|
|
self.epoch = epoch |
|
|
|
|
|
|
|
|
class DistributedSampler(Sampler): |
|
|
"""Sampler that restricts data loading to a subset of the dataset. |
|
|
It is especially useful in conjunction with |
|
|
:class:`torch.nn.parallel.DistributedDataParallel`. In such case, each |
|
|
process can pass a DistributedSampler instance as a DataLoader sampler, |
|
|
and load a subset of the original dataset that is exclusive to it. |
|
|
.. note:: |
|
|
Dataset is assumed to be of constant size. |
|
|
Arguments: |
|
|
dataset: Dataset used for sampling. |
|
|
num_replicas (optional): Number of processes participating in |
|
|
distributed training. |
|
|
rank (optional): Rank of the current process within num_replicas. |
|
|
""" |
|
|
|
|
|
def __init__(self, dataset, num_replicas=None, rank=None, shuffle=True): |
|
|
if num_replicas is None: |
|
|
if not dist.is_available(): |
|
|
raise RuntimeError( |
|
|
"Requires distributed package to be available") |
|
|
num_replicas = dist.get_world_size() |
|
|
if rank is None: |
|
|
if not dist.is_available(): |
|
|
raise RuntimeError( |
|
|
"Requires distributed package to be available") |
|
|
rank = dist.get_rank() |
|
|
self.dataset = dataset |
|
|
self.num_replicas = num_replicas |
|
|
self.rank = rank |
|
|
self.epoch = 0 |
|
|
self.num_samples = int( |
|
|
math.ceil(len(self.dataset) * 1.0 / self.num_replicas)) |
|
|
self.total_size = self.num_samples * self.num_replicas |
|
|
self.shuffle = shuffle |
|
|
|
|
|
def __iter__(self): |
|
|
if self.shuffle: |
|
|
|
|
|
g = torch.Generator() |
|
|
g.manual_seed(self.epoch) |
|
|
indices = torch.randperm(len(self.dataset)).tolist() |
|
|
else: |
|
|
indices = torch.arange(len(self.dataset)).tolist() |
|
|
|
|
|
|
|
|
indices += indices[: (self.total_size - len(indices))] |
|
|
assert len(indices) == self.total_size |
|
|
|
|
|
|
|
|
offset = self.num_samples * self.rank |
|
|
indices = indices[offset: offset + self.num_samples] |
|
|
assert len(indices) == self.num_samples |
|
|
|
|
|
return iter(indices) |
|
|
|
|
|
def __len__(self): |
|
|
return self.num_samples |
|
|
|
|
|
def set_epoch(self, epoch): |
|
|
self.epoch = epoch |
|
|
|
|
|
|
|
|
class InfiniteOrderedSampler(Sampler): |
|
|
def __init__(self, data_source, limit_size): |
|
|
self.data_source = data_source |
|
|
self.limit_size = limit_size |
|
|
|
|
|
def __iter__(self): |
|
|
n = len(self.data_source) |
|
|
|
|
|
def wrapper(): |
|
|
cnt = 0 |
|
|
while cnt < self.limit_size: |
|
|
if cnt % n == 0: |
|
|
idx = torch.randperm(n).tolist() |
|
|
yield idx[cnt % n] |
|
|
cnt += 1 |
|
|
return wrapper() |
|
|
|
|
|
def __len__(self): |
|
|
return self.limit_size |
|
|
|
|
|
|
|
|
class InfiniteDataLoader(Configurable, torch.utils.data.DataLoader): |
|
|
dataset = State() |
|
|
batch_size = State(default=256) |
|
|
num_workers = State(default=10) |
|
|
limit_size = State(default=2 ** 31) |
|
|
|
|
|
def __init__(self, **kwargs): |
|
|
self.load_all(**kwargs) |
|
|
|
|
|
cmd = kwargs['cmd'] |
|
|
if 'batch_size' in cmd: |
|
|
self.batch_size = cmd['batch_size'] |
|
|
|
|
|
sampler = InfiniteOrderedSampler(self.dataset, self.limit_size) |
|
|
|
|
|
torch.utils.data.DataLoader.__init__( |
|
|
self, self.dataset, |
|
|
batch_size=self.batch_size, num_workers=self.num_workers, |
|
|
sampler=sampler, worker_init_fn=default_worker_init_fn, |
|
|
) |
|
|
|
|
|
|
|
|
class RandomSampleSampler(Sampler): |
|
|
def __init__(self, data_source, weights=None, size=2 ** 31): |
|
|
self.data_source = data_source |
|
|
if weights is None: |
|
|
self.probabilities = np.full(len(data_source), 1 / len(data_source)) |
|
|
else: |
|
|
self.probabilities = np.array(weights) / np.sum(weights) |
|
|
self.cum_prob = np.cumsum(self.probabilities) |
|
|
self.size = size |
|
|
|
|
|
def __iter__(self): |
|
|
def wrapper(): |
|
|
for i in range(self.size): |
|
|
yield bisect.bisect(self.cum_prob, torch.rand(1)[0], hi=len(self.data_source) - 1) |
|
|
return wrapper() |
|
|
|
|
|
def __len__(self): |
|
|
return self.size |
|
|
|
|
|
|
|
|
class RandomSampleDataLoader(Configurable, torch.utils.data.DataLoader): |
|
|
datasets = State() |
|
|
weights = State() |
|
|
batch_size = State(default=256) |
|
|
num_workers = State(default=10) |
|
|
size = State(default=2 ** 31) |
|
|
|
|
|
def __init__(self, **kwargs): |
|
|
self.load_all(**kwargs) |
|
|
|
|
|
cmd = kwargs['cmd'] |
|
|
if 'batch_size' in cmd: |
|
|
self.batch_size = cmd['batch_size'] |
|
|
|
|
|
probs = [] |
|
|
for dataset, weight in zip(self.datasets, self.weights): |
|
|
probs.append(np.full(len(dataset), weight / len(dataset))) |
|
|
|
|
|
dataset = ConcatDataset(self.datasets) |
|
|
probs = np.concatenate(probs) |
|
|
assert(len(dataset) == len(probs)) |
|
|
|
|
|
sampler = RandomSampleSampler(dataset, probs, self.size) |
|
|
|
|
|
torch.utils.data.DataLoader.__init__( |
|
|
self, dataset, |
|
|
batch_size=self.batch_size, num_workers=self.num_workers, |
|
|
sampler=sampler, worker_init_fn=default_worker_init_fn, |
|
|
) |
|
|
|