|
|
import torch |
|
|
import math |
|
|
import numpy as np |
|
|
|
|
|
class SequentialDistributedSampler(torch.utils.data.sampler.Sampler): |
|
|
""" |
|
|
Distributed Sampler that subsamples indicies sequentially, |
|
|
making it easier to collate all results at the end. |
|
|
Even though we only use this sampler for eval and predict (no training), |
|
|
which means that the model params won't have to be synced (i.e. will not hang |
|
|
for synchronization even if varied number of forward passes), we still add extra |
|
|
samples to the sampler to make it evenly divisible (like in `DistributedSampler`) |
|
|
to make it easy to `gather` or `reduce` resulting tensors at the end of the loop. |
|
|
""" |
|
|
|
|
|
def __init__(self, dataset, batch_size, rank=None, num_replicas=None): |
|
|
if num_replicas is None: |
|
|
if not torch.distributed.is_available(): |
|
|
raise RuntimeError("Requires distributed package to be available") |
|
|
num_replicas = torch.distributed.get_world_size() |
|
|
if rank is None: |
|
|
if not torch.distributed.is_available(): |
|
|
raise RuntimeError("Requires distributed package to be available") |
|
|
rank = torch.distributed.get_rank() |
|
|
self.dataset = dataset |
|
|
self.num_replicas = num_replicas |
|
|
self.rank = rank |
|
|
self.batch_size = batch_size |
|
|
self.num_samples = int(math.ceil(len(self.dataset) * 1.0 / self.batch_size / self.num_replicas)) * self.batch_size |
|
|
self.total_size = self.num_samples * self.num_replicas |
|
|
|
|
|
def __iter__(self): |
|
|
indices = list(range(len(self.dataset))) |
|
|
|
|
|
indices += [indices[-1]] * (self.total_size - len(indices)) |
|
|
|
|
|
indices = indices[self.rank * self.num_samples : (self.rank + 1) * self.num_samples] |
|
|
return iter(indices) |
|
|
|
|
|
def __len__(self): |
|
|
return self.num_samples |
|
|
|
|
|
|
|
|
def distributed_concat(tensor, num_total_examples): |
|
|
output_tensors = [tensor.clone() for _ in range(torch.distributed.get_world_size())] |
|
|
torch.distributed.all_gather(output_tensors, tensor) |
|
|
concat = torch.cat(output_tensors, dim=0) |
|
|
return concat[:num_total_examples] |