| import io | |
| import os | |
| import torch | |
| import torch.distributed as dist | |
| _print = print | |
| def get_world_size(): return int(os.getenv('WORLD_SIZE', 1)) | |
| def get_rank(): return int(os.getenv('RANK', 0)) | |
| def get_local_rank(): return int(os.getenv('LOCAL_RANK', 0)) | |
| def is_dist(): | |
| return dist.is_available() and dist.is_initialized() and get_world_size() > 1 | |
| def print(*argc, all=False, **kwargs): | |
| if not is_dist(): | |
| _print(*argc, **kwargs) | |
| return | |
| if not all and get_local_rank() != 0: | |
| return | |
| output = io.StringIO() | |
| kwargs['end'] = '' | |
| kwargs['file'] = output | |
| kwargs['flush'] = True | |
| _print(*argc, **kwargs) | |
| s = output.getvalue() | |
| output.close() | |
| s = '[rank {}] {}'.format(dist.get_rank(), s) | |
| _print(s) | |
| def reduce_mean(tensor, nprocs=None): | |
| if not is_dist(): | |
| return tensor | |
| if not isinstance(tensor, torch.Tensor): | |
| device = torch.cuda.current_device() | |
| rt = torch.tensor(tensor, device=device) | |
| else: | |
| rt = tensor.clone() | |
| dist.all_reduce(rt, op=dist.ReduceOp.SUM) | |
| nprocs = nprocs if nprocs else dist.get_world_size() | |
| rt = rt / nprocs | |
| if not isinstance(tensor, torch.Tensor): | |
| rt = rt.item() | |
| return rt | |
| def reduce_sum(tensor): | |
| if not is_dist(): | |
| return tensor | |
| if not isinstance(tensor, torch.Tensor): | |
| device = torch.cuda.current_device() | |
| rt = torch.tensor(tensor, device=device) | |
| else: | |
| rt = tensor.clone() | |
| dist.all_reduce(rt, op=dist.ReduceOp.SUM) | |
| if not isinstance(tensor, torch.Tensor): | |
| rt = rt.item() | |
| return rt | |
| def barrier(): | |
| if not is_dist(): | |
| return | |
| dist.barrier() |