Masond / jukebox /utils /dist_utils.py
MasonCrinr's picture
Upload 331 files
814c8cf
import os
from time import sleep
import torch
import jukebox.utils.dist_adapter as dist
def print_once(msg):
if (not dist.is_available()) or dist.get_rank()==0:
print(msg)
def print_all(msg):
if (not dist.is_available()):
print(msg)
elif dist.get_rank()%8==0:
print(f'{dist.get_rank()//8}: {msg}')
def allgather(x):
xs = [torch.empty_like(x) for _ in range(dist.get_world_size())]
dist.all_gather(xs, x)
xs = torch.cat(xs, dim=0)
return xs
def allreduce(x, op=dist.ReduceOp.SUM):
x = torch.tensor(x).float().cuda()
dist.all_reduce(x, op=op)
return x.item()
def allgather_lists(xs):
bs = len(xs)
total_bs = dist.get_world_size()*len(xs)
lengths = torch.tensor([len(x) for x in xs], dtype=t.long, device='cuda')
lengths = allgather(lengths)
assert lengths.shape == (total_bs,)
max_length = torch.max(lengths).item()
xs = torch.tensor([[*x, *[0]*(max_length - len(x))] for x in xs], device='cuda')
assert xs.shape == (bs, max_length), f'Expected {(bs, max_length)}, got {xs.shape}'
xs = allgather(xs)
assert xs.shape == (total_bs,max_length), f'Expected {(total_bs, max_length)}, got {xs.shape}'
return [xs[i][:lengths[i]].cpu().numpy().tolist() for i in range(total_bs)]
def setup_dist_from_mpi(
master_addr="127.0.0.1", backend="nccl", port=29500, n_attempts=5, verbose=False
):
if dist.is_available():
return _setup_dist_from_mpi(master_addr, backend, port, n_attempts, verbose)
else:
use_cuda = torch.cuda.is_available()
print(f'Using cuda {use_cuda}')
mpi_rank = 0
local_rank = 0
device = torch.device("cuda", local_rank) if use_cuda else torch.device("cpu")
torch.cuda.set_device(local_rank)
return mpi_rank, local_rank, device
def _setup_dist_from_mpi(master_addr, backend, port, n_attempts, verbose):
from mpi4py import MPI # This must be imported in order to get e rrors from all ranks to show up
mpi_rank = MPI.COMM_WORLD.Get_rank()
mpi_size = MPI.COMM_WORLD.Get_size()
os.environ["RANK"] = str(mpi_rank)
os.environ["WORLD_SIZE"] = str(mpi_size)
os.environ["MASTER_ADDR"] = master_addr
os.environ["MASTER_PORT"] = str(port)
os.environ["NCCL_LL_THRESHOLD"] = "0"
os.environ["NCCL_NSOCKS_PERTHREAD"] = "2"
os.environ["NCCL_SOCKET_NTHREADS"] = "8"
# Pin this rank to a specific GPU on the node
local_rank = mpi_rank % 8
if torch.cuda.is_available():
torch.cuda.set_device(local_rank)
if verbose:
print(f"Connecting to master_addr: {master_addr}")
# There is a race condition when initializing NCCL with a large number of ranks (e.g 500 ranks)
# We guard against the failure and then retry
for attempt_idx in range(n_attempts):
try:
dist.init_process_group(backend=backend, init_method=f"env://")
assert dist.get_rank() == mpi_rank
use_cuda = torch.cuda.is_available()
print(f'Using cuda {use_cuda}')
local_rank = mpi_rank % 8
device = torch.device("cuda", local_rank) if use_cuda else torch.device("cpu")
torch.cuda.set_device(local_rank)
return mpi_rank, local_rank, device
except RuntimeError as e:
print(f"Caught error during NCCL init (attempt {attempt_idx} of {n_attempts}): {e}")
sleep(1 + (0.01 * mpi_rank)) # Sleep to avoid thundering herd
pass
raise RuntimeError("Failed to initialize NCCL")