| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| """ |
| This example shows how to execute distributed training based on PyTorch native `DistributedDataParallel` module. |
| It can run on several nodes with multiple GPU devices on every node. |
| Main steps to set up the distributed training: |
| |
| - Execute `torch.distributed.launch` to create processes on every node for every GPU. |
| It receives parameters as below: |
| `--nproc_per_node=NUM_GPUS_PER_NODE` |
| `--nnodes=NUM_NODES` |
| `--node_rank=INDEX_CURRENT_NODE` |
| `--master_addr="192.168.1.1"` |
| `--master_port=1234` |
| For more details, refer to https://github.com/pytorch/pytorch/blob/master/torch/distributed/launch.py. |
| Alternatively, we can also use `torch.multiprocessing.spawn` to start program, but it that case, need to handle |
| all the above parameters and compute `rank` manually, then set to `init_process_group`, etc. |
| `torch.distributed.launch` is even more efficient than `torch.multiprocessing.spawn` during training. |
| - Use `init_process_group` to initialize every process, every GPU runs in a separate process with unique rank. |
| Here we use `NVIDIA NCCL` as the backend and must set `init_method="env://"` if use `torch.distributed.launch`. |
| - Wrap the model with `DistributedDataParallel` after moving to expected device. |
| - Wrap Dataset with `DistributedSampler`, and disable the `shuffle` in DataLoader. |
| Instead, shuffle data by `train_sampler.set_epoch(epoch)` before every epoch. |
| |
| Note: |
| `torch.distributed.launch` will launch `nnodes * nproc_per_node = world_size` processes in total. |
| Suggest setting exactly the same software environment for every node, especially `PyTorch`, `nccl`, etc. |
| A good practice is to use the same MONAI docker image for all nodes directly. |
| Example script to execute this program on every node: |
| python -m torch.distributed.launch --nproc_per_node=NUM_GPUS_PER_NODE |
| --nnodes=NUM_NODES --node_rank=INDEX_CURRENT_NODE |
| --master_addr="192.168.1.1" --master_port=1234 |
| unet_training_ddp.py -d DIR_OF_TESTDATA |
| |
| This example was tested with [Ubuntu 16.04/20.04], [NCCL 2.6.3]. |
| |
| Referring to: https://pytorch.org/tutorials/intermediate/ddp_tutorial.html |
| |
| """ |
|
|
| import argparse |
| import os |
| import sys |
| from glob import glob |
|
|
| import nibabel as nib |
| import numpy as np |
| import torch |
| import torch.distributed as dist |
| from torch.nn.parallel import DistributedDataParallel |
| from torch.utils.data.distributed import DistributedSampler |
|
|
| import monai |
| from monai.data import DataLoader, Dataset, create_test_image_3d |
| from monai.transforms import ( |
| AsChannelFirstd, |
| Compose, |
| LoadNiftid, |
| RandCropByPosNegLabeld, |
| RandRotate90d, |
| ScaleIntensityd, |
| ToTensord, |
| ) |
|
|
|
|
| def train(args): |
| |
| if args.local_rank != 0: |
| f = open(os.devnull, "w") |
| sys.stdout = sys.stderr = f |
| elif not os.path.exists(args.dir): |
| |
| print(f"generating synthetic data to {args.dir} (this may take a while)") |
| os.makedirs(args.dir) |
| |
| np.random.seed(seed=0) |
| for i in range(40): |
| im, seg = create_test_image_3d(128, 128, 128, num_seg_classes=1, channel_dim=-1) |
| n = nib.Nifti1Image(im, np.eye(4)) |
| nib.save(n, os.path.join(args.dir, f"img{i:d}.nii.gz")) |
| n = nib.Nifti1Image(seg, np.eye(4)) |
| nib.save(n, os.path.join(args.dir, f"seg{i:d}.nii.gz")) |
|
|
| |
| dist.init_process_group(backend="nccl", init_method="env://") |
|
|
| images = sorted(glob(os.path.join(args.dir, "img*.nii.gz"))) |
| segs = sorted(glob(os.path.join(args.dir, "seg*.nii.gz"))) |
| train_files = [{"img": img, "seg": seg} for img, seg in zip(images, segs)] |
|
|
| |
| train_transforms = Compose( |
| [ |
| LoadNiftid(keys=["img", "seg"]), |
| AsChannelFirstd(keys=["img", "seg"], channel_dim=-1), |
| ScaleIntensityd(keys="img"), |
| RandCropByPosNegLabeld( |
| keys=["img", "seg"], label_key="seg", spatial_size=[96, 96, 96], pos=1, neg=1, num_samples=4 |
| ), |
| RandRotate90d(keys=["img", "seg"], prob=0.5, spatial_axes=[0, 2]), |
| ToTensord(keys=["img", "seg"]), |
| ] |
| ) |
|
|
| |
| train_ds = Dataset(data=train_files, transform=train_transforms) |
| |
| train_sampler = DistributedSampler(train_ds) |
| |
| train_loader = DataLoader( |
| train_ds, |
| batch_size=2, |
| shuffle=False, |
| num_workers=2, |
| pin_memory=True, |
| sampler=train_sampler, |
| ) |
|
|
| |
| device = torch.device(f"cuda:{args.local_rank}") |
| model = monai.networks.nets.UNet( |
| dimensions=3, |
| in_channels=1, |
| out_channels=1, |
| channels=(16, 32, 64, 128, 256), |
| strides=(2, 2, 2, 2), |
| num_res_units=2, |
| ).to(device) |
| loss_function = monai.losses.DiceLoss(sigmoid=True).to(device) |
| optimizer = torch.optim.Adam(model.parameters(), 1e-3) |
| |
| model = DistributedDataParallel(model, device_ids=[args.local_rank]) |
|
|
| |
| epoch_loss_values = list() |
| for epoch in range(5): |
| print("-" * 10) |
| print(f"epoch {epoch + 1}/{5}") |
| model.train() |
| epoch_loss = 0 |
| step = 0 |
| train_sampler.set_epoch(epoch) |
| for batch_data in train_loader: |
| step += 1 |
| inputs, labels = batch_data["img"].to(device), batch_data["seg"].to(device) |
| optimizer.zero_grad() |
| outputs = model(inputs) |
| loss = loss_function(outputs, labels) |
| loss.backward() |
| optimizer.step() |
| epoch_loss += loss.item() |
| epoch_len = len(train_ds) // train_loader.batch_size |
| print(f"{step}/{epoch_len}, train_loss: {loss.item():.4f}") |
| epoch_loss /= step |
| epoch_loss_values.append(epoch_loss) |
| print(f"epoch {epoch + 1} average loss: {epoch_loss:.4f}") |
| print(f"train completed, epoch losses: {epoch_loss_values}") |
| if dist.get_rank() == 0: |
| |
| |
| |
| torch.save(model.state_dict(), "final_model.pth") |
| dist.destroy_process_group() |
|
|
|
|
| def main(): |
| parser = argparse.ArgumentParser() |
| parser.add_argument("-d", "--dir", default="./testdata", type=str, help="directory to create random data") |
| |
| parser.add_argument("--local_rank", type=int) |
| args = parser.parse_args() |
|
|
| train(args=args) |
|
|
|
|
| |
|
|
| |
| |
| |
| |
|
|
| if __name__ == "__main__": |
| main() |
|
|