| | from torch.distributed.fsdp.wrap import transformer_auto_wrap_policy |
| | from torch.distributed.fsdp.wrap import size_based_auto_wrap_policy |
| | from torch.distributed.fsdp.wrap import ModuleWrapPolicy |
| | from torch.distributed.fsdp import ( |
| | FullStateDictConfig, |
| | StateDictType, |
| | MixedPrecision, |
| | ShardingStrategy, |
| | FullyShardedDataParallel as FSDP |
| | ) |
| | from torchvision.utils import make_grid |
| | from datetime import timedelta, datetime |
| | import torch.distributed as dist |
| | from omegaconf import OmegaConf |
| | from functools import partial |
| | import numpy as np |
| | import random |
| | import torch |
| | import wandb |
| | import os |
| |
|
| |
|
| | def launch_distributed_job(backend: str = "nccl"): |
| | rank = int(os.environ["RANK"]) |
| | local_rank = int(os.environ["LOCAL_RANK"]) |
| | world_size = int(os.environ["WORLD_SIZE"]) |
| | host = os.environ["MASTER_ADDR"] |
| | port = int(os.environ["MASTER_PORT"]) |
| |
|
| | if ":" in host: |
| | init_method = f"tcp://[{host}]:{port}" |
| | else: |
| | init_method = f"tcp://{host}:{port}" |
| | dist.init_process_group(rank=rank, world_size=world_size, backend=backend, |
| | init_method=init_method, timeout=timedelta(minutes=30)) |
| | torch.cuda.set_device(local_rank) |
| |
|
| |
|
| | def set_seed(seed: int, deterministic: bool = False): |
| | """ |
| | Helper function for reproducible behavior to set the seed in `random`, `numpy`, `torch`. |
| | |
| | Args: |
| | seed (`int`): |
| | The seed to set. |
| | deterministic (`bool`, *optional*, defaults to `False`): |
| | Whether to use deterministic algorithms where available. Can slow down training. |
| | """ |
| | random.seed(seed) |
| | np.random.seed(seed) |
| | torch.manual_seed(seed) |
| | torch.cuda.manual_seed_all(seed) |
| |
|
| | if deterministic: |
| | torch.use_deterministic_algorithms(True) |
| |
|
| |
|
| | def init_logging_folder(args): |
| | date = str(datetime.now()).replace(" ", "-").replace(":", "-") |
| | output_path = os.path.join( |
| | args.output_path, |
| | f"{date}_seed{args.seed}" |
| | ) |
| | os.makedirs(output_path, exist_ok=False) |
| |
|
| | os.makedirs(args.output_path, exist_ok=True) |
| | wandb.login(host=args.wandb_host, key=args.wandb_key) |
| | run = wandb.init(config=OmegaConf.to_container(args, resolve=True), dir=args.output_path, ** |
| | {"mode": "online", "entity": args.wandb_entity, "project": args.wandb_project}) |
| | wandb.run.log_code(".") |
| | wandb.run.name = args.wandb_name |
| | print(f"run dir: {run.dir}") |
| | wandb_folder = run.dir |
| | os.makedirs(wandb_folder, exist_ok=True) |
| |
|
| | return output_path, wandb_folder |
| |
|
| |
|
| | def fsdp_wrap(module, sharding_strategy="full", mixed_precision=False, wrap_strategy="size", min_num_params=int(5e7), transformer_module=None): |
| | if mixed_precision: |
| | mixed_precision_policy = MixedPrecision( |
| | param_dtype=torch.bfloat16, |
| | reduce_dtype=torch.float32, |
| | buffer_dtype=torch.float32, |
| | cast_forward_inputs=False |
| | ) |
| | else: |
| | mixed_precision_policy = None |
| |
|
| | if wrap_strategy == "transformer": |
| | auto_wrap_policy = partial( |
| | transformer_auto_wrap_policy, |
| | transformer_layer_cls=transformer_module |
| | ) |
| | elif wrap_strategy == "size": |
| | auto_wrap_policy = partial( |
| | size_based_auto_wrap_policy, |
| | min_num_params=min_num_params |
| | ) |
| | else: |
| | raise ValueError(f"Invalid wrap strategy: {wrap_strategy}") |
| |
|
| | os.environ["NCCL_CROSS_NIC"] = "1" |
| |
|
| | sharding_strategy = { |
| | "full": ShardingStrategy.FULL_SHARD, |
| | "hybrid_full": ShardingStrategy.HYBRID_SHARD, |
| | "hybrid_zero2": ShardingStrategy._HYBRID_SHARD_ZERO2, |
| | "no_shard": ShardingStrategy.NO_SHARD, |
| | }[sharding_strategy] |
| |
|
| | module = FSDP( |
| | module, |
| | auto_wrap_policy=auto_wrap_policy, |
| | sharding_strategy=sharding_strategy, |
| | mixed_precision=mixed_precision_policy, |
| | device_id=torch.cuda.current_device(), |
| | limit_all_gathers=True, |
| | sync_module_states=False |
| | ) |
| | return module |
| |
|
| |
|
| | def cycle(dl): |
| | while True: |
| | for data in dl: |
| | yield data |
| |
|
| |
|
| | def fsdp_state_dict(model): |
| | fsdp_fullstate_save_policy = FullStateDictConfig( |
| | offload_to_cpu=True, rank0_only=True |
| | ) |
| | with FSDP.state_dict_type( |
| | model, StateDictType.FULL_STATE_DICT, fsdp_fullstate_save_policy |
| | ): |
| | checkpoint = model.state_dict() |
| |
|
| | return checkpoint |
| |
|
| |
|
| | def barrier(): |
| | if dist.is_initialized(): |
| | dist.barrier() |
| |
|
| |
|
| | def prepare_for_saving(tensor, fps=16, caption=None): |
| | |
| | tensor = (tensor * 0.5 + 0.5).clamp(0, 1).detach() |
| |
|
| | if tensor.ndim == 4: |
| | |
| | tensor = make_grid(tensor, 4, padding=0, normalize=False) |
| | return wandb.Image((tensor * 255).cpu().numpy().astype(np.uint8), caption=caption) |
| | elif tensor.ndim == 5: |
| | |
| | return wandb.Video((tensor * 255).cpu().numpy().astype(np.uint8), fps=fps, format="webm", caption=caption) |
| | else: |
| | raise ValueError("Unsupported tensor shape for saving. Expected 4D (image) or 5D (video) tensor.") |
| |
|