| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | import os |
| | import sys |
| |
|
| | import torch |
| |
|
| | from ..utils import is_torch_version |
| | from .dataclasses import DistributedType |
| |
|
| |
|
| | def get_launch_prefix(): |
| | """ |
| | Grabs the correct launcher for starting a distributed command, such as either `torchrun`, `python -m |
| | torch.distributed.run`, etc |
| | """ |
| | if is_torch_version(">=", "1.10.0"): |
| | cmd = ["torchrun"] |
| | elif is_torch_version(">=", "1.9.0"): |
| | cmd = [sys.executable, "-m", "torch.distributed.run"] |
| | else: |
| | cmd = [sys.executable, "-m", "torch.distributed.launch", "--use_env"] |
| | return cmd |
| |
|
| |
|
| | def _filter_args(args): |
| | """ |
| | Filters out all `accelerate` specific args |
| | """ |
| | if is_torch_version(">=", "1.9.1"): |
| | import torch.distributed.run as distrib_run |
| | distrib_args = distrib_run.get_args_parser() |
| | new_args, _ = distrib_args.parse_known_args() |
| |
|
| | for key, value in vars(args).items(): |
| | if key in vars(new_args).keys(): |
| | setattr(new_args, key, value) |
| | return new_args |
| |
|
| |
|
| | def env_var_path_add(env_var_name, path_to_add): |
| | """ |
| | Extends a path-based environment variable's value with a new path and returns the updated value. It's up to the |
| | caller to set it in os.environ. |
| | """ |
| | paths = [p for p in os.environ.get(env_var_name, "").split(":") if len(p) > 0] |
| | paths.append(str(path_to_add)) |
| | return ":".join(paths) |
| |
|
| |
|
| | class PrepareForLaunch: |
| | """ |
| | Prepare a function that will launched in a distributed setup. |
| | |
| | Args: |
| | launcher (`Callable`): |
| | The function to launch. |
| | distributed_type ([`~state.DistributedType`]): |
| | The distributed type to prepare for. |
| | debug (`bool`, *optional*, defaults to `False`): |
| | Whether or not this is a debug launch. |
| | """ |
| |
|
| | def __init__(self, launcher, distributed_type="NO", debug=False): |
| | self.launcher = launcher |
| | self.distributed_type = DistributedType(distributed_type) |
| | self.debug = debug |
| |
|
| | def __call__(self, index, *args): |
| | if self.debug: |
| | world_size = int(os.environ.get("WORLD_SIZE")) |
| | rdv_file = os.environ.get("ACCELERATE_DEBUG_RDV_FILE") |
| | torch.distributed.init_process_group( |
| | "gloo", |
| | rank=index, |
| | store=torch.distributed.FileStore(rdv_file, world_size), |
| | world_size=world_size, |
| | ) |
| | elif self.distributed_type in (DistributedType.MULTI_GPU, DistributedType.MULTI_CPU): |
| | |
| | os.environ["LOCAL_RANK"] = str(index) |
| | os.environ["RANK"] = str(index) |
| |
|
| | os.environ["FORK_LAUNCHED"] = str(1) |
| | self.launcher(*args) |
| |
|