| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """ |
| | the class for Worker |
| | """ |
| |
|
| | import os |
| | import socket |
| | from dataclasses import dataclass |
| | from typing import Tuple |
| |
|
| | import ray |
| | import torch |
| |
|
| | from .decorator import Dispatch, Execute, register |
| | from .register_center.ray import create_worker_group_register_center |
| |
|
| |
|
| | @dataclass |
| | class DistRankInfo: |
| | tp_rank: int |
| | dp_rank: int |
| | pp_rank: int |
| |
|
| |
|
| | @dataclass |
| | class DistGlobalInfo: |
| | tp_size: int |
| | dp_size: int |
| | pp_size: int |
| |
|
| |
|
| | class WorkerHelper: |
| | def _get_node_ip(self) -> str: |
| | host_ipv4 = os.getenv("MY_HOST_IP", None) |
| | host_ipv6 = os.getenv("MY_HOST_IPV6", None) |
| | host_ip_by_env = host_ipv4 or host_ipv6 |
| | host_ip_by_sdk = ray._private.services.get_node_ip_address() |
| |
|
| | host_ip = host_ip_by_env or host_ip_by_sdk |
| | return host_ip |
| |
|
| | def _get_free_port(self) -> int: |
| | with socket.socket() as sock: |
| | sock.bind(("", 0)) |
| | return sock.getsockname()[1] |
| |
|
| | def get_availale_master_addr_port(self) -> Tuple[str, str]: |
| | return self._get_node_ip(), str(self._get_free_port()) |
| |
|
| | def _get_pid(self): |
| | return |
| |
|
| |
|
| | class WorkerMeta: |
| | keys = [ |
| | "WORLD_SIZE", |
| | "RANK", |
| | "LOCAL_WORLD_SIZE", |
| | "LOCAL_RANK", |
| | "MASTER_ADDR", |
| | "MASTER_PORT", |
| | "CUDA_VISIBLE_DEVICES", |
| | ] |
| |
|
| | def __init__(self, store) -> None: |
| | self._store = store |
| |
|
| | def to_dict(self): |
| | return {f"_{key.lower()}": self._store.get(f"_{key.lower()}", None) for key in WorkerMeta.keys} |
| |
|
| |
|
| | |
| | class Worker(WorkerHelper): |
| | """A (distributed) worker.""" |
| |
|
| | _world_size: int |
| | _rank: int |
| | _local_world_size: int |
| | _local_rank: int |
| | _master_addr: str |
| | _master_port: str |
| | _cuda_visible_devices: str |
| |
|
| | def __new__(cls, *args, **kwargs): |
| | instance = super().__new__(cls) |
| |
|
| | |
| | disable_worker_init = int(os.getenv("DISABLE_WORKER_INIT", 0)) |
| | if disable_worker_init: |
| | return instance |
| |
|
| | rank = os.getenv("RANK", None) |
| | worker_group_prefix = os.getenv("WG_PREFIX", None) |
| |
|
| | |
| | if None not in [rank, worker_group_prefix] and "ActorClass(" not in cls.__name__: |
| | instance._configure_before_init(f"{worker_group_prefix}_register_center", int(rank)) |
| |
|
| | return instance |
| |
|
| | def _configure_before_init(self, register_center_name: str, rank: int): |
| | assert isinstance(rank, int), f"rank must be int, instead of {type(rank)}" |
| |
|
| | if rank == 0: |
| | master_addr, master_port = self.get_availale_master_addr_port() |
| | rank_zero_info = { |
| | "MASTER_ADDR": master_addr, |
| | "MASTER_PORT": master_port, |
| | } |
| | self.register_center = create_worker_group_register_center(name=register_center_name, info=rank_zero_info) |
| | os.environ.update(rank_zero_info) |
| |
|
| | def __init__(self, cuda_visible_devices=None) -> None: |
| | |
| | world_size = int(os.getenv("WORLD_SIZE")) |
| | rank = int(os.getenv("RANK")) |
| | self._rank = rank |
| | self._world_size = world_size |
| |
|
| | if "AMD" in torch.cuda.get_device_name(): |
| | os.environ["CUDA_VISIBLE_DEVICES"] = os.getenv("ROCR_VISIBLE_DEVICES") |
| | os.environ["LOCAL_RANK"] = os.getenv("RAY_LOCAL_RANK") |
| | cuda_visible_devices = os.getenv("LOCAL_RANK", "0") |
| | torch.cuda.set_device(int(cuda_visible_devices)) |
| |
|
| | master_addr = os.getenv("MASTER_ADDR") |
| | master_port = os.getenv("MASTER_PORT") |
| |
|
| | local_world_size = int(os.getenv("LOCAL_WORLD_SIZE", "1")) |
| | local_rank = int(os.getenv("LOCAL_RANK", "0")) |
| |
|
| | store = { |
| | "_world_size": world_size, |
| | "_rank": rank, |
| | "_local_world_size": local_world_size, |
| | "_local_rank": local_rank, |
| | "_master_addr": master_addr, |
| | "_master_port": master_port, |
| | } |
| | if cuda_visible_devices is not None: |
| | store["_cuda_visible_devices"] = cuda_visible_devices |
| |
|
| | meta = WorkerMeta(store=store) |
| | self._configure_with_meta(meta=meta) |
| |
|
| | def _configure_with_meta(self, meta: WorkerMeta): |
| | """ |
| | This function should only be called inside by WorkerGroup |
| | """ |
| | assert isinstance(meta, WorkerMeta) |
| | self.__dict__.update(meta.to_dict()) |
| | |
| | for key in WorkerMeta.keys: |
| | val = self.__dict__.get(f"_{key.lower()}", None) |
| | if val is not None: |
| | |
| | os.environ[key] = str(val) |
| |
|
| | os.environ["REDIS_STORE_SERVER_HOST"] = ( |
| | str(self._master_addr).replace("[", "").replace("]", "") if self._master_addr else "" |
| | ) |
| |
|
| | def get_master_addr_port(self): |
| | return self._master_addr, self._master_port |
| |
|
| | def get_cuda_visible_devices(self): |
| | cuda_visible_devices = os.getenv("CUDA_VISIBLE_DEVICES", "not set") |
| | return cuda_visible_devices |
| |
|
| | def print_rank0(self, *args, **kwargs): |
| | if self.rank == 0: |
| | print(*args, **kwargs) |
| |
|
| | @property |
| | def world_size(self): |
| | return self._world_size |
| |
|
| | @property |
| | def rank(self): |
| | return self._rank |
| |
|
| | @register(dispatch_mode=Dispatch.DP_COMPUTE_PROTO_WITH_FUNC) |
| | def execute_with_func_generator(self, func, *args, **kwargs): |
| | ret_proto = func(self, *args, **kwargs) |
| | return ret_proto |
| |
|
| | @register(dispatch_mode=Dispatch.ALL_TO_ALL, execute_mode=Execute.RANK_ZERO) |
| | def execute_func_rank_zero(self, func, *args, **kwargs): |
| | result = func(*args, **kwargs) |
| | return result |
| |
|