File size: 3,616 Bytes
b4d7ac8 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 |
# Copyright 2020 The Microsoft DeepSpeed Team
"""
sailing runner is the main front-end to launching multi-worker
training jobs with DeepSpeed. By default this uses pdsh to parallel
ssh into multiple worker nodes and launch all the necessary processes
per rank for training.
"""
import os
import sys
import json
import subprocess
import collections
import socket
import signal
import logging
import torch.distributed as dist
def fetch_hostfile(hostfile_path):
if not os.path.isfile(hostfile_path):
print("Unable to find hostfile, will proceed with training "
"with local resources only.")
return None
# e.g., worker-0 slots=16
with open(hostfile_path, 'r') as fd:
resource_pool = collections.OrderedDict()
for line in fd.readlines():
line = line.strip()
if line == '':
# skip empty lines
continue
try:
hostname, slots = line.split()
_, slot_count = slots.split("=")
slot_count = int(slot_count)
except ValueError as err:
raise err
if hostname in resource_pool:
raise ValueError(f"host {hostname} is already defined")
resource_pool[hostname] = slot_count
return resource_pool
def cmd_load_hyperparam(config_path=None, format="json", encoding="utf-8"):
"""
shell load arguments form argparse and config file
"""
# config_path='config/config_block_large_chinese.json'
format = config_path.rsplit('.')[-1]
with open(config_path, 'r', encoding=encoding) as f:
if format == "json":
config_dict = json.load(f)
else:
raise NameError("current format%s for hyperparam file is invalid" %
format)
config_cmd = []
for key in config_dict:
if len(str(config_dict[key])) == 0:
config_cmd.append('--' + key)
else:
config_cmd.append('--' + key)
config_cmd.append(str(config_dict[key]))
return config_cmd
def launch_dist(
env_type="DDP",
num_nodes=1,
gpus_per_node=1,
master_addr='localhost',
master_port=17500,
training_script='train.py',
):
if num_nodes != 1:
print("多机多卡待测试。暂不支持。")
os._exit(0)
if env_type == "DDP":
cmd_launch = []
cmd_launch.extend([
# 'export NUM_NODES=' + str(num_nodes) + ';',
# 'export GPUS_PER_NODE=' + str(gpus_per_node) + ';',
# sys.executable,
# "python",
# '-m',
"torchrun"
# 'torch.distributed.launch'
])
torch_distributed_args = [
'--nproc_per_node',
str(gpus_per_node),
'--nnodes',
str(num_nodes),
'--node_rank',
str(0),
'--master_addr',
master_addr,
'--master_port',
str(master_port),
]
cmd_launch.extend(torch_distributed_args)
cmd_launch.append(training_script)
cmd_launch.append('--not_call_launch')
run_cmd = ' '.join(cmd_launch)
p = subprocess.Popen(run_cmd, shell=True, preexec_fn=os.setsid)
def signal_handler(signal, frame):
os.killpg(os.getpgid(p.pid), 9)
signal.signal(signal.SIGINT, signal_handler)
p.wait()
print ('finish')
else :
print("不支持的env_type")
os._exit(0)
|