braindeck
Initial commit
bcdf9fa
# Copyright 2024 Bytedance Ltd. and/or its affiliates
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import os
from importlib.metadata import PackageNotFoundError, version
from packaging.version import Version
def get_version(pkg):
try:
return version(pkg)
except PackageNotFoundError:
return None
vllm_package_name = "vllm"
vllm_package_version = get_version(vllm_package_name)
if vllm_package_version is None:
raise PackageNotFoundError("To use vllm rollout, please ensure the 'vllm' package is properly installed. See https://verl.readthedocs.io/en/latest/start/install.html for more details")
###
# package_version = get_version(package_name)
# [SUPPORT AMD:]
# Do not call any torch.cuda* API here, or ray actor creation import class will fail.
if "ROCM_PATH" in os.environ:
import re
match = re.match(r"(\d+\.\d+\.?\d*)", vllm_package_version)
if match:
vllm_package_version = match.group(1)
else:
raise ValueError(f"Warning: Could not parse version format: {vllm_package_version}")
###
if Version(vllm_package_version) <= Version("0.6.3"):
vllm_mode = "customized"
from .fire_vllm_rollout import FIREvLLMRollout # noqa: F401
from .vllm_rollout import vLLMRollout # noqa: F401
else:
vllm_mode = "spmd"
from .vllm_rollout_spmd import vLLMAsyncRollout, vLLMRollout # noqa: F401