|
|
import logging |
|
|
import os |
|
|
from pathlib import Path |
|
|
from typing import Dict, Iterable, Iterator |
|
|
|
|
|
import torch |
|
|
|
|
|
from bitsandbytes.cextension import get_cuda_bnb_library_path |
|
|
from bitsandbytes.consts import NONPYTORCH_DOC_URL |
|
|
from bitsandbytes.cuda_specs import CUDASpecs |
|
|
from bitsandbytes.diagnostics.utils import print_dedented |
|
|
|
|
|
CUDART_PATH_PREFERRED_ENVVARS = ("CONDA_PREFIX", "LD_LIBRARY_PATH") |
|
|
|
|
|
CUDART_PATH_IGNORED_ENVVARS = { |
|
|
"DBUS_SESSION_BUS_ADDRESS", |
|
|
"GOOGLE_VM_CONFIG_LOCK_FILE", |
|
|
"HOME", |
|
|
"LESSCLOSE", |
|
|
"LESSOPEN", |
|
|
"MAIL", |
|
|
"OLDPWD", |
|
|
"PATH", |
|
|
"PWD", |
|
|
"SHELL", |
|
|
"SSH_AUTH_SOCK", |
|
|
"SSH_TTY", |
|
|
"TMUX", |
|
|
"XDG_DATA_DIRS", |
|
|
"XDG_GREETER_DATA_DIR", |
|
|
"XDG_RUNTIME_DIR", |
|
|
"_", |
|
|
} |
|
|
|
|
|
CUDA_RUNTIME_LIB_PATTERNS = ( |
|
|
"cudart64*.dll", |
|
|
"libcudart*.so*", |
|
|
"nvcuda*.dll", |
|
|
) |
|
|
|
|
|
logger = logging.getLogger(__name__) |
|
|
|
|
|
|
|
|
def find_cuda_libraries_in_path_list(paths_list_candidate: str) -> Iterable[Path]: |
|
|
for dir_string in paths_list_candidate.split(os.pathsep): |
|
|
if not dir_string: |
|
|
continue |
|
|
if os.sep not in dir_string: |
|
|
continue |
|
|
try: |
|
|
dir = Path(dir_string) |
|
|
try: |
|
|
if not dir.exists(): |
|
|
logger.warning(f"The directory listed in your path is found to be non-existent: {dir}") |
|
|
continue |
|
|
except OSError: |
|
|
pass |
|
|
for lib_pattern in CUDA_RUNTIME_LIB_PATTERNS: |
|
|
for pth in dir.glob(lib_pattern): |
|
|
if pth.is_file(): |
|
|
yield pth |
|
|
except (OSError, PermissionError): |
|
|
pass |
|
|
|
|
|
|
|
|
def is_relevant_candidate_env_var(env_var: str, value: str) -> bool: |
|
|
return ( |
|
|
env_var in CUDART_PATH_PREFERRED_ENVVARS |
|
|
or ( |
|
|
os.sep in value |
|
|
and env_var not in CUDART_PATH_IGNORED_ENVVARS |
|
|
and "CONDA" not in env_var |
|
|
and "BASH_FUNC" not in env_var |
|
|
and "\n" not in value |
|
|
) |
|
|
) |
|
|
|
|
|
|
|
|
def get_potentially_lib_path_containing_env_vars() -> Dict[str, str]: |
|
|
return {env_var: value for env_var, value in os.environ.items() if is_relevant_candidate_env_var(env_var, value)} |
|
|
|
|
|
|
|
|
def find_cudart_libraries() -> Iterator[Path]: |
|
|
""" |
|
|
Searches for a cuda installations, in the following order of priority: |
|
|
1. active conda env |
|
|
2. LD_LIBRARY_PATH |
|
|
3. any other env vars, while ignoring those that |
|
|
- are known to be unrelated |
|
|
- don't contain the path separator `/` |
|
|
|
|
|
If multiple libraries are found in part 3, we optimistically try one, |
|
|
while giving a warning message. |
|
|
""" |
|
|
candidate_env_vars = get_potentially_lib_path_containing_env_vars() |
|
|
|
|
|
for envvar in CUDART_PATH_PREFERRED_ENVVARS: |
|
|
if envvar in candidate_env_vars: |
|
|
directory = candidate_env_vars[envvar] |
|
|
yield from find_cuda_libraries_in_path_list(directory) |
|
|
candidate_env_vars.pop(envvar) |
|
|
|
|
|
for env_var, value in candidate_env_vars.items(): |
|
|
yield from find_cuda_libraries_in_path_list(value) |
|
|
|
|
|
|
|
|
def print_cuda_diagnostics(cuda_specs: CUDASpecs) -> None: |
|
|
print( |
|
|
f"PyTorch settings found: CUDA_VERSION={cuda_specs.cuda_version_string}, " |
|
|
f"Highest Compute Capability: {cuda_specs.highest_compute_capability}.", |
|
|
) |
|
|
|
|
|
binary_path = get_cuda_bnb_library_path(cuda_specs) |
|
|
if not binary_path.exists(): |
|
|
print_dedented( |
|
|
f""" |
|
|
Library not found: {binary_path}. Maybe you need to compile it from source? |
|
|
If you compiled from source, try again with `make CUDA_VERSION=DETECTED_CUDA_VERSION`, |
|
|
for example, `make CUDA_VERSION=113`. |
|
|
|
|
|
The CUDA version for the compile might depend on your conda install, if using conda. |
|
|
Inspect CUDA version via `conda list | grep cuda`. |
|
|
""", |
|
|
) |
|
|
|
|
|
cuda_major, cuda_minor = cuda_specs.cuda_version_tuple |
|
|
if cuda_major < 11: |
|
|
print_dedented( |
|
|
""" |
|
|
WARNING: CUDA versions lower than 11 are currently not supported for LLM.int8(). |
|
|
You will be only to use 8-bit optimizers and quantization routines! |
|
|
""", |
|
|
) |
|
|
|
|
|
print(f"To manually override the PyTorch CUDA version please see: {NONPYTORCH_DOC_URL}") |
|
|
|
|
|
|
|
|
if not cuda_specs.has_cublaslt: |
|
|
print_dedented( |
|
|
""" |
|
|
WARNING: Compute capability < 7.5 detected! Only slow 8-bit matmul is supported for your GPU! |
|
|
If you run into issues with 8-bit matmul, you can try 4-bit quantization: |
|
|
https://huggingface.co/blog/4bit-transformers-bitsandbytes |
|
|
""", |
|
|
) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def print_cuda_runtime_diagnostics() -> None: |
|
|
cudart_paths = list(find_cudart_libraries()) |
|
|
if not cudart_paths: |
|
|
print("CUDA SETUP: WARNING! CUDA runtime files not found in any environmental path.") |
|
|
elif len(cudart_paths) > 1: |
|
|
print_dedented( |
|
|
f""" |
|
|
Found duplicate CUDA runtime files (see below). |
|
|
|
|
|
We select the PyTorch default CUDA runtime, which is {torch.version.cuda}, |
|
|
but this might mismatch with the CUDA version that is needed for bitsandbytes. |
|
|
To override this behavior set the `BNB_CUDA_VERSION=<version string, e.g. 122>` environmental variable. |
|
|
|
|
|
For example, if you want to use the CUDA version 122, |
|
|
BNB_CUDA_VERSION=122 python ... |
|
|
|
|
|
OR set the environmental variable in your .bashrc: |
|
|
export BNB_CUDA_VERSION=122 |
|
|
|
|
|
In the case of a manual override, make sure you set LD_LIBRARY_PATH, e.g. |
|
|
export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/usr/local/cuda-11.2, |
|
|
""", |
|
|
) |
|
|
for pth in cudart_paths: |
|
|
print(f"* Found CUDA runtime at: {pth}") |
|
|
|