# Copyright (c) MONAI Consortium # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from __future__ import annotations import platform from _thread import interrupt_main from contextlib import contextmanager from glob import glob from os import path from threading import Timer from types import ModuleType import torch from monai.utils.module import get_torch_version_tuple, optional_import dir_path = path.dirname(path.realpath(__file__)) @contextmanager def timeout(time, message): timer = None try: timer = Timer(time, interrupt_main) timer.daemon = True timer.start() yield except KeyboardInterrupt as e: if timer is not None and timer.is_alive(): raise e # interrupt from user? raise TimeoutError(message) from e finally: if timer is not None: try: timer.cancel() finally: pass def load_module( module_name: str, defines: dict | None = None, verbose_build: bool = False, build_timeout: int = 300 ) -> ModuleType: """ Handles the loading of c++ extension modules. Args: module_name: Name of the module to load. Must match the name of the relevant source directory in the `_extensions` directory. defines: Dictionary containing names and values of compilation defines. verbose_build: Set to true to enable build logging. build_timeout: Time in seconds before the build will throw an exception to prevent hanging. """ # Ensuring named module exists in _extensions directory. module_dir = path.join(dir_path, module_name) if not path.exists(module_dir): raise ValueError(f"No extension module named {module_name}") platform_str = f"_{platform.system()}_{platform.python_version()}_" platform_str += "".join(f"{v}" for v in get_torch_version_tuple()[:2]) # Adding configuration to module name. if defines is not None: module_name = "_".join([module_name] + [f"{v}" for v in defines.values()]) # Gathering source files. source = glob(path.join(module_dir, "**", "*.cpp"), recursive=True) if torch.cuda.is_available(): source += glob(path.join(module_dir, "**", "*.cu"), recursive=True) platform_str += f"_{torch.version.cuda}" # Constructing compilation argument list. define_args = [] if not defines else [f"-D {key}={defines[key]}" for key in defines] # Ninja may be blocked by something out of our control. # This will error if the build takes longer than expected. with timeout(build_timeout, "Build appears to be blocked. Is there a stopped process building the same extension?"): load, _ = optional_import("torch.utils.cpp_extension", name="load") # main trigger some JIT config in pytorch # This will either run the build or return the existing .so object. name = module_name + platform_str.replace(".", "_") module = load( name=name, sources=source, extra_cflags=define_args, extra_cuda_cflags=define_args, verbose=verbose_build ) return module # type: ignore[no-any-return]