|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
""" |
|
|
This module provides a main DSL class for any Dialect. |
|
|
The DSL should be inherited as a new class, and its initialization requires dialects. |
|
|
It handles most of the mechanics for the DSL in an agnostic way, |
|
|
for example, it can handle various dialect-specific tasks. |
|
|
""" |
|
|
|
|
|
|
|
|
|
|
|
from dataclasses import dataclass, field |
|
|
import atexit |
|
|
import os |
|
|
import io |
|
|
import sys |
|
|
import errno |
|
|
import ctypes |
|
|
import re |
|
|
import inspect |
|
|
import argparse |
|
|
import hashlib |
|
|
from functools import lru_cache, wraps |
|
|
from collections import namedtuple |
|
|
from abc import ABC, abstractmethod |
|
|
from typing import Any, Union, Tuple, get_origin, get_args |
|
|
from types import FunctionType |
|
|
import warnings |
|
|
|
|
|
from . import typing as t |
|
|
from .env_manager import EnvironmentVarManager |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from ..base_dsl._mlir_helpers.arith import const |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from .cache_helpers import * |
|
|
from .jit_executor import JitExecutor |
|
|
from .utils.timer import timer |
|
|
from .utils.logger import setup_log, log |
|
|
from .utils.stacktrace import filter_exception, walk_to_top_module, filter_stackframe |
|
|
from .runtime.jit_arg_adapters import is_argument_constexpr, JitArgAdapterRegistry |
|
|
from .runtime.tensor_descriptor import TensorDescriptor |
|
|
from .ast_preprocessor import DSLPreprocessor |
|
|
from .common import * |
|
|
from .typing import ( |
|
|
get_c_pointers, |
|
|
get_mlir_types, |
|
|
) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from .._mlir import ir |
|
|
from .._mlir import runtime as rt |
|
|
from .._mlir.extras import types as T |
|
|
from .._mlir.dialects import arith, math, func |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from .runtime.dlpack_runtime import dlpack_to_tensor_desc, mark_layout_dynamic |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
MLIR_DYNAMIC = -9223372036854775808 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def _numpy_type_to_mlir_type(dtype): |
|
|
if dtype == np.float64: |
|
|
return T.f64() |
|
|
if dtype == np.float16: |
|
|
return T.f16() |
|
|
if dtype == np.float32: |
|
|
return T.f32() |
|
|
if dtype == np.int64: |
|
|
return T.i64() |
|
|
if dtype == np.int32: |
|
|
return T.i32() |
|
|
if dtype == np.int16: |
|
|
return T.i16() |
|
|
if dtype == np.int8: |
|
|
return T.i8() |
|
|
if dtype == np.uint64: |
|
|
return T.ui64() |
|
|
if dtype == np.uint32: |
|
|
return T.ui32() |
|
|
if dtype == np.uint16: |
|
|
return T.ui16() |
|
|
if dtype == np.uint8: |
|
|
return T.ui8() |
|
|
if dtype == np.bool_: |
|
|
return T.bool() |
|
|
if dtype == f8E5M2: |
|
|
return T.f8E5M2() |
|
|
if dtype == f8E4M3FN: |
|
|
return T.f8E4M3FN() |
|
|
if dtype == f8E8M0FNU: |
|
|
return T.f8E8M0FNU() |
|
|
if dtype == f6E3M2FN: |
|
|
return T.f6E3M2FN() |
|
|
if dtype == f6E2M3FN: |
|
|
return T.f6E2M3FN() |
|
|
if dtype == f4E2M1FN: |
|
|
return T.f4E2M1FN() |
|
|
assert False, f"Unknown type {type}" |
|
|
|
|
|
|
|
|
def _mlir_type_to_numpy_type(type): |
|
|
if type == T.f64(): |
|
|
return np.float64 |
|
|
if type == T.f16(): |
|
|
return np.float16 |
|
|
if type == T.f32(): |
|
|
return np.float32 |
|
|
if type == T.i64(): |
|
|
return np.int64 |
|
|
if type == T.i32(): |
|
|
return np.int32 |
|
|
if type == T.i16(): |
|
|
return np.int16 |
|
|
if type == T.i8(): |
|
|
return np.int8 |
|
|
if type == T.ui64(): |
|
|
return np.uint64 |
|
|
if type == T.ui32(): |
|
|
return np.uint32 |
|
|
if type == T.ui16(): |
|
|
return np.uint16 |
|
|
if type == T.ui8(): |
|
|
return np.uint8 |
|
|
if type == T.bool(): |
|
|
return np.bool_ |
|
|
assert False, f"Unknown type {type}" |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def is_dynamic_expression(value): |
|
|
""" |
|
|
Check if the value is an MLIR's SSA value. |
|
|
""" |
|
|
|
|
|
|
|
|
return ( |
|
|
isinstance(value, ir.Value) |
|
|
or hasattr(value, "__extract_mlir_values__") |
|
|
or len(extract_mlir_values(value)) > 0 |
|
|
) |
|
|
|
|
|
|
|
|
def extract_mlir_values(obj): |
|
|
""" |
|
|
Given the `obj`, recursively go through it to extract all contained IR values as list of MLIR values |
|
|
""" |
|
|
res = [] |
|
|
if hasattr(obj, "__extract_mlir_values__"): |
|
|
res = obj.__extract_mlir_values__() |
|
|
elif isinstance(obj, (tuple, list)): |
|
|
res = sum((extract_mlir_values(x) for x in obj), []) |
|
|
|
|
|
elif isinstance(obj, set): |
|
|
raise DSLRuntimeError( |
|
|
"Sets are not supported in extract_mlir_values to ensure order preservation", |
|
|
context="The DSL attempted to generate JIT function argument(s) for an argument of type set but failed.", |
|
|
suggestion="Consider using a list or tuple instead", |
|
|
) |
|
|
elif isinstance(obj, ir.Value): |
|
|
res = [obj] |
|
|
elif isinstance(obj, ir.BlockArgumentList): |
|
|
res = list(obj) |
|
|
|
|
|
return res |
|
|
|
|
|
|
|
|
def new_from_mlir_values(obj, values): |
|
|
""" |
|
|
Create a new python object by populating containing MLIR values with list of new values |
|
|
""" |
|
|
if hasattr(obj, "__new_from_mlir_values__"): |
|
|
return obj.__new_from_mlir_values__(values) |
|
|
elif isinstance(obj, (tuple, list)): |
|
|
res = [] |
|
|
for x in obj: |
|
|
n_items = len(get_mlir_types(x)) |
|
|
res.append(new_from_mlir_values(x, values[:n_items])) |
|
|
values = values[n_items:] |
|
|
obj_ty = type(obj) |
|
|
return obj_ty(res) |
|
|
elif isinstance(obj, set): |
|
|
raise DSLRuntimeError( |
|
|
"Sets are not supported in new_from_mlir_values to ensure order preservation", |
|
|
context="The DSL attempted to generate JIT function argument(s) for an argument of type set but failed.", |
|
|
suggestion="Consider using a list or tuple instead", |
|
|
) |
|
|
elif is_dynamic_expression(obj): |
|
|
|
|
|
if len(values) == 0: |
|
|
return obj |
|
|
|
|
|
assert len(values) == 1 |
|
|
return values[0] |
|
|
else: |
|
|
assert len(values) == 0, f"{obj} expects 0 values, but got {values}" |
|
|
return obj |
|
|
|
|
|
|
|
|
class BaseDSL: |
|
|
gpu_module = None |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
name: str, |
|
|
compiler_provider: Any, |
|
|
pass_sm_arch_name: str, |
|
|
device_compilation_only=False, |
|
|
preprocess=False, |
|
|
): |
|
|
""" |
|
|
Constructor for initializing the class with required providers and environment settings. |
|
|
|
|
|
Parameters: |
|
|
- name (str): Name of DSL, used for environment variables and logging. |
|
|
- compiler_provider (MLIR dialect): Provider for compiler. |
|
|
- pass_sm_arch_name (str): The keyword name of the SM. |
|
|
- device_compilation_only (bool) : Only device code, and call it via cuda driver |
|
|
- preprocess (bool): Enable AST transformation. |
|
|
|
|
|
This constructs a DSL instance and sets up environment management, |
|
|
warning configurations, and logging functionalities. It reads |
|
|
environment variables using `EnvironmentVarManager` and configures |
|
|
a logger with settings from the environment. If environment warnings |
|
|
are detected, they are escalated to errors to ensure strict handling. |
|
|
""" |
|
|
|
|
|
if not all([name, compiler_provider, pass_sm_arch_name]): |
|
|
raise DSLRuntimeError( |
|
|
"All required parameters must be provided and non-empty" |
|
|
) |
|
|
|
|
|
self.name = name |
|
|
self.compiler_provider = compiler_provider |
|
|
self.pass_sm_arch_name = pass_sm_arch_name |
|
|
self.frame = None |
|
|
self.no_cache = False |
|
|
self.device_compilation_only = device_compilation_only |
|
|
self.num_kernels = 0 |
|
|
|
|
|
self.envar = EnvironmentVarManager(self.name) |
|
|
self.enable_preprocessor = preprocess |
|
|
|
|
|
self.jit_cache = ( |
|
|
dict() |
|
|
if self.envar.disable_file_caching |
|
|
else load_cache_from_path(self.name, self.envar.file_caching_capacity) |
|
|
) |
|
|
self.host_jit_decorator_name = f"@{BaseDSL.jit.__name__}" |
|
|
self.device_jit_decorator_name = f"@{BaseDSL.kernel.__name__}" |
|
|
|
|
|
|
|
|
if self.envar.warnings_as_errors: |
|
|
warnings.filterwarnings("error") |
|
|
if self.envar.warnings_ignore: |
|
|
warnings.filterwarnings("ignore") |
|
|
|
|
|
|
|
|
if self.envar.log_to_console == False and self.envar.jitTimeProfiling: |
|
|
self.envar.log_to_console = True |
|
|
self.envar.log_level = 20 |
|
|
setup_log( |
|
|
self.name, |
|
|
self.envar.log_to_console, |
|
|
self.envar.log_to_file, |
|
|
f"{self.name}.log", |
|
|
self.envar.log_level, |
|
|
) |
|
|
|
|
|
|
|
|
self.kernel_symbols = [] |
|
|
|
|
|
self.launch_inner_count = 0 |
|
|
|
|
|
if preprocess: |
|
|
self.preprocessor = DSLPreprocessor() |
|
|
log().info(f"Initializing {name} DSL") |
|
|
log().debug(f"Logger initialized for {self.name}") |
|
|
|
|
|
|
|
|
if self.envar.filterStacktrace: |
|
|
origin_excepthook = sys.excepthook |
|
|
module_dir = walk_to_top_module(os.path.dirname(os.path.abspath(__file__))) |
|
|
|
|
|
def excepthook(excep_type, value, traceback): |
|
|
filter_exception(value, module_dir) |
|
|
if hasattr(value, "__traceback__"): |
|
|
origin_excepthook(excep_type, value, value.__traceback__) |
|
|
else: |
|
|
origin_excepthook( |
|
|
excep_type, value, filter_stackframe(traceback, module_dir) |
|
|
) |
|
|
|
|
|
sys.excepthook = excepthook |
|
|
|
|
|
|
|
|
def restore_excepthook(hook): |
|
|
sys.excepthook = hook |
|
|
|
|
|
atexit.register(restore_excepthook, origin_excepthook) |
|
|
|
|
|
def dump_cache(self): |
|
|
if not self.envar.disable_file_caching: |
|
|
dump_cache_to_path( |
|
|
self.name, self.jit_cache, self.envar.file_caching_capacity |
|
|
) |
|
|
|
|
|
@lru_cache(maxsize=1) |
|
|
def print_warning_once(self, message): |
|
|
log().warning(f"Warning: {message}") |
|
|
warnings.warn(message, UserWarning) |
|
|
|
|
|
def print_warning(self, message): |
|
|
log().warning(f"Warning: {message}") |
|
|
warnings.warn(message, UserWarning) |
|
|
|
|
|
@classmethod |
|
|
@lru_cache(maxsize=1) |
|
|
def _get_dsl(cls): |
|
|
|
|
|
main_dsl = cls() |
|
|
if not main_dsl.no_cache: |
|
|
|
|
|
atexit.register(main_dsl.dump_cache) |
|
|
return main_dsl |
|
|
|
|
|
@staticmethod |
|
|
def _can_preprocess(**dkwargs): |
|
|
""" |
|
|
Check if AST transformation is enabled or not for `jit` and `kernel` decorators. |
|
|
""" |
|
|
return dkwargs.pop("preprocess", True) |
|
|
|
|
|
@staticmethod |
|
|
def _get_original_function(fcn_ptr, name): |
|
|
""" |
|
|
Get the original function from the decorated function |
|
|
""" |
|
|
while fcn_ptr.__name__ != name: |
|
|
|
|
|
if hasattr(fcn_ptr, "__wrapped__"): |
|
|
fcn_ptr = fcn_ptr.__wrapped__ |
|
|
|
|
|
elif callable(fcn_ptr.__closure__[0].cell_contents): |
|
|
fcn_ptr = fcn_ptr.__closure__[0].cell_contents |
|
|
else: |
|
|
raise DSLRuntimeError( |
|
|
f"Cannot find the original function {name} in the closure chain" |
|
|
) |
|
|
return fcn_ptr |
|
|
|
|
|
@staticmethod |
|
|
def _preprocess_and_execute(func): |
|
|
""" |
|
|
Run ast transformation and return the materialized function pointer |
|
|
""" |
|
|
if hasattr(func, "_transformed_ast"): |
|
|
|
|
|
func._dsl_object.frame = func._decorator_frame |
|
|
|
|
|
if func._transformed_ast is None: |
|
|
func._transformed_ast = func._dsl_object.run_preprocessor(func) |
|
|
if func._transformed_ast is None: |
|
|
del func._decorator_frame |
|
|
del func._transformed_ast |
|
|
return func |
|
|
|
|
|
fcn_ptr = func._dsl_object.get_function_ptr(func, func._transformed_ast) |
|
|
|
|
|
fcn_ptr = BaseDSL._get_original_function(fcn_ptr, func.__name__) |
|
|
return fcn_ptr |
|
|
return func |
|
|
|
|
|
def jit_runner(self, frame, executor, *dargs, **dkwargs): |
|
|
""" |
|
|
Decorator to mark a function for JIT compilation. |
|
|
""" |
|
|
|
|
|
self.frame = frame |
|
|
log().info("jit_runner") |
|
|
|
|
|
def jit_runner_decorator(func): |
|
|
func._dsl_object = self |
|
|
|
|
|
if self.enable_preprocessor and BaseDSL._can_preprocess(**dkwargs): |
|
|
|
|
|
|
|
|
func._decorator_frame = frame |
|
|
|
|
|
func._transformed_ast = None |
|
|
|
|
|
@wraps(func) |
|
|
def jit_wrapper(*args, **kwargs): |
|
|
func_ptr = BaseDSL._preprocess_and_execute(func) |
|
|
return executor(func_ptr, *args, **kwargs) |
|
|
|
|
|
return jit_wrapper |
|
|
|
|
|
if len(dargs) == 1 and callable(dargs[0]): |
|
|
return jit_runner_decorator(dargs[0]) |
|
|
else: |
|
|
return jit_runner_decorator |
|
|
|
|
|
@classmethod |
|
|
def jit(cls, *dargs, **dkwargs): |
|
|
""" |
|
|
Decorator to mark a function for JIT compilation for Host code. |
|
|
""" |
|
|
frame = inspect.currentframe().f_back |
|
|
|
|
|
main_dsl = cls._get_dsl() |
|
|
return main_dsl.jit_runner(frame, main_dsl._func, *dargs, **dkwargs) |
|
|
|
|
|
@classmethod |
|
|
def kernel(cls, *dargs, **dkwargs): |
|
|
""" |
|
|
Decorator to mark a function for JIT compilation for GPU. |
|
|
""" |
|
|
frame = inspect.currentframe().f_back |
|
|
|
|
|
main_dsl = cls._get_dsl() |
|
|
return main_dsl.jit_runner(frame, main_dsl._kernel_helper, *dargs, **dkwargs) |
|
|
|
|
|
@abstractmethod |
|
|
def _kernel_helper(self, func, *args, **kwargs): |
|
|
""" |
|
|
Helper function to handle kernel generation logic |
|
|
""" |
|
|
pass |
|
|
|
|
|
@abstractmethod |
|
|
def _build_gpu_module(self, attrs): |
|
|
""" |
|
|
Build the module op that contains the kernels. |
|
|
""" |
|
|
pass |
|
|
|
|
|
@abstractmethod |
|
|
def _get_pipeline(self, pipeline): |
|
|
""" |
|
|
Get the pipeline from the other configuration options. |
|
|
""" |
|
|
if pipeline != None: |
|
|
return pipeline |
|
|
return None |
|
|
|
|
|
@staticmethod |
|
|
def log_additions(func_type, operands=None, types=None, arg_attrs=None): |
|
|
if operands is not None and operands != []: |
|
|
log().debug( |
|
|
f"Added {func_type} operands: [%s]", ", ".join(map(str, operands)) |
|
|
) |
|
|
if types is not None: |
|
|
log().debug( |
|
|
f"Added {func_type} arg_types: [%s]", ", ".join(map(str, types)) |
|
|
) |
|
|
if arg_attrs is not None: |
|
|
log().debug( |
|
|
f"Added {func_type} arg_attrs: [%s]", ", ".join(map(str, arg_attrs)) |
|
|
) |
|
|
|
|
|
def mangle_name(self, function_name, args, args_spec: inspect.FullArgSpec): |
|
|
"""Does simple name mangling""" |
|
|
|
|
|
for spec_arg, arg in zip(args_spec.args, args): |
|
|
spec_ty = args_spec.annotations.get(spec_arg, None) |
|
|
if spec_ty != None: |
|
|
if issubclass(type(spec_ty), (t.IRValue, t.IRVariadic)): |
|
|
continue |
|
|
if isinstance(spec_ty, (ir.Type, ir.Value)): |
|
|
continue |
|
|
if isinstance(arg, (ir.Type, ir.Value, ir.OpResult)): |
|
|
continue |
|
|
if isinstance(type(arg), (ir.Type, ir.Value, ir.OpResult)): |
|
|
continue |
|
|
if self._is_tensor_descriptor(arg): |
|
|
continue |
|
|
if inspect.isclass(spec_ty): |
|
|
class_name = str(arg).replace("class", "") |
|
|
class_name = class_name.replace(" ", "") |
|
|
function_name = f"{function_name}_{class_name}" |
|
|
elif isinstance(arg, (list, tuple)): |
|
|
function_name = f"{function_name}_{'_'.join(map(str, arg))}" |
|
|
else: |
|
|
function_name = f"{function_name}_{arg}" |
|
|
|
|
|
unwanted_chars = r"'-![]#,.<>()\":{}=%?@;" |
|
|
translation_table = str.maketrans("", "", unwanted_chars) |
|
|
function_name = function_name.translate(translation_table) |
|
|
|
|
|
function_name = re.sub(r"0x[a-f0-9]{8,16}", "", function_name) |
|
|
function_name = re.sub(r"\s+", " ", function_name) |
|
|
function_name = function_name.replace(" ", "_") |
|
|
function_name = function_name.replace("\n", "_") |
|
|
|
|
|
function_name = function_name[:180] |
|
|
log().info(f"Final mangled function name: {function_name}") |
|
|
return function_name |
|
|
|
|
|
def _generate_execution_arguments_for_known_types( |
|
|
self, arg, arg_spec, arg_name, i, fop_args, iv_block_args |
|
|
): |
|
|
""" |
|
|
Generate MLIR arguments for known types. |
|
|
|
|
|
Sub-DSLs can override this method to handle types that are not |
|
|
natively supported by the Base DSL. |
|
|
""" |
|
|
ir_arg = [] |
|
|
if is_argument_constexpr(arg, arg_spec, arg_name, i, func): |
|
|
ir_arg.append(arg) |
|
|
|
|
|
return ir_arg, iv_block_args |
|
|
|
|
|
def generate_execution_arguments( |
|
|
self, |
|
|
args, |
|
|
kwargs, |
|
|
fop, |
|
|
args_spec: inspect.FullArgSpec, |
|
|
): |
|
|
"""Create list of arguments that will be passed to MLIR's func.func op""" |
|
|
|
|
|
def gen_exec_args(input_args, arg_names, annotations, fop_args): |
|
|
assert len(input_args) == len(arg_names) |
|
|
|
|
|
ir_args = [] |
|
|
iv_block_args = 0 |
|
|
for i, arg in enumerate(input_args): |
|
|
arg_name = arg_names[i] |
|
|
arg_spec = annotations.get(arg_name, None) |
|
|
log().debug("Processing [%d] Argument [%s : %s]", i, arg_name, arg_spec) |
|
|
|
|
|
|
|
|
if isinstance(arg_spec, t.NumericMeta) and not isinstance( |
|
|
arg, arg_spec |
|
|
): |
|
|
arg = t.cast(arg, arg_spec) |
|
|
|
|
|
ir_arg, iv_block_args = ( |
|
|
self._generate_execution_arguments_for_known_types( |
|
|
arg, arg_spec, arg_name, i, fop_args, iv_block_args |
|
|
) |
|
|
) |
|
|
|
|
|
if not ir_arg: |
|
|
|
|
|
|
|
|
adapter = JitArgAdapterRegistry.get_registered_adapter(type(arg)) |
|
|
arg = adapter(arg) if adapter else arg |
|
|
|
|
|
n_args = len(get_mlir_types(arg)) |
|
|
blk_args = fop_args[iv_block_args : iv_block_args + n_args] |
|
|
ir_arg.append(new_from_mlir_values(arg, blk_args)) |
|
|
iv_block_args += n_args |
|
|
|
|
|
self.log_additions(ir_arg) |
|
|
ir_args.extend(ir_arg) |
|
|
|
|
|
return ir_args, iv_block_args |
|
|
|
|
|
fop_args = list(fop.regions[0].blocks[0].arguments) |
|
|
ir_args, iv_block_args = gen_exec_args( |
|
|
args, args_spec.args, args_spec.annotations, fop_args |
|
|
) |
|
|
ir_kwargs, _ = gen_exec_args( |
|
|
[kwargs[arg] for arg in args_spec.kwonlyargs], |
|
|
args_spec.kwonlyargs, |
|
|
args_spec.annotations, |
|
|
fop_args[iv_block_args:], |
|
|
) |
|
|
ir_kwargs = {k: v for k, v in zip(args_spec.kwonlyargs, ir_kwargs)} |
|
|
|
|
|
log().debug("execution args: %s", ", ".join(map(str, ir_args))) |
|
|
log().debug("execution kwargs: %s", ", ".join(map(str, ir_kwargs))) |
|
|
return ir_args, ir_kwargs |
|
|
|
|
|
@abstractmethod |
|
|
def _generate_mlir_type_for_tensor_descriptor(self, tensor: TensorDescriptor): |
|
|
""" |
|
|
Generate MLIR type for the tensor descriptor. |
|
|
""" |
|
|
pass |
|
|
|
|
|
@abstractmethod |
|
|
def _generate_executable_arg_for_tensor_descriptor( |
|
|
self, mlir_value=None, ptr_tensor_ty=None, tensor=None |
|
|
): |
|
|
""" |
|
|
Generates executable value for the given tensor descriptor. |
|
|
""" |
|
|
pass |
|
|
|
|
|
@abstractmethod |
|
|
def _get_globals(self): |
|
|
""" |
|
|
Combines global and local variables from the current context and the |
|
|
caller's frame comes. This includes the current module's globals, the |
|
|
global variables from the caller's frame, and the local variables from |
|
|
the caller's frame. |
|
|
|
|
|
"self.frame" is used to fetch the caller's frame. |
|
|
|
|
|
AST preprocessor generates a new python code, so the resulting globals |
|
|
dictionary is used to execute the python code. |
|
|
""" |
|
|
pass |
|
|
|
|
|
def _is_tensor_descriptor(self, maybe_tensor_descriptor) -> bool: |
|
|
return isinstance( |
|
|
maybe_tensor_descriptor, TensorDescriptor |
|
|
) or TensorDescriptor.can_transformed_to_dlpack(maybe_tensor_descriptor) |
|
|
|
|
|
def _handle_tensor_descriptor( |
|
|
self, maybe_tensor, arg_name: str, need_gpu_memory: bool |
|
|
) -> TensorDescriptor: |
|
|
if self._is_tensor_descriptor(maybe_tensor): |
|
|
tensor = ( |
|
|
maybe_tensor |
|
|
if isinstance(maybe_tensor, TensorDescriptor) |
|
|
else TensorDescriptor(maybe_tensor) |
|
|
) |
|
|
if need_gpu_memory and not tensor.is_in_device: |
|
|
log().info( |
|
|
"FAIL name=[%s] tensor=[%s] in_gpu=[%s]", |
|
|
arg_name, |
|
|
tensor, |
|
|
tensor.is_in_device, |
|
|
) |
|
|
raise DSLRuntimeError( |
|
|
f'Tensor "{arg_name}" is tensor "{tensor}" ' |
|
|
"is not in the GPU memory. " |
|
|
) |
|
|
|
|
|
return tensor |
|
|
|
|
|
raise DSLRuntimeError( |
|
|
f"Argument {arg_name} could not be transformed into a TensorDescriptor." |
|
|
) |
|
|
|
|
|
def _validate_arg(self, arg, arg_index, arg_name, arg_spec): |
|
|
""" |
|
|
Validates if the arg is really of the annotated type for type safety. |
|
|
|
|
|
The default implementation is empty. Subclasses can override this method to add more validation logic. |
|
|
Returns None if validation passes, otherwise returns an error derived from DSLBaseError. |
|
|
""" |
|
|
pass |
|
|
|
|
|
def _generate_jit_func_args_for_known_types( |
|
|
self, |
|
|
func, |
|
|
arg, |
|
|
arg_name, |
|
|
arg_spec, |
|
|
arg_index, |
|
|
*, |
|
|
is_host=True, |
|
|
): |
|
|
""" |
|
|
Generate JIT function arguments for known types. |
|
|
|
|
|
Sub-DSLs can override this method to handle types that are not |
|
|
natively supported by the Base DSL. |
|
|
""" |
|
|
|
|
|
jit_arg_type, jit_arg_attr, jit_exec_arg = [], [], [] |
|
|
default_attr = ir.DictAttr.get({}) |
|
|
|
|
|
if is_argument_constexpr(arg, arg_spec, arg_name, arg_index, func): |
|
|
jit_exec_arg = jit_arg_type = jit_arg_attr = None |
|
|
|
|
|
return jit_exec_arg, jit_arg_type, jit_arg_attr |
|
|
|
|
|
def _generate_jit_func_args( |
|
|
self, |
|
|
func, |
|
|
function_name, |
|
|
args, |
|
|
kwargs, |
|
|
args_spec: inspect.FullArgSpec, |
|
|
*, |
|
|
is_host=True, |
|
|
): |
|
|
"""Generate JIT function arguments.""" |
|
|
|
|
|
assert len(args) == len(args_spec.args) and len(kwargs) == len( |
|
|
args_spec.kwonlyargs |
|
|
), ( |
|
|
f"Input args {len(args)=} and kwargs {len(kwargs)=} must match arg_spec.args " |
|
|
f"{len(args_spec.args)=} and arg_spec.kwonlyargs {len(args_spec.kwonlyargs)=}" |
|
|
) |
|
|
|
|
|
jit_arg_types, jit_arg_attrs, jit_exec_args = [], [], [] |
|
|
default_attr = ir.DictAttr.get({}) |
|
|
|
|
|
input_args = [*args, *kwargs.values()] |
|
|
input_arg_names = [*args_spec.args, *args_spec.kwonlyargs] |
|
|
for i, (arg_name, arg) in enumerate(zip(input_arg_names, input_args)): |
|
|
spec_ty = args_spec.annotations.get(arg_name, None) |
|
|
log().debug("Processing [%d] Argument [%s : %s]", i, arg_name, spec_ty) |
|
|
|
|
|
|
|
|
if isinstance(spec_ty, t.NumericMeta) and not isinstance(arg, spec_ty): |
|
|
arg = t.cast(arg, spec_ty) |
|
|
|
|
|
|
|
|
if spec_ty is not None: |
|
|
err = self._validate_arg(arg, i, arg_name, spec_ty) |
|
|
if err is not None: |
|
|
raise err |
|
|
|
|
|
jit_exec_arg, jit_arg_type, jit_arg_attr = ( |
|
|
self._generate_jit_func_args_for_known_types( |
|
|
func, |
|
|
arg, |
|
|
arg_name, |
|
|
spec_ty, |
|
|
i, |
|
|
is_host=is_host, |
|
|
) |
|
|
) |
|
|
|
|
|
if jit_arg_type is not None and len(jit_arg_type) == 0: |
|
|
|
|
|
|
|
|
adapter = JitArgAdapterRegistry.get_registered_adapter(type(arg)) |
|
|
arg = adapter(arg) if adapter else arg |
|
|
|
|
|
if is_host: |
|
|
jit_exec_arg.extend(get_c_pointers(arg)) |
|
|
jit_arg_type.extend(get_mlir_types(arg)) |
|
|
else: |
|
|
dyn_vals = extract_mlir_values(arg) |
|
|
jit_exec_arg.extend(dyn_vals) |
|
|
jit_arg_type.extend([v.type for v in dyn_vals]) |
|
|
|
|
|
if not jit_arg_type or not jit_exec_arg: |
|
|
if (is_host and hasattr(arg, "__c_pointers__")) or ( |
|
|
not is_host |
|
|
and hasattr(arg, "__extract_mlir_values__") |
|
|
and hasattr(arg, "__new_from_mlir_values__") |
|
|
): |
|
|
pass |
|
|
else: |
|
|
raise DSLRuntimeError( |
|
|
f"failed to generate argument #{i+1} ({arg_name}) for JIT function '{function_name}'.", |
|
|
context={ |
|
|
f"Argument {arg_name}": "The DSL attempted to convert it into Dynamic Expression (aka MLIR values) but failed.", |
|
|
f"Call-site argument value": arg, |
|
|
f"Call-site argument type": type(arg), |
|
|
}, |
|
|
suggestion=f"Consider annotating the argument with `{arg_name} : Constexpr` " |
|
|
"if it's a value known at compile-time. " |
|
|
f"Otherwise, implement the {'`JitArgument`' if is_host else '`DynamicExpression`'} " |
|
|
f"protocol or register a custom JIT argument adapter for type `{type(arg)}` to " |
|
|
"enable dynamic value conversion at runtime.", |
|
|
) |
|
|
|
|
|
jit_arg_attr.extend([default_attr] * len(jit_arg_type)) |
|
|
|
|
|
if jit_arg_type is not None: |
|
|
jit_exec_args.extend(jit_exec_arg) |
|
|
jit_arg_types.extend(jit_arg_type) |
|
|
jit_arg_attrs.extend(jit_arg_attr) |
|
|
|
|
|
return jit_exec_args, jit_arg_types, jit_arg_attrs |
|
|
|
|
|
def generate_mlir_function_types( |
|
|
self, func, function_name, input_args, kwargs, args_spec: inspect.FullArgSpec |
|
|
): |
|
|
"""Convert input arguments to MLIR function signature also convert numpy arrays to memref.""" |
|
|
|
|
|
exe_args, types, _ = self._generate_jit_func_args( |
|
|
func, function_name, input_args, kwargs, args_spec, is_host=True |
|
|
) |
|
|
|
|
|
log().debug("Execution Arguments: %s", ", ".join(map(str, exe_args))) |
|
|
log().debug("Types: %s", ", ".join(map(str, types))) |
|
|
|
|
|
assert len(exe_args) == len( |
|
|
types |
|
|
), "expects the same number of arguments and function parameters" |
|
|
|
|
|
return exe_args, types |
|
|
|
|
|
@dataclass |
|
|
class LaunchConfig: |
|
|
cluster: list = None |
|
|
grid: list = field(default_factory=lambda: [1, 1, 1]) |
|
|
block: list = field(default_factory=lambda: [1, 1, 1]) |
|
|
smem: int = 0 |
|
|
async_deps: list = field(default_factory=list) |
|
|
has_cluster: bool = False |
|
|
min_blocks_per_mp: int = 0 |
|
|
|
|
|
def __post_init__(self): |
|
|
if len(self.grid) != 3: |
|
|
raise DSLRuntimeError(f"Expect 3d grid!") |
|
|
if len(self.block) != 3: |
|
|
raise DSLRuntimeError(f"Expect 3d block!") |
|
|
|
|
|
self.has_cluster = self.cluster is not None |
|
|
if self.cluster is None: |
|
|
self.cluster = [None, None, None] |
|
|
elif len(self.cluster) != 3: |
|
|
raise DSLRuntimeError(f"Expect 3d cluster!") |
|
|
|
|
|
def diagnostic(self): |
|
|
"""Check command line parameters and enables diagnostic""" |
|
|
|
|
|
parser = argparse.ArgumentParser(description="Process diagnostic status.") |
|
|
parser.add_argument( |
|
|
"-diagnostic", |
|
|
nargs="?", |
|
|
const="all", |
|
|
choices=["all", "fail", "success", "info", "suggestion"], |
|
|
help="Set diagnostic status (fail, success, info, suggestion).", |
|
|
) |
|
|
|
|
|
args, _ = parser.parse_known_args() |
|
|
ctx = ir.Context.current |
|
|
|
|
|
def callback(d): |
|
|
print(f" [{self.name} Diagnostic] : {d.message}") |
|
|
|
|
|
ctx.attach_diagnostic_handler(callback) |
|
|
|
|
|
|
|
|
if args.diagnostic is None: |
|
|
return |
|
|
|
|
|
|
|
|
ctx.emit_error_diagnostics = True |
|
|
ir._GlobalDebug.flag = True |
|
|
if args.diagnostic == "all": |
|
|
ir._GlobalDebug.set_types("diagnostic") |
|
|
else: |
|
|
ir._GlobalDebug.set_types(f"diagnostic-{args.diagnostic}") |
|
|
|
|
|
def get_location(self): |
|
|
""" |
|
|
Get python location information and generate MLIR location |
|
|
""" |
|
|
|
|
|
frame = self.frame |
|
|
if frame is None: |
|
|
print("Frame is None") |
|
|
return None |
|
|
|
|
|
file_loc = ir.Location.file(frame.f_code.co_filename, frame.f_lineno, 0) |
|
|
|
|
|
def print_all_frames(): |
|
|
for i, frame in enumerate(inspect.stack()): |
|
|
print( |
|
|
f"Frame {i}: {frame.function} in {frame.filename}, line {frame.lineno}" |
|
|
) |
|
|
|
|
|
loc = ir.Location.name(frame.f_code.co_name, childLoc=file_loc) |
|
|
return loc |
|
|
|
|
|
def compile_and_jit(self, module, pipeline, shared_libs, function_name=""): |
|
|
""" |
|
|
Compile and JIT an MLIR module. |
|
|
""" |
|
|
|
|
|
try: |
|
|
self.diagnostic() |
|
|
|
|
|
orig_stdout = sys.stdout |
|
|
orig_stderr = sys.stderr |
|
|
sys.stderr = redirect_stderr = io.StringIO() |
|
|
sys.stdout = redirect_stdout = io.StringIO() |
|
|
|
|
|
try: |
|
|
kernel = self.compiler_provider.compile_and_jit( |
|
|
module, |
|
|
pipeline, |
|
|
shared_libs=shared_libs, |
|
|
cuda_toolkit=self.envar.cuda_toolkit, |
|
|
arch=self.envar.arch, |
|
|
) |
|
|
|
|
|
finally: |
|
|
sys.stdout = orig_stdout |
|
|
sys.stderr = orig_stderr |
|
|
ir._GlobalDebug.flag = False |
|
|
|
|
|
|
|
|
print(redirect_stdout.getvalue(), file=sys.stdout, end="") |
|
|
print(redirect_stderr.getvalue(), file=sys.stderr, end="") |
|
|
|
|
|
return kernel |
|
|
|
|
|
except Exception as e: |
|
|
raise DSLRuntimeError("🧊🧊🧊 ICE 🧊🧊🧊", cause=e) |
|
|
finally: |
|
|
pass |
|
|
|
|
|
def preprocess_pipeline(self, pipeline, arch) -> str: |
|
|
|
|
|
if self.envar.cuda_toolkit is None: |
|
|
self.print_warning( |
|
|
"CUDA_TOOLKIT_PATH environment variable is not set. Cannot set toolkitPath." |
|
|
) |
|
|
|
|
|
options = { |
|
|
"toolkitPath": self.envar.cuda_toolkit if self.envar.cuda_toolkit else None, |
|
|
self.pass_sm_arch_name: arch, |
|
|
} |
|
|
|
|
|
opt_str = "" |
|
|
for k, v in options.items(): |
|
|
if v: |
|
|
opt_str += f"{k}={v} " |
|
|
|
|
|
if opt_str: |
|
|
|
|
|
pattern = re.compile(r"{(.+)}") |
|
|
match = pattern.search(pipeline) |
|
|
if match: |
|
|
opt_str = f"{{{match[1]} {opt_str}}}" |
|
|
pipeline = re.sub(r"{.+}", opt_str, pipeline) |
|
|
else: |
|
|
pipeline = pipeline.rstrip(")") + f"{{{opt_str}}})" |
|
|
log().debug(f"Using pipeline = {pipeline}") |
|
|
return pipeline |
|
|
|
|
|
def get_shared_libs(self) -> list: |
|
|
shared_libs = [] |
|
|
support_libs = self.envar.shared_libs |
|
|
if support_libs is not None: |
|
|
_libs = support_libs.split(":") |
|
|
for lib in _libs: |
|
|
if not os.path.exists(lib): |
|
|
raise FileNotFoundError( |
|
|
errno.ENOENT, os.strerror(errno.ENOENT), lib |
|
|
) |
|
|
shared_libs.append(lib) |
|
|
else: |
|
|
self.print_warning(f"{self.name}_LIBS environment variable is not set") |
|
|
|
|
|
return shared_libs |
|
|
|
|
|
@lru_cache(maxsize=1) |
|
|
def get_version(self): |
|
|
version_hash = hashlib.sha256() |
|
|
|
|
|
return version_hash |
|
|
|
|
|
def get_module_hash(self, module, function_name): |
|
|
s = io.BytesIO() |
|
|
module.operation.write_bytecode(s) |
|
|
for attr, value in self.envar.__dict__.items(): |
|
|
if value is not None: |
|
|
s.write(str(value).encode()) |
|
|
module_hash = self.get_version().copy() |
|
|
module_hash.update(s.getvalue()) |
|
|
module_hash = module_hash.hexdigest() |
|
|
|
|
|
log().debug("Bytecode=[%s]", s.getvalue().hex()) |
|
|
log().debug("Version=[%s]", self.get_version().hexdigest()) |
|
|
log().info( |
|
|
"Function=[%s] Computed module_hash=[%s]", function_name, module_hash |
|
|
) |
|
|
return module_hash |
|
|
|
|
|
def build_module(self, module, function_name: str): |
|
|
""" |
|
|
Build the MLIR module, verify and return the module |
|
|
""" |
|
|
|
|
|
|
|
|
if self.envar.keepIR: |
|
|
save_ir(self.name, module, function_name) |
|
|
|
|
|
if self.envar.printIR: |
|
|
print("\n//===--- ------ Generated IR ------ ---====\n") |
|
|
module.operation.print( |
|
|
enable_debug_info=self.envar.generate_source_location |
|
|
) |
|
|
print("\n//===--- --- End of Generated IR -- ---====\n") |
|
|
|
|
|
|
|
|
try: |
|
|
module.operation.verify() |
|
|
except Exception as e: |
|
|
raise DSLRuntimeError(f"🧊🧊🧊 ICE IR Verification Failed 🧊🧊🧊", cause=e) |
|
|
|
|
|
return module |
|
|
|
|
|
def generate_original_ir( |
|
|
self, |
|
|
ir, |
|
|
func, |
|
|
funcBody, |
|
|
kwargs, |
|
|
function_name, |
|
|
func_types, |
|
|
gpu_module_attrs, |
|
|
args, |
|
|
args_spec, |
|
|
): |
|
|
|
|
|
|
|
|
|
|
|
loc = None |
|
|
|
|
|
def build_ir_module(): |
|
|
module = ir.Module.create(loc=loc) |
|
|
unit_attr = ir.UnitAttr.get() |
|
|
module.operation.attributes["gpu.container_module"] = unit_attr |
|
|
|
|
|
with ir.InsertionPoint(module.body): |
|
|
|
|
|
self._build_gpu_module(gpu_module_attrs) |
|
|
|
|
|
fop = func.FuncOp(function_name, (func_types, []), loc=loc) |
|
|
fop.attributes["llvm.emit_c_interface"] = ir.UnitAttr.get() |
|
|
log().debug("Generated Function OP [%s]", fop) |
|
|
with ir.InsertionPoint(fop.add_entry_block()): |
|
|
ir_args, ir_kwargs = self.generate_execution_arguments( |
|
|
args, kwargs, fop, args_spec |
|
|
) |
|
|
|
|
|
try: |
|
|
result = funcBody(*ir_args, **ir_kwargs) |
|
|
func.ReturnOp([]) |
|
|
except DSLAstPreprocessorError as pp_error: |
|
|
raise pp_error |
|
|
except NameError as name_error: |
|
|
raise DSLRuntimeError( |
|
|
f"💥💥💥 Error during runtime code generation for function `{funcBody.__name__}` 💥💥💥", |
|
|
cause=name_error, |
|
|
suggestion="Using variables defined in dynamic control flow is not supported. Please give an initial value before control flow.", |
|
|
) |
|
|
except DSLRuntimeError as dsl_error: |
|
|
|
|
|
raise dsl_error |
|
|
except Exception as general_e: |
|
|
|
|
|
raise DSLRuntimeError( |
|
|
f"💥💥💥 Error during runtime code generation for function `{funcBody.__name__}` 💥💥💥" |
|
|
) from general_e |
|
|
return module, result |
|
|
|
|
|
|
|
|
profiler = timer(enable=self.envar.jitTimeProfiling) |
|
|
module, result = profiler(build_ir_module)() |
|
|
module_hash = self.get_module_hash(module, function_name) |
|
|
|
|
|
module = self.build_module(module, function_name) |
|
|
|
|
|
return module, module_hash, result |
|
|
|
|
|
def compile_and_cache( |
|
|
self, module, module_hash, function_name, pipeline, args_spec, no_cache |
|
|
): |
|
|
arch = self.envar.arch |
|
|
pipeline = self.preprocess_pipeline(self._get_pipeline(pipeline), arch) |
|
|
shared_libs = self.get_shared_libs() |
|
|
profiler = timer(enable=self.envar.jitTimeProfiling) |
|
|
if ( |
|
|
no_cache |
|
|
or module_hash not in self.jit_cache |
|
|
or self.jit_cache[module_hash].ir_module is None |
|
|
): |
|
|
log().info( |
|
|
"JIT cache miss function=[%s] module_hash=[%s]", |
|
|
function_name, |
|
|
module_hash, |
|
|
) |
|
|
|
|
|
engine = profiler(self.compile_and_jit)( |
|
|
module, pipeline, shared_libs, function_name=function_name |
|
|
) |
|
|
else: |
|
|
log().info( |
|
|
"JIT cache hit IN-FILE function=[%s] module_hash=[%s]", |
|
|
function_name, |
|
|
module_hash, |
|
|
) |
|
|
module = self.jit_cache[module_hash].ir_module |
|
|
engine = self.compiler_provider.jit(module, shared_libs=shared_libs) |
|
|
capi_func = profiler(engine.lookup)(function_name) |
|
|
jit_executor = JitExecutor( |
|
|
self, |
|
|
engine, |
|
|
capi_func, |
|
|
module, |
|
|
args_spec, |
|
|
function_name, |
|
|
jit_time_profiling=self.envar.jitTimeProfiling, |
|
|
) |
|
|
jit_executor = jit_executor.update_jit_cuda_modules(self.kernel_symbols) |
|
|
|
|
|
if not no_cache: |
|
|
|
|
|
self.jit_cache[module_hash] = jit_executor |
|
|
|
|
|
return jit_executor |
|
|
|
|
|
def post_compilation_cleanup(self): |
|
|
"""Clean up some internal state after one compilation is completed.""" |
|
|
|
|
|
self.kernel_symbols = [] |
|
|
self.launch_inner_count = 0 |
|
|
|
|
|
self.num_kernels = 0 |
|
|
|
|
|
def generate_mlir( |
|
|
self, |
|
|
funcBody, |
|
|
kwargs, |
|
|
function_name, |
|
|
gpu_module_attrs, |
|
|
args, |
|
|
args_spec, |
|
|
pipeline, |
|
|
no_cache, |
|
|
compile_only, |
|
|
loc=None, |
|
|
): |
|
|
"""Generate MLIR module and compile iself.T_provider.""" |
|
|
with ir.Context(), ir.Location.unknown(): |
|
|
|
|
|
exe_args, func_types = self.generate_mlir_function_types( |
|
|
funcBody, function_name, args, kwargs, args_spec |
|
|
) |
|
|
|
|
|
|
|
|
module, module_hash, result = self.generate_original_ir( |
|
|
ir, |
|
|
func, |
|
|
funcBody, |
|
|
kwargs, |
|
|
function_name, |
|
|
func_types, |
|
|
gpu_module_attrs, |
|
|
args, |
|
|
args_spec, |
|
|
) |
|
|
|
|
|
|
|
|
if self.envar.dryrun: |
|
|
return result |
|
|
|
|
|
if ( |
|
|
no_cache |
|
|
or module_hash not in self.jit_cache |
|
|
or self.jit_cache[module_hash].capi_func is None |
|
|
): |
|
|
|
|
|
jit_executor = self.compile_and_cache( |
|
|
module, module_hash, function_name, pipeline, args_spec, no_cache |
|
|
) |
|
|
else: |
|
|
|
|
|
log().info( |
|
|
"JIT cache hit IN-MEMORY function=[%s] module_hash=[%s]", |
|
|
function_name, |
|
|
module_hash, |
|
|
) |
|
|
jit_executor = self.jit_cache[module_hash] |
|
|
|
|
|
self.post_compilation_cleanup() |
|
|
|
|
|
if compile_only: |
|
|
return jit_executor |
|
|
|
|
|
jit_executor.run_compiled_program(exe_args) |
|
|
|
|
|
return result |
|
|
|
|
|
def run_preprocessor(self, funcBody): |
|
|
if not hasattr(funcBody, "_preprocessed"): |
|
|
function_name = funcBody.__name__ |
|
|
self.funcBody = funcBody |
|
|
log().info("Started preprocessing [%s]", function_name) |
|
|
exec_globals = self._get_globals() |
|
|
transformed_ast = self.preprocessor.transform(funcBody, exec_globals) |
|
|
if self.envar.print_after_preprocessor: |
|
|
log().info( |
|
|
f"# Printing unparsed AST after preprocess of func=`{function_name}` id=`{id(funcBody)}`" |
|
|
) |
|
|
DSLPreprocessor.print_ast(transformed_ast) |
|
|
funcBody._preprocessed = True |
|
|
return transformed_ast |
|
|
return None |
|
|
|
|
|
def get_function_ptr(self, original_function, transformed_ast): |
|
|
file_name = inspect.getsourcefile(original_function) |
|
|
code_object = compile(transformed_ast, filename=file_name, mode="exec") |
|
|
return self.preprocessor.exec( |
|
|
original_function.__name__, |
|
|
original_function, |
|
|
code_object, |
|
|
self._get_globals(), |
|
|
) |
|
|
|
|
|
@lru_cache(maxsize=None) |
|
|
def _get_function_signature(self, func): |
|
|
return inspect.signature(func) |
|
|
|
|
|
def _get_function_bound_args(self, sig, func_name, *args, **kwargs): |
|
|
""" |
|
|
Binds provided arguments to a function's signature and applies default values. |
|
|
|
|
|
E.g. given a function signature `def foo(a, b=2, c=3)`, and at call-site if we do |
|
|
`foo(a=1, c=4)`, the returned BoundArguments object will have args = `[1]` |
|
|
and kwargs = `{'b': 2, 'c': 4}` |
|
|
|
|
|
An exception will be raised if binding fails. |
|
|
""" |
|
|
try: |
|
|
bound_args = sig.bind_partial(*args, **kwargs) |
|
|
bound_args.apply_defaults() |
|
|
except Exception as e: |
|
|
raise DSLRuntimeError( |
|
|
f"Failed to bind arguments to function `{func_name}` with signature `{sig}`", |
|
|
cause=e, |
|
|
) |
|
|
return bound_args |
|
|
|
|
|
def _canonicalize_args(self, *args, **kwargs): |
|
|
""" |
|
|
Canonicalize the input arguments so that returned args only contain |
|
|
positional arguments and kwargs only contain keyword arguments. |
|
|
""" |
|
|
sig = self._get_function_signature(self.funcBody) |
|
|
function_name = self.funcBody.__name__ |
|
|
bound_args = self._get_function_bound_args(sig, function_name, *args, **kwargs) |
|
|
canonicalized_args = bound_args.args |
|
|
canonicalized_kwargs = bound_args.kwargs |
|
|
return canonicalized_args, canonicalized_kwargs |
|
|
|
|
|
def _check_arg_count(self, *args, **kwargs): |
|
|
if not self.funcBody: |
|
|
raise DSLRuntimeError("Function body is not set.") |
|
|
|
|
|
|
|
|
sig = self._get_function_signature(self.funcBody) |
|
|
function_name = self.funcBody.__name__ |
|
|
|
|
|
bound_args = self._get_function_bound_args(sig, function_name, *args, **kwargs) |
|
|
|
|
|
|
|
|
for param in sig.parameters.values(): |
|
|
if ( |
|
|
param.default is inspect.Parameter.empty |
|
|
and param.name not in bound_args.arguments |
|
|
): |
|
|
raise DSLRuntimeError( |
|
|
f"Missing required argument in `{function_name}`: '{param.name}'" |
|
|
) |
|
|
|
|
|
def _func(self, funcBody, *args, **kwargs): |
|
|
"""Decorator for MLIR functions. |
|
|
It cuts the boilerplate code, does the following: |
|
|
1. Generates `func.func` |
|
|
2. Types translation (numpy arrays -> cute.memref, float -> <f32>, etc.) |
|
|
3. Compiles and JITs the MLIR module |
|
|
4. Invokes the generated function |
|
|
5. Operator overloading (a + b --> arith.addi a, b) |
|
|
6. Generates GPU kernel function with GPU module and kernel attributes baked |
|
|
""" |
|
|
if ir.Context.current is None: |
|
|
pass |
|
|
elif ir.InsertionPoint.current is not None: |
|
|
return funcBody(*args, **kwargs) |
|
|
|
|
|
function_name = funcBody.__name__ |
|
|
self.funcBody = funcBody |
|
|
|
|
|
pipeline = kwargs.pop("pipeline", None) |
|
|
gpu_module_attrs = kwargs.pop("gpu_module_attrs", {}) |
|
|
|
|
|
|
|
|
no_cache = kwargs.pop("no_cache", False) |
|
|
|
|
|
|
|
|
compile_only = kwargs.pop("compile_only", False) |
|
|
|
|
|
if not no_cache and compile_only: |
|
|
no_cache = True |
|
|
self.print_warning("Cache is disabled as user wants to compile only.") |
|
|
|
|
|
|
|
|
self._check_arg_count(*args, **kwargs) |
|
|
|
|
|
args_spec = inspect.getfullargspec(funcBody) |
|
|
|
|
|
|
|
|
canonicalized_args, canonicalized_kwargs = self._canonicalize_args( |
|
|
*args, **kwargs |
|
|
) |
|
|
|
|
|
|
|
|
function_name = self.mangle_name(function_name, canonicalized_args, args_spec) |
|
|
|
|
|
|
|
|
log().debug(f"Generating MLIR for function '{function_name}'") |
|
|
result = self.generate_mlir( |
|
|
funcBody, |
|
|
canonicalized_kwargs, |
|
|
function_name, |
|
|
gpu_module_attrs, |
|
|
canonicalized_args, |
|
|
args_spec, |
|
|
pipeline, |
|
|
no_cache, |
|
|
compile_only, |
|
|
) |
|
|
|
|
|
return result |
|
|
|
|
|
class _KernelGenHelper(ABC): |
|
|
def __init__(self): |
|
|
self.func_op = None |
|
|
self.func_type = None |
|
|
|
|
|
@abstractmethod |
|
|
def generate_func_op(self, arg_types, arg_attrs, kernel_name, loc=None): |
|
|
assert arg_types is not None, "Invalid arg_types!" |
|
|
assert kernel_name is not None, "kernel name is empty" |
|
|
pass |
|
|
|
|
|
@abstractmethod |
|
|
def generate_func_ret_op(self): |
|
|
pass |
|
|
|
|
|
@abstractmethod |
|
|
def generate_launch_op(self, *args, **kwargs): |
|
|
pass |
|
|
|
|
|
@abstractmethod |
|
|
def get_func_body_start(self): |
|
|
pass |
|
|
|
|
|
@abstractmethod |
|
|
def enter_gpu_module(module): |
|
|
"""Compute the insertion point into the given module.""" |
|
|
pass |
|
|
|
|
|
@lru_cache(maxsize=1) |
|
|
def _get_default_stream(self): |
|
|
"""Returns the default stream 0""" |
|
|
from .runtime import cuda as cuda_helpers |
|
|
|
|
|
return cuda_helpers.stream_create() |
|
|
|
|
|
def _execute_cuda( |
|
|
self, fname_cubin, kernel_name, grid_size, block_size, stream=None |
|
|
): |
|
|
""" |
|
|
Executes a specified CUDA kernel from a cubin file, handling module loading, |
|
|
kernel retrieval, stream creation, kernel launch, and synchronization. |
|
|
""" |
|
|
from .runtime import cuda as cuda_helpers |
|
|
|
|
|
|
|
|
module = cuda_helpers.load_cubin_module(fname_cubin) |
|
|
|
|
|
kernel_ptr = cuda_helpers.get_kernel_function(module, kernel_name) |
|
|
|
|
|
sync_execution_default = False |
|
|
if stream is None: |
|
|
stream = self._get_default_stream() |
|
|
sync_execution_default = True |
|
|
|
|
|
|
|
|
cuda_helpers.launch_kernel( |
|
|
kernel_ptr, |
|
|
grid_size, |
|
|
block_size, |
|
|
stream, |
|
|
smem_size=16000, |
|
|
kernel_args=self.exe_args, |
|
|
) |
|
|
|
|
|
if sync_execution_default: |
|
|
|
|
|
cuda_helpers.stream_sync(stream) |
|
|
|
|
|
def _execute_by_cuda_driver( |
|
|
self, kernel_generator, generate_cubin, grid_size, block_size, stream=None |
|
|
): |
|
|
""" |
|
|
This function builds IR and execute the module using cuda driver. |
|
|
It doesn't use mlir's cuda runtime |
|
|
""" |
|
|
ret = None |
|
|
|
|
|
|
|
|
with ir.Context(), ir.Location.unknown(): |
|
|
loc = self.get_location() |
|
|
module = ir.Module.create(loc=loc) |
|
|
unit_attr = ir.UnitAttr.get() |
|
|
module.operation.attributes["gpu.container_module"] = unit_attr |
|
|
with ir.InsertionPoint(module.body): |
|
|
self._build_gpu_module() |
|
|
ret, kernel_name = kernel_generator() |
|
|
log().debug( |
|
|
f"Kernel generator returned: ret={ret}, kernel_name={kernel_name}" |
|
|
) |
|
|
|
|
|
module = self.build_module(module, kernel_name) |
|
|
|
|
|
|
|
|
if self.envar.dryrun: |
|
|
return ret |
|
|
|
|
|
|
|
|
fname_cubin = generate_cubin(module, kernel_name) |
|
|
|
|
|
|
|
|
if block_size is None: |
|
|
|
|
|
block_size = self.block_size |
|
|
self._execute_cuda(fname_cubin, kernel_name, grid_size, block_size, stream) |
|
|
|
|
|
return ret |
|
|
|
|
|
def generate_kernel_operands_and_types( |
|
|
self, kernel_func, kernel_name, args_spec, args, kwargs |
|
|
): |
|
|
""" |
|
|
Generate the operands and types for the kernel function |
|
|
""" |
|
|
|
|
|
kernel_operands, kernel_arg_types, kernel_arg_attrs = [], [], [] |
|
|
|
|
|
log().debug( |
|
|
"Processing GPU kernel call in [%s] mode", |
|
|
( |
|
|
f"Only {self.device_jit_decorator_name}" |
|
|
if self.device_compilation_only |
|
|
else f"{self.host_jit_decorator_name} + {self.device_jit_decorator_name}" |
|
|
), |
|
|
) |
|
|
|
|
|
if self.device_compilation_only: |
|
|
return kernel_operands, kernel_arg_types, kernel_arg_attrs |
|
|
|
|
|
kernel_operands, kernel_arg_types, kernel_arg_attrs = ( |
|
|
self._generate_jit_func_args( |
|
|
kernel_func, kernel_name, args, kwargs, args_spec, is_host=False |
|
|
) |
|
|
) |
|
|
|
|
|
log().debug("Final kernel_operands: %s", ", ".join(map(str, kernel_operands))) |
|
|
log().debug("Final kernel_arg_types: %s", ", ".join(map(str, kernel_arg_types))) |
|
|
log().debug("Final kernel_arg_attrs: %s", ", ".join(map(str, kernel_arg_attrs))) |
|
|
|
|
|
assert ( |
|
|
len(kernel_operands) == len(kernel_arg_types) == len(kernel_arg_attrs) |
|
|
), "Size of kernel_operands, kernel_arg_types and kernel_arg_attrs must be equal" |
|
|
|
|
|
return kernel_operands, kernel_arg_types, kernel_arg_attrs |
|
|
|
|
|
def kernel_launcher(self, *dargs, **dkwargs): |
|
|
def decorator(funcBody): |
|
|
@wraps(funcBody) |
|
|
def kernel_wrapper(*args, **kwargs): |
|
|
""" |
|
|
Base decorator for generating kernel function |
|
|
|
|
|
This decorator provides a template for kernel function generation |
|
|
including kernel function header/body and kernel launch op at call site |
|
|
|
|
|
Optional arguments (with default value in <>): |
|
|
- requiredArgs <[]>: specifies the mandatory arguments that must present in kernel function signature |
|
|
the args will be validated and collected as a namedtuple |
|
|
- optionalArgs <[]>: specifies the optional arguments that might present in kernel function signature |
|
|
the args will be collected (if present) as a namedtuple |
|
|
- unitAttrNames <[]>: specifies the name(s) of ir.UnitAttr to be set for kernel function op |
|
|
- valueAttrDict <{}>: specifies the name(s) and value(s) of ir.Attribute to be set for kernel function op |
|
|
- kernelGenHelper <None>: specifies the mandatory customized kernel generation helper class (derived from _KernelGenHelper) |
|
|
|
|
|
Return value: |
|
|
A namedtuple "KernelReturns" is returned with following fields: |
|
|
- kernel_func_ret: the return of the kernel function |
|
|
- launch_op_ret: the return of the launch op |
|
|
""" |
|
|
|
|
|
requiredArgs = dkwargs.get("requiredArgs", []) |
|
|
optionalArgs = dkwargs.get("optionalArgs", []) |
|
|
unitAttrNames = dkwargs.get("unitAttrNames", []) |
|
|
valueAttrDict = dkwargs.get("valueAttrDict", {}) |
|
|
kernelGenHelper = dkwargs.get("kernelGenHelper", None) |
|
|
|
|
|
kernel_name = funcBody.__name__ |
|
|
args_spec = inspect.getfullargspec(funcBody) |
|
|
self.funcBody = funcBody |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
kernel_name = f"kernel_{self.mangle_name(kernel_name, args, args_spec)}_{self.num_kernels}" |
|
|
self.num_kernels += 1 |
|
|
|
|
|
|
|
|
def extract_args(argNames, assertIfNone=False) -> list: |
|
|
extracted = [] |
|
|
for name in argNames: |
|
|
value = kwargs.pop(name, None) |
|
|
if assertIfNone and value is None: |
|
|
raise DSLRuntimeError( |
|
|
f"{name} is required for {kernel_name}" |
|
|
) |
|
|
extracted.append(value) |
|
|
|
|
|
return extracted |
|
|
|
|
|
RequiredArgs = namedtuple("RequiredArgs", requiredArgs) |
|
|
req_args = ( |
|
|
RequiredArgs._make(extract_args(requiredArgs, assertIfNone=True)) |
|
|
if requiredArgs |
|
|
else None |
|
|
) |
|
|
OptionalArgs = namedtuple("OptionalArgs", optionalArgs) |
|
|
opt_args = ( |
|
|
OptionalArgs._make(extract_args(optionalArgs)) |
|
|
if optionalArgs |
|
|
else None |
|
|
) |
|
|
assert ( |
|
|
kernelGenHelper is not None |
|
|
), "kernelGenHelper should be explicitly specified!" |
|
|
|
|
|
|
|
|
self._check_arg_count(*args, **kwargs) |
|
|
|
|
|
|
|
|
canonicalized_args, canonicalized_kwargs = self._canonicalize_args( |
|
|
*args, **kwargs |
|
|
) |
|
|
|
|
|
kernel_operands, kernel_types, kernel_arg_attrs = ( |
|
|
self.generate_kernel_operands_and_types( |
|
|
funcBody, |
|
|
kernel_name, |
|
|
args_spec, |
|
|
canonicalized_args, |
|
|
canonicalized_kwargs, |
|
|
) |
|
|
) |
|
|
|
|
|
with self._enter_gpu_module(): |
|
|
log().debug("Generating device kernel") |
|
|
if self.device_compilation_only: |
|
|
log().debug("Generating cuda-python arguments") |
|
|
|
|
|
self.exe_args, kernel_types = self.generate_mlir_function_types( |
|
|
funcBody, |
|
|
kernel_name, |
|
|
canonicalized_args, |
|
|
canonicalized_kwargs, |
|
|
args_spec, |
|
|
) |
|
|
|
|
|
helper = kernelGenHelper() |
|
|
loc = self.get_location() |
|
|
fop = helper.generate_func_op( |
|
|
kernel_types, kernel_arg_attrs, kernel_name, loc |
|
|
) |
|
|
log().debug(f"Kernel function op: {fop}") |
|
|
for attr in unitAttrNames: |
|
|
fop.attributes[attr] = ir.UnitAttr.get() |
|
|
for key, val in valueAttrDict.items(): |
|
|
fop.attributes[key] = val |
|
|
|
|
|
fop.sym_visibility = ir.StringAttr.get("public") |
|
|
with ir.InsertionPoint(helper.get_func_body_start()): |
|
|
ir_args, ir_kwargs = self.generate_execution_arguments( |
|
|
canonicalized_args, canonicalized_kwargs, fop, args_spec |
|
|
) |
|
|
log().debug( |
|
|
f"IR arguments - args: {ir_args} ; kwargs: {ir_kwargs}" |
|
|
) |
|
|
|
|
|
kernel_ret = funcBody(*ir_args, **ir_kwargs) |
|
|
helper.generate_func_ret_op() |
|
|
|
|
|
|
|
|
kernel_sym = ir.SymbolRefAttr.get(["kernels", kernel_name]) |
|
|
launch_ret = helper.generate_launch_op( |
|
|
kernelSym=kernel_sym, |
|
|
kernelOperands=kernel_operands, |
|
|
requiredArgs=req_args, |
|
|
optionalArgs=opt_args, |
|
|
) |
|
|
|
|
|
KernelReturns = namedtuple( |
|
|
"KernelReturns", ["kernel_func_ret", "launch_op_ret"] |
|
|
) |
|
|
result = KernelReturns( |
|
|
kernel_func_ret=kernel_ret, launch_op_ret=launch_ret |
|
|
) |
|
|
log().debug(f"Kernel result: {result}, kernel name: {kernel_name}") |
|
|
return result, kernel_name |
|
|
|
|
|
return kernel_wrapper |
|
|
|
|
|
if len(dargs) == 1 and callable(dargs[0]): |
|
|
return decorator(dargs[0]) |
|
|
else: |
|
|
return decorator |
|
|
|