VidChain-exercise
/
VTimeLLM
/flash-attention
/csrc
/cutlass
/python
/CuTeDSL
/cutlass_dsl
/cutlass.py
| # SPDX-FileCopyrightText: Copyright (c) 2025 NVIDIA CORPORATION & AFFILIATES. All rights reserved. | |
| # SPDX-License-Identifier: LicenseRef-NvidiaProprietary | |
| # | |
| # Use of this software is governed by the terms and conditions of the | |
| # NVIDIA End User License Agreement (EULA), available at: | |
| # https://docs.nvidia.com/cutlass/media/docs/pythonDSL/license.html | |
| # | |
| # Any use, reproduction, disclosure, or distribution of this software | |
| # and related documentation outside the scope permitted by the EULA | |
| # is strictly prohibited. | |
| """ | |
| This module provides a DSL for Cutlass Dialects. It also includes utils with | |
| regarding to that dialect. | |
| """ | |
| # Local module imports | |
| from typing import Callable, Union, Type, List, Union, Sequence, ForwardRef | |
| from inspect import isclass | |
| import functools | |
| import pkgutil | |
| from dataclasses import is_dataclass | |
| from ..base_dsl import * | |
| from ..base_dsl import compiler | |
| from ..base_dsl.dsl import is_dynamic_expression, extract_mlir_values | |
| from ..base_dsl.typing import * | |
| from ..base_dsl.typing import DynamicExpression, get_mlir_types | |
| from ..base_dsl.runtime.jit_arg_adapters import is_arg_spec_constexpr | |
| from ..base_dsl.ast_helpers import const_expr | |
| # MLIR Imports | |
| from cutlass._mlir import ir, execution_engine, passmanager | |
| from cutlass._mlir.dialects import arith, func, gpu, scf, cute, gpu as cutlass_gpu | |
| from cutlass._mlir.dialects._ods_common import ( | |
| get_op_result_or_op_results as _get_op_result_or_op_results, | |
| ) | |
| from cutlass._mlir.extras import types as T | |
| # Helpers | |
| from ..base_dsl._mlir_helpers import arith as cutlass_arith | |
| from ..base_dsl._mlir_helpers import lru_cache_ir | |
| from ..base_dsl.ast_helpers import ( | |
| loop_selector, | |
| executor, | |
| if_selector, | |
| if_executor, | |
| while_selector, | |
| while_executor, | |
| assert_executor, | |
| bool_cast, | |
| ) | |
| from ..base_dsl.runtime.dlpack_runtime import ( | |
| get_cute_tensor_c_pointer, | |
| get_tensor_desc_shape_all, | |
| get_tensor_desc_stride_all, | |
| get_tensor_desc_element_type, | |
| get_tensor_desc_is_in_device, | |
| get_tensor_desc_assumed_align, | |
| ) | |
| from .cutlass_ast_decorators import ( | |
| _loop_execute_range_dynamic, | |
| _if_execute_dynamic, | |
| _while_execute_dynamic, | |
| ) | |
| # ============================================================================= | |
| # Set the AST decorator | |
| # ============================================================================= | |
| # Set the DSL specific functions | |
| executor.set_functions( | |
| is_dynamic_expression, | |
| _loop_execute_range_dynamic, | |
| _if_execute_dynamic, | |
| _while_execute_dynamic, | |
| ) | |
| # ============================================================================= | |
| # Cutlass DSL Base Abstract Class | |
| # ============================================================================= | |
| # Return a ctype class that represents the in-memory layout expected | |
| # for a CuTe hierarchical tuple type. | |
| def get_sparse_tuple_ctype(dyn): | |
| # When there is a single dynamic value, the sparse CuTe | |
| # representation is a single integer. | |
| if isinstance(dyn, int): | |
| return ctypes.c_int32 | |
| # For zero or greater than 1 dynamic values, the tuple | |
| # representation will be a struct with a field for each dynamic | |
| # value. The representation is flattened, even for hierarchical CuTe | |
| # profiles (although we are only dealing with depth 1 inputs here). | |
| class TupleDescriptor(ctypes.Structure): | |
| _fields_ = [(f"x{idx}", ctypes.c_int32) for idx in range(len(dyn))] | |
| def __str__(self): | |
| return f"struct<{str(self._fields_)}>" | |
| return TupleDescriptor | |
| def is_cute_algebra_type(arg_spec): | |
| # Walk through the arg_spec to check if it's a cute algebra type | |
| _cute_algebra_type_aliases = ( | |
| "Shape", | |
| "Stride", | |
| "Coord", | |
| "Tile", | |
| "IntTuple", | |
| ) | |
| origin = get_origin(arg_spec) | |
| if origin is Union: | |
| for sub_ty in get_args(arg_spec): | |
| sub_origin = get_origin(sub_ty) | |
| if sub_origin is Tuple or ( | |
| type(sub_origin) is type and issubclass(sub_origin, tuple) | |
| ): | |
| tuple_arg0 = get_args(sub_ty)[0] | |
| if isinstance( | |
| tuple_arg0, ForwardRef | |
| ) and tuple_arg0.__forward_arg__ in (_cute_algebra_type_aliases): | |
| return True | |
| return False | |
| class CutlassBaseDSL(BaseDSL): | |
| """This abstract class provides a DSL for Cutlass.""" | |
| def __init__( | |
| self, | |
| name: str, | |
| compiler_provider: Any, | |
| pass_sm_arch_name: str, | |
| device_compilation_only: bool = False, | |
| preprocess: bool = False, | |
| ): | |
| super().__init__( | |
| name, | |
| compiler_provider, | |
| pass_sm_arch_name, | |
| device_compilation_only, | |
| preprocess, | |
| ) | |
| def _is_tensor_descriptor(self, maybe_tensor_descriptor) -> bool: | |
| return False | |
| def _build_gpu_module(self, attrs): | |
| self.gpu_module = gpu.GPUModuleOp(ir.StringAttr.get("kernels")) | |
| with ir.InsertionPoint(self.gpu_module.bodyRegion.blocks.append(*[])): | |
| pass | |
| for attr_name in attrs: | |
| self.gpu_module.attributes[attr_name] = ir.Attribute.parse(attrs[attr_name]) | |
| def _get_pipeline(self, pipeline): | |
| pipeline = super()._get_pipeline(pipeline) | |
| if pipeline == None: | |
| # cubin format is required to be cubin as we launch cuda module at python level. | |
| return "builtin.module(cute-to-nvvm{cubin-format=bin opt-level=3})" | |
| return pipeline | |
| def preprocess_pipeline(self, pipeline, arch) -> str: | |
| pipeline = super().preprocess_pipeline(pipeline, arch) | |
| pipeline = pipeline.rstrip(")") + ",external-kernel-for-gpu-launch)" | |
| return pipeline | |
| def _enter_gpu_module(self): | |
| return ir.InsertionPoint(self.gpu_module.bodyRegion.blocks[0]) | |
| def _generate_kernel_attrs(self, config: BaseDSL.LaunchConfig) -> dict: | |
| assert isinstance( | |
| config, BaseDSL.LaunchConfig | |
| ), f"Expect LaunchConfig for @kernel, but got {type(config)}" | |
| ret = {} | |
| # generate launch bound attr from LaunchConfig | |
| max_threads = ", ".join(map(str, config.block)) | |
| ret["nvvm.reqntid"] = ir.Attribute.parse(f"array<i32 : {max_threads}>") | |
| # min_blocks_per_mp is optional for kernel | |
| min_blocks = config.min_blocks_per_mp | |
| if min_blocks > 0: | |
| ret["nvvm.minctasm"] = ir.Attribute.parse(f"{min_blocks} : i32") | |
| return ret | |
| def get_version(self): | |
| """ | |
| Get the version of cutlass dsl, used for computing the hash key of the cache. | |
| Including source python files and the shared library. | |
| """ | |
| dsl_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) | |
| # get the version hash of the cutlass shared library | |
| version_hash = hashlib.sha256() | |
| # update the version hash of the source python files | |
| for lib in pkgutil.walk_packages([dsl_path], prefix="cutlass."): | |
| try: | |
| with open(lib.module_finder.find_spec(lib.name).origin, "rb") as f: | |
| version_hash.update(f.read()) | |
| except Exception: | |
| raise DSLRuntimeError( | |
| f"Failed to read module file {lib.name}. The file may not exist or may not be readable." | |
| "Please re-install the package." | |
| ) | |
| try: | |
| # update the version hash of the cutlass shared library | |
| with open( | |
| os.path.join(dsl_path, "_mlir/_mlir_libs/libCutlassIRPythonCAPI.so"), | |
| "rb", | |
| ) as f: | |
| while True: | |
| chunk = f.read(1024**2) | |
| if not chunk: | |
| break | |
| version_hash.update(chunk) | |
| except Exception: | |
| raise DSLRuntimeError( | |
| f"Failed to read the shared library file libCutlassIRPythonCAPI.so." | |
| "The file may not exist or may not be readable." | |
| "Please re-install the package." | |
| ) | |
| return version_hash | |
| def _kernel_helper(self, funcBody, *args, **kwargs): | |
| class _CutlassIrKernelGenHelper(BaseDSL._KernelGenHelper): | |
| def generate_func_op(self, arg_types, arg_attrs, kernel_name, loc=None): | |
| super().generate_func_op(arg_types, arg_attrs, kernel_name) | |
| self.func_op = func.FuncOp( | |
| kernel_name, ir.FunctionType.get(arg_types, []), loc=loc | |
| ) | |
| if arg_attrs is not None: | |
| log().debug(arg_attrs) | |
| self.func_op.arg_attrs = arg_attrs | |
| return self.func_op | |
| def generate_func_ret_op(self): | |
| return func.ReturnOp([]) | |
| def get_func_body_start(self): | |
| assert self.func_op is not None, "Invalid func_op is not expected!" | |
| return self.func_op.add_entry_block() | |
| def generate_launch_op(self, *args, **kwargs): | |
| # Extract args and do validation | |
| kernelSym = kwargs.get("kernelSym", None) | |
| kernelOperands = kwargs.get("kernelOperands", None) | |
| requiredArgs = kwargs.get("requiredArgs", None) | |
| assert kernelSym is not None, "kernelSym being None is not expected!" | |
| assert ( | |
| requiredArgs is not None | |
| ), "requiredArgs being None is not expected!" | |
| assert ( | |
| kernelOperands is not None | |
| ), "kernelOperands being None is not expected!" | |
| assert isinstance( | |
| requiredArgs.config, BaseDSL.LaunchConfig | |
| ), f"Expect LaunchConfig for @kernel, but got {type(requiredArgs.config)}" | |
| cfg = requiredArgs.config | |
| # Apply to grid, block, and cluster if present | |
| cfg.grid = [to_index(size) for size in cfg.grid] | |
| cfg.block = [to_index(size) for size in cfg.block] | |
| if cfg.has_cluster: | |
| cfg.cluster = [to_index(size) for size in cfg.cluster] | |
| cfg.smem = const(cfg.smem) | |
| if not isinstance(cfg.async_deps, (list, tuple)): | |
| cfg.async_deps = [cfg.async_deps] | |
| is_async = len(cfg.async_deps) > 0 | |
| token = gpu.launch_func( | |
| gpu.AsyncTokenType.get() if is_async else None, | |
| cfg.async_deps, | |
| kernelSym, | |
| *cfg.grid, | |
| *cfg.block, | |
| kernelOperands, | |
| **dict( | |
| zip( | |
| ("cluster_size_x", "cluster_size_y", "cluster_size_z"), | |
| tuple(cfg.cluster), | |
| ) | |
| ), | |
| dynamic_shared_memory_size=cfg.smem, | |
| ) | |
| return token if is_async else None | |
| return KernelLauncher( | |
| self, _CutlassIrKernelGenHelper, funcBody, *args, **kwargs | |
| ) | |
| def _get_globals(self): | |
| caller_globals = self.frame.f_globals | |
| caller_locals = self.frame.f_locals | |
| all_globals = globals().copy() | |
| all_globals.update(caller_globals) | |
| all_globals.update(caller_locals) | |
| return all_globals | |
| def _preprocess_launch_config_args(self, args, kwargs): | |
| """Helper to preprocess args and kwargs for LaunchConfig""" | |
| if "stream" in kwargs: | |
| kwargs["async_deps"] = kwargs.pop("stream") | |
| def mangle_name(self, function_name, args, args_spec: inspect.FullArgSpec): | |
| """Mangle the name of the function to avoid conflicts with other functions""" | |
| function_name = "cutlass_" + function_name | |
| return super().mangle_name(function_name, args, args_spec) | |
| def _validate_arg(self, arg, arg_index, arg_name, arg_annotation): | |
| """ | |
| Validates if the arg is really of the annotated type. | |
| """ | |
| if is_arg_spec_constexpr(arg_annotation, arg_name, arg_index, None): | |
| pass | |
| else: | |
| origin = get_origin(arg_annotation) | |
| # Handle special case where annotation is Type[X] but arg is an actual type | |
| if origin is type and isinstance(arg, type): | |
| # Get the expected base type from Type[X] | |
| expected_base = get_args(arg_annotation)[0] | |
| if not issubclass(arg, expected_base): | |
| return DSLRuntimeError( | |
| f"expects argument #{arg_index+1} ({arg_name}) to be Type[{expected_base}], but got {arg}" | |
| ) | |
| # Handle Union types and generic types | |
| elif origin is Union: | |
| # For Union types, check if arg matches any of the allowed types | |
| allowed_types = get_args(arg_annotation) | |
| if not any( | |
| (isinstance(ty, type) and isinstance(arg, ty)) | |
| or (get_origin(ty) is tuple and isinstance(arg, tuple)) | |
| for ty in allowed_types | |
| ): | |
| return DSLRuntimeError( | |
| f"expects argument #{arg_index+1} ({arg_name}) to be one of {allowed_types}, but got {type(arg)}" | |
| ) | |
| elif isinstance(arg_annotation, type): | |
| # Handle simple type annotations | |
| if not isinstance(arg, arg_annotation) and arg is not None: | |
| return DSLRuntimeError( | |
| f"expects argument #{arg_index+1} ({arg_name}) to be {arg_annotation}, but got {type(arg)}" | |
| ) | |
| # Everything looks good if we are here | |
| return None | |
| def _generate_jit_func_args_for_known_types( | |
| self, | |
| func, | |
| arg, | |
| arg_name, | |
| arg_spec, | |
| arg_index, | |
| *, | |
| is_host=True, | |
| ): | |
| jit_arg_type, jit_arg_attr, jit_exec_arg = [], [], [] | |
| default_attr = ir.DictAttr.get({}) | |
| ( | |
| jit_exec_arg, | |
| jit_arg_type, | |
| jit_arg_attr, | |
| ) = super()._generate_jit_func_args_for_known_types( | |
| func, arg, arg_name, arg_spec, arg_index, is_host=is_host | |
| ) | |
| if jit_arg_type is not None and len(jit_arg_type) == 0: | |
| # Handle DSL specific types | |
| if is_cute_algebra_type(arg_spec): | |
| dyn_vals = extract_mlir_values(arg) | |
| if dyn_vals: | |
| # Handle dynamic types | |
| jit_arg_type.extend([v.type for v in dyn_vals]) | |
| jit_arg_attr.extend([default_attr] * len(dyn_vals)) | |
| jit_exec_arg.extend(get_c_pointers(arg) if is_host else dyn_vals) | |
| else: | |
| jit_exec_arg = jit_arg_type = jit_arg_attr = None | |
| return jit_exec_arg, jit_arg_type, jit_arg_attr | |
| def _generate_execution_arguments_for_known_types( | |
| self, arg, arg_spec, arg_name, i, fop_args, iv_block_args | |
| ): | |
| ir_arg, iv_block_args = super()._generate_execution_arguments_for_known_types( | |
| arg, arg_spec, arg_name, i, fop_args, iv_block_args | |
| ) | |
| if not ir_arg: | |
| # Handling DSL specific types | |
| if is_cute_algebra_type(arg_spec): | |
| n_args = len(get_mlir_types(arg)) | |
| blk_args = fop_args[iv_block_args : iv_block_args + n_args] | |
| ir_arg.append(new_from_mlir_values(arg, blk_args)) | |
| iv_block_args += n_args | |
| return ir_arg, iv_block_args | |
| # ============================================================================= | |
| # Cute DSL Class | |
| # ============================================================================= | |
| class CuTeDSL(CutlassBaseDSL): | |
| """ | |
| This is a concrete DSL subclass for the CuTe dialect. | |
| """ | |
| def __init__(self): | |
| name = "CUTE_DSL" | |
| compiler_provider = compiler.Compiler(passmanager, execution_engine) | |
| pass_sm_arch_name = "cubin-chip" | |
| super().__init__(name, compiler_provider, pass_sm_arch_name, preprocess=True) | |
| # ============================================================================= | |
| # KernelLauncher | |
| # ============================================================================= | |
| class KernelLauncher: | |
| """ | |
| This class is used to launch a kernel function. | |
| Usage: | |
| ```python | |
| @cute.kernel | |
| def kernel(arg1, arg2, ...): | |
| ... | |
| @cute.jit | |
| def launch_kernel(): | |
| kernel(arg1, arg2, ...).launch(grid=[1, 1, 1], block=[1, 1, 1], ...) | |
| # or | |
| kernel(arg1, arg2, ...)(grid=[1, 1, 1], block=[1, 1, 1], ...) | |
| ``` | |
| """ | |
| def __init__( | |
| self, | |
| dsl: "CutlassBaseDSL", | |
| kernelGenHelper: BaseDSL._KernelGenHelper, | |
| funcBody, | |
| *func_args, | |
| **func_kwargs, | |
| ): | |
| self.dsl = dsl | |
| self.kernelGenHelper = kernelGenHelper | |
| self.funcBody = funcBody | |
| self.func_args = func_args | |
| self.func_kwargs = func_kwargs | |
| self._check_func_args(funcBody, *func_args, **func_kwargs) | |
| def _check_func_args(self, funcBody, *func_args, **func_kwargs): | |
| # Get function signature | |
| sig = inspect.signature(funcBody) | |
| # func_args and func_kwargs should match funcBody's signature, | |
| # no extra or missing arguments. | |
| try: | |
| sig.bind(*func_args, **func_kwargs) | |
| except TypeError as e: | |
| raise DSLRuntimeError( | |
| f"Failed to bind arguments to function `{funcBody.__name__}` with signature `{sig}`", | |
| cause=e, | |
| ) | |
| def launch(self, *args, **kwargs): | |
| self.dsl.frame = inspect.currentframe().f_back | |
| self.dsl._preprocess_launch_config_args(args, kwargs) | |
| config = self.dsl.LaunchConfig(*args, **kwargs) | |
| kernel_generator = self.dsl.kernel_launcher( | |
| requiredArgs=["config"], | |
| unitAttrNames=["gpu.kernel", "cute.kernel"], | |
| valueAttrDict=self.dsl._generate_kernel_attrs(config), | |
| kernelGenHelper=self.kernelGenHelper, | |
| )(self.funcBody) | |
| ret, name = kernel_generator(*self.func_args, **self.func_kwargs, config=config) | |
| self.dsl.kernel_symbols.append(name) | |
| return ret.launch_op_ret | |
| def __call__(self, *args, **kwargs): | |
| return self.launch(*args, **kwargs) | |
| # ============================================================================= | |
| # Utils | |
| # ============================================================================= | |
| def is_frozen_dataclass(obj_or_cls) -> bool: | |
| """ | |
| Return True if obj_or_cls is a dataclass (class or instance) declared with frozen=True, | |
| otherwise False. | |
| """ | |
| if not isinstance(obj_or_cls, type): | |
| # If it's an instance, get its class | |
| obj_or_cls = obj_or_cls.__class__ | |
| # Must be a dataclass, and __dataclass_params__.frozen must be True | |
| return ( | |
| is_dataclass(obj_or_cls) | |
| and getattr(obj_or_cls, "__dataclass_params__", None) is not None | |
| and obj_or_cls.__dataclass_params__.frozen | |
| ) | |
| def pack_from_irvalue( | |
| ir_values: List["ir.Value"], | |
| indices: Dict[int, Tuple[int, int]], | |
| class_types: List[Any], | |
| ) -> List[Any]: | |
| """ | |
| Packs MLIR values into a list of mixed values. | |
| """ | |
| log().debug("===--- Values Pack (%d)", len(ir_values)) | |
| for idx, packed in enumerate(ir_values): | |
| log().debug("[%d]: will-packed: %s", idx, ir_values) | |
| for idx, unpacked in indices.items(): | |
| log().debug("[%d]: indices: %s", idx, unpacked) | |
| for idx, c in enumerate(class_types): | |
| log().debug("[%d]: obj-types: %s", idx, type(c)) | |
| mixed_values = [None] * len(indices) | |
| for idx, (start, length) in sorted(indices.items()): | |
| chunk = ir_values[start : start + length] | |
| obj = class_types[idx] | |
| if is_frozen_dataclass(obj): | |
| mixed_values[idx] = obj | |
| elif not isinstance(obj, type) and hasattr(obj, "__new_from_mlir_values__"): | |
| mixed_values[idx] = obj.__new_from_mlir_values__(chunk) | |
| else: | |
| try: | |
| if isinstance(chunk, list) and chunk[0] is None: | |
| mixed_values[idx] = class_types[idx] | |
| else: | |
| mixed_values[idx] = t.as_numeric(chunk[0]) | |
| except DSLRuntimeError as e: | |
| mixed_values[idx] = chunk[0] | |
| log().debug("------------------ ") | |
| for idx, packed in enumerate(mixed_values): | |
| log().debug("[%d]: packed: %s", idx, packed) | |
| log().debug("------------------ ") | |
| return mixed_values | |
| def unpack_to_irvalue( | |
| mixed_values: List[Any], body_name: str | |
| ) -> Tuple[List[ir.Value], List[Any], Dict[int, Tuple[int, int]], List[Any]]: | |
| """ | |
| Unpacks mixed values into ir.Value values. | |
| """ | |
| unpacked_values = [] | |
| ir_values = [] | |
| indices = {} | |
| class_types = [] | |
| current_offset = 0 | |
| log().debug("===--- Values UNPack (%d)", len(mixed_values)) | |
| for idx, packed in enumerate(mixed_values): | |
| log().debug("[%d]: will-unpacked: [type:%s] %s", idx, type(packed), packed) | |
| for idx, item in enumerate(mixed_values): | |
| class_types.append(item) | |
| try: | |
| if is_frozen_dataclass(item): | |
| extracted_vals = [None] | |
| else: | |
| extracted_vals = extract_mlir_values(item) | |
| # it's consexpr (python value), so we create mlir value for it | |
| if extracted_vals == []: | |
| if item is None: | |
| extracted_vals = [None] | |
| else: | |
| dyn_expr = t.as_numeric(item) | |
| extracted_vals = extract_mlir_values(dyn_expr) | |
| ir_values.extend(extracted_vals) | |
| else: | |
| ir_values.extend(extracted_vals) | |
| unpacked_values.extend(extracted_vals) | |
| length = len(extracted_vals) | |
| indices[idx] = (current_offset, length) | |
| current_offset += length | |
| except Exception as e: | |
| raise DSLRuntimeError( | |
| f"The '{body_name}' statement encountered a user-defined Python object, which cannot be automatically converted into an dynamic expression (aka MLIR value).", | |
| context={ | |
| item: ( | |
| f"All expressions within '{body_name}' must be dynamic expressions, " | |
| "mixing Python objects and dynamic expressions (aka MLIR values) is not supported. " | |
| "The DSL failed to convert the Python object into MLIR values." | |
| ) | |
| }, | |
| suggestion=( | |
| f"Please ensure '{item}' implements the '{DynamicExpression.__name__}', " | |
| f"so it can be treated as a valid dynamic expression or mark '{body_name}' as a constant expression if conditions are Python objects." | |
| ), | |
| ) from e | |
| log().debug("------------------ ") | |
| for idx, unpacked in enumerate(unpacked_values): | |
| log().debug("[%d]: unpacked values: %s", idx, unpacked) | |
| for idx, unpacked in enumerate(ir_values): | |
| log().debug("[%d]: unpacked ir_values: %s", idx, unpacked) | |
| for idx, unpacked in indices.items(): | |
| log().debug("[%d]: indices: %s", idx, unpacked) | |
| for idx, unpacked in enumerate(class_types): | |
| log().debug("[%d]: initial-class-types: %s", idx, unpacked) | |
| log().debug("------------------ ") | |
| return ir_values, unpacked_values, indices, class_types | |
| def to_index(value): | |
| """Converts a value to an index, either by casting or coercing to int.""" | |
| if is_dynamic_expression(value): | |
| if isinstance(value, Numeric): | |
| value = value.ir_value() | |
| assert ir.IntegerType.isinstance( | |
| value.type | |
| ), f"expects integer type, but got {value.type}" | |
| res = arith.index_cast(T.index(), value) | |
| else: | |
| res = const(int(value), ty=T.index()) | |
| return res | |
| def _validate_iter_args_structure(iter_args, ir_values): | |
| """ | |
| Validates that iter_args structure contains the same number of atomic values | |
| as there are IR values. | |
| Args: | |
| iter_args: Original iteration arguments, possibly nested sequences | |
| ir_values: Flattened MLIR values extracted from iter_args | |
| Returns: | |
| bool: True if the number of atomic values in iter_args matches | |
| the number of values in ir_values | |
| """ | |
| # Handle non-sequence case | |
| if not isinstance(iter_args, (tuple, list, set)): | |
| return not isinstance(ir_values, (tuple, list, set)) or len(ir_values) == 1 | |
| # If we have a sequence but ir_values isn't one, there's a mismatch | |
| if not isinstance(ir_values, (tuple, list, set)): | |
| return False | |
| # Count all non-sequence values recursively | |
| def count_values(args): | |
| if not isinstance(args, (tuple, list, set)): | |
| return 1 | |
| else: | |
| return sum(count_values(arg) for arg in args) | |
| return count_values(iter_args) == len(ir_values) | |
| # ============================================================================= | |
| # DSL implementation of Python Build-in Operators | |
| # ============================================================================= | |
| def _minmax(op, *args, loc=None, ip=None): | |
| """Computes the minimum or maximum value from the provided arguments.""" | |
| from ..base_dsl.typing import _binary_op, _binary_op_type_promote | |
| # AST Traversal doesn't support early exit in if executor | |
| x = None | |
| res = None | |
| if len(args) == 1: | |
| # Handle case for min([a, b, c, d, ..]) | |
| if hasattr(args[0], "__iter__"): | |
| x = op(*tuple(args[0])) | |
| # Handle case for min(a) | |
| else: | |
| x = args[0] | |
| # Handle case for min(a, b, c, ...) and min([x, y], [b]) and min(a, (x, y, z)) | |
| elif len(args) > 1: | |
| res, *xs = tuple(args) | |
| for x in xs: | |
| lhs = as_numeric(op(res, loc=loc, ip=ip)) | |
| rhs = as_numeric(op(x, loc=loc, ip=ip)) | |
| emitter = getattr(cutlass_arith, f"_{op.__name__}") | |
| lhs, rhs, res_type = _binary_op_type_promote(lhs, rhs, promote_bool=True) | |
| if isinstance(lhs.value, cutlass_arith.ArithValue) and isinstance( | |
| lhs, Integer | |
| ): | |
| lhs_val = lhs.value.with_signedness(lhs.signed) | |
| else: | |
| lhs_val = lhs.value | |
| if isinstance(rhs.value, cutlass_arith.ArithValue) and isinstance( | |
| rhs, Integer | |
| ): | |
| rhs_val = rhs.value.with_signedness(rhs.signed) | |
| else: | |
| rhs_val = rhs.value | |
| res = res_type(emitter(lhs_val, rhs_val), loc=loc, ip=ip) | |
| x = res | |
| else: | |
| raise DSLNotImplemented(f"{type(args)} is not supported") | |
| return x | |
| def min(*args, loc=None, ip=None): | |
| """Computes the minimum value from the provided arguments. | |
| This function differs from Python's built-in min() in that the return type | |
| is determined by the static types of the inputs, not their dynamic values. | |
| :param args: One or more values or iterables to find the minimum of | |
| :type args: tuple | |
| :param loc: Source location for MLIR operation tracking | |
| :type loc: object, optional | |
| :param ip: Insertion point for MLIR operation | |
| :type ip: object, optional | |
| :return: The minimum value among all inputs | |
| :rtype: Numeric | |
| :raises DSLNotImplemented: If the input type is not supported | |
| Supports multiple calling patterns: | |
| - min(a): Returns a | |
| - min([a, b, c, ...]): Returns minimum of all elements in the iterable | |
| - min(a, b, c, ...): Returns minimum of all arguments | |
| - min([x, y], [b]): Returns minimum across all elements in all iterables | |
| - min(a, (x, y, z)): Returns minimum across all elements | |
| Examples: | |
| .. code-block:: python | |
| # Find minimum of two values | |
| result = min(x, y) | |
| # Find minimum of multiple values | |
| result = min(a, b, c, d) | |
| # Find minimum of values in a list | |
| values = [a, b, c, d] | |
| result = min(values) | |
| # Find minimum across mixed arguments | |
| result = min(x, [y, z]) | |
| Difference from Python's built-in min(): | |
| .. code-block:: python | |
| # In Python, the return type depends on the dynamic values: | |
| a = 5 | |
| b = 3.14 | |
| result = min(a, b) # Returns 3.14 (float) | |
| # In this DSL implementation, the return type is determined statically: | |
| a = Int32(5) | |
| b = Float32(3.14) | |
| result = min(a, b) # Return type is determined by the type of operands, not values | |
| """ | |
| return _minmax(min, *args, loc=loc, ip=ip) | |
| def max(*args, loc=None, ip=None): | |
| """Computes the maximum value from the provided arguments. | |
| This function differs from Python's built-in max() in that the return type | |
| is determined by the static types of the inputs, not their dynamic values. | |
| :param args: One or more values or iterables to find the maximum of | |
| :type args: tuple | |
| :param loc: Source location for MLIR operation tracking | |
| :type loc: object, optional | |
| :param ip: Insertion point for MLIR operation | |
| :type ip: object, optional | |
| :return: The maximum value among all inputs | |
| :rtype: Numeric | |
| :raises DSLNotImplemented: If the input type is not supported | |
| Supports multiple calling patterns: | |
| - max(a): Returns a | |
| - max([a, b, c, ...]): Returns maximum of all elements in the iterable | |
| - max(a, b, c, ...): Returns maximum of all arguments | |
| - max([x, y], [b]): Returns maximum across all elements in all iterables | |
| - max(a, (x, y, z)): Returns maximum across all elements | |
| Examples: | |
| .. code-block:: python | |
| # Find maximum of two values | |
| result = max(x, y) | |
| # Find maximum of multiple values | |
| result = max(a, b, c, d) | |
| # Find maximum of values in a list | |
| values = [a, b, c, d] | |
| result = max(values) | |
| # Find maximum across mixed arguments | |
| result = max(x, [y, z]) | |
| Difference from Python's built-in max(): | |
| .. code-block:: python | |
| # In Python, the return type depends on the dynamic values: | |
| a = 5 | |
| b = 3.14 | |
| result = max(a, b) # Returns 5 (int) | |
| # In this DSL implementation, the return type is determined statically: | |
| a = Int32(5) | |
| b = Float32(3.14) | |
| result = max(a, b) # Return type is determined by the type of operands, not values | |
| """ | |
| return _minmax(max, *args, loc=loc, ip=ip) | |
| def and_(*args, loc=None, ip=None): | |
| """AND operation for value in DSL numeric types. | |
| :param *args: One or more numeric values to AND together | |
| :type *args: Numeric | |
| :param loc: Source location for MLIR operation tracking | |
| :type loc: object, optional | |
| :param ip: Insertion point for MLIR operation | |
| :type ip: object, optional | |
| :return: The result of the logical AND operation | |
| :rtype: Numeric | |
| :raises ValueError: If no arguments are provided | |
| Supports multiple calling patterns: | |
| - and_(a): Returns a | |
| - and_(a, b, c, ...): if a is truthy, returns and_(b, c, ...), otherwise returns a | |
| All arguments must be of the same type. | |
| Examples: | |
| .. code-block:: python | |
| # In Python, 'and' returns the second operand if the first is truthy, | |
| # otherwise it returns the first operand | |
| a = 5 | |
| b = 3 | |
| result = a and b # Returns 3 | |
| # In this DSL implementation, the behavior is similar but works with DSL types | |
| a = Int32(5) | |
| b = Int32(3) | |
| result = and_(a, b) # Returns b | |
| """ | |
| if len(args) == 0: | |
| raise ValueError("and_() requires at least one argument") | |
| if len(args) == 1: | |
| return args[0] | |
| def and_op(lhs, rhs): | |
| if not isinstance(lhs, (Numeric, cutlass_arith.ArithValue, int, float, bool)): | |
| raise DSLNotImplemented(f"{type(lhs)} is not supported") | |
| elif isinstance(lhs, (int, float, bool)) and isinstance( | |
| rhs, (int, float, bool) | |
| ): | |
| return lhs and rhs | |
| else: | |
| return as_numeric(lhs).__dsl_and__(as_numeric(rhs)) | |
| return functools.reduce(and_op, args[1:], args[0]) | |
| def or_(*args, loc=None, ip=None): | |
| """Logical OR operation for DSL numeric types. | |
| :param *args: One or more numeric values to OR together | |
| :type *args: Numeric | |
| :param loc: Source location for MLIR operation tracking | |
| :type loc: object, optional | |
| :param ip: Insertion point for MLIR operation | |
| :type ip: object, optional | |
| :return: The result of the logical OR operation | |
| :rtype: Numeric | |
| :raises ValueError: If no arguments are provided | |
| Supports multiple calling patterns: | |
| - or_(a): Returns a | |
| - or_(a, b, c, ...): if a is truthy, returns a, otherwise returns or_(b, c, ...) | |
| Examples: | |
| .. code-block:: python | |
| # In Python, 'or' returns the first operand if it's truthy, | |
| # otherwise it returns the second operand | |
| a = 5 | |
| b = 3 | |
| result = a or b # Returns 5 | |
| # In this DSL implementation, the behavior is similar but works with DSL types | |
| a = Int32(5) | |
| b = Int32(3) | |
| result = or_(a, b) # Returns a | |
| """ | |
| if len(args) == 0: | |
| raise ValueError("or_() requires at least one argument") | |
| if len(args) == 1: | |
| return args[0] | |
| def or_op(lhs, rhs): | |
| if not isinstance(lhs, (Numeric, cutlass_arith.ArithValue, int, float, bool)): | |
| raise DSLNotImplemented(f"{type(lhs)} is not supported") | |
| elif isinstance(lhs, (int, float, bool)) and isinstance( | |
| rhs, (int, float, bool) | |
| ): | |
| return lhs or rhs | |
| else: | |
| return as_numeric(lhs).__dsl_or__(as_numeric(rhs)) | |
| return functools.reduce(or_op, args[1:], args[0]) | |
| def all_(iterable): | |
| """Logical AND operation for all elements in an iterable. | |
| Returns True if all elements in the iterable are truthy, otherwise False. | |
| This is the DSL equivalent of Python's built-in all() function. | |
| :param iterable: An iterable containing values to check | |
| :type iterable: Iterable | |
| :return: True if all elements are truthy, False otherwise | |
| :rtype: Boolean | |
| Examples: | |
| .. code-block:: python | |
| # Check if all values are non-zero | |
| values = [Int32(1), Int32(2), Int32(3)] | |
| result = all_(values) # Returns True | |
| # Check if all conditions are met | |
| conditions = [a > 0, b < 10, c != 0] | |
| result = all_(conditions) # Returns True if all conditions are met | |
| """ | |
| bool_iterable = [Boolean(i) for i in iterable] | |
| return functools.reduce( | |
| lambda lhs, rhs: lhs.__dsl_and__(rhs) if hasattr(lhs, "__dsl_and__") else lhs, | |
| bool_iterable, | |
| Boolean(True), | |
| ) | |
| def any_(iterable): | |
| """Logical OR operation for any element in an iterable. | |
| Returns True if any element in the iterable is truthy, otherwise False. | |
| This is the DSL equivalent of Python's built-in any() function. | |
| :param iterable: An iterable containing values to check | |
| :type iterable: Iterable | |
| :return: True if any element is truthy, False otherwise | |
| :rtype: Boolean | |
| Examples: | |
| .. code-block:: python | |
| # Check if any value is non-zero | |
| values = [Int32(0), Int32(0), Int32(3)] | |
| result = any_(values) # Returns True | |
| # Check if any condition is met | |
| conditions = [a > 10, b < 0, c != 0] | |
| result = any_(conditions) # Returns True if any condition is met | |
| """ | |
| bool_iterable = [Boolean(i) for i in iterable] | |
| return functools.reduce( | |
| lambda lhs, rhs: lhs.__dsl_or__(rhs) if hasattr(lhs, "__dsl_or__") else lhs, | |
| bool_iterable, | |
| Boolean(False), | |
| ) | |
| # ============================================================================= | |
| # Conditional Expression | |
| # ============================================================================= | |
| def select_(cond, if_value, else_value): | |
| def _as_scalar(value): | |
| if const_expr(isinstance(value, list)): | |
| if const_expr(len(value) == 1): | |
| return value[0] | |
| else: | |
| raise DSLRuntimeError( | |
| "Conditional expression must have exactly one value in all expressions" | |
| ) | |
| return value | |
| # Non-DSL dynamic cond should be handled before this. | |
| if const_expr(not is_dynamic_expression(cond)): | |
| raise DSLRuntimeError("Conditional expression must be dynamic") | |
| # Extract MLIR values | |
| cond = extract_mlir_values(cond) | |
| if const_expr(is_dynamic_expression(if_value)): | |
| if_value = extract_mlir_values(if_value) | |
| else: | |
| if_value = const(if_value) | |
| if const_expr(is_dynamic_expression(else_value)): | |
| else_value = extract_mlir_values(else_value) | |
| else: | |
| else_value = const(else_value) | |
| return arith.SelectOp( | |
| _as_scalar(cond), _as_scalar(if_value), _as_scalar(else_value) | |
| ).result | |
| # ============================================================================= | |
| # Terminator | |
| # ============================================================================= | |
| def yield_out(args=[], loc=None, ip=None): | |
| """ | |
| Generate a yield operation. It it used to return values from a loop, if-else, or while region. | |
| """ | |
| scf.yield_(extract_mlir_values(args), loc=loc, ip=ip) | |
| # ============================================================================= | |
| # For Loop | |
| # ============================================================================= | |
| class LoopUnroll(ir.Attribute): | |
| def __init__(self, **kwargs): | |
| valid_keys = set(["count", "full"]) | |
| def to_mlir_attr(val): | |
| if isinstance(val, bool): | |
| return "true" if val else "false" | |
| elif isinstance(val, int): | |
| return f"{val} : i32" | |
| else: | |
| raise DSLNotImplemented(f"{type(val)} is not supported") | |
| cfg = {key: to_mlir_attr(kwargs[key]) for key in valid_keys if key in kwargs} | |
| if kwargs.get("count", None) == 1: | |
| cfg["disable"] = "true" | |
| unroll = "<" + ", ".join(f"{key} = {value}" for key, value in cfg.items()) + ">" | |
| super().__init__( | |
| ir.Attribute.parse(f"#llvm.loop_annotation<unroll = {unroll}>") | |
| ) | |
| def for_generate( | |
| start, | |
| stop=None, | |
| step=None, | |
| iter_args: Optional[Sequence[ir.Value]] = None, | |
| *, | |
| unroll: LoopUnroll = None, | |
| loc=None, | |
| ip=None, | |
| ): | |
| """ | |
| scf.for with yield support | |
| """ | |
| if step is None: | |
| step = 1 | |
| if stop is None: | |
| stop = start | |
| start = 0 | |
| start = const(start) | |
| params = [start, stop, step] | |
| for i, p in enumerate(params): | |
| if isinstance(p, int): | |
| p = const(p) | |
| elif isinstance(p, float): | |
| raise DSLRuntimeError(f"{p=} must be int.") | |
| elif isinstance(p, Integer): | |
| p = p.ir_value() | |
| params[i] = p | |
| start, stop, step = params | |
| def _createI32Attr(value): | |
| if not isinstance(value, int): | |
| raise DSLRuntimeError(f"value must be int.") | |
| return ir.IntegerAttr.get(ir.IntegerType.get_signless(32), value) | |
| ir_iter_args = extract_mlir_values(iter_args) if iter_args is not None else None | |
| if not _validate_iter_args_structure(iter_args, ir_iter_args): | |
| raise DSLRuntimeError("iter_args: Elements should be extractable as ir.Value.") | |
| for_op = scf.ForOp(start, stop, step, ir_iter_args, loc=loc, ip=ip) | |
| if unroll is not None: | |
| for_op.attributes["loop_annotation"] = unroll | |
| iv = for_op.induction_variable | |
| new_results = new_from_mlir_values(iter_args, for_op.results) | |
| new_iter_args = new_from_mlir_values(iter_args, for_op.inner_iter_args) | |
| new_iter_args = () if new_iter_args is None else tuple(new_iter_args) | |
| with ir.InsertionPoint(for_op.body): | |
| if len(new_iter_args) > 1: | |
| yield iv, new_iter_args, new_results | |
| elif len(new_iter_args) == 1: | |
| yield iv, new_iter_args[0], new_results[0] | |
| else: | |
| yield iv | |
| # ============================================================================= | |
| # Logical Operators | |
| # ============================================================================= | |
| def not_(lhs: Union[ir.Value, bool], *, loc=None, ip=None): | |
| """ | |
| Logical Not | |
| """ | |
| res = None | |
| # Handle Python bool first to prevent infinite recursion | |
| if const_expr(type(lhs) == bool): | |
| res = lhs ^ True | |
| elif const_expr(hasattr(lhs, "__dsl_not__")): | |
| res = lhs.__dsl_not__(loc=loc, ip=ip) | |
| elif const_expr(is_dynamic_expression(lhs)): | |
| # If lhs is MLIR value, compute not using xor | |
| res = arith.XOrIOp(lhs, const(1, lhs.type)).result | |
| else: | |
| res = bool(lhs) ^ True | |
| return res | |
| # ============================================================================= | |
| # If/Else | |
| # ============================================================================= | |
| def if_generate( | |
| cond: Boolean, | |
| then_body: Callable, | |
| else_body: Optional[Callable] = None, | |
| input_args: List[DslType] = None, | |
| return_types: List[DslType] = None, | |
| *, | |
| loc=None, | |
| ip=None, | |
| ) -> List: | |
| """ | |
| Generate an IfOp with optional else branch and return values. | |
| Args: | |
| cond: The condition expression | |
| then_body: Function to execute in then branch | |
| else_body: Optional function to execute in else branch | |
| input_args: Arguments to pass to branch bodies | |
| return_types: Expected return types for the operation | |
| loc: Optional location information | |
| ip: Optional insertion point | |
| Returns: | |
| List of DSL typed results | |
| """ | |
| input_args = input_args or [] | |
| mlir_return_types = [] | |
| # Validate and collect MLIR return types (if provided). | |
| if return_types is not None: | |
| for t in return_types: | |
| if not isinstance(t, DslType): | |
| raise DSLRuntimeError(f"{t=} must be a DslType.") | |
| mlir_return_types.append(t.mlir_type) | |
| # Determine whether there's an else branch. | |
| has_else = else_body is not None | |
| # Create the IfOp. | |
| if_op = scf.IfOp( | |
| Boolean(cond).ir_value(), mlir_return_types, hasElse=has_else, loc=loc, ip=ip | |
| ) | |
| def _execute_and_yield_out(body, input_args): | |
| yield_vals = body(*input_args) | |
| if return_types is not None: | |
| if not isinstance(yield_vals, Iterable): | |
| # body only return single element | |
| yield_vals = [yield_vals] | |
| yield_vals = [t(r) for t, r in zip(return_types, yield_vals)] | |
| yield_out(yield_vals) | |
| # Generate the body for 'then'. | |
| with ir.InsertionPoint(if_op.then_block): | |
| _execute_and_yield_out(then_body, input_args) | |
| # Generate the body for 'else' if provided. | |
| if has_else: | |
| with ir.InsertionPoint(if_op.else_block): | |
| _execute_and_yield_out(else_body, input_args) | |
| # Collect MLIR results. | |
| mlir_results = _get_op_result_or_op_results(if_op) | |
| if not isinstance(mlir_results, list): | |
| mlir_results = [mlir_results] | |
| # Wrap the results with their DSL types. | |
| if return_types is None: | |
| return [] | |
| vals = [t(r) for t, r in zip(return_types, mlir_results)] | |
| if len(vals) == 1: | |
| return vals[0] | |
| return vals | |
| # ============================================================================= | |
| # While Loop | |
| # ============================================================================= | |
| class WhileLoopContext: | |
| """ | |
| Context manager for a dynamic while loop. | |
| """ | |
| def __init__( | |
| self, | |
| inputs: Sequence[Union[ir.Value, Numeric]], | |
| condition: Callable[[Sequence[ir.Value]], ir.Value], | |
| *, | |
| loc=None, | |
| ip=None, | |
| ): | |
| # Keep original inputs and allow recover original type information | |
| self.inputs = inputs | |
| self.input_ir_values = extract_mlir_values(inputs) | |
| if not _validate_iter_args_structure(inputs, self.input_ir_values): | |
| raise DSLRuntimeError("inputs: Elements should be extractable as ir.Value.") | |
| self.condition = condition | |
| self.input_ir_types = [i.type for i in self.input_ir_values] | |
| self.while_op = scf.WhileOp( | |
| self.input_ir_types, self.input_ir_values, loc=loc, ip=ip | |
| ) | |
| self.before_region = self.while_op.before | |
| self.after_region = self.while_op.after | |
| self.before_region.blocks.append(*self.input_ir_types) | |
| self.before_block = self.before_region.blocks[0] | |
| self.after_region.blocks.append(*self.input_ir_types) | |
| self.after_block = self.after_region.blocks[0] | |
| def __enter__(self): | |
| with ir.InsertionPoint(self.before_block): | |
| args = new_from_mlir_values(self.inputs, self.before_block.arguments) | |
| cond = self.condition(*args) | |
| cond_ir_val = extract_mlir_values(cond) | |
| scf.ConditionOp(cond_ir_val[0], [*self.before_block.arguments]) | |
| self.ipoint_op = ir.InsertionPoint(self.after_block) | |
| self.ipoint_op.__enter__() | |
| return new_from_mlir_values(self.inputs, self.after_block.arguments) | |
| def __exit__(self, exc_type, exc_value, traceback): | |
| self.ipoint_op.__exit__(exc_type, exc_value, traceback) | |
| def results(self): | |
| return new_from_mlir_values(self.inputs, self.while_op.results_) | |
| def while_generate( | |
| inputs: Sequence[Union[ir.Value, Numeric]], | |
| condition: Callable[[Sequence[Union[ir.Value, Numeric]]], Union[ir.Value, Numeric]], | |
| *, | |
| loc=None, | |
| ip=None, | |
| ) -> WhileLoopContext: | |
| """ | |
| Generate a WhileLoopContext for a dynamic loop. | |
| """ | |
| return WhileLoopContext(inputs, condition, loc=loc, ip=ip) | |