// SPDX-License-Identifier: MIT // Copyright (c) 2023-2024, Advanced Micro Devices, Inc. All rights reserved. #pragma once #include "ck/wrapper/utils/layout_utils.hpp" // Disable from doxygen docs generation /// @cond INTERNAL namespace ck { namespace wrapper { /// @endcond /** * \brief Layout wrapper that performs the tensor descriptor logic. * * \tparam Shape Tuple of Number<> (for compile-time layout) or index_t * (dynamic layout). It is possible to pass nested shapes * (e.g. ((4, 2), 2)), nested dimensions are merged. * \tparam UnrolledDescriptorType Tensor descriptor for unnested shape dims. */ template struct Layout { // Disable from doxygen docs generation /// @cond INTERNAL private: static constexpr auto I0 = Number<0>{}; static constexpr auto I1 = Number<1>{}; /** * \brief Generate default indices tuple (idx with all merged nested shapes) * * \param shape Shape to align. * \return Multi idx tuple with zeros. */ template __host__ __device__ constexpr static auto GenerateDefaultIdxsTuple([[maybe_unused]] const Tuple& shape) { return generate_tuple( [&](auto) { if constexpr(!remove_cvref_t::IsKnownAtCompileTime()) { // runtime layout return index_t(0); } else { // compiletime layout return I0; } }, Number::Size()>{}); } /** * \brief Generate lower dims in compile-time for the Merge transform using * provided type. If element of nested Tuple is also a tuple, then * merge (generate sequence for merge). If tuple is element, then pass * through (sequence with one element). * * \param shape Shape to align. * \return LowerDims for MergeTrasform. */ template __host__ __device__ constexpr static auto GenerateLowerDim([[maybe_unused]] const Tuple& shape) { if constexpr(Idx::value == 0) { if constexpr(is_detected>>::value) { // Return Sequence for the first tuple constexpr index_t merge_nelems = decltype(UnrollNestedTuple( tuple_element_t>{}))::Size(); using LowerDimsSequence = typename arithmetic_sequence_gen<0, merge_nelems, 1>::type; return LowerDimsSequence::Reverse(); } else { // Return first element return Sequence<0>{}; } } else { // Get previous element using recurence (in compile-time) using PreviousSeqT = decltype(GenerateLowerDim>(Tuple{})); const auto next_seq_val = PreviousSeqT::At(I0) + 1; if constexpr(is_detected>>::value) { constexpr index_t merge_nelems = decltype(UnrollNestedTuple( tuple_element_t>{}))::Size(); using LowerDimsSequence = typename arithmetic_sequence_gen:: type; return LowerDimsSequence::Reverse(); } else { return Sequence{}; } } } /** * \brief Iterate over the nested tuples in the shape. * Unroll nested tuples to align Tuple to Tuple * Example idx: (1, 1), 1, 1 * Example shape: (2, (2, 2)), 2, (2, 2) * Unrolled shape: 2, (2, 2), 2, (2, 2) * * \param shape Layout shape. * \param idx Idx to align. * \return Algined shape. */ template __host__ __device__ constexpr static auto AlignShapeToIdx(const Tuple& shape, const Tuple& idx) { if constexpr(!IsNestedTuple(Tuple{})) { // Index unrolled to flatten, return shape return shape; } else { // Iterate over shape tuple elements: // 1. If corresponding idx element is tuple then return (will be unrolled) // 2. If no, pack in tuple. It will be restored during unroll. auto aligned_shape = generate_tuple( [&](auto i) { if constexpr(is_detected>>::value) { return shape.At(i); } else { return make_tuple(shape.At(i)); } }, Number::Size()>{}); // Unroll and process next step return AlignShapeToIdx(UnrollNestedTuple<0, 1>(aligned_shape), UnrollNestedTuple<0, 1>(idx)); } } /** * \brief Merge descriptor to 1D. * * \param shape Layout shape. * \param desc Descriptor to merge. * \return 1D descriptor. */ template __host__ __device__ constexpr static auto MakeMerge1d(const Tuple& shape, const DescriptorToMerge& desc) { // Reverse each element in tuple const auto merge_elems = TupleReverse(UnrollNestedTuple(shape)); // Generate reverted indexes (column major traverse) using MergeElemsSequence = typename arithmetic_sequence_gen<0, merge_elems.Size(), 1>::type; const auto lower_dims = make_tuple(MergeElemsSequence::Reverse()); const auto upper_dims = make_tuple(Sequence<0>{}); // Merge to 1d if constexpr(!remove_cvref_t::IsKnownAtCompileTime()) { return transform_tensor_descriptor( desc, make_tuple(make_merge_transform(merge_elems)), lower_dims, upper_dims); } else { // If the descriptor is known at the compilation time, // use `make_merge_transform_v1_carry_check` because it doesn't use // memcpy. return transform_tensor_descriptor( desc, make_tuple(make_merge_transform_v1_carry_check(merge_elems)), lower_dims, upper_dims); } } /** * \brief Merge nested shape dims when corresponding index is also merged. * Input desc shape: 2, 2, 2, 2, 2, 2 * Example idx: 1, 1, 1, (1, 1) * Example shape: 2, (2, 2), 2, (2, 2) * Merged shape: 2, 4, 2, 2, 2 * * \param shape Layout shape. * \param idxs Indexes to align descriptor. * \param desc Descriptor to merge. * \return Aligned descriptor to idx. */ template __host__ __device__ constexpr static auto CreateMergedDescriptor(const Tuple& shape, [[maybe_unused]] const Tuple& idxs, DescriptorToMerge& desc) { const auto transforms = generate_tuple( [&](auto i) { // Compare Idx with shape if constexpr(is_detected>>::value && !is_detected>>::value) { // If shape element is tuple and idx element is Number, then merge // Unroll and reverse tuple to traverse column-major const auto merge_elems = TupleReverse(UnrollNestedTuple(shape.At(i))); if constexpr(!remove_cvref_t::IsKnownAtCompileTime()) { return make_merge_transform(merge_elems); } else { // If the descriptor is known at the compilation time, // use `make_merge_transform_v1_carry_check` because // it doesn't use memcpy. return make_merge_transform_v1_carry_check(merge_elems); } } else { // If shape element is integer and idx element is tuple, passed idx is wrong static_assert( !(!is_detected>>::value && is_detected>>::value), "Wrong Idx for layout()"); // If shape element has the same type as idx element, then pass through return make_pass_through_transform(shape.At(i)); } }, Number::Size()>{}); const auto lower_dims = generate_tuple([&](auto i) { return GenerateLowerDim>(shape); }, Number::Size()>{}); const auto upper_dims = generate_tuple([&](auto i) { return Sequence{}; }, Number::Size()>{}); return transform_tensor_descriptor(desc, transforms, lower_dims, upper_dims); } using Descriptor1dType = remove_cvref_t; using DefaultIdxsTupleType = remove_cvref_t; /// @endcond public: using LayoutShape = Shape; using LayoutUnrolledDescriptorType = UnrolledDescriptorType; /** * \brief Transform descriptor to align to passed indexes. * * \param shape Layout shape. * \param idxs Indexes to align descriptor. * \param naive_descriptor Descriptor to merge. * \return Aligned descriptor to idx. */ template __host__ __device__ constexpr static auto TransformDesc(const Tuple& shape, const Tuple& idxs, const UnrolledDescriptorType& naive_descriptor) { if constexpr(Tuple::Size() == I1) { // 1d idx path return MakeMerge1d(shape, naive_descriptor); } else { // Merge nested shape dims // Example idx: (1, 1), 1, 1 // Example shape: (2, (2, 2)), 2, (2, 2) // Merged shape: (2, 4), 2, 4 static_assert(Tuple::Size() == Tuple::Size(), "Idx rank and Shape rank must be the same (except 1d)."); // Unroll while IdxDims is nested const auto aligned_shape = AlignShapeToIdx(shape, idxs); // Transform correct form of shape return CreateMergedDescriptor(aligned_shape, UnrollNestedTuple(idxs), naive_descriptor); } } using MergedNestsDescriptorType = remove_cvref_t; __host__ __device__ constexpr auto GetElementSpaceSize() const { return unrolled_descriptor_.GetElementSpaceSize(); } __host__ __device__ Layout() = delete; /** * \brief Layout constructor. * * \param shape Shape for layout. * \param unnested_descriptor Descriptor */ __host__ __device__ constexpr Layout(const Shape& shape, const UnrolledDescriptorType& unnested_descriptor) : unrolled_descriptor_(unnested_descriptor), shape_(shape) { // Construct if runtime mode if constexpr(!remove_cvref_t::IsKnownAtCompileTime()) { descriptor_1d_ = MakeMerge1d(shape_, unrolled_descriptor_); merged_nests_descriptor_ = TransformDesc(shape_, DefaultIdxsTupleType{}, unrolled_descriptor_); } } /** * \brief Returns real offset to element in runtime. * * \tparam Idxs Tuple of indexes. * \return Calculated offset. */ template __host__ __device__ constexpr index_t operator()() const { static_assert(remove_cvref_t::IsKnownAtCompileTime(), "Compiletime operator used on runtime layout."); using TransformedDesc = decltype(TransformDesc(Shape{}, Idxs{}, UnrolledDescriptorType{})); using UnrolledIdx = decltype(UnrollNestedTuple(Idxs{})); return TransformedDesc{}.CalculateOffset(UnrolledIdx{}); } /** * \brief Returns real offset to element in compile time. * * \param Idx Tuple of indexes. * \return Calculated offset. */ template __host__ __device__ index_t operator()(const Tuple& Idx) const { if constexpr(!IsNestedTuple(Tuple{}) && Tuple::Size() == 1) { // if 1d access return descriptor_1d_.CalculateOffset(Idx); } else if constexpr(!IsNestedTuple(Tuple{}) && Tuple::Size() == Shape::Size()) { // if Shape::Size() access (merged nested shapes) return merged_nests_descriptor_.CalculateOffset(UnrollNestedTuple(Idx)); } else { // Custom index, need to transform descriptor const auto transformed_desc = TransformDesc(shape_, Idx, unrolled_descriptor_); return transformed_desc.CalculateOffset(UnrollNestedTuple(Idx)); } } /** * \brief Length getter (product if tuple). * * \tparam IDim Tuple of indexes or index. * \return Calculated size. */ template __host__ __device__ constexpr auto GetLength() const { const auto elem = shape_.At(Number{}); if constexpr(is_detected>::value) { const auto unrolled_element = UnrollNestedTuple(elem); return TupleReduce( [](auto x, auto y) { return x * y; }, unrolled_element); } else { return elem; } } /** * \brief Layout size getter (product of shape). * * \return Calculated size. */ __host__ __device__ constexpr auto GetLengths() const { const auto unrolled_shape = UnrollNestedTuple(shape_); return TupleReduce([](auto x, auto y) { return x * y; }, unrolled_shape); } /** * \brief Shape getter. * * \return Shape. */ __host__ __device__ constexpr const Shape& GetShape() const { return shape_; } /** * \brief Get default lengths (tuple filled with Shape length elements). * * \return Default lengths. */ __host__ __device__ constexpr auto GetDefaultLengthsTuple() const { return generate_tuple([&](auto i) { return GetLength(); }, Number{}); } /** * \brief Get default start idx (tuple filled with 0s of the same size as Shape). * * \return Default start idx. */ __host__ __device__ constexpr auto GetDefaultStartIdxs() const { return GenerateDefaultIdxsTuple(shape_); } /** * \brief Get descriptor with all nested dimensions merged. * Example, shape: ((2, 2), 2) * Descriptor lengths: (4, 2) * * \note The size of merged descriptor is the same as Layout's shape. * * \return Merged nests descriptor. */ __host__ __device__ constexpr const MergedNestsDescriptorType& GetMergedNestingDescriptor() const { return merged_nests_descriptor_; } /** * \brief Get descriptor with all dimensions are merged (1D). * Example, shape: ((2, 2), 2) * Descriptor lengths: (8) * * \return 1D descriptor. */ __host__ __device__ constexpr const Descriptor1dType& Get1DDescriptor() const { return descriptor_1d_; } /** * \brief Get unnested descriptor (with unrolled dims) * Example, shape: ((2, 2), 2) * Descriptor lengths: (2, 2, 2) * * \return Flattened descriptor. */ __host__ __device__ constexpr const UnrolledDescriptorType& GetUnrolledDescriptor() const { return unrolled_descriptor_; } // Disable from doxygen docs generation /// @cond INTERNAL private: // All dimensions are unrolled UnrolledDescriptorType unrolled_descriptor_; // 1D descriptor Descriptor1dType descriptor_1d_; // All nesting are merged MergedNestsDescriptorType merged_nests_descriptor_; // Example, shape: ((2, 2), 2) // UnrolledDescriptorType lengths: (2, 2, 2) // Descriptor1dType lengths: (8) // MergedNestsDescriptorType lengths: (4, 2) const Shape shape_; /// @endcond }; } // namespace wrapper } // namespace ck