Matt300209's picture
Add files using upload-large-folder tool
a3f2e4d verified
# Copyright 2018 The apache/tvm Authors. All Rights Reserved.
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
#
# Modifications Copyright (c) Microsoft.
# The code below is mostly copied from apache/tvm reduction.py in dlight.
"""A rule for reduction. """
from typing import List, Optional, Tuple, Union
from tvm import arith, ir, tir
from tvm.target import Target
from ..base import (
BlockInfo,
normalize_prim_func,
try_inline_contiguous_spatial,
detect_dominant_read,
is_broadcast_epilogue,
)
from . import utils
from .base import GPUScheduleRule
def _get_reduction_expr(block: tir.Block) -> Optional[tir.PrimExpr]:
# Detect and return `Y` in `X[...] = X[...] + Y`
buffer_store = block.body
if not isinstance(buffer_store, tir.BufferStore):
return None
if not isinstance(buffer_store.value, tir.Add):
return None
if not ir.structural_equal(
buffer_store.value.a,
tir.BufferLoad(buffer_store.buffer, block.body.indices),
map_free_vars=True,
):
return None
return buffer_store.value.b
class Reduction(GPUScheduleRule):
"""A rule for Reduction."""
def apply( # pylint: disable=too-many-locals,too-many-branches,too-many-return-statements
self,
func: tir.PrimFunc,
target: Target,
_: bool,
) -> Union[None, tir.Schedule, List[tir.Schedule]]:
if not isinstance(func, tir.PrimFunc) or not self.is_target_available(target):
return None
sch = tir.Schedule(func)
block_infos = normalize_prim_func(sch)
if block_infos is None:
return None
block_infos = try_inline_contiguous_spatial(sch, block_infos)
if len(block_infos) == 1:
epilogue = None
elif len(block_infos) == 2:
epilogue = block_infos[1]
if not epilogue.is_injective():
return None
else:
return None
block_info = block_infos[0]
block = block_info.block_rv
block_stmt = sch.get(block)
# Step 1. Check reduction block
if (
(not block_info.is_reduction())
or len(block_stmt.writes) != 1
or _get_reduction_expr(block_stmt) is None
):
return None
# Step 2. Normalize the block, merge spatial and reduction iters
is_inner_reduction, c_factor, loop_order, s_split_index = self._normalize(
sch,
block_info,
arith.normalize_to_iter_sum(
detect_dominant_read(block_stmt),
input_iters={i.var: i.dom for i in block_stmt.iter_vars},
),
)
if is_inner_reduction is None and c_factor is None:
return None
# Step 3. Do the scheduling
if is_inner_reduction:
self._sch_inner_reduction(
sch, target, block, c_factor, epilogue, loop_order, s_split_index
)
else:
self._sch_inner_spatial(
sch, target, block, block_info, c_factor, epilogue, loop_order, s_split_index
)
return sch
def _normalize( # pylint: disable=too-many-branches
self,
sch: tir.Schedule,
block_info: BlockInfo,
access: arith.IterSumExpr,
) -> Tuple[Optional[bool], Optional[int]]:
if access.base != 0:
return None, None, None, None
iter_to_info = {i.var: i for i in block_info.iters}
s_loops, r_loops, c_loops, c_factor = [], [], [], None
s_split_loop, s_split_index = None, None
for split_expr in access.args:
var = split_expr.source.source
info = iter_to_info.pop(var)
loop = info.loop_rv
is_inner_reduction = info.kind == "R"
if split_expr.lower_factor > 1:
if c_loops:
return None, None, None, None
s_split_loop = loop
s_split_index = len(s_loops)
loop, c_loop = sch.split(loop, factors=[None, split_expr.lower_factor])
c_loops.append(c_loop)
if not is_inner_reduction:
c_factor = split_expr.lower_factor
if is_inner_reduction:
r_loops.append(loop)
else:
s_loops.append(loop)
if iter_to_info:
for var, info in iter_to_info.items():
if info.kind == "S" and info.dom.extent == 1:
s_loops.append(info.loop_rv)
else:
return None, None, None, None
loop_order = {}
s_block_var_loops = []
for i in block_info.iters:
if i.loop_rv in s_loops or i.loop_rv == s_split_loop:
s_block_var_loops.append(i.loop_rv)
for i in range(len(s_block_var_loops)):
for j in range(len(s_loops)):
if s_block_var_loops[i] == s_loops[j]:
loop_order[i] = j
break
if s_block_var_loops[i] == s_split_loop:
loop_order[i] = s_split_index
break
assert s_loops
assert r_loops
if len(s_loops) != len([i for i in block_info.iters if i.kind == "S"]):
return None, None
if not c_loops:
c_loops = [sch.add_unit_loop(block_info.block_rv)]
sch.reorder(*s_loops, *r_loops, *c_loops)
sch.fuse(*s_loops)
sch.fuse(*r_loops)
return is_inner_reduction, c_factor, loop_order, s_split_index
def _sch_inner_reduction( # pylint: disable=too-many-arguments
self,
sch: tir.Schedule,
target: Target,
block: tir.schedule.BlockRV,
unroll_spatial_factor: Optional[int],
epilogue_info: Optional[BlockInfo],
loop_order,
s_split_index,
):
# pylint: disable=invalid-name
_, r, _ = sch.get_loops(block)
(len_tx,) = utils.suggest_threads_per_block( # pylint: disable=unbalanced-tuple-unpacking
target, [sch.get(r)]
)
_, tx = sch.split(r, factors=[None, len_tx])
# Schedule the RF block
rf = sch.rfactor(tx, 0)
bx, r, tx, _ = sch.get_loops(rf)
sch.reorder(bx, tx, r)
sch.bind(bx, "blockIdx.x")
sch.bind(tx, "threadIdx.x")
sch.annotate(tx, ann_key="pragma_auto_unroll_max_step", ann_val=256)
sch.annotate(tx, ann_key="pragma_unroll_explicit", ann_val=1)
sch.set_scope(rf, 0, "local")
sch.decompose_reduction(rf, r)
# Schedule the write back block
sch.reverse_compute_at(block, bx, preserve_unit_loops=True)
_, tx, *s = sch.get_loops(block)
if unroll_spatial_factor:
assert len(s) == len(loop_order)
new_order_s = [s[loop_order[i]] for i in range(len(s))]
sch.reorder(*new_order_s)
new_order_s[s_split_index], c = sch.split(
new_order_s[s_split_index], factors=[None, unroll_spatial_factor]
)
sch.reorder(*new_order_s, c)
s = sch.fuse(*new_order_s)
sch.reorder(s, tx, c)
else:
s = sch.fuse(*s)
sch.reorder(s, tx)
sch.bind(tx, "threadIdx.x")
# Schedule epilogue
if epilogue_info is not None:
epilogue = epilogue_info.block_rv
sch.reverse_compute_at(epilogue, bx)
if is_broadcast_epilogue(sch, block, epilogue):
sch.set_scope(block, 0, "shared")
_, *s = sch.get_loops(epilogue) # pylint: disable=invalid-name
_, tx = sch.split(sch.fuse(*s), factors=[None, len_tx])
sch.bind(tx, "threadIdx.x")
else:
sch.set_scope(block, 0, "local")
# pylint: enable=invalid-name
def _sch_inner_spatial(
self,
sch: tir.Schedule,
_: Target,
block: tir.schedule.BlockRV,
block_info: BlockInfo,
unroll_spatial_factor: Optional[int],
epilogue_info: Optional[BlockInfo],
loop_order,
s_split_index,
):
# pylint: disable=invalid-name
s, r, _ = sch.get_loops(block)
len_tx, len_ty = 16, 16
s_factor = [i.dom.extent for i in block_info.iters if i.kind == "S"][-1]
# get perfect spatial factor, spatial factor should be divide the innermost spatial loop so
# that the block after r_factor and be reversed compute at the original scope
while len_tx > 1:
if s_factor % len_tx == 0:
break
len_tx -= 1
_, _ = sch.split(s, factors=[None, len_tx])
_, ty = sch.split(r, factors=[None, len_ty])
# Schedule the RF block
rf = sch.rfactor(ty, 0)
bx, tx, r, ty, _ = sch.get_loops(rf)
sch.reorder(bx, tx, ty, r)
sch.bind(tx, "threadIdx.x")
sch.bind(ty, "threadIdx.y")
sch.bind(bx, "blockIdx.x")
sch.set_scope(rf, 0, "local")
sch.decompose_reduction(rf, r)
# Schedule the write back block
sch.reverse_compute_at(block, bx, preserve_unit_loops=True)
_, r, *s = sch.get_loops(block)
if unroll_spatial_factor:
assert len(s) == len(loop_order)
new_order_s = [s[loop_order[i]] for i in range(len(s))]
sch.reorder(*new_order_s)
new_order_s[s_split_index], c = sch.split(
new_order_s[s_split_index], factors=[None, unroll_spatial_factor]
)
sch.reorder(*new_order_s, c)
s = sch.fuse(*new_order_s)
sch.reorder(s, c, r)
else:
s = sch.fuse(*s)
sch.reorder(s, r)
sch.bind(s, "threadIdx.x")
sch.bind(r, "threadIdx.y")
# Schedule epilogue
if epilogue_info is not None:
epilogue = epilogue_info.block_rv
sch.reverse_compute_at(epilogue, bx)
if is_broadcast_epilogue(sch, block, epilogue):
sch.set_scope(block, 0, "shared")
_, *s = sch.get_loops(epilogue) # pylint: disable=invalid-name
_, tx, ty = sch.split(sch.fuse(*s), factors=[None, len_tx, len_ty])
sch.bind(tx, "threadIdx.x")
sch.bind(ty, "threadIdx.y")
else:
# The epilogue is element-wise without broadcasting.
# Thus the remaining spatial part should be bind to tx.
sch.set_scope(block, 0, "local")
_, *s = sch.get_loops(epilogue) # pylint: disable=invalid-name
tx, _ = sch.split(sch.fuse(*s), factors=[len_tx, None])
sch.bind(tx, "threadIdx.x")
# pylint: enable=invalid-name