|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""A rule for reduction. """ |
|
|
from typing import List, Optional, Tuple, Union |
|
|
|
|
|
from tvm import arith, ir, tir |
|
|
from tvm.target import Target |
|
|
|
|
|
from ..base import ( |
|
|
BlockInfo, |
|
|
normalize_prim_func, |
|
|
try_inline_contiguous_spatial, |
|
|
detect_dominant_read, |
|
|
is_broadcast_epilogue, |
|
|
) |
|
|
from . import utils |
|
|
from .base import GPUScheduleRule |
|
|
|
|
|
|
|
|
def _get_reduction_expr(block: tir.Block) -> Optional[tir.PrimExpr]: |
|
|
|
|
|
buffer_store = block.body |
|
|
if not isinstance(buffer_store, tir.BufferStore): |
|
|
return None |
|
|
if not isinstance(buffer_store.value, tir.Add): |
|
|
return None |
|
|
if not ir.structural_equal( |
|
|
buffer_store.value.a, |
|
|
tir.BufferLoad(buffer_store.buffer, block.body.indices), |
|
|
map_free_vars=True, |
|
|
): |
|
|
return None |
|
|
return buffer_store.value.b |
|
|
|
|
|
|
|
|
class Reduction(GPUScheduleRule): |
|
|
"""A rule for Reduction.""" |
|
|
|
|
|
def apply( |
|
|
self, |
|
|
func: tir.PrimFunc, |
|
|
target: Target, |
|
|
_: bool, |
|
|
) -> Union[None, tir.Schedule, List[tir.Schedule]]: |
|
|
if not isinstance(func, tir.PrimFunc) or not self.is_target_available(target): |
|
|
return None |
|
|
sch = tir.Schedule(func) |
|
|
block_infos = normalize_prim_func(sch) |
|
|
if block_infos is None: |
|
|
return None |
|
|
block_infos = try_inline_contiguous_spatial(sch, block_infos) |
|
|
if len(block_infos) == 1: |
|
|
epilogue = None |
|
|
elif len(block_infos) == 2: |
|
|
epilogue = block_infos[1] |
|
|
if not epilogue.is_injective(): |
|
|
return None |
|
|
else: |
|
|
return None |
|
|
|
|
|
block_info = block_infos[0] |
|
|
block = block_info.block_rv |
|
|
block_stmt = sch.get(block) |
|
|
|
|
|
|
|
|
if ( |
|
|
(not block_info.is_reduction()) |
|
|
or len(block_stmt.writes) != 1 |
|
|
or _get_reduction_expr(block_stmt) is None |
|
|
): |
|
|
return None |
|
|
|
|
|
is_inner_reduction, c_factor, loop_order, s_split_index = self._normalize( |
|
|
sch, |
|
|
block_info, |
|
|
arith.normalize_to_iter_sum( |
|
|
detect_dominant_read(block_stmt), |
|
|
input_iters={i.var: i.dom for i in block_stmt.iter_vars}, |
|
|
), |
|
|
) |
|
|
if is_inner_reduction is None and c_factor is None: |
|
|
return None |
|
|
|
|
|
if is_inner_reduction: |
|
|
self._sch_inner_reduction( |
|
|
sch, target, block, c_factor, epilogue, loop_order, s_split_index |
|
|
) |
|
|
else: |
|
|
self._sch_inner_spatial( |
|
|
sch, target, block, block_info, c_factor, epilogue, loop_order, s_split_index |
|
|
) |
|
|
return sch |
|
|
|
|
|
def _normalize( |
|
|
self, |
|
|
sch: tir.Schedule, |
|
|
block_info: BlockInfo, |
|
|
access: arith.IterSumExpr, |
|
|
) -> Tuple[Optional[bool], Optional[int]]: |
|
|
if access.base != 0: |
|
|
return None, None, None, None |
|
|
iter_to_info = {i.var: i for i in block_info.iters} |
|
|
s_loops, r_loops, c_loops, c_factor = [], [], [], None |
|
|
s_split_loop, s_split_index = None, None |
|
|
for split_expr in access.args: |
|
|
var = split_expr.source.source |
|
|
info = iter_to_info.pop(var) |
|
|
loop = info.loop_rv |
|
|
is_inner_reduction = info.kind == "R" |
|
|
if split_expr.lower_factor > 1: |
|
|
if c_loops: |
|
|
return None, None, None, None |
|
|
s_split_loop = loop |
|
|
s_split_index = len(s_loops) |
|
|
loop, c_loop = sch.split(loop, factors=[None, split_expr.lower_factor]) |
|
|
c_loops.append(c_loop) |
|
|
if not is_inner_reduction: |
|
|
c_factor = split_expr.lower_factor |
|
|
if is_inner_reduction: |
|
|
r_loops.append(loop) |
|
|
else: |
|
|
s_loops.append(loop) |
|
|
|
|
|
if iter_to_info: |
|
|
for var, info in iter_to_info.items(): |
|
|
if info.kind == "S" and info.dom.extent == 1: |
|
|
s_loops.append(info.loop_rv) |
|
|
else: |
|
|
return None, None, None, None |
|
|
|
|
|
loop_order = {} |
|
|
s_block_var_loops = [] |
|
|
for i in block_info.iters: |
|
|
if i.loop_rv in s_loops or i.loop_rv == s_split_loop: |
|
|
s_block_var_loops.append(i.loop_rv) |
|
|
|
|
|
for i in range(len(s_block_var_loops)): |
|
|
for j in range(len(s_loops)): |
|
|
if s_block_var_loops[i] == s_loops[j]: |
|
|
loop_order[i] = j |
|
|
break |
|
|
if s_block_var_loops[i] == s_split_loop: |
|
|
loop_order[i] = s_split_index |
|
|
break |
|
|
|
|
|
assert s_loops |
|
|
assert r_loops |
|
|
if len(s_loops) != len([i for i in block_info.iters if i.kind == "S"]): |
|
|
return None, None |
|
|
if not c_loops: |
|
|
c_loops = [sch.add_unit_loop(block_info.block_rv)] |
|
|
sch.reorder(*s_loops, *r_loops, *c_loops) |
|
|
sch.fuse(*s_loops) |
|
|
sch.fuse(*r_loops) |
|
|
return is_inner_reduction, c_factor, loop_order, s_split_index |
|
|
|
|
|
def _sch_inner_reduction( |
|
|
self, |
|
|
sch: tir.Schedule, |
|
|
target: Target, |
|
|
block: tir.schedule.BlockRV, |
|
|
unroll_spatial_factor: Optional[int], |
|
|
epilogue_info: Optional[BlockInfo], |
|
|
loop_order, |
|
|
s_split_index, |
|
|
): |
|
|
|
|
|
_, r, _ = sch.get_loops(block) |
|
|
(len_tx,) = utils.suggest_threads_per_block( |
|
|
target, [sch.get(r)] |
|
|
) |
|
|
|
|
|
_, tx = sch.split(r, factors=[None, len_tx]) |
|
|
|
|
|
rf = sch.rfactor(tx, 0) |
|
|
bx, r, tx, _ = sch.get_loops(rf) |
|
|
sch.reorder(bx, tx, r) |
|
|
sch.bind(bx, "blockIdx.x") |
|
|
sch.bind(tx, "threadIdx.x") |
|
|
sch.annotate(tx, ann_key="pragma_auto_unroll_max_step", ann_val=256) |
|
|
sch.annotate(tx, ann_key="pragma_unroll_explicit", ann_val=1) |
|
|
sch.set_scope(rf, 0, "local") |
|
|
sch.decompose_reduction(rf, r) |
|
|
|
|
|
sch.reverse_compute_at(block, bx, preserve_unit_loops=True) |
|
|
_, tx, *s = sch.get_loops(block) |
|
|
|
|
|
if unroll_spatial_factor: |
|
|
assert len(s) == len(loop_order) |
|
|
new_order_s = [s[loop_order[i]] for i in range(len(s))] |
|
|
sch.reorder(*new_order_s) |
|
|
new_order_s[s_split_index], c = sch.split( |
|
|
new_order_s[s_split_index], factors=[None, unroll_spatial_factor] |
|
|
) |
|
|
sch.reorder(*new_order_s, c) |
|
|
s = sch.fuse(*new_order_s) |
|
|
sch.reorder(s, tx, c) |
|
|
else: |
|
|
s = sch.fuse(*s) |
|
|
sch.reorder(s, tx) |
|
|
sch.bind(tx, "threadIdx.x") |
|
|
|
|
|
if epilogue_info is not None: |
|
|
epilogue = epilogue_info.block_rv |
|
|
sch.reverse_compute_at(epilogue, bx) |
|
|
if is_broadcast_epilogue(sch, block, epilogue): |
|
|
sch.set_scope(block, 0, "shared") |
|
|
_, *s = sch.get_loops(epilogue) |
|
|
_, tx = sch.split(sch.fuse(*s), factors=[None, len_tx]) |
|
|
sch.bind(tx, "threadIdx.x") |
|
|
else: |
|
|
sch.set_scope(block, 0, "local") |
|
|
|
|
|
|
|
|
def _sch_inner_spatial( |
|
|
self, |
|
|
sch: tir.Schedule, |
|
|
_: Target, |
|
|
block: tir.schedule.BlockRV, |
|
|
block_info: BlockInfo, |
|
|
unroll_spatial_factor: Optional[int], |
|
|
epilogue_info: Optional[BlockInfo], |
|
|
loop_order, |
|
|
s_split_index, |
|
|
): |
|
|
|
|
|
s, r, _ = sch.get_loops(block) |
|
|
len_tx, len_ty = 16, 16 |
|
|
s_factor = [i.dom.extent for i in block_info.iters if i.kind == "S"][-1] |
|
|
|
|
|
|
|
|
while len_tx > 1: |
|
|
if s_factor % len_tx == 0: |
|
|
break |
|
|
len_tx -= 1 |
|
|
_, _ = sch.split(s, factors=[None, len_tx]) |
|
|
_, ty = sch.split(r, factors=[None, len_ty]) |
|
|
|
|
|
rf = sch.rfactor(ty, 0) |
|
|
bx, tx, r, ty, _ = sch.get_loops(rf) |
|
|
sch.reorder(bx, tx, ty, r) |
|
|
sch.bind(tx, "threadIdx.x") |
|
|
sch.bind(ty, "threadIdx.y") |
|
|
sch.bind(bx, "blockIdx.x") |
|
|
sch.set_scope(rf, 0, "local") |
|
|
sch.decompose_reduction(rf, r) |
|
|
|
|
|
sch.reverse_compute_at(block, bx, preserve_unit_loops=True) |
|
|
_, r, *s = sch.get_loops(block) |
|
|
if unroll_spatial_factor: |
|
|
assert len(s) == len(loop_order) |
|
|
new_order_s = [s[loop_order[i]] for i in range(len(s))] |
|
|
sch.reorder(*new_order_s) |
|
|
new_order_s[s_split_index], c = sch.split( |
|
|
new_order_s[s_split_index], factors=[None, unroll_spatial_factor] |
|
|
) |
|
|
sch.reorder(*new_order_s, c) |
|
|
s = sch.fuse(*new_order_s) |
|
|
sch.reorder(s, c, r) |
|
|
else: |
|
|
s = sch.fuse(*s) |
|
|
sch.reorder(s, r) |
|
|
sch.bind(s, "threadIdx.x") |
|
|
sch.bind(r, "threadIdx.y") |
|
|
|
|
|
|
|
|
if epilogue_info is not None: |
|
|
epilogue = epilogue_info.block_rv |
|
|
sch.reverse_compute_at(epilogue, bx) |
|
|
if is_broadcast_epilogue(sch, block, epilogue): |
|
|
sch.set_scope(block, 0, "shared") |
|
|
_, *s = sch.get_loops(epilogue) |
|
|
_, tx, ty = sch.split(sch.fuse(*s), factors=[None, len_tx, len_ty]) |
|
|
sch.bind(tx, "threadIdx.x") |
|
|
sch.bind(ty, "threadIdx.y") |
|
|
else: |
|
|
|
|
|
|
|
|
sch.set_scope(block, 0, "local") |
|
|
_, *s = sch.get_loops(epilogue) |
|
|
tx, _ = sch.split(sch.fuse(*s), factors=[len_tx, None]) |
|
|
sch.bind(tx, "threadIdx.x") |
|
|
|
|
|
|