|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import math |
|
|
from typing import Optional, Tuple |
|
|
|
|
|
import torch |
|
|
|
|
|
import cuda.bindings.driver as cuda |
|
|
|
|
|
import cutlass |
|
|
import cutlass.cute as cute |
|
|
from cutlass.cute.runtime import from_dlpack |
|
|
|
|
|
from flash_attn.cute import utils |
|
|
from flash_attn.cute.flash_fwd import FlashAttentionForwardSm80, FlashAttentionForwardSm90 |
|
|
from flash_attn.cute.flash_fwd_sm100 import FlashAttentionForwardSm100 |
|
|
from flash_attn.cute.flash_bwd_preprocess import FlashAttentionBackwardPreprocess |
|
|
from flash_attn.cute.flash_bwd import FlashAttentionBackwardSm80 |
|
|
from flash_attn.cute.flash_bwd_postprocess import FlashAttentionBackwardPostprocess |
|
|
from flash_attn.cute.flash_fwd_combine import FlashAttentionForwardCombine |
|
|
|
|
|
|
|
|
def maybe_contiguous(x): |
|
|
return x.contiguous() if x is not None and x.stride(-1) != 1 else x |
|
|
|
|
|
|
|
|
torch2cute_dtype_map = { |
|
|
torch.float16: cutlass.Float16, |
|
|
torch.bfloat16: cutlass.BFloat16, |
|
|
torch.float32: cutlass.Float32, |
|
|
} |
|
|
|
|
|
|
|
|
def _flash_attn_fwd( |
|
|
q: torch.Tensor, |
|
|
k: torch.Tensor, |
|
|
v: torch.Tensor, |
|
|
cu_seqlens_q: Optional[torch.Tensor] = None, |
|
|
cu_seqlens_k: Optional[torch.Tensor] = None, |
|
|
seqused_q: Optional[torch.Tensor] = None, |
|
|
seqused_k: Optional[torch.Tensor] = None, |
|
|
page_table: Optional[torch.Tensor] = None, |
|
|
softmax_scale: Optional[float] = None, |
|
|
causal: bool = False, |
|
|
softcap: Optional[float] = None, |
|
|
window_size_left: Optional[int] = None, |
|
|
window_size_right: Optional[int] = None, |
|
|
learnable_sink: Optional[torch.Tensor] = None, |
|
|
|
|
|
|
|
|
|
|
|
m_block_size: int = 128, |
|
|
n_block_size: int = 128, |
|
|
num_threads: int = 384, |
|
|
pack_gqa: Optional[bool] = None, |
|
|
_compute_capability: Optional[int] = None, |
|
|
) -> Tuple[torch.Tensor, torch.Tensor]: |
|
|
q, k, v = [maybe_contiguous(t) for t in (q, k, v)] |
|
|
num_head, head_dim = q.shape[-2:] |
|
|
if cu_seqlens_q is None: |
|
|
batch_size, seqlen_q = q.shape[:2] |
|
|
total_q = batch_size * seqlen_q |
|
|
else: |
|
|
batch_size = cu_seqlens_q.shape[0] - 1 |
|
|
seqlen_q = None |
|
|
total_q = q.shape[0] |
|
|
if page_table is not None: |
|
|
assert cu_seqlens_k is None, "page_table is not supported with cu_seqlens_k" |
|
|
assert page_table.dtype == torch.int32, "page_table must be int32" |
|
|
assert page_table.stride(-1) == 1, "page_table must be contiguous in the last dimension" |
|
|
max_num_pages_per_seq = page_table.shape[1] |
|
|
assert page_table.shape == (batch_size, max_num_pages_per_seq) |
|
|
num_pages, page_size = k.shape[:2] |
|
|
seqlen_k = num_pages * page_size |
|
|
else: |
|
|
num_pages, page_size = None, None |
|
|
seqlen_k = k.shape[-3] |
|
|
num_head_kv = k.shape[-2] |
|
|
head_dim_v = v.shape[-1] |
|
|
if cu_seqlens_k is None: |
|
|
if page_table is None: |
|
|
assert k.shape == (batch_size, seqlen_k, num_head_kv, head_dim) |
|
|
assert v.shape == (batch_size, seqlen_k, num_head_kv, head_dim_v) |
|
|
else: |
|
|
assert k.shape == (num_pages, page_size, num_head_kv, head_dim) |
|
|
assert v.shape == (num_pages, page_size, num_head_kv, head_dim_v) |
|
|
else: |
|
|
assert k.shape == (seqlen_k, num_head_kv, head_dim) |
|
|
assert v.shape == (seqlen_k, num_head_kv, head_dim_v) |
|
|
assert cu_seqlens_k.shape == (batch_size + 1,), "cu_seqlens_k must have shape (batch_size + 1,)" |
|
|
if cu_seqlens_q is not None: |
|
|
assert cu_seqlens_q.shape == (batch_size + 1,), "cu_seqlens_q must have shape (batch_size + 1,)" |
|
|
assert seqused_q is None or seqused_q.shape == (batch_size,), "seqused_q must have shape (batch_size,)" |
|
|
assert seqused_k is None or seqused_k.shape == (batch_size,), "seqused_k must have shape (batch_size,)" |
|
|
assert q.dtype in [torch.float16, torch.bfloat16], "inputs must be float16 or bfloat16" |
|
|
assert q.dtype == k.dtype == v.dtype, "inputs must have the same dtype" |
|
|
for t in [cu_seqlens_q, cu_seqlens_k, seqused_q, seqused_k]: |
|
|
if t is not None: |
|
|
assert t.dtype == torch.int32, "cu_seqlens_q, cu_seqlens_k, seqused_q, seqused_k must be int32" |
|
|
assert t.stride(0) == 1, "cu_seqlens_q, cu_seqlens_k, seqused_q, seqused_k must be contiguous" |
|
|
if learnable_sink is not None: |
|
|
assert learnable_sink.shape == (num_head,) |
|
|
assert learnable_sink.dtype == torch.bfloat16, "learnable_sink must be bfloat16" |
|
|
assert all(t is None or t.is_cuda for t in (q, k, v, cu_seqlens_q, cu_seqlens_k, seqused_q, seqused_k, page_table, learnable_sink)), "inputs must be on CUDA device" |
|
|
assert num_head % num_head_kv == 0, "num_head must be divisible by num_head_kv" |
|
|
assert head_dim <= 256, "head_dim must be less than or equal to 256" |
|
|
alignment = 16 // q.element_size() |
|
|
assert head_dim % alignment == 0, f"head_dim must be divisible by {alignment}" |
|
|
assert head_dim_v % alignment == 0, f"head_dim_v must be divisible by {alignment}" |
|
|
if softmax_scale is None: |
|
|
softmax_scale = 1.0 / math.sqrt(head_dim) |
|
|
if softcap == 0.0: |
|
|
softcap = None |
|
|
qhead_per_kvhead = num_head // num_head_kv |
|
|
if pack_gqa is None: |
|
|
pack_gqa = qhead_per_kvhead > 1 |
|
|
|
|
|
out_torch_dtype = q.dtype |
|
|
device = q.device |
|
|
q_batch_seqlen_shape = (batch_size, seqlen_q) if cu_seqlens_q is None else (total_q,) |
|
|
out = torch.empty(*q_batch_seqlen_shape, num_head, head_dim_v, dtype=out_torch_dtype, device=device) |
|
|
lse_shape = (batch_size, num_head, seqlen_q) if cu_seqlens_q is None else (num_head, total_q) |
|
|
requires_grad = q.requires_grad or k.requires_grad or v.requires_grad |
|
|
lse = torch.empty(lse_shape, dtype=torch.float32, device=device) if requires_grad else None |
|
|
|
|
|
dtype = torch2cute_dtype_map[q.dtype] |
|
|
q_tensor, k_tensor, v_tensor, o_tensor = [ |
|
|
from_dlpack(t.detach(), assumed_align=16).mark_layout_dynamic(leading_dim=t.ndim - 1) |
|
|
for t in (q, k, v, out) |
|
|
] |
|
|
lse_tensor = from_dlpack(lse.detach(), assumed_align=4).mark_layout_dynamic(leading_dim=lse.ndim - 1) if lse is not None else None |
|
|
cu_seqlens_q_tensor, cu_seqlens_k_tensor, seqused_q_tensor, seqused_k_tensor, learnable_sink_tensor = [ |
|
|
from_dlpack(t.detach(), assumed_align=4).mark_layout_dynamic(leading_dim=0) if t is not None else None |
|
|
for t in (cu_seqlens_q, cu_seqlens_k, seqused_q, seqused_k, learnable_sink) |
|
|
] |
|
|
page_table_tensor = from_dlpack(page_table.detach(), assumed_align=4).mark_layout_dynamic(leading_dim=1) if page_table is not None else None |
|
|
if causal: |
|
|
window_size_right = 0 |
|
|
local = window_size_left is not None or window_size_right is not None |
|
|
if window_size_left is not None or window_size_right is not None: |
|
|
if window_size_left is None and window_size_right == 0: |
|
|
causal, local = True, False |
|
|
else: |
|
|
causal, local = False, True |
|
|
compute_capability = torch.cuda.get_device_capability()[0] if _compute_capability is None else _compute_capability |
|
|
assert compute_capability in [9, 10], "Unsupported compute capability. Supported: 9.x, 10.x" |
|
|
current_stream = cuda.CUstream(torch.cuda.current_stream().cuda_stream) |
|
|
|
|
|
if compute_capability == 9: |
|
|
if head_dim == head_dim_v == 128 and not causal and not local: |
|
|
n_block_size = 192 |
|
|
if compute_capability == 10: |
|
|
|
|
|
if pack_gqa and (128 % qhead_per_kvhead != 0) or (cu_seqlens_q is not None or seqused_q is not None): |
|
|
pack_gqa = False |
|
|
|
|
|
compile_key = ( |
|
|
dtype, head_dim, head_dim_v, qhead_per_kvhead, causal, softcap is not None, |
|
|
lse is None, cu_seqlens_q is None, cu_seqlens_k is None, seqused_q is None, seqused_k is None, |
|
|
page_table is not None, |
|
|
window_size_left is not None, window_size_right is not None, |
|
|
learnable_sink is not None, |
|
|
m_block_size, n_block_size, num_threads, pack_gqa, |
|
|
compute_capability, |
|
|
) |
|
|
if compile_key not in _flash_attn_fwd.compile_cache: |
|
|
if compute_capability == 9: |
|
|
assert page_table is None, "paged KV not supported on SM 9.0" |
|
|
|
|
|
fa_fwd = FlashAttentionForwardSm90( |
|
|
dtype, |
|
|
head_dim, |
|
|
head_dim_v, |
|
|
qhead_per_kvhead, |
|
|
is_causal=causal, |
|
|
is_local=local, |
|
|
pack_gqa=pack_gqa, |
|
|
m_block_size=m_block_size, |
|
|
n_block_size=n_block_size, |
|
|
|
|
|
num_stages=2, |
|
|
num_threads=num_threads, |
|
|
Q_in_regs=False, |
|
|
) |
|
|
elif compute_capability == 10: |
|
|
assert page_size in [None, 128], "Only page_size=128 is supported for paged KV on SM 10.0" |
|
|
fa_fwd = FlashAttentionForwardSm100( |
|
|
head_dim, |
|
|
head_dim_v, |
|
|
qhead_per_kvhead=qhead_per_kvhead, |
|
|
is_causal=causal, |
|
|
is_local=local, |
|
|
pack_gqa=pack_gqa, |
|
|
is_persistent=not causal and not local and cu_seqlens_q is None and seqused_q is None, |
|
|
) |
|
|
else: |
|
|
raise ValueError(f"Unsupported compute capability: {compute_capability}. Supported: 9.x, 10.x") |
|
|
|
|
|
_flash_attn_fwd.compile_cache[compile_key] = cute.compile( |
|
|
fa_fwd, q_tensor, k_tensor, v_tensor, o_tensor, lse_tensor, softmax_scale, current_stream, |
|
|
cu_seqlens_q_tensor, cu_seqlens_k_tensor, seqused_q_tensor, seqused_k_tensor, |
|
|
page_table_tensor, |
|
|
softcap, window_size_left, window_size_right, learnable_sink_tensor, |
|
|
) |
|
|
_flash_attn_fwd.compile_cache[compile_key]( |
|
|
q_tensor, k_tensor, v_tensor, o_tensor, lse_tensor, softmax_scale, current_stream, |
|
|
cu_seqlens_q_tensor, cu_seqlens_k_tensor, seqused_q_tensor, seqused_k_tensor, |
|
|
page_table_tensor, |
|
|
softcap, window_size_left, window_size_right, learnable_sink_tensor, |
|
|
) |
|
|
return out, lse |
|
|
|
|
|
|
|
|
_flash_attn_fwd.compile_cache = {} |
|
|
|
|
|
|
|
|
def _flash_attn_bwd( |
|
|
q: torch.Tensor, |
|
|
k: torch.Tensor, |
|
|
v: torch.Tensor, |
|
|
out: torch.Tensor, |
|
|
dout: torch.Tensor, |
|
|
lse: torch.Tensor, |
|
|
softmax_scale: Optional[float] = None, |
|
|
causal: bool = False, |
|
|
softcap: float = 0.0, |
|
|
m_block_size: int = 64, |
|
|
n_block_size: int = 128, |
|
|
num_threads: int = 256, |
|
|
num_stages_Q: int = 2, |
|
|
num_stages_dO: int = 2, |
|
|
SdP_swapAB: bool = False, |
|
|
dKV_swapAB: bool = False, |
|
|
dQ_swapAB: bool = False, |
|
|
AtomLayoutMSdP: int = 2, |
|
|
AtomLayoutNdKV: int = 2, |
|
|
AtomLayoutMdQ: int = 2, |
|
|
V_in_regs: bool = False, |
|
|
) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: |
|
|
q, k, v, out, dout, lse = [maybe_contiguous(t) for t in (q, k, v, out, dout, lse)] |
|
|
batch_size, seqlen_q, num_head, head_dim = q.shape |
|
|
_, seqlen_k, num_head_kv, _ = k.shape |
|
|
_, _, _, head_dim_v = v.shape |
|
|
assert k.shape == (batch_size, seqlen_k, num_head_kv, head_dim) |
|
|
assert v.shape == (batch_size, seqlen_k, num_head_kv, head_dim_v) |
|
|
assert out.shape == (batch_size, seqlen_q, num_head, head_dim_v) |
|
|
assert dout.shape == (batch_size, seqlen_q, num_head, head_dim_v) |
|
|
assert lse.shape == (batch_size, num_head, seqlen_q), "lse must have shape (batch_size, num_head, seqlen_q)" |
|
|
assert q.dtype in [torch.float16, torch.bfloat16], "inputs must be float16 or bfloat16" |
|
|
assert q.dtype == k.dtype == v.dtype == out.dtype == dout.dtype, "inputs must have the same dtype" |
|
|
assert lse.dtype == torch.float32, "lse must be float32" |
|
|
assert all(t.is_cuda for t in (q, k, v, out, dout, lse)), "inputs must be on CUDA device" |
|
|
assert num_head % num_head_kv == 0, "num_head must be divisible by num_head_kv" |
|
|
assert head_dim <= 256, "head_dim must be less than or equal to 256" |
|
|
alignment = 16 // q.element_size() |
|
|
assert head_dim % alignment == 0, f"head_dim must be divisible by {alignment}" |
|
|
assert head_dim_v % alignment == 0, f"head_dim_v must be divisible by {alignment}" |
|
|
if softmax_scale is None: |
|
|
softmax_scale = 1.0 / math.sqrt(head_dim) |
|
|
qhead_per_kvhead = num_head // num_head_kv |
|
|
|
|
|
device = q.device |
|
|
|
|
|
seqlen_q_rounded = (seqlen_q + m_block_size - 1) // m_block_size * m_block_size |
|
|
head_dim_rounded = (head_dim + 32 - 1) // 32 * 32 |
|
|
dq = torch.empty_like(q) |
|
|
dk = torch.empty_like(k) |
|
|
dv = torch.empty_like(v) |
|
|
dq_accum = torch.empty(batch_size, num_head, seqlen_q_rounded * head_dim_rounded, dtype=torch.float32, device=device) |
|
|
dpsum = torch.empty(batch_size, num_head, seqlen_q_rounded, dtype=torch.float32, device=device) |
|
|
lse_log2 = torch.empty(batch_size, num_head, seqlen_q_rounded, dtype=torch.float32, device=device) |
|
|
if qhead_per_kvhead > 1: |
|
|
seqlen_k_rounded = (seqlen_k + n_block_size - 1) // n_block_size * n_block_size |
|
|
head_dim_v_rounded = (head_dim_v + 32 - 1) // 32 * 32 |
|
|
dk_accum = torch.zeros(batch_size, num_head_kv, seqlen_k_rounded * head_dim_rounded, dtype=torch.float32, device=device) |
|
|
dv_accum = torch.zeros(batch_size, num_head_kv, seqlen_k_rounded * head_dim_v_rounded, dtype=torch.float32, device=device) |
|
|
|
|
|
dtype = torch2cute_dtype_map[q.dtype] |
|
|
q_tensor, k_tensor, v_tensor, o_tensor, do_tensor, dq_tensor, dk_tensor, dv_tensor = [ |
|
|
from_dlpack(t.detach(), assumed_align=16).mark_layout_dynamic(leading_dim=t.ndim - 1) |
|
|
for t in (q, k, v, out, dout, dq, dk, dv) |
|
|
] |
|
|
lse_tensor = from_dlpack(lse.detach(), assumed_align=4).mark_layout_dynamic(leading_dim=2) |
|
|
dq_accum_tensor, dpsum_tensor, lse_log2_tensor = [ |
|
|
from_dlpack(t.detach(), assumed_align=16).mark_layout_dynamic(leading_dim=2) |
|
|
for t in (dq_accum, dpsum, lse_log2) |
|
|
] |
|
|
if qhead_per_kvhead > 1: |
|
|
dk_accum_tensor, dv_accum_tensor = [ |
|
|
from_dlpack(t.detach(), assumed_align=16).mark_layout_dynamic(leading_dim=2) |
|
|
for t in (dk_accum, dv_accum) |
|
|
] |
|
|
current_stream = cuda.CUstream(torch.cuda.current_stream().cuda_stream) |
|
|
|
|
|
|
|
|
compile_key_pre = (dtype, head_dim_v, m_block_size, num_threads) |
|
|
if compile_key_pre not in _flash_attn_bwd.compile_cache_pre: |
|
|
fa_bwd_pre = FlashAttentionBackwardPreprocess( |
|
|
dtype, head_dim_v, m_block_size, num_threads=num_threads, |
|
|
) |
|
|
|
|
|
_flash_attn_bwd.compile_cache_pre[compile_key_pre] = cute.compile( |
|
|
fa_bwd_pre, o_tensor, do_tensor, dpsum_tensor, lse_tensor, lse_log2_tensor, |
|
|
dq_accum_tensor, current_stream |
|
|
) |
|
|
_flash_attn_bwd.compile_cache_pre[compile_key_pre]( |
|
|
o_tensor, do_tensor, dpsum_tensor, lse_tensor, lse_log2_tensor, dq_accum_tensor, current_stream |
|
|
) |
|
|
|
|
|
|
|
|
compile_key = ( |
|
|
dtype, head_dim, head_dim_v, qhead_per_kvhead, causal, softcap != 0.0, m_block_size, |
|
|
n_block_size, num_threads, num_stages_Q, num_stages_dO, SdP_swapAB, dKV_swapAB, dQ_swapAB, |
|
|
AtomLayoutMSdP, AtomLayoutNdKV, AtomLayoutMdQ, V_in_regs |
|
|
) |
|
|
if compile_key not in _flash_attn_bwd.compile_cache: |
|
|
fa_bwd_sm80 = FlashAttentionBackwardSm80( |
|
|
dtype, |
|
|
head_dim, |
|
|
head_dim_v, |
|
|
qhead_per_kvhead, |
|
|
m_block_size, |
|
|
n_block_size, |
|
|
num_stages_Q, |
|
|
num_stages_dO, |
|
|
num_threads, |
|
|
causal, |
|
|
SdP_swapAB, |
|
|
dKV_swapAB, |
|
|
dQ_swapAB, |
|
|
AtomLayoutMSdP, |
|
|
AtomLayoutNdKV, |
|
|
AtomLayoutMdQ, |
|
|
V_in_regs=V_in_regs, |
|
|
) |
|
|
|
|
|
_flash_attn_bwd.compile_cache[compile_key] = cute.compile( |
|
|
fa_bwd_sm80, q_tensor, k_tensor, v_tensor, do_tensor, lse_log2_tensor, dpsum_tensor, |
|
|
dq_accum_tensor, |
|
|
dk_tensor if qhead_per_kvhead == 1 else dk_accum_tensor, |
|
|
dv_tensor if qhead_per_kvhead == 1 else dv_accum_tensor, |
|
|
softmax_scale, current_stream |
|
|
) |
|
|
_flash_attn_bwd.compile_cache[compile_key]( |
|
|
q_tensor, k_tensor, v_tensor, do_tensor, lse_log2_tensor, dpsum_tensor, |
|
|
dq_accum_tensor, |
|
|
dk_tensor if qhead_per_kvhead == 1 else dk_accum_tensor, |
|
|
dv_tensor if qhead_per_kvhead == 1 else dv_accum_tensor, |
|
|
softmax_scale, current_stream |
|
|
) |
|
|
|
|
|
|
|
|
compile_key_post = (dtype, head_dim, m_block_size, num_threads, AtomLayoutMdQ, dQ_swapAB) |
|
|
if compile_key_post not in _flash_attn_bwd.compile_cache_post: |
|
|
fa_bwd_post = FlashAttentionBackwardPostprocess( |
|
|
dtype, head_dim, m_block_size, num_threads, AtomLayoutMdQ, dQ_swapAB |
|
|
) |
|
|
|
|
|
_flash_attn_bwd.compile_cache_post[compile_key_post] = cute.compile( |
|
|
fa_bwd_post, dq_accum_tensor, dq_tensor, softmax_scale, current_stream |
|
|
) |
|
|
_flash_attn_bwd.compile_cache_post[compile_key_post]( |
|
|
dq_accum_tensor, dq_tensor, softmax_scale, current_stream |
|
|
) |
|
|
|
|
|
if qhead_per_kvhead > 1: |
|
|
|
|
|
compile_key_post = (dtype, head_dim, n_block_size, num_threads, AtomLayoutNdKV, dKV_swapAB) |
|
|
if compile_key_post not in _flash_attn_bwd.compile_cache_post: |
|
|
fa_bwd_post = FlashAttentionBackwardPostprocess( |
|
|
dtype, head_dim, n_block_size, num_threads, AtomLayoutNdKV, dKV_swapAB |
|
|
) |
|
|
|
|
|
_flash_attn_bwd.compile_cache_post[compile_key_post] = cute.compile( |
|
|
fa_bwd_post, dk_accum_tensor, dk_tensor, softmax_scale, current_stream |
|
|
) |
|
|
_flash_attn_bwd.compile_cache_post[compile_key_post]( |
|
|
dk_accum_tensor, dk_tensor, softmax_scale, current_stream |
|
|
) |
|
|
compile_key_post = (dtype, head_dim_v, n_block_size, num_threads, AtomLayoutNdKV, dKV_swapAB) |
|
|
if compile_key_post not in _flash_attn_bwd.compile_cache_post: |
|
|
fa_bwd_post = FlashAttentionBackwardPostprocess( |
|
|
dtype, head_dim_v, n_block_size, num_threads, AtomLayoutNdKV, dKV_swapAB |
|
|
) |
|
|
|
|
|
_flash_attn_bwd.compile_cache_post[compile_key_post] = cute.compile( |
|
|
fa_bwd_post, dv_accum_tensor, dv_tensor, cutlass.Float32(1.0), current_stream |
|
|
) |
|
|
_flash_attn_bwd.compile_cache_post[compile_key_post]( |
|
|
dv_accum_tensor, dv_tensor, cutlass.Float32(1.0), current_stream |
|
|
) |
|
|
|
|
|
return dq, dk, dv |
|
|
|
|
|
|
|
|
_flash_attn_bwd.compile_cache_pre = {} |
|
|
_flash_attn_bwd.compile_cache = {} |
|
|
_flash_attn_bwd.compile_cache_post = {} |
|
|
|
|
|
|
|
|
class FlashAttnFunc(torch.autograd.Function): |
|
|
|
|
|
@staticmethod |
|
|
def forward( |
|
|
ctx, |
|
|
q: torch.Tensor, |
|
|
k: torch.Tensor, |
|
|
v: torch.Tensor, |
|
|
softmax_scale: Optional[float] = None, |
|
|
causal: bool = False, |
|
|
window_size: Tuple[Optional[int], Optional[int]] = (None, None), |
|
|
learnable_sink: Optional[torch.Tensor] = None, |
|
|
softcap: float = 0.0, |
|
|
pack_gqa: Optional[bool] = None, |
|
|
): |
|
|
out, lse = _flash_attn_fwd( |
|
|
q, |
|
|
k, |
|
|
v, |
|
|
softmax_scale=softmax_scale, |
|
|
causal=causal, |
|
|
window_size_left=window_size[0], |
|
|
window_size_right=window_size[1], |
|
|
learnable_sink=learnable_sink, |
|
|
softcap=softcap, |
|
|
pack_gqa=pack_gqa, |
|
|
) |
|
|
ctx.save_for_backward(q, k, v, out, lse) |
|
|
ctx.softmax_scale = softmax_scale |
|
|
ctx.causal = causal |
|
|
ctx.window_size = window_size |
|
|
ctx.softcap = softcap |
|
|
return out, lse |
|
|
|
|
|
@staticmethod |
|
|
def backward(ctx, dout, *args): |
|
|
q, k, v, out, lse = ctx.saved_tensors |
|
|
dq, dk, dv = _flash_attn_bwd( |
|
|
q, |
|
|
k, |
|
|
v, |
|
|
out, |
|
|
dout, |
|
|
lse, |
|
|
ctx.softmax_scale, |
|
|
ctx.causal, |
|
|
ctx.softcap, |
|
|
) |
|
|
return dq, dk, dv, *((None,) * 5) |
|
|
|
|
|
|
|
|
class FlashAttnVarlenFunc(torch.autograd.Function): |
|
|
|
|
|
@staticmethod |
|
|
def forward( |
|
|
ctx, |
|
|
q: torch.Tensor, |
|
|
k: torch.Tensor, |
|
|
v: torch.Tensor, |
|
|
cu_seqlens_q: Optional[torch.Tensor], |
|
|
cu_seqlens_k: Optional[torch.Tensor], |
|
|
seqused_q: Optional[torch.Tensor] = None, |
|
|
seqused_k: Optional[torch.Tensor] = None, |
|
|
page_table: Optional[torch.Tensor] = None, |
|
|
softmax_scale: Optional[float] = None, |
|
|
causal: bool = False, |
|
|
window_size: Tuple[Optional[int], Optional[int]] = (None, None), |
|
|
learnable_sink: Optional[torch.Tensor] = None, |
|
|
softcap: float = 0.0, |
|
|
pack_gqa: Optional[bool] = None, |
|
|
): |
|
|
out, lse = _flash_attn_fwd( |
|
|
q, |
|
|
k, |
|
|
v, |
|
|
cu_seqlens_q, |
|
|
cu_seqlens_k, |
|
|
seqused_q, |
|
|
seqused_k, |
|
|
page_table=page_table, |
|
|
softmax_scale=softmax_scale, |
|
|
causal=causal, |
|
|
window_size_left=window_size[0], |
|
|
window_size_right=window_size[1], |
|
|
learnable_sink=learnable_sink, |
|
|
softcap=softcap, |
|
|
pack_gqa=pack_gqa, |
|
|
) |
|
|
ctx.save_for_backward(q, k, v, out, lse, cu_seqlens_q, cu_seqlens_k, seqused_q, seqused_k) |
|
|
ctx.softmax_scale = softmax_scale |
|
|
ctx.causal = causal |
|
|
ctx.window_size = window_size |
|
|
ctx.softcap = softcap |
|
|
return out, lse |
|
|
|
|
|
@staticmethod |
|
|
def backward(ctx, dout, *args): |
|
|
q, k, v, out, lse, cu_seqlens_q, cu_seqlens_k, seqused_q, seqused_k = ctx.saved_tensors |
|
|
raise NotImplementedError( |
|
|
"Backward pass for FlashAttention with variable length sequences is not implemented yet." |
|
|
) |
|
|
|
|
|
|
|
|
def flash_attn_func( |
|
|
q: torch.Tensor, |
|
|
k: torch.Tensor, |
|
|
v: torch.Tensor, |
|
|
softmax_scale: Optional[float] = None, |
|
|
causal: bool = False, |
|
|
window_size: Tuple[Optional[int], Optional[int]] = (None, None), |
|
|
learnable_sink: Optional[torch.Tensor] = None, |
|
|
softcap: float = 0.0, |
|
|
pack_gqa: Optional[bool] = None, |
|
|
): |
|
|
return FlashAttnFunc.apply( |
|
|
q, |
|
|
k, |
|
|
v, |
|
|
softmax_scale, |
|
|
causal, |
|
|
window_size, |
|
|
learnable_sink, |
|
|
softcap, |
|
|
pack_gqa, |
|
|
) |
|
|
|
|
|
|
|
|
def flash_attn_varlen_func( |
|
|
q: torch.Tensor, |
|
|
k: torch.Tensor, |
|
|
v: torch.Tensor, |
|
|
cu_seqlens_q: Optional[torch.Tensor] = None, |
|
|
cu_seqlens_k: Optional[torch.Tensor] = None, |
|
|
seqused_q: Optional[torch.Tensor] = None, |
|
|
seqused_k: Optional[torch.Tensor] = None, |
|
|
page_table: Optional[torch.Tensor] = None, |
|
|
softmax_scale: Optional[float] = None, |
|
|
causal: bool = False, |
|
|
window_size: Tuple[Optional[int], Optional[int]] = (None, None), |
|
|
learnable_sink: Optional[torch.Tensor] = None, |
|
|
softcap: float = 0.0, |
|
|
pack_gqa: Optional[bool] = None, |
|
|
): |
|
|
return FlashAttnVarlenFunc.apply( |
|
|
q, |
|
|
k, |
|
|
v, |
|
|
cu_seqlens_q, |
|
|
cu_seqlens_k, |
|
|
seqused_q, |
|
|
seqused_k, |
|
|
page_table, |
|
|
softmax_scale, |
|
|
causal, |
|
|
window_size, |
|
|
learnable_sink, |
|
|
softcap, |
|
|
pack_gqa, |
|
|
) |
|
|
|
|
|
|
|
|
def _flash_attn_fwd_combine( |
|
|
out_partial: torch.Tensor, |
|
|
lse_partial: torch.Tensor, |
|
|
out: torch.Tensor, |
|
|
lse: Optional[torch.Tensor] = None, |
|
|
cu_seqlens: Optional[torch.Tensor] = None, |
|
|
seqused: Optional[torch.Tensor] = None, |
|
|
num_splits_dynamic_ptr: Optional[torch.Tensor] = None, |
|
|
semaphore_to_reset: Optional[torch.Tensor] = None, |
|
|
) -> None: |
|
|
"""Forward combine kernel for split attention computation. |
|
|
|
|
|
Combines partial outputs and log-sum-exp values from multiple splits |
|
|
of attention computation into final outputs. |
|
|
|
|
|
Args: |
|
|
out_partial: Partial outputs tensor (num_splits, batch, seqlen, nheads, headdim) or |
|
|
(num_splits, total_q, nheads, headdim) if there's cu_seqlens |
|
|
lse_partial: Partial LSE tensor (num_splits, batch, seqlen, nheads) or |
|
|
(num_splits, total_q, nheads) if there's cu_seqlens |
|
|
out: Output tensor (batch, seqlen, nheads, headdim) or (total_q, nheads, headdim) if there's cu_seqlens |
|
|
lse: Output LSE tensor (batch, seqlen, nheads) or (total_q, nheads) if there's cu_seqlens. |
|
|
cu_seqlens: Cumulative sequence lengths for variable length sequences |
|
|
seqused: Used sequence lengths for each batch |
|
|
num_splits_dynamic_ptr: Dynamic number of splits per batch |
|
|
semaphore_to_reset: Semaphore for synchronization |
|
|
k_block_size: Block size for head dimension |
|
|
|
|
|
Returns: |
|
|
None |
|
|
""" |
|
|
|
|
|
assert out_partial.dim() in [4, 5], "out_partial must have 4 or 5 dimensions" |
|
|
assert lse_partial.dim() in [3, 4], "lse_partial must have 3 or 4 dimensions" |
|
|
assert out_partial.dtype in [torch.float16, torch.bfloat16, torch.float32], "out_partial must be fp16, bf16, or fp32" |
|
|
assert lse_partial.dtype == torch.float32, "lse_partial must be fp32" |
|
|
assert out_partial.is_cuda and lse_partial.is_cuda, "tensors must be on CUDA device" |
|
|
assert out_partial.stride(-1) == 1, "out_partial must be contiguous in the last dimension" |
|
|
assert lse_partial.stride(-2) == 1, "lse_partial must be contiguous in the seqlen dimension" |
|
|
assert lse_partial.shape == out_partial.shape[:-1] |
|
|
|
|
|
|
|
|
is_varlen = out_partial.dim() == 4 |
|
|
|
|
|
|
|
|
assert out.shape == out_partial.shape[1:], "out shape mismatch" |
|
|
if lse is not None: |
|
|
assert lse.shape == lse_partial.shape[1:], "lse shape mismatch" |
|
|
assert lse.dtype == torch.float32, "lse must be fp32" |
|
|
|
|
|
|
|
|
for t, name in [(cu_seqlens, "cu_seqlens"), (seqused, "seqused"), (num_splits_dynamic_ptr, "num_splits_dynamic_ptr")]: |
|
|
if t is not None: |
|
|
assert t.dtype == torch.int32, f"{name} must be int32" |
|
|
assert t.is_cuda, f"{name} must be on CUDA device" |
|
|
assert t.is_contiguous(), f"{name} must be contiguous" |
|
|
|
|
|
head_dim = out_partial.shape[-1] |
|
|
num_splits = out_partial.shape[0] |
|
|
assert num_splits <= 256 |
|
|
|
|
|
|
|
|
k_block_size = 64 if head_dim <= 64 else 128 |
|
|
|
|
|
|
|
|
m_block_size = 8 if k_block_size % 128 == 0 else (16 if k_block_size % 64 == 0 else 32) |
|
|
log_max_splits = max(math.ceil(math.log2(num_splits)), 4) |
|
|
if m_block_size == 8: |
|
|
|
|
|
|
|
|
log_max_splits = max(log_max_splits, 5) |
|
|
|
|
|
|
|
|
out_partial_tensor = from_dlpack(out_partial.detach(), assumed_align=16).mark_layout_dynamic(leading_dim=4) |
|
|
lse_partial_tensor = from_dlpack(lse_partial.detach(), assumed_align=4).mark_layout_dynamic(leading_dim=lse_partial.ndim - 2) |
|
|
out_tensor = from_dlpack(out.detach(), assumed_align=16).mark_layout_dynamic(leading_dim=3) |
|
|
lse_tensor = from_dlpack(lse.detach(), assumed_align=4).mark_layout_dynamic(leading_dim=lse.ndim - 2) if lse is not None else None |
|
|
|
|
|
optional_tensors = [ |
|
|
from_dlpack(t.detach(), assumed_align=4).mark_layout_dynamic(leading_dim=0) if t is not None else None |
|
|
for t in (cu_seqlens, seqused, num_splits_dynamic_ptr, semaphore_to_reset) |
|
|
] |
|
|
cu_seqlens_tensor, seqused_tensor, num_splits_dynamic_tensor, semaphore_tensor = optional_tensors |
|
|
|
|
|
current_stream = cuda.CUstream(torch.cuda.current_stream().cuda_stream) |
|
|
|
|
|
|
|
|
dtype = torch2cute_dtype_map[out.dtype] |
|
|
dtype_partial = torch2cute_dtype_map[out_partial.dtype] |
|
|
|
|
|
compile_key = ( |
|
|
dtype, dtype_partial, head_dim, m_block_size, k_block_size, |
|
|
log_max_splits, |
|
|
cu_seqlens is not None, seqused is not None, lse is not None, |
|
|
) |
|
|
|
|
|
if compile_key not in _flash_attn_fwd_combine.compile_cache: |
|
|
fa_combine = FlashAttentionForwardCombine( |
|
|
dtype=dtype, |
|
|
dtype_partial=dtype_partial, |
|
|
head_dim=head_dim, |
|
|
m_block_size=m_block_size, |
|
|
k_block_size=k_block_size, |
|
|
log_max_splits=log_max_splits, |
|
|
) |
|
|
|
|
|
|
|
|
if not fa_combine.can_implement( |
|
|
dtype, dtype_partial, head_dim, m_block_size, k_block_size, log_max_splits, num_threads=256 |
|
|
): |
|
|
raise RuntimeError(f"FlashAttention combine kernel cannot be implemented with given parameters") |
|
|
|
|
|
_flash_attn_fwd_combine.compile_cache[compile_key] = cute.compile( |
|
|
fa_combine, |
|
|
out_partial_tensor, |
|
|
lse_partial_tensor, |
|
|
out_tensor, |
|
|
lse_tensor, |
|
|
cu_seqlens_tensor, |
|
|
seqused_tensor, |
|
|
num_splits_dynamic_tensor, |
|
|
semaphore_tensor, |
|
|
current_stream |
|
|
) |
|
|
|
|
|
_flash_attn_fwd_combine.compile_cache[compile_key]( |
|
|
out_partial_tensor, |
|
|
lse_partial_tensor, |
|
|
out_tensor, |
|
|
lse_tensor, |
|
|
cu_seqlens_tensor, |
|
|
seqused_tensor, |
|
|
num_splits_dynamic_tensor, |
|
|
semaphore_tensor, |
|
|
current_stream |
|
|
) |
|
|
|
|
|
|
|
|
_flash_attn_fwd_combine.compile_cache = {} |
|
|
|
|
|
|
|
|
def flash_attn_combine( |
|
|
out_partial: torch.Tensor, |
|
|
lse_partial: torch.Tensor, |
|
|
out: Optional[torch.Tensor] = None, |
|
|
out_dtype: Optional[torch.dtype] = None, |
|
|
return_lse: bool = True, |
|
|
) -> Tuple[torch.Tensor, Optional[torch.Tensor]]: |
|
|
"""Flash Attention combine function for split attention computation. |
|
|
|
|
|
Combines partial outputs and log-sum-exp values from multiple splits |
|
|
of attention computation into final outputs. This is the main user-facing |
|
|
interface for the combine kernel. |
|
|
|
|
|
Args: |
|
|
out_partial: Partial outputs tensor with shape: |
|
|
- (num_splits, batch_size, seqlen, num_heads, head_size) for regular batched input |
|
|
- (num_splits, total_q, num_heads, head_size) for variable length input |
|
|
lse_partial: Partial LSE tensor with shape: |
|
|
- (num_splits, batch_size, seqlen, num_heads) for regular batched input |
|
|
- (num_splits, total_q, num_heads) for variable length input |
|
|
out: Optional output tensor. If None, will be created automatically. |
|
|
out_dtype: Optional output dtype. If None, will use fp16/bf16 based on input. |
|
|
return_lse: Whether to return the combined LSE tensor. Default is True. |
|
|
|
|
|
Returns: |
|
|
Tuple of (out, lse) where: |
|
|
- out: Combined output tensor with shape (batch_size, seqlen, num_heads, head_size) |
|
|
or (total_q, num_heads, head_size) for varlen |
|
|
- lse: Combined log-sum-exp tensor with shape (batch_size, seqlen, num_heads) |
|
|
or (total_q, num_heads) for varlen. None if return_lse=False |
|
|
|
|
|
Note: |
|
|
This function expects the input tensors to be in the format produced by |
|
|
split attention computation, where the first dimension is num_splits. |
|
|
The permuting from user format to kernel format is now done inside the kernel. |
|
|
""" |
|
|
|
|
|
assert out_partial.dim() in [4, 5], "out_partial must have 4 or 5 dimensions" |
|
|
assert lse_partial.dim() in [3, 4], "lse_partial must have 3 or 4 dimensions" |
|
|
assert out_partial.dtype == torch.float32, "out_partial must be fp32 (from accumulation)" |
|
|
assert lse_partial.dtype == torch.float32, "lse_partial must be fp32" |
|
|
|
|
|
|
|
|
is_varlen = out_partial.dim() == 4 |
|
|
|
|
|
if is_varlen: |
|
|
|
|
|
num_splits, total_q, num_heads, head_size = out_partial.shape |
|
|
assert lse_partial.shape == (num_splits, total_q, num_heads), "lse_partial shape mismatch for varlen" |
|
|
batch_size = 1 |
|
|
seqlen = total_q |
|
|
else: |
|
|
|
|
|
num_splits, batch_size, seqlen, num_heads, head_size = out_partial.shape |
|
|
assert lse_partial.shape == (num_splits, batch_size, seqlen, num_heads), "lse_partial shape mismatch" |
|
|
|
|
|
|
|
|
if out_dtype is None: |
|
|
out_dtype = out_partial.dtype |
|
|
|
|
|
|
|
|
device = out_partial.device |
|
|
if out is None: |
|
|
if is_varlen: |
|
|
out = torch.empty(total_q, num_heads, head_size, dtype=out_dtype, device=device) |
|
|
else: |
|
|
out = torch.empty(batch_size, seqlen, num_heads, head_size, dtype=out_dtype, device=device) |
|
|
|
|
|
|
|
|
if return_lse: |
|
|
if is_varlen: |
|
|
lse = torch.empty(num_heads, total_q, dtype=torch.float32, device=device).transpose(0, 1) |
|
|
else: |
|
|
lse = torch.empty(batch_size, num_heads, seqlen, dtype=torch.float32, device=device).transpose(1, 2) |
|
|
else: |
|
|
lse = None |
|
|
|
|
|
_flash_attn_fwd_combine(out_partial, lse_partial, out, lse) |
|
|
return out, lse |
|
|
|