File size: 333 Bytes
62dca4c | 1 2 3 4 5 6 7 8 9 10 11 12 13 | # adapt from https://github.com/feifeibear/long-context-attention/tree/main/yunchang
from .ring_flash_attn import (
ring_flash_attn_func,
ring_flash_attn_kvpacked_func,
ring_flash_attn_qkvpacked_func,
)
__all__ = [
"ring_flash_attn_func",
"ring_flash_attn_kvpacked_func",
"ring_flash_attn_qkvpacked_func",
]
|