--- library_name: kernels license: apache-2.0 --- This is the repository card of {repo_id} that has been pushed on the Hub. It was built to be used with the [`kernels` library](https://github.com/huggingface/kernels). This card was automatically generated. ## How to use ```python # make sure `kernels` is installed: `pip install -U kernels` from kernels import get_kernel kernel_module = get_kernel("REPO_ID") # <- change the ID if needed flash_attn_combine = kernel_module.flash_attn_combine flash_attn_combine(...) ``` ## Available functions - `flash_attn_combine` - `flash_attn_func` - `flash_attn_qkvpacked_func` - `flash_attn_varlen_func` - `flash_attn_with_kvcache` - `get_scheduler_metadata` ## Supported backends - cuda ## CUDA Capabilities - 9.0a - 8.0 ## Benchmarks [TODO: provide benchmarks if available] ## Code source [TODO: provide original code source and other relevant citations if available] ## Notes [TODO: provide additional notes about this kernel if needed]