Custom Flash Attention 2.8.3 Wheel for Windows + Python 3.13 + Blackwell
This is a from-source build of flash-attn v2.8.3 targeting:
- Windows 11 (win_amd64)
- Python 3.13 (cp313)
- Blackwell architecture (sm_120 support via TORCH_CUDA_ARCH_LIST=12.0)
- PyTorch with CUDA 13.0 (cu130) โ built for RTX PRO 60000 Blackwell Workstation
Installation
pip install https://huggingface.co/IxaOne/flash-attn-blackwell-win-cp313/resolve/main/flash_attn-2.8.3-cp313-cp313-win_amd64.whl
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support