# SageAttention Wheels (CUDA 13.x) ๐Ÿš€ Prebuilt SageAttention 2.2.0 wheels compiled for Linux x86_64 with CUDA 13.x support. This repository provides ready-to-use binary wheels for different Python and PyTorch versions, optimized for modern NVIDIA GPUs (Ada / Hopper / Ampere). --- # ๐Ÿ“ฆ Available Wheels | Python Version | PyTorch Version | CUDA | File | |----------------|----------------|------|------| | 3.11 | 2.10 | cu13 | sageattention-2.2.0-python3.11-pytorch2.10-cu13-linux_x86_64.whl | | 3.12 | 2.10 | cu13 | sageattention-2.2.0-python3.12-pytorch2.10-cu13-linux_x86_64.whl | | 3.12 | 2.11 | cu13 | sageattention-2.2.0-python3.12-pytorch2.11-cu13-linux_x86_64.whl | | 3.13 | 2.11 | cu13 | sageattention-2.2.0-python3.13-pytorch2.11-cu13-linux_x86_64.whl | --- # โšก Requirements - Linux x86_64 - NVIDIA GPU (Ada / Ampere / Hopper tested) - CUDA 13.x runtime / toolkit - PyTorch matching wheel version - Python version matching wheel --- # ๐Ÿง  Installation ## 1. Create virtual environment python3.12 -m venv venv source venv/bin/activate --- ## 2. Install PyTorch (CUDA 13) pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu130 --- ## 3. Install SageAttention wheel pip install sageattention-2.2.0-python3.12-pytorch2.11-cu13-linux_x86_64.whl --- # ๐Ÿงช Quick Test python -c "import torch; print(torch.cuda.is_available())" python -c "import sageattention; print('SageAttention loaded successfully')" --- # ๐Ÿš€ Notes - Wheels are precompiled for performance - Must match Python + PyTorch versions exactly - CUDA 13.x required - Optimized for sm_80+ GPUs --- # โš ๏ธ Troubleshooting CUDA not found: export CUDA_HOME=/opt/cuda export PATH=$CUDA_HOME/bin:$PATH export LD_LIBRARY_PATH=$CUDA_HOME/lib64:$LD_LIBRARY_PATH --- # ๐Ÿ’ฌ Support Matrix Network @aimiko:mochiart.moe --- # ๐Ÿ“œ License Refer to upstream SageAttention repository. This repo contains only prebuilt binaries.