ltx2 / Wan2GP /docs /INSTALLATION.md
vidfom's picture
Upload folder using huggingface_hub
31112ad verified

Manual Installation Guide For Windows & Linux

This guide covers installation for different GPU generations and operating systems.

Requirements

- Compatible GPU (GTX 10XX - RTX 50XX)

Installation for Nvidia GTX 10XX - RTX QUADRO - 50XX (Stable)

This installation uses PyTorch 2.6.0, Cuda 12.6 for GTX 10XX - RTX 30XX & PyTorch 2.7.1, Cuda 12.8 for RTX 40XX - 50XX which are well-tested and stable.

Unless you need absolutely to use Pytorch compilation (with RTX 50xx), it is not recommeneded to use PytTorch 2.8.0 as some System RAM memory leaks have been observed when switching models.

If you want to use the NV FP4 optimized kernels for RTX 50xx, you will need PyTorch 2.9.1 with Cuda 13.0

Download Repo and Setup Conda Environment

Clone the repository

First, Create a folder named Wan2GP, then open it, then right click & select "open in terminal", then copy & paste the following commands, one at a time.

git clone https://github.com/deepbeepmeep/Wan2GP.git

Create Python 3.10.9 environment using Conda

conda create -n wan2gp python=3.10.9

Activate Conda Environment

conda activate wan2gp

NOW CHOOSE INSTALLATION ACCORDING TO YOUR GPU

Windows Installation for GTX 10XX -16XX Only

Windows Install PyTorch 2.6.0 with CUDA 12.6 for GTX 10XX -16XX Only

pip install torch==2.6.0+cu126 torchvision==0.21.0+cu126 torchaudio==2.6.0+cu126 --index-url https://download.pytorch.org/whl/cu126

Windows Install requirements.txt for GTX 10XX -16XX Only

pip install -r requirements.txt

Windows Installation for RTX QUADRO - 20XX Only

Windows Install PyTorch 2.6.0 with CUDA 12.6 for RTX QUADRO - 20XX Only

pip install torch==2.6.0+cu126 torchvision==0.21.0+cu126 torchaudio==2.6.0+cu126 --index-url https://download.pytorch.org/whl/cu126

Windows Install Triton for RTX QUADRO - 20XX Only

pip install -U "triton-windows<3.3"

Windows Install Sage1 Attention for RTX QUADRO - 20XX Only

pip install sageattention==1.0.6

Windows Install requirements.txt for RTX QUADRO - 20XX Only

pip install -r requirements.txt

Windows Installation for RTX 30XX Only

Windows Install PyTorch 2.6.0 with CUDA 12.6 for RTX 30XX Only

pip install torch==2.6.0+cu126 torchvision==0.21.0+cu126 torchaudio==2.6.0+cu126 --index-url https://download.pytorch.org/whl/cu126

Windows Install Triton for RTX 30XX Only

pip install -U "triton-windows<3.3"

Windows Install Sage2 Attention for RTX 30XX Only

pip install https://github.com/woct0rdho/SageAttention/releases/download/v2.1.1-windows/sageattention-2.1.1+cu126torch2.6.0-cp310-cp310-win_amd64.whl

Windows Install requirements.txt for RTX 30XX Only

pip install -r requirements.txt

Installation for RTX 40XX, 50XX Only

Windows Install PyTorch 2.7.1 with CUDA 12.8 for RTX 40XX - 50XX Only

pip install torch==2.7.1 torchvision==0.22.1 torchaudio==2.7.1 --index-url https://download.pytorch.org/whl/cu128

Windows Install Triton for RTX 40XX, 50XX Only

pip install -U "triton-windows<3.4"

Windows Install Sage2 Attention for RTX 40XX, 50XX Only

pip install https://github.com/woct0rdho/SageAttention/releases/download/v2.2.0-windows/sageattention-2.2.0+cu128torch2.7.1-cp310-cp310-win_amd64.whl

Windows Install requirements.txt for RTX 40XX, 50XX Only

pip install -r requirements.txt

Installation for 50XX Only PyTorch 2.9.1 Cuda 13.. for NVFP4 optimized kernels

Windows Install PyTorch 2.9.1 with CUDA 13.0 for RTX 50XX Only

pip install torch==2.9.1 torchvision torchaudio --index-url https://download.pytorch.org/whl/cu130

Windows Install Triton for RTX 50XX Only

pip install -U "triton-windows<3.4"

Windows Install Sage2 Attention for RTX 50XX Only

pip install https://github.com/woct0rdho/SageAttention/releases/download/v2.2.0-windows.post4/sageattention-2.2.0+cu130torch2.9.0andhigher.post4-cp39-abi3-win_amd64.whl

Windows Install requirements.txt for RTX 50XX Only

pip install -r requirements.txt

Optional

Flash Attention

Windows

pip install https://github.com/Redtash1/Flash_Attention_2_Windows/releases/download/v2.7.0-v2.7.4/flash_attn-2.7.4.post1+cu128torch2.7.0cxx11abiFALSE-cp310-cp310-win_amd64.whl

Linux Installation

Step 1: Download Repo and Setup Conda Environment

Clone the repository

git clone https://github.com/deepbeepmeep/Wan2GP.git

Change directory

cd Wan2GP

Create Python 3.10.9 environment using Conda

conda create -n wan2gp python=3.10.9

Activate Conda Environment

conda activate wan2gp

Installation for RTX 10XX -16XX Only

Install PyTorch 2.6.0 with CUDA 12.6 for RTX 10XX -16XX Only

pip install torch==2.6.0+cu126 torchvision==0.21.0+cu126 torchaudio==2.6.0+cu126 --index-url https://download.pytorch.org/whl/cu126

Install requirements.txt for RTX 30XX Only

pip install -r requirements.txt

Installation for RTX QUADRO - 20XX Only

Install PyTorch 2.6.0 with CUDA 12.6 for RTX QUADRO - 20XX Only

pip install torch==2.6.0+cu126 torchvision==0.21.0+cu126 torchaudio==2.6.0+cu126 --index-url https://download.pytorch.org/whl/cu126

Install Triton for RTX QUADRO - 20XX Only

pip install -U "triton<3.3"

Install Sage1 Attention for RTX QUADRO - 20XX Only

pip install sageattention==1.0.6

Install requirements.txt for RTX QUADRO - 20XX Only

pip install -r requirements.txt

Installation for RTX 30XX Only

Install PyTorch 2.6.0 with CUDA 12.6 for RTX 30XX Only

pip install torch==2.6.0+cu126 torchvision==0.21.0+cu126 torchaudio==2.6.0+cu126 --index-url https://download.pytorch.org/whl/cu126

Install Triton for RTX 30XX Only

pip install -U "triton<3.3"

Install Sage2 Attention for RTX 30XX Only. Make sure it's Sage 2.1.1

python -m pip install "setuptools<=75.8.2" --force-reinstall
git clone https://github.com/thu-ml/SageAttention
cd SageAttention 
pip install -e .

Install requirements.txt for RTX 30XX Only

pip install -r requirements.txt

Installation for RTX 40XX, 50XX Only

Install PyTorch 2.7.1 with CUDA 12.8 for RTX 40XX - 50XX Only

pip install torch==2.7.1 torchvision==0.22.1 torchaudio==2.7.1 --index-url https://download.pytorch.org/whl/cu128

Install Triton for RTX 40XX, 50XX Only

pip install -U "triton<3.4"

Install Sage Attention for RTX 40XX, 50XX Only. Make sure it's Sage 2.2.0

python -m pip install "setuptools<=75.8.2" --force-reinstall
git clone https://github.com/thu-ml/SageAttention
cd SageAttention 
pip install -e .

Install requirements.txt for RTX 40XX, 50XX Only

pip install -r requirements.txt

Optional

Flash Attention

Linux

pip install flash-attn==2.7.2.post1

Attention Modes

WanGP supports several attention implementations:

  • SDPA (default): Available by default with PyTorch
  • Sage: 30% speed boost with small quality cost
  • Sage2: 40% speed boost
  • Flash: Good performance, may be complex to install on Windows

Attention GPU Compatibility

  • RTX 10XX: SDPA
  • RTX 20XX: SPDA, Sage1
  • RTX 30XX, 40XX: SDPA, Flash Attention, Xformers, Sage1, Sage2/Sage2++
  • RTX 50XX: SDPA, Flash Attention, Xformers, Sage2/Sage2++ / Sage3

Performance Profiles

Choose a profile based on your hardware:

  • Profile 3 (LowRAM_HighVRAM): Loads entire model in VRAM, requires 24GB VRAM for 8-bit quantized 14B model
  • Profile 4 (LowRAM_LowVRAM): Default, loads model parts as needed, slower but lower VRAM requirement

Troubleshooting

Sage Attention Issues

If Sage attention doesn't work:

  1. Check if Triton is properly installed
  2. Clear Triton cache
  3. Fallback to SDPA attention:
    python wgp.py --attention sdpa
    

Memory Issues

  • Use lower resolution or shorter videos
  • Enable quantization (default)
  • Use Profile 4 for lower VRAM usage
  • Consider using 1.3B models instead of 14B models

For more troubleshooting, see TROUBLESHOOTING.md

Optional Kernels for INT4 / FP4 quantized support

These kernels will offer optimized INT4 / FP4 dequantization.

Please Note FP4 support is hardware dependent and will work only with RTX 50xx / sm120+ GPUs

Light2xv NVP4 Kernels Wheels for Windows / Python 3.10 / Pytorch 2.9.1 / Cuda 13.8 (RTX 50xx / sm120+ only !)

  • Windows

    pip install https://github.com/deepbeepmeep/kernels/releases/download/Light2xv/lightx2v_kernel-0.0.1+torch2.9.1-cp39-abi3-win_amd64.whl
    
  • Linux

    pip install https://github.com/deepbeepmeep/kernels/releases/download/Light2xv/lightx2v_kernel-0.0.1+torch2.9.1cu130-cp39-abi3-linux_x86_64.whl
    

Nunchaku INT4/FP4 Kernels Wheels for Python 3.10

  • Windows (Pytorch 2.7.1 / Cuda 12.8)
    
    

pip install https://github.com/deepbeepmeep/kernels/releases/download/v1.2.0_Nunchaku/nunchaku-1.2.0+torch2.7-cp310-cp310-win_amd64.whl

- Windows (Pytorch 2.9.1 / Cuda 13)

pip install https://github.com/deepbeepmeep/kernels/releases/download/v1.2.0_Nunchaku/nunchaku-1.2.0+torch2.9-cp310-cp310-win_amd64.whl


- Linux (Pytorch 2.7.1 / Cuda 12.8) 

pip install https://github.com/deepbeepmeep/kernels/releases/download/v1.2.0_Nunchaku/nunchaku-1.2.0+torch2.7-cp310-cp310-linux_x86_64.whl

- Windows (Pytorch 2.9.1 / Cuda 13)

pip install https://github.com/deepbeepmeep/kernels/releases/download/v1.2.0_Nunchaku/nunchaku-1.2.0+torch2.9-cp310-cp310-linux_x86_64.whl ```