Spaces:
Sleeping
Sleeping
| # Installation | |
| ## Installing from source | |
| Installing from source requires the latest CUDA Toolkit that matches the major.minor of CUDA Python installed. | |
| Prior to installing the CUTLASS Python interface, one may optionally set the following environment variables: | |
| * `CUTLASS_PATH`: the path to the cloned CUTLASS repository | |
| * `CUDA_INSTALL_PATH`: the path to the installation of CUDA | |
| If these environment variables are not set, the installation process will infer them to be the following: | |
| * `CUTLASS_PATH`: one directory level above the current directory (i.e., `$(pwd)/..`) | |
| * `CUDA_INSTALL_PATH`: the directory holding `/bin/nvcc` for the first version of `nvcc` on `$PATH` (i.e., `which nvcc | awk -F'/bin/nvcc' '{print $1}'`) | |
| **NOTE:** The version of `cuda-python` installed must match the CUDA version in `CUDA_INSTALL_PATH`. | |
| ### Installing a developer-mode package | |
| The CUTLASS Python interface can currently be installed via: | |
| ```bash | |
| python setup.py develop --user | |
| ``` | |
| This will allow changes to the Python interface source to be reflected when using the Python interface. | |
| We plan to add support for installing via `python setup.py install` in a future release. | |
| ## Docker | |
| To ensure that you have all of the necessary Python modules for running the examples using the | |
| CUTLASS Python interface, we recommend using one of the Docker images located in the docker directory. | |
| For example, to build and launch a container that uses CUDA 12.1 via an NGC PyTorch container, run: | |
| ```bash | |
| docker build -t cutlass-cuda12.1:latest -f docker/Dockerfile-cuda12.1-pytorch . | |
| docker run --gpus all -it --rm cutlass-cuda12.1:latest | |
| ``` | |
| The CUTLASS Python interface has been tested with CUDA 11.8, 12.0, and 12.1 on Python 3.8.10 and 3.9.7. | |