update [markdown] - [cleaned] ✅
#9
by
prithivMLmods - opened
README.md
CHANGED
|
@@ -4,12 +4,15 @@ language:
|
|
| 4 |
tags:
|
| 5 |
- flash_attention_2
|
| 6 |
- pytorch
|
|
|
|
| 7 |
- text-generation-inference
|
| 8 |
license: bsd-3-clause
|
| 9 |
---
|
| 10 |
|
| 11 |

|
| 12 |
|
|
|
|
|
|
|
| 13 |
# **Flash Attention 2 Pre-built Wheels**
|
| 14 |
|
| 15 |
> [!IMPORTANT]
|
|
@@ -17,19 +20,19 @@ license: bsd-3-clause
|
|
| 17 |
|
| 18 |
This repository provides pre-built wheels for `flash-attn` version **2.8.3** for various PyTorch versions, Python versions, and architectures (compiled with CUDA 12). You can install these directly using `pip install <url>` or add the provided strings directly to your `requirements.txt`.
|
| 19 |
|
| 20 |
-
> [!
|
| 21 |
> The detailed categories and structured view of the `strangertoolshf/flash_attention_2_wheelhouse` folders and files on the hf-tree shareable link are available here: [huggingface-tree](https://strangertoolshf-huggingface-tree.static.hf.space/index.html#models/strangertoolshf/flash_attention_2_wheelhouse/main)
|
| 22 |
|
| 23 |
<div style="
|
| 24 |
-
background: rgba(
|
| 25 |
padding: 16px;
|
| 26 |
border-radius: 6px;
|
| 27 |
-
border: 1px solid rgba(
|
| 28 |
margin: 16px 0;
|
| 29 |
">
|
| 30 |
|
| 31 |
<details>
|
| 32 |
-
<summary>
|
| 33 |
|
| 34 |

|
| 35 |
|
|
@@ -39,6 +42,27 @@ This repository provides pre-built wheels for `flash-attn` version **2.8.3** for
|
|
| 39 |
|
| 40 |
---
|
| 41 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 42 |
## Linux x86_64
|
| 43 |
|
| 44 |
### **<span style="color:orangered;">Torch 2.9</span>**
|
|
|
|
| 4 |
tags:
|
| 5 |
- flash_attention_2
|
| 6 |
- pytorch
|
| 7 |
+
- flash-attn @ <>
|
| 8 |
- text-generation-inference
|
| 9 |
license: bsd-3-clause
|
| 10 |
---
|
| 11 |
|
| 12 |

|
| 13 |
|
| 14 |
+
`flash-attn @ <>`
|
| 15 |
+
|
| 16 |
# **Flash Attention 2 Pre-built Wheels**
|
| 17 |
|
| 18 |
> [!IMPORTANT]
|
|
|
|
| 20 |
|
| 21 |
This repository provides pre-built wheels for `flash-attn` version **2.8.3** for various PyTorch versions, Python versions, and architectures (compiled with CUDA 12). You can install these directly using `pip install <url>` or add the provided strings directly to your `requirements.txt`.
|
| 22 |
|
| 23 |
+
> [!IMPORTANT]
|
| 24 |
> The detailed categories and structured view of the `strangertoolshf/flash_attention_2_wheelhouse` folders and files on the hf-tree shareable link are available here: [huggingface-tree](https://strangertoolshf-huggingface-tree.static.hf.space/index.html#models/strangertoolshf/flash_attention_2_wheelhouse/main)
|
| 25 |
|
| 26 |
<div style="
|
| 27 |
+
background: rgba(61, 122, 255, 0.15);
|
| 28 |
padding: 16px;
|
| 29 |
border-radius: 6px;
|
| 30 |
+
border: 1px solid rgba(0, 102, 255, 0.3);
|
| 31 |
margin: 16px 0;
|
| 32 |
">
|
| 33 |
|
| 34 |
<details>
|
| 35 |
+
<summary>Open the sample hf-tree screenshot</summary>
|
| 36 |
|
| 37 |

|
| 38 |
|
|
|
|
| 42 |
|
| 43 |
---
|
| 44 |
|
| 45 |
+
## **Quick Start with <span style="color:orangered;">StrangerTools-HF</span>**
|
| 46 |
+
|
| 47 |
+
```
|
| 48 |
+
flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-<FLASH_ATTN_VERSION>/<ARCHITECTURE>/torch<TORCH_VERSION>/cu<CUDA_VERSION>/abi<ABI_FLAG>/<PYTHON_TAG>/flash_attn-<FLASH_ATTN_VERSION>+cu<CUDA_VERSION>torch<TORCH_VERSION>cxx11abi<ABI_FLAG>-<PYTHON_TAG>-<PYTHON_TAG>-<ARCHITECTURE>.whl
|
| 49 |
+
|
| 50 |
+
```
|
| 51 |
+
|
| 52 |
+
### Breakdown of the Variables:
|
| 53 |
+
|
| 54 |
+
* **`<FLASH_ATTN_VERSION>`**: The version of Flash Attention (e.g., `2.8.3`).
|
| 55 |
+
* **`<ARCHITECTURE>`**: The OS and system architecture (e.g., `linux_x86_64`, `linux_aarch64`).
|
| 56 |
+
* **`<TORCH_VERSION>`**: The PyTorch version without the minor patch (e.g., `2.4`, `2.5`, `2.8`).
|
| 57 |
+
* **`<CUDA_VERSION>`**: The CUDA major version (e.g., `12`).
|
| 58 |
+
* **`<ABI_FLAG>`**: The CXX11 ABI boolean flag (e.g., `TRUE`, `FALSE`).
|
| 59 |
+
* **`<PYTHON_TAG>`**: The specific CPython version tag (e.g., `cp39`, `cp310`, `cp312`).
|
| 60 |
+
|
| 61 |
+
> [!IMPORTANT]
|
| 62 |
+
Find all the provided pre-built wheels below.👇
|
| 63 |
+
|
| 64 |
+
---
|
| 65 |
+
|
| 66 |
## Linux x86_64
|
| 67 |
|
| 68 |
### **<span style="color:orangered;">Torch 2.9</span>**
|