update [markdown] - [cleaned] ✅

#9
Files changed (1) hide show
  1. README.md +28 -4
README.md CHANGED
@@ -4,12 +4,15 @@ language:
4
  tags:
5
  - flash_attention_2
6
  - pytorch
 
7
  - text-generation-inference
8
  license: bsd-3-clause
9
  ---
10
 
11
  ![flashattention_logo](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/yQ8ugTupP2yLUV9sZ4HXt.png)
12
 
 
 
13
  # **Flash Attention 2 Pre-built Wheels**
14
 
15
  > [!IMPORTANT]
@@ -17,19 +20,19 @@ license: bsd-3-clause
17
 
18
  This repository provides pre-built wheels for `flash-attn` version **2.8.3** for various PyTorch versions, Python versions, and architectures (compiled with CUDA 12). You can install these directly using `pip install <url>` or add the provided strings directly to your `requirements.txt`.
19
 
20
- > [!NOTE]
21
  > The detailed categories and structured view of the `strangertoolshf/flash_attention_2_wheelhouse` folders and files on the hf-tree shareable link are available here: [huggingface-tree](https://strangertoolshf-huggingface-tree.static.hf.space/index.html#models/strangertoolshf/flash_attention_2_wheelhouse/main)
22
 
23
  <div style="
24
- background: rgba(255, 61, 61, 0.15);
25
  padding: 16px;
26
  border-radius: 6px;
27
- border: 1px solid rgba(255, 0, 0, 0.3);
28
  margin: 16px 0;
29
  ">
30
 
31
  <details>
32
- <summary>open the sample hf-tree screenshot</summary>
33
 
34
  ![Screenshot 2026-02-17 213525](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/SDJRQ7mKxpY21rVLaneSD.png)
35
 
@@ -39,6 +42,27 @@ This repository provides pre-built wheels for `flash-attn` version **2.8.3** for
39
 
40
  ---
41
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
42
  ## Linux x86_64
43
 
44
  ### **<span style="color:orangered;">Torch 2.9</span>**
 
4
  tags:
5
  - flash_attention_2
6
  - pytorch
7
+ - flash-attn @ <>
8
  - text-generation-inference
9
  license: bsd-3-clause
10
  ---
11
 
12
  ![flashattention_logo](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/yQ8ugTupP2yLUV9sZ4HXt.png)
13
 
14
+ `flash-attn @ <>`
15
+
16
  # **Flash Attention 2 Pre-built Wheels**
17
 
18
  > [!IMPORTANT]
 
20
 
21
  This repository provides pre-built wheels for `flash-attn` version **2.8.3** for various PyTorch versions, Python versions, and architectures (compiled with CUDA 12). You can install these directly using `pip install <url>` or add the provided strings directly to your `requirements.txt`.
22
 
23
+ > [!IMPORTANT]
24
  > The detailed categories and structured view of the `strangertoolshf/flash_attention_2_wheelhouse` folders and files on the hf-tree shareable link are available here: [huggingface-tree](https://strangertoolshf-huggingface-tree.static.hf.space/index.html#models/strangertoolshf/flash_attention_2_wheelhouse/main)
25
 
26
  <div style="
27
+ background: rgba(61, 122, 255, 0.15);
28
  padding: 16px;
29
  border-radius: 6px;
30
+ border: 1px solid rgba(0, 102, 255, 0.3);
31
  margin: 16px 0;
32
  ">
33
 
34
  <details>
35
+ <summary>Open the sample hf-tree screenshot</summary>
36
 
37
  ![Screenshot 2026-02-17 213525](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/SDJRQ7mKxpY21rVLaneSD.png)
38
 
 
42
 
43
  ---
44
 
45
+ ## **Quick Start with <span style="color:orangered;">StrangerTools-HF</span>**
46
+
47
+ ```
48
+ flash-attn @ https://huggingface.co/strangertoolshf/flash_attention_2_wheelhouse/resolve/main/wheelhouse-flash_attn-<FLASH_ATTN_VERSION>/<ARCHITECTURE>/torch<TORCH_VERSION>/cu<CUDA_VERSION>/abi<ABI_FLAG>/<PYTHON_TAG>/flash_attn-<FLASH_ATTN_VERSION>+cu<CUDA_VERSION>torch<TORCH_VERSION>cxx11abi<ABI_FLAG>-<PYTHON_TAG>-<PYTHON_TAG>-<ARCHITECTURE>.whl
49
+
50
+ ```
51
+
52
+ ### Breakdown of the Variables:
53
+
54
+ * **`<FLASH_ATTN_VERSION>`**: The version of Flash Attention (e.g., `2.8.3`).
55
+ * **`<ARCHITECTURE>`**: The OS and system architecture (e.g., `linux_x86_64`, `linux_aarch64`).
56
+ * **`<TORCH_VERSION>`**: The PyTorch version without the minor patch (e.g., `2.4`, `2.5`, `2.8`).
57
+ * **`<CUDA_VERSION>`**: The CUDA major version (e.g., `12`).
58
+ * **`<ABI_FLAG>`**: The CXX11 ABI boolean flag (e.g., `TRUE`, `FALSE`).
59
+ * **`<PYTHON_TAG>`**: The specific CPython version tag (e.g., `cp39`, `cp310`, `cp312`).
60
+
61
+ > [!IMPORTANT]
62
+ Find all the provided pre-built wheels below.👇
63
+
64
+ ---
65
+
66
  ## Linux x86_64
67
 
68
  ### **<span style="color:orangered;">Torch 2.9</span>**