Spaces:
Sleeping
Sleeping
Commit ·
f5837a1
1
Parent(s): 6b04183
1. Add flash-attn 2.8.3 release link with ABI=False since it works on HF Spaces
Browse files2. Drop torch (2.9.0 -> 2.8) since 2.9+ABI=False binary unavailable
3. Drop torchvision to 0.23 to try match with torch==2.8
- README.md +1 -1
- requirements.txt +3 -3
README.md
CHANGED
|
@@ -6,7 +6,7 @@ colorTo: blue
|
|
| 6 |
sdk: gradio
|
| 7 |
sdk_version: 6.5.1
|
| 8 |
app_file: app.py
|
| 9 |
-
python_version: "3.
|
| 10 |
pinned: false
|
| 11 |
license: apache-2.0
|
| 12 |
---
|
|
|
|
| 6 |
sdk: gradio
|
| 7 |
sdk_version: 6.5.1
|
| 8 |
app_file: app.py
|
| 9 |
+
python_version: "3.12"
|
| 10 |
pinned: false
|
| 11 |
license: apache-2.0
|
| 12 |
---
|
requirements.txt
CHANGED
|
@@ -17,8 +17,8 @@ nvidia-cuda-runtime-cu12==12.8.90
|
|
| 17 |
nvidia-cudnn-cu12==9.10.2.21
|
| 18 |
nvidia-nccl-cu12==2.27.5
|
| 19 |
nvidia-nvtx-cu12==12.8.90
|
| 20 |
-
torch==2.
|
| 21 |
-
torchvision==0.
|
| 22 |
|
| 23 |
# Progress and utilities
|
| 24 |
tqdm==4.67.1
|
|
@@ -34,7 +34,7 @@ transformers==5.1.0
|
|
| 34 |
accelerate==1.11.0
|
| 35 |
datasets==4.5.0
|
| 36 |
einops==0.8.2
|
| 37 |
-
flash-attn
|
| 38 |
|
| 39 |
# Additional deep learning utilities
|
| 40 |
bitsandbytes==0.49.1
|
|
|
|
| 17 |
nvidia-cudnn-cu12==9.10.2.21
|
| 18 |
nvidia-nccl-cu12==2.27.5
|
| 19 |
nvidia-nvtx-cu12==12.8.90
|
| 20 |
+
torch==2.8
|
| 21 |
+
torchvision==0.23
|
| 22 |
|
| 23 |
# Progress and utilities
|
| 24 |
tqdm==4.67.1
|
|
|
|
| 34 |
accelerate==1.11.0
|
| 35 |
datasets==4.5.0
|
| 36 |
einops==0.8.2
|
| 37 |
+
flash-attn @ https://github.com/Dao-AILab/flash-attention/releases/download/v2.8.3/flash_attn-2.8.3+cu12torch2.8cxx11abiFALSE-cp312-cp312-linux_x86_64.whl
|
| 38 |
|
| 39 |
# Additional deep learning utilities
|
| 40 |
bitsandbytes==0.49.1
|