jena-shreyas commited on
Commit
f5837a1
·
1 Parent(s): 6b04183

1. Add flash-attn 2.8.3 release link with ABI=False since it works on HF Spaces

Browse files

2. Drop torch (2.9.0 -> 2.8) since 2.9+ABI=False binary unavailable
3. Drop torchvision to 0.23 to try match with torch==2.8

Files changed (2) hide show
  1. README.md +1 -1
  2. requirements.txt +3 -3
README.md CHANGED
@@ -6,7 +6,7 @@ colorTo: blue
6
  sdk: gradio
7
  sdk_version: 6.5.1
8
  app_file: app.py
9
- python_version: "3.10"
10
  pinned: false
11
  license: apache-2.0
12
  ---
 
6
  sdk: gradio
7
  sdk_version: 6.5.1
8
  app_file: app.py
9
+ python_version: "3.12"
10
  pinned: false
11
  license: apache-2.0
12
  ---
requirements.txt CHANGED
@@ -17,8 +17,8 @@ nvidia-cuda-runtime-cu12==12.8.90
17
  nvidia-cudnn-cu12==9.10.2.21
18
  nvidia-nccl-cu12==2.27.5
19
  nvidia-nvtx-cu12==12.8.90
20
- torch==2.9.0
21
- torchvision==0.24.0
22
 
23
  # Progress and utilities
24
  tqdm==4.67.1
@@ -34,7 +34,7 @@ transformers==5.1.0
34
  accelerate==1.11.0
35
  datasets==4.5.0
36
  einops==0.8.2
37
- flash-attn==2.8.3
38
 
39
  # Additional deep learning utilities
40
  bitsandbytes==0.49.1
 
17
  nvidia-cudnn-cu12==9.10.2.21
18
  nvidia-nccl-cu12==2.27.5
19
  nvidia-nvtx-cu12==12.8.90
20
+ torch==2.8
21
+ torchvision==0.23
22
 
23
  # Progress and utilities
24
  tqdm==4.67.1
 
34
  accelerate==1.11.0
35
  datasets==4.5.0
36
  einops==0.8.2
37
+ flash-attn @ https://github.com/Dao-AILab/flash-attention/releases/download/v2.8.3/flash_attn-2.8.3+cu12torch2.8cxx11abiFALSE-cp312-cp312-linux_x86_64.whl
38
 
39
  # Additional deep learning utilities
40
  bitsandbytes==0.49.1