Upload 9 files
Browse files- .gitattributes +6 -0
- flash_attn-2.8.3+torch2.9.1.cuda13.1-cp310-cp310-linux_x86_64.whl +3 -0
- flash_attn-2.8.3+torch2.9.1.cuda13.1-cp311-cp311-win_amd64.whl +3 -0
- flash_attn-2.8.3+torch2.9.1.cuda13.1-cp312-cp312-win_amd64.whl +3 -0
- flash_attn-2.8.3+torch2.9.1.cuda13.1-cp313-cp313-win_amd64.whl +3 -0
- sageattention-2.2.0+torch2.9.1.cuda13.1-cp39-abi3-linux_x86_64.whl +3 -0
- xformers-0.0.34+41531cee.d20260109-cp39-abi3-linux_x86_64.whl +3 -0
.gitattributes
CHANGED
|
@@ -125,3 +125,9 @@ cudnn_9.17.1_windows_x86_64.exe filter=lfs diff=lfs merge=lfs -text
|
|
| 125 |
flash_attn-2.8.3+torch2.9.1.cuda13.1-cp310-cp310-win_amd64.whl filter=lfs diff=lfs merge=lfs -text
|
| 126 |
sageattention-2.2.0+torch2.9.1.cuda13.1-cp39-abi3-win_amd64.whl filter=lfs diff=lfs merge=lfs -text
|
| 127 |
xformers-0.0.34+41531cee.d20260109-cp39-abi3-win_amd64.whl filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 125 |
flash_attn-2.8.3+torch2.9.1.cuda13.1-cp310-cp310-win_amd64.whl filter=lfs diff=lfs merge=lfs -text
|
| 126 |
sageattention-2.2.0+torch2.9.1.cuda13.1-cp39-abi3-win_amd64.whl filter=lfs diff=lfs merge=lfs -text
|
| 127 |
xformers-0.0.34+41531cee.d20260109-cp39-abi3-win_amd64.whl filter=lfs diff=lfs merge=lfs -text
|
| 128 |
+
flash_attn-2.8.3+torch2.9.1.cuda13.1-cp310-cp310-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
| 129 |
+
flash_attn-2.8.3+torch2.9.1.cuda13.1-cp311-cp311-win_amd64.whl filter=lfs diff=lfs merge=lfs -text
|
| 130 |
+
flash_attn-2.8.3+torch2.9.1.cuda13.1-cp312-cp312-win_amd64.whl filter=lfs diff=lfs merge=lfs -text
|
| 131 |
+
flash_attn-2.8.3+torch2.9.1.cuda13.1-cp313-cp313-win_amd64.whl filter=lfs diff=lfs merge=lfs -text
|
| 132 |
+
sageattention-2.2.0+torch2.9.1.cuda13.1-cp39-abi3-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
| 133 |
+
xformers-0.0.34+41531cee.d20260109-cp39-abi3-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
flash_attn-2.8.3+torch2.9.1.cuda13.1-cp310-cp310-linux_x86_64.whl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1eab7fa5fc05b6ece00121ab03a9ed08f7d4a0e0db8afdc7609f1e3674cb7a89
|
| 3 |
+
size 378737092
|
flash_attn-2.8.3+torch2.9.1.cuda13.1-cp311-cp311-win_amd64.whl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:438342f82485bd92ce65533398d331490b460327c97ea7e7f2a25effb6682f30
|
| 3 |
+
size 375657566
|
flash_attn-2.8.3+torch2.9.1.cuda13.1-cp312-cp312-win_amd64.whl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cc7e1899b127d64ef9a28f5c35308b52547f7adab9b9aebeafb749ef6a71630b
|
| 3 |
+
size 375662974
|
flash_attn-2.8.3+torch2.9.1.cuda13.1-cp313-cp313-win_amd64.whl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fcb58148d4d7382fedca41bd0e8fbb4d55e9c0543afaef5a75fb65d0c6202427
|
| 3 |
+
size 375662955
|
sageattention-2.2.0+torch2.9.1.cuda13.1-cp39-abi3-linux_x86_64.whl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:db5791e88f13d4728b1e0897fd1aa0104afde989a60b393905ee579ca7c8428f
|
| 3 |
+
size 19984366
|
xformers-0.0.34+41531cee.d20260109-cp39-abi3-linux_x86_64.whl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:636f777cc162a28efd5390727c96eee04d81f632bf39f898cca4f6b17b9c906e
|
| 3 |
+
size 91205834
|