aliensmn commited on
Commit
21e0c24
·
verified ·
1 Parent(s): e433ec2

Mirror from Kijai/PrecompiledWheels

Browse files
.gitattributes CHANGED
@@ -33,3 +33,12 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ sageattention-2.1.0-cp312-cp312-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
37
+ triton-3.2.0-cp312-cp312-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
38
+ sageattention-2.1.1-cp312-cp312-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
39
+ flash_attn-2.7.4.post1-cp312-cp312-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
40
+ triton-3.3.0-cp312-cp312-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
41
+ flash_attn-2.7.4+cu128torch2.6.0cxx11abiFALSE-cp312-cp312-win_amd64.whl filter=lfs diff=lfs merge=lfs -text
42
+ sageattention-2.2.0-cp312-cp312-win_amd64.whl filter=lfs diff=lfs merge=lfs -text
43
+ sageattention-2.2.0-cp312-cp312-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
44
+ spas_sage_attn-0.1.0-cp312-cp312-win_amd64.whl filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ For testing Blackwell torch.compile and sageattention. Confirmed working.
2
+
3
+ Compiled on Debian 13 testing with torch 2.7 nightly, cu128
flash_attn-2.7.4+cu128torch2.6.0cxx11abiFALSE-cp312-cp312-win_amd64.whl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:961dc02509cf6c7ec6a31e88450946d8d7e9070f5f20daba5d2ea2a8e6b7b954
3
+ size 403043198
flash_attn-2.7.4.post1-cp312-cp312-linux_x86_64.whl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4bd742601695c31d541dc7b576d7f5cf4fece99b37e224daf47bef034938abc
3
+ size 583520957
sageattention-2.1.0-cp312-cp312-linux_x86_64.whl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a919e917730b3d0d1d245359e1b866df0f3fbd6b9e90fc2fb24c06025b8f7f0
3
+ size 11539848
sageattention-2.1.1-cp312-cp312-linux_x86_64.whl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a7b4a47271eacc1c772fe2186374853c8a738fd2a440a0e3312fb7fab41c234
3
+ size 11539980
sageattention-2.2.0-cp312-cp312-linux_x86_64.whl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78bd9b58b948e53b358160c65c265f9f9c1788b81f9a3b771c810aca9bf77bf5
3
+ size 14571398
sageattention-2.2.0-cp312-cp312-win_amd64.whl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0aec2881718f98cf3e4ff9ed241d08390dfe4087601a6aaed41cc8806e211316
3
+ size 7135099
sparse_sageattn-0.1.0-py3-none-any.whl ADDED
Binary file (11.8 kB). View file
 
spas_sage_attn-0.1.0-cp312-cp312-win_amd64.whl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dec6a3e27f16d316b2bdfd07193078636aeb1b5ce2bc0af8ad8bdac32e56267a
3
+ size 3034418
triton-3.2.0-cp312-cp312-linux_x86_64.whl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8153ad08343525e81558fc056d0a41f510008c1e5fc55402c6a4594e7b48b340
3
+ size 250718730
triton-3.3.0-cp312-cp312-linux_x86_64.whl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e84f2992ba9b0cf02ed7d95633ef1147efe774e86c8936a88448ca9316aebb0
3
+ size 258715429