Kernels
danieldk HF Staff commited on
Commit
79c7050
·
verified ·
1 Parent(s): af1ed51

Build uploaded using `kernels`.

Browse files
Files changed (35) hide show
  1. build/torch210-cxx11-cu126-x86_64-linux/_ops.py +3 -3
  2. build/torch210-cxx11-cu126-x86_64-linux/{_paged_attention_7c06fa7.abi3.so → _paged_attention_cuda_c6404f1.abi3.so} +2 -2
  3. build/torch210-cxx11-cu126-x86_64-linux/metadata.json +16 -2
  4. build/torch210-cxx11-cu128-x86_64-linux/_ops.py +3 -3
  5. build/torch210-cxx11-cu128-x86_64-linux/{_paged_attention_7c06fa7.abi3.so → _paged_attention_cuda_c6404f1.abi3.so} +2 -2
  6. build/torch210-cxx11-cu128-x86_64-linux/metadata.json +19 -2
  7. build/torch210-cxx11-cu130-x86_64-linux/_ops.py +3 -3
  8. build/{torch29-cxx11-cu130-x86_64-linux/_paged_attention_7c06fa7.abi3.so → torch210-cxx11-cu130-x86_64-linux/_paged_attention_cuda_c6404f1.abi3.so} +2 -2
  9. build/torch210-cxx11-cu130-x86_64-linux/metadata.json +17 -2
  10. build/torch210-cxx11-rocm70-x86_64-linux/_ops.py +3 -3
  11. build/{torch210-cxx11-cu130-x86_64-linux/_paged_attention_7c06fa7.abi3.so → torch210-cxx11-rocm70-x86_64-linux/_paged_attention_rocm_c6404f1.abi3.so} +2 -2
  12. build/torch210-cxx11-rocm70-x86_64-linux/metadata.json +15 -2
  13. build/torch210-cxx11-rocm71-x86_64-linux/_ops.py +3 -3
  14. build/torch210-cxx11-rocm71-x86_64-linux/_paged_attention_7c06fa7.abi3.so +0 -3
  15. build/{torch210-cxx11-rocm70-x86_64-linux/_paged_attention_7c06fa7.abi3.so → torch210-cxx11-rocm71-x86_64-linux/_paged_attention_rocm_c6404f1.abi3.so} +2 -2
  16. build/torch210-cxx11-rocm71-x86_64-linux/metadata.json +15 -2
  17. build/torch29-cxx11-cu126-x86_64-linux/_ops.py +3 -3
  18. build/torch29-cxx11-cu126-x86_64-linux/_paged_attention_7c06fa7.abi3.so +0 -3
  19. build/torch29-cxx11-cu126-x86_64-linux/_paged_attention_cuda_c6404f1.abi3.so +3 -0
  20. build/torch29-cxx11-cu126-x86_64-linux/metadata.json +16 -2
  21. build/torch29-cxx11-cu128-x86_64-linux/_ops.py +3 -3
  22. build/torch29-cxx11-cu128-x86_64-linux/_paged_attention_7c06fa7.abi3.so +0 -3
  23. build/torch29-cxx11-cu128-x86_64-linux/_paged_attention_cuda_c6404f1.abi3.so +3 -0
  24. build/torch29-cxx11-cu128-x86_64-linux/metadata.json +19 -2
  25. build/torch29-cxx11-cu130-x86_64-linux/_ops.py +3 -3
  26. build/torch29-cxx11-cu130-x86_64-linux/_paged_attention_cuda_c6404f1.abi3.so +3 -0
  27. build/torch29-cxx11-cu130-x86_64-linux/metadata.json +17 -2
  28. build/torch29-cxx11-rocm63-x86_64-linux/_ops.py +3 -3
  29. build/torch29-cxx11-rocm63-x86_64-linux/_paged_attention_7c06fa7.abi3.so +0 -3
  30. build/torch29-cxx11-rocm63-x86_64-linux/_paged_attention_rocm_c6404f1.abi3.so +3 -0
  31. build/torch29-cxx11-rocm63-x86_64-linux/metadata.json +15 -2
  32. build/torch29-cxx11-rocm64-x86_64-linux/_ops.py +3 -3
  33. build/torch29-cxx11-rocm64-x86_64-linux/_paged_attention_7c06fa7.abi3.so +0 -3
  34. build/torch29-cxx11-rocm64-x86_64-linux/_paged_attention_rocm_c6404f1.abi3.so +3 -0
  35. build/torch29-cxx11-rocm64-x86_64-linux/metadata.json +15 -2
build/torch210-cxx11-cu126-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _paged_attention_7c06fa7
3
- ops = torch.ops._paged_attention_7c06fa7
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_paged_attention_7c06fa7::{op_name}"
 
1
  import torch
2
+ from . import _paged_attention_cuda_c6404f1
3
+ ops = torch.ops._paged_attention_cuda_c6404f1
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_paged_attention_cuda_c6404f1::{op_name}"
build/torch210-cxx11-cu126-x86_64-linux/{_paged_attention_7c06fa7.abi3.so → _paged_attention_cuda_c6404f1.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6ae9817078adf5def219bc7e9b378db9941bba44afb42db6f6a54f95ddd1e797
3
- size 140162672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ab6d2e49a9f6a1b06335b4c18d09192b754f2a57d5c71ee8b6d30b5b0db4259
3
+ size 140162704
build/torch210-cxx11-cu126-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,18 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "cuda",
7
+ "archs": [
8
+ "7.0",
9
+ "7.2",
10
+ "7.5",
11
+ "8.0",
12
+ "8.6",
13
+ "8.7",
14
+ "8.9",
15
+ "9.0+PTX"
16
+ ]
17
+ }
18
+ }
build/torch210-cxx11-cu128-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _paged_attention_7c06fa7
3
- ops = torch.ops._paged_attention_7c06fa7
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_paged_attention_7c06fa7::{op_name}"
 
1
  import torch
2
+ from . import _paged_attention_cuda_c6404f1
3
+ ops = torch.ops._paged_attention_cuda_c6404f1
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_paged_attention_cuda_c6404f1::{op_name}"
build/torch210-cxx11-cu128-x86_64-linux/{_paged_attention_7c06fa7.abi3.so → _paged_attention_cuda_c6404f1.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b5c4e1a932c63c6b9d5b4914c788bb8097315b2c4ac8c98c8e2ca3fc4dc9ec4e
3
- size 167726072
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13321e903d6ffb8b39126a53e30d9509f5e78ac84a0393d2b8825590b8022216
3
+ size 167726096
build/torch210-cxx11-cu128-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,21 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "cuda",
7
+ "archs": [
8
+ "10.0",
9
+ "10.1",
10
+ "12.0+PTX",
11
+ "7.0",
12
+ "7.2",
13
+ "7.5",
14
+ "8.0",
15
+ "8.6",
16
+ "8.7",
17
+ "8.9",
18
+ "9.0"
19
+ ]
20
+ }
21
+ }
build/torch210-cxx11-cu130-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _paged_attention_7c06fa7
3
- ops = torch.ops._paged_attention_7c06fa7
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_paged_attention_7c06fa7::{op_name}"
 
1
  import torch
2
+ from . import _paged_attention_cuda_c6404f1
3
+ ops = torch.ops._paged_attention_cuda_c6404f1
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_paged_attention_cuda_c6404f1::{op_name}"
build/{torch29-cxx11-cu130-x86_64-linux/_paged_attention_7c06fa7.abi3.so → torch210-cxx11-cu130-x86_64-linux/_paged_attention_cuda_c6404f1.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:abe0886aae7fb2eba00149a0966eef1a21e53f58480d934396c8efcb29aa4656
3
- size 86544192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8487f185f808c2f07993a5ddd9b66ddac8c86379ea2218fe9694908754468c2
3
+ size 86563792
build/torch210-cxx11-cu130-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,19 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "cuda",
7
+ "archs": [
8
+ "10.0",
9
+ "11.0",
10
+ "12.0+PTX",
11
+ "7.5",
12
+ "8.0",
13
+ "8.6",
14
+ "8.7",
15
+ "8.9",
16
+ "9.0"
17
+ ]
18
+ }
19
+ }
build/torch210-cxx11-rocm70-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _paged_attention_7c06fa7
3
- ops = torch.ops._paged_attention_7c06fa7
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_paged_attention_7c06fa7::{op_name}"
 
1
  import torch
2
+ from . import _paged_attention_rocm_c6404f1
3
+ ops = torch.ops._paged_attention_rocm_c6404f1
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_paged_attention_rocm_c6404f1::{op_name}"
build/{torch210-cxx11-cu130-x86_64-linux/_paged_attention_7c06fa7.abi3.so → torch210-cxx11-rocm70-x86_64-linux/_paged_attention_rocm_c6404f1.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa4562017b329f203ceae4d209c6500540c02d01e545ac7a3a7be8ec5c429aaa
3
- size 86563760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c574035cf116ab7ce297c150ca7b36d4f1ea43e9abbf15d5872e1da38d95e5e
3
+ size 58992416
build/torch210-cxx11-rocm70-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,17 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "rocm",
7
+ "archs": [
8
+ "gfx1030",
9
+ "gfx1100",
10
+ "gfx1101",
11
+ "gfx906",
12
+ "gfx908",
13
+ "gfx90a",
14
+ "gfx942"
15
+ ]
16
+ }
17
+ }
build/torch210-cxx11-rocm71-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _paged_attention_7c06fa7
3
- ops = torch.ops._paged_attention_7c06fa7
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_paged_attention_7c06fa7::{op_name}"
 
1
  import torch
2
+ from . import _paged_attention_rocm_c6404f1
3
+ ops = torch.ops._paged_attention_rocm_c6404f1
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_paged_attention_rocm_c6404f1::{op_name}"
build/torch210-cxx11-rocm71-x86_64-linux/_paged_attention_7c06fa7.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:87076139e88ed1200757ae53d0644cddfb3dad3080c579fb069def2d5a72842b
3
- size 118089384
 
 
 
 
build/{torch210-cxx11-rocm70-x86_64-linux/_paged_attention_7c06fa7.abi3.so → torch210-cxx11-rocm71-x86_64-linux/_paged_attention_rocm_c6404f1.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f6f45d5f4f2788ebd424b0d50360e60808dad63792f25cd3dbb60fdd9e9762a5
3
- size 118183688
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f24c67c8ead8e4c43e57c7318d588d4957a315363f85c9e3257794f244e7224
3
+ size 58971840
build/torch210-cxx11-rocm71-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,17 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "rocm",
7
+ "archs": [
8
+ "gfx1030",
9
+ "gfx1100",
10
+ "gfx1101",
11
+ "gfx906",
12
+ "gfx908",
13
+ "gfx90a",
14
+ "gfx942"
15
+ ]
16
+ }
17
+ }
build/torch29-cxx11-cu126-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _paged_attention_7c06fa7
3
- ops = torch.ops._paged_attention_7c06fa7
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_paged_attention_7c06fa7::{op_name}"
 
1
  import torch
2
+ from . import _paged_attention_cuda_c6404f1
3
+ ops = torch.ops._paged_attention_cuda_c6404f1
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_paged_attention_cuda_c6404f1::{op_name}"
build/torch29-cxx11-cu126-x86_64-linux/_paged_attention_7c06fa7.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:fd2f92bd86b0ebb2667fb27675bbce9d49e013fefda6fb8c8e0e77b2b874427e
3
- size 140143480
 
 
 
 
build/torch29-cxx11-cu126-x86_64-linux/_paged_attention_cuda_c6404f1.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89325f1b97efcf0d22e55ee4a68bb7590296190384a5af34702b355566b9140c
3
+ size 140143504
build/torch29-cxx11-cu126-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,18 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "cuda",
7
+ "archs": [
8
+ "7.0",
9
+ "7.2",
10
+ "7.5",
11
+ "8.0",
12
+ "8.6",
13
+ "8.7",
14
+ "8.9",
15
+ "9.0+PTX"
16
+ ]
17
+ }
18
+ }
build/torch29-cxx11-cu128-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _paged_attention_7c06fa7
3
- ops = torch.ops._paged_attention_7c06fa7
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_paged_attention_7c06fa7::{op_name}"
 
1
  import torch
2
+ from . import _paged_attention_cuda_c6404f1
3
+ ops = torch.ops._paged_attention_cuda_c6404f1
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_paged_attention_cuda_c6404f1::{op_name}"
build/torch29-cxx11-cu128-x86_64-linux/_paged_attention_7c06fa7.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a5fefc761e1c29ba770f5af0b95b1749de8184dbc46b39e5b66c52971ef9f47f
3
- size 167710744
 
 
 
 
build/torch29-cxx11-cu128-x86_64-linux/_paged_attention_cuda_c6404f1.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb574b6d123b5a419063568d05231969f4a70a5fb2d4903c9d1df60c8d5fc9a2
3
+ size 167710768
build/torch29-cxx11-cu128-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,21 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "cuda",
7
+ "archs": [
8
+ "10.0",
9
+ "10.1",
10
+ "12.0+PTX",
11
+ "7.0",
12
+ "7.2",
13
+ "7.5",
14
+ "8.0",
15
+ "8.6",
16
+ "8.7",
17
+ "8.9",
18
+ "9.0"
19
+ ]
20
+ }
21
+ }
build/torch29-cxx11-cu130-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _paged_attention_7c06fa7
3
- ops = torch.ops._paged_attention_7c06fa7
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_paged_attention_7c06fa7::{op_name}"
 
1
  import torch
2
+ from . import _paged_attention_cuda_c6404f1
3
+ ops = torch.ops._paged_attention_cuda_c6404f1
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_paged_attention_cuda_c6404f1::{op_name}"
build/torch29-cxx11-cu130-x86_64-linux/_paged_attention_cuda_c6404f1.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8b109c24c9a2f74e945b6d5bd8634b53b791fbc3ad60bbe7304e2f52097f51e
3
+ size 86544216
build/torch29-cxx11-cu130-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,19 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "cuda",
7
+ "archs": [
8
+ "10.0",
9
+ "11.0",
10
+ "12.0+PTX",
11
+ "7.5",
12
+ "8.0",
13
+ "8.6",
14
+ "8.7",
15
+ "8.9",
16
+ "9.0"
17
+ ]
18
+ }
19
+ }
build/torch29-cxx11-rocm63-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _paged_attention_7c06fa7
3
- ops = torch.ops._paged_attention_7c06fa7
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_paged_attention_7c06fa7::{op_name}"
 
1
  import torch
2
+ from . import _paged_attention_rocm_c6404f1
3
+ ops = torch.ops._paged_attention_rocm_c6404f1
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_paged_attention_rocm_c6404f1::{op_name}"
build/torch29-cxx11-rocm63-x86_64-linux/_paged_attention_7c06fa7.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d4980d89cf19812b85f0fb06a9f7cbf2179831b3d0644baea8f281a380f054e9
3
- size 120196920
 
 
 
 
build/torch29-cxx11-rocm63-x86_64-linux/_paged_attention_rocm_c6404f1.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a74db0000a94f64fff130180677a0b3284058226460b58b15ff0c619919d133
3
+ size 58641744
build/torch29-cxx11-rocm63-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,17 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "rocm",
7
+ "archs": [
8
+ "gfx1030",
9
+ "gfx1100",
10
+ "gfx1101",
11
+ "gfx906",
12
+ "gfx908",
13
+ "gfx90a",
14
+ "gfx942"
15
+ ]
16
+ }
17
+ }
build/torch29-cxx11-rocm64-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _paged_attention_7c06fa7
3
- ops = torch.ops._paged_attention_7c06fa7
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_paged_attention_7c06fa7::{op_name}"
 
1
  import torch
2
+ from . import _paged_attention_rocm_c6404f1
3
+ ops = torch.ops._paged_attention_rocm_c6404f1
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_paged_attention_rocm_c6404f1::{op_name}"
build/torch29-cxx11-rocm64-x86_64-linux/_paged_attention_7c06fa7.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:dbed4ca210df74630817bb16af3051bf56da9707421a3b5f23e70b3e21f314a4
3
- size 121030520
 
 
 
 
build/torch29-cxx11-rocm64-x86_64-linux/_paged_attention_rocm_c6404f1.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac5e2466c6cd858996e10bce9e0abce1d5e935cb82158cb32474267e06f323df
3
+ size 57972360
build/torch29-cxx11-rocm64-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,17 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "rocm",
7
+ "archs": [
8
+ "gfx1030",
9
+ "gfx1100",
10
+ "gfx1101",
11
+ "gfx906",
12
+ "gfx908",
13
+ "gfx90a",
14
+ "gfx942"
15
+ ]
16
+ }
17
+ }