danieldk HF Staff commited on
Commit
f7e26d5
·
verified ·
1 Parent(s): c0b7b5c

Build uploaded using `kernels`.

Browse files
build/torch210-cxx11-cu128-x86_64-linux/{_flash_mla_cuda_n5cwpkebdn7tu.abi3.so → _flash_mla_cuda_09f70ef.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6447a21ee1d2be7d769ed0ad7a6ed8259b3a6ad1b2818be8e1f5a6c0a812deaa
3
- size 3534104
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:159f6907fce13b98e5b9efd95c8c51f14087d5e4014d2b79fea6c9ad4453f372
3
+ size 3534072
build/torch210-cxx11-cu128-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_mla_cuda_n5cwpkebdn7tu
3
- ops = torch.ops._flash_mla_cuda_n5cwpkebdn7tu
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_mla_cuda_n5cwpkebdn7tu::{op_name}"
 
1
  import torch
2
+ from . import _flash_mla_cuda_09f70ef
3
+ ops = torch.ops._flash_mla_cuda_09f70ef
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_mla_cuda_09f70ef::{op_name}"
build/torch210-cxx11-cu128-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,11 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "MIT",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "cuda",
7
+ "archs": [
8
+ "9.0a"
9
+ ]
10
+ }
11
+ }
build/torch210-cxx11-cu130-x86_64-linux/{_flash_mla_cuda_jeu3ybctgatdg.abi3.so → _flash_mla_cuda_09f70ef.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f45c42c89859a7011ace4c0d38f54ff2381d2f1c6a1dc12ab19a556e1936b1f
3
- size 9395248
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1f914866fdfdcf91845a2d11c0218309b99020025fd35ed547230ae8e1ba519
3
+ size 9395216
build/torch210-cxx11-cu130-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_mla_cuda_jeu3ybctgatdg
3
- ops = torch.ops._flash_mla_cuda_jeu3ybctgatdg
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_mla_cuda_jeu3ybctgatdg::{op_name}"
 
1
  import torch
2
+ from . import _flash_mla_cuda_09f70ef
3
+ ops = torch.ops._flash_mla_cuda_09f70ef
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_mla_cuda_09f70ef::{op_name}"
build/torch210-cxx11-cu130-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,12 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "MIT",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "cuda",
7
+ "archs": [
8
+ "10.0f",
9
+ "9.0a"
10
+ ]
11
+ }
12
+ }
build/torch29-cxx11-cu128-x86_64-linux/{_flash_mla_cuda_sq54danr2ngz4.abi3.so → _flash_mla_cuda_09f70ef.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c9af57ce65dbf293c052de51d2b5f601dfd2fb667f350f1608cbb1ba8c006fd
3
- size 3523120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22efc944dccf34d020a87fe713033bfae026d91eda6ecc86a4d491abe38edc51
3
+ size 3523096
build/torch29-cxx11-cu128-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_mla_cuda_sq54danr2ngz4
3
- ops = torch.ops._flash_mla_cuda_sq54danr2ngz4
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_mla_cuda_sq54danr2ngz4::{op_name}"
 
1
  import torch
2
+ from . import _flash_mla_cuda_09f70ef
3
+ ops = torch.ops._flash_mla_cuda_09f70ef
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_mla_cuda_09f70ef::{op_name}"
build/torch29-cxx11-cu128-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,11 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "MIT",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "cuda",
7
+ "archs": [
8
+ "9.0a"
9
+ ]
10
+ }
11
+ }
build/torch29-cxx11-cu130-x86_64-linux/{_flash_mla_cuda_ytyeqvwwmlcyi.abi3.so → _flash_mla_cuda_09f70ef.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6289152d635de44ae70af9c051ff64ba8ead54dfd03f6d34ba4cb1c09b28e362
3
- size 9380408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:243747242a38da9e2034590bfd52c3f883683b4faec3ccfd6ec76a0f97addf43
3
+ size 9380384
build/torch29-cxx11-cu130-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_mla_cuda_ytyeqvwwmlcyi
3
- ops = torch.ops._flash_mla_cuda_ytyeqvwwmlcyi
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_mla_cuda_ytyeqvwwmlcyi::{op_name}"
 
1
  import torch
2
+ from . import _flash_mla_cuda_09f70ef
3
+ ops = torch.ops._flash_mla_cuda_09f70ef
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_mla_cuda_09f70ef::{op_name}"
build/torch29-cxx11-cu130-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,12 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "MIT",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "cuda",
7
+ "archs": [
8
+ "10.0f",
9
+ "9.0a"
10
+ ]
11
+ }
12
+ }