danieldk HF Staff commited on
Commit
5df262f
·
verified ·
1 Parent(s): 018bb8d

Build uploaded using `kernels`.

Browse files
.gitattributes CHANGED
@@ -244,3 +244,4 @@ build/torch29-cxx11-rocm63-x86_64-linux/_relu_rocm_vox4kgutt6m6i.abi3.so filter=
244
  build/torch29-cxx11-rocm64-x86_64-linux/_relu_rocm_dd3ars5bdez2c.abi3.so filter=lfs diff=lfs merge=lfs -text
245
  build/torch29-cxx11-xpu20252-x86_64-linux/_relu_xpu_wcav6tli7hepe.abi3.so filter=lfs diff=lfs merge=lfs -text
246
  build/torch210-cu128-x86_64-windows/_relu_cuda_9ba4f37.pyd filter=lfs diff=lfs merge=lfs -text
 
 
244
  build/torch29-cxx11-rocm64-x86_64-linux/_relu_rocm_dd3ars5bdez2c.abi3.so filter=lfs diff=lfs merge=lfs -text
245
  build/torch29-cxx11-xpu20252-x86_64-linux/_relu_xpu_wcav6tli7hepe.abi3.so filter=lfs diff=lfs merge=lfs -text
246
  build/torch210-cu128-x86_64-windows/_relu_cuda_9ba4f37.pyd filter=lfs diff=lfs merge=lfs -text
247
+ build/torch210-xpu20253-x86_64-windows/_relu_xpu_9ba4f37.pyd filter=lfs diff=lfs merge=lfs -text
build/torch210-xpu20253-x86_64-windows/__init__.py ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from typing import Optional
2
+
3
+ import torch
4
+
5
+ from ._ops import ops
6
+
7
+
8
+ def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
9
+ if out is None:
10
+ out = torch.empty_like(x)
11
+ ops.relu(out, x)
12
+ return out
build/torch210-xpu20253-x86_64-windows/{relu/_ops.py → _ops.py} RENAMED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _relu_5629432
3
- ops = torch.ops._relu_5629432
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_relu_5629432::{op_name}"
 
1
  import torch
2
+ from . import _relu_xpu_9ba4f37
3
+ ops = torch.ops._relu_xpu_9ba4f37
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_relu_xpu_9ba4f37::{op_name}"
build/torch210-xpu20253-x86_64-windows/{relu/_relu_5629432.pyd → _relu_xpu_9ba4f37.pyd} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8979a8a08bb73d24f072975c05f25c3555aa999576c90873ba2e3a08a2e11e56
3
  size 166400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40c78ca3ff6cde26f1784af71073b6869eeecbc9333eff32b771e80088fdda09
3
  size 166400
build/torch210-xpu20253-x86_64-windows/relu/__init__.py CHANGED
@@ -1,12 +1,26 @@
1
- from typing import Optional
 
2
 
3
- import torch
 
 
4
 
5
- from ._ops import ops
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
 
7
 
8
- def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
9
- if out is None:
10
- out = torch.empty_like(x)
11
- ops.relu(out, x)
12
- return out
 
1
+ import ctypes
2
+ import sys
3
 
4
+ import importlib
5
+ from pathlib import Path
6
+ from types import ModuleType
7
 
8
+ def _import_from_path(file_path: Path) -> ModuleType:
9
+ # We cannot use the module name as-is, after adding it to `sys.modules`,
10
+ # it would also be used for other imports. So, we make a module name that
11
+ # depends on the path for it to be unique using the hex-encoded hash of
12
+ # the path.
13
+ path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value)
14
+ module_name = path_hash
15
+ spec = importlib.util.spec_from_file_location(module_name, file_path)
16
+ if spec is None:
17
+ raise ImportError(f"Cannot load spec for {module_name} from {file_path}")
18
+ module = importlib.util.module_from_spec(spec)
19
+ if module is None:
20
+ raise ImportError(f"Cannot load module {module_name} from spec")
21
+ sys.modules[module_name] = module
22
+ spec.loader.exec_module(module) # type: ignore
23
+ return module
24
 
25
 
26
+ globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py")))