Kernels
danieldk HF Staff commited on
Commit
48f2c67
Β·
verified Β·
1 Parent(s): 663614f

Build uploaded using `kernels`.

Browse files
This view is limited to 50 files because it contains too many changes. Β  See raw diff
Files changed (50) hide show
  1. build/torch210-cxx11-cu126-x86_64-linux/__init__.py +4 -0
  2. build/{torch28-cxx11-cu126-x86_64-linux/causal_conv1d/_causal_conv1d_90f5a60.abi3.so β†’ torch210-cxx11-cu126-x86_64-linux/_causal_conv1d_1b44a8e.abi3.so} +2 -2
  3. build/{torch28-cxx11-cu126-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu126-x86_64-linux}/_ops.py +3 -3
  4. build/torch210-cxx11-cu126-x86_64-linux/causal_conv1d/__init__.py +26 -0
  5. build/{torch28-cxx11-cu126-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu126-x86_64-linux}/causal_conv1d_interface.py +0 -0
  6. build/{torch28-cxx11-cu126-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu126-x86_64-linux}/causal_conv1d_varlen.py +0 -0
  7. build/{torch28-cxx11-cu126-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu126-x86_64-linux}/cpp_functions.py +0 -0
  8. build/torch210-cxx11-cu126-x86_64-linux/metadata.json +1 -0
  9. build/torch210-cxx11-cu128-x86_64-linux/__init__.py +4 -0
  10. build/torch210-cxx11-cu128-x86_64-linux/_causal_conv1d_1b44a8e.abi3.so +3 -0
  11. build/{torch28-cxx11-cu128-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu128-x86_64-linux}/_ops.py +3 -3
  12. build/torch210-cxx11-cu128-x86_64-linux/causal_conv1d/__init__.py +26 -0
  13. build/{torch28-cxx11-cu128-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu128-x86_64-linux}/causal_conv1d_interface.py +0 -0
  14. build/{torch28-cxx11-cu128-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu128-x86_64-linux}/causal_conv1d_varlen.py +0 -0
  15. build/{torch28-cxx11-cu128-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu128-x86_64-linux}/cpp_functions.py +0 -0
  16. build/torch210-cxx11-cu128-x86_64-linux/metadata.json +1 -0
  17. build/torch210-cxx11-cu130-x86_64-linux/__init__.py +4 -0
  18. build/{torch28-cxx11-cu128-x86_64-linux/causal_conv1d/_causal_conv1d_90f5a60.abi3.so β†’ torch210-cxx11-cu130-x86_64-linux/_causal_conv1d_1b44a8e.abi3.so} +2 -2
  19. build/{torch28-cxx11-cu129-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu130-x86_64-linux}/_ops.py +3 -3
  20. build/torch210-cxx11-cu130-x86_64-linux/causal_conv1d/__init__.py +26 -0
  21. build/{torch28-cxx11-cu129-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu130-x86_64-linux}/causal_conv1d_interface.py +0 -0
  22. build/{torch28-cxx11-cu129-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu130-x86_64-linux}/causal_conv1d_varlen.py +0 -0
  23. build/{torch28-cxx11-cu129-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu130-x86_64-linux}/cpp_functions.py +0 -0
  24. build/torch210-cxx11-cu130-x86_64-linux/metadata.json +1 -0
  25. build/torch28-cxx11-cu126-x86_64-linux/__init__.py +4 -0
  26. build/{torch28-cxx11-cu129-x86_64-linux/causal_conv1d/_causal_conv1d_90f5a60.abi3.so β†’ torch28-cxx11-cu126-x86_64-linux/_causal_conv1d_1b44a8e.abi3.so} +2 -2
  27. build/{torch29-cxx11-cu126-x86_64-linux/causal_conv1d β†’ torch28-cxx11-cu126-x86_64-linux}/_ops.py +3 -3
  28. build/torch28-cxx11-cu126-x86_64-linux/causal_conv1d/__init__.py +25 -3
  29. build/torch28-cxx11-cu126-x86_64-linux/causal_conv1d/__pycache__/__init__.cpython-313.pyc +0 -0
  30. build/torch28-cxx11-cu126-x86_64-linux/causal_conv1d/__pycache__/_ops.cpython-313.pyc +0 -0
  31. build/torch28-cxx11-cu126-x86_64-linux/causal_conv1d/__pycache__/causal_conv1d_interface.cpython-313.pyc +0 -0
  32. build/torch28-cxx11-cu126-x86_64-linux/causal_conv1d/__pycache__/causal_conv1d_varlen.cpython-313.pyc +0 -0
  33. build/torch28-cxx11-cu126-x86_64-linux/causal_conv1d/__pycache__/cpp_functions.cpython-313.pyc +0 -0
  34. build/{torch29-cxx11-cu126-x86_64-linux/causal_conv1d β†’ torch28-cxx11-cu126-x86_64-linux}/causal_conv1d_interface.py +0 -0
  35. build/{torch29-cxx11-cu126-x86_64-linux/causal_conv1d β†’ torch28-cxx11-cu126-x86_64-linux}/causal_conv1d_varlen.py +0 -0
  36. build/{torch29-cxx11-cu126-x86_64-linux/causal_conv1d β†’ torch28-cxx11-cu126-x86_64-linux}/cpp_functions.py +0 -0
  37. build/torch28-cxx11-cu126-x86_64-linux/metadata.json +1 -0
  38. build/torch28-cxx11-cu128-x86_64-linux/__init__.py +4 -0
  39. build/torch28-cxx11-cu128-x86_64-linux/_causal_conv1d_1b44a8e.abi3.so +3 -0
  40. build/torch28-cxx11-cu128-x86_64-linux/_ops.py +9 -0
  41. build/torch28-cxx11-cu128-x86_64-linux/causal_conv1d/__init__.py +25 -3
  42. build/torch28-cxx11-cu128-x86_64-linux/causal_conv1d/__pycache__/__init__.cpython-313.pyc +0 -0
  43. build/torch28-cxx11-cu128-x86_64-linux/causal_conv1d/__pycache__/_ops.cpython-313.pyc +0 -0
  44. build/torch28-cxx11-cu128-x86_64-linux/causal_conv1d/__pycache__/causal_conv1d_interface.cpython-313.pyc +0 -0
  45. build/torch28-cxx11-cu128-x86_64-linux/causal_conv1d/__pycache__/causal_conv1d_varlen.cpython-313.pyc +0 -0
  46. build/torch28-cxx11-cu128-x86_64-linux/causal_conv1d/__pycache__/cpp_functions.cpython-313.pyc +0 -0
  47. build/{torch29-cxx11-cu128-x86_64-linux/causal_conv1d β†’ torch28-cxx11-cu128-x86_64-linux}/causal_conv1d_interface.py +0 -0
  48. build/{torch29-cxx11-cu128-x86_64-linux/causal_conv1d β†’ torch28-cxx11-cu128-x86_64-linux}/causal_conv1d_varlen.py +0 -0
  49. build/{torch29-cxx11-cu128-x86_64-linux/causal_conv1d β†’ torch28-cxx11-cu128-x86_64-linux}/cpp_functions.py +0 -0
  50. build/torch28-cxx11-cu128-x86_64-linux/metadata.json +1 -0
build/torch210-cxx11-cu126-x86_64-linux/__init__.py ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ from .causal_conv1d_interface import causal_conv1d_fn, causal_conv1d_update
2
+ from .causal_conv1d_varlen import causal_conv1d_varlen_states
3
+
4
+ __all__ = ["causal_conv1d_fn", "causal_conv1d_update", "causal_conv1d_varlen_states"]
build/{torch28-cxx11-cu126-x86_64-linux/causal_conv1d/_causal_conv1d_90f5a60.abi3.so β†’ torch210-cxx11-cu126-x86_64-linux/_causal_conv1d_1b44a8e.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e0f4f7ddbf2822ad6dc2a89b513a4eb74700cff343ee3728952a8e35616978cf
3
- size 64213792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6eb0fdb8827538d27d0822e22dd968059657aafdd8dca77b99d606e0026ae43b
3
+ size 80694456
build/{torch28-cxx11-cu126-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu126-x86_64-linux}/_ops.py RENAMED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _causal_conv1d_90f5a60
3
- ops = torch.ops._causal_conv1d_90f5a60
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_causal_conv1d_90f5a60::{op_name}"
 
1
  import torch
2
+ from . import _causal_conv1d_1b44a8e
3
+ ops = torch.ops._causal_conv1d_1b44a8e
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_causal_conv1d_1b44a8e::{op_name}"
build/torch210-cxx11-cu126-x86_64-linux/causal_conv1d/__init__.py ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import ctypes
2
+ import sys
3
+
4
+ import importlib
5
+ from pathlib import Path
6
+ from types import ModuleType
7
+
8
+ def _import_from_path(file_path: Path) -> ModuleType:
9
+ # We cannot use the module name as-is, after adding it to `sys.modules`,
10
+ # it would also be used for other imports. So, we make a module name that
11
+ # depends on the path for it to be unique using the hex-encoded hash of
12
+ # the path.
13
+ path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value)
14
+ module_name = path_hash
15
+ spec = importlib.util.spec_from_file_location(module_name, file_path)
16
+ if spec is None:
17
+ raise ImportError(f"Cannot load spec for {module_name} from {file_path}")
18
+ module = importlib.util.module_from_spec(spec)
19
+ if module is None:
20
+ raise ImportError(f"Cannot load module {module_name} from spec")
21
+ sys.modules[module_name] = module
22
+ spec.loader.exec_module(module) # type: ignore
23
+ return module
24
+
25
+
26
+ globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py")))
build/{torch28-cxx11-cu126-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu126-x86_64-linux}/causal_conv1d_interface.py RENAMED
File without changes
build/{torch28-cxx11-cu126-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu126-x86_64-linux}/causal_conv1d_varlen.py RENAMED
File without changes
build/{torch28-cxx11-cu126-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu126-x86_64-linux}/cpp_functions.py RENAMED
File without changes
build/torch210-cxx11-cu126-x86_64-linux/metadata.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"python-depends":[]}
build/torch210-cxx11-cu128-x86_64-linux/__init__.py ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ from .causal_conv1d_interface import causal_conv1d_fn, causal_conv1d_update
2
+ from .causal_conv1d_varlen import causal_conv1d_varlen_states
3
+
4
+ __all__ = ["causal_conv1d_fn", "causal_conv1d_update", "causal_conv1d_varlen_states"]
build/torch210-cxx11-cu128-x86_64-linux/_causal_conv1d_1b44a8e.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78531cef5f05968a528ae8bc7a5a348b2abad1b180ac90142dd7df2491cef608
3
+ size 107169824
build/{torch28-cxx11-cu128-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu128-x86_64-linux}/_ops.py RENAMED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _causal_conv1d_90f5a60
3
- ops = torch.ops._causal_conv1d_90f5a60
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_causal_conv1d_90f5a60::{op_name}"
 
1
  import torch
2
+ from . import _causal_conv1d_1b44a8e
3
+ ops = torch.ops._causal_conv1d_1b44a8e
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_causal_conv1d_1b44a8e::{op_name}"
build/torch210-cxx11-cu128-x86_64-linux/causal_conv1d/__init__.py ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import ctypes
2
+ import sys
3
+
4
+ import importlib
5
+ from pathlib import Path
6
+ from types import ModuleType
7
+
8
+ def _import_from_path(file_path: Path) -> ModuleType:
9
+ # We cannot use the module name as-is, after adding it to `sys.modules`,
10
+ # it would also be used for other imports. So, we make a module name that
11
+ # depends on the path for it to be unique using the hex-encoded hash of
12
+ # the path.
13
+ path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value)
14
+ module_name = path_hash
15
+ spec = importlib.util.spec_from_file_location(module_name, file_path)
16
+ if spec is None:
17
+ raise ImportError(f"Cannot load spec for {module_name} from {file_path}")
18
+ module = importlib.util.module_from_spec(spec)
19
+ if module is None:
20
+ raise ImportError(f"Cannot load module {module_name} from spec")
21
+ sys.modules[module_name] = module
22
+ spec.loader.exec_module(module) # type: ignore
23
+ return module
24
+
25
+
26
+ globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py")))
build/{torch28-cxx11-cu128-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu128-x86_64-linux}/causal_conv1d_interface.py RENAMED
File without changes
build/{torch28-cxx11-cu128-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu128-x86_64-linux}/causal_conv1d_varlen.py RENAMED
File without changes
build/{torch28-cxx11-cu128-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu128-x86_64-linux}/cpp_functions.py RENAMED
File without changes
build/torch210-cxx11-cu128-x86_64-linux/metadata.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"python-depends":[]}
build/torch210-cxx11-cu130-x86_64-linux/__init__.py ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ from .causal_conv1d_interface import causal_conv1d_fn, causal_conv1d_update
2
+ from .causal_conv1d_varlen import causal_conv1d_varlen_states
3
+
4
+ __all__ = ["causal_conv1d_fn", "causal_conv1d_update", "causal_conv1d_varlen_states"]
build/{torch28-cxx11-cu128-x86_64-linux/causal_conv1d/_causal_conv1d_90f5a60.abi3.so β†’ torch210-cxx11-cu130-x86_64-linux/_causal_conv1d_1b44a8e.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:47aabbc543b8f9899f8447973962253cf0201f88c7dff965b2b387bf369fb204
3
- size 90660576
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8746e8c1e94e2022fe638316ba9cf89489d45d0d92047cafe54e554297a2c701
3
+ size 64618464
build/{torch28-cxx11-cu129-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu130-x86_64-linux}/_ops.py RENAMED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _causal_conv1d_90f5a60
3
- ops = torch.ops._causal_conv1d_90f5a60
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_causal_conv1d_90f5a60::{op_name}"
 
1
  import torch
2
+ from . import _causal_conv1d_1b44a8e
3
+ ops = torch.ops._causal_conv1d_1b44a8e
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_causal_conv1d_1b44a8e::{op_name}"
build/torch210-cxx11-cu130-x86_64-linux/causal_conv1d/__init__.py ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import ctypes
2
+ import sys
3
+
4
+ import importlib
5
+ from pathlib import Path
6
+ from types import ModuleType
7
+
8
+ def _import_from_path(file_path: Path) -> ModuleType:
9
+ # We cannot use the module name as-is, after adding it to `sys.modules`,
10
+ # it would also be used for other imports. So, we make a module name that
11
+ # depends on the path for it to be unique using the hex-encoded hash of
12
+ # the path.
13
+ path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value)
14
+ module_name = path_hash
15
+ spec = importlib.util.spec_from_file_location(module_name, file_path)
16
+ if spec is None:
17
+ raise ImportError(f"Cannot load spec for {module_name} from {file_path}")
18
+ module = importlib.util.module_from_spec(spec)
19
+ if module is None:
20
+ raise ImportError(f"Cannot load module {module_name} from spec")
21
+ sys.modules[module_name] = module
22
+ spec.loader.exec_module(module) # type: ignore
23
+ return module
24
+
25
+
26
+ globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py")))
build/{torch28-cxx11-cu129-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu130-x86_64-linux}/causal_conv1d_interface.py RENAMED
File without changes
build/{torch28-cxx11-cu129-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu130-x86_64-linux}/causal_conv1d_varlen.py RENAMED
File without changes
build/{torch28-cxx11-cu129-x86_64-linux/causal_conv1d β†’ torch210-cxx11-cu130-x86_64-linux}/cpp_functions.py RENAMED
File without changes
build/torch210-cxx11-cu130-x86_64-linux/metadata.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"python-depends":[]}
build/torch28-cxx11-cu126-x86_64-linux/__init__.py ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ from .causal_conv1d_interface import causal_conv1d_fn, causal_conv1d_update
2
+ from .causal_conv1d_varlen import causal_conv1d_varlen_states
3
+
4
+ __all__ = ["causal_conv1d_fn", "causal_conv1d_update", "causal_conv1d_varlen_states"]
build/{torch28-cxx11-cu129-x86_64-linux/causal_conv1d/_causal_conv1d_90f5a60.abi3.so β†’ torch28-cxx11-cu126-x86_64-linux/_causal_conv1d_1b44a8e.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd59adbb4d01c1d3b986f6ab63e3873a57cc8f4cae885c7ad83fe7c9df16b395
3
- size 97498136
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49a73bdc1f6d9a32c2e107610f5ba22c2ca054a3efc1237a8291118af3191e7b
3
+ size 80684768
build/{torch29-cxx11-cu126-x86_64-linux/causal_conv1d β†’ torch28-cxx11-cu126-x86_64-linux}/_ops.py RENAMED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _causal_conv1d_90f5a60
3
- ops = torch.ops._causal_conv1d_90f5a60
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_causal_conv1d_90f5a60::{op_name}"
 
1
  import torch
2
+ from . import _causal_conv1d_1b44a8e
3
+ ops = torch.ops._causal_conv1d_1b44a8e
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_causal_conv1d_1b44a8e::{op_name}"
build/torch28-cxx11-cu126-x86_64-linux/causal_conv1d/__init__.py CHANGED
@@ -1,4 +1,26 @@
1
- from .causal_conv1d_interface import causal_conv1d_fn, causal_conv1d_update
2
- from .causal_conv1d_varlen import causal_conv1d_varlen_states
3
 
4
- __all__ = ["causal_conv1d_fn", "causal_conv1d_update", "causal_conv1d_varlen_states"]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import ctypes
2
+ import sys
3
 
4
+ import importlib
5
+ from pathlib import Path
6
+ from types import ModuleType
7
+
8
+ def _import_from_path(file_path: Path) -> ModuleType:
9
+ # We cannot use the module name as-is, after adding it to `sys.modules`,
10
+ # it would also be used for other imports. So, we make a module name that
11
+ # depends on the path for it to be unique using the hex-encoded hash of
12
+ # the path.
13
+ path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value)
14
+ module_name = path_hash
15
+ spec = importlib.util.spec_from_file_location(module_name, file_path)
16
+ if spec is None:
17
+ raise ImportError(f"Cannot load spec for {module_name} from {file_path}")
18
+ module = importlib.util.module_from_spec(spec)
19
+ if module is None:
20
+ raise ImportError(f"Cannot load module {module_name} from spec")
21
+ sys.modules[module_name] = module
22
+ spec.loader.exec_module(module) # type: ignore
23
+ return module
24
+
25
+
26
+ globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py")))
build/torch28-cxx11-cu126-x86_64-linux/causal_conv1d/__pycache__/__init__.cpython-313.pyc DELETED
Binary file (390 Bytes)
 
build/torch28-cxx11-cu126-x86_64-linux/causal_conv1d/__pycache__/_ops.cpython-313.pyc DELETED
Binary file (535 Bytes)
 
build/torch28-cxx11-cu126-x86_64-linux/causal_conv1d/__pycache__/causal_conv1d_interface.cpython-313.pyc DELETED
Binary file (9.94 kB)
 
build/torch28-cxx11-cu126-x86_64-linux/causal_conv1d/__pycache__/causal_conv1d_varlen.cpython-313.pyc DELETED
Binary file (4.92 kB)
 
build/torch28-cxx11-cu126-x86_64-linux/causal_conv1d/__pycache__/cpp_functions.cpython-313.pyc DELETED
Binary file (3.62 kB)
 
build/{torch29-cxx11-cu126-x86_64-linux/causal_conv1d β†’ torch28-cxx11-cu126-x86_64-linux}/causal_conv1d_interface.py RENAMED
File without changes
build/{torch29-cxx11-cu126-x86_64-linux/causal_conv1d β†’ torch28-cxx11-cu126-x86_64-linux}/causal_conv1d_varlen.py RENAMED
File without changes
build/{torch29-cxx11-cu126-x86_64-linux/causal_conv1d β†’ torch28-cxx11-cu126-x86_64-linux}/cpp_functions.py RENAMED
File without changes
build/torch28-cxx11-cu126-x86_64-linux/metadata.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"python-depends":[]}
build/torch28-cxx11-cu128-x86_64-linux/__init__.py ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ from .causal_conv1d_interface import causal_conv1d_fn, causal_conv1d_update
2
+ from .causal_conv1d_varlen import causal_conv1d_varlen_states
3
+
4
+ __all__ = ["causal_conv1d_fn", "causal_conv1d_update", "causal_conv1d_varlen_states"]
build/torch28-cxx11-cu128-x86_64-linux/_causal_conv1d_1b44a8e.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43ea19b486dc11d1eb780e7c1c4944ad27d27713ab41b8824b14add98c5eb645
3
+ size 107168432
build/torch28-cxx11-cu128-x86_64-linux/_ops.py ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ from . import _causal_conv1d_1b44a8e
3
+ ops = torch.ops._causal_conv1d_1b44a8e
4
+
5
+ def add_op_namespace_prefix(op_name: str):
6
+ """
7
+ Prefix op by namespace.
8
+ """
9
+ return f"_causal_conv1d_1b44a8e::{op_name}"
build/torch28-cxx11-cu128-x86_64-linux/causal_conv1d/__init__.py CHANGED
@@ -1,4 +1,26 @@
1
- from .causal_conv1d_interface import causal_conv1d_fn, causal_conv1d_update
2
- from .causal_conv1d_varlen import causal_conv1d_varlen_states
3
 
4
- __all__ = ["causal_conv1d_fn", "causal_conv1d_update", "causal_conv1d_varlen_states"]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import ctypes
2
+ import sys
3
 
4
+ import importlib
5
+ from pathlib import Path
6
+ from types import ModuleType
7
+
8
+ def _import_from_path(file_path: Path) -> ModuleType:
9
+ # We cannot use the module name as-is, after adding it to `sys.modules`,
10
+ # it would also be used for other imports. So, we make a module name that
11
+ # depends on the path for it to be unique using the hex-encoded hash of
12
+ # the path.
13
+ path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value)
14
+ module_name = path_hash
15
+ spec = importlib.util.spec_from_file_location(module_name, file_path)
16
+ if spec is None:
17
+ raise ImportError(f"Cannot load spec for {module_name} from {file_path}")
18
+ module = importlib.util.module_from_spec(spec)
19
+ if module is None:
20
+ raise ImportError(f"Cannot load module {module_name} from spec")
21
+ sys.modules[module_name] = module
22
+ spec.loader.exec_module(module) # type: ignore
23
+ return module
24
+
25
+
26
+ globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py")))
build/torch28-cxx11-cu128-x86_64-linux/causal_conv1d/__pycache__/__init__.cpython-313.pyc DELETED
Binary file (390 Bytes)
 
build/torch28-cxx11-cu128-x86_64-linux/causal_conv1d/__pycache__/_ops.cpython-313.pyc DELETED
Binary file (535 Bytes)
 
build/torch28-cxx11-cu128-x86_64-linux/causal_conv1d/__pycache__/causal_conv1d_interface.cpython-313.pyc DELETED
Binary file (9.94 kB)
 
build/torch28-cxx11-cu128-x86_64-linux/causal_conv1d/__pycache__/causal_conv1d_varlen.cpython-313.pyc DELETED
Binary file (4.92 kB)
 
build/torch28-cxx11-cu128-x86_64-linux/causal_conv1d/__pycache__/cpp_functions.cpython-313.pyc DELETED
Binary file (3.62 kB)
 
build/{torch29-cxx11-cu128-x86_64-linux/causal_conv1d β†’ torch28-cxx11-cu128-x86_64-linux}/causal_conv1d_interface.py RENAMED
File without changes
build/{torch29-cxx11-cu128-x86_64-linux/causal_conv1d β†’ torch28-cxx11-cu128-x86_64-linux}/causal_conv1d_varlen.py RENAMED
File without changes
build/{torch29-cxx11-cu128-x86_64-linux/causal_conv1d β†’ torch28-cxx11-cu128-x86_64-linux}/cpp_functions.py RENAMED
File without changes
build/torch28-cxx11-cu128-x86_64-linux/metadata.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"python-depends":[]}