diff --git a/build/torch210-cxx11-cu126-x86_64-linux/__init__.py b/build/torch210-cxx11-cu126-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu126-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch210-cxx11-cu126-x86_64-linux/_ops.py b/build/torch210-cxx11-cu126-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu126-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch210-cxx11-cu126-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch210-cxx11-cu126-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index 4938cf01d4f35abd8c4bde198fbe9c090b5b7be2..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu126-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:4331a70e73fe4493c4a77e411ceb88fbdf91fe22988cf387ba888457bc95428a -size 2019872 diff --git a/build/torch210-cxx11-cu126-x86_64-linux/metadata.json b/build/torch210-cxx11-cu126-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu126-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch210-cxx11-cu126-x86_64-linux/relu/__init__.py b/build/torch210-cxx11-cu126-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu126-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py"))) diff --git a/build/torch210-cxx11-cu128-x86_64-linux/__init__.py b/build/torch210-cxx11-cu128-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu128-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch210-cxx11-cu128-x86_64-linux/_ops.py b/build/torch210-cxx11-cu128-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu128-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch210-cxx11-cu128-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch210-cxx11-cu128-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index bcfb9ba6a82fcc681bcf1933fff7af9d8a78cd15..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu128-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:fb9d4b04679a723ed97e12592f1c9977d3e858d896c527b0e99302de5dcbf00d -size 2111824 diff --git a/build/torch210-cxx11-cu128-x86_64-linux/metadata.json b/build/torch210-cxx11-cu128-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu128-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch210-cxx11-cu128-x86_64-linux/relu/__init__.py b/build/torch210-cxx11-cu128-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu128-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py"))) diff --git a/build/torch210-cxx11-cu130-x86_64-linux/__init__.py b/build/torch210-cxx11-cu130-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu130-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch210-cxx11-cu130-x86_64-linux/_ops.py b/build/torch210-cxx11-cu130-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu130-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch210-cxx11-cu130-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch210-cxx11-cu130-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index 847cdd8ff6e3276bb336aa32756de59a80d35943..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu130-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:2983b028177a3807d423750a37ca685c7ebfc8100b5f6598df69667afe8f35ec -size 2129640 diff --git a/build/torch210-cxx11-cu130-x86_64-linux/metadata.json b/build/torch210-cxx11-cu130-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu130-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch210-cxx11-cu130-x86_64-linux/relu/__init__.py b/build/torch210-cxx11-cu130-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu130-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py"))) diff --git a/build/torch210-cxx11-rocm70-x86_64-linux/__init__.py b/build/torch210-cxx11-rocm70-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-rocm70-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch210-cxx11-rocm70-x86_64-linux/_ops.py b/build/torch210-cxx11-rocm70-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-rocm70-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch210-cxx11-rocm70-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch210-cxx11-rocm70-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index 945f4c9656a7d805beb984ae035885c080e38eb3..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-rocm70-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:d0ee62fbd93bbf70a5239bd94669ec10c5466bd871d225dd42d89f68055335c1 -size 1990320 diff --git a/build/torch210-cxx11-rocm70-x86_64-linux/metadata.json b/build/torch210-cxx11-rocm70-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-rocm70-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch210-cxx11-rocm70-x86_64-linux/relu/__init__.py b/build/torch210-cxx11-rocm70-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-rocm70-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py"))) diff --git a/build/torch210-cxx11-rocm71-x86_64-linux/__init__.py b/build/torch210-cxx11-rocm71-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-rocm71-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch210-cxx11-rocm71-x86_64-linux/_ops.py b/build/torch210-cxx11-rocm71-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-rocm71-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch210-cxx11-rocm71-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch210-cxx11-rocm71-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index a6e6a213854673b3049c7ce78dd537f41ced4591..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-rocm71-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:87f0d86aeedffc637b070645822e6688be5744263c5fda3d0f92c4a827d8a37b -size 1990224 diff --git a/build/torch210-cxx11-rocm71-x86_64-linux/metadata.json b/build/torch210-cxx11-rocm71-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-rocm71-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch210-cxx11-rocm71-x86_64-linux/relu/__init__.py b/build/torch210-cxx11-rocm71-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-rocm71-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py"))) diff --git a/build/torch210-cxx11-xpu20253-x86_64-linux/__init__.py b/build/torch210-cxx11-xpu20253-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-xpu20253-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch210-cxx11-xpu20253-x86_64-linux/_ops.py b/build/torch210-cxx11-xpu20253-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-xpu20253-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch210-cxx11-xpu20253-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch210-cxx11-xpu20253-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index 34acfa59082fae91f396406494aee92d9cb55d70..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-xpu20253-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:3b8050743b1530ad911c74eef35ec5ffb76c2dda91790c56e083635d49aca10e -size 2050440 diff --git a/build/torch210-cxx11-xpu20253-x86_64-linux/metadata.json b/build/torch210-cxx11-xpu20253-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-xpu20253-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch210-cxx11-xpu20253-x86_64-linux/relu/__init__.py b/build/torch210-cxx11-xpu20253-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-xpu20253-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py"))) diff --git a/build/torch28-cxx11-cu126-x86_64-linux/__init__.py b/build/torch28-cxx11-cu126-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch28-cxx11-cu126-x86_64-linux/_ops.py b/build/torch28-cxx11-cu126-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch28-cxx11-cu126-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch28-cxx11-cu126-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index 7e95ed652184da6136e07c58e249912e9ea65c43..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:7ce61ae869ce8392e5ffcbf234b073368eed20808fedcf7236fa10690133d5a2 -size 2013928 diff --git a/build/torch28-cxx11-cu126-x86_64-linux/metadata.json b/build/torch28-cxx11-cu126-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch28-cxx11-cu126-x86_64-linux/relu/__init__.py b/build/torch28-cxx11-cu126-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py"))) diff --git a/build/torch28-cxx11-cu128-x86_64-linux/__init__.py b/build/torch28-cxx11-cu128-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch28-cxx11-cu128-x86_64-linux/_ops.py b/build/torch28-cxx11-cu128-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch28-cxx11-cu128-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch28-cxx11-cu128-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index cbfafde4dd46a2085f80e351ab2c5deba50b338c..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:f53b59d3863275f40114ae9d415e8dfbd1af68ea11bd924ed60088e76467a645 -size 2101656 diff --git a/build/torch28-cxx11-cu128-x86_64-linux/metadata.json b/build/torch28-cxx11-cu128-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch28-cxx11-cu128-x86_64-linux/relu/__init__.py b/build/torch28-cxx11-cu128-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py"))) diff --git a/build/torch28-cxx11-cu129-x86_64-linux/__init__.py b/build/torch28-cxx11-cu129-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu129-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch28-cxx11-cu129-x86_64-linux/_ops.py b/build/torch28-cxx11-cu129-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu129-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch28-cxx11-cu129-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch28-cxx11-cu129-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index c7da8370f256f34e08475735d75059eee92eee55..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu129-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c4cb02e5b63c1236dbbfc491dc79ac88da1172389f8402f23cbe76e76f1dd0a0 -size 2135440 diff --git a/build/torch28-cxx11-cu129-x86_64-linux/metadata.json b/build/torch28-cxx11-cu129-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu129-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch28-cxx11-cu129-x86_64-linux/relu/__init__.py b/build/torch28-cxx11-cu129-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu129-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py"))) diff --git a/build/torch28-cxx11-rocm63-x86_64-linux/__init__.py b/build/torch28-cxx11-rocm63-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-rocm63-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch28-cxx11-rocm63-x86_64-linux/_ops.py b/build/torch28-cxx11-rocm63-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-rocm63-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch28-cxx11-rocm63-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch28-cxx11-rocm63-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index f6749cbda5edabb1dd8e590103f8aa5003722327..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-rocm63-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:fc1207b128f1d74a74c0dc454c74c20894a2b80568b6abec71b12effc660f95f -size 1970264 diff --git a/build/torch28-cxx11-rocm63-x86_64-linux/metadata.json b/build/torch28-cxx11-rocm63-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-rocm63-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch28-cxx11-rocm63-x86_64-linux/relu/__init__.py b/build/torch28-cxx11-rocm63-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-rocm63-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py"))) diff --git a/build/torch28-cxx11-rocm64-x86_64-linux/__init__.py b/build/torch28-cxx11-rocm64-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-rocm64-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch28-cxx11-rocm64-x86_64-linux/_ops.py b/build/torch28-cxx11-rocm64-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-rocm64-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch28-cxx11-rocm64-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch28-cxx11-rocm64-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index 36567e023c3721b4eab8f9cd52aba40ffa804952..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-rocm64-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:4ac4301301b2a1bd30317797d6deb03c3797b822f377e320192263c9297f8b87 -size 1984936 diff --git a/build/torch28-cxx11-rocm64-x86_64-linux/metadata.json b/build/torch28-cxx11-rocm64-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-rocm64-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch28-cxx11-rocm64-x86_64-linux/relu/__init__.py b/build/torch28-cxx11-rocm64-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-rocm64-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py"))) diff --git a/build/torch28-cxx11-xpu20251-x86_64-linux/__init__.py b/build/torch28-cxx11-xpu20251-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-xpu20251-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch28-cxx11-xpu20251-x86_64-linux/_ops.py b/build/torch28-cxx11-xpu20251-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-xpu20251-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch28-cxx11-xpu20251-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch28-cxx11-xpu20251-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index df2e506c73da9cd0a181049172978e831f655cc2..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-xpu20251-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:580355a61256f6977fb85ee1f2ba6d1cd9d1a02a76eec4d325fd33381c51c8f9 -size 2049544 diff --git a/build/torch28-cxx11-xpu20251-x86_64-linux/metadata.json b/build/torch28-cxx11-xpu20251-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-xpu20251-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch28-cxx11-xpu20251-x86_64-linux/relu/__init__.py b/build/torch28-cxx11-xpu20251-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-xpu20251-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py"))) diff --git a/build/torch29-cxx11-cu126-x86_64-linux/__init__.py b/build/torch29-cxx11-cu126-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu126-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch29-cxx11-cu126-x86_64-linux/_ops.py b/build/torch29-cxx11-cu126-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu126-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch29-cxx11-cu126-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch29-cxx11-cu126-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index 333e3b6b2c884c739e0fe10891ab357800342059..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu126-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:adaa09fda0b8c00fe8032f58106f01e6599952d91abb7ab5c5cb1da085dd5cef -size 2014176 diff --git a/build/torch29-cxx11-cu126-x86_64-linux/metadata.json b/build/torch29-cxx11-cu126-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu126-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch29-cxx11-cu126-x86_64-linux/relu/__init__.py b/build/torch29-cxx11-cu126-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu126-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py"))) diff --git a/build/torch29-cxx11-cu128-x86_64-linux/__init__.py b/build/torch29-cxx11-cu128-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu128-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch29-cxx11-cu128-x86_64-linux/_ops.py b/build/torch29-cxx11-cu128-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu128-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch29-cxx11-cu128-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch29-cxx11-cu128-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index e4a1345aa29e88719fba1103c638f4b37bbeedc3..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu128-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:538cbc98afbcdd872169260f5073f18063dcf88c44b1ec87759a276b9849a0b3 -size 2101904 diff --git a/build/torch29-cxx11-cu128-x86_64-linux/metadata.json b/build/torch29-cxx11-cu128-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu128-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch29-cxx11-cu128-x86_64-linux/relu/__init__.py b/build/torch29-cxx11-cu128-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu128-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py"))) diff --git a/build/torch29-cxx11-cu130-x86_64-linux/__init__.py b/build/torch29-cxx11-cu130-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu130-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch29-cxx11-cu130-x86_64-linux/_ops.py b/build/torch29-cxx11-cu130-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu130-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch29-cxx11-cu130-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch29-cxx11-cu130-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index 4cad79fe6bc25240df9110d04a6908dff747bd0c..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu130-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:351b103bddc869b606a6d6dea56c4bb828177340c6288dcdda77beba6fc261f8 -size 2127912 diff --git a/build/torch29-cxx11-cu130-x86_64-linux/metadata.json b/build/torch29-cxx11-cu130-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu130-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch29-cxx11-cu130-x86_64-linux/relu/__init__.py b/build/torch29-cxx11-cu130-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu130-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py"))) diff --git a/build/torch29-cxx11-rocm63-x86_64-linux/__init__.py b/build/torch29-cxx11-rocm63-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-rocm63-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch29-cxx11-rocm63-x86_64-linux/_ops.py b/build/torch29-cxx11-rocm63-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-rocm63-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch29-cxx11-rocm63-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch29-cxx11-rocm63-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index 2e5c2f0c83e449a1386cba6b1b6be59817369938..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-rocm63-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:e12e09f5304da27ac819c109dc8d82ac2a6dfac5635ba5c845d53238668380a4 -size 1970448 diff --git a/build/torch29-cxx11-rocm63-x86_64-linux/metadata.json b/build/torch29-cxx11-rocm63-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-rocm63-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch29-cxx11-rocm63-x86_64-linux/relu/__init__.py b/build/torch29-cxx11-rocm63-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-rocm63-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py"))) diff --git a/build/torch29-cxx11-rocm64-x86_64-linux/__init__.py b/build/torch29-cxx11-rocm64-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-rocm64-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch29-cxx11-rocm64-x86_64-linux/_ops.py b/build/torch29-cxx11-rocm64-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-rocm64-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch29-cxx11-rocm64-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch29-cxx11-rocm64-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index 10a0c479c6265385691edf70b33dc969f29e5191..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-rocm64-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:2e66a1c949ead2bc606247f97b2f2d4ba9bedcf533984ad8a79a3cab2e2c620c -size 1989224 diff --git a/build/torch29-cxx11-rocm64-x86_64-linux/metadata.json b/build/torch29-cxx11-rocm64-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-rocm64-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch29-cxx11-rocm64-x86_64-linux/relu/__init__.py b/build/torch29-cxx11-rocm64-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-rocm64-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py"))) diff --git a/build/torch29-cxx11-xpu20252-x86_64-linux/__init__.py b/build/torch29-cxx11-xpu20252-x86_64-linux/__init__.py deleted file mode 100644 index 8050dfd765d1a50720e99d17870c2c854d9e2574..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-xpu20252-x86_64-linux/__init__.py +++ /dev/null @@ -1,12 +0,0 @@ -from typing import Optional - -import torch - -from ._ops import ops - - -def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: - if out is None: - out = torch.empty_like(x) - ops.relu(out, x) - return out \ No newline at end of file diff --git a/build/torch29-cxx11-xpu20252-x86_64-linux/_ops.py b/build/torch29-cxx11-xpu20252-x86_64-linux/_ops.py deleted file mode 100644 index 2e5edf901b80e2f17196d66460df1c077e0491ac..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-xpu20252-x86_64-linux/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _relu_9f1f069 -ops = torch.ops._relu_9f1f069 - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_relu_9f1f069::{op_name}" \ No newline at end of file diff --git a/build/torch29-cxx11-xpu20252-x86_64-linux/_relu_9f1f069.abi3.so b/build/torch29-cxx11-xpu20252-x86_64-linux/_relu_9f1f069.abi3.so deleted file mode 100644 index 30edb2baf8aa39ce3c36084398571f3a8f15faba..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-xpu20252-x86_64-linux/_relu_9f1f069.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:61e2960692c64b89d41bdc120c6738c248b50b06a19a388135852bbe862a18bb -size 2048320 diff --git a/build/torch29-cxx11-xpu20252-x86_64-linux/metadata.json b/build/torch29-cxx11-xpu20252-x86_64-linux/metadata.json deleted file mode 100644 index 76bafa5f33b6818aa6bb4cab04be811b87519b44..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-xpu20252-x86_64-linux/metadata.json +++ /dev/null @@ -1 +0,0 @@ -{"python-depends":[]} \ No newline at end of file diff --git a/build/torch29-cxx11-xpu20252-x86_64-linux/relu/__init__.py b/build/torch29-cxx11-xpu20252-x86_64-linux/relu/__init__.py deleted file mode 100644 index 03dbc1afe1cf156661a2b1b22003cd5f599a0309..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-xpu20252-x86_64-linux/relu/__init__.py +++ /dev/null @@ -1,26 +0,0 @@ -import ctypes -import sys - -import importlib -from pathlib import Path -from types import ModuleType - -def _import_from_path(file_path: Path) -> ModuleType: - # We cannot use the module name as-is, after adding it to `sys.modules`, - # it would also be used for other imports. So, we make a module name that - # depends on the path for it to be unique using the hex-encoded hash of - # the path. - path_hash = "{:x}".format(ctypes.c_size_t(hash(file_path.absolute())).value) - module_name = path_hash - spec = importlib.util.spec_from_file_location(module_name, file_path) - if spec is None: - raise ImportError(f"Cannot load spec for {module_name} from {file_path}") - module = importlib.util.module_from_spec(spec) - if module is None: - raise ImportError(f"Cannot load module {module_name} from spec") - sys.modules[module_name] = module - spec.loader.exec_module(module) # type: ignore - return module - - -globals().update(vars(_import_from_path(Path(__file__).parent.parent / "__init__.py")))