|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
""" |
|
|
Utilities and types for defining networks, these depend on PyTorch. |
|
|
""" |
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
import io |
|
|
import re |
|
|
import warnings |
|
|
from collections import OrderedDict |
|
|
from collections.abc import Callable, Mapping, Sequence |
|
|
from contextlib import contextmanager |
|
|
from copy import deepcopy |
|
|
from typing import Any |
|
|
|
|
|
import numpy as np |
|
|
import torch |
|
|
import torch.nn as nn |
|
|
|
|
|
from monai.apps.utils import get_logger |
|
|
from monai.config import PathLike |
|
|
from monai.utils.misc import ensure_tuple, save_obj, set_determinism |
|
|
from monai.utils.module import look_up_option, optional_import, pytorch_after |
|
|
from monai.utils.type_conversion import convert_to_dst_type, convert_to_tensor |
|
|
|
|
|
onnx, _ = optional_import("onnx") |
|
|
onnxreference, _ = optional_import("onnx.reference") |
|
|
onnxruntime, _ = optional_import("onnxruntime") |
|
|
|
|
|
__all__ = [ |
|
|
"one_hot", |
|
|
"predict_segmentation", |
|
|
"normalize_transform", |
|
|
"to_norm_affine", |
|
|
"normal_init", |
|
|
"icnr_init", |
|
|
"pixelshuffle", |
|
|
"eval_mode", |
|
|
"train_mode", |
|
|
"get_state_dict", |
|
|
"copy_model_state", |
|
|
"save_state", |
|
|
"convert_to_onnx", |
|
|
"convert_to_torchscript", |
|
|
"convert_to_trt", |
|
|
"meshgrid_ij", |
|
|
"meshgrid_xy", |
|
|
"replace_modules", |
|
|
"replace_modules_temp", |
|
|
"look_up_named_module", |
|
|
"set_named_module", |
|
|
"has_nvfuser_instance_norm", |
|
|
] |
|
|
|
|
|
logger = get_logger(module_name=__name__) |
|
|
|
|
|
_has_nvfuser = None |
|
|
|
|
|
|
|
|
def has_nvfuser_instance_norm(): |
|
|
"""whether the current environment has InstanceNorm3dNVFuser |
|
|
https://github.com/NVIDIA/apex/blob/23.05-devel/apex/normalization/instance_norm.py#L15-L16 |
|
|
""" |
|
|
global _has_nvfuser |
|
|
if _has_nvfuser is not None: |
|
|
return _has_nvfuser |
|
|
|
|
|
_, _has_nvfuser = optional_import("apex.normalization", name="InstanceNorm3dNVFuser") |
|
|
if not _has_nvfuser: |
|
|
return False |
|
|
try: |
|
|
import importlib |
|
|
|
|
|
importlib.import_module("instance_norm_nvfuser_cuda") |
|
|
except ImportError: |
|
|
_has_nvfuser = False |
|
|
return _has_nvfuser |
|
|
|
|
|
|
|
|
def look_up_named_module(name: str, mod, print_all_options=False): |
|
|
""" |
|
|
get the named module in `mod` by the attribute name, |
|
|
for example ``look_up_named_module(net, "features.3.1.attn")`` |
|
|
|
|
|
Args: |
|
|
name: a string representing the module attribute. |
|
|
mod: a pytorch module to be searched (in ``mod.named_modules()``). |
|
|
print_all_options: whether to print all named modules when `name` is not found in `mod`. Defaults to False. |
|
|
|
|
|
Returns: |
|
|
the corresponding pytorch module's subcomponent such as ``net.features[3][1].attn`` |
|
|
""" |
|
|
name_str = look_up_option( |
|
|
name, {n[0] for n in mod.named_modules()}, default=None, print_all_options=print_all_options |
|
|
) |
|
|
if name_str is None: |
|
|
return None |
|
|
if name_str == "": |
|
|
return mod |
|
|
for n in name_str.split("."): |
|
|
if n.isdigit(): |
|
|
mod = mod[int(n)] |
|
|
else: |
|
|
n = look_up_option(n, {item[0] for item in mod.named_modules()}, default=None, print_all_options=False) |
|
|
if n is None: |
|
|
return None |
|
|
mod = getattr(mod, n) |
|
|
return mod |
|
|
|
|
|
|
|
|
def set_named_module(mod, name: str, new_layer): |
|
|
""" |
|
|
look up `name` in `mod` and replace the layer with `new_layer`, return the updated `mod`. |
|
|
|
|
|
Args: |
|
|
mod: a pytorch module to be updated. |
|
|
name: a string representing the target module attribute. |
|
|
new_layer: a new module replacing the corresponding layer at ``mod.name``. |
|
|
|
|
|
Returns: |
|
|
an updated ``mod`` |
|
|
|
|
|
See also: :py:func:`monai.networks.utils.look_up_named_module`. |
|
|
""" |
|
|
mods_attr = name.rsplit(".", 1) |
|
|
submods, attr = mods_attr if len(mods_attr) == 2 else ("", name) |
|
|
if not attr: |
|
|
return new_layer |
|
|
_mod = look_up_named_module(submods, mod) |
|
|
setattr(_mod, attr, new_layer) |
|
|
return mod |
|
|
|
|
|
|
|
|
def one_hot(labels: torch.Tensor, num_classes: int, dtype: torch.dtype = torch.float, dim: int = 1) -> torch.Tensor: |
|
|
""" |
|
|
For every value v in `labels`, the value in the output will be either 1 or 0. Each vector along the `dim`-th |
|
|
dimension has the "one-hot" format, i.e., it has a total length of `num_classes`, |
|
|
with a one and `num_class-1` zeros. |
|
|
Note that this will include the background label, thus a binary mask should be treated as having two classes. |
|
|
|
|
|
Args: |
|
|
labels: input tensor of integers to be converted into the 'one-hot' format. Internally `labels` will be |
|
|
converted into integers `labels.long()`. |
|
|
num_classes: number of output channels, the corresponding length of `labels[dim]` will be converted to |
|
|
`num_classes` from `1`. |
|
|
dtype: the data type of the output one_hot label. |
|
|
dim: the dimension to be converted to `num_classes` channels from `1` channel, should be non-negative number. |
|
|
|
|
|
Example: |
|
|
|
|
|
For a tensor `labels` of dimensions [B]1[spatial_dims], return a tensor of dimensions `[B]N[spatial_dims]` |
|
|
when `num_classes=N` number of classes and `dim=1`. |
|
|
|
|
|
.. code-block:: python |
|
|
|
|
|
from monai.networks.utils import one_hot |
|
|
import torch |
|
|
|
|
|
a = torch.randint(0, 2, size=(1, 2, 2, 2)) |
|
|
out = one_hot(a, num_classes=2, dim=0) |
|
|
print(out.shape) # torch.Size([2, 2, 2, 2]) |
|
|
|
|
|
a = torch.randint(0, 2, size=(2, 1, 2, 2, 2)) |
|
|
out = one_hot(a, num_classes=2, dim=1) |
|
|
print(out.shape) # torch.Size([2, 2, 2, 2, 2]) |
|
|
|
|
|
""" |
|
|
|
|
|
|
|
|
if labels.ndim < dim + 1: |
|
|
shape = list(labels.shape) + [1] * (dim + 1 - len(labels.shape)) |
|
|
labels = torch.reshape(labels, shape) |
|
|
|
|
|
sh = list(labels.shape) |
|
|
|
|
|
if sh[dim] != 1: |
|
|
raise AssertionError("labels should have a channel with length equal to one.") |
|
|
|
|
|
sh[dim] = num_classes |
|
|
|
|
|
o = torch.zeros(size=sh, dtype=dtype, device=labels.device) |
|
|
labels = o.scatter_(dim=dim, index=labels.long(), value=1) |
|
|
|
|
|
return labels |
|
|
|
|
|
|
|
|
def predict_segmentation(logits: torch.Tensor, mutually_exclusive: bool = False, threshold: float = 0.0) -> Any: |
|
|
""" |
|
|
Given the logits from a network, computing the segmentation by thresholding all values above 0 |
|
|
if multi-labels task, computing the `argmax` along the channel axis if multi-classes task, |
|
|
logits has shape `BCHW[D]`. |
|
|
|
|
|
Args: |
|
|
logits: raw data of model output. |
|
|
mutually_exclusive: if True, `logits` will be converted into a binary matrix using |
|
|
a combination of argmax, which is suitable for multi-classes task. Defaults to False. |
|
|
threshold: thresholding the prediction values if multi-labels task. |
|
|
""" |
|
|
if not mutually_exclusive: |
|
|
return (logits >= threshold).int() |
|
|
if logits.shape[1] == 1: |
|
|
warnings.warn("single channel prediction, `mutually_exclusive=True` ignored, use threshold instead.") |
|
|
return (logits >= threshold).int() |
|
|
return logits.argmax(1, keepdim=True) |
|
|
|
|
|
|
|
|
def normalize_transform( |
|
|
shape, |
|
|
device: torch.device | str | None = None, |
|
|
dtype: torch.dtype | None = None, |
|
|
align_corners: bool = False, |
|
|
zero_centered: bool = False, |
|
|
) -> torch.Tensor: |
|
|
""" |
|
|
Compute an affine matrix according to the input shape. |
|
|
The transform normalizes the homogeneous image coordinates to the |
|
|
range of `[-1, 1]`. Currently the following source coordinates are supported: |
|
|
|
|
|
- `align_corners=False`, `zero_centered=False`, normalizing from ``[-0.5, d-0.5]``. |
|
|
- `align_corners=True`, `zero_centered=False`, normalizing from ``[0, d-1]``. |
|
|
- `align_corners=False`, `zero_centered=True`, normalizing from ``[-(d-1)/2, (d-1)/2]``. |
|
|
- `align_corners=True`, `zero_centered=True`, normalizing from ``[-d/2, d/2]``. |
|
|
|
|
|
Args: |
|
|
shape: input spatial shape, a sequence of integers. |
|
|
device: device on which the returned affine will be allocated. |
|
|
dtype: data type of the returned affine |
|
|
align_corners: if True, consider -1 and 1 to refer to the centers of the |
|
|
corner pixels rather than the image corners. |
|
|
See also: https://pytorch.org/docs/stable/nn.functional.html#torch.nn.functional.grid_sample |
|
|
zero_centered: whether the coordinates are normalized from a zero-centered range, default to `False`. |
|
|
Setting this flag and `align_corners` will jointly specify the normalization source range. |
|
|
""" |
|
|
shape = convert_to_tensor(shape, torch.float64, device=device, wrap_sequence=True, track_meta=False) |
|
|
norm = shape.clone().detach().to(dtype=torch.float64, device=device) |
|
|
if align_corners: |
|
|
norm[norm <= 1.0] = 2.0 |
|
|
norm = 2.0 / (norm if zero_centered else norm - 1.0) |
|
|
norm = torch.diag(torch.cat((norm, torch.ones((1,), dtype=torch.float64, device=device)))) |
|
|
if not zero_centered: |
|
|
norm[:-1, -1] = -1.0 |
|
|
else: |
|
|
norm[norm <= 0.0] = 2.0 |
|
|
norm = 2.0 / (norm - 1.0 if zero_centered else norm) |
|
|
norm = torch.diag(torch.cat((norm, torch.ones((1,), dtype=torch.float64, device=device)))) |
|
|
if not zero_centered: |
|
|
norm[:-1, -1] = 1.0 / shape - 1.0 |
|
|
norm = norm.unsqueeze(0).to(dtype=dtype) |
|
|
norm.requires_grad = False |
|
|
return norm |
|
|
|
|
|
|
|
|
def to_norm_affine( |
|
|
affine: torch.Tensor, |
|
|
src_size: Sequence[int], |
|
|
dst_size: Sequence[int], |
|
|
align_corners: bool = False, |
|
|
zero_centered: bool = False, |
|
|
) -> torch.Tensor: |
|
|
""" |
|
|
Given ``affine`` defined for coordinates in the pixel space, compute the corresponding affine |
|
|
for the normalized coordinates. |
|
|
|
|
|
Args: |
|
|
affine: Nxdxd batched square matrix |
|
|
src_size: source image spatial shape |
|
|
dst_size: target image spatial shape |
|
|
align_corners: if True, consider -1 and 1 to refer to the centers of the |
|
|
corner pixels rather than the image corners. |
|
|
See also: https://pytorch.org/docs/stable/nn.functional.html#torch.nn.functional.grid_sample |
|
|
zero_centered: whether the coordinates are normalized from a zero-centered range, default to `False`. |
|
|
See also: :py:func:`monai.networks.utils.normalize_transform`. |
|
|
|
|
|
Raises: |
|
|
TypeError: When ``affine`` is not a ``torch.Tensor``. |
|
|
ValueError: When ``affine`` is not Nxdxd. |
|
|
ValueError: When ``src_size`` or ``dst_size`` dimensions differ from ``affine``. |
|
|
|
|
|
""" |
|
|
if not isinstance(affine, torch.Tensor): |
|
|
raise TypeError(f"affine must be a torch.Tensor but is {type(affine).__name__}.") |
|
|
if affine.ndimension() != 3 or affine.shape[1] != affine.shape[2]: |
|
|
raise ValueError(f"affine must be Nxdxd, got {tuple(affine.shape)}.") |
|
|
sr = affine.shape[1] - 1 |
|
|
if sr != len(src_size) or sr != len(dst_size): |
|
|
raise ValueError(f"affine suggests {sr}D, got src={len(src_size)}D, dst={len(dst_size)}D.") |
|
|
|
|
|
src_xform = normalize_transform(src_size, affine.device, affine.dtype, align_corners, zero_centered) |
|
|
dst_xform = normalize_transform(dst_size, "cpu", affine.dtype, align_corners, zero_centered) |
|
|
return src_xform @ affine @ convert_to_dst_type(np.linalg.inv(dst_xform.numpy()), dst=affine)[0] |
|
|
|
|
|
|
|
|
def normal_init( |
|
|
m, std: float = 0.02, normal_func: Callable[[torch.Tensor, float, float], Any] = torch.nn.init.normal_ |
|
|
) -> None: |
|
|
""" |
|
|
Initialize the weight and bias tensors of `m' and its submodules to values from a normal distribution with a |
|
|
stddev of `std'. Weight tensors of convolution and linear modules are initialized with a mean of 0, batch |
|
|
norm modules with a mean of 1. The callable `normal_func', used to assign values, should have the same arguments |
|
|
as its default normal_(). This can be used with `nn.Module.apply` to visit submodules of a network. |
|
|
""" |
|
|
cname = m.__class__.__name__ |
|
|
|
|
|
if getattr(m, "weight", None) is not None and (cname.find("Conv") != -1 or cname.find("Linear") != -1): |
|
|
normal_func(m.weight.data, 0.0, std) |
|
|
if getattr(m, "bias", None) is not None: |
|
|
nn.init.constant_(m.bias.data, 0.0) |
|
|
|
|
|
elif cname.find("BatchNorm") != -1: |
|
|
normal_func(m.weight.data, 1.0, std) |
|
|
nn.init.constant_(m.bias.data, 0) |
|
|
|
|
|
|
|
|
def icnr_init(conv, upsample_factor, init=nn.init.kaiming_normal_): |
|
|
""" |
|
|
ICNR initialization for 2D/3D kernels adapted from Aitken et al.,2017 , "Checkerboard artifact free |
|
|
sub-pixel convolution". |
|
|
""" |
|
|
out_channels, in_channels, *dims = conv.weight.shape |
|
|
scale_factor = upsample_factor ** len(dims) |
|
|
|
|
|
oc2 = int(out_channels / scale_factor) |
|
|
|
|
|
kernel = torch.zeros([oc2, in_channels] + dims) |
|
|
kernel = init(kernel) |
|
|
kernel = kernel.transpose(0, 1) |
|
|
kernel = kernel.reshape(oc2, in_channels, -1) |
|
|
kernel = kernel.repeat(1, 1, scale_factor) |
|
|
kernel = kernel.reshape([in_channels, out_channels] + dims) |
|
|
kernel = kernel.transpose(0, 1) |
|
|
conv.weight.data.copy_(kernel) |
|
|
|
|
|
|
|
|
def pixelshuffle(x: torch.Tensor, spatial_dims: int, scale_factor: int) -> torch.Tensor: |
|
|
""" |
|
|
Apply pixel shuffle to the tensor `x` with spatial dimensions `spatial_dims` and scaling factor `scale_factor`. |
|
|
|
|
|
See: Shi et al., 2016, "Real-Time Single Image and Video Super-Resolution |
|
|
Using a nEfficient Sub-Pixel Convolutional Neural Network." |
|
|
|
|
|
See: Aitken et al., 2017, "Checkerboard artifact free sub-pixel convolution". |
|
|
|
|
|
Args: |
|
|
x: Input tensor |
|
|
spatial_dims: number of spatial dimensions, typically 2 or 3 for 2D or 3D |
|
|
scale_factor: factor to rescale the spatial dimensions by, must be >=1 |
|
|
|
|
|
Returns: |
|
|
Reshuffled version of `x`. |
|
|
|
|
|
Raises: |
|
|
ValueError: When input channels of `x` are not divisible by (scale_factor ** spatial_dims) |
|
|
""" |
|
|
dim, factor = spatial_dims, scale_factor |
|
|
input_size = list(x.size()) |
|
|
batch_size, channels = input_size[:2] |
|
|
scale_divisor = factor**dim |
|
|
|
|
|
if channels % scale_divisor != 0: |
|
|
raise ValueError( |
|
|
f"Number of input channels ({channels}) must be evenly " |
|
|
f"divisible by scale_factor ** dimensions ({factor}**{dim}={scale_divisor})." |
|
|
) |
|
|
|
|
|
org_channels = int(channels // scale_divisor) |
|
|
output_size = [batch_size, org_channels] + [d * factor for d in input_size[2:]] |
|
|
|
|
|
indices = list(range(2, 2 + 2 * dim)) |
|
|
indices = indices[dim:] + indices[:dim] |
|
|
permute_indices = [0, 1] |
|
|
for idx in range(dim): |
|
|
permute_indices.extend(indices[idx::dim]) |
|
|
|
|
|
x = x.reshape([batch_size, org_channels] + [factor] * dim + input_size[2:]) |
|
|
x = x.permute(permute_indices).reshape(output_size) |
|
|
return x |
|
|
|
|
|
|
|
|
@contextmanager |
|
|
def eval_mode(*nets: nn.Module): |
|
|
""" |
|
|
Set network(s) to eval mode and then return to original state at the end. |
|
|
|
|
|
Args: |
|
|
nets: Input network(s) |
|
|
|
|
|
Examples |
|
|
|
|
|
.. code-block:: python |
|
|
|
|
|
t=torch.rand(1,1,16,16) |
|
|
p=torch.nn.Conv2d(1,1,3) |
|
|
print(p.training) # True |
|
|
with eval_mode(p): |
|
|
print(p.training) # False |
|
|
print(p(t).sum().backward()) # will correctly raise an exception as gradients are calculated |
|
|
""" |
|
|
|
|
|
|
|
|
|
|
|
training = [n for n in nets if hasattr(n, "training") and n.training] |
|
|
|
|
|
try: |
|
|
|
|
|
with torch.no_grad(): |
|
|
yield [n.eval() if hasattr(n, "eval") else n for n in nets] |
|
|
finally: |
|
|
|
|
|
for n in training: |
|
|
if hasattr(n, "train"): |
|
|
n.train() |
|
|
|
|
|
|
|
|
@contextmanager |
|
|
def train_mode(*nets: nn.Module): |
|
|
""" |
|
|
Set network(s) to train mode and then return to original state at the end. |
|
|
|
|
|
Args: |
|
|
nets: Input network(s) |
|
|
|
|
|
Examples |
|
|
|
|
|
.. code-block:: python |
|
|
|
|
|
t=torch.rand(1,1,16,16) |
|
|
p=torch.nn.Conv2d(1,1,3) |
|
|
p.eval() |
|
|
print(p.training) # False |
|
|
with train_mode(p): |
|
|
print(p.training) # True |
|
|
print(p(t).sum().backward()) # No exception |
|
|
""" |
|
|
|
|
|
|
|
|
|
|
|
eval_list = [n for n in nets if hasattr(n, "training") and (not n.training)] |
|
|
|
|
|
try: |
|
|
|
|
|
with torch.set_grad_enabled(True): |
|
|
yield [n.train() if hasattr(n, "train") else n for n in nets] |
|
|
finally: |
|
|
|
|
|
for n in eval_list: |
|
|
if hasattr(n, "eval"): |
|
|
n.eval() |
|
|
|
|
|
|
|
|
def get_state_dict(obj: torch.nn.Module | Mapping): |
|
|
""" |
|
|
Get the state dict of input object if has `state_dict`, otherwise, return object directly. |
|
|
For data parallel model, automatically convert it to regular model first. |
|
|
|
|
|
Args: |
|
|
obj: input object to check and get the state_dict. |
|
|
|
|
|
""" |
|
|
if isinstance(obj, (nn.DataParallel, nn.parallel.DistributedDataParallel)): |
|
|
obj = obj.module |
|
|
return obj.state_dict() if hasattr(obj, "state_dict") else obj |
|
|
|
|
|
|
|
|
def copy_model_state( |
|
|
dst: torch.nn.Module | Mapping, |
|
|
src: torch.nn.Module | Mapping, |
|
|
dst_prefix="", |
|
|
mapping=None, |
|
|
exclude_vars=None, |
|
|
inplace=True, |
|
|
filter_func=None, |
|
|
): |
|
|
""" |
|
|
Compute a module state_dict, of which the keys are the same as `dst`. The values of `dst` are overwritten |
|
|
by the ones from `src` whenever their keys match. The method provides additional `dst_prefix` for |
|
|
the `dst` key when matching them. `mapping` can be a `{"src_key": "dst_key"}` dict, indicating |
|
|
`dst[dst_prefix + dst_key] = src[src_key]`. |
|
|
This function is mainly to return a model state dict |
|
|
for loading the `src` model state into the `dst` model, `src` and `dst` can have different dict keys, but |
|
|
their corresponding values normally have the same shape. |
|
|
|
|
|
Args: |
|
|
dst: a pytorch module or state dict to be updated. |
|
|
src: a pytorch module or state dict used to get the values used for the update. |
|
|
dst_prefix: `dst` key prefix, so that `dst[dst_prefix + src_key]` |
|
|
will be assigned to the value of `src[src_key]`. |
|
|
mapping: a `{"src_key": "dst_key"}` dict, indicating that `dst[dst_prefix + dst_key]` |
|
|
to be assigned to the value of `src[src_key]`. |
|
|
exclude_vars: a regular expression to match the `dst` variable names, |
|
|
so that their values are not overwritten by `src`. |
|
|
inplace: whether to set the `dst` module with the updated `state_dict` via `load_state_dict`. |
|
|
This option is only available when `dst` is a `torch.nn.Module`. |
|
|
filter_func: a filter function used to filter the weights to be loaded. |
|
|
See 'filter_swinunetr' in "monai.networks.nets.swin_unetr.py". |
|
|
|
|
|
Examples: |
|
|
.. code-block:: python |
|
|
|
|
|
from monai.networks.nets import BasicUNet |
|
|
from monai.networks.utils import copy_model_state |
|
|
|
|
|
model_a = BasicUNet(in_channels=1, out_channels=4) |
|
|
model_b = BasicUNet(in_channels=1, out_channels=2) |
|
|
model_a_b, changed, unchanged = copy_model_state( |
|
|
model_a, model_b, exclude_vars="conv_0.conv_0", inplace=False) |
|
|
# dst model updated: 76 of 82 variables. |
|
|
model_a.load_state_dict(model_a_b) |
|
|
# <All keys matched successfully> |
|
|
|
|
|
Returns: an OrderedDict of the updated `dst` state, the changed, and unchanged keys. |
|
|
|
|
|
""" |
|
|
src_dict = get_state_dict(src) |
|
|
dst_dict = OrderedDict(get_state_dict(dst)) |
|
|
|
|
|
to_skip = {s_key for s_key in src_dict if exclude_vars and re.compile(exclude_vars).search(s_key)} |
|
|
|
|
|
|
|
|
all_keys, updated_keys = list(dst_dict), list() |
|
|
for s, val in src_dict.items(): |
|
|
dst_key = f"{dst_prefix}{s}" |
|
|
if dst_key in dst_dict and dst_key not in to_skip and dst_dict[dst_key].shape == val.shape: |
|
|
dst_dict[dst_key] = val |
|
|
updated_keys.append(dst_key) |
|
|
for s in mapping if mapping else {}: |
|
|
dst_key = f"{dst_prefix}{mapping[s]}" |
|
|
if dst_key in dst_dict and dst_key not in to_skip: |
|
|
if dst_dict[dst_key].shape != src_dict[s].shape: |
|
|
warnings.warn(f"Param. shape changed from {dst_dict[dst_key].shape} to {src_dict[s].shape}.") |
|
|
dst_dict[dst_key] = src_dict[s] |
|
|
updated_keys.append(dst_key) |
|
|
if filter_func is not None: |
|
|
for key, value in src_dict.items(): |
|
|
new_pair = filter_func(key, value) |
|
|
if new_pair is not None and new_pair[0] not in to_skip: |
|
|
dst_dict[new_pair[0]] = new_pair[1] |
|
|
updated_keys.append(new_pair[0]) |
|
|
|
|
|
updated_keys = sorted(set(updated_keys)) |
|
|
unchanged_keys = sorted(set(all_keys).difference(updated_keys)) |
|
|
logger.info(f"'dst' model updated: {len(updated_keys)} of {len(dst_dict)} variables.") |
|
|
if inplace and isinstance(dst, torch.nn.Module): |
|
|
if isinstance(dst, (nn.DataParallel, nn.parallel.DistributedDataParallel)): |
|
|
dst = dst.module |
|
|
dst.load_state_dict(dst_dict) |
|
|
return dst_dict, updated_keys, unchanged_keys |
|
|
|
|
|
|
|
|
def save_state(src: torch.nn.Module | dict, path: PathLike, **kwargs): |
|
|
""" |
|
|
Save the state dict of input source data with PyTorch `save`. |
|
|
It can save `nn.Module`, `state_dict`, a dictionary of `nn.Module` or `state_dict`. |
|
|
And automatically convert the data parallel module to regular module. |
|
|
For example:: |
|
|
|
|
|
save_state(net, path) |
|
|
save_state(net.state_dict(), path) |
|
|
save_state({"net": net, "opt": opt}, path) |
|
|
net_dp = torch.nn.DataParallel(net) |
|
|
save_state(net_dp, path) |
|
|
|
|
|
Refer to: https://pytorch.org/ignite/v0.4.8/generated/ignite.handlers.DiskSaver.html. |
|
|
|
|
|
Args: |
|
|
src: input data to save, can be `nn.Module`, `state_dict`, a dictionary of `nn.Module` or `state_dict`. |
|
|
path: target file path to save the input object. |
|
|
kwargs: other args for the `save_obj` except for the `obj` and `path`. |
|
|
default `func` is `torch.save()`, details of the args: |
|
|
https://pytorch.org/docs/stable/generated/torch.save.html. |
|
|
|
|
|
""" |
|
|
|
|
|
ckpt: dict = {} |
|
|
if isinstance(src, dict): |
|
|
for k, v in src.items(): |
|
|
ckpt[k] = get_state_dict(v) |
|
|
else: |
|
|
ckpt = get_state_dict(src) |
|
|
|
|
|
save_obj(obj=ckpt, path=path, **kwargs) |
|
|
|
|
|
|
|
|
def convert_to_onnx( |
|
|
model: nn.Module, |
|
|
inputs: Sequence[Any], |
|
|
input_names: Sequence[str] | None = None, |
|
|
output_names: Sequence[str] | None = None, |
|
|
opset_version: int | None = None, |
|
|
dynamic_axes: Mapping[str, Mapping[int, str]] | Mapping[str, Sequence[int]] | None = None, |
|
|
filename: Any | None = None, |
|
|
verify: bool = False, |
|
|
device: torch.device | None = None, |
|
|
use_ort: bool = False, |
|
|
ort_provider: Sequence[str] | None = None, |
|
|
rtol: float = 1e-4, |
|
|
atol: float = 0.0, |
|
|
use_trace: bool = True, |
|
|
**kwargs, |
|
|
): |
|
|
""" |
|
|
Utility to convert a model into ONNX model and optionally verify with ONNX or onnxruntime. |
|
|
See also: https://pytorch.org/docs/stable/onnx.html for how to convert a PyTorch model to ONNX. |
|
|
|
|
|
Args: |
|
|
model: source PyTorch model to save. |
|
|
inputs: input sample data used by pytorch.onnx.export. It is also used in ONNX model verification. |
|
|
input_names: optional input names of the ONNX model. |
|
|
output_names: optional output names of the ONNX model. |
|
|
opset_version: version of the (ai.onnx) opset to target. Must be >= 7 and not exceed |
|
|
the latest opset version supported by PyTorch, for more details: |
|
|
https://github.com/onnx/onnx/blob/main/docs/Operators.md and |
|
|
https://github.com/pytorch/pytorch/blob/master/torch/onnx/_constants.py |
|
|
dynamic_axes: specifies axes of tensors as dynamic (i.e. known only at run-time). If set to None, |
|
|
the exported model will have the shapes of all input and output tensors set to match given |
|
|
ones, for more details: https://pytorch.org/docs/stable/onnx.html#torch.onnx.export. |
|
|
filename: optional filename to save the ONNX model, if None, don't save the ONNX model. |
|
|
verify: whether to verify the ONNX model with ONNX or onnxruntime. |
|
|
device: target PyTorch device to verify the model, if None, use CUDA if available. |
|
|
use_ort: whether to use onnxruntime to verify the model. |
|
|
ort_provider": onnxruntime provider to use, default is ["CPUExecutionProvider"]. |
|
|
rtol: the relative tolerance when comparing the outputs of PyTorch model and TorchScript model. |
|
|
atol: the absolute tolerance when comparing the outputs of PyTorch model and TorchScript model. |
|
|
use_trace: whether to use `torch.jit.trace` to export the torchscript model. |
|
|
kwargs: other arguments except `obj` for `torch.jit.script()` to convert model, for more details: |
|
|
https://pytorch.org/docs/master/generated/torch.jit.script.html. |
|
|
|
|
|
""" |
|
|
model.eval() |
|
|
with torch.no_grad(): |
|
|
torch_versioned_kwargs = {} |
|
|
if use_trace: |
|
|
|
|
|
mode_to_export = model |
|
|
else: |
|
|
if not pytorch_after(1, 10): |
|
|
if "example_outputs" not in kwargs: |
|
|
|
|
|
raise TypeError( |
|
|
"example_outputs is required in scripting mode before PyTorch 1.10." |
|
|
"Please provide example outputs or use trace mode to export onnx model." |
|
|
) |
|
|
torch_versioned_kwargs["example_outputs"] = kwargs["example_outputs"] |
|
|
del kwargs["example_outputs"] |
|
|
mode_to_export = torch.jit.script(model, **kwargs) |
|
|
|
|
|
if filename is None: |
|
|
f = io.BytesIO() |
|
|
torch.onnx.export( |
|
|
mode_to_export, |
|
|
tuple(inputs), |
|
|
f=f, |
|
|
input_names=input_names, |
|
|
output_names=output_names, |
|
|
dynamic_axes=dynamic_axes, |
|
|
opset_version=opset_version, |
|
|
**torch_versioned_kwargs, |
|
|
) |
|
|
onnx_model = onnx.load_model_from_string(f.getvalue()) |
|
|
else: |
|
|
torch.onnx.export( |
|
|
mode_to_export, |
|
|
tuple(inputs), |
|
|
f=filename, |
|
|
input_names=input_names, |
|
|
output_names=output_names, |
|
|
dynamic_axes=dynamic_axes, |
|
|
opset_version=opset_version, |
|
|
**torch_versioned_kwargs, |
|
|
) |
|
|
onnx_model = onnx.load(filename) |
|
|
|
|
|
if verify: |
|
|
if device is None: |
|
|
device = torch.device("cuda" if torch.cuda.is_available() else "cpu") |
|
|
|
|
|
inputs = [i.to(device) if isinstance(i, torch.Tensor) else i for i in inputs] |
|
|
model = model.to(device) |
|
|
|
|
|
with torch.no_grad(): |
|
|
set_determinism(seed=0) |
|
|
torch_out = ensure_tuple(model(*inputs), True) |
|
|
|
|
|
set_determinism(seed=0) |
|
|
model_input_names = [i.name for i in onnx_model.graph.input] |
|
|
input_dict = dict(zip(model_input_names, [i.cpu().numpy() for i in inputs])) |
|
|
if use_ort: |
|
|
ort_sess = onnxruntime.InferenceSession( |
|
|
onnx_model.SerializeToString(), providers=ort_provider if ort_provider else ["CPUExecutionProvider"] |
|
|
) |
|
|
onnx_out = ort_sess.run(None, input_dict) |
|
|
else: |
|
|
sess = onnxreference.ReferenceEvaluator(onnx_model) |
|
|
onnx_out = sess.run(None, input_dict) |
|
|
set_determinism(seed=None) |
|
|
|
|
|
for r1, r2 in zip(torch_out, onnx_out): |
|
|
if isinstance(r1, torch.Tensor): |
|
|
assert_fn = torch.testing.assert_close if pytorch_after(1, 11) else torch.testing.assert_allclose |
|
|
assert_fn(r1.cpu(), convert_to_tensor(r2, dtype=r1.dtype), rtol=rtol, atol=atol) |
|
|
|
|
|
return onnx_model |
|
|
|
|
|
|
|
|
def convert_to_torchscript( |
|
|
model: nn.Module, |
|
|
filename_or_obj: Any | None = None, |
|
|
extra_files: dict | None = None, |
|
|
verify: bool = False, |
|
|
inputs: Sequence[Any] | None = None, |
|
|
device: torch.device | None = None, |
|
|
rtol: float = 1e-4, |
|
|
atol: float = 0.0, |
|
|
use_trace: bool = False, |
|
|
**kwargs, |
|
|
): |
|
|
""" |
|
|
Utility to convert a model into TorchScript model and save to file, |
|
|
with optional input / output data verification. |
|
|
|
|
|
Args: |
|
|
model: source PyTorch model to save. |
|
|
filename_or_obj: if not None, specify a file-like object (has to implement write and flush) |
|
|
or a string containing a file path name to save the TorchScript model. |
|
|
extra_files: map from filename to contents which will be stored as part of the save model file. |
|
|
for more details: https://pytorch.org/docs/stable/generated/torch.jit.save.html. |
|
|
verify: whether to verify the input and output of TorchScript model. |
|
|
if `filename_or_obj` is not None, load the saved TorchScript model and verify. |
|
|
inputs: input test data to verify model, should be a sequence of data, every item maps to a argument |
|
|
of `model()` function. |
|
|
device: target device to verify the model, if None, use CUDA if available. |
|
|
rtol: the relative tolerance when comparing the outputs of PyTorch model and TorchScript model. |
|
|
atol: the absolute tolerance when comparing the outputs of PyTorch model and TorchScript model. |
|
|
use_trace: whether to use `torch.jit.trace` to export the TorchScript model. |
|
|
kwargs: other arguments except `obj` for `torch.jit.script()` or `torch.jit.trace()` (if use_trace is True) |
|
|
to convert model, for more details: https://pytorch.org/docs/master/generated/torch.jit.script.html. |
|
|
|
|
|
""" |
|
|
model.eval() |
|
|
with torch.no_grad(): |
|
|
if use_trace: |
|
|
if inputs is None: |
|
|
raise ValueError("Missing input data for tracing convert.") |
|
|
script_module = torch.jit.trace(model, example_inputs=inputs, **kwargs) |
|
|
else: |
|
|
script_module = torch.jit.script(model, **kwargs) |
|
|
if filename_or_obj is not None: |
|
|
torch.jit.save(m=script_module, f=filename_or_obj, _extra_files=extra_files) |
|
|
|
|
|
if verify: |
|
|
if device is None: |
|
|
device = torch.device("cuda" if torch.cuda.is_available() else "cpu") |
|
|
if inputs is None: |
|
|
raise ValueError("Missing input data for verification.") |
|
|
|
|
|
inputs = [i.to(device) if isinstance(i, torch.Tensor) else i for i in inputs] |
|
|
ts_model = torch.jit.load(filename_or_obj) if filename_or_obj is not None else script_module |
|
|
ts_model.eval().to(device) |
|
|
model = model.to(device) |
|
|
|
|
|
with torch.no_grad(): |
|
|
set_determinism(seed=0) |
|
|
torch_out = ensure_tuple(model(*inputs)) |
|
|
set_determinism(seed=0) |
|
|
torchscript_out = ensure_tuple(ts_model(*inputs)) |
|
|
set_determinism(seed=None) |
|
|
|
|
|
for r1, r2 in zip(torch_out, torchscript_out): |
|
|
if isinstance(r1, torch.Tensor) or isinstance(r2, torch.Tensor): |
|
|
assert_fn = torch.testing.assert_close if pytorch_after(1, 11) else torch.testing.assert_allclose |
|
|
assert_fn(r1, r2, rtol=rtol, atol=atol) |
|
|
|
|
|
return script_module |
|
|
|
|
|
|
|
|
def _onnx_trt_compile( |
|
|
onnx_model, |
|
|
min_shape: Sequence[int], |
|
|
opt_shape: Sequence[int], |
|
|
max_shape: Sequence[int], |
|
|
device: int, |
|
|
precision: str, |
|
|
input_names: Sequence[str] | None, |
|
|
output_names: Sequence[str] | None, |
|
|
): |
|
|
""" |
|
|
This function takes an ONNX model as input, exports it to a TensorRT engine, wraps the TensorRT engine |
|
|
to a TensorRT engine-based TorchScript model and return the TorchScript model. |
|
|
|
|
|
Args: |
|
|
onnx_model: the source ONNX model to compile. |
|
|
min_shape: the minimum input shape of the converted TensorRT model. |
|
|
opt_shape: the optimization input shape of the model, on which the TensorRT optimizes. |
|
|
max_shape: the maximum input shape of the converted TensorRT model. |
|
|
device: the target GPU index to convert and verify the model. |
|
|
precision: the weight precision of the converted TensorRT engine-based TorchScript model. |
|
|
Should be 'fp32' or 'fp16'. |
|
|
input_names: optional input names of the ONNX model. Should be a sequence like |
|
|
`['input_0', 'input_1', ..., 'input_N']` where N equals to the number of the |
|
|
model inputs. |
|
|
output_names: optional output names of the ONNX model. Should be a sequence like |
|
|
`['output_0', 'output_1', ..., 'output_N']` where N equals to the number of |
|
|
the model outputs. |
|
|
|
|
|
""" |
|
|
trt, _ = optional_import("tensorrt", "8.5.3") |
|
|
torch_tensorrt, _ = optional_import("torch_tensorrt", "1.4.0") |
|
|
|
|
|
input_shapes = (min_shape, opt_shape, max_shape) |
|
|
|
|
|
input_names = [] if not input_names else input_names |
|
|
output_names = [] if not output_names else output_names |
|
|
|
|
|
|
|
|
torch_tensorrt.set_device(device) |
|
|
logger = trt.Logger(trt.Logger.WARNING) |
|
|
builder = trt.Builder(logger) |
|
|
network = builder.create_network(1 << int(trt.NetworkDefinitionCreationFlag.EXPLICIT_BATCH)) |
|
|
profile = builder.create_optimization_profile() |
|
|
if input_names: |
|
|
profile.set_shape(input_names[0], *input_shapes) |
|
|
|
|
|
|
|
|
parser = trt.OnnxParser(network, logger) |
|
|
success = parser.parse(onnx_model.SerializeToString()) |
|
|
if not success: |
|
|
parser_error_message = "" |
|
|
for idx in range(parser.num_errors): |
|
|
parser_error_message += parser.get_error(idx).desc() + "\n" |
|
|
raise Exception(f"TensorRT cannot parse the ONNX model, due to:\n{parser_error_message}") |
|
|
|
|
|
|
|
|
config = builder.create_builder_config() |
|
|
config.add_optimization_profile(profile) |
|
|
if precision == "fp16": |
|
|
config.set_flag(trt.BuilderFlag.FP16) |
|
|
serialized_engine = builder.build_serialized_network(network, config) |
|
|
f = io.BytesIO() |
|
|
f.write(serialized_engine) |
|
|
|
|
|
|
|
|
trt_model = torch_tensorrt.ts.embed_engine_in_new_module( |
|
|
f.getvalue(), |
|
|
device=torch.device(f"cuda:{device}"), |
|
|
input_binding_names=input_names, |
|
|
output_binding_names=output_names, |
|
|
) |
|
|
return trt_model |
|
|
|
|
|
|
|
|
def convert_to_trt( |
|
|
model: nn.Module, |
|
|
precision: str, |
|
|
input_shape: Sequence[int], |
|
|
dynamic_batchsize: Sequence[int] | None = None, |
|
|
use_trace: bool = False, |
|
|
filename_or_obj: Any | None = None, |
|
|
verify: bool = False, |
|
|
device: int | None = None, |
|
|
use_onnx: bool | None = False, |
|
|
onnx_input_names: Sequence[str] | None = ("input_0",), |
|
|
onnx_output_names: Sequence[str] | None = ("output_0",), |
|
|
rtol: float = 1e-2, |
|
|
atol: float = 0.0, |
|
|
**kwargs, |
|
|
): |
|
|
""" |
|
|
Utility to export a model into a TensorRT engine-based TorchScript model with optional input / output data verification. |
|
|
|
|
|
There are two ways to export a model: |
|
|
1, Torch-TensorRT way: PyTorch module ---> TorchScript module ---> TensorRT engine-based TorchScript. |
|
|
2, ONNX-TensorRT way: PyTorch module ---> TorchScript module ---> ONNX model ---> TensorRT engine ---> |
|
|
TensorRT engine-based TorchScript. |
|
|
|
|
|
When exporting through the first way, some models suffer from the slowdown problem, since Torch-TensorRT |
|
|
may only convert a little part of the PyTorch model to the TensorRT engine. However when exporting through |
|
|
the second way, some Python data structures like `dict` are not supported. And some TorchScript models are |
|
|
not supported by the ONNX if exported through `torch.jit.script`. |
|
|
|
|
|
Args: |
|
|
model: a source PyTorch model to convert. |
|
|
precision: the weight precision of the converted TensorRT engine based TorchScript models. Should be 'fp32' or 'fp16'. |
|
|
input_shape: the input shape that is used to convert the model. Should be a list like [N, C, H, W] or |
|
|
[N, C, H, W, D]. |
|
|
dynamic_batchsize: a sequence with three elements to define the batch size range of the input for the model to be |
|
|
converted. Should be a sequence like [MIN_BATCH, OPT_BATCH, MAX_BATCH]. After converted, the batchsize of model |
|
|
input should between `MIN_BATCH` and `MAX_BATCH` and the `OPT_BATCH` is the best performance batchsize that the |
|
|
TensorRT tries to fit. The `OPT_BATCH` should be the most frequently used input batchsize in the application, |
|
|
default to None. |
|
|
use_trace: whether using `torch.jit.trace` to convert the PyTorch model to a TorchScript model and then convert to |
|
|
a TensorRT engine based TorchScript model or an ONNX model (if `use_onnx` is True), default to False. |
|
|
filename_or_obj: if not None, specify a file-like object (has to implement write and flush) or a string containing a |
|
|
file path name to load the TensorRT engine based TorchScript model for verifying. |
|
|
verify: whether to verify the input and output of the TensorRT engine based TorchScript model. |
|
|
device: the target GPU index to convert and verify the model. If None, use #0 GPU. |
|
|
use_onnx: whether to use the ONNX-TensorRT way to export the TensorRT engine-based TorchScript model. |
|
|
onnx_input_names: optional input names of the ONNX model. This arg is only useful when `use_onnx` is True. Should be |
|
|
a sequence like `('input_0', 'input_1', ..., 'input_N')` where N equals to the number of the model inputs. If not |
|
|
given, will use `('input_0',)`, which supposes the model only has one input. |
|
|
onnx_output_names: optional output names of the ONNX model. This arg is only useful when `use_onnx` is True. Should be |
|
|
a sequence like `('output_0', 'output_1', ..., 'output_N')` where N equals to the number of the model outputs. If |
|
|
not given, will use `('output_0',)`, which supposes the model only has one output. |
|
|
rtol: the relative tolerance when comparing the outputs between the PyTorch model and TensorRT model. |
|
|
atol: the absolute tolerance when comparing the outputs between the PyTorch model and TensorRT model. |
|
|
kwargs: other arguments except `module`, `inputs`, `enabled_precisions` and `device` for `torch_tensorrt.compile()` |
|
|
to compile model, for more details: https://pytorch.org/TensorRT/py_api/torch_tensorrt.html#torch-tensorrt-py. |
|
|
""" |
|
|
|
|
|
torch_tensorrt, _ = optional_import("torch_tensorrt", version="1.4.0") |
|
|
|
|
|
if not torch.cuda.is_available(): |
|
|
raise Exception("Cannot find any GPU devices.") |
|
|
|
|
|
if not input_shape: |
|
|
raise ValueError("Missing the input shape for model convert.") |
|
|
|
|
|
if not dynamic_batchsize: |
|
|
warnings.warn(f"There is no dynamic batch range. The converted model only takes {input_shape} shape input.") |
|
|
|
|
|
if (dynamic_batchsize is not None) and (len(dynamic_batchsize) != 3): |
|
|
warnings.warn(f"The dynamic batch range sequence should have 3 elements, but got {dynamic_batchsize} elements.") |
|
|
|
|
|
device = device if device else 0 |
|
|
target_device = torch.device(f"cuda:{device}") if device else torch.device("cuda:0") |
|
|
convert_precision = torch.float32 if precision == "fp32" else torch.half |
|
|
inputs = [torch.rand(ensure_tuple(input_shape)).to(target_device)] |
|
|
|
|
|
def scale_batch_size(input_shape: Sequence[int], scale_num: int): |
|
|
scale_shape = [*input_shape] |
|
|
scale_shape[0] *= scale_num |
|
|
return scale_shape |
|
|
|
|
|
|
|
|
if dynamic_batchsize: |
|
|
min_input_shape = scale_batch_size(input_shape, dynamic_batchsize[0]) |
|
|
opt_input_shape = scale_batch_size(input_shape, dynamic_batchsize[1]) |
|
|
max_input_shape = scale_batch_size(input_shape, dynamic_batchsize[2]) |
|
|
else: |
|
|
min_input_shape = opt_input_shape = max_input_shape = input_shape |
|
|
|
|
|
|
|
|
model = model.eval().to(target_device) |
|
|
ir_model = convert_to_torchscript(model, device=target_device, inputs=inputs, use_trace=use_trace) |
|
|
ir_model.eval() |
|
|
|
|
|
if use_onnx: |
|
|
|
|
|
dynamic_axes = {k: {0: "batchsize"} for k in onnx_input_names} if onnx_input_names else {} |
|
|
dynamic_axes.update({k: {0: "batchsize"} for k in onnx_output_names} if onnx_output_names else {}) |
|
|
ir_model = convert_to_onnx( |
|
|
model, inputs, onnx_input_names, onnx_output_names, use_trace=use_trace, dynamic_axes=dynamic_axes |
|
|
) |
|
|
|
|
|
|
|
|
trt_model = _onnx_trt_compile( |
|
|
ir_model, |
|
|
min_shape=min_input_shape, |
|
|
opt_shape=opt_input_shape, |
|
|
max_shape=max_input_shape, |
|
|
device=device, |
|
|
precision=precision, |
|
|
input_names=onnx_input_names, |
|
|
output_names=onnx_output_names, |
|
|
) |
|
|
else: |
|
|
|
|
|
ir_model.to(target_device) |
|
|
with torch.no_grad(): |
|
|
with torch.cuda.device(device=device): |
|
|
input_placeholder = [ |
|
|
torch_tensorrt.Input( |
|
|
min_shape=min_input_shape, opt_shape=opt_input_shape, max_shape=max_input_shape |
|
|
) |
|
|
] |
|
|
trt_model = torch_tensorrt.compile( |
|
|
ir_model, |
|
|
inputs=input_placeholder, |
|
|
enabled_precisions=convert_precision, |
|
|
device=target_device, |
|
|
ir="torchscript", |
|
|
**kwargs, |
|
|
) |
|
|
|
|
|
|
|
|
if verify: |
|
|
if inputs is None: |
|
|
raise ValueError("Missing input data for verification.") |
|
|
|
|
|
trt_model = torch.jit.load(filename_or_obj) if filename_or_obj is not None else trt_model |
|
|
|
|
|
with torch.no_grad(): |
|
|
set_determinism(seed=0) |
|
|
torch_out = ensure_tuple(model(*inputs)) |
|
|
set_determinism(seed=0) |
|
|
trt_out = ensure_tuple(trt_model(*inputs)) |
|
|
set_determinism(seed=None) |
|
|
|
|
|
for r1, r2 in zip(torch_out, trt_out): |
|
|
if isinstance(r1, torch.Tensor) or isinstance(r2, torch.Tensor): |
|
|
assert_fn = torch.testing.assert_close if pytorch_after(1, 11) else torch.testing.assert_allclose |
|
|
assert_fn(r1, r2, rtol=rtol, atol=atol) |
|
|
|
|
|
return trt_model |
|
|
|
|
|
|
|
|
def meshgrid_ij(*tensors): |
|
|
if torch.meshgrid.__kwdefaults__ is not None and "indexing" in torch.meshgrid.__kwdefaults__: |
|
|
return torch.meshgrid(*tensors, indexing="ij") |
|
|
|
|
|
return torch.meshgrid(*tensors) |
|
|
|
|
|
|
|
|
def meshgrid_xy(*tensors): |
|
|
if torch.meshgrid.__kwdefaults__ is not None and "indexing" in torch.meshgrid.__kwdefaults__: |
|
|
return torch.meshgrid(*tensors, indexing="xy") |
|
|
|
|
|
return torch.meshgrid(tensors[1], tensors[0], *tensors[2:]) |
|
|
|
|
|
|
|
|
def _replace_modules( |
|
|
parent: torch.nn.Module, |
|
|
name: str, |
|
|
new_module: torch.nn.Module, |
|
|
out: list[tuple[str, torch.nn.Module]], |
|
|
strict_match: bool = True, |
|
|
match_device: bool = True, |
|
|
) -> None: |
|
|
""" |
|
|
Helper function for :py:class:`monai.networks.utils.replace_modules`. |
|
|
""" |
|
|
if match_device: |
|
|
devices = list({i.device for i in parent.parameters()}) |
|
|
|
|
|
if len(devices) == 1: |
|
|
new_module.to(devices[0]) |
|
|
idx = name.find(".") |
|
|
|
|
|
if idx != -1: |
|
|
parent_name = name[:idx] |
|
|
parent = getattr(parent, parent_name) |
|
|
name = name[idx + 1 :] |
|
|
_out: list[tuple[str, torch.nn.Module]] = [] |
|
|
_replace_modules(parent, name, new_module, _out) |
|
|
|
|
|
out += [(f"{parent_name}.{r[0]}", r[1]) for r in _out] |
|
|
|
|
|
else: |
|
|
if strict_match: |
|
|
old_module = getattr(parent, name) |
|
|
setattr(parent, name, new_module) |
|
|
out += [(name, old_module)] |
|
|
else: |
|
|
for mod_name, _ in parent.named_modules(): |
|
|
if name in mod_name: |
|
|
_replace_modules(parent, mod_name, deepcopy(new_module), out, strict_match=True) |
|
|
|
|
|
|
|
|
def replace_modules( |
|
|
parent: torch.nn.Module, |
|
|
name: str, |
|
|
new_module: torch.nn.Module, |
|
|
strict_match: bool = True, |
|
|
match_device: bool = True, |
|
|
) -> list[tuple[str, torch.nn.Module]]: |
|
|
""" |
|
|
Replace sub-module(s) in a parent module. |
|
|
|
|
|
The name of the module to be replace can be nested e.g., |
|
|
`features.denseblock1.denselayer1.layers.relu1`. If this is the case (there are "." |
|
|
in the module name), then this function will recursively call itself. |
|
|
|
|
|
Args: |
|
|
parent: module that contains the module to be replaced |
|
|
name: name of module to be replaced. Can include ".". |
|
|
new_module: `torch.nn.Module` to be placed at position `name` inside `parent`. This will |
|
|
be deep copied if `strict_match == False` multiple instances are independent. |
|
|
strict_match: if `True`, module name must `== name`. If false then |
|
|
`name in named_modules()` will be used. `True` can be used to change just |
|
|
one module, whereas `False` can be used to replace all modules with similar |
|
|
name (e.g., `relu`). |
|
|
match_device: if `True`, the device of the new module will match the model. Requires all |
|
|
of `parent` to be on the same device. |
|
|
|
|
|
Returns: |
|
|
List of tuples of replaced modules. Element 0 is module name, element 1 is the replaced module. |
|
|
|
|
|
Raises: |
|
|
AttributeError: if `strict_match` is `True` and `name` is not a named module in `parent`. |
|
|
""" |
|
|
out: list[tuple[str, torch.nn.Module]] = [] |
|
|
_replace_modules(parent, name, new_module, out, strict_match, match_device) |
|
|
return out |
|
|
|
|
|
|
|
|
@contextmanager |
|
|
def replace_modules_temp( |
|
|
parent: torch.nn.Module, |
|
|
name: str, |
|
|
new_module: torch.nn.Module, |
|
|
strict_match: bool = True, |
|
|
match_device: bool = True, |
|
|
): |
|
|
""" |
|
|
Temporarily replace sub-module(s) in a parent module (context manager). |
|
|
|
|
|
See :py:class:`monai.networks.utils.replace_modules`. |
|
|
""" |
|
|
replaced: list[tuple[str, torch.nn.Module]] = [] |
|
|
try: |
|
|
|
|
|
_replace_modules(parent, name, new_module, replaced, strict_match, match_device) |
|
|
yield |
|
|
finally: |
|
|
|
|
|
for name, module in replaced: |
|
|
_replace_modules(parent, name, module, [], strict_match=True, match_device=match_device) |
|
|
|
|
|
|
|
|
def freeze_layers(model: nn.Module, freeze_vars=None, exclude_vars=None): |
|
|
""" |
|
|
A utilty function to help freeze specific layers. |
|
|
|
|
|
Args: |
|
|
model: a source PyTorch model to freeze layer. |
|
|
freeze_vars: a regular expression to match the `model` variable names, |
|
|
so that their `requires_grad` will set to `False`. |
|
|
exclude_vars: a regular expression to match the `model` variable names, |
|
|
except for matched variable names, other `requires_grad` will set to `False`. |
|
|
|
|
|
Raises: |
|
|
ValueError: when freeze_vars and exclude_vars are both specified. |
|
|
|
|
|
""" |
|
|
if freeze_vars is not None and exclude_vars is not None: |
|
|
raise ValueError("Incompatible values: freeze_vars and exclude_vars are both specified.") |
|
|
src_dict = get_state_dict(model) |
|
|
|
|
|
frozen_keys = list() |
|
|
if freeze_vars is not None: |
|
|
to_freeze = {s_key for s_key in src_dict if freeze_vars and re.compile(freeze_vars).search(s_key)} |
|
|
for name, param in model.named_parameters(): |
|
|
if name in to_freeze: |
|
|
param.requires_grad = False |
|
|
frozen_keys.append(name) |
|
|
elif not param.requires_grad: |
|
|
param.requires_grad = True |
|
|
warnings.warn( |
|
|
f"The freeze_vars does not include {param}, but requires_grad is False, change it to True." |
|
|
) |
|
|
if exclude_vars is not None: |
|
|
to_exclude = {s_key for s_key in src_dict if exclude_vars and re.compile(exclude_vars).search(s_key)} |
|
|
for name, param in model.named_parameters(): |
|
|
if name not in to_exclude: |
|
|
param.requires_grad = False |
|
|
frozen_keys.append(name) |
|
|
elif not param.requires_grad: |
|
|
param.requires_grad = True |
|
|
warnings.warn(f"The exclude_vars includes {param}, but requires_grad is False, change it to True.") |
|
|
|
|
|
logger.info(f"{len(frozen_keys)} of {len(src_dict)} variables frozen.") |
|
|
|