|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
from abc import abstractmethod |
|
|
from math import ceil, sqrt |
|
|
|
|
|
import torch |
|
|
|
|
|
from ..transform import RandomizableTransform |
|
|
|
|
|
__all__ = ["MixUp", "CutMix", "CutOut", "Mixer"] |
|
|
|
|
|
|
|
|
class Mixer(RandomizableTransform): |
|
|
|
|
|
def __init__(self, batch_size: int, alpha: float = 1.0) -> None: |
|
|
""" |
|
|
Mixer is a base class providing the basic logic for the mixup-class of |
|
|
augmentations. In all cases, we need to sample the mixing weights for each |
|
|
sample (lambda in the notation used in the papers). Also, pairs of samples |
|
|
being mixed are picked by randomly shuffling the batch samples. |
|
|
|
|
|
Args: |
|
|
batch_size (int): number of samples per batch. That is, samples are expected tp |
|
|
be of size batchsize x channels [x depth] x height x width. |
|
|
alpha (float, optional): mixing weights are sampled from the Beta(alpha, alpha) |
|
|
distribution. Defaults to 1.0, the uniform distribution. |
|
|
""" |
|
|
super().__init__() |
|
|
if alpha <= 0: |
|
|
raise ValueError(f"Expected positive number, but got {alpha = }") |
|
|
self.alpha = alpha |
|
|
self.batch_size = batch_size |
|
|
|
|
|
@abstractmethod |
|
|
def apply(self, data: torch.Tensor): |
|
|
raise NotImplementedError() |
|
|
|
|
|
def randomize(self, data=None) -> None: |
|
|
""" |
|
|
Sometimes you need may to apply the same transform to different tensors. |
|
|
The idea is to get a sample and then apply it with apply() as often |
|
|
as needed. You need to call this method everytime you apply the transform to a new |
|
|
batch. |
|
|
""" |
|
|
self._params = ( |
|
|
torch.from_numpy(self.R.beta(self.alpha, self.alpha, self.batch_size)).type(torch.float32), |
|
|
self.R.permutation(self.batch_size), |
|
|
) |
|
|
|
|
|
|
|
|
class MixUp(Mixer): |
|
|
"""MixUp as described in: |
|
|
Hongyi Zhang, Moustapha Cisse, Yann N. Dauphin, David Lopez-Paz. |
|
|
mixup: Beyond Empirical Risk Minimization, ICLR 2018 |
|
|
|
|
|
Class derived from :py:class:`monai.transforms.Mixer`. See corresponding |
|
|
documentation for details on the constructor parameters. |
|
|
""" |
|
|
|
|
|
def apply(self, data: torch.Tensor): |
|
|
weight, perm = self._params |
|
|
nsamples, *dims = data.shape |
|
|
if len(weight) != nsamples: |
|
|
raise ValueError(f"Expected batch of size: {len(weight)}, but got {nsamples}") |
|
|
|
|
|
if len(dims) not in [3, 4]: |
|
|
raise ValueError("Unexpected number of dimensions") |
|
|
|
|
|
mixweight = weight[(Ellipsis,) + (None,) * len(dims)] |
|
|
return mixweight * data + (1 - mixweight) * data[perm, ...] |
|
|
|
|
|
def __call__(self, data: torch.Tensor, labels: torch.Tensor | None = None): |
|
|
self.randomize() |
|
|
if labels is None: |
|
|
return self.apply(data) |
|
|
return self.apply(data), self.apply(labels) |
|
|
|
|
|
|
|
|
class CutMix(Mixer): |
|
|
"""CutMix augmentation as described in: |
|
|
Sangdoo Yun, Dongyoon Han, Seong Joon Oh, Sanghyuk Chun, Junsuk Choe, Youngjoon Yoo. |
|
|
CutMix: Regularization Strategy to Train Strong Classifiers with Localizable Features, |
|
|
ICCV 2019 |
|
|
|
|
|
Class derived from :py:class:`monai.transforms.Mixer`. See corresponding |
|
|
documentation for details on the constructor parameters. Here, alpha not only determines |
|
|
the mixing weight but also the size of the random rectangles used during for mixing. |
|
|
Please refer to the paper for details. |
|
|
|
|
|
The most common use case is something close to: |
|
|
|
|
|
.. code-block:: python |
|
|
|
|
|
cm = CutMix(batch_size=8, alpha=0.5) |
|
|
for batch in loader: |
|
|
images, labels = batch |
|
|
augimg, auglabels = cm(images, labels) |
|
|
output = model(augimg) |
|
|
loss = loss_function(output, auglabels) |
|
|
... |
|
|
|
|
|
""" |
|
|
|
|
|
def apply(self, data: torch.Tensor): |
|
|
weights, perm = self._params |
|
|
nsamples, _, *dims = data.shape |
|
|
if len(weights) != nsamples: |
|
|
raise ValueError(f"Expected batch of size: {len(weights)}, but got {nsamples}") |
|
|
|
|
|
mask = torch.ones_like(data) |
|
|
for s, weight in enumerate(weights): |
|
|
coords = [torch.randint(0, d, size=(1,)) for d in dims] |
|
|
lengths = [d * sqrt(1 - weight) for d in dims] |
|
|
idx = [slice(None)] + [slice(c, min(ceil(c + ln), d)) for c, ln, d in zip(coords, lengths, dims)] |
|
|
mask[s][idx] = 0 |
|
|
|
|
|
return mask * data + (1 - mask) * data[perm, ...] |
|
|
|
|
|
def apply_on_labels(self, labels: torch.Tensor): |
|
|
weights, perm = self._params |
|
|
nsamples, *dims = labels.shape |
|
|
if len(weights) != nsamples: |
|
|
raise ValueError(f"Expected batch of size: {len(weights)}, but got {nsamples}") |
|
|
|
|
|
mixweight = weights[(Ellipsis,) + (None,) * len(dims)] |
|
|
return mixweight * labels + (1 - mixweight) * labels[perm, ...] |
|
|
|
|
|
def __call__(self, data: torch.Tensor, labels: torch.Tensor | None = None): |
|
|
self.randomize() |
|
|
augmented = self.apply(data) |
|
|
return (augmented, self.apply_on_labels(labels)) if labels is not None else augmented |
|
|
|
|
|
|
|
|
class CutOut(Mixer): |
|
|
"""Cutout as described in the paper: |
|
|
Terrance DeVries, Graham W. Taylor. |
|
|
Improved Regularization of Convolutional Neural Networks with Cutout, |
|
|
arXiv:1708.04552 |
|
|
|
|
|
Class derived from :py:class:`monai.transforms.Mixer`. See corresponding |
|
|
documentation for details on the constructor parameters. Here, alpha not only determines |
|
|
the mixing weight but also the size of the random rectangles being cut put. |
|
|
Please refer to the paper for details. |
|
|
""" |
|
|
|
|
|
def apply(self, data: torch.Tensor): |
|
|
weights, _ = self._params |
|
|
nsamples, _, *dims = data.shape |
|
|
if len(weights) != nsamples: |
|
|
raise ValueError(f"Expected batch of size: {len(weights)}, but got {nsamples}") |
|
|
|
|
|
mask = torch.ones_like(data) |
|
|
for s, weight in enumerate(weights): |
|
|
coords = [torch.randint(0, d, size=(1,)) for d in dims] |
|
|
lengths = [d * sqrt(1 - weight) for d in dims] |
|
|
idx = [slice(None)] + [slice(c, min(ceil(c + ln), d)) for c, ln, d in zip(coords, lengths, dims)] |
|
|
mask[s][idx] = 0 |
|
|
|
|
|
return mask * data |
|
|
|
|
|
def __call__(self, data: torch.Tensor): |
|
|
self.randomize() |
|
|
return self.apply(data) |
|
|
|