| | |
| |
|
| |
|
| | import logging |
| | import random |
| | import math |
| | from typing import Optional, Tuple, Union |
| |
|
| | import torch |
| | import torch.nn as nn |
| | from torch import Tensor |
| |
|
| | class Transpose(nn.Identity): |
| | """(N, T, D) -> (N, D, T)""" |
| |
|
| | def forward(self, input: torch.Tensor) -> torch.Tensor: |
| | return input.transpose(1, 2) |
| |
|
| |
|
| | class ActivationBalancerFunction(torch.autograd.Function): |
| | @staticmethod |
| | def forward( |
| | ctx, |
| | x: Tensor, |
| | scale_factor: Tensor, |
| | sign_factor: Optional[Tensor], |
| | channel_dim: int, |
| | ) -> Tensor: |
| | if channel_dim < 0: |
| | channel_dim += x.ndim |
| | ctx.channel_dim = channel_dim |
| | xgt0 = x > 0 |
| | if sign_factor is None: |
| | ctx.save_for_backward(xgt0, scale_factor) |
| | else: |
| | ctx.save_for_backward(xgt0, scale_factor, sign_factor) |
| | return x |
| |
|
| | @staticmethod |
| | def backward(ctx, x_grad: Tensor) -> Tuple[Tensor, None, None, None]: |
| | if len(ctx.saved_tensors) == 3: |
| | xgt0, scale_factor, sign_factor = ctx.saved_tensors |
| | for _ in range(ctx.channel_dim, x_grad.ndim - 1): |
| | scale_factor = scale_factor.unsqueeze(-1) |
| | sign_factor = sign_factor.unsqueeze(-1) |
| | factor = sign_factor + scale_factor * (xgt0.to(x_grad.dtype) - 0.5) |
| | else: |
| | xgt0, scale_factor = ctx.saved_tensors |
| | for _ in range(ctx.channel_dim, x_grad.ndim - 1): |
| | scale_factor = scale_factor.unsqueeze(-1) |
| | factor = scale_factor * (xgt0.to(x_grad.dtype) - 0.5) |
| | neg_delta_grad = x_grad.abs() * factor |
| | return ( |
| | x_grad - neg_delta_grad, |
| | None, |
| | None, |
| | None, |
| | ) |
| |
|
| |
|
| | def _compute_scale_factor( |
| | x: Tensor, |
| | channel_dim: int, |
| | min_abs: float, |
| | max_abs: float, |
| | gain_factor: float, |
| | max_factor: float, |
| | ) -> Tensor: |
| | if channel_dim < 0: |
| | channel_dim += x.ndim |
| | sum_dims = [d for d in range(x.ndim) if d != channel_dim] |
| | x_abs_mean = torch.mean(x.abs(), dim=sum_dims).to(torch.float32) |
| |
|
| | if min_abs == 0.0: |
| | below_threshold = 0.0 |
| | else: |
| | |
| | |
| | below_threshold = ( |
| | (min_abs - x_abs_mean) * (gain_factor / min_abs) |
| | ).clamp(min=0, max=max_factor) |
| |
|
| | above_threshold = ((x_abs_mean - max_abs) * (gain_factor / max_abs)).clamp( |
| | min=0, max=max_factor |
| | ) |
| |
|
| | return below_threshold - above_threshold |
| |
|
| |
|
| | def _compute_sign_factor( |
| | x: Tensor, |
| | channel_dim: int, |
| | min_positive: float, |
| | max_positive: float, |
| | gain_factor: float, |
| | max_factor: float, |
| | ) -> Tensor: |
| | if channel_dim < 0: |
| | channel_dim += x.ndim |
| | sum_dims = [d for d in range(x.ndim) if d != channel_dim] |
| | proportion_positive = torch.mean((x > 0).to(torch.float32), dim=sum_dims) |
| | if min_positive == 0.0: |
| | factor1 = 0.0 |
| | else: |
| | |
| | |
| | factor1 = ( |
| | (min_positive - proportion_positive) * (gain_factor / min_positive) |
| | ).clamp_(min=0, max=max_factor) |
| |
|
| | if max_positive == 1.0: |
| | factor2 = 0.0 |
| | else: |
| | |
| | |
| | factor2 = ( |
| | (proportion_positive - max_positive) |
| | * (gain_factor / (1.0 - max_positive)) |
| | ).clamp_(min=0, max=max_factor) |
| | sign_factor = factor1 - factor2 |
| | |
| | assert not isinstance(sign_factor, float) |
| | return sign_factor |
| |
|
| |
|
| | class ActivationScaleBalancerFunction(torch.autograd.Function): |
| | """ |
| | This object is used in class ActivationBalancer when the user specified |
| | min_positive=0, max_positive=1, so there are no constraints on the signs |
| | of the activations and only the absolute value has a constraint. |
| | """ |
| |
|
| | @staticmethod |
| | def forward( |
| | ctx, |
| | x: Tensor, |
| | sign_factor: Tensor, |
| | scale_factor: Tensor, |
| | channel_dim: int, |
| | ) -> Tensor: |
| | if channel_dim < 0: |
| | channel_dim += x.ndim |
| | ctx.channel_dim = channel_dim |
| | xgt0 = x > 0 |
| | ctx.save_for_backward(xgt0, sign_factor, scale_factor) |
| | return x |
| |
|
| | @staticmethod |
| | def backward(ctx, x_grad: Tensor) -> Tuple[Tensor, None, None, None]: |
| | xgt0, sign_factor, scale_factor = ctx.saved_tensors |
| | for _ in range(ctx.channel_dim, x_grad.ndim - 1): |
| | sign_factor = sign_factor.unsqueeze(-1) |
| | scale_factor = scale_factor.unsqueeze(-1) |
| |
|
| | factor = sign_factor + scale_factor * (xgt0.to(x_grad.dtype) - 0.5) |
| | neg_delta_grad = x_grad.abs() * factor |
| | return ( |
| | x_grad - neg_delta_grad, |
| | None, |
| | None, |
| | None, |
| | ) |
| |
|
| |
|
| | class RandomClampFunction(torch.autograd.Function): |
| | @staticmethod |
| | def forward( |
| | ctx, |
| | x: Tensor, |
| | min: Optional[float], |
| | max: Optional[float], |
| | prob: float, |
| | reflect: float, |
| | ) -> Tensor: |
| | x_clamped = torch.clamp(x, min=min, max=max) |
| | mask = torch.rand_like(x) < prob |
| | ans = torch.where(mask, x_clamped, x) |
| | if x.requires_grad: |
| | ctx.save_for_backward(ans == x) |
| | ctx.reflect = reflect |
| | if reflect != 0.0: |
| | ans = ans * (1.0 + reflect) - (x * reflect) |
| | return ans |
| |
|
| | @staticmethod |
| | def backward( |
| | ctx, ans_grad: Tensor |
| | ) -> Tuple[Tensor, None, None, None, None]: |
| | (is_same,) = ctx.saved_tensors |
| | x_grad = ans_grad * is_same.to(ans_grad.dtype) |
| | reflect = ctx.reflect |
| | if reflect != 0.0: |
| | x_grad = x_grad * (1.0 + reflect) - (ans_grad * reflect) |
| | return x_grad, None, None, None, None |
| |
|
| |
|
| | def random_clamp( |
| | x: Tensor, |
| | min: Optional[float] = None, |
| | max: Optional[float] = None, |
| | prob: float = 0.5, |
| | reflect: float = 0.0, |
| | ): |
| | return RandomClampFunction.apply(x, min, max, prob, reflect) |
| |
|
| |
|
| | def random_cast_to_half(x: Tensor, min_abs: float = 5.0e-06) -> Tensor: |
| | """ |
| | A randomized way of casting a floating point value to half precision. |
| | """ |
| | if x.dtype == torch.float16: |
| | return x |
| | x_abs = x.abs() |
| | is_too_small = x_abs < min_abs |
| | |
| | |
| | |
| | random_val = min_abs * x.sign() * (torch.rand_like(x) * min_abs < x_abs) |
| | return torch.where(is_too_small, random_val, x).to(torch.float16) |
| |
|
| |
|
| | class RandomGradFunction(torch.autograd.Function): |
| | """ |
| | Does nothing in forward pass; in backward pass, gets rid of very small grads using |
| | randomized approach that preserves expectations (intended to reduce roundoff). |
| | """ |
| |
|
| | @staticmethod |
| | def forward(ctx, x: Tensor, min_abs: float) -> Tensor: |
| | ctx.min_abs = min_abs |
| | return x |
| |
|
| | @staticmethod |
| | def backward(ctx, ans_grad: Tensor) -> Tuple[Tensor, None]: |
| | if ans_grad.dtype == torch.float16: |
| | return ( |
| | random_cast_to_half( |
| | ans_grad.to(torch.float32), min_abs=ctx.min_abs |
| | ), |
| | None, |
| | ) |
| | else: |
| | return ans_grad, None |
| |
|
| |
|
| | class RandomGrad(torch.nn.Module): |
| | """ |
| | Gets rid of very small gradients using an expectation-preserving method, intended to increase |
| | accuracy of training when using amp (automatic mixed precision) |
| | """ |
| |
|
| | def __init__(self, min_abs: float = 5.0e-06): |
| | super(RandomGrad, self).__init__() |
| | self.min_abs = min_abs |
| |
|
| | def forward(self, x: Tensor): |
| | if ( |
| | torch.jit.is_scripting() |
| | or not self.training |
| | or torch.jit.is_tracing() |
| | ): |
| | return x |
| | else: |
| | return RandomGradFunction.apply(x, self.min_abs) |
| |
|
| |
|
| | class SoftmaxFunction(torch.autograd.Function): |
| | """ |
| | Tries to handle half-precision derivatives in a randomized way that should |
| | be more accurate for training than the default behavior. |
| | """ |
| |
|
| | @staticmethod |
| | def forward(ctx, x: Tensor, dim: int): |
| | ans = x.softmax(dim=dim) |
| | |
| | |
| | |
| | if torch.is_autocast_enabled(): |
| | ans = ans.to(torch.float16) |
| | ctx.save_for_backward(ans) |
| | ctx.x_dtype = x.dtype |
| | ctx.dim = dim |
| | return ans |
| |
|
| | @staticmethod |
| | def backward(ctx, ans_grad: Tensor): |
| | (ans,) = ctx.saved_tensors |
| | with torch.cuda.amp.autocast(enabled=False): |
| | ans_grad = ans_grad.to(torch.float32) |
| | ans = ans.to(torch.float32) |
| | x_grad = ans_grad * ans |
| | x_grad = x_grad - ans * x_grad.sum(dim=ctx.dim, keepdim=True) |
| | return x_grad, None |
| |
|
| |
|
| | def softmax(x: Tensor, dim: int): |
| | if torch.jit.is_scripting() or torch.jit.is_tracing(): |
| | return x.softmax(dim) |
| |
|
| | return SoftmaxFunction.apply(x, dim) |
| |
|
| |
|
| | class MaxEigLimiterFunction(torch.autograd.Function): |
| | @staticmethod |
| | def forward( |
| | ctx, |
| | x: Tensor, |
| | coeffs: Tensor, |
| | direction: Tensor, |
| | channel_dim: int, |
| | grad_scale: float, |
| | ) -> Tensor: |
| | ctx.channel_dim = channel_dim |
| | ctx.grad_scale = grad_scale |
| | ctx.save_for_backward(x.detach(), coeffs.detach(), direction.detach()) |
| | return x |
| |
|
| | @staticmethod |
| | def backward(ctx, x_grad, *args): |
| | with torch.enable_grad(): |
| | (x_orig, coeffs, new_direction) = ctx.saved_tensors |
| | x_orig.requires_grad = True |
| | num_channels = x_orig.shape[ctx.channel_dim] |
| | x = x_orig.transpose(ctx.channel_dim, -1).reshape(-1, num_channels) |
| | new_direction.requires_grad = False |
| | x = x - x.mean(dim=0) |
| | x_var = (x ** 2).mean() |
| | x_residual = x - coeffs * new_direction |
| | x_residual_var = (x_residual ** 2).mean() |
| | |
| | |
| | variance_proportion = (x_var - x_residual_var) / (x_var + 1.0e-20) |
| | variance_proportion.backward() |
| | x_orig_grad = x_orig.grad |
| | x_extra_grad = ( |
| | x_orig.grad |
| | * ctx.grad_scale |
| | * x_grad.norm() |
| | / (x_orig_grad.norm() + 1.0e-20) |
| | ) |
| | return x_grad + x_extra_grad.detach(), None, None, None, None |
| |
|
| |
|
| | class BasicNorm(torch.nn.Module): |
| | """ |
| | This is intended to be a simpler, and hopefully cheaper, replacement for |
| | LayerNorm. The observation this is based on, is that Transformer-type |
| | networks, especially with pre-norm, sometimes seem to set one of the |
| | feature dimensions to a large constant value (e.g. 50), which "defeats" |
| | the LayerNorm because the output magnitude is then not strongly dependent |
| | on the other (useful) features. Presumably the weight and bias of the |
| | LayerNorm are required to allow it to do this. |
| | |
| | So the idea is to introduce this large constant value as an explicit |
| | parameter, that takes the role of the "eps" in LayerNorm, so the network |
| | doesn't have to do this trick. We make the "eps" learnable. |
| | |
| | Args: |
| | num_channels: the number of channels, e.g. 512. |
| | channel_dim: the axis/dimension corresponding to the channel, |
| | interprted as an offset from the input's ndim if negative. |
| | shis is NOT the num_channels; it should typically be one of |
| | {-2, -1, 0, 1, 2, 3}. |
| | eps: the initial "epsilon" that we add as ballast in: |
| | scale = ((input_vec**2).mean() + epsilon)**-0.5 |
| | Note: our epsilon is actually large, but we keep the name |
| | to indicate the connection with conventional LayerNorm. |
| | learn_eps: if true, we learn epsilon; if false, we keep it |
| | at the initial value. |
| | eps_min: float |
| | eps_max: float |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | num_channels: int, |
| | channel_dim: int = -1, |
| | eps: float = 0.25, |
| | learn_eps: bool = True, |
| | eps_min: float = -3.0, |
| | eps_max: float = 3.0, |
| | ) -> None: |
| | super(BasicNorm, self).__init__() |
| | self.num_channels = num_channels |
| | self.channel_dim = channel_dim |
| | if learn_eps: |
| | self.eps = nn.Parameter(torch.tensor(eps).log().detach()) |
| | else: |
| | self.register_buffer("eps", torch.tensor(eps).log().detach()) |
| | self.eps_min = eps_min |
| | self.eps_max = eps_max |
| |
|
| | def forward(self, x: Tensor) -> Tensor: |
| | assert x.shape[self.channel_dim] == self.num_channels |
| | eps = self.eps |
| | if self.training and random.random() < 0.25: |
| | |
| | |
| | |
| | |
| |
|
| | |
| | |
| | eps = eps.clamp(min=self.eps_min, max=self.eps_max) |
| | scales = ( |
| | torch.mean(x ** 2, dim=self.channel_dim, keepdim=True) + eps.exp() |
| | ) ** -0.5 |
| | return x * scales |
| |
|
| |
|
| | def ScaledLinear(*args, initial_scale: float = 1.0, **kwargs) -> nn.Linear: |
| | """ |
| | Behaves like a constructor of a modified version of nn.Linear |
| | that gives an easy way to set the default initial parameter scale. |
| | |
| | Args: |
| | Accepts the standard args and kwargs that nn.Linear accepts |
| | e.g. in_features, out_features, bias=False. |
| | |
| | initial_scale: you can override this if you want to increase |
| | or decrease the initial magnitude of the module's output |
| | (affects the initialization of weight_scale and bias_scale). |
| | Another option, if you want to do something like this, is |
| | to re-initialize the parameters. |
| | """ |
| | ans = nn.Linear(*args, **kwargs) |
| | with torch.no_grad(): |
| | ans.weight[:] *= initial_scale |
| | if ans.bias is not None: |
| | torch.nn.init.uniform_( |
| | ans.bias, -0.1 * initial_scale, 0.1 * initial_scale |
| | ) |
| | return ans |
| |
|
| |
|
| | def ScaledConv1d( |
| | *args, |
| | initial_scale: float = 1.0, |
| | kernel_size: int = 3, |
| | padding: str = "same", |
| | **kwargs, |
| | ) -> nn.Conv1d: |
| | """ |
| | Behaves like a constructor of a modified version of nn.Conv1d |
| | that gives an easy way to set the default initial parameter scale. |
| | |
| | Args: |
| | Accepts the standard args and kwargs that nn.Linear accepts |
| | e.g. in_features, out_features, bias=False. |
| | |
| | initial_scale: you can override this if you want to increase |
| | or decrease the initial magnitude of the module's output |
| | (affects the initialization of weight_scale and bias_scale). |
| | Another option, if you want to do something like this, is |
| | to re-initialize the parameters. |
| | """ |
| | ans = nn.Conv1d(*args, kernel_size=kernel_size, padding=padding, **kwargs) |
| | with torch.no_grad(): |
| | ans.weight[:] *= initial_scale |
| | if ans.bias is not None: |
| | torch.nn.init.uniform_( |
| | ans.bias, -0.1 * initial_scale, 0.1 * initial_scale |
| | ) |
| | return ans |
| |
|
| |
|
| | def TransposeScaledConv1d( |
| | *args, |
| | initial_scale: float = 1.0, |
| | kernel_size: int = 3, |
| | padding: str = "same", |
| | **kwargs, |
| | ) -> nn.Sequential: |
| | """ |
| | Transpose -> ScaledConv1d |
| | """ |
| | return nn.Sequential( |
| | Transpose(), |
| | ScaledConv1d( |
| | *args, |
| | initial_scale=initial_scale, |
| | kernel_size=kernel_size, |
| | padding=padding, |
| | **kwargs, |
| | ), |
| | ) |
| |
|
| |
|
| | def ScaledConv1dTranspose( |
| | *args, |
| | initial_scale: float = 1.0, |
| | kernel_size: int = 3, |
| | padding: str = "same", |
| | **kwargs, |
| | ) -> nn.Sequential: |
| | """ |
| | Transpose -> ScaledConv1d |
| | """ |
| | return nn.Sequential( |
| | ScaledConv1d( |
| | *args, |
| | initial_scale=initial_scale, |
| | kernel_size=kernel_size, |
| | padding=padding, |
| | **kwargs, |
| | ), |
| | Transpose(), |
| | ) |
| |
|
| |
|
| | def TransposeConv1d( |
| | *args, kernel_size: int = 3, padding: str = "same", **kwargs |
| | ) -> nn.Sequential: |
| | """ |
| | Transpose -> Conv1d |
| | """ |
| | return nn.Sequential( |
| | Transpose(), |
| | nn.Conv1d(*args, kernel_size=kernel_size, padding=padding, **kwargs), |
| | ) |
| |
|
| |
|
| | def Conv1dTranspose( |
| | *args, kernel_size: int = 3, padding: str = "same", **kwargs |
| | ) -> nn.Sequential: |
| | """ |
| | ScaledConv1d -> Transpose |
| | """ |
| | return nn.Sequential( |
| | nn.Conv1d(*args, kernel_size=kernel_size, padding=padding, **kwargs), |
| | Transpose(), |
| | ) |
| |
|
| |
|
| | class SRLinear(nn.Linear): |
| | """https://arxiv.org/abs/2303.06296 |
| | Stabilizing Transformer Training by Preventing Attention Entropy Collapse |
| | """ |
| |
|
| | def __init__(self, in_features, out_features, bias=True, **kwargs): |
| | super().__init__(in_features, out_features, bias=bias, **kwargs) |
| | self.register_buffer( |
| | "u", nn.functional.normalize(torch.randn(in_features), dim=0) |
| | ) |
| | with torch.no_grad(): |
| | sigma = self.get_sigma() |
| | self.register_buffer("spectral_norm", sigma) |
| | self.sigma = nn.Parameter(torch.ones(1)) |
| |
|
| | def get_sigma(self): |
| | with torch.no_grad(): |
| | u = self.u |
| | v = self.weight.mv(u) |
| | v = nn.functional.normalize(v, dim=0) |
| | u = self.weight.T.mv(v) |
| | u = nn.functional.normalize(u, dim=0) |
| | self.u.data.copy_(u) |
| | return torch.einsum("c,cd,d->", v, self.weight, u) |
| |
|
| | def get_weight(self): |
| | sigma = self.get_sigma() |
| | if self.training: |
| | self.spectral_norm.data.copy_(sigma) |
| | weight = (self.sigma / sigma) * self.weight |
| | return weight |
| |
|
| | def forward(self, x): |
| | return nn.functional.linear(x, self.get_weight(), self.bias) |
| |
|
| |
|
| | class SRConv1d(SRLinear): |
| | def __init__( |
| | self, |
| | in_features, |
| | out_features, |
| | kernel_size, |
| | stride: int = 1, |
| | padding: str = "same", |
| | bias: bool = True, |
| | **kwargs, |
| | ): |
| | in_features = in_features * kernel_size |
| | super().__init__(in_features, out_features, bias=bias, **kwargs) |
| | nn.init.kaiming_uniform_(self.weight, a=math.sqrt(5)) |
| | self.kernel_size = kernel_size |
| | self.stride = stride |
| | self.padding = padding |
| |
|
| | def forward(self, x): |
| | in_features = self.in_features // self.kernel_size |
| | weight = self.get_weight().view( |
| | self.out_features, in_features, self.kernel_size |
| | ) |
| | return nn.functional.conv1d( |
| | x, weight, bias=self.bias, stride=self.stride, padding=self.padding |
| | ) |
| |
|
| |
|
| | def TransposeSRConv1d( |
| | *args, kernel_size: int = 3, padding: str = "same", **kwargs |
| | ) -> nn.Sequential: |
| | """ |
| | Transpose -> SRConv1d |
| | """ |
| | return nn.Sequential( |
| | Transpose(), |
| | SRConv1d(*args, kernel_size=kernel_size, padding=padding, **kwargs), |
| | ) |
| |
|
| |
|
| | def SRConv1dTranspose( |
| | *args, kernel_size: int = 3, padding: str = "same", **kwargs |
| | ) -> nn.Sequential: |
| | """ |
| | SRConv1d -> Transpose |
| | """ |
| | return nn.Sequential( |
| | SRConv1d(*args, kernel_size=kernel_size, padding=padding, **kwargs), |
| | Transpose(), |
| | ) |
| |
|
| |
|
| | class ActivationBalancer(torch.nn.Module): |
| | """ |
| | Modifies the backpropped derivatives of a function to try to encourage, for |
| | each channel, that it is positive at least a proportion `threshold` of the |
| | time. It does this by multiplying negative derivative values by up to |
| | (1+max_factor), and positive derivative values by up to (1-max_factor), |
| | interpolated from 1 at the threshold to those extremal values when none |
| | of the inputs are positive. |
| | |
| | Args: |
| | num_channels: the number of channels |
| | channel_dim: the dimension/axis corresponding to the channel, e.g. |
| | -1, 0, 1, 2; will be interpreted as an offset from x.ndim if negative. |
| | min_positive: the minimum, per channel, of the proportion of the time |
| | that (x > 0), below which we start to modify the derivatives. |
| | max_positive: the maximum, per channel, of the proportion of the time |
| | that (x > 0), above which we start to modify the derivatives. |
| | max_factor: the maximum factor by which we modify the derivatives for |
| | either the sign constraint or the magnitude constraint; |
| | e.g. with max_factor=0.02, the the derivatives would be multiplied by |
| | values in the range [0.98..1.02]. |
| | sign_gain_factor: determines the 'gain' with which we increase the |
| | change in gradient once the constraints on min_positive and max_positive |
| | are violated. |
| | scale_gain_factor: determines the 'gain' with which we increase the |
| | change in gradient once the constraints on min_abs and max_abs |
| | are violated. |
| | min_abs: the minimum average-absolute-value difference from the mean |
| | value per channel, which we allow, before we start to modify |
| | the derivatives to prevent this. |
| | max_abs: the maximum average-absolute-value difference from the mean |
| | value per channel, which we allow, before we start to modify |
| | the derivatives to prevent this. |
| | min_prob: determines the minimum probability with which we modify the |
| | gradients for the {min,max}_positive and {min,max}_abs constraints, |
| | on each forward(). This is done randomly to prevent all layers |
| | from doing it at the same time. Early in training we may use |
| | higher probabilities than this; it will decay to this value. |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | num_channels: int, |
| | channel_dim: int, |
| | min_positive: float = 0.05, |
| | max_positive: float = 0.95, |
| | max_factor: float = 0.04, |
| | sign_gain_factor: float = 0.01, |
| | scale_gain_factor: float = 0.02, |
| | min_abs: float = 0.2, |
| | max_abs: float = 100.0, |
| | min_prob: float = 0.1, |
| | ): |
| | super(ActivationBalancer, self).__init__() |
| | self.num_channels = num_channels |
| | self.channel_dim = channel_dim |
| | self.min_positive = min_positive |
| | self.max_positive = max_positive |
| | self.max_factor = max_factor |
| | self.min_abs = min_abs |
| | self.max_abs = max_abs |
| | self.min_prob = min_prob |
| | self.sign_gain_factor = sign_gain_factor |
| | self.scale_gain_factor = scale_gain_factor |
| |
|
| | |
| | |
| | |
| | self.cpu_count = 0 |
| | self.register_buffer("count", torch.tensor(0, dtype=torch.int64)) |
| |
|
| | def forward(self, x: Tensor) -> Tensor: |
| | if ( |
| | torch.jit.is_scripting() |
| | or not x.requires_grad |
| | or torch.jit.is_tracing() |
| | ): |
| | return _no_op(x) |
| |
|
| | count = self.cpu_count |
| | self.cpu_count += 1 |
| |
|
| | if random.random() < 0.01: |
| | |
| | |
| | |
| | self.cpu_count = max(self.cpu_count, self.count.item()) |
| | self.count.fill_(self.cpu_count) |
| |
|
| | |
| | |
| | prob = max(self.min_prob, 0.5 ** (1 + (count / 4000.0))) |
| |
|
| | if random.random() < prob: |
| | sign_gain_factor = 0.5 |
| | if self.min_positive != 0.0 or self.max_positive != 1.0: |
| | sign_factor = _compute_sign_factor( |
| | x, |
| | self.channel_dim, |
| | self.min_positive, |
| | self.max_positive, |
| | gain_factor=self.sign_gain_factor / prob, |
| | max_factor=self.max_factor, |
| | ) |
| | else: |
| | sign_factor = None |
| |
|
| | scale_factor = _compute_scale_factor( |
| | x.detach(), |
| | self.channel_dim, |
| | min_abs=self.min_abs, |
| | max_abs=self.max_abs, |
| | gain_factor=self.scale_gain_factor / prob, |
| | max_factor=self.max_factor, |
| | ) |
| | return ActivationBalancerFunction.apply( |
| | x, |
| | scale_factor, |
| | sign_factor, |
| | self.channel_dim, |
| | ) |
| | else: |
| | return _no_op(x) |
| |
|
| |
|
| | def penalize_abs_values_gt(x: Tensor, limit: float, penalty: float) -> Tensor: |
| | """ |
| | Returns x unmodified, but in backprop will put a penalty for the excess of |
| | the absolute values of elements of x over the limit "limit". E.g. if |
| | limit == 10.0, then if x has any values over 10 it will get a penalty. |
| | |
| | Caution: the value of this penalty will be affected by grad scaling used |
| | in automatic mixed precision training. For this reasons we use this, |
| | it shouldn't really matter, or may even be helpful; we just use this |
| | to disallow really implausible values of scores to be given to softmax. |
| | """ |
| | x_sign = x.sign() |
| | over_limit = (x.abs() - limit) > 0 |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | aux_loss = penalty * ((x_sign * over_limit).to(torch.int8) * x) |
| | |
| | |
| | x = with_loss(x, aux_loss) |
| | |
| | return x |
| |
|
| |
|
| | def _diag(x: Tensor): |
| | if x.ndim == 2: |
| | return x.diag() |
| | else: |
| | (batch, dim, dim) = x.shape |
| | x = x.reshape(batch, dim * dim) |
| | x = x[:, :: dim + 1] |
| | assert x.shape == (batch, dim) |
| | return x |
| |
|
| |
|
| | def _whitening_metric(x: Tensor, num_groups: int): |
| | """ |
| | Computes the "whitening metric", a value which will be 1.0 if all the eigenvalues of |
| | of the centered feature covariance are the same within each group's covariance matrix |
| | and also between groups. |
| | Args: |
| | x: a Tensor of shape (*, num_channels) |
| | num_groups: the number of groups of channels, a number >=1 that divides num_channels |
| | Returns: |
| | Returns a scalar Tensor that will be 1.0 if the data is "perfectly white" and |
| | greater than 1.0 otherwise. |
| | """ |
| | assert x.dtype != torch.float16 |
| | x = x.reshape(-1, x.shape[-1]) |
| | (num_frames, num_channels) = x.shape |
| | assert num_channels % num_groups == 0 |
| | channels_per_group = num_channels // num_groups |
| | x = x.reshape(num_frames, num_groups, channels_per_group).transpose(0, 1) |
| | |
| | |
| | |
| | |
| | |
| | x = x - x.mean(dim=1, keepdim=True) |
| | |
| | x_covar = torch.matmul(x.transpose(1, 2), x) |
| | x_covar_mean_diag = _diag(x_covar).mean() |
| | |
| | |
| | |
| | x_covarsq_mean_diag = (x_covar ** 2).sum() / ( |
| | num_groups * channels_per_group |
| | ) |
| | |
| | metric = x_covarsq_mean_diag / (x_covar_mean_diag ** 2 + 1.0e-20) |
| | return metric |
| |
|
| |
|
| | class WhiteningPenaltyFunction(torch.autograd.Function): |
| | @staticmethod |
| | def forward( |
| | ctx, |
| | x: Tensor, |
| | num_groups: int, |
| | whitening_limit: float, |
| | grad_scale: float, |
| | ) -> Tensor: |
| | ctx.save_for_backward(x) |
| | ctx.num_groups = num_groups |
| | ctx.whitening_limit = whitening_limit |
| | ctx.grad_scale = grad_scale |
| | return x |
| |
|
| | @staticmethod |
| | def backward(ctx, x_grad: Tensor): |
| | (x_orig,) = ctx.saved_tensors |
| | with torch.enable_grad(): |
| | with torch.cuda.amp.autocast(enabled=False): |
| | x_detached = x_orig.to(torch.float32).detach() |
| | x_detached.requires_grad = True |
| |
|
| | metric = _whitening_metric(x_detached, ctx.num_groups) |
| |
|
| | if random.random() < 0.005 or __name__ == "__main__": |
| | logging.info( |
| | f"Whitening: num_groups={ctx.num_groups}, num_channels={x_orig.shape[-1]}, " |
| | f"metric={metric.item():.2f} vs. limit={ctx.whitening_limit}" |
| | ) |
| |
|
| | (metric - ctx.whitening_limit).relu().backward() |
| | penalty_grad = x_detached.grad |
| | scale = ctx.grad_scale * ( |
| | x_grad.to(torch.float32).norm() |
| | / (penalty_grad.norm() + 1.0e-20) |
| | ) |
| | penalty_grad = penalty_grad * scale |
| | return x_grad + penalty_grad.to(x_grad.dtype), None, None, None |
| |
|
| |
|
| | class Whiten(nn.Module): |
| | def __init__( |
| | self, |
| | num_groups: int, |
| | whitening_limit: float, |
| | prob: Union[float, Tuple[float, float]], |
| | grad_scale: float, |
| | ): |
| | """ |
| | Args: |
| | num_groups: the number of groups to divide the channel dim into before |
| | whitening. We will attempt to make the feature covariance |
| | within each group, after mean subtraction, as "white" as possible, |
| | while having the same trace across all groups. |
| | whitening_limit: a value greater than 1.0, that dictates how much |
| | freedom we have to violate the constraints. 1.0 would mean perfectly |
| | white, with exactly the same trace across groups; larger values |
| | give more freedom. E.g. 2.0. |
| | prob: the probability with which we apply the gradient modification |
| | (also affects the grad scale). May be supplied as a float, |
| | or as a pair (min_prob, max_prob) |
| | |
| | grad_scale: determines the scale on the gradient term from this object, |
| | relative to the rest of the gradient on the attention weights. |
| | E.g. 0.02 (you may want to use smaller values than this if prob is large) |
| | """ |
| | super(Whiten, self).__init__() |
| | assert num_groups >= 1 |
| | assert whitening_limit >= 1 |
| | assert grad_scale >= 0 |
| | self.num_groups = num_groups |
| | self.whitening_limit = whitening_limit |
| | if isinstance(prob, float): |
| | assert 0 < prob <= 1 |
| | self.prob = prob |
| | else: |
| | (self.min_prob, self.max_prob) = prob |
| | assert 0 < self.min_prob < self.max_prob <= 1 |
| | self.prob = self.max_prob |
| |
|
| | self.grad_scale = grad_scale |
| |
|
| | def forward(self, x: Tensor) -> Tensor: |
| | """ |
| | In the forward pass, this function just returns the input unmodified. |
| | In the backward pass, it will modify the gradients to ensure that the |
| | distribution in each group has close to (lambda times I) as the covariance |
| | after mean subtraction, with the same lambda across groups. |
| | For whitening_limit > 1, there will be more freedom to violate this |
| | constraint. |
| | |
| | Args: |
| | x: the input of shape (*, num_channels) |
| | |
| | Returns: |
| | x, unmodified. You should make sure |
| | you use the returned value, or the graph will be freed |
| | and nothing will happen in backprop. |
| | """ |
| | if ( |
| | not x.requires_grad |
| | or random.random() > self.prob |
| | or self.grad_scale == 0 |
| | ): |
| | return _no_op(x) |
| | else: |
| | if hasattr(self, "min_prob") and random.random() < 0.25: |
| | |
| | |
| | if ( |
| | _whitening_metric(x.to(torch.float32), self.num_groups) |
| | > self.whitening_limit |
| | ): |
| | |
| | self.prob = self.max_prob |
| | else: |
| | self.prob = self.min_prob |
| |
|
| | return WhiteningPenaltyFunction.apply( |
| | x, self.num_groups, self.whitening_limit, self.grad_scale |
| | ) |
| |
|
| |
|
| | class WithLoss(torch.autograd.Function): |
| | @staticmethod |
| | def forward(ctx, x: Tensor, y: Tensor): |
| | ctx.y_shape = y.shape |
| | return x |
| |
|
| | @staticmethod |
| | def backward(ctx, ans_grad: Tensor): |
| | return ans_grad, torch.ones( |
| | ctx.y_shape, dtype=ans_grad.dtype, device=ans_grad.device |
| | ) |
| |
|
| |
|
| | def with_loss(x, y): |
| | if torch.jit.is_scripting() or torch.jit.is_tracing(): |
| | return x |
| | |
| | return WithLoss.apply(x, y) |
| |
|
| |
|
| | def _no_op(x: Tensor) -> Tensor: |
| | if torch.jit.is_scripting() or torch.jit.is_tracing(): |
| | return x |
| | else: |
| | |
| | |
| | return x.chunk(1, dim=-1)[0] |
| |
|
| |
|
| | class Identity(torch.nn.Module): |
| | def __init__(self): |
| | super(Identity, self).__init__() |
| |
|
| | def forward(self, x): |
| | return _no_op(x) |
| |
|
| |
|
| | class MaxEig(torch.nn.Module): |
| | """ |
| | Modifies the backpropped derivatives of a function to try to discourage |
| | that any given direction in activation space accounts for more than |
| | a specified proportion of the covariance (e.g. 0.2). |
| | |
| | |
| | Args: |
| | num_channels: the number of channels |
| | channel_dim: the dimension/axis corresponding to the channel, e.g. |
| | -1, 0, 1, 2; will be interpreted as an offset from x.ndim if negative. |
| | max_var_per_eig: the maximum proportion of the variance of the |
| | features/channels, after mean subtraction, that can come from |
| | any given eigenvalue. |
| | min_prob: the minimum probability with which we apply this during any invocation |
| | of forward(), assuming last time we applied the constraint it was |
| | not active; supplied for speed. |
| | scale: determines the scale with which we modify the gradients, relative |
| | to the existing / unmodified gradients |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | num_channels: int, |
| | channel_dim: int, |
| | max_var_per_eig: float = 0.2, |
| | min_prob: float = 0.01, |
| | scale: float = 0.01, |
| | ): |
| | super(MaxEig, self).__init__() |
| | self.num_channels = num_channels |
| | self.channel_dim = channel_dim |
| | self.scale = scale |
| | assert max_var_per_eig == 0.0 or max_var_per_eig > 1.0 / num_channels |
| | self.max_var_per_eig = max_var_per_eig |
| |
|
| | |
| | |
| | with torch.no_grad(): |
| | |
| | |
| | direction = torch.arange(num_channels).to(torch.float) |
| | direction = direction / direction.norm() |
| | self.register_buffer("max_eig_direction", direction) |
| |
|
| | self.min_prob = min_prob |
| | |
| | |
| | |
| | self.cur_prob = 1.0 |
| |
|
| | def forward(self, x: Tensor) -> Tensor: |
| | if ( |
| | torch.jit.is_scripting() |
| | or self.max_var_per_eig <= 0 |
| | or random.random() > self.cur_prob |
| | or torch.jit.is_tracing() |
| | ): |
| | return _no_op(x) |
| |
|
| | with torch.cuda.amp.autocast(enabled=False): |
| | eps = 1.0e-20 |
| | orig_x = x |
| | x = x.to(torch.float32) |
| | with torch.no_grad(): |
| | x = x.transpose(self.channel_dim, -1).reshape( |
| | -1, self.num_channels |
| | ) |
| | x = x - x.mean(dim=0) |
| | new_direction, coeffs = self._find_direction_coeffs( |
| | x, self.max_eig_direction |
| | ) |
| | x_var = (x ** 2).mean() |
| | x_residual = x - coeffs * new_direction |
| | x_residual_var = (x_residual ** 2).mean() |
| |
|
| | |
| | |
| | variance_proportion = (x_var - x_residual_var) / ( |
| | x_var + 1.0e-20 |
| | ) |
| |
|
| | |
| | self._set_direction( |
| | 0.1 * self.max_eig_direction + new_direction |
| | ) |
| |
|
| | if random.random() < 0.01 or __name__ == "__main__": |
| | logging.info( |
| | f"variance_proportion = {variance_proportion.item()}, shape={tuple(orig_x.shape)}, cur_prob={self.cur_prob}" |
| | ) |
| |
|
| | if variance_proportion >= self.max_var_per_eig: |
| | |
| | |
| | |
| | cur_prob = self.cur_prob |
| | self.cur_prob = ( |
| | 1.0 |
| | ) |
| | return MaxEigLimiterFunction.apply( |
| | orig_x, coeffs, new_direction, self.channel_dim, self.scale |
| | ) |
| | else: |
| | |
| | |
| | self.cur_prob = 0.75 * self.cur_prob + 0.25 * self.min_prob |
| | return orig_x |
| |
|
| | def _set_direction(self, direction: Tensor): |
| | """ |
| | Sets self.max_eig_direction to a normalized version of `direction` |
| | """ |
| | direction = direction.detach() |
| | direction = direction / direction.norm() |
| | direction_sum = direction.sum().item() |
| | if direction_sum - direction_sum == 0: |
| | self.max_eig_direction[:] = direction |
| | else: |
| | logging.info( |
| | f"Warning: sum of direction in MaxEig is {direction_sum}, " |
| | "num_channels={self.num_channels}, channel_dim={self.channel_dim}" |
| | ) |
| |
|
| | def _find_direction_coeffs( |
| | self, x: Tensor, prev_direction: Tensor |
| | ) -> Tuple[Tensor, Tensor, Tensor]: |
| | """ |
| | Figure out (an approximation to) the proportion of the variance of a set of |
| | feature vectors that can be attributed to the top eigen-direction. |
| | Args: |
| | x: a Tensor of shape (num_frames, num_channels), with num_frames > 1. |
| | prev_direction: a Tensor of shape (num_channels,), that is our previous estimate |
| | of the top eigen-direction, or a random direction if this is the first |
| | iteration. Does not have to be normalized, but should be nonzero. |
| | |
| | Returns: (cur_direction, coeffs), where: |
| | cur_direction: a Tensor of shape (num_channels,) that is the current |
| | estimate of the top eigen-direction. |
| | coeffs: a Tensor of shape (num_frames, 1) that minimizes, or |
| | approximately minimizes, (x - coeffs * cur_direction).norm() |
| | """ |
| | (num_frames, num_channels) = x.shape |
| | assert num_channels > 1 and num_frames > 1 |
| | assert prev_direction.shape == (num_channels,) |
| | |
| | |
| | coeffs = (x * prev_direction).sum(dim=1, keepdim=True) + 1.0e-10 |
| | cur_direction = (x * coeffs).sum(dim=0) / ( |
| | (coeffs ** 2).sum() + 1.0e-20 |
| | ) |
| | return cur_direction, coeffs |
| |
|
| |
|
| | class DoubleSwishFunction(torch.autograd.Function): |
| | """ |
| | double_swish(x) = x * torch.sigmoid(x-1) |
| | This is a definition, originally motivated by its close numerical |
| | similarity to swish(swish(x)), where swish(x) = x * sigmoid(x). |
| | |
| | Memory-efficient derivative computation: |
| | double_swish(x) = x * s, where s(x) = torch.sigmoid(x-1) |
| | double_swish'(x) = d/dx double_swish(x) = x * s'(x) + x' * s(x) = x * s'(x) + s(x). |
| | Now, s'(x) = s(x) * (1-s(x)). |
| | double_swish'(x) = x * s'(x) + s(x). |
| | = x * s(x) * (1-s(x)) + s(x). |
| | = double_swish(x) * (1-s(x)) + s(x) |
| | ... so we just need to remember s(x) but not x itself. |
| | """ |
| |
|
| | @staticmethod |
| | def forward(ctx, x: Tensor) -> Tensor: |
| | requires_grad = x.requires_grad |
| | x_dtype = x.dtype |
| | if x.dtype == torch.float16: |
| | x = x.to(torch.float32) |
| |
|
| | s = torch.sigmoid(x - 1.0) |
| | y = x * s |
| |
|
| | if requires_grad: |
| | deriv = y * (1 - s) + s |
| | |
| | |
| | |
| | |
| | |
| | |
| | floor = -0.043637 |
| | ceil = 1.2 |
| | d_scaled = (deriv - floor) * ( |
| | 255.0 / (ceil - floor) |
| | ) + torch.rand_like(deriv) |
| | if __name__ == "__main__": |
| | |
| | assert d_scaled.min() >= 0.0 |
| | assert d_scaled.max() < 256.0 |
| | d_int = d_scaled.to(torch.uint8) |
| | ctx.save_for_backward(d_int) |
| | if x.dtype == torch.float16 or torch.is_autocast_enabled(): |
| | y = y.to(torch.float16) |
| | return y |
| |
|
| | @staticmethod |
| | def backward(ctx, y_grad: Tensor) -> Tensor: |
| | (d,) = ctx.saved_tensors |
| | |
| | floor = -0.043637 |
| | ceil = 1.2 |
| | d = d * ((ceil - floor) / 255.0) + floor |
| | return y_grad * d |
| |
|
| |
|
| | class DoubleSwish(torch.nn.Module): |
| | def forward(self, x: Tensor) -> Tensor: |
| | """Return double-swish activation function which is an approximation to Swish(Swish(x)), |
| | that we approximate closely with x * sigmoid(x-1). |
| | """ |
| | if torch.jit.is_scripting() or torch.jit.is_tracing(): |
| | return x * torch.sigmoid(x - 1.0) |
| | return DoubleSwishFunction.apply(x) |
| |
|
| |
|
| | def BalancedDoubleSwish( |
| | d_model, channel_dim=-1, max_abs=10.0, min_prob=0.25 |
| | ) -> nn.Sequential: |
| | """ |
| | ActivationBalancer -> DoubleSwish |
| | """ |
| | balancer = ActivationBalancer( |
| | d_model, channel_dim=channel_dim, max_abs=max_abs, min_prob=min_prob |
| | ) |
| | return nn.Sequential( |
| | balancer, |
| | DoubleSwish(), |
| | ) |
| |
|
| |
|
| | def _test_max_eig(): |
| | for proportion in [0.1, 0.5, 10.0]: |
| | logging.info(f"proportion = {proportion}") |
| | x = torch.randn(100, 128) |
| | direction = torch.randn(128) |
| | coeffs = torch.randn(100, 1) |
| | x += proportion * direction * coeffs |
| |
|
| | x.requires_grad = True |
| |
|
| | num_channels = 128 |
| | m = MaxEig( |
| | num_channels, 1, 0.5, scale=0.1 |
| | ) |
| |
|
| | for _ in range(4): |
| | y = m(x) |
| |
|
| | y_grad = torch.randn_like(x) |
| | y.backward(gradient=y_grad) |
| |
|
| | if proportion < 0.2: |
| | assert torch.allclose(x.grad, y_grad, atol=1.0e-02) |
| | elif proportion > 1.0: |
| | assert not torch.allclose(x.grad, y_grad) |
| |
|
| |
|
| | def _test_whiten(): |
| | for proportion in [0.1, 0.5, 10.0]: |
| | logging.info(f"_test_whiten(): proportion = {proportion}") |
| | x = torch.randn(100, 128) |
| | direction = torch.randn(128) |
| | coeffs = torch.randn(100, 1) |
| | x += proportion * direction * coeffs |
| |
|
| | x.requires_grad = True |
| |
|
| | num_channels = 128 |
| | m = Whiten( |
| | 1, 5.0, prob=1.0, grad_scale=0.1 |
| | ) |
| |
|
| | for _ in range(4): |
| | y = m(x) |
| |
|
| | y_grad = torch.randn_like(x) |
| | y.backward(gradient=y_grad) |
| |
|
| | if proportion < 0.2: |
| | assert torch.allclose(x.grad, y_grad) |
| | elif proportion > 1.0: |
| | assert not torch.allclose(x.grad, y_grad) |
| |
|
| |
|
| | def _test_activation_balancer_sign(): |
| | probs = torch.arange(0, 1, 0.01) |
| | N = 1000 |
| | x = 1.0 * ( |
| | (2.0 * (torch.rand(probs.numel(), N) < probs.unsqueeze(-1))) - 1.0 |
| | ) |
| | x = x.detach() |
| | x.requires_grad = True |
| | m = ActivationBalancer( |
| | probs.numel(), |
| | channel_dim=0, |
| | min_positive=0.05, |
| | max_positive=0.95, |
| | max_factor=0.2, |
| | min_abs=0.0, |
| | ) |
| |
|
| | y_grad = torch.sign(torch.randn(probs.numel(), N)) |
| |
|
| | y = m(x) |
| | y.backward(gradient=y_grad) |
| | print("_test_activation_balancer_sign: x = ", x) |
| | print("_test_activation_balancer_sign: y grad = ", y_grad) |
| | print("_test_activation_balancer_sign: x grad = ", x.grad) |
| |
|
| |
|
| | def _test_activation_balancer_magnitude(): |
| | magnitudes = torch.arange(0, 1, 0.01) |
| | N = 1000 |
| | x = torch.sign(torch.randn(magnitudes.numel(), N)) * magnitudes.unsqueeze( |
| | -1 |
| | ) |
| | x = x.detach() |
| | x.requires_grad = True |
| | m = ActivationBalancer( |
| | magnitudes.numel(), |
| | channel_dim=0, |
| | min_positive=0.0, |
| | max_positive=1.0, |
| | max_factor=0.2, |
| | min_abs=0.2, |
| | max_abs=0.8, |
| | min_prob=1.0, |
| | ) |
| |
|
| | y_grad = torch.sign(torch.randn(magnitudes.numel(), N)) |
| |
|
| | y = m(x) |
| | y.backward(gradient=y_grad) |
| | print("_test_activation_balancer_magnitude: x = ", x) |
| | print("_test_activation_balancer_magnitude: y grad = ", y_grad) |
| | print("_test_activation_balancer_magnitude: x grad = ", x.grad) |
| |
|
| |
|
| | def _test_basic_norm(): |
| | num_channels = 128 |
| | m = BasicNorm(num_channels=num_channels, channel_dim=1) |
| |
|
| | x = torch.randn(500, num_channels) |
| |
|
| | y = m(x) |
| |
|
| | assert y.shape == x.shape |
| | x_rms = (x ** 2).mean().sqrt() |
| | y_rms = (y ** 2).mean().sqrt() |
| | print("x rms = ", x_rms) |
| | print("y rms = ", y_rms) |
| | assert y_rms < x_rms |
| | assert y_rms > 0.5 * x_rms |
| |
|
| |
|
| | def _test_double_swish_deriv(): |
| | x = torch.randn(10, 12, dtype=torch.double) * 3.0 |
| | x.requires_grad = True |
| | m = DoubleSwish() |
| |
|
| | tol = (1.2 - (-0.043637)) / 255.0 |
| | torch.autograd.gradcheck(m, x, atol=tol) |
| |
|
| | |
| | x = torch.randn(1000, 1000, dtype=torch.double) * 3.0 |
| | x.requires_grad = True |
| | y = m(x) |
| |
|
| |
|
| | def _test_softmax(): |
| | a = torch.randn(2, 10, dtype=torch.float64) |
| | b = a.clone() |
| | a.requires_grad = True |
| | b.requires_grad = True |
| | a.softmax(dim=1)[:, 0].sum().backward() |
| | print("a grad = ", a.grad) |
| | softmax(b, dim=1)[:, 0].sum().backward() |
| | print("b grad = ", b.grad) |
| | assert torch.allclose(a.grad, b.grad) |
| |
|
| |
|
| | if __name__ == "__main__": |
| | logging.getLogger().setLevel(logging.INFO) |
| | torch.set_num_threads(1) |
| | torch.set_num_interop_threads(1) |
| | _test_softmax() |
| | _test_whiten() |
| | _test_max_eig() |
| | _test_activation_balancer_sign() |
| | _test_activation_balancer_magnitude() |
| | _test_basic_norm() |
| | _test_double_swish_deriv() |
| |
|