python_code
stringlengths
0
679k
repo_name
stringlengths
9
41
file_path
stringlengths
6
149
import torch from torch.optim.optimizer import Optimizer, required from apex.multi_tensor_apply import multi_tensor_applier class FusedSGD(Optimizer): r"""Implements stochastic gradient descent (optionally with momentum). Currently GPU-only. Requires Apex to be installed via ``pip install -v --no-cache-dir --global-option="--cpp_ext" --global-option="--cuda_ext" ./``. This version of fused SGD implements 2 fusions. * Fusion of the SGD update's elementwise operations * A multi-tensor apply launch that batches the elementwise updates applied to all the model's parameters into one or a few kernel launches. :class:`apex.optimizers.FusedSGD` may be used as a drop-in replacement for ``torch.optim.SGD``:: opt = apex.optimizers.FusedSGD(model.parameters(), lr = ....) ... opt.step() :class:`apex.optimizers.FusedSGD` may be used with or without Amp. If you wish to use :class:`FusedSGD` with Amp, you may choose any ``opt_level``:: opt = apex.optimizers.FusedSGD(model.parameters(), lr = ....) model, opt = amp.initialize(model, opt, opt_level="O0" or "O1 or "O2") ... opt.step() In general, ``opt_level="O1"`` is recommended. Nesterov momentum is based on the formula from `On the importance of initialization and momentum in deep learning`__. Args: params (iterable): iterable of parameters to optimize or dicts defining parameter groups lr (float): learning rate momentum (float, optional): momentum factor (default: 0) weight_decay (float, optional): weight decay (L2 penalty) (default: 0) dampening (float, optional): dampening for momentum (default: 0) nesterov (bool, optional): enables Nesterov momentum (default: False) Example: >>> optimizer = torch.optim.SGD(model.parameters(), lr=0.1, momentum=0.9) >>> optimizer.zero_grad() >>> loss_fn(model(input), target).backward() >>> optimizer.step() __ http://www.cs.toronto.edu/%7Ehinton/absps/momentum.pdf .. note:: The implementation of SGD with Momentum/Nesterov subtly differs from Sutskever et. al. and implementations in some other frameworks. Considering the specific case of Momentum, the update can be written as .. math:: v = \rho * v + g \\ p = p - lr * v where p, g, v and :math:`\rho` denote the parameters, gradient, velocity, and momentum respectively. This is in contrast to Sutskever et. al. and other frameworks which employ an update of the form .. math:: v = \rho * v + lr * g \\ p = p - v The Nesterov version is analogously modified. """ def __init__(self, params, lr=required, momentum=0, dampening=0, weight_decay=0, nesterov=False, wd_after_momentum=False, materialize_master_grads=True, set_grad_none=False): if lr is not required and lr < 0.0: raise ValueError("Invalid learning rate: {}".format(lr)) if momentum < 0.0: raise ValueError("Invalid momentum value: {}".format(momentum)) if weight_decay < 0.0: raise ValueError("Invalid weight_decay value: {}".format(weight_decay)) defaults = dict(lr=lr, momentum=momentum, dampening=dampening, weight_decay=weight_decay, nesterov=nesterov) if nesterov and (momentum <= 0 or dampening != 0): raise ValueError("Nesterov momentum requires a momentum and zero dampening") super(FusedSGD, self).__init__(params, defaults) self.wd_after_momentum = wd_after_momentum self.materialize_master_grads = materialize_master_grads self.most_recent_scale = 1.0 self.scale_set_by_backward = False self.set_grad_none = set_grad_none if multi_tensor_applier.available: import amp_C # Skip buffer self._dummy_overflow_buf = torch.tensor([0], dtype=torch.int, device=self.param_groups[0]["params"][0].device) self.multi_tensor_sgd = amp_C.multi_tensor_sgd else: raise RuntimeError('apex.optimizers.FusedSGD requires cuda extensions') def __setstate__(self, state): super(FusedSGD, self).__setstate__(state) for group in self.param_groups: group.setdefault('nesterov', False) def zero_grad(self): if self.set_grad_none: for group in self.param_groups: for p in group['params']: p.grad = None else: super(FusedSGD, self).zero_grad() def get_momentums(self, params): momentums = [] first_run = True for p in params: param_state = self.state[p] # torch.optim.SGD initializes momentum in the main loop, we have # to do it here, and track whether or not we've done so, so that # momentum application can be skipped in the main kernel. if 'momentum_buffer' not in param_state: first_run = True buf = param_state['momentum_buffer'] = torch.zeros_like(p.data) momentums.append(buf) else: first_run = False momentums.append(param_state['momentum_buffer']) return momentums, first_run def step(self, closure=None): """Performs a single optimization step. Arguments: closure (callable, optional): A closure that reevaluates the model and returns the loss. """ loss = None if closure is not None: loss = closure() explicit_master_params = (hasattr(self, "_amp_stash") and hasattr(self._amp_stash, "fp32_from_fp16_groups")) for gid, group in enumerate(self.param_groups): weight_decay = group['weight_decay'] momentum = group['momentum'] dampening = group['dampening'] nesterov = group['nesterov'] # For each group, there are 3 possible combinations we need to consider: # grad_type, param_to_update_type, momentum_type, requires_fp16_model_copy # 1. fp16, fp16, fp16, No # 2. fp32, fp32, fp32, No # 3. fp16, fp32, fp32, Yes first_runs = [True, True] # I think a bit of code divergence in exchange for naming clarity is worthwhile if explicit_master_params: stash = self._amp_stash fp32_params = [p for p in stash.fp32_from_fp32_groups[gid] if p.grad is not None] fp32_grads = [p.grad for p in stash.fp32_from_fp32_groups[gid] if p.grad is not None] fp32_momentums, first_runs[1] = self.get_momentums(fp32_params) if self.materialize_master_grads: fp16_model_params = [p for i, p in enumerate( stash.fp16_groups[gid]) if stash.fp32_from_fp16_groups[gid][i].grad is not None] fp32_from_fp16_grads = [p.grad for p in stash.fp32_from_fp16_groups[gid] if p.grad is not None] fp32_from_fp16_params = [p for p in stash.fp32_from_fp16_groups[gid] if p.grad is not None] fp32_from_fp16_momentums, first_runs[0] = self.get_momentums(fp32_from_fp16_params) fp16_set = [fp32_from_fp16_grads, fp32_from_fp16_params, fp32_from_fp16_momentums, fp16_model_params] else: fp16_model_params = [p for p in stash.fp16_groups[gid] if p.grad is not None] fp16_model_grads = [p.grad for p in stash.fp16_groups[gid] if p.grad is not None] fp32_from_fp16_params = [p for i, p in enumerate( stash.fp32_from_fp16_groups[gid]) if stash.fp16_groups[gid][i].grad is not None] fp32_from_fp16_momentums, first_runs[0] = self.get_momentums(fp32_from_fp16_params) fp16_set = [fp16_model_grads, fp32_from_fp16_params, fp32_from_fp16_momentums, fp16_model_params] launch_sets= [fp16_set, [fp32_grads, fp32_params, fp32_momentums]] else: fp16_params = [p for p in group['params'] if (p.dtype == torch.float16 and p.grad is not None)] fp16_grads = [p.grad for p in group['params'] if (p.dtype == torch.float16 and p.grad is not None)] fp16_momentums, first_runs[0] = self.get_momentums(fp16_params) fp32_params = [p for p in group['params'] if (p.dtype == torch.float32 and p.grad is not None)] fp32_grads = [p.grad for p in group['params'] if (p.dtype == torch.float32 and p.grad is not None)] fp32_momentums, first_runs[1] = self.get_momentums(fp32_params) launch_sets = [[fp16_grads, fp16_params, fp16_momentums], [fp32_grads, fp32_params, fp32_momentums]] for s, (launch_set, first_run) in enumerate(zip(launch_sets, first_runs)): assert len(launch_set[0]) == len(launch_set[1]) assert len(launch_set[0]) == len(launch_set[2]) if len(launch_set[0]) > 0: multi_tensor_applier( self.multi_tensor_sgd, self._dummy_overflow_buf, launch_set, weight_decay, momentum, dampening, group['lr'], nesterov, first_run, self.wd_after_momentum, 1.0/self.most_recent_scale) self.most_recent_scale = 1.0 self.scale_set_by_backward = False return loss
apex-master
apex/optimizers/fused_sgd.py
import types from ..fp16_utils import master_params_to_model_params from ..multi_tensor_apply import multi_tensor_applier from ._amp_state import maybe_print import torch from ..optimizers import FusedSGD class AmpOptimizerState(object): def __init__(self): pass def _master_params_to_model_params(self): stash = self._amp_stash if multi_tensor_applier.available: if len(stash.all_fp16_params) > 0: multi_tensor_applier( stash.multi_tensor_scale, stash.dummy_overflow_buf, [stash.all_fp32_from_fp16_params, stash.all_fp16_params], 1.0) else: for fp16_group, fp32_from_fp16_group in zip(stash.fp16_groups, stash.fp32_from_fp16_groups): master_params_to_model_params(fp16_group, fp32_from_fp16_group) def lazy_init_with_master_weights(self): stash = self._amp_stash stash.fp16_groups = [] stash.fp32_from_fp16_groups = [] stash.fp32_from_fp32_groups = [] for i, param_group in enumerate(self.param_groups): # maybe_print("FP16_Optimizer processing param group {}:".format(i)) fp16_params_this_group = [] fp32_params_this_group = [] fp32_from_fp16_params_this_group = [] for i, param in enumerate(param_group['params']): if param.requires_grad: if param.type() == 'torch.cuda.HalfTensor': # maybe_print("FP16_Optimizer received torch.cuda.HalfTensor with {}" # .format(param.size())) fp16_params_this_group.append(param) master_param = param.detach().clone().float() master_param.requires_grad = True param_group['params'][i] = master_param fp32_from_fp16_params_this_group.append(master_param) # Reset existing state dict key to the new master param. # We still need to recast per-param state tensors, if any, to FP32. if param in self.state: self.state[master_param] = self.state.pop(param) elif param.type() == 'torch.cuda.FloatTensor': # maybe_print("FP16_Optimizer received torch.cuda.FloatTensor with {}" # .format(param.size())) fp32_params_this_group.append(param) param_group['params'][i] = param else: raise TypeError("Optimizer's parameters must be either " "torch.cuda.FloatTensor or torch.cuda.HalfTensor. " "Received {}".format(param.type())) stash.fp16_groups.append(fp16_params_this_group) stash.fp32_from_fp16_groups.append(fp32_from_fp16_params_this_group) stash.fp32_from_fp32_groups.append(fp32_params_this_group) stash.all_fp16_params = [] for group in stash.fp16_groups: stash.all_fp16_params += group stash.all_fp32_from_fp16_params = [] for group in stash.fp32_from_fp16_groups: stash.all_fp32_from_fp16_params += group stash.all_fp32_from_fp32_params = [] for group in stash.fp32_from_fp32_groups: stash.all_fp32_from_fp32_params += group # all_fp16_grad_stash is only needed for fused optimizers. stash.all_fp16_grad_stash = [None for _ in stash.all_fp16_params] # stash.all_fp32_from_fp16_grad_stash = [None for _ in stash.all_fp32_from_fp16_params] stash.all_fp32_from_fp32_grad_stash = [None for _ in stash.all_fp32_from_fp32_params] for param in stash.all_fp32_from_fp16_params: param.grad = None for param in stash.all_fp32_from_fp32_params: param.grad = None # Leverage state_dict() and load_state_dict() to recast preexisting per-param state tensors self.load_state_dict(self.state_dict()) def post_backward_models_are_masters(scaler, params, stashed_grads, scale_override=None): grads_have_scale, stashed_have_scale, out_scale = scaler.loss_scale(), 1.0, 1.0 # not much to do if scale == 1.0 and static scaling if scaler.loss_scale() == 1.0 and not scaler.dynamic: # Clear the stash. for i in range(len(stashed_grads)): stashed_grads[i] = None return if scale_override is not None: grads_have_scale, stashed_have_scale, out_scale = scale_override # This is a lot of python overhead... grads_needing_unscale = [] grads_needing_unscale_with_stash = [] stashed = [] for param, stashed_grad in zip(params, stashed_grads): if param.grad is None and stashed_grad is not None: param.grad = stashed_grad elif param.grad is not None and stashed_grad is None: grads_needing_unscale.append(param.grad) elif param.grad is not None and stashed_grad is not None: grads_needing_unscale_with_stash.append(param.grad) stashed.append(stashed_grad) else: # param.grad is None and stashed_grad is None continue # unscale() implements grads*(1/scale), so "scale" should be grads_have_scale/out_scale. if len(grads_needing_unscale) > 0: scaler.unscale( grads_needing_unscale, grads_needing_unscale, None, # unused_scale, currently present to avoid API breakage elsewhere models_are_masters=True, scale_override=grads_have_scale/out_scale) if len(grads_needing_unscale_with_stash) > 0: scaler.unscale_with_stashed( grads_needing_unscale_with_stash, stashed, grads_needing_unscale_with_stash, scale_override=(grads_have_scale, stashed_have_scale, out_scale)) # Clear the stash. for i in range(len(stashed_grads)): stashed_grads[i] = None def prepare_backward_with_master_weights(self): stash = self._amp_stash self._amp_lazy_init() for i, param in enumerate(stash.all_fp16_params): # Set up to leverage grad copy elision. # This may behave differently from an unpatched optimizer if zero_grad is used and the param is unused. param.grad = None # for i, param in enumerate(stash.all_fp32_from_fp16_params): # stash.all_fp32_from_fp16_grad_stash[i] = param.grad for i, param in enumerate(stash.all_fp32_from_fp32_params): stash.all_fp32_from_fp32_grad_stash[i] = param.grad # Set up to leverage grad copy elision: param.grad = None def post_backward_with_master_weights(self, scaler): stash = self._amp_stash self._amp_lazy_init() # This is a lot of python overhead... fp16_grads_needing_unscale = [] new_fp32_grads = [] fp16_grads_needing_unscale_with_stash = [] preexisting_fp32_grads = [] for fp16_param, fp32_param in zip(stash.all_fp16_params, stash.all_fp32_from_fp16_params): if fp16_param.grad is None and fp32_param.grad is not None: continue elif fp16_param.grad is not None and fp32_param.grad is None: fp32_param.grad = torch.empty_like(fp32_param) fp16_grads_needing_unscale.append(fp16_param.grad) new_fp32_grads.append(fp32_param.grad) elif fp16_param.grad is not None and fp32_param.grad is not None: fp16_grads_needing_unscale_with_stash.append(fp16_param.grad) preexisting_fp32_grads.append(fp32_param.grad) else: # fp16_param.grad is None and fp32_param.grad is None: continue if len(fp16_grads_needing_unscale) > 0: scaler.unscale( fp16_grads_needing_unscale, new_fp32_grads, scaler.loss_scale(), models_are_masters=False) if len(fp16_grads_needing_unscale_with_stash) > 0: scaler.unscale_with_stashed( fp16_grads_needing_unscale_with_stash, preexisting_fp32_grads, preexisting_fp32_grads) # fp32 params can be treated as they would be in the "no_master_weights" case. post_backward_models_are_masters( scaler, stash.all_fp32_from_fp32_params, stash.all_fp32_from_fp32_grad_stash) def lazy_init_no_master_weights(self): stash = self._amp_stash stash.all_fp16_params = [] stash.all_fp32_params = [] for i, param_group in enumerate(self.param_groups): for i, param in enumerate(param_group['params']): if param.type() == 'torch.cuda.HalfTensor': stash.all_fp16_params.append(param) elif param.type() == 'torch.cuda.FloatTensor': stash.all_fp32_params.append(param) else: raise TypeError("Optimizer's parameters must be either " "torch.cuda.FloatTensor or torch.cuda.HalfTensor. " "Received {}".format(param.type())) stash.all_fp16_grad_stash = [None for _ in stash.all_fp16_params] stash.all_fp32_grad_stash = [None for _ in stash.all_fp32_params] def prepare_backward_no_master_weights(self): stash = self._amp_stash self._amp_lazy_init() for i, param in enumerate(stash.all_fp16_params): stash.all_fp16_grad_stash[i] = param.grad # Set up to leverage grad copy elision: param.grad = None for i, param in enumerate(stash.all_fp32_params): stash.all_fp32_grad_stash[i] = param.grad # Set up to leverage grad copy elision: param.grad = None def post_backward_no_master_weights(self, scaler): stash = self._amp_stash self._amp_lazy_init() split_types = ((stash.all_fp16_params, stash.all_fp16_grad_stash), (stash.all_fp32_params, stash.all_fp32_grad_stash)) for params, stashed_grads in split_types: post_backward_models_are_masters(scaler, params, stashed_grads) ##################################################################################### # FusedSGD versions ##################################################################################### # FusedSGD never explicitly materializes the fp32 gradients for "fp32 from fp16" master params # outside the kernel, so we must accumulate directly into the model grads. def prepare_backward_with_master_weights_FusedSGD(self): if self.materialize_master_grads: prepare_backward_with_master_weights(self) else: stash = self._amp_stash self._amp_lazy_init() for i, param in enumerate(stash.all_fp16_params): stash.all_fp16_grad_stash[i] = param.grad # Set up to leverage grad copy elision: param.grad = None for i, param in enumerate(stash.all_fp32_from_fp32_params): stash.all_fp32_from_fp32_grad_stash[i] = param.grad # Set up to leverage grad copy elision: param.grad = None def post_backward_with_master_weights_FusedSGD(self, scaler): if self.materialize_master_grads: post_backward_with_master_weights(self, scaler) else: stash = self._amp_stash self._amp_lazy_init() grads_have_scale = scaler.loss_scale() stashed_have_scale = self.most_recent_scale out_scale = grads_have_scale if self.scale_set_by_backward: out_scale = min(grads_have_scale, self.most_recent_scale) split_types = ((stash.all_fp16_params, stash.all_fp16_grad_stash), (stash.all_fp32_from_fp32_params, stash.all_fp32_from_fp32_grad_stash)) # unscale_with_stashed() implements grads*1/scale + stashed_grads*1. # stashed_grads are scaled by self.most_recent_scale. for params, stashed_grads in split_types: post_backward_models_are_masters(scaler, params, stashed_grads, (grads_have_scale, stashed_have_scale, out_scale)) self.most_recent_scale = out_scale self.scale_set_by_backward = True def prepare_backward_no_master_weights_FusedSGD(self): prepare_backward_no_master_weights(self) def post_backward_no_master_weights_FusedSGD(self, scaler): post_backward_no_master_weights(self, scaler) def _amp_lazy_init(self): stash = self._amp_stash if not stash.lazy_init_called: self._lazy_init_maybe_master_weights() stash.lazy_init_called = True def _process_optimizer(optimizer, properties): if hasattr(optimizer, "_amp_stash"): raise RuntimeError("A given optimizer should only be passed through amp.initialize once.") else: optimizer._amp_stash = AmpOptimizerState() optimizer._amp_stash.lazy_init_called = False optimizer._amp_stash.already_patched = False optimizer._amp_stash.params_have_scaled_gradients = False for name in ("_lazy_init_maybe_master_weights", "_master_params_to_model_params", "_prepare_amp_backward", "_post_amp_backward", "_amp_lazy_init"): if hasattr(optimizer, name): raise RuntimeError("Incoming optimizer already has {} defined.".format(name)) # TODO: Centralize exposure and import error checking for the C backend. if multi_tensor_applier.available: import amp_C optimizer._amp_stash.multi_tensor_scale = amp_C.multi_tensor_scale optimizer._amp_stash.multi_tensor_l2norm = amp_C.multi_tensor_l2norm optimizer._amp_stash.dummy_overflow_buf = torch.cuda.IntTensor([0]); if properties.master_weights: optimizer._lazy_init_maybe_master_weights = types.MethodType( lazy_init_with_master_weights, optimizer) optimizer._master_params_to_model_params = types.MethodType( _master_params_to_model_params, optimizer) old_step = optimizer.step def new_step(self, closure=None): if closure is not None: raise RuntimeError("Currently, Amp does not support closure use with optimizers.") retval = old_step() if not isinstance(self, FusedSGD): self._master_params_to_model_params() # Clear the master grads that wouldn't be zeroed by model.zero_grad() for param in self._amp_stash.all_fp32_from_fp16_params: param.grad = None return retval optimizer.step = types.MethodType(new_step, optimizer) old_zero_grad = optimizer.zero_grad def new_zero_grad(self): stash = self._amp_stash self._amp_lazy_init() # Zero the model grads. for param in stash.all_fp16_params: if param.grad is not None: param.grad.detach_() param.grad.zero_() for param in stash.all_fp32_from_fp32_params: if param.grad is not None: param.grad.detach_() param.grad.zero_() # Clear the master grads that are independent of model grads for param in self._amp_stash.all_fp32_from_fp16_params: param.grad = None optimizer.zero_grad = types.MethodType(new_zero_grad, optimizer) if isinstance(optimizer, FusedSGD): optimizer._prepare_amp_backward = types.MethodType( prepare_backward_with_master_weights_FusedSGD, optimizer) optimizer._post_amp_backward = types.MethodType( post_backward_with_master_weights_FusedSGD, optimizer) else: optimizer._prepare_amp_backward = types.MethodType( prepare_backward_with_master_weights, optimizer) optimizer._post_amp_backward = types.MethodType( post_backward_with_master_weights, optimizer) else: optimizer._lazy_init_maybe_master_weights = types.MethodType( lazy_init_no_master_weights, optimizer) if isinstance(optimizer, FusedSGD): optimizer._prepare_amp_backward = types.MethodType( prepare_backward_no_master_weights_FusedSGD, optimizer) optimizer._post_amp_backward = types.MethodType( post_backward_no_master_weights_FusedSGD, optimizer) else: optimizer._prepare_amp_backward = types.MethodType( prepare_backward_no_master_weights, optimizer) optimizer._post_amp_backward = types.MethodType( post_backward_no_master_weights, optimizer) optimizer._amp_lazy_init = types.MethodType(_amp_lazy_init, optimizer) old_add_param_group = optimizer.add_param_group def new_add_param_group(self, new_group): stash = self._amp_stash if not stash.lazy_init_called: self._lazy_init_maybe_master_weights() stash.lazy_init_called = True assert isinstance(new_group, dict), "param group must be a dict" new_params = new_group['params'] if isinstance(new_params, torch.Tensor): new_group['params'] = [new_params] elif isinstance(new_params, set): raise TypeError('optimizer parameters need to be organized in ordered collections, but ' 'the ordering of tensors in sets will change between runs. Please use a list instead.') else: new_group['params'] = list(new_params) if properties.master_weights: # Mutate new_group in-place to use FP32 master params fp16_params_this_group = [] fp32_params_this_group = [] fp32_from_fp16_params_this_group = [] for i, param in enumerate(new_group['params']): if param.requires_grad: if param.type() == 'torch.cuda.HalfTensor': fp16_params_this_group.append(param) master_param = param.detach().clone().float() master_param.requires_grad = True new_group['params'][i] = master_param fp32_from_fp16_params_this_group.append(master_param) elif param.type() == 'torch.cuda.FloatTensor': fp32_params_this_group.append(param) new_group['params'][i] = param else: raise TypeError("Optimizer's parameters must be either " "torch.cuda.FloatTensor or torch.cuda.HalfTensor. " "Received {}".format(param.type())) stash.fp16_groups.append(fp16_params_this_group) stash.fp32_from_fp16_groups.append(fp32_from_fp16_params_this_group) stash.fp32_from_fp32_groups.append(fp32_params_this_group) stash.all_fp16_params += fp16_params_this_group stash.all_fp32_from_fp16_params += fp32_from_fp16_params_this_group stash.all_fp32_from_fp32_params += fp32_params_this_group # stash.all_fp32_from_fp16_grad_stash = [None for _ in stash.all_fp32_from_fp16_params] stash.all_fp32_from_fp32_grad_stash += [None for _ in fp32_params_this_group] # It should be ok to let params be added with existing .grad attributes. # for param in fp16_params_this_group: # param.grad = None # for param in fp32_from_fp16_params_this_group: # param.grad = None # for param in stash.fp32_params_this_group: # param.grad = None else: for param in new_group['params']: if param.type() == 'torch.cuda.HalfTensor': stash.all_fp16_params.append(param) stash.all_fp16_grad_stash.append(None) elif param.type() == 'torch.cuda.FloatTensor': stash.all_fp32_params.append(param) stash.all_fp32_grad_stash.append(None) else: raise TypeError("Optimizer's parameters must be either " "torch.cuda.FloatTensor or torch.cuda.HalfTensor. " "Received {}".format(param.type())) old_add_param_group(new_group) optimizer.add_param_group = types.MethodType(new_add_param_group, optimizer) return optimizer
apex-master
apex/amp/_process_optimizer.py
import torch # True for post-0.4, when Variables/Tensors merged. def variable_is_tensor(): v = torch.autograd.Variable() return isinstance(v, torch.Tensor) def tensor_is_variable(): x = torch.Tensor() return type(x) == torch.autograd.Variable # False for post-0.4 def tensor_is_float_tensor(): x = torch.Tensor() return type(x) == torch.FloatTensor # Akin to `torch.is_tensor`, but returns True for Variable # objects in pre-0.4. def is_tensor_like(x): return torch.is_tensor(x) or isinstance(x, torch.autograd.Variable) # Wraps `torch.is_floating_point` if present, otherwise checks # the suffix of `x.type()`. def is_floating_point(x): if hasattr(torch, 'is_floating_point'): return torch.is_floating_point(x) try: torch_type = x.type() return torch_type.endswith('FloatTensor') or \ torch_type.endswith('HalfTensor') or \ torch_type.endswith('DoubleTensor') except AttributeError: return False def scalar_python_val(x): if hasattr(x, 'item'): return x.item() else: if isinstance(x, torch.autograd.Variable): return x.data[0] else: return x[0] # Accounts for the possibility that some ops may be removed from a namespace. def filter_attrs(module, attrs): return list(attrname for attrname in attrs if hasattr(module, attrname))
apex-master
apex/amp/compat.py
import contextlib import warnings import sys import torch from . import utils from .opt import OptimWrapper from .scaler import LossScaler from ._amp_state import _amp_state, master_params, maybe_print if torch.distributed.is_available(): from ..parallel.LARC import LARC # There's no reason to expose the notion of a "handle". Everything can happen through amp.* calls. @contextlib.contextmanager def scale_loss(loss, optimizers, loss_id=0, model=None, delay_unscale=False, delay_overflow_check=False): """ On context manager entrance, creates ``scaled_loss = (loss.float())*current loss scale``. ``scaled_loss`` is yielded so that the user can call ``scaled_loss.backward()``:: with amp.scale_loss(loss, optimizer) as scaled_loss: scaled_loss.backward() On context manager exit (if ``delay_unscale=False``), the gradients are checked for infs/NaNs and unscaled, so that ``optimizer.step()`` can be called. .. note:: If Amp is using explicit FP32 master params (which is the default for ``opt_level=O2``, and can also be manually enabled by supplying ``master_weights=True`` to ``amp.initialize``) any FP16 gradients are copied to FP32 master gradients before being unscaled. ``optimizer.step()`` will then apply the unscaled master gradients to the master params. .. warning:: If Amp is using explicit FP32 master params, only the FP32 master gradients will be unscaled. The direct ``.grad`` attributes of any FP16 model params will remain scaled after context manager exit. This subtlety affects gradient clipping. See "Gradient clipping" under `Advanced Amp Usage`_ for best practices. Args: loss(Tensor): Typically a scalar Tensor. The ``scaled_loss`` that the context manager yields is simply ``loss.float()*loss_scale``, so in principle ``loss`` could have more than one element, as long as you call ``backward()`` on ``scaled_loss`` appropriately within the context manager body. optimizers: All optimizer(s) for which the current backward pass is creating gradients. Must be an optimizer or list of optimizers returned from an earlier call to ``amp.initialize``. For example use with multiple optimizers, see "Multiple models/optimizers/losses" under `Advanced Amp Usage`_. loss_id(int, optional, default=0): When used in conjunction with the ``num_losses`` argument to ``amp.initialize``, enables Amp to use a different loss scale per loss. ``loss_id`` must be an integer between 0 and ``num_losses`` that tells Amp which loss is being used for the current backward pass. See "Multiple models/optimizers/losses" under `Advanced Amp Usage`_ for examples. If ``loss_id`` is left unspecified, Amp will use the default global loss scaler for this backward pass. model(torch.nn.Module, optional, default=None): Currently unused, reserved to enable future optimizations. delay_unscale(bool, optional, default=False): ``delay_unscale`` is never necessary, and the default value of ``False`` is strongly recommended. If ``True``, Amp will not unscale the gradients or perform model->master gradient copies on context manager exit. ``delay_unscale=True`` is a minor ninja performance optimization and can result in weird gotchas (especially with multiple models/optimizers/losses), so only use it if you know what you're doing. "Gradient accumulation across iterations" under `Advanced Amp Usage`_ illustrates a situation where this CAN (but does not need to) be used. .. warning:: If ``delay_unscale`` is ``True`` for a given backward pass, ``optimizer.step()`` cannot be called yet after context manager exit, and must wait for another, later backward context manager invocation with ``delay_unscale`` left to False. .. _`Advanced Amp Usage`: https://nvidia.github.io/apex/advanced.html """ if not hasattr(_amp_state, "opt_properties"): raise RuntimeError("Invoked 'with amp.scale_loss`, but internal Amp state has not been initialized. " "model, optimizer = amp.initialize(model, optimizer, opt_level=...) must be called " "before `with amp.scale_loss`.") if not _amp_state.opt_properties.enabled: yield loss return if isinstance(optimizers, torch.optim.Optimizer) or ('LARC' in globals() and isinstance(optimizers, LARC)): optimizers = [optimizers] loss_scaler = _amp_state.loss_scalers[loss_id] loss_scale = loss_scaler.loss_scale() if ((not _amp_state.opt_properties.master_weights) and (not loss_scaler.dynamic) and loss_scale == 1.0): yield loss.float() # Needing to drop the cache here as well is an ugly gotcha. # But for now I think it's necessary to short-circuit. # Probably ok to skip this if not delay_unscale if _amp_state.opt_properties.patch_torch_functions: _amp_state.handle._clear_cache() return if not delay_unscale: if isinstance(optimizers, list): for optimizer in optimizers: if not optimizer._amp_stash.params_have_scaled_gradients: optimizer._prepare_amp_backward() yield (loss.float())*loss_scale if delay_unscale: for optimizer in optimizers: optimizer._amp_stash.params_have_scaled_gradients = True else: # FusedSGD may take care of unscaling as part of their step() methods. # if not isinstance(optimizers, FP16_Optimizer_for_fused): loss_scaler.clear_overflow_state() for optimizer in optimizers: optimizer._post_amp_backward(loss_scaler) optimizer._amp_stash.params_have_scaled_gradients = False # For future fused optimizers that enable sync-free dynamic loss scaling, # should_skip will always be False. should_skip = False if delay_overflow_check else loss_scaler.update_scale() if should_skip: for optimizer in optimizers: if not optimizer._amp_stash.already_patched: # Close on loss_scaler and loss_id as well, to be safe. Probably not # necessary because amp.scale_loss is already creating a temporary scope. def patch_step(opt, loss_scaler, loss_id): opt_step = opt.step def skip_step(closure=None): if closure is not None: raise RuntimeError("Currently, Amp does not support closure use with optimizers.") maybe_print(("Gradient overflow. Skipping step, loss scaler " + "{} reducing loss scale to {}").format(loss_id, loss_scaler.loss_scale())) # TODO: I don't like the special casing for different optimizer implementations. # Maybe skip should delegate to a method owned by the optimizers themselves. if hasattr(opt._amp_stash, "all_fp32_from_fp16_params"): # Clear the master grads that wouldn't be zeroed by model.zero_grad() for param in opt._amp_stash.all_fp32_from_fp16_params: param.grad = None if hasattr(opt, "most_recent_scale"): opt.most_recent_scale = 1.0 opt.scale_set_by_backward = False opt.step = opt_step opt._amp_stash.already_patched = False return skip_step optimizer.step = patch_step(optimizer, loss_scaler, loss_id) optimizer._amp_stash.already_patched = True # Probably ok to skip this if not delay_unscale if _amp_state.opt_properties.patch_torch_functions: _amp_state.handle._clear_cache() # Free function version of AmpHandle.disable_casts, another step on the # path to removing the concept of "AmpHandle" @contextlib.contextmanager def disable_casts(): _amp_state.handle._is_active = False yield _amp_state.handle._is_active = True class AmpHandle(object): def __init__(self, loss_scale="dynamic", enable_caching=True, verbose=False): self._enable_caching = enable_caching self._verbose = verbose self._cache = dict() self._default_scaler = LossScaler(loss_scale) self._is_active = True self._all_wrappers = [] def is_active(self): return self._is_active @contextlib.contextmanager def _disable_casts(self): self._is_active = False yield self._is_active = True def wrap_optimizer(self, optimizer, num_loss=1): self._default_scaler = None return OptimWrapper(optimizer, self, num_loss) @contextlib.contextmanager def scale_loss(self, loss, optimizer): raise RuntimeError("The old Amp API is no longer supported. Please move to the new API, " "documented here: https://nvidia.github.io/apex/amp.html. Transition guide: " "https://nvidia.github.io/apex/amp.html#transition-guide-for-old-api-users") if not self.is_active(): yield loss return if self._default_scaler is None: raise RuntimeError( 'After calling `handle.wrap_optimizer()`, you must explicitly ' + 'use `optimizer.scale_loss(loss)`.') # TODO: this code block is duplicated here and `opt.py`. Unify. loss_scale = self._default_scaler.loss_scale() yield loss * loss_scale self._default_scaler.clear_overflow_state() self._default_scaler.unscale( master_params(optimizer), master_params(optimizer), loss_scale) should_skip = self._default_scaler.update_scale() if should_skip: optimizer_step = optimizer.step def skip_step(): maybe_print('Gradient overflow, skipping update') optimizer.step = optimizer_step optimizer.step = skip_step self._clear_cache() def _clear_cache(self): self._cache.clear() # Experimental support for saving / restoring uncasted versions of functions def _save_func(self, mod, fn, func): self._all_wrappers.append((mod, fn, func)) def _deactivate(self): for mod, fn, func in self._all_wrappers: utils.set_func(mod, fn, func) self._all_wrappers = [] @property def has_cache(self): return self._enable_caching @property def cache(self): return self._cache def remove_cache(self, param): if self.has_cache and param in self.cache: del self.cache[param] @property def verbose(self): return self._verbose class NoOpHandle(object): def is_active(self): return False @contextlib.contextmanager def _disable_casts(self): yield def wrap_optimizer(self, optimizer, num_loss=1): return OptimWrapper(optimizer, self, num_loss) @contextlib.contextmanager def scale_loss(self, loss, optimizer): yield loss @property def has_cache(self): return False @property def verbose(self): return False def _clear_cache(self): pass def _deactivate(self): pass
apex-master
apex/amp/handle.py
import collections.abc as container_abcs from types import MethodType import functools import sys import warnings import numpy as np import torch from ._amp_state import _amp_state, warn_or_err from .handle import disable_casts from .scaler import LossScaler from ._process_optimizer import _process_optimizer from apex.fp16_utils import convert_network from ..fp16_utils import FP16_Optimizer as FP16_Optimizer_general from ..contrib.optimizers import FP16_Optimizer as FP16_Optimizer_for_fused if torch.distributed.is_available(): from ..parallel import DistributedDataParallel as apex_DDP from ..parallel.LARC import LARC def to_type(dtype, t): if isinstance(t, torch.Tensor): if not t.is_cuda: # This should not be a hard error, since it may be legitimate. warnings.warn("An input tensor was not cuda.") # GANs require this. # if t.requires_grad: # warn_or_err("input data requires grad. Since input data is not a model parameter,\n" # "its gradients will not be properly allreduced by DDP.") if t.is_floating_point(): return t.to(dtype) return t else: # Trust the user's custom batch type, that's all I can do here. return t.to(dtype) # Modified from torch.optim.optimizer.py. This is a bit more general than casted_args in utils.py. def applier(value, fn): if isinstance(value, torch.Tensor): return fn(value) elif isinstance(value, str): return value elif isinstance(value, np.ndarray): return value elif hasattr(value, "to"): # Allow handling of custom batch classes return fn(value) elif isinstance(value, container_abcs.Mapping): return {applier(k, fn) : applier(v, fn) for k, v in value.items()} elif isinstance(value, container_abcs.Iterable): return type(value)(applier(v, fn) for v in value) else: # Do I want this to fire off even if someone chooses to pass something ordinary like # an int or float? May be more annoying than it's worth. # print("Warning: unrecognized type in applier. If your input data is a custom class, " # "provide it with a .to(dtype) method which converts its floating-point Tensors to dtype. " # "Amp will check for your custom to() and invoke it to cast the batch's " # "floating-point Tensors to the appropriate type. " # "Also, if your data is a custom class, it is your responsibility to ensure that " # "any Tensors you want to be cuda are already cuda." return value def check_models(models): for model in models: parallel_type = None if isinstance(model, torch.nn.parallel.DistributedDataParallel): parallel_type = "torch.nn.parallel.DistributedDataParallel" if ('apex_DDP' in sys.modules) and isinstance(model, apex_DDP): parallel_type = "apex.parallel.DistributedDataParallel" if isinstance(model, torch.nn.parallel.DataParallel): parallel_type = "torch.nn.parallel.DataParallel" if parallel_type is not None: raise RuntimeError("Incoming model is an instance of {}. ".format(parallel_type) + "Parallel wrappers should only be applied to the model(s) AFTER \n" "the model(s) have been returned from amp.initialize.") def check_params_fp32(models): for model in models: for name, param in model.named_parameters(): if param.is_floating_point(): if 'Half' in param.type(): warn_or_err("Found param {} with type {}, expected torch.cuda.FloatTensor.\n" "When using amp.initialize, you do not need to call .half() on your model\n" "before passing it, no matter what optimization level you choose.".format( name, param.type())) elif not param.is_cuda: warn_or_err("Found param {} with type {}, expected torch.cuda.FloatTensor.\n" "When using amp.initialize, you need to provide a model with parameters\n" "located on a CUDA device before passing it no matter what optimization level\n" "you chose. Use model.to('cuda') to use the default device.".format( name, param.type())) # Backward compatibility for PyTorch 0.4 if hasattr(model, 'named_buffers'): buf_iter = model.named_buffers() else: buf_iter = model._buffers for obj in buf_iter: if type(obj)==tuple: name, buf = obj else: name, buf = obj, buf_iter[obj] if buf.is_floating_point(): if 'Half' in buf.type(): warn_or_err("Found buffer {} with type {}, expected torch.cuda.FloatTensor.\n" "When using amp.initialize, you do not need to call .half() on your model\n" "before passing it, no matter what optimization level you choose.".format( name, buf.type())) elif not buf.is_cuda: warn_or_err("Found buffer {} with type {}, expected torch.cuda.FloatTensor.\n" "When using amp.initialize, you need to provide a model with buffers\n" "located on a CUDA device before passing it no matter what optimization level\n" "you chose. Use model.to('cuda') to use the default device.".format( name, buf.type())) def check_optimizers(optimizers): for optim in optimizers: bad_optim_type = None if isinstance(optim, FP16_Optimizer_general): bad_optim_type = "apex.fp16_utils.FP16_Optimizer" if isinstance(optim, FP16_Optimizer_for_fused): bad_optim_type = "apex.optimizers.FP16_Optimizer" if bad_optim_type is not None: raise RuntimeError("An incoming optimizer is an instance of {}. ".format(bad_optim_type) + "The optimizer(s) passed to amp.initialize() must be bare \n" "instances of either ordinary Pytorch optimizers, or Apex fused \n" "optimizers.\n") class O2StateDictHook(object): def __init__(self, fn): self.fn = fn def __call__(self, module, state_dict, prefix, local_metadata): for key in state_dict: param = state_dict[key] if 'Half' in param.type(): param = param.to(torch.float32) state_dict[key] = param def _initialize(models, optimizers, properties, num_losses=1, cast_model_outputs=None): from .amp import init as amp_init optimizers_was_list = False if isinstance(optimizers, torch.optim.Optimizer) or ('LARC' in globals() and isinstance(optimizers, LARC)): optimizers = [optimizers] elif optimizers is None: optimizers = [] elif isinstance(optimizers, list): optimizers_was_list = True check_optimizers(optimizers) else: check_optimizers([optimizers]) raise TypeError("optimizers must be either a single optimizer or a list of optimizers.") if isinstance(models, torch.nn.Module): models_was_list = False models = [models] elif isinstance(models, list): models_was_list = True else: raise TypeError("models must be either a single model or a list of models.") check_models(models) if not _amp_state.allow_incoming_model_not_fp32: check_params_fp32(models) # In the future, when FP16_Optimizer can be deprecated and master weights can # become an attribute, remember to stash master weights before casting the model. if properties.cast_model_type: if properties.keep_batchnorm_fp32: for model in models: convert_network(model, properties.cast_model_type) else: for model in models: model.to(properties.cast_model_type) input_caster = functools.partial(to_type, properties.cast_model_type) if cast_model_outputs is not None: output_caster = functools.partial(to_type, cast_model_outputs) else: output_caster = functools.partial(to_type, torch.float32) for model in models: # Patch the forward method to cast incoming data to the correct type, and # outgoing data to float32, so "the user never needs to call .half()." # I like writing things explicitly more than decorators. def patch_forward(old_fwd): def new_fwd(*args, **kwargs): output = old_fwd(*applier(args, input_caster), **applier(kwargs, input_caster)) return applier(output, output_caster) return new_fwd model.forward = patch_forward(model.forward) # State dict trick to recast any preexisting per-param state tensors for optimizer in optimizers: optimizer.load_state_dict(optimizer.state_dict()) # patch model.state_dict() to return float32 params for model in models: for module in model.modules(): module._register_state_dict_hook(O2StateDictHook(functools.partial(to_type, torch.float32))) elif cast_model_outputs is not None: output_caster = functools.partial(to_type, cast_model_outputs) for model in models: def patch_forward(old_fwd): def new_fwd(*args, **kwargs): output = old_fwd(*args, **kwargs) return applier(output, output_caster) return new_fwd model.forward = patch_forward(model.forward) for i, optimizer in enumerate(optimizers): optimizers[i] = _process_optimizer(optimizer, properties) _amp_state.loss_scalers = [] for _ in range(num_losses): _amp_state.loss_scalers.append(LossScaler(properties.loss_scale, min_loss_scale=_amp_state.min_loss_scale, max_loss_scale=_amp_state.max_loss_scale)) if properties.patch_torch_functions: # handle is unused here. It's accessible later through a global value anyway. handle = amp_init(loss_scale=properties.loss_scale, verbose=(_amp_state.verbosity == 2)) for optimizer in optimizers: # Disable Amp casting for the optimizer step, because it should only be # applied to FP32 master params anyway. def patch_step(old_step): def new_step(self, *args, **kwargs): with disable_casts(): output = old_step(*args, **kwargs) return output return new_step optimizer.step = MethodType(patch_step(optimizer.step), optimizer) if optimizers_was_list: if models_was_list: return models, optimizers else: return models[0], optimizers else: if models_was_list: if len(optimizers) == 0: return models else: return models, optimizers[0] else: if len(optimizers) == 0: return models[0] else: return models[0], optimizers[0]
apex-master
apex/amp/_initialize.py
import functools import itertools import torch from . import compat, rnn_compat, utils, wrap from .handle import AmpHandle, NoOpHandle from .lists import functional_overrides, torch_overrides, tensor_overrides from ._amp_state import _amp_state from .frontend import * _DECORATOR_HANDLE = None _USER_CAST_REGISTRY = set() _USER_PROMOTE_REGISTRY = set() def _decorator_helper(orig_fn, cast_fn, wrap_fn): def wrapper(*args, **kwargs): handle = _DECORATOR_HANDLE if handle is None or not handle.is_active(): return orig_fn(*args, **kwargs) inner_cast_fn = utils.verbosify(cast_fn, orig_fn.__name__, handle.verbose) return wrap_fn(orig_fn, inner_cast_fn, handle)(*args, **kwargs) return wrapper # Decorator form def half_function(fn): from apex import deprecated_warning deprecated_warning("apex.amp is deprecated and will be removed by the end of February 2023. Use [PyTorch AMP](https://pytorch.org/docs/stable/amp.html)") wrap_fn = functools.partial(wrap.make_cast_wrapper, try_caching=True) return _decorator_helper(fn, utils.maybe_half, wrap_fn) def float_function(fn): from apex import deprecated_warning deprecated_warning("apex.amp is deprecated and will be removed by the end of February 2023. Use [PyTorch AMP](https://pytorch.org/docs/stable/amp.html)") wrap_fn = functools.partial(wrap.make_cast_wrapper, try_caching=False) return _decorator_helper(fn, utils.maybe_float, wrap_fn) def promote_function(fn): from apex import deprecated_warning deprecated_warning("apex.amp is deprecated and will be removed by the end of February 2023. Use [PyTorch AMP](https://pytorch.org/docs/stable/amp.html)") wrap_fn = functools.partial(wrap.make_promote_wrapper) return _decorator_helper(fn, utils.maybe_float, wrap_fn) # Registry form def register_half_function(module, name): if not hasattr(module, name): raise ValueError('No function named {} in module {}.'.format( name, module)) _USER_CAST_REGISTRY.add((module, name, utils.maybe_half)) def register_float_function(module, name): if not hasattr(module, name): raise ValueError('No function named {} in module {}.'.format( name, module)) _USER_CAST_REGISTRY.add((module, name, utils.maybe_float)) def register_promote_function(module, name): if not hasattr(module, name): raise ValueError('No function named {} in module {}.'.format( name, module)) _USER_PROMOTE_REGISTRY.add((module, name)) # Top-level function to insert _all_ the hooks. def init(enabled=True, loss_scale="dynamic", enable_caching=True, verbose=False, allow_banned=False): global _DECORATOR_HANDLE if not enabled: handle = NoOpHandle() _DECORATOR_HANDLE = handle return handle handle = AmpHandle(loss_scale, enable_caching, verbose) # 0) Force-{fp16, fp32} for user-annotated functions for mod, fn, cast_fn in _USER_CAST_REGISTRY: try_caching = (cast_fn == utils.maybe_half) wrap.cached_cast(mod, fn, cast_fn, handle, try_caching, verbose) _USER_CAST_REGISTRY.clear() # 0.5) Force-promote for user-annotated functions for mod, fn in _USER_PROMOTE_REGISTRY: wrap.promote(mod, fn, handle, verbose) _USER_PROMOTE_REGISTRY.clear() # 1) Force-{fp16, fp32} on white- / black-list functions override_modules = [functional_overrides, torch_overrides, tensor_overrides] cast_table = [('FP16_FUNCS', utils.maybe_half), ('FP32_FUNCS', utils.maybe_float)] for module, (list_name, cast_fn) in itertools.product(override_modules, cast_table): for fn in getattr(module, list_name): try_caching = (cast_fn == utils.maybe_half) wrap.cached_cast(module.MODULE, fn, cast_fn, handle, try_caching, verbose) # 1.5) Pre-0.4, put the blacklist methods on HalfTensor and whitelist # methods on FloatTensor, since they're distinct types. if compat.tensor_is_float_tensor(): for fn in tensor_overrides.FP16_FUNCS: wrap.cached_cast(torch.cuda.FloatTensor, fn, utils.maybe_half, handle, try_caching=True, verbose=verbose) for fn in tensor_overrides.FP32_FUNCS: wrap.cached_cast(torch.cuda.HalfTensor, fn, utils.maybe_float, handle, try_caching=False, verbose=verbose) # 2) Enable type-promotion on multi-arg functions and methods. # NB: special handling for sequence fns (e.g. `torch.cat`). promote_modules = [torch_overrides, tensor_overrides] promote_table = [('CASTS', wrap.promote), ('SEQUENCE_CASTS', wrap.sequence_promote)] for promote_mod, (list_name, promote_fn) in itertools.product(promote_modules, promote_table): for fn in getattr(promote_mod, list_name): promote_fn(promote_mod.MODULE, fn, handle, verbose) # 2.5) Pre-0.4, add blacklist methods directly to HalfTensor and FloatTensor types if compat.tensor_is_float_tensor(): for cls, (list_name, promote_fn) in itertools.product([torch.cuda.FloatTensor, torch.cuda.HalfTensor], promote_table): for fn in getattr(tensor_overrides, list_name): promote_fn(cls, fn, handle, verbose) # 3) For any in-place version of a blacklist function, error if any input is fp16. # NB: this is overly conservative. for fn in utils.as_inplace(torch_overrides.FP32_FUNCS): wrap.err_if_any_half(torch_overrides.MODULE, fn, handle) # 3.5) For any in-place blacklist method, error if called on fp16 tensor for fn in utils.as_inplace(tensor_overrides.FP32_FUNCS): wrap.err_if_arg0_half(tensor_overrides.MODULE, fn, handle, verbose) if compat.tensor_is_float_tensor(): wrap.err_if_arg0_half(torch.cuda.HalfTensor, fn, handle, verbose) # 4) For other in-place methods, match the type of self tensor for fn in utils.as_inplace(itertools.chain( tensor_overrides.FP16_FUNCS, tensor_overrides.CASTS)): wrap.promote_match_arg0(tensor_overrides.MODULE, fn, handle, verbose) if compat.tensor_is_float_tensor(): wrap.promote_match_arg0(torch.cuda.HalfTensor, fn, handle, verbose) wrap.promote_match_arg0(torch.cuda.FloatTensor, fn, handle, verbose) # 5) RNNs + RNN cells are whitelisted specially if rnn_compat.has_old_rnns(): wrap.rnn_cast(torch.nn.backends.thnn.backend, 'RNN', handle, verbose) if not rnn_compat.has_old_rnns(): # Patch in our own indirection of `_VF` in modules/rnn s.t. it is mutable. torch.nn.modules.rnn._VF = rnn_compat.VariableFunctionsShim() # Wrap all the rnns for x in rnn_compat.RNN_NAMES: wrap.new_rnn_cast(x.upper(), handle, verbose) # Wrap all the RNN cells rnn_compat.whitelist_rnn_cells(handle, verbose) # 6) Place error+print message on banned functions. # Or, if allow_banned, then cast to FP32. for fn, err_msg in functional_overrides.BANNED_FUNCS: if allow_banned: wrap.cached_cast(functional_overrides.MODULE, fn, utils.maybe_float, handle, try_caching=True, verbose=verbose) else: wrap.err_if_any_half(functional_overrides.MODULE, fn, handle, err_msg) _DECORATOR_HANDLE = handle _amp_state.handle = handle return handle
apex-master
apex/amp/amp.py
from collections import OrderedDict import torch from ._initialize import _initialize from ._amp_state import _amp_state, warn_or_err, maybe_print class Properties(object): """ This class has two purposes: to establish a set of default properties, and to route setting of these attributes through __setattr__ so that (in theory) they can be checked for consistency with other existing args. """ def __init__(self): self.options = { "enabled" : False, "opt_level" : None, "cast_model_type" : None, "patch_torch_functions" : False, "keep_batchnorm_fp32" : None, "master_weights" : None, "loss_scale" : 1.0, # Reserved for future functionality # "fused_optimizer" : False, # "enable_ddp_interop" : False, } """ This function allows updating several options at a time without routing through __setattr__ checks, to avoid "you can't get there from here" scenarios. Currently not intended to be exposed; users are expected to select an opt_level and apply consistent modifications. """ def _update_options_dict(self, new_options): for k, v in new_options: if k in self.options: self.options[k] = v else: raise ValueError("Tried to set unexpected option {}".format(k)) """ The members of "options" are not direct attributes of self, so access attempts will roll down to __getattr__. This borrows from the logic in torch.nn.Module. """ def __getattr__(self, name): if "options" in self.__dict__: options = self.__dict__["options"] if name in options: return options[name] raise AttributeError("'{}' object has no attribute '{}'".format( type(self).__name__, name)) def __setattr__(self, name, value): if "options" in self.__dict__: if name in self.options: # print("setting {} {}".format(name, value)) if name == "cast_model_type": if self.opt_level == "O1" and value is not None: if value is not False: if value is not torch.float32: warn_or_err("O1 inserts casts around Torch functions rather than " "model weights, so with O1, the model weights themselves " "should remain FP32. If you wish to cast the model to a " "different type, use opt_level='O2' or 'O3'. " + "cast_model_type was {}".format(value)) self.options[name] = value elif name == "patch_torch_functions": if self.opt_level != "O1" and value: warn_or_err("Currently, patch_torch_functions=True should only be set by " "selecting opt_level='O1'.") self.options[name] = value elif name == "keep_batchnorm_fp32": if self.opt_level == "O1" and value is not None: warn_or_err("With opt_level O1, batchnorm functions are automatically patched " "to run in FP32, so keep_batchnorm_fp32 should be None." + " keep_batchnorm_fp32 was {}".format(value)) if value == "False": self.options[name] = False elif value == "True": self.options[name] = True else: assert (value is True or value is False or value is None),\ "keep_batchnorm_fp32 must be a boolean, the string 'True' or 'False', "\ "or None, found keep_batchnorm_fp32={}".format(value) self.options[name] = value elif name == "master_weights": if self.opt_level == "O1" and value is not None: warn_or_err("It doesn't make sense to use master_weights with O1. " "With O1, your model weights themselves should be FP32.") self.options[name] = value elif name == "loss_scale": if value == "dynamic": self.options[name] = value else: self.options[name] = float(value) else: self.options[name] = value else: super(Properties, self).__setattr__(name, value) """ O0-O3 are convenience wrappers to establish defaults for typically used mixed precision options. """ class O3: brief = "O3: Pure FP16 training." more = "Calls .half() on your model, converting the entire model to FP16.\n"\ "A casting operation is also inserted to cast incoming Tensors to FP16,\n"\ "so you don't need to change your data pipeline.\n"\ "This mode is useful for establishing a performance ceiling.\n"\ "It's also possible training may 'just work' in this mode.\n"\ "If not, try other optimization levels." def __call__(self, properties): properties.enabled = True properties.opt_level = "O3" properties.cast_model_type = torch.float16 properties.patch_torch_functions = False properties.keep_batchnorm_fp32 = False properties.master_weights = False properties.loss_scale = 1.0 # properties.fused_optimizer = False # properties.enable_ddp_interop = False return properties # modified in place so this isn't really necessary class O2: brief = "O2: FP16 training with FP32 batchnorm and FP32 master weights.\n" more = "Calls .half() on your model, converting the entire model (except for batchnorms)\n"\ "to FP16. Batchnorms are retained in FP32 for additional stability.\n"\ "The forward pass is patched to cast incoming Tensors to FP16, so you don't need to change\n"\ "your data pipeline.\n"\ "O2 creates FP32 master weights outside the model and patches any optimizers to update\n"\ "these master weights, then copy the master weights into the FP16 model weights.\n"\ "Master weights can also improve convergence and stability." def __call__(self, properties): properties.enabled = True properties.opt_level = "O2" properties.cast_model_type = torch.float16 properties.patch_torch_functions = False properties.keep_batchnorm_fp32 = True properties.master_weights = True properties.loss_scale = "dynamic" # properties.fused_optimizer = False # properties.enable_ddp_interop = False return properties # modified in place so this isn't really necessary class O1: brief = "O1: Insert automatic casts around Pytorch functions and Tensor methods.\n" more = "The type of your model's weights is not altered. However, internally,\n"\ "Pytorch functions are patched to cast any Tensor Core-friendly ops to FP16 for speed,\n"\ "while operations that might benefit from the additional stability of FP32 are patched\n"\ "to cast their inputs to fp32.\n"\ "O1 is the safest way to try mixed precision training, and is recommended when\n"\ "trying mixed precision training for the first time." def __call__(self, properties): properties.enabled = True properties.opt_level = "O1" properties.cast_model_type = None properties.patch_torch_functions = True properties.keep_batchnorm_fp32 = None properties.master_weights = None properties.loss_scale = "dynamic" # properties.fused_optimizer = False # properties.enable_ddp_interop = False return properties # modified in place so this isn't really necessary class O0: brief = "O0: Pure FP32 training.\n" more = "Your models are checked to make sure parameters are FP32, but otherwise the\n"\ "types of weights and internal Pytorch operations are not altered. This mode disables any\n"\ "FP16 arithmetic, although other optimizations like DDP interop may still be requested.\n" def __call__(self, properties): properties.enabled = True properties.opt_level = "O0" properties.cast_model_type = torch.float32 properties.patch_torch_functions = False properties.keep_batchnorm_fp32 = None properties.master_weights = False properties.loss_scale = 1.0 # properties.fused_optimizer = False # properties.enable_ddp_interop = False return properties # modified in place so this isn't really necessary opt_levels = {"O3": O3(), "O2": O2(), "O1": O1(), "O0": O0()} # allow user to directly pass Properties struct as well? def initialize( models, optimizers=None, enabled=True, opt_level="O1", cast_model_type=None, patch_torch_functions=None, keep_batchnorm_fp32=None, master_weights=None, loss_scale=None, cast_model_outputs=None, num_losses=1, verbosity=1, min_loss_scale=None, max_loss_scale=2.**24 ): """ Initialize your models, optimizers, and the Torch tensor and functional namespace according to the chosen ``opt_level`` and overridden properties, if any. ``amp.initialize`` should be called **after** you have finished constructing your model(s) and optimizer(s), but **before** you send your model through any DistributedDataParallel wrapper. See `Distributed training`_ in the Imagenet example. Currently, ``amp.initialize`` should only be called **once**, although it can process an arbitrary number of models and optimizers (see the corresponding `Advanced Amp Usage topic`_). If you think your use case requires ``amp.initialize`` to be called more than once, `let us know`_. Any property keyword argument that is not ``None`` will be interpreted as a manual override. To prevent having to rewrite anything else in your script, name the returned models/optimizers to replace the passed models/optimizers, as in the code sample below. Args: models (torch.nn.Module or list of torch.nn.Modules): Models to modify/cast. optimizers (optional, torch.optim.Optimizer or list of torch.optim.Optimizers): Optimizers to modify/cast. REQUIRED for training, optional for inference. enabled (bool, optional, default=True): If False, renders all Amp calls no-ops, so your script should run as if Amp were not present. opt_level (str, optional, default="O1"): Pure or mixed precision optimization level. Accepted values are "O0", "O1", "O2", and "O3", explained in detail above. cast_model_type (``torch.dtype``, optional, default=None): Optional property override, see above. patch_torch_functions (bool, optional, default=None): Optional property override. keep_batchnorm_fp32 (bool or str, optional, default=None): Optional property override. If passed as a string, must be the string "True" or "False". master_weights (bool, optional, default=None): Optional property override. loss_scale (float or str, optional, default=None): Optional property override. If passed as a string, must be a string representing a number, e.g., "128.0", or the string "dynamic". cast_model_outputs (torch.dtype, optional, default=None): Option to ensure that the outputs of your model(s) are always cast to a particular type regardless of ``opt_level``. num_losses (int, optional, default=1): Option to tell Amp in advance how many losses/backward passes you plan to use. When used in conjunction with the ``loss_id`` argument to ``amp.scale_loss``, enables Amp to use a different loss scale per loss/backward pass, which can improve stability. See "Multiple models/optimizers/losses" under `Advanced Amp Usage`_ for examples. If ``num_losses`` is left to 1, Amp will still support multiple losses/backward passes, but use a single global loss scale for all of them. verbosity (int, default=1): Set to 0 to suppress Amp-related output. min_loss_scale (float, default=None): Sets a floor for the loss scale values that can be chosen by dynamic loss scaling. The default value of None means that no floor is imposed. If dynamic loss scaling is not used, `min_loss_scale` is ignored. max_loss_scale (float, default=2.**24): Sets a ceiling for the loss scale values that can be chosen by dynamic loss scaling. If dynamic loss scaling is not used, `max_loss_scale` is ignored. Returns: Model(s) and optimizer(s) modified according to the ``opt_level``. If either the ``models`` or ``optimizers`` args were lists, the corresponding return value will also be a list. Permissible invocations:: model, optim = amp.initialize(model, optim,...) model, [optim1, optim2] = amp.initialize(model, [optim1, optim2],...) [model1, model2], optim = amp.initialize([model1, model2], optim,...) [model1, model2], [optim1, optim2] = amp.initialize([model1, model2], [optim1, optim2],...) # This is not an exhaustive list of the cross product of options that are possible, # just a set of examples. model, optim = amp.initialize(model, optim, opt_level="O0") model, optim = amp.initialize(model, optim, opt_level="O0", loss_scale="dynamic"|128.0|"128.0") model, optim = amp.initialize(model, optim, opt_level="O1") # uses "loss_scale="dynamic" default model, optim = amp.initialize(model, optim, opt_level="O1", loss_scale=128.0|"128.0") model, optim = amp.initialize(model, optim, opt_level="O2") # uses "loss_scale="dynamic" default model, optim = amp.initialize(model, optim, opt_level="O2", loss_scale=128.0|"128.0") model, optim = amp.initialize(model, optim, opt_level="O2", keep_batchnorm_fp32=True|False|"True"|"False") model, optim = amp.initialize(model, optim, opt_level="O3") # uses loss_scale=1.0 default model, optim = amp.initialize(model, optim, opt_level="O3", loss_scale="dynamic"|128.0|"128.0") model, optim = amp.initialize(model, optim, opt_level="O3", keep_batchnorm_fp32=True|False|"True"|"False") The `Imagenet example`_ demonstrates live use of various opt_levels and overrides. .. _`Distributed training`: https://github.com/NVIDIA/apex/tree/master/examples/imagenet#distributed-training .. _`Imagenet example`: https://github.com/NVIDIA/apex/tree/master/examples/imagenet .. _`Advanced Amp Usage`: https://nvidia.github.io/apex/advanced.html .. _`Advanced Amp Usage topic`: https://nvidia.github.io/apex/advanced.html#multiple-models-optimizers-losses .. _`let us know`: https://github.com/NVIDIA/apex/issues """ from apex import deprecated_warning deprecated_warning("apex.amp is deprecated and will be removed by the end of February 2023. Use [PyTorch AMP](https://pytorch.org/docs/stable/amp.html)") _amp_state.opt_properties = Properties() _amp_state.verbosity = verbosity if not enabled: if optimizers is None: return models else: return models, optimizers if not torch.backends.cudnn.enabled: raise RuntimeError( "Amp requires torch.backends.cudnn.enabled = True") if opt_level not in opt_levels: raise RuntimeError( "Unexpected optimization level {}. ".format(opt_level) + "Options are 'O0', 'O1', 'O2', 'O3'. Note that in `O0`, `O1`, etc., the prefix O is the letter O, " + "not the number zero.") else: _amp_state.opt_properties = opt_levels[opt_level](_amp_state.opt_properties) maybe_print("Selected optimization level {}".format(opt_levels[opt_level].brief), True) maybe_print("Defaults for this optimization level are:", True) for k, v in _amp_state.opt_properties.options.items(): maybe_print("{:22} : {}".format(k, v), True) _amp_state.min_loss_scale = min_loss_scale _amp_state.max_loss_scale = max_loss_scale maybe_print("Processing user overrides (additional kwargs that are not None)...", True) # I chose to have the keyword arguments listed directly in the argument list, # instead of **kwargs, so I can't use kwargs.items() here. if enabled is not None: _amp_state.opt_properties.enabled = enabled if opt_level is not None: _amp_state.opt_properties.opt_level = opt_level if cast_model_type is not None: _amp_state.opt_properties.cast_model_type = cast_model_type if patch_torch_functions is not None: _amp_state.opt_properties.patch_torch_functions = patch_torch_functions if keep_batchnorm_fp32 is not None: _amp_state.opt_properties.keep_batchnorm_fp32 = keep_batchnorm_fp32 if master_weights is not None: _amp_state.opt_properties.master_weights = master_weights if loss_scale is not None: _amp_state.opt_properties.loss_scale = loss_scale maybe_print("After processing overrides, optimization options are:", True) for k, v in _amp_state.opt_properties.options.items(): maybe_print("{:22} : {}".format(k, v), True) return _initialize(models, optimizers, _amp_state.opt_properties, num_losses, cast_model_outputs) def state_dict(destination=None): if destination is None: destination = OrderedDict() for idx, loss_scaler in enumerate(_amp_state.loss_scalers): destination['loss_scaler%d' % idx] = { 'loss_scale': loss_scaler.loss_scale(), 'unskipped': loss_scaler._unskipped, } return destination def load_state_dict(state_dict): # Check if state_dict containes the same number of loss_scalers as current setup if len(state_dict) != len(_amp_state.loss_scalers): print('Warning: state_dict contains {} entries, while {} loss_scalers are used'.format( len(state_dict), len(_amp_state.loss_scalers))) state_dict = state_dict.copy() nb_loss_scalers = len(_amp_state.loss_scalers) unexpected_keys = [] # Initialize idx outside, since unexpected_keys will increase it if enumerate is used idx = 0 for key in state_dict: if 'loss_scaler' not in key: unexpected_keys.append(key) else: if idx > (nb_loss_scalers - 1): print('Skipping loss_scaler[{}], since num_losses was set to {}'.format( idx, nb_loss_scalers)) break _amp_state.loss_scalers[idx]._loss_scale = state_dict[key]['loss_scale'] _amp_state.loss_scalers[idx]._unskipped = state_dict[key]['unskipped'] idx += 1 if len(unexpected_keys) > 0: raise RuntimeError( 'Error(s) in loading state_dict. Unexpected key(s) in state_dict: {}. '.format( ', '.join('"{}"'.format(k) for k in unexpected_keys))) # TODO: is this necessary/useful? # def check_option_consistency(enabled=True, # opt_level=None, # cast_model_type=None, # patch_torch_functions=None, # keep_batchnorm_fp32=None, # master_weights=None, # loss_scale=None, # enable_ddp_interop=None, # hard_override=False): # """ # Utility function that enables users to quickly check if the option combination they intend # to use is permitted. ``check_option_consistency`` does not require models or optimizers # to be constructed, and can be called at any point in the script. ``check_option_consistency`` # is totally self-contained; it does not set any amp global state or affect anything outside # of itself. # """ # # if not enabled: # return # # if opt_level not in opt_levels: # raise RuntimeError("Unexpected optimization level. Options are 'O0', 'O1', 'O2', 'O3'.") # else: # opt_properties = opt_levels[opt_level](Properties()) # print("Selected optimization level {}", opt_levels[opt_level].brief) # print("Defaults for this optimization level are:") # for k, v in opt_properties.options: # print("{:22} : {}".format(k, v)) # # print("Processing user overrides (additional kwargs that are not None)...") # for k, v in kwargs: # if k not in _amp_state.opt_properties.options: # raise RuntimeError("Unexpected kwarg {}".format(k)) # if v is not None: # setattr(opt_properties, k, v) # # print("After processing overrides, optimization options are:") # for k, v in opt_properties.options: # print("{:22} : {}".format(k, v))
apex-master
apex/amp/frontend.py
from .amp import init, half_function, float_function, promote_function,\ register_half_function, register_float_function, register_promote_function from .handle import scale_loss, disable_casts from .frontend import initialize, state_dict, load_state_dict from ._amp_state import master_params, _amp_state
apex-master
apex/amp/__init__.py
import torch from ..multi_tensor_apply import multi_tensor_applier from ._amp_state import _amp_state, master_params, maybe_print from itertools import product def scale_check_overflow_python(model_grad, master_grad, scale, check_overflow=False): # Exception handling for 18.04 compatibility if check_overflow: cpu_sum = float(model_grad.float().sum()) if cpu_sum == float('inf') or cpu_sum == -float('inf') or cpu_sum != cpu_sum: return True if master_grad is not model_grad: # copy_ probably internally short-circuits this master_grad.copy_(model_grad) if scale != 1.0: master_grad.mul_(scale) return False def axpby_check_overflow_python(model_grad, stashed_grad, master_grad, a, b, check_overflow=False): # Exception handling for 18.04 compatibility if check_overflow: cpu_sum = float(model_grad.float().sum()) if cpu_sum == float('inf') or cpu_sum == -float('inf') or cpu_sum != cpu_sum: return True # if master_grad is not model_grad: # copy_ probably internally short-circuits this # master_grad.copy_(model_grad) assert stashed_grad.dtype == master_grad.dtype converted_model_grad = model_grad.data.to(master_grad.dtype) master_grad.data = a*converted_model_grad.data + b*stashed_grad.data return False class LossScaler(object): warned_no_fused_kernel = False warned_unscaling_non_fp32_grad = False has_fused_kernel = False def __init__(self, loss_scale, init_scale=2.**16, scale_factor=2., scale_window=2000, min_loss_scale=None, max_loss_scale=2.**24): if loss_scale == "dynamic": self.dynamic = True self._loss_scale = min(max_loss_scale, init_scale) else: self.dynamic = False self._loss_scale = loss_scale self._max_loss_scale = max_loss_scale self._min_loss_scale = min_loss_scale self._scale_seq_len = scale_window self._unskipped = 0 self._has_overflow = False self._overflow_buf = torch.cuda.IntTensor([0]) if multi_tensor_applier.available: import amp_C LossScaler.has_fused_kernel = multi_tensor_applier.available LossScaler.multi_tensor_scale_cuda = amp_C.multi_tensor_scale LossScaler.multi_tensor_axpby_cuda = amp_C.multi_tensor_axpby else: if not LossScaler.warned_no_fused_kernel: maybe_print( "Warning: multi_tensor_applier fused unscale kernel is unavailable, " "possibly because apex was installed without --cuda_ext --cpp_ext. " "Using Python fallback. Original ImportError was: " + repr(multi_tensor_applier.import_err), True) LossScaler.has_fused_kernel = False LossScaler.warned_no_fused_kernel = True def loss_scale(self): return self._loss_scale def unscale_python(self, model_grads, master_grads, scale): for model, master in zip(model_grads, master_grads): if model is not None: if not LossScaler.warned_unscaling_non_fp32_grad: if master.dtype != torch.float32: maybe_print( "Attempting to unscale a grad with type {} ".format(master.type()) + "Unscaling non-fp32 grads may indicate an error. " "When using Amp, you don't need to call .half() on your model.") LossScaler.warned_unscaling_non_fp32_grad = True self._has_overflow = scale_check_overflow_python(model, master, 1./scale, self.dynamic) if self._has_overflow and self.dynamic: break # unused_scale keeps some of the old API alive for hopefully a short time. def unscale(self, model_grads, master_grads, unused_scale, models_are_masters=False, scale_override=None): if self._has_overflow: return scale = self._loss_scale if scale_override is not None: scale = scale_override if scale == 1.0 and models_are_masters and not self.dynamic: return if LossScaler.has_fused_kernel: # if (not LossScaler.warned_unscaling_non_fp32_grad # and master_grads[0].dtype == torch.float16): # print("Warning: unscaling grads that are not FP32. " # "Unscaling non-fp32 grads may indicate an error. " # "When using Amp, you don't need to call .half() on your model.") # # Setting this to True unconditionally allows the possibility of an escape # # if never-before-seen non-fp32 grads are created in some later iteration. # LossScaler.warned_unscaling_non_fp32_grad = True multi_tensor_applier(LossScaler.multi_tensor_scale_cuda, self._overflow_buf, [model_grads, master_grads], 1./scale) else: self.unscale_python(model_grads, master_grads, scale) # Defer to update_scale # If the fused kernel is available, we only need one D2H memcopy and sync. # if LossScaler.has_fused_kernel and self.dynamic and not self._has_overflow: # self._has_overflow = self._overflow_buf.item() def unscale_with_stashed_python(self, model_grads, stashed_master_grads, master_grads, a, b): for model, stashed, master in zip(model_grads, stashed_master_grads, master_grads): if model is None and stashed is None: continue else: if not LossScaler.warned_unscaling_non_fp32_grad: if master.dtype != torch.float32: maybe_print( "Attempting to unscale a grad with type {} ".format(master.type()) + "Unscaling non-fp32 grads may indicate an error. " "When using Amp, you don't need to call .half() on your model.") LossScaler.warned_unscaling_non_fp32_grad = True self._has_overflow = axpby_check_overflow_python(model, stashed, master, a, b, self.dynamic) if self._has_overflow and self.dynamic: break def unscale_with_stashed(self, model_grads, stashed_master_grads, master_grads, scale_override=None): if self._has_overflow: return grads_have_scale, stashed_have_scale, out_scale = self._loss_scale, 1.0, 1.0 if scale_override is not None: grads_have_scale, stashed_have_scale, out_scale = scale_override if LossScaler.has_fused_kernel: if (not LossScaler.warned_unscaling_non_fp32_grad and master_grads[0].dtype == torch.float16): print("Warning: unscaling grads that are not FP32. " "Unscaling non-fp32 grads may indicate an error. " "When using Amp, you don't need to call .half() on your model.") # Setting this to True unconditionally allows the possibility of an escape # if never-before-seen non-fp32 grads are created in some later iteration. LossScaler.warned_unscaling_non_fp32_grad = True multi_tensor_applier(LossScaler.multi_tensor_axpby_cuda, self._overflow_buf, [model_grads, stashed_master_grads, master_grads], out_scale/grads_have_scale, # 1./scale, out_scale/stashed_have_scale, # 1.0, 0) # check only arg 0, aka the incoming model grads, for infs else: self.unscale_with_stashed_python(model_grads, stashed_master_grads, master_grads, out_scale/grads_have_scale, out_scale/stashed_have_scale) # Defer to update_scale # If the fused kernel is available, we only need one D2H memcopy and sync. # if LossScaler.has_fused_kernel and self.dynamic and not self._has_overflow: # self._has_overflow = self._overflow_buf.item() def clear_overflow_state(self): self._has_overflow = False if self.has_fused_kernel: self._overflow_buf.zero_() # Separate so unscale() can be called more that once before updating. def update_scale(self): # If the fused kernel is available, we only need one D2H memcopy and sync. if LossScaler.has_fused_kernel and self.dynamic and not self._has_overflow: self._has_overflow = self._overflow_buf.item() if self._has_overflow and self.dynamic: should_skip = True if(self._min_loss_scale): self._loss_scale = max(self._min_loss_scale, self._loss_scale/2.) else: self._loss_scale = self._loss_scale/2. self._unskipped = 0 else: should_skip = False self._unskipped += 1 if self._unskipped == self._scale_seq_len and self.dynamic: self._loss_scale = min(self._max_loss_scale, self._loss_scale*2.) self._unskipped = 0 return should_skip
apex-master
apex/amp/scaler.py
VERSION = (0, 1, 0) __version__ = '.'.join(map(str, VERSION))
apex-master
apex/amp/__version__.py
import contextlib import warnings from .scaler import LossScaler, master_params from ._amp_state import maybe_print import numpy as np class OptimWrapper(object): def __init__(self, optimizer, amp_handle, num_loss): self._optimizer = optimizer self._amp_handle = amp_handle self._num_loss = num_loss self._loss_idx = 0 self._skip_next = [False] * num_loss self._loss_scaler = [LossScaler('dynamic') for _ in range(num_loss)] @contextlib.contextmanager def scale_loss(self, loss): if not self._amp_handle.is_active(): yield loss return # When there are multiple losses per-optimizer, we need # to save out current grad accumulation, since we won't be # able to unscale this particulare loss once the grads are # all mixed together. cached_grads = [] if self._loss_idx > 0: for p in master_params(self._optimizer): if p.grad is not None: cached_grads.append(p.grad.data.detach().clone()) else: cached_grads.append(None) self._optimizer.zero_grad() loss_scale = self._cur_loss_scaler().loss_scale() yield loss * loss_scale self._cur_loss_scaler().clear_overflow_state() self._cur_loss_scaler().unscale( master_params(self._optimizer), master_params(self._optimizer), loss_scale) self._skip_next[self._loss_idx] = self._cur_loss_scaler().update_scale() self._loss_idx += 1 if len(cached_grads) > 0: for p, cached_grad in zip(master_params(self._optimizer), cached_grads): if cached_grad is not None: p.grad.data.add_(cached_grad) cached_grads = [] def _cur_loss_scaler(self): assert 0 <= self._loss_idx < self._num_loss return self._loss_scaler[self._loss_idx] def step(self, closure=None): if not self._amp_handle.is_active(): return self._optimizer.step(closure=closure) self._loss_idx = 0 for group in self._optimizer.param_groups: for p in group['params']: self._amp_handle.remove_cache(p) if closure is not None: raise NotImplementedError( 'The `closure` argument is unsupported by the amp ' + 'optimizer wrapper.') if any(self._skip_next): maybe_print('Gradient overflow, skipping update') self._skip_next = [False] * self._num_loss else: return self._optimizer.step(closure=closure) # Forward any attribute lookups def __getattr__(self, attr): return getattr(self._optimizer, attr) # Forward all torch.optim.Optimizer methods def __getstate__(self): return self._optimizer.__getstate__() def __setstate__(self): return self._optimizer.__setstate__() def __repr__(self): return self._optimizer.__repr__() def state_dict(self): return self._optimizer.state_dict() def load_state_dict(self, state_dict): return self._optimizer.load_state_dict(state_dict) def zero_grad(self): return self._optimizer.zero_grad() def add_param_group(self, param_group): return self._optimizer.add_param_group(param_group)
apex-master
apex/amp/opt.py
# This is a "header object" that allows different amp modules to communicate. # I'm a C++ guy, not a python guy. I decided this approach because it seemed most C++-like. # But apparently it's ok: # http://effbot.org/pyfaq/how-do-i-share-global-variables-across-modules.htm import torch class AmpState(object): def __init__(self): self.hard_override=False self.allow_incoming_model_not_fp32 = False self.verbosity=1 # Attribute stash. Could also just stash things as global module attributes. _amp_state = AmpState() def warn_or_err(msg): if _amp_state.hard_override: print("Warning: " + msg) else: raise RuntimeError(msg) # I'm not sure if allowing hard_override is a good idea. # + " If you're sure you know what you're doing, supply " + # "hard_override=True to amp.initialize.") def maybe_print(msg, rank0=False): distributed = torch.distributed.is_available() and \ torch.distributed.is_initialized() and \ torch.distributed.get_world_size() > 1 if _amp_state.verbosity > 0: if rank0: if distributed: if torch.distributed.get_rank() == 0: print(msg) else: print(msg) else: print(msg) # def iter_params(param_groups): # for group in param_groups: # for p in group['params']: # yield p def master_params(optimizer): """ Generator expression that iterates over the params owned by ``optimizer``. Args: optimizer: An optimizer previously returned from ``amp.initialize``. """ for group in optimizer.param_groups: for p in group['params']: yield p
apex-master
apex/amp/_amp_state.py
from . import compat import functools import itertools import torch def is_cuda_enabled(): return torch.version.cuda is not None def get_cuda_version(): return tuple(int(x) for x in torch.version.cuda.split('.')) def is_fp_tensor(x): if is_nested(x): # Fast-fail version of all(is_fp_tensor) for y in x: if not is_fp_tensor(y): return False return True return compat.is_tensor_like(x) and compat.is_floating_point(x) def is_nested(x): return isinstance(x, tuple) or isinstance(x, list) def should_cache(x): if is_nested(x): # Fast-fail version of all(should_cache) for y in x: if not should_cache(y): return False return True return isinstance(x, torch.nn.parameter.Parameter) and \ type_string(x) == 'FloatTensor' def collect_fp_tensor_types(args, kwargs): def collect_types(x, types): if is_nested(x): for y in x: collect_types(y, types) else: types.add(type_string(x)) all_args = itertools.chain(args, kwargs.values()) types = set() for x in all_args: if is_fp_tensor(x): collect_types(x, types) return types def type_string(x): return x.type().split('.')[-1] def maybe_half(x, name='', verbose=False): if is_nested(x): return type(x)([maybe_half(y) for y in x]) if not x.is_cuda or type_string(x) == 'HalfTensor': return x else: if verbose: print('Float->Half ({})'.format(name)) return x.half() def maybe_float(x, name='', verbose=False): if is_nested(x): return type(x)([maybe_float(y) for y in x]) if not x.is_cuda or type_string(x) == 'FloatTensor': return x else: if verbose: print('Half->Float ({})'.format(name)) return x.float() # NB: returneds casted `args`, mutates `kwargs` in-place def casted_args(cast_fn, args, kwargs): new_args = [] for x in args: if is_fp_tensor(x): new_args.append(cast_fn(x)) else: new_args.append(x) for k in kwargs: val = kwargs[k] if is_fp_tensor(val): kwargs[k] = cast_fn(val) return new_args def cached_cast(cast_fn, x, cache): if is_nested(x): return type(x)([cached_cast(y) for y in x]) if x in cache: cached_x = cache[x] if x.requires_grad and cached_x.requires_grad: # Make sure x is actually cached_x's autograd parent. if cached_x.grad_fn.next_functions[1][0].variable is not x: raise RuntimeError("x and cache[x] both require grad, but x is not " "cache[x]'s parent. This is likely an error.") # During eval, it's possible to end up caching casted weights with # requires_grad=False. On the next training iter, if cached_x is found # and reused from the cache, it will not actually have x as its parent. # Therefore, we choose to invalidate the cache (and force refreshing the cast) # if x.requires_grad and cached_x.requires_grad do not match. # # During eval (i.e. running under with torch.no_grad()) the invalidation # check would cause the cached value to be dropped every time, because # cached_x would always be created with requires_grad=False, while x would # still have requires_grad=True. This would render the cache effectively # useless during eval. Therefore, if we are running under the no_grad() # context manager (torch.is_grad_enabled=False) we elide the invalidation # check, and use the cached value even though its requires_grad flag doesn't # match. During eval, we don't care that there's no autograd-graph # connection between x and cached_x. if torch.is_grad_enabled() and x.requires_grad != cached_x.requires_grad: del cache[x] else: return cached_x casted_x = cast_fn(x) cache[x] = casted_x return casted_x def verbosify(cast_fn, fn_name, verbose): if verbose: return functools.partial(cast_fn, name=fn_name, verbose=verbose) else: return cast_fn def as_inplace(fns): for x in fns: yield x + '_' def has_func(mod, fn): if isinstance(mod, dict): return fn in mod else: return hasattr(mod, fn) def get_func(mod, fn): if isinstance(mod, dict): return mod[fn] else: return getattr(mod, fn) def set_func(mod, fn, new_fn): if isinstance(mod, dict): mod[fn] = new_fn else: setattr(mod, fn, new_fn) def set_func_save(handle, mod, fn, new_fn): cur_fn = get_func(mod, fn) handle._save_func(mod, fn, cur_fn) set_func(mod, fn, new_fn) # A couple problems get solved here: # - The flat_weight buffer is disconnected from autograd graph, # so the fp16 weights need to be derived from the input weights # to this forward call, not the flat buffer. # - The ordering of weights in the flat buffer is...idiosyncratic. # First problem is solved with combination of set_ (to set up # correct storage) and copy_ (so the fp16 weight derives from the # fp32 one in autograd. # Second is solved by doing ptr arithmetic on the fp32 weights # to derive the correct offset. # # TODO: maybe this should actually use # `torch._cudnn_rnn_flatten_weight`? But then I need to call # on first iter and cache the right offsets. Ugh. def synthesize_flattened_rnn_weights(fp32_weights, fp16_flat_tensor, rnn_fn='', verbose=False): fp16_weights = [] fp32_base_ptr = fp32_weights[0][0].data_ptr() for layer_weights in fp32_weights: fp16_layer_weights = [] for w_fp32 in layer_weights: w_fp16 = w_fp32.new().half() offset = (w_fp32.data_ptr() - fp32_base_ptr) // w_fp32.element_size() w_fp16.set_(fp16_flat_tensor.storage(), offset, w_fp32.shape) w_fp16.copy_(w_fp32) if verbose: print('Float->Half ({})'.format(rnn_fn)) fp16_layer_weights.append(w_fp16) fp16_weights.append(fp16_layer_weights) return fp16_weights # Roughly same as above, just the `fp32_weights` aren't nested. # Code kept separate for readability. def new_synthesize_flattened_rnn_weights(fp32_weights, fp16_flat_tensor, rnn_fn='', verbose=False): fp16_weights = [] fp32_base_ptr = fp32_weights[0].data_ptr() for w_fp32 in fp32_weights: w_fp16 = w_fp32.new().half() offset = (w_fp32.data_ptr() - fp32_base_ptr) // w_fp32.element_size() w_fp16.set_(fp16_flat_tensor.storage(), offset, w_fp32.shape) w_fp16.copy_(w_fp32) if verbose: print('Float->Half ({})'.format(rnn_fn)) fp16_weights.append(w_fp16) return fp16_weights
apex-master
apex/amp/utils.py
from . import compat from . import utils from ._amp_state import _amp_state from . import rnn_compat import functools import torch def make_cast_wrapper(orig_fn, cast_fn, handle, try_caching=False): @functools.wraps(orig_fn) def wrapper(*args, **kwargs): if not handle.is_active(): return orig_fn(*args, **kwargs) if try_caching and handle.has_cache: args = list(args) for i in range(len(args)): if utils.should_cache(args[i]): args[i] = utils.cached_cast(cast_fn, args[i], handle.cache) for k in kwargs: if utils.should_cache(kwargs[k]): kwargs[k] = utils.cached_cast(cast_fn, kwargs[k], handle.cache) new_args = utils.casted_args(cast_fn, args, kwargs) return orig_fn(*new_args, **kwargs) return wrapper def cached_cast(mod, fn, cast_fn, handle, try_caching=False, verbose=False): if not utils.has_func(mod, fn): return orig_fn = utils.get_func(mod, fn) cast_fn = utils.verbosify(cast_fn, fn, verbose) wrapper = make_cast_wrapper(orig_fn, cast_fn, handle, try_caching) utils.set_func_save(handle, mod, fn, wrapper) # `handle` arg is unused, but simplifies API to make `make_cast_wrapper` # Annoyingly, make_promote_wrapper still uses the global handle. Once everyone # is on the new API and I am free to get rid of handle, I can clean this up. def make_promote_wrapper(orig_fn, cast_fn, handle=None): @functools.wraps(orig_fn) def wrapper(*args, **kwargs): if not _amp_state.handle.is_active(): return orig_fn(*args, **kwargs) types = utils.collect_fp_tensor_types(args, kwargs) if len(types) <= 1: return orig_fn(*args, **kwargs) elif len(types) == 2 and types == set(['HalfTensor', 'FloatTensor']): new_args = utils.casted_args(cast_fn, args, kwargs) return orig_fn(*new_args, **kwargs) else: raise NotImplementedError('Do not know how to handle ' + 'these types to promote: {}' .format(types)) return wrapper def promote(mod, fn, handle, verbose=False): orig_fn = utils.get_func(mod, fn) maybe_float = utils.verbosify(utils.maybe_float, fn, verbose) wrapper = make_promote_wrapper(orig_fn, maybe_float) utils.set_func_save(handle, mod, fn, wrapper) def sequence_promote(mod, fn, handle, verbose=False): orig_fn = utils.get_func(mod, fn) maybe_float = utils.verbosify(utils.maybe_float, fn, verbose) @functools.wraps(orig_fn) def wrapper(seq, *args, **kwargs): if not _amp_state.handle.is_active(): return orig_fn(seq, *args, **kwargs) types = set([utils.type_string(x) for x in seq]) if len(types) <= 1: return orig_fn(seq, *args, **kwargs) elif types == set(['HalfTensor', 'FloatTensor']): cast_seq = utils.casted_args(maybe_float, seq, {}) return orig_fn(cast_seq, *args, **kwargs) else: # TODO: other mixed-type cases aren't due to amp. # Just pass through? return orig_fn(seq, *args, **kwargs) utils.set_func_save(handle, mod, fn, wrapper) def promote_match_arg0(mod, fn, handle, verbose=False): if not utils.has_func(mod, fn): return orig_fn = utils.get_func(mod, fn) @functools.wraps(orig_fn) def wrapper(arg0, *args, **kwargs): assert compat.is_tensor_like(arg0) if not _amp_state.handle.is_active(): return orig_fn(arg0, *args, **kwargs) if utils.type_string(arg0) == 'HalfTensor': cast_fn = utils.maybe_half elif utils.type_string(arg0) == 'FloatTensor': cast_fn = utils.maybe_float else: return orig_fn(arg0, *args, **kwargs) cast_fn = utils.verbosify(cast_fn, fn, verbose) new_args = utils.casted_args(cast_fn, args, kwargs) return orig_fn(arg0, *new_args, **kwargs) utils.set_func_save(handle, mod, fn, wrapper) def err_if_any_half(mod, fn, handle, custom_err_msg=None): if not utils.has_func(mod, fn): return orig_fn = utils.get_func(mod, fn) @functools.wraps(orig_fn) def wrapper(*args, **kwargs): types = utils.collect_fp_tensor_types(args, kwargs) if 'HalfTensor' in types: if custom_err_msg: raise NotImplementedError(custom_err_msg) else: raise NotImplementedError('Cannot call in-place function ' + '{} with fp16 arguments.'.format(fn)) else: return orig_fn(*args, **kwargs) utils.set_func_save(handle, mod, fn, wrapper) def err_if_arg0_half(mod, fn, handle, verbose=False): if not utils.has_func(mod, fn): return orig_fn = utils.get_func(mod, fn) @functools.wraps(orig_fn) def wrapper(arg0, *args, **kwargs): assert compat.is_tensor_like(arg0) if utils.type_string(arg0) == 'HalfTensor': raise NotImplementedError('Cannot call in-place method ' + '{} on fp16 Tensors.'.format(fn)) else: cast_fn = utils.verbosify(utils.maybe_float, fn, verbose) new_args = utils.casted_args(cast_fn, args, kwargs) return orig_fn(arg0, *new_args, **kwargs) utils.set_func_save(handle, mod, fn, wrapper) # Current RNN approach: # - Wrap top-level `RNN` function in thnn backend # - Will call into either CudnnRNN or AutogradRNN # - Each of these are factory functions that return a per-iter # `forward` function # - We interpose on the factory function to: # 1) Interpose on the actual forward function and put in casts # 2) Insert an fp16 `flat_weight` if necessary def rnn_cast(backend, fn, handle, verbose=False): orig_rnn = utils.get_func(backend, fn) @functools.wraps(orig_rnn) def rnn_wrapper(*args, **kwargs): flat_weight = kwargs.get('flat_weight') if flat_weight is not None: # We replace `flat_weight` with an uninitialized fp16 # Tensor. The "actual" weight tensors (provided in `forward`), # will then be set up as ptrs into the buffer and have the # corresponding fp32 values copied in. # We need to call `copy` on the "actual" weights so that the # autograd graph correctly backprops from the wgrads computed # inside cuDNN (on fp16 weights) into the fp32 weights. assert utils.type_string(flat_weight) == 'FloatTensor' if compat.tensor_is_float_tensor() or compat.tensor_is_variable(): # Pre-0.4. A little slower, since it zeros out memory. flat_weight_fp16 = flat_weight.new().half().resize_(flat_weight.shape) else: flat_weight_fp16 = torch.empty_like(flat_weight, dtype=torch.float16) kwargs['flat_weight'] = flat_weight_fp16 else: flat_weight_fp16 = None forward = orig_rnn(*args, **kwargs) @functools.wraps(forward) def fwd_wrapper(*fargs, **fkwargs): assert len(fargs) == 3 or len(fargs) == 4 inputs, weights, hiddens = fargs[:3] assert utils.is_fp_tensor(inputs) assert isinstance(weights, list) cast_fn = utils.verbosify(utils.maybe_half, fn, verbose) new_args = [] # 0) Inputs new_args.append(cast_fn(inputs)) # 1) Weights if flat_weight_fp16 is not None: fp16_weights = utils.synthesize_flattened_rnn_weights( weights, flat_weight_fp16, fn, verbose) else: fp16_weights = [[cast_fn(w) for w in layer] for layer in weights] new_args.append(fp16_weights) # 2) Inputs: either a tuple (for LSTM) or single tensor if isinstance(hiddens, tuple): new_args.append(tuple(cast_fn(x) for x in hiddens)) elif utils.is_fp_tensor(hiddens): new_args.append(cast_fn(hiddens)) else: # Hiddens can, in principle, be `None` -- pass through new_args.append(hiddens) # 3) Batch sizes (0.4 or later only) if len(fargs) == 4: new_args.append(fargs[3]) return forward(*new_args, **fkwargs) return fwd_wrapper utils.set_func_save(handle, backend, fn, rnn_wrapper) def new_rnn_cast(fn, handle, verbose=False): # Forward+backward compatibility around https://github.com/pytorch/pytorch/pull/15744 # For rnn backend calls that route through _rnn_impls, we must patch the ref # that _rnn_impls stashed. For rnn backend calls that directly invoke # _VF.<backend>, e.g. _VF.lstm, we can patch onto VariableFunctionsShim, # which in turn has patched the ref named "_VF" in torch.nn.modules.rnn. if utils.has_func(torch.nn.modules.rnn._rnn_impls, fn): mod = torch.nn.modules.rnn._rnn_impls else: mod = torch.nn.modules.rnn._VF assert isinstance(mod, rnn_compat.VariableFunctionsShim) fn = fn.lower() orig_fn = utils.get_func(mod, fn) cast_fn = utils.verbosify(utils.maybe_half, fn, verbose) @functools.wraps(orig_fn) def wrapper(*args, **kwargs): # Exact call signature from modules/rnn.py assert len(args) == 9 assert len(kwargs) == 0 if not _amp_state.handle.is_active(): return orig_fn(*args, **kwargs) if isinstance(args[6], bool): params_idx = 2 # Not PackedSequence case else: params_idx = 3 # PackedSequence case new_args = [] for i, arg in enumerate(args): if i == params_idx: num_params = sum([x.numel() for x in arg]) fp16_weight_buf = args[0].new_empty((num_params,), dtype=torch.half) casted_weights = utils.new_synthesize_flattened_rnn_weights( arg, fp16_weight_buf, fn, verbose) new_args.append(casted_weights) elif utils.is_fp_tensor(arg): new_args.append(cast_fn(arg)) else: new_args.append(arg) return orig_fn(*new_args) utils.set_func_save(handle, mod, fn, wrapper) def disable_casts(mod, fn, handle): if not utils.has_func(mod, fn): return orig_fn = utils.get_func(mod, fn) @functools.wraps(orig_fn) def wrapper(*args, **kwargs): with handle._disable_casts(): return orig_fn(*args, **kwargs) utils.set_func_save(handle, mod, fn, wrapper)
apex-master
apex/amp/wrap.py
from . import utils, wrap import torch _VF = torch._C._VariableFunctions RNN_NAMES = ['rnn_relu', 'rnn_tanh', 'gru', 'lstm'] def _gen_VF_wrapper(name): def wrapper(*args, **kwargs): return getattr(_VF, name)(*args, **kwargs) return wrapper # Some python magic to generate an object that has the rnn cell functions # defined on it, all of which call into corresponding _VF version. # Intended to patch torch.nn.modules.rnn._VF (aka, the ref named "_VF" # imported at module scope within torch.nn.modules.rnn). This should # not affect third-party importers of _VF.py. class VariableFunctionsShim(object): def __init__(self): for name in RNN_NAMES: for suffix in ['', '_cell']: fn_name = name + suffix setattr(self, fn_name, _gen_VF_wrapper(fn_name)) def has_old_rnns(): try: torch.nn.backends.thnn.backend.LSTMCell return True except: return False def whitelist_rnn_cells(handle, verbose): # Different module + function names in old/new RNN cases if has_old_rnns(): fn_names = ['RNNReLUCell', 'RNNTanhCell', 'LSTMCell', 'GRUCell'] mod = torch.nn.backends.thnn.backend else: fn_names = [x + '_cell' for x in RNN_NAMES] mod = torch.nn.modules.rnn._VF assert isinstance(mod, VariableFunctionsShim) # Insert casts on cell functions for fn in fn_names: wrap.cached_cast(mod, fn, utils.maybe_half, handle, try_caching=True, verbose=verbose) if has_old_rnns(): # Special handling of `backward` for fused gru / lstm: # The `backward` method calls Tensor.sum() (blacklist) internally, # and then the resulting grad_input has the wrong type. # TODO: where else is this a problem? for rnn_type in ['GRUFused', 'LSTMFused']: mod = getattr(torch.nn._functions.thnn.rnnFusedPointwise, rnn_type) wrap.disable_casts(mod, 'backward', handle)
apex-master
apex/amp/rnn_compat.py
apex-master
apex/amp/lists/__init__.py
import torch from .. import utils MODULE = torch FP16_FUNCS = [ # Low level functions wrapped by torch.nn layers. # The wrapper layers contain the weights which are then passed in as a parameter # to these functions. 'conv1d', 'conv2d', 'conv3d', 'conv_transpose1d', 'conv_transpose2d', 'conv_transpose3d', 'conv_tbc', 'prelu', # BLAS 'addmm', 'addmv', 'addr', 'matmul', 'mm', 'mv', ] FP32_FUNCS = [ # Pointwise 'acos', 'asin', 'cosh', 'erfinv', 'exp', 'expm1', 'log', 'log10', 'log2', 'reciprocal', 'rsqrt', 'sinh', 'tan', # Other math 'pow', # Reduction 'cumprod', 'cumsum', 'dist', # 'mean', 'norm', 'prod', 'std', 'sum', 'var', # Misc 'renorm' ] version_strings = torch.__version__.split('.') version_major = version_strings[0] version_minor = version_strings[1] version_num = float(version_major + "." + version_minor) # Before torch 1.1, mean must be blacklisted. if version_num < 1.1: FP32_FUNCS.append('mean') # Before CUDA 9.1, batched matmul was missing fast FP16 kernels. We # check the CUDA version -- if at least 9.1, then put the bmm # functions on the fp16 list. Otherwise, put them on the fp32 list. _bmms = ['addbmm', 'baddbmm', 'bmm'] if utils.is_cuda_enabled(): # workaround https://github.com/facebookresearch/maskrcnn-benchmark/issues/802 if utils.get_cuda_version() >= (9, 1, 0): FP16_FUNCS.extend(_bmms) else: FP32_FUNCS.extend(_bmms) # Multi-tensor fns that may need type promotion CASTS = [ # Multi-tensor math 'addcdiv', 'addcmul', 'atan2', 'cross', 'bilinear', 'dot', # Element-wise _or_ tensor-wise math 'add', 'div', 'mul', # Comparison 'eq', 'equal', 'ge', 'gt', 'le', 'lt', 'ne' ] # Functions that take sequence arguments. We need to inspect the whole # sequence and cast to the widest type. SEQUENCE_CASTS = [ 'cat', 'stack' ]
apex-master
apex/amp/lists/torch_overrides.py
# TODO: think about the following two. They do weird things. # - torch.nn.utils.clip_grad (but it should always be fp32 anyway) # - torch.nn.utils.weight_norm # Notes: # F.instance_norm uses batch_norm internally. Which correctly handles # fp16 in/out with fp32 weights. So we shouldn't do anything for # either of these. # F.normalize calls `input.norm()` internally, so it's redundant, but # kept here in case impl. changes. # F.cosine_similarity is same: calls `x.norm()` internally. import torch.nn.functional MODULE = torch.nn.functional FP16_FUNCS = [ 'conv1d', 'conv2d', 'conv3d', 'conv_transpose1d', 'conv_transpose2d', 'conv_transpose3d', 'conv_tbc', # Undocumented / maybe new? 'linear', ] FP32_FUNCS = [ # Interpolation/Upsampling TODO: Remove for 1.2 'interpolate', 'grid_sample', # Pointwise 'softplus', 'softmin', 'log_softmax', 'softmax', 'gelu', # Normalization 'layer_norm', 'group_norm', 'local_response_norm', 'normalize', 'cosine_similarity', # Loss functions # TODO: which of these can be fp16? 'poisson_nll_loss', 'cosine_embedding_loss', 'cross_entropy', 'hinge_embedding_loss', 'kl_div', 'l1_loss', 'mse_loss', 'margin_ranking_loss', 'multilabel_margin_loss', 'multilabel_soft_margin_loss', 'multi_margin_loss', 'nll_loss', 'binary_cross_entropy_with_logits', 'smooth_l1_loss', 'soft_margin_loss', 'triplet_margin_loss', 'ctc_loss' ] BANNED_FUNCS = [ ('binary_cross_entropy', ("\namp does not work out-of-the-box with `F.binary_cross_entropy` or `torch.nn.BCELoss.` " "It requires that the output of the previous function be already a FloatTensor. \n\n" "Most models have a Sigmoid right before BCELoss. In that case, you can use\n" " torch.nn.BCEWithLogitsLoss\nto combine Sigmoid+BCELoss into a single layer " "that is compatible with amp.\nAnother option is to add\n" " amp.register_float_function(torch, 'sigmoid')\nbefore calling `amp.init()`.\n" "If you _really_ know what you are doing, you can disable this warning by passing " "allow_banned=True to `amp.init()`.")) ]
apex-master
apex/amp/lists/functional_overrides.py
from .. import compat from . import torch_overrides import importlib import torch # if compat.variable_is_tensor() and not compat.tensor_is_variable(): MODULE = torch.Tensor # else: # MODULE = torch.autograd.Variable FP16_FUNCS = compat.filter_attrs(MODULE, [ '__matmul__', ]) FP32_FUNCS = compat.filter_attrs(MODULE, [ '__ipow__', '__pow__', '__rpow__', # Cast to fp32 before transfer to CPU 'cpu', ]) CASTS = compat.filter_attrs(MODULE, [ '__add__', '__div__', '__eq__', '__ge__', '__gt__', '__iadd__', '__idiv__', '__imul__', '__isub__', '__itruediv__', '__le__', '__lt__', '__mul__', '__ne__', '__radd__', '__rdiv__', '__rmul__', '__rsub__', '__rtruediv__', '__sub__', '__truediv__', ]) # None of these, but here to make code cleaner. SEQUENCE_CASTS = [] # We need to grab all the methods from torch_overrides and add them to # the Tensor lists as well, as almost all methods are duplicated # between `torch` and `torch.Tensor` (and check with `hasattr`, # because a few random ones aren't defined on Tensor) _self_mod = importlib.import_module(__name__) for attrname in ['FP16_FUNCS', 'FP32_FUNCS', 'CASTS', 'SEQUENCE_CASTS']: lst = getattr(_self_mod, attrname) for fn in getattr(torch_overrides, attrname): if hasattr(MODULE, fn): lst.append(fn)
apex-master
apex/amp/lists/tensor_overrides.py
import torch import torch.nn as nn from torch.autograd import Variable import torch.nn.functional as F import math def is_iterable(maybe_iterable): return isinstance(maybe_iterable, list) or isinstance(maybe_iterable, tuple) def flatten_list(tens_list): """ flatten_list """ if not is_iterable(tens_list): return tens_list return torch.cat(tens_list, dim=0).view(len(tens_list), *tens_list[0].size() ) #These modules always assumes batch_first class bidirectionalRNN(nn.Module): """ bidirectionalRNN """ def __init__(self, inputRNN, num_layers=1, dropout = 0): super(bidirectionalRNN, self).__init__() self.dropout = dropout self.fwd = stackedRNN(inputRNN, num_layers=num_layers, dropout = dropout) self.bckwrd = stackedRNN(inputRNN.new_like(), num_layers=num_layers, dropout = dropout) self.rnns = nn.ModuleList([self.fwd, self.bckwrd]) #collect hidden option will return all hidden/cell states from entire RNN def forward(self, input, collect_hidden=False): """ forward() """ seq_len = input.size(0) bsz = input.size(1) fwd_out, fwd_hiddens = list(self.fwd(input, collect_hidden = collect_hidden)) bckwrd_out, bckwrd_hiddens = list(self.bckwrd(input, reverse=True, collect_hidden = collect_hidden)) output = torch.cat( [fwd_out, bckwrd_out], -1 ) hiddens = tuple( torch.cat(hidden, -1) for hidden in zip( fwd_hiddens, bckwrd_hiddens) ) return output, hiddens def reset_parameters(self): """ reset_parameters() """ for rnn in self.rnns: rnn.reset_parameters() def init_hidden(self, bsz): """ init_hidden() """ for rnn in self.rnns: rnn.init_hidden(bsz) def detach_hidden(self): """ detach_hidden() """ for rnn in self.rnns: rnn.detachHidden() def reset_hidden(self, bsz): """ reset_hidden() """ for rnn in self.rnns: rnn.reset_hidden(bsz) def init_inference(self, bsz): """ init_inference() """ for rnn in self.rnns: rnn.init_inference(bsz) #assumes hidden_state[0] of inputRNN is output hidden state #constructor either takes an RNNCell or list of RNN layers class stackedRNN(nn.Module): """ stackedRNN """ def __init__(self, inputRNN, num_layers=1, dropout=0): super(stackedRNN, self).__init__() self.dropout = dropout if isinstance(inputRNN, RNNCell): self.rnns = [inputRNN] for i in range(num_layers-1): self.rnns.append(inputRNN.new_like(inputRNN.output_size)) elif isinstance(inputRNN, list): assert len(inputRNN) == num_layers, "RNN list length must be equal to num_layers" self.rnns=inputRNN else: raise RuntimeError() self.nLayers = len(self.rnns) self.rnns = nn.ModuleList(self.rnns) ''' Returns output as hidden_state[0] Tensor([sequence steps][batch size][features]) If collect hidden will also return Tuple( [n_hidden_states][sequence steps] Tensor([layer][batch size][features]) ) If not collect hidden will also return Tuple( [n_hidden_states] Tensor([layer][batch size][features]) ''' def forward(self, input, collect_hidden=False, reverse=False): """ forward() """ seq_len = input.size(0) bsz = input.size(1) inp_iter = reversed(range(seq_len)) if reverse else range(seq_len) hidden_states = [[] for i in range(self.nLayers)] outputs = [] for seq in inp_iter: for layer in range(self.nLayers): if layer == 0: prev_out = input[seq] outs = self.rnns[layer](prev_out) if collect_hidden: hidden_states[layer].append(outs) elif seq == seq_len-1: hidden_states[layer].append(outs) prev_out = outs[0] outputs.append(prev_out) if reverse: outputs = list(reversed(outputs)) ''' At this point outputs is in format: list( [seq_length] x Tensor([bsz][features]) ) need to convert it to: list( Tensor([seq_length][bsz][features]) ) ''' output = flatten_list(outputs) ''' hidden_states at this point is in format: list( [layer][seq_length][hidden_states] x Tensor([bsz][features]) ) need to convert it to: For not collect hidden: list( [hidden_states] x Tensor([layer][bsz][features]) ) For collect hidden: list( [hidden_states][seq_length] x Tensor([layer][bsz][features]) ) ''' if not collect_hidden: seq_len = 1 n_hid = self.rnns[0].n_hidden_states new_hidden = [ [ [ None for k in range(self.nLayers)] for j in range(seq_len) ] for i in range(n_hid) ] for i in range(n_hid): for j in range(seq_len): for k in range(self.nLayers): new_hidden[i][j][k] = hidden_states[k][j][i] hidden_states = new_hidden #Now in format list( [hidden_states][seq_length][layer] x Tensor([bsz][features]) ) #Reverse seq_length if reverse if reverse: hidden_states = list( list(reversed(list(entry))) for entry in hidden_states) #flatten layer dimension into tensor hiddens = list( list( flatten_list(seq) for seq in hidden ) for hidden in hidden_states ) #Now in format list( [hidden_states][seq_length] x Tensor([layer][bsz][features]) ) #Remove seq_length dimension if not collect_hidden if not collect_hidden: hidden_states = list( entry[0] for entry in hidden_states) return output, hidden_states def reset_parameters(self): """ reset_parameters() """ for rnn in self.rnns: rnn.reset_parameters() def init_hidden(self, bsz): """ init_hidden() """ for rnn in self.rnns: rnn.init_hidden(bsz) def detach_hidden(self): """ detach_hidden() """ for rnn in self.rnns: rnn.detach_hidden() def reset_hidden(self, bsz): """ reset_hidden() """ for rnn in self.rnns: rnn.reset_hidden(bsz) def init_inference(self, bsz): """ init_inference() """ for rnn in self.rnns: rnn.init_inference(bsz) class RNNCell(nn.Module): """ RNNCell gate_multiplier is related to the architecture you're working with For LSTM-like it will be 4 and GRU-like will be 3. Always assumes input is NOT batch_first. Output size that's not hidden size will use output projection Hidden_states is number of hidden states that are needed for cell if one will go directly to cell as tensor, if more will go as list """ def __init__(self, gate_multiplier, input_size, hidden_size, cell, n_hidden_states = 2, bias = False, output_size = None): super(RNNCell, self).__init__() self.gate_multiplier = gate_multiplier self.input_size = input_size self.hidden_size = hidden_size self.cell = cell self.bias = bias self.output_size = output_size if output_size is None: self.output_size = hidden_size self.gate_size = gate_multiplier * self.hidden_size self.n_hidden_states = n_hidden_states self.w_ih = nn.Parameter(torch.empty(self.gate_size, self.input_size)) self.w_hh = nn.Parameter(torch.empty(self.gate_size, self.output_size)) #Check if there's recurrent projection if(self.output_size != self.hidden_size): self.w_ho = nn.Parameter(torch.empty(self.output_size, self.hidden_size)) self.b_ih = self.b_hh = None if self.bias: self.b_ih = nn.Parameter(torch.empty(self.gate_size)) self.b_hh = nn.Parameter(torch.empty(self.gate_size)) #hidden states for forward self.hidden = [ None for states in range(self.n_hidden_states)] self.reset_parameters() def new_like(self, new_input_size=None): """ new_like() """ if new_input_size is None: new_input_size = self.input_size return type(self)(self.gate_multiplier, new_input_size, self.hidden_size, self.cell, self.n_hidden_states, self.bias, self.output_size) #Use xavier where we can (weights), otherwise use uniform (bias) def reset_parameters(self, gain=1): """ reset_parameters() """ stdev = 1.0 / math.sqrt(self.hidden_size) for param in self.parameters(): param.data.uniform_(-stdev, stdev) ''' Xavier reset: def reset_parameters(self, gain=1): stdv = 1.0 / math.sqrt(self.gate_size) for param in self.parameters(): if (param.dim() > 1): torch.nn.init.xavier_normal(param, gain) else: param.data.uniform_(-stdv, stdv) ''' def init_hidden(self, bsz): """ init_hidden() """ for param in self.parameters(): if param is not None: a_param = param break for i, _ in enumerate(self.hidden): if(self.hidden[i] is None or self.hidden[i].data.size()[0] != bsz): if i==0: hidden_size = self.output_size else: hidden_size = self.hidden_size tens = a_param.data.new(bsz, hidden_size).zero_() self.hidden[i] = Variable(tens, requires_grad=False) def reset_hidden(self, bsz): """ reset_hidden() """ for i, _ in enumerate(self.hidden): self.hidden[i] = None self.init_hidden(bsz) def detach_hidden(self): """ detach_hidden() """ for i, _ in enumerate(self.hidden): if self.hidden[i] is None: raise RuntimeError("Must initialize hidden state before you can detach it") for i, _ in enumerate(self.hidden): self.hidden[i] = self.hidden[i].detach() def forward(self, input): """ forward() if not inited or bsz has changed this will create hidden states """ self.init_hidden(input.size()[0]) hidden_state = self.hidden[0] if self.n_hidden_states == 1 else self.hidden self.hidden = self.cell(input, hidden_state, self.w_ih, self.w_hh, b_ih=self.b_ih, b_hh=self.b_hh) if(self.n_hidden_states > 1): self.hidden = list(self.hidden) else: self.hidden=[self.hidden] if self.output_size != self.hidden_size: self.hidden[0] = F.linear(self.hidden[0], self.w_ho) return tuple(self.hidden)
apex-master
apex/RNN/RNNBackend.py
import torch from torch.nn._functions.rnn import LSTMCell, RNNReLUCell, RNNTanhCell, GRUCell from apex import deprecated_warning from .RNNBackend import bidirectionalRNN, stackedRNN, RNNCell from .cells import mLSTMRNNCell, mLSTMCell def toRNNBackend(inputRNN, num_layers, bidirectional=False, dropout = 0): """ :class:`toRNNBackend` """ deprecated_warning("`apex.RNN` is deprecated and will be removed by the end of February 2023.") if bidirectional: return bidirectionalRNN(inputRNN, num_layers, dropout = dropout) else: return stackedRNN(inputRNN, num_layers, dropout = dropout) def LSTM(input_size, hidden_size, num_layers, bias=True, batch_first=False, dropout=0, bidirectional=False, output_size = None): """ :class:`LSTM` """ inputRNN = RNNCell(4, input_size, hidden_size, LSTMCell, 2, bias, output_size) return toRNNBackend(inputRNN, num_layers, bidirectional, dropout=dropout) def GRU(input_size, hidden_size, num_layers, bias=True, batch_first=False, dropout=0, bidirectional=False, output_size = None): """ :class:`GRU` """ inputRNN = RNNCell(3, input_size, hidden_size, GRUCell, 1, bias, output_size) return toRNNBackend(inputRNN, num_layers, bidirectional, dropout=dropout) def ReLU(input_size, hidden_size, num_layers, bias=True, batch_first=False, dropout=0, bidirectional=False, output_size = None): """ :class:`ReLU` """ inputRNN = RNNCell(1, input_size, hidden_size, RNNReLUCell, 1, bias, output_size) return toRNNBackend(inputRNN, num_layers, bidirectional, dropout=dropout) def Tanh(input_size, hidden_size, num_layers, bias=True, batch_first=False, dropout=0, bidirectional=False, output_size = None): """ :class:`Tanh` """ inputRNN = RNNCell(1, input_size, hidden_size, RNNTanhCell, 1, bias, output_size) return toRNNBackend(inputRNN, num_layers, bidirectional, dropout=dropout) def mLSTM(input_size, hidden_size, num_layers, bias=True, batch_first=False, dropout=0, bidirectional=False, output_size = None): """ :class:`mLSTM` """ inputRNN = mLSTMRNNCell(input_size, hidden_size, bias=bias, output_size=output_size) return toRNNBackend(inputRNN, num_layers, bidirectional, dropout=dropout)
apex-master
apex/RNN/models.py
from .models import LSTM, GRU, ReLU, Tanh, mLSTM __all__ = ['models']
apex-master
apex/RNN/__init__.py
import torch import torch.nn as nn import torch.nn.functional as F from .RNNBackend import RNNCell from torch.nn._functions.thnn import rnnFusedPointwise as fusedBackend import math class mLSTMRNNCell(RNNCell): """ mLSTMRNNCell """ def __init__(self, input_size, hidden_size, bias = False, output_size = None): gate_multiplier = 4 super(mLSTMRNNCell, self).__init__(gate_multiplier, input_size, hidden_size, mLSTMCell, n_hidden_states = 2, bias = bias, output_size = output_size) self.w_mih = nn.Parameter(torch.empty(self.output_size, self.input_size)) self.w_mhh = nn.Parameter(torch.empty(self.output_size, self.output_size)) self.reset_parameters() def forward(self, input): """ mLSTMRNNCell.forward() """ #if not inited or bsz has changed this will create hidden states self.init_hidden(input.size()[0]) hidden_state = self.hidden[0] if self.n_hidden_states == 1 else self.hidden self.hidden = list( self.cell(input, hidden_state, self.w_ih, self.w_hh, self.w_mih, self.w_mhh, b_ih=self.b_ih, b_hh=self.b_hh) ) if self.output_size != self.hidden_size: self.hidden[0] = F.linear(self.hidden[0], self.w_ho) return tuple(self.hidden) def new_like(self, new_input_size=None): if new_input_size is None: new_input_size = self.input_size return type(self)( new_input_size, self.hidden_size, self.bias, self.output_size) def mLSTMCell(input, hidden, w_ih, w_hh, w_mih, w_mhh, b_ih=None, b_hh=None): """ mLSTMCell """ if input.is_cuda: igates = F.linear(input, w_ih) m = F.linear(input, w_mih) * F.linear(hidden[0], w_mhh) hgates = F.linear(m, w_hh) state = fusedBackend.LSTMFused.apply return state(igates, hgates, hidden[1], b_ih, b_hh) hx, cx = hidden m = F.linear(input, w_mih) * F.linear(hidden[0], w_mhh) gates = F.linear(input, w_ih, b_ih) + F.linear(m, w_hh, b_hh) ingate, forgetgate, cellgate, outgate = gates.chunk(4, 1) ingate = F.sigmoid(ingate) forgetgate = F.sigmoid(forgetgate) cellgate = F.tanh(cellgate) outgate = F.sigmoid(outgate) cy = (forgetgate * cx) + (ingate * cellgate) hy = outgate * F.tanh(cy) return hy, cy
apex-master
apex/RNN/cells.py
from .mlp import *
apex-master
apex/mlp/__init__.py
from copy import copy import math import torch from torch import nn from apex._autocast_utils import _cast_if_autocast_enabled import mlp_cuda class MlpFunction(torch.autograd.Function): @staticmethod def forward(ctx, bias, activation, *args): output = mlp_cuda.forward(bias, activation, args) ctx.save_for_backward(*args) ctx.outputs = output ctx.bias = bias ctx.activation = activation return output[0] @staticmethod def backward(ctx, grad_o): grads = mlp_cuda.backward(ctx.bias, ctx.activation, grad_o, ctx.outputs, ctx.saved_tensors) del ctx.outputs return (None, None, *grads) def mlp_function(bias, activation, *args): autocast_args = _cast_if_autocast_enabled(bias, activation, *args) return MlpFunction.apply(*autocast_args) class MLP(torch.nn.Module): """Launch MLP in C++ Args: mlp_sizes (list of int): MLP sizes. Example: [1024,1024,1024] will create 2 MLP layers with shape 1024x1024 bias (bool): Default True: relu (bool): Default True """ def __init__(self, mlp_sizes, bias=True, activation='relu'): super().__init__() self.num_layers = len(mlp_sizes) - 1 self.mlp_sizes = copy(mlp_sizes) self.bias = 1 if bias else 0 if activation == 'none': self.activation = 0 elif activation == 'relu': self.activation = 1 elif activation == 'sigmoid': self.activation = 2 else: raise TypeError("activation must be relu or none.") self.weights = [] self.biases = [] for i in range(self.num_layers): w = torch.nn.Parameter(torch.empty(mlp_sizes[i+1], mlp_sizes[i])) self.weights.append(w) name = 'weight_{}'.format(i) setattr(self, name, w) if self.bias: b = torch.nn.Parameter(torch.empty(mlp_sizes[i+1])) self.biases.append(b) name = 'bias_{}'.format(i) setattr(self, name, b) self.reset_parameters() def reset_parameters(self): for weight in self.weights: dimsum = weight.size(0) + weight.size(1) std = math.sqrt(2. / float(dimsum)) nn.init.normal_(weight, 0., std) if self.bias: for bias in self.biases: std = math.sqrt(1. / float(bias.size(0))) nn.init.normal_(bias, 0., std) def forward(self, input): return mlp_function(self.bias, self.activation, input, *self.weights, *self.biases) def extra_repr(self): s = F"MLP sizes: {self.mlp_sizes}, Bias={self.bias}, activation={self.activation}" return s
apex-master
apex/mlp/mlp.py
import torch import torch.distributed as dist from torch.nn import Parameter from torch.nn import Module from apex.parallel import DistributedDataParallel as DDP import argparse import os parser = argparse.ArgumentParser(description='allreduce hook example') parser.add_argument("--local_rank", default=0, type=int) args = parser.parse_args() args.distributed = False if 'WORLD_SIZE' in os.environ: args.distributed = int(os.environ['WORLD_SIZE']) > 1 if args.distributed: args.gpu = args.local_rank % torch.cuda.device_count() torch.cuda.set_device(args.gpu) torch.distributed.init_process_group(backend='nccl', init_method='env://') args.world_size = torch.distributed.get_world_size() torch.set_printoptions(precision=10) torch.manual_seed(args.local_rank) class Model(Module): def __init__(self): super(Model, self).__init__() self.a = Parameter(torch.cuda.FloatTensor(4096*4096).fill_(1.0)) self.b = Parameter(torch.cuda.FloatTensor(4096*4096).fill_(2.0)) def forward(self, input): return (input*self.a)*self.b model = Model() # model = DDP(model, message_size=1, gradient_predivide_factor=8.0) # model = DDP(model, delay_allreduce=True) # model = DDP(model, message_size=1, allreduce_trigger_params=[model.b]) model = DDP(model, message_size=1, allreduce_trigger_params=[model.b], num_allreduce_streams=3) x = torch.cuda.FloatTensor(4096*4096) passed = True torch.cuda.cudart().cudaProfilerStart() for i in range(10): x.fill_(i + args.local_rank) # fill x with new values every iteration for sanity model.zero_grad() out = model(x) loss = out.sum() # torch.cuda.nvtx.range_push("backward") loss.backward() # torch.cuda.nvtx.range_pop() # torch.cuda.nvtx.range_push("synchronize() + info") # torch.cuda.synchronize() print("i = {}".format(i)) def info(name, param, val): expected = val*4096*4096*(2.*i+1)/2. actual = param.grad.data.sum().item() print(name+": grad.data_ptr() = {}, expected sum {}, got {}".format( param.grad.data_ptr(), expected, actual)) return (expected == actual) if not info("model.a", model.module.a, 2.): passed = False if not info("model.b", model.module.b, 1.): passed = False # torch.cuda.nvtx.range_pop() torch.cuda.cudart().cudaProfilerStop() print("passed = ", passed)
apex-master
tests/distributed/DDP/ddp_race_condition_test.py
import torch import argparse import os from apex import amp # FOR DISTRIBUTED: (can also use torch.nn.parallel.DistributedDataParallel instead) from apex.parallel import DistributedDataParallel parser = argparse.ArgumentParser() # FOR DISTRIBUTED: Parse for the local_rank argument, which will be supplied # automatically by torch.distributed.launch. parser.add_argument("--local_rank", default=0, type=int) args = parser.parse_args() # FOR DISTRIBUTED: If we are running under torch.distributed.launch, # the 'WORLD_SIZE' environment variable will also be set automatically. args.distributed = False if 'WORLD_SIZE' in os.environ: args.distributed = int(os.environ['WORLD_SIZE']) > 1 if args.distributed: # FOR DISTRIBUTED: Set the device according to local_rank. torch.cuda.set_device(args.local_rank) # FOR DISTRIBUTED: Initialize the backend. torch.distributed.launch will provide # environment variables, and requires that you use init_method=`env://`. torch.distributed.init_process_group(backend='nccl', init_method='env://') torch.manual_seed(torch.distributed.get_rank()) torch.backends.cudnn.benchmark = True N, D_in, D_out = 64, 1024, 16 # Each process receives its own batch of "fake input data" and "fake target data." # The "training loop" in each process just uses this fake batch over and over. # https://github.com/NVIDIA/apex/tree/master/examples/imagenet provides a more realistic # example of distributed data sampling for both training and validation. x = torch.randn(N, D_in, device='cuda') y = torch.randn(N, D_out, device='cuda') model = torch.nn.Linear(D_in, D_out).cuda() optimizer = torch.optim.SGD(model.parameters(), lr=1e-3) model, optimizer = amp.initialize(model, optimizer, opt_level="O2") if args.distributed: # FOR DISTRIBUTED: After amp.initialize, wrap the model with # apex.parallel.DistributedDataParallel. model = DistributedDataParallel(model) # torch.nn.parallel.DistributedDataParallel is also fine, with some added args: # model = torch.nn.parallel.DistributedDataParallel(model, # device_ids=[args.local_rank], # output_device=args.local_rank) loss_fn = torch.nn.MSELoss() for t in range(500): optimizer.zero_grad() y_pred = model(x) loss = loss_fn(y_pred, y) with amp.scale_loss(loss, optimizer) as scaled_loss: scaled_loss.backward() optimizer.step() if args.local_rank == 0: print("final loss = ", loss) torch.save(list(model.parameters()), "rank{}model.pth".format(torch.distributed.get_rank())) torch.save(list(amp.master_params(optimizer)), "rank{}master.pth".format(torch.distributed.get_rank()))
apex-master
tests/distributed/amp_master_params/amp_master_params.py
import torch model_params_rank0 = torch.load("rank0model.pth", map_location = lambda storage, loc: storage.cuda(0)) model_params_rank1 = torch.load("rank1model.pth", map_location = lambda storage, loc: storage.cuda(0)) master_params_rank0 = torch.load("rank0master.pth", map_location = lambda storage, loc: storage.cuda(0)) master_params_rank1 = torch.load("rank1master.pth", map_location = lambda storage, loc: storage.cuda(0)) for model_rank0, model_rank1, master_rank0, master_rank1 in zip( model_params_rank0, model_params_rank1, master_params_rank0, master_params_rank1): assert torch.allclose(model_rank0, model_rank1), "Model param mismatch" assert torch.allclose(master_rank0, master_rank1), "Master param mismatch" # Some debugging/investigation assistance code: # maxval, maxind = torch.max(((torch.abs(model_rank0).float())/torch.abs(master_rank0)).view(-1), 0) # offending_val_half = model_rank0.view(-1)[maxind.item()] # offending_val_float = master_rank0.view(-1)[maxind.item()] # print(maxval.item(), maxind.item(), offending_val_half.item(), offending_val_float.item(), # offending_val_float.half().item()) # rtol needs to be > 2^-11 because of denormals... assert torch.allclose(model_rank0, master_rank0.half(), rtol=.005), "Model-master mismatch" print("OK: Model and master params match across ranks.")
apex-master
tests/distributed/amp_master_params/compare.py
import torch import apex model = apex.parallel.SyncBatchNorm(4).cuda() model.weight.data.uniform_() model.bias.data.uniform_() data = torch.rand((8,4)).cuda() model_ref = torch.nn.BatchNorm1d(4).cuda() model_ref.load_state_dict(model.state_dict()) data_ref = data.clone() output = model(data) output_ref = model_ref(data_ref) assert(output.allclose(output_ref)) assert(model.running_mean.allclose(model_ref.running_mean)) assert(model.running_var.allclose(model_ref.running_var))
apex-master
tests/distributed/synced_batchnorm/test_batchnorm1d.py
import torch import numpy as np import apex import syncbn import os import argparse import torch.optim as optim def compare(desc, inp1, inp2, error): a = inp1.clone().detach().cpu().numpy() b = inp2.clone().detach().cpu().numpy() close = np.allclose(a,b, error, error) if not close: print(desc, close) z = a - b index = (np.abs(z) >= error + error * np.abs(b)).nonzero() print("dif : ", z[index]) print("inp1 : ", a[index]) print("inp2 : ", b[index]) return close feature_size = 10 space_size = 40 batch_size = 32 from apex.parallel import DistributedDataParallel as DDP parser = argparse.ArgumentParser() parser.add_argument("--local_rank", default=0, type=int) parser.add_argument("--fp16", action='store_true', default=False) parser.add_argument("--fp64", action='store_true', default=False) args = parser.parse_args() args.world_size = int(os.environ['WORLD_SIZE']) torch.cuda.set_device(args.local_rank) torch.distributed.init_process_group(backend='nccl', init_method='env://') start = args.local_rank * batch_size//args.world_size finish = (args.local_rank + 1) * batch_size//args.world_size error = 1e-5 dtype = np.float32 if args.fp16: error = 1e-3 dtype = np.float16 elif args.fp64: error = 1e-8 dtype = np.float64 np.random.seed(18) inp = np.random.randn(batch_size, feature_size, space_size, space_size).astype(dtype) grad = np.random.randn(batch_size, feature_size, space_size, space_size).astype(dtype) weight = np.random.randn(feature_size).astype(dtype) bias = np.random.randn(feature_size).astype(dtype) type_tensor = torch.cuda.FloatTensor if args.fp16: type_tensor = torch.cuda.HalfTensor if args.fp64: type_tensor = torch.cuda.DoubleTensor ref_tensor = torch.cuda.DoubleTensor inp_t = type_tensor(inp) weight_t = type_tensor(weight) bias_t = type_tensor(bias) inp_r = ref_tensor(inp.transpose(1, 0, 2, 3).reshape(feature_size, -1)) inp2_r = ref_tensor(inp) weight_r = ref_tensor(weight).view(-1, 1, 1) bias_r = ref_tensor(bias).view(-1, 1, 1) grad_output_t = type_tensor(grad) m = inp_r.mean(1) b_v = inp_r.var(1, unbiased=False) unb_v = inp_r.var(1, unbiased=True) eps = 1e-5 mean, var_biased = syncbn.welford_mean_var(inp_t) inv_std = 1.0 / torch.sqrt(var_biased + eps) bn = torch.nn.BatchNorm2d(feature_size).cuda() bn.momentum = 1.0 bn.weight.data = weight_t.clone() bn.bias.data = bias_t.clone() if args.fp16: bn.half() if args.fp64: bn.double() inp_bn = inp_t.clone().requires_grad_() grad_bn = grad_output_t.clone().detach() out_bn = bn(inp_bn) out_bn.backward(grad_bn) # compensating the averaging over processes done by DDP # in order to produce mathematically equivalent result # https://github.com/NVIDIA/apex/issues/134#issuecomment-458307368 for param in bn.parameters(): param.grad = param.grad / args.world_size bn_opt = optim.SGD(bn.parameters(), lr=1.0) sbn = apex.parallel.SyncBatchNorm(feature_size).cuda() sbn.momentum = 1.0 sbn.weight.data = weight_t.clone() sbn.bias.data = bias_t.clone() if args.fp16: sbn.half() if args.fp64: sbn.double() sbn = DDP(sbn) sbn_opt = optim.SGD(sbn.parameters(), lr=1.0) inp_sbn = inp_t.clone().requires_grad_() grad_sbn = grad_output_t.clone().detach() out_sbn = sbn(inp_sbn[start:finish]) out_sbn.backward(grad_sbn[start:finish]) count = [ space_size**2 * ( (i+1) * batch_size // args.world_size - i * batch_size // args.world_size ) for i in range(0, args.world_size)] count = torch.cuda.IntTensor(count) print("--- count : " , count) sbn_result = True bn_result = True if args.local_rank == 0: sbn_result = compare("comparing mean: ", mean, m, error) and sbn_result sbn_result = compare("comparing biased variance: ", var_biased, b_v, error) and sbn_result out = syncbn.batchnorm_forward(inp_t, mean, inv_std, weight_t, bias_t) out_r = weight_r * (inp2_r - m.view(-1, 1, 1)) * torch.rsqrt(b_v.view(-1,1,1) + eps) + bias_r if args.local_rank == 0: sbn_result = compare("comparing output: ", out, out_r, error) and sbn_result compare("comparing bn output: ", out_bn, out_r, error) grad_output_t = type_tensor(grad) grad_output_r = ref_tensor(grad.transpose(1, 0, 2, 3).reshape(feature_size, -1)) grad_output2_r = ref_tensor(grad) grad_bias_r = grad_output_r.sum(1) grad_weight_r = ((inp2_r - m.view(-1, 1, 1)) * torch.rsqrt(b_v.view(-1,1,1) + eps) * grad_output2_r).transpose(1,0).contiguous().view(feature_size, -1).sum(1) sum_dy_r = grad_output_r.sum(1) mean_dy_r = grad_output_r.mean(1) mean_dy_xmu_r = ((inp2_r - m.view(-1, 1, 1)) * grad_output2_r).transpose(1,0).contiguous().view(feature_size, -1).mean(1) sum_dy_xmu_r = ((inp2_r - m.view(-1, 1, 1)) * grad_output2_r).transpose(1,0).contiguous().view(feature_size, -1).sum(1) grad_input_r = (grad_output2_r - mean_dy_r.view(-1, 1, 1) - (inp2_r - m.view(-1, 1, 1)) / (b_v.view(-1,1,1) + eps) * mean_dy_xmu_r.view(-1, 1, 1) ) * torch.rsqrt(b_v.view(-1,1,1) + eps) * weight_r.view(-1,1,1) sum_dy, sum_dy_xmu, grad_weight, grad_bias = syncbn.reduce_bn(grad_output_t, inp_t, mean, inv_std, weight_t) grad_input = syncbn.batchnorm_backward(grad_output_t, inp_t, mean, inv_std, weight_t, sum_dy, sum_dy_xmu, count) if args.local_rank == 0: sbn_result = compare("comparing bias grad: ", grad_bias, grad_bias_r, error) and sbn_result sbn_result = compare("comparing weight grad: ", grad_weight, grad_weight_r, error) and sbn_result sbn_result = compare("comparing sum_dy grad: ", sum_dy, sum_dy_r, error) and sbn_result sbn_result = compare("comparing sum_dy_xmu grad: ", sum_dy_xmu, sum_dy_xmu_r, error) and sbn_result sbn_result = compare("comparing input grad: ", grad_input, grad_input_r, error) and sbn_result compare("comparing bn input grad: ", inp_bn.grad, grad_input_r, error) if args.local_rank == 0: sbn_result = compare("comparing running_mean: ", bn.running_mean.data, sbn.module.running_mean.data, error) and sbn_result sbn_result = compare("comparing running_variance: ", bn.running_var.data, sbn.module.running_var.data, error) and sbn_result # execute by both compare("comparing layers output: ", out_bn[start:finish], out_sbn, error) and sbn_result compare("comparing layers grad_input: ", inp_bn.grad[start:finish], inp_sbn.grad[start:finish], error) and sbn_result bn_opt.step() sbn_opt.step() if args.local_rank == 0: compare("comparing bn vs sbn bias: ", bn.bias, sbn.module.bias, error) compare("comparing bn vs sbn weight: ", bn.weight, sbn.module.weight, error) if sbn_result: print("====SBN two gpu passed tests") else: print("*SBN two gpu failed*")
apex-master
tests/distributed/synced_batchnorm/two_gpu_unit_test.py
import torch import torch.nn as nn from torch.nn.parallel import DistributedDataParallel as DDP from apex.parallel import SyncBatchNorm as ApexSyncBatchNorm import argparse import os import numpy as np var_batch = 16 def compare(desc, inp1, inp2, error= 1e-5): a = inp1.clone().detach().cpu().numpy() b = inp2.clone().detach().cpu().numpy() close = np.allclose(a,b, error, error) if not close: print(desc, close) z = a - b index = (np.abs(z) >= error + error * np.abs(b)).nonzero() print("dif : ", z[index]) print("inp1 : ", a[index]) print("inp2 : ", b[index]) return close parser = argparse.ArgumentParser() parser.add_argument('--local_rank', type=int, default=0) parser.add_argument('--apex', action='store_true') args = parser.parse_args() torch.manual_seed(2809) # Setup DDP torch.cuda.set_device(args.local_rank) device = torch.device('cuda:{}'.format(args.local_rank)) torch.distributed.init_process_group( 'nccl', init_method='env://', rank=args.local_rank, ) # Setup model if args.apex: model = nn.Sequential( nn.Conv2d(3, 6, 3, 1, 1), ApexSyncBatchNorm(6) ) else: model = nn.Sequential( nn.Conv2d(3, 6, 3, 1, 1), nn.SyncBatchNorm(6) ) # Setup reference model model_reference = nn.Sequential( nn.Conv2d(3, 6, 3, 1, 1), nn.BatchNorm2d(6) ) with torch.no_grad(): model_reference[0].weight.copy_(model[0].weight) model_reference[0].bias.copy_(model[0].bias) model_reference.to(device) model = model.to(device) model = DDP(model, device_ids=[args.local_rank], output_device=args.local_rank) global_batch_size = var_batch + 8 # Create random data if args.local_rank == 0: data = torch.randn(var_batch, 3, 8, 8, device=device, dtype=torch.float) * 50.0 grad = torch.randint(0, 10, (var_batch, 6, 8, 8), device=device, dtype=torch.float) / 10.0 else: data = torch.randn(8, 3, 8, 8, device=device) grad = torch.randint(0, 10, (8, 6, 8, 8), device=device, dtype=torch.float) / 10.0 data.requires_grad_() data.retain_grad = True weighted_gradient = True # DDP forward/backward output = model(data) if weighted_gradient: output.backward(grad * 2 / global_batch_size) else: output.backward(grad / output.size(0)) d_list = [torch.randn(8, 3, 8, 8, device=device) for i in range(int(os.environ['WORLD_SIZE']))] y_list = [torch.randn(8, 6, 8, 8, device=device) for i in range(int(os.environ['WORLD_SIZE']))] dgrad_list = [torch.randn(8, 3, 8, 8, device=device) for i in range(int(os.environ['WORLD_SIZE']))] grad_list = [torch.randn(8, 6, 8, 8, device=device) for i in range(int(os.environ['WORLD_SIZE']))] if args.local_rank == 0: # placeholder, these random data will later be discarded. torch.distributed.all_gather(d_list, torch.randn(8, 3, 8, 8, device=device)) torch.distributed.all_gather(y_list, torch.randn(8, 6, 8, 8, device=device)) torch.distributed.all_gather(dgrad_list, torch.randn(8, 3, 8, 8, device=device)) torch.distributed.all_gather(grad_list, torch.randn(8, 6, 8, 8, device=device)) else: torch.distributed.all_gather(d_list, data) torch.distributed.all_gather(y_list, output) torch.distributed.all_gather(dgrad_list, data.grad) torch.distributed.all_gather(grad_list, grad) torch.distributed.barrier() if args.local_rank == 0: ref_tensor = d_list[1:] ref_tensor.insert(0, data) assert(ref_tensor[0].equal(data)) ref_tensor = torch.cat(ref_tensor, 0) ref_tensor = ref_tensor.detach() ref_tensor.requires_grad_() ref_tensor.retain_grad() # Reference forward/backward output_reference = model_reference(ref_tensor) grad_tensor = grad_list[1:] grad_tensor.insert(0, grad) assert(grad_tensor[0].equal(grad)) grad_tensor = torch.cat(grad_tensor, 0) if weighted_gradient: output_reference.backward(grad_tensor / output_reference.size(0)) else: output_reference.backward(grad_tensor / output_reference.size(0)) dgrad_tensor = dgrad_list[1:] dgrad_tensor.insert(0, data.grad) dgrad_tensor = torch.cat(dgrad_tensor, 0) # check output output_tensor = y_list[1:] output_tensor.insert(0, output) output_tensor = torch.cat(output_tensor, 0) passed = True passed = passed and compare("check output", output_tensor, output_reference) # check stats passed = passed and compare("check running mean failed", model_reference[1].running_mean, model.module[1].running_mean) passed = passed and compare("check running var failed", model_reference[1].running_var, model.module[1].running_var) passed = passed and compare("bn wgrad check failed!", model_reference[1].weight.grad, model.module[1].weight.grad, 1e-6) passed = passed and compare("conv wgrad check failed!", model_reference[0].weight.grad, model.module[0].weight.grad) # can't really compare dgrad directly, as we need to scale it to account for # DDP # passed = passed and compare("dgrad check failed!", ref_tensor.grad, dgrad_tensor) if passed: print("====SBN two gpu with different batches test passed") else: assert("*failed two gpu with different batches tests*")
apex-master
tests/distributed/synced_batchnorm/two_gpu_test_different_batch_size.py
import torch import numpy as np import apex if True: print("using setup tools") import syncbn else: print("using jit") from torch.utils.cpp_extension import load syncbn = load(name='syncbn', sources=['../../csrc/syncbn.cpp', '../../csrc/welford.cu']) def compare(desc, inp1, inp2, error): a = inp1.clone().detach().cpu().numpy() b = inp2.clone().detach().cpu().numpy() close = np.allclose(a,b, error, error) if not close: print(desc, close) z = a - b index = (np.abs(z) >= error + error * np.abs(b)).nonzero() print("dif : ", z[index]) print("inp1 : ", a[index]) print("inp2 : ", b[index]) return close feature_size = 10 space_size = 16 batch_size = 5 error = 1e-5 np.random.seed(1) dtype = np.float32 inp = (np.random.randn(batch_size, feature_size, space_size, space_size)).astype(dtype) grad = (np.random.randn(batch_size, feature_size, space_size, space_size)).astype(dtype) weight = (np.random.randn(feature_size)).astype(dtype) bias = (np.random.randn(feature_size)).astype(dtype) count = torch.cuda.IntTensor([batch_size*space_size**2]) type_tensor = torch.cuda.FloatTensor ref_tensor = torch.cuda.DoubleTensor inp_t = type_tensor(inp) weight_t = type_tensor(weight) bias_t = type_tensor(bias) inp_r = ref_tensor(inp.transpose(1, 0, 2, 3).reshape(feature_size, -1)) inp2_r = ref_tensor(inp) weight_r = ref_tensor(weight).view(-1, 1, 1) bias_r = ref_tensor(bias).view(-1, 1, 1) grad_output_t = type_tensor(grad) m = inp_r.mean(1) b_v = inp_r.var(1, unbiased=False) unb_v = inp_r.var(1, unbiased=True) eps = 1e-5 #mean, var, var_biased = syncbn.welford_mean_var(inp_t) mean, var_biased = syncbn.welford_mean_var(inp_t) inv_std = 1.0 / torch.sqrt(var_biased + eps) bn = torch.nn.BatchNorm2d(feature_size).cuda() bn.momentum = 1.0 bn.weight.data = weight_t.clone() bn.bias.data = bias_t.clone() inp_bn = inp_t.clone().requires_grad_() grad_bn = grad_output_t.clone().detach() out_bn = bn(inp_bn) out_bn.backward(grad_bn) sbn = apex.parallel.SyncBatchNorm(feature_size).cuda() sbn.momentum = 1.0 sbn.weight.data = weight_t.clone() sbn.bias.data = bias_t.clone() inp_sbn = inp_t.clone().requires_grad_() grad_sbn = grad_output_t.clone().detach() out_sbn = sbn(inp_sbn) out_sbn.backward(grad_sbn) sbn_c_last = apex.parallel.SyncBatchNorm(feature_size, channel_last=True).cuda() sbn_c_last.momentum = 1.0 sbn_c_last.weight.data = weight_t.clone() sbn_c_last.bias.data = bias_t.clone() inp_sbn_c_last = inp_t.clone().transpose(-1, 1).contiguous().requires_grad_() grad_sbn_c_last = grad_output_t.clone().transpose(-1, 1).contiguous().detach() out_sbn_c_last = sbn_c_last(inp_sbn_c_last) out_sbn_c_last.backward(grad_sbn_c_last) sbn_result = True sbn_result_c_last = True bn_result = True sbn_result = compare("comparing mean: ", mean, m, error) and sbn_result #sbn_result = compare("comparing variance: ", var, unb_v, error) and sbn_result sbn_result = compare("comparing biased variance: ", var_biased, b_v, error) and sbn_result out = syncbn.batchnorm_forward(inp_t, mean, inv_std, weight_t, bias_t) out_r = weight_r * (inp2_r - m.view(-1, 1, 1)) * torch.rsqrt(b_v.view(-1,1,1) + eps) + bias_r sbn_result = compare("comparing output: ", out, out_r, error) and sbn_result compare("comparing bn output: ", out_bn, out_r, error) grad_output_t = type_tensor(grad) grad_output_r = ref_tensor(grad.transpose(1, 0, 2, 3).reshape(feature_size, -1)) grad_output2_r = ref_tensor(grad) grad_bias_r = grad_output_r.sum(1) grad_weight_r = ((inp2_r - m.view(-1, 1, 1)) * torch.rsqrt(b_v.view(-1,1,1) + eps) * grad_output2_r).transpose(1,0).contiguous().view(feature_size, -1).sum(1) sum_dy_r = grad_output_r.sum(1) mean_dy_r = grad_output_r.mean(1) sum_dy_xmu_r = ((inp2_r - m.view(-1, 1, 1)) * grad_output2_r).transpose(1,0).contiguous().view(feature_size, -1).sum(1) mean_dy_xmu_r = ((inp2_r - m.view(-1, 1, 1)) * grad_output2_r).transpose(1,0).contiguous().view(feature_size, -1).mean(1) grad_input_r = (grad_output2_r - mean_dy_r.view(-1, 1, 1) - (inp2_r - m.view(-1, 1, 1)) / (b_v.view(-1,1,1) + eps) * mean_dy_xmu_r.view(-1, 1, 1) ) * torch.rsqrt(b_v.view(-1,1,1) + eps) * weight_r.view(-1,1,1) sum_dy, sum_dy_xmu, grad_weight, grad_bias = syncbn.reduce_bn(grad_output_t, inp_t, mean, inv_std, weight_t) grad_input = syncbn.batchnorm_backward(grad_output_t, inp_t, mean, inv_std, weight_t, sum_dy, sum_dy_xmu, count) sbn_result = compare("comparing bias grad: ", grad_bias, grad_bias_r, error) and sbn_result sbn_result = compare("comparing weight grad: ", grad_weight, grad_weight_r, error) and sbn_result sbn_result = compare("comparing sum_dy grad: ", sum_dy, sum_dy_r, error) and sbn_result sbn_result = compare("comparing sum_dy_xmu grad: ", sum_dy_xmu, sum_dy_xmu_r, error) and sbn_result sbn_result = compare("comparing input grad: ", grad_input, grad_input_r, error) and sbn_result compare("comparing bn input grad: ", inp_bn.grad, grad_input_r, error) sbn_result = compare("comparing sbn input grad: ", inp_sbn.grad, grad_input_r, error) and sbn_result compare("comparing bn/sbn output: ", out_bn, out_sbn, error) sbn_result = compare("comparing running_mean: ", bn.running_mean.data, sbn.running_mean.data, error) and sbn_result sbn_result = compare("comparing running_variance: ", bn.running_var.data, sbn.running_var.data, error) and sbn_result compare("comparing grad_input: ", inp_bn.grad, inp_sbn.grad, error) compare("comparing grad_bias: ", bn.bias.grad, sbn.bias.grad, error) compare("comparing grad_bias bn to ref: ", bn.bias.grad, grad_bias_r, error) sbn_result = compare("comparing grad_bias sbn to ref: ", sbn.bias.grad, grad_bias_r, error) and sbn_result compare("comparing grad_weight: ", bn.weight.grad, sbn.weight.grad, error) compare("comparing grad_weight bn to ref: ", bn.weight.grad, grad_weight_r, error) sbn_result = compare("comparing grad_weight sbn to ref: ", sbn.weight.grad, grad_weight_r, error) and sbn_result compare("comparing channel last bn/sbn output: ", out_bn, out_sbn_c_last.transpose(-1, 1).contiguous(), error) sbn_result_c_last = compare("comparing channel last running_mean: ", bn.running_mean.data, sbn_c_last.running_mean.data, error) and sbn_result_c_last sbn_result_c_last = compare("comparing channel last running_variance: ", bn.running_var.data, sbn_c_last.running_var.data, error) and sbn_result_c_last compare("comparing channel last grad_input: ", inp_bn.grad, inp_sbn_c_last.grad.transpose(-1, 1).contiguous(), error) compare("comparing channel last grad_bias: ", bn.bias.grad, sbn_c_last.bias.grad, error) sbn_result_c_last = compare("comparing channel last grad_bias sbn to ref: ", sbn_c_last.bias.grad, grad_bias_r, error) and sbn_result_c_last compare("comparing channel last grad_weight: ", bn.weight.grad, sbn_c_last.weight.grad, error) sbn_result_c_last = compare("comparing channel last grad_weight sbn to ref: ", sbn_c_last.weight.grad, grad_weight_r, error) and sbn_result_c_last if sbn_result: print("====SBN single gpu passed tests") else: print("*SBN single gpu failed*") if sbn_result_c_last: print("====SBN channel last single gpu passed tests") else: print("*SBN channel last single gpu failed*")
apex-master
tests/distributed/synced_batchnorm/single_gpu_unit_test.py
import torch import numpy as np import apex import syncbn import os import argparse import torch.optim as optim def compare(desc, inp1, inp2, error): a = inp1.clone().detach().cpu().numpy() b = inp2.clone().detach().cpu().numpy() close = np.allclose(a,b, error, error) if not close: print(desc, close) z = a - b index = (np.abs(z) >= error + error * np.abs(b)).nonzero() print("dif : ", z[index]) print("inp1 : ", a[index]) print("inp2 : ", b[index]) return close feature_size = 10 space_size = 40 batch_size = 32 from apex.parallel import DistributedDataParallel as DDP parser = argparse.ArgumentParser() parser.add_argument("--local_rank", default=0, type=int) parser.add_argument("--fp16", action='store_true', default=False) parser.add_argument("--fp64", action='store_true', default=False) parser.add_argument("--group_size", default=0, type=int) args = parser.parse_args() try: args.world_size = int(os.environ['WORLD_SIZE']) except: print("This is a multi-gpu test. To run it please use 'python -m torch.distributed.launch --nproc_per_node=<num gpus> test_groups.py <more options>'") exit(1) torch.cuda.set_device(args.local_rank) torch.distributed.init_process_group(backend='nccl', init_method='env://') start = (args.local_rank%args.group_size) * batch_size//args.group_size finish = (args.local_rank%args.group_size + 1) * batch_size//args.group_size error = 1e-5 dtype = np.float32 if args.fp16: error = 1e-3 dtype = np.float16 elif args.fp64: error = 1e-8 dtype = np.float64 np.random.seed(18 + args.local_rank//args.group_size) inp = np.random.randn(batch_size, feature_size, space_size, space_size).astype(dtype) grad = np.random.randn(batch_size, feature_size, space_size, space_size).astype(dtype) weight = np.random.randn(feature_size).astype(dtype) bias = np.random.randn(feature_size).astype(dtype) type_tensor = torch.cuda.FloatTensor if args.fp16: type_tensor = torch.cuda.HalfTensor if args.fp64: type_tensor = torch.cuda.DoubleTensor ref_tensor = torch.cuda.DoubleTensor inp_t = type_tensor(inp) weight_t = type_tensor(weight) bias_t = type_tensor(bias) inp_r = ref_tensor(inp.transpose(1, 0, 2, 3).reshape(feature_size, -1)) inp2_r = ref_tensor(inp) weight_r = ref_tensor(weight).view(-1, 1, 1) bias_r = ref_tensor(bias).view(-1, 1, 1) grad_output_t = type_tensor(grad) m = inp_r.mean(1) b_v = inp_r.var(1, unbiased=False) unb_v = inp_r.var(1, unbiased=True) eps = 1e-5 mean, var_biased = syncbn.welford_mean_var(inp_t) inv_std = 1.0 / torch.sqrt(var_biased + eps) bn = torch.nn.BatchNorm2d(feature_size).cuda() bn.momentum = 1.0 bn.weight.data = weight_t.clone() bn.bias.data = bias_t.clone() if args.fp16: bn.half() if args.fp64: bn.double() bn = DDP(bn) inp_bn = inp_t.clone().requires_grad_() grad_bn = grad_output_t.clone().detach() out_bn = bn(inp_bn) out_bn.backward(grad_bn) # compensating the averaging over processes done by DDP # in order to produce mathematically equivalent result # https://github.com/NVIDIA/apex/issues/134#issuecomment-458307368 for param in bn.parameters(): param.grad = param.grad / args.group_size bn_opt = optim.SGD(bn.parameters(), lr=1.0) sbn = apex.parallel.SyncBatchNorm(feature_size, process_group=apex.parallel.create_syncbn_process_group(args.group_size)).cuda() sbn.momentum = 1.0 sbn.weight.data = weight_t.clone() sbn.bias.data = bias_t.clone() if args.fp16: sbn.half() if args.fp64: sbn.double() sbn = DDP(sbn) sbn_opt = optim.SGD(sbn.parameters(), lr=1.0) inp_sbn = inp_t.clone().requires_grad_() grad_sbn = grad_output_t.clone().detach() out_sbn = sbn(inp_sbn[start:finish]) out_sbn.backward(grad_sbn[start:finish]) sbn_result = True bn_result = True if args.local_rank == 0: sbn_result = compare("comparing mean: ", mean, m, error) and sbn_result sbn_result = compare("comparing biased variance: ", var_biased, b_v, error) and sbn_result out = syncbn.batchnorm_forward(inp_t, mean, inv_std, weight_t, bias_t) out_r = weight_r * (inp2_r - m.view(-1, 1, 1)) * torch.rsqrt(b_v.view(-1,1,1) + eps) + bias_r if args.local_rank == 0: sbn_result = compare("comparing output: ", out, out_r, error) and sbn_result compare("comparing bn output: ", out_bn, out_r, error) grad_output_t = type_tensor(grad) grad_output_r = ref_tensor(grad.transpose(1, 0, 2, 3).reshape(feature_size, -1)) grad_output2_r = ref_tensor(grad) grad_bias_r = grad_output_r.sum(1) grad_weight_r = ((inp2_r - m.view(-1, 1, 1)) * torch.rsqrt(b_v.view(-1,1,1) + eps) * grad_output2_r).transpose(1,0).contiguous().view(feature_size, -1).sum(1) mean_dy_r = grad_output_r.mean(1) mean_dy_xmu_r = ((inp2_r - m.view(-1, 1, 1)) * grad_output2_r).transpose(1,0).contiguous().view(feature_size, -1).mean(1) grad_input_r = (grad_output2_r - mean_dy_r.view(-1, 1, 1) - (inp2_r - m.view(-1, 1, 1)) / (b_v.view(-1,1,1) + eps) * mean_dy_xmu_r.view(-1, 1, 1) ) * torch.rsqrt(b_v.view(-1,1,1) + eps) * weight_r.view(-1,1,1) mean_dy, mean_dy_xmu, grad_weight, grad_bias = syncbn.reduce_bn(grad_output_t, inp_t, mean, inv_std, weight_t) grad_input = syncbn.batchnorm_backward(grad_output_t, inp_t, mean, inv_std, weight_t, mean_dy, mean_dy_xmu) if args.local_rank == 0: sbn_result = compare("comparing bias grad: ", grad_bias, grad_bias_r, error) and sbn_result sbn_result = compare("comparing weight grad: ", grad_weight, grad_weight_r, error) and sbn_result sbn_result = compare("comparing mean_dy grad: ", mean_dy, mean_dy_r, error) and sbn_result sbn_result = compare("comparing mean_dy_xmu grad: ", mean_dy_xmu, mean_dy_xmu_r, error) and sbn_result sbn_result = compare("comparing input grad: ", grad_input, grad_input_r, error) and sbn_result compare("comparing bn input grad: ", inp_bn.grad, grad_input_r, error) if args.local_rank == 0: sbn_result = compare("comparing running_mean: ", bn.module.running_mean.data, sbn.module.running_mean.data, error) and sbn_result sbn_result = compare("comparing running_variance: ", bn.module.running_var.data, sbn.module.running_var.data, error) and sbn_result # execute by both compare("comparing layers output: ", out_bn[start:finish], out_sbn, error) and sbn_result compare("comparing layers grad_input: ", inp_bn.grad[start:finish], inp_sbn.grad[start:finish], error) and sbn_result bn_opt.step() sbn_opt.step() if args.local_rank == 0: compare("comparing bn vs sbn bias: ", bn.module.bias, sbn.module.bias, error) compare("comparing bn vs sbn weight: ", bn.module.weight, sbn.module.weight, error) if sbn_result: print("====SBN group test passed") else: print("*SBN group test failed*")
apex-master
tests/distributed/synced_batchnorm/test_groups.py
import torch import numpy as np import apex def compare(desc, inp1, inp2, error): a = inp1.clone().detach().cpu().numpy() b = inp2.clone().detach().cpu().numpy() close = np.allclose(a,b, error, error) if not close: print(desc, close) z = a - b index = (np.abs(z) >= error + error * np.abs(b)).nonzero() print("dif : ", z[index]) print("inp1 : ", a[index]) print("inp2 : ", b[index]) return close feature_size = 10 space_size = 16 batch_size = 5 error = 1e-5 np.random.seed(1) dtype = np.float32 inp = (np.random.randn(batch_size, feature_size, space_size, space_size)).astype(dtype) grad = (np.random.randn(batch_size, feature_size, space_size, space_size)).astype(dtype) weight = (np.random.randn(feature_size)).astype(dtype) bias = (np.random.randn(feature_size)).astype(dtype) type_tensor = torch.cuda.FloatTensor ref_tensor = torch.cuda.DoubleTensor inp_t = type_tensor(inp) weight_t = type_tensor(weight) bias_t = type_tensor(bias) inp_r = ref_tensor(inp.transpose(1, 0, 2, 3).reshape(feature_size, -1)) inp2_r = ref_tensor(inp) weight_r = ref_tensor(weight).view(-1, 1, 1) bias_r = ref_tensor(bias).view(-1, 1, 1) grad_output_t = type_tensor(grad) m = inp_r.mean(1) b_v = inp_r.var(1, unbiased=False) unb_v = inp_r.var(1, unbiased=True) eps = 1e-5 bn = torch.nn.BatchNorm2d(feature_size).cuda() bn.momentum = 1.0 bn.weight.data = weight_t.clone() bn.bias.data = bias_t.clone() inp_bn = inp_t.clone().requires_grad_() grad_bn = grad_output_t.clone().detach() out_bn = bn(inp_bn) out_bn.backward(grad_bn) from apex.parallel.sync_batchnorm import SyncBatchNorm sbn = SyncBatchNorm(feature_size).cuda() sbn.momentum = 1.0 sbn.weight.data = weight_t.clone() sbn.bias.data = bias_t.clone() inp_sbn = inp_t.clone().requires_grad_() grad_sbn = grad_output_t.clone().detach() out_sbn = sbn(inp_sbn) out_sbn.backward(grad_sbn) sbn_result = True sbn_result_c_last = True bn_result = True out_r = weight_r * (inp2_r - m.view(-1, 1, 1)) * torch.rsqrt(b_v.view(-1,1,1) + eps) + bias_r compare("comparing bn output: ", out_bn, out_r, error) grad_output_t = type_tensor(grad) grad_output_r = ref_tensor(grad.transpose(1, 0, 2, 3).reshape(feature_size, -1)) grad_output2_r = ref_tensor(grad) grad_bias_r = grad_output_r.sum(1) grad_weight_r = ((inp2_r - m.view(-1, 1, 1)) * torch.rsqrt(b_v.view(-1,1,1) + eps) * grad_output2_r).transpose(1,0).contiguous().view(feature_size, -1).sum(1) mean_dy_r = grad_output_r.mean(1) mean_dy_xmu_r = ((inp2_r - m.view(-1, 1, 1)) * grad_output2_r).transpose(1,0).contiguous().view(feature_size, -1).mean(1) grad_input_r = (grad_output2_r - mean_dy_r.view(-1, 1, 1) - (inp2_r - m.view(-1, 1, 1)) / (b_v.view(-1,1,1) + eps) * mean_dy_xmu_r.view(-1, 1, 1) ) * torch.rsqrt(b_v.view(-1,1,1) + eps) * weight_r.view(-1,1,1) compare("comparing bn input grad: ", inp_bn.grad, grad_input_r, error) sbn_result = compare("comparing sbn input grad: ", inp_sbn.grad, grad_input_r, error) and sbn_result compare("comparing bn/sbn output: ", out_bn, out_sbn, error) sbn_result = compare("comparing running_mean: ", bn.running_mean.data, sbn.running_mean.data, error) and sbn_result sbn_result = compare("comparing running_variance: ", bn.running_var.data, sbn.running_var.data, error) and sbn_result compare("comparing grad_input: ", inp_bn.grad, inp_sbn.grad, error) compare("comparing grad_bias: ", bn.bias.grad, sbn.bias.grad, error) compare("comparing grad_bias bn to ref: ", bn.bias.grad, grad_bias_r, error) sbn_result = compare("comparing grad_bias sbn to ref: ", sbn.bias.grad, grad_bias_r, error) and sbn_result compare("comparing grad_weight: ", bn.weight.grad, sbn.weight.grad, error) compare("comparing grad_weight bn to ref: ", bn.weight.grad, grad_weight_r, error) sbn_result = compare("comparing grad_weight sbn to ref: ", sbn.weight.grad, grad_weight_r, error) and sbn_result if sbn_result: print("====SBN single gpu passed tests") else: print("*SBN single gpu failed*")
apex-master
tests/distributed/synced_batchnorm/python_single_gpu_unit_test.py
"""L0 Tests Runner. How to run this script? 1. Run all the tests: `python /path/to/apex/tests/L0/run_test.py` If you want an xml report, pass `--xml-report`, i.e. `python /path/to/apex/tests/L0/run_test.py --xml-report` and the file is created in `/path/to/apex/tests/L0`. 2. Run one of the tests (e.g. fused layer norm): `python /path/to/apex/tests/L0/run_test.py --include run_fused_layer_norm` 3. Run two or more of the tests (e.g. optimizers and fused layer norm): `python /path/to/apex/tests/L0/run_test.py --include run_optimizers run_fused_layer_norm` """ import argparse import os import unittest import sys TEST_ROOT = os.path.dirname(os.path.abspath(__file__)) TEST_DIRS = [ "run_amp", "run_deprecated", "run_fp16util", "run_optimizers", "run_fused_layer_norm", "run_mlp", "run_transformer", ] DEFAULT_TEST_DIRS = [ "run_optimizers", "run_fused_layer_norm", "run_mlp", "run_transformer", ] def parse_args(): parser = argparse.ArgumentParser( description="L0 test runner", formatter_class=argparse.ArgumentDefaultsHelpFormatter, ) parser.add_argument( "--include", nargs="+", choices=TEST_DIRS, default=DEFAULT_TEST_DIRS, help="select a set of tests to run (defaults to ALL tests).", ) parser.add_argument( "--xml-report", default=None, action="store_true", help="[deprecated] pass this argument to get a junit xml report. Use `--xml-dir`. (requires `xmlrunner`)", ) parser.add_argument( "--xml-dir", default=None, type=str, help="Directory to save junit test reports. (requires `xmlrunner`)", ) args, _ = parser.parse_known_args() return args def main(args: argparse.Namespace) -> None: test_runner_kwargs = {"verbosity": 2} Runner = unittest.TextTestRunner xml_dir = None if (args.xml_report is not None) or (args.xml_dir is not None): if args.xml_report is not None: import warnings warnings.warn("The option of `--xml-report` is deprecated", FutureWarning) import xmlrunner from datetime import date # NOQA Runner = xmlrunner.XMLTestRunner if args.xml_report: xml_dir = os.path.abspath(os.path.dirname(__file__)) else: xml_dir = os.path.abspath(args.xml_dir) if not os.path.exists(xml_dir): os.makedirs(xml_dir) errcode = 0 for test_dir in args.include: if xml_dir is not None: xml_output = os.path.join( xml_dir, f"""TEST_{test_dir}_{date.today().strftime("%y%m%d")}""", ) if not os.path.exists(xml_output): os.makedirs(xml_output) test_runner_kwargs["output"] = xml_output runner = Runner(**test_runner_kwargs) test_dir = os.path.join(TEST_ROOT, test_dir) suite = unittest.TestLoader().discover(test_dir) print("\nExecuting tests from " + test_dir) result = runner.run(suite) if not result.wasSuccessful(): errcode = 1 sys.exit(errcode) if __name__ == '__main__': args = parse_args() main(args)
apex-master
tests/L0/run_test.py
import torch from apex.normalization import FusedLayerNorm from apex.normalization import FusedRMSNorm from apex.normalization import MixedFusedLayerNorm from apex.normalization import MixedFusedRMSNorm from torch.testing._internal import common_utils from torch.testing._internal.common_device_type import instantiate_device_type_tests from itertools import product def _prep_inputs(batch_size, normalized_shape, dtype): shape = (batch_size, *normalized_shape) fused = torch.randn(shape).cuda().requires_grad_(True) with torch.no_grad(): native = fused.clone().to(dtype).requires_grad_(True) return native, fused autocast_dtypes = (torch.half, torch.bfloat16) if torch.cuda.is_bf16_supported() else (torch.half,) class TestFusedLayerNorm(common_utils.TestCase): def _test_fused_layer_norm( self, batch_size, contiguous, elementwise_affine, mixed_fused, dtype, fwd_thresholds=dict(rtol=None, atol=None), bwd_thresholds=dict(rtol=None, atol=None) ): normalized_shape = [32, 16] if not mixed_fused: module_cpu_ = FusedLayerNorm( normalized_shape=normalized_shape, elementwise_affine=elementwise_affine).cpu() module_cuda_ = FusedLayerNorm( normalized_shape=normalized_shape, elementwise_affine=elementwise_affine).to(device="cuda", dtype=dtype) else: assert elementwise_affine module_cpu_ = MixedFusedLayerNorm( normalized_shape=normalized_shape).cpu() module_cuda_ = MixedFusedLayerNorm( normalized_shape=normalized_shape).to(device="cuda", dtype=dtype) torch.cuda.manual_seed(42) if contiguous: input_shape = [batch_size] + normalized_shape input_ = torch.randn(input_shape, device="cpu").requires_grad_(True) input_cuda_ = input_.to(device="cuda", dtype=dtype).detach().requires_grad_(True) self.assertTrue(input_.is_contiguous()) self.assertTrue(input_cuda_.is_contiguous()) else: input_shape = [batch_size] + normalized_shape input_shape = [batch_size * 3] + [normalized_shape[0] * 5, normalized_shape[1] * 3] input_src_ = torch.randn(input_shape, device="cpu") input_ = input_src_[::3, ::5, ::3].detach().requires_grad_(True) input_cuda_ = input_src_.to(device="cuda", dtype=dtype)[::3, ::5, ::3].detach().requires_grad_(True) # make sure that tensors are NOT contiguous. self.assertFalse(input_.is_contiguous()) self.assertFalse(input_cuda_.is_contiguous()) out_cpu_ = module_cpu_(input_) gO = torch.rand_like(out_cpu_) out_cpu_.backward(gO) out_cuda_ = module_cuda_(input_cuda_) gO = gO.to(device="cuda", dtype=dtype) out_cuda_.backward(gO) self.assertFalse(out_cpu_.is_cuda) self.assertTrue(out_cuda_.is_cuda) torch.testing.assert_close( out_cpu_.to(device="cuda", dtype=dtype), out_cuda_, **fwd_thresholds) torch.testing.assert_close( input_.grad.to(device="cuda", dtype=dtype), input_cuda_.grad, **bwd_thresholds) def _test_fused_rms_norm( self, batch_size, contiguous, elementwise_affine, mixed_fused, dtype, fwd_thresholds=dict(rtol=None, atol=None), bwd_thresholds=dict(rtol=None, atol=None) ): normalized_shape = [32, 16] if not mixed_fused: module_cpu_ = FusedRMSNorm( normalized_shape=normalized_shape, elementwise_affine=elementwise_affine).cpu() module_cuda_ = FusedRMSNorm( normalized_shape=normalized_shape, elementwise_affine=elementwise_affine).to(device="cuda", dtype=dtype) else: assert elementwise_affine module_cpu_ = MixedFusedRMSNorm( normalized_shape=normalized_shape).cpu() module_cuda_ = MixedFusedRMSNorm( normalized_shape=normalized_shape).to(device="cuda", dtype=dtype) torch.cuda.manual_seed(42) if contiguous: input_shape = [batch_size] + normalized_shape input_ = torch.randn(input_shape, device="cpu").requires_grad_(True) input_cuda_ = input_.to(device="cuda", dtype=dtype).detach().requires_grad_(True) self.assertTrue(input_.is_contiguous()) self.assertTrue(input_cuda_.is_contiguous()) else: input_shape = [batch_size] + normalized_shape input_shape = [batch_size * 3] + [normalized_shape[0] * 5, normalized_shape[1] * 3] input_src_ = torch.randn(input_shape, device="cpu") input_ = input_src_[::3, ::5, ::3].detach().requires_grad_(True) input_cuda_ = input_src_.to(device="cuda", dtype=dtype)[::3, ::5, ::3].detach().requires_grad_(True) # make sure that tensors are NOT contiguous. self.assertFalse(input_.is_contiguous()) self.assertFalse(input_cuda_.is_contiguous()) out_cpu_ = module_cpu_(input_) gO = torch.rand_like(out_cpu_) out_cpu_.backward(gO) out_cuda_ = module_cuda_(input_cuda_) torch.testing.assert_close( out_cpu_.to(device="cuda", dtype=dtype), out_cuda_.clone().detach(), **fwd_thresholds) gO = gO.to(device="cuda", dtype=dtype) out_cuda_.backward(gO) self.assertFalse(out_cpu_.is_cuda) self.assertTrue(out_cuda_.is_cuda) torch.testing.assert_close( input_.grad.to(device="cuda", dtype=dtype), input_cuda_.grad, **bwd_thresholds) if elementwise_affine: torch.testing.assert_close(module_cpu_.weight.grad.to(device="cuda", dtype=dtype), module_cuda_.weight.grad, **bwd_thresholds) # layer norm tests @common_utils.parametrize( "batch_size, contiguous, elementwise_affine, mixed_fused, dtype", list(product((16, 65536), (True, False), (False,), (False,), (torch.float,))) ) def test_layer_norm_regular(self, batch_size, contiguous, elementwise_affine, mixed_fused, dtype): self._test_fused_layer_norm(batch_size, contiguous, elementwise_affine, mixed_fused, dtype) @common_utils.parametrize( "batch_size, contiguous, elementwise_affine, mixed_fused, dtype", list(product((16, 65536), (True, False), (True,), (False,), (torch.float,))) ) def test_layer_norm_elemwise(self, batch_size, contiguous, elementwise_affine, mixed_fused, dtype): self._test_fused_layer_norm(batch_size, contiguous, elementwise_affine, mixed_fused, dtype) @common_utils.parametrize( "batch_size, contiguous, elementwise_affine, mixed_fused, dtype", list(product((16, 65536), (True, False), (True,), (True,), (torch.float,))) ) def test_layer_norm_mixed(self, batch_size, contiguous, elementwise_affine, mixed_fused, dtype): self._test_fused_layer_norm(batch_size, contiguous, elementwise_affine, mixed_fused, dtype) @common_utils.parametrize( "batch_size, contiguous, elementwise_affine, mixed_fused, dtype", list(product((16,), (True, False), (True,), (False,), (torch.half,))) ) def test_layer_norm_half(self, batch_size, contiguous, elementwise_affine, mixed_fused, dtype): self._test_fused_layer_norm(batch_size, contiguous, elementwise_affine, mixed_fused, dtype, fwd_thresholds=dict(rtol=1e-3, atol=1e-3), bwd_thresholds=dict(rtol=1e-3, atol=1e-3)) @common_utils.parametrize( "batch_size, contiguous, elementwise_affine, mixed_fused, dtype", list(product((16,), (True, False), (True,), (False,), (torch.bfloat16,))) ) def test_layer_norm_bfloat16(self, batch_size, contiguous, elementwise_affine, mixed_fused, dtype): self._test_fused_layer_norm(batch_size, contiguous, elementwise_affine, mixed_fused, dtype, fwd_thresholds=dict(rtol=1.6e-2, atol=3e-4), bwd_thresholds=dict(rtol=1.6e-2, atol=3e-3)) # rms norm tests @common_utils.parametrize( "batch_size, contiguous, elementwise_affine, mixed_fused, dtype", list(product((16, 65536), (True, False), (False,), (False,), (torch.float,))) ) def test_rms_norm_regular(self, batch_size, contiguous, elementwise_affine, mixed_fused, dtype): self._test_fused_rms_norm(batch_size, contiguous, elementwise_affine, mixed_fused, dtype) @common_utils.parametrize( "batch_size, contiguous, elementwise_affine, mixed_fused, dtype", list(product((16, 65536), (True, False), (True,), (False,), (torch.float,))) ) def test_rms_norm_elemwise(self, batch_size, contiguous, elementwise_affine, mixed_fused, dtype): self._test_fused_rms_norm(batch_size, contiguous, elementwise_affine, mixed_fused, dtype, bwd_thresholds=dict(rtol=2e-3, atol=2e-4)) @common_utils.parametrize( "batch_size, contiguous, elementwise_affine, mixed_fused, dtype", list(product((16, 65536), (True, False), (True,), (True,), (torch.float,))) ) def test_rms_norm_mixed(self, batch_size, contiguous, elementwise_affine, mixed_fused, dtype): self._test_fused_rms_norm(batch_size, contiguous, elementwise_affine, mixed_fused, dtype, bwd_thresholds=dict(rtol=2e-3, atol=2e-4)) @common_utils.parametrize( "batch_size, contiguous, elementwise_affine, mixed_fused, dtype", list(product((16,), (True, False), (True,), (False,), (torch.half,))) ) def test_rms_norm_half(self, batch_size, contiguous, elementwise_affine, mixed_fused, dtype): self._test_fused_rms_norm(batch_size, contiguous, elementwise_affine, mixed_fused, dtype, bwd_thresholds = dict(rtol=1.6e-2, atol=3e-3)) @common_utils.parametrize( "batch_size, contiguous, elementwise_affine, mixed_fused, dtype", list(product((16,), (True, False), (True,), (False,), (torch.bfloat16,))) ) def test_rms_norm_bfloat16(self, batch_size, contiguous, elementwise_affine, mixed_fused, dtype): self._test_fused_rms_norm(batch_size, contiguous, elementwise_affine, mixed_fused, dtype, fwd_thresholds=dict(rtol=1.6e-2, atol=3e-4), bwd_thresholds=dict(rtol=1.6e-2, atol=3e-2)) @common_utils.parametrize( "dtype, elementwise_affine", list(product(autocast_dtypes, (True, False))) ) def test_autocast_fused_layer_norm(self, dtype, elementwise_affine): bf16_fwd_thresholds = dict(rtol=1.6e-2, atol=3e-4) bf16_bwd_thresholds = dict(rtol=1.6e-2, atol=3e-3) batch_size = 16 normalized_shape = [32, 16] native = torch.nn.LayerNorm( normalized_shape=normalized_shape, elementwise_affine=elementwise_affine ).to(device="cuda", dtype=dtype) fused = FusedLayerNorm( normalized_shape=normalized_shape, elementwise_affine=elementwise_affine ).cuda() native_x, fused_x = _prep_inputs(batch_size, normalized_shape, dtype) expected = native(native_x) with torch.cuda.amp.autocast(dtype=dtype): actual = fused(fused_x) tols = {'rtol': None, 'atol': None} if dtype == torch.half else bf16_fwd_thresholds # original tests used torch.testing.assert_allclose, which disables dtype checking by default. # link to issue here: https://github.com/pytorch/pytorch/issues/61844 torch.testing.assert_close(actual, expected, **tols, check_dtype=False) g_native = torch.rand_like(expected) with torch.no_grad(): g_fused = g_native.clone() expected.backward(g_native) actual.backward(g_fused) tols = {'rtol': None, 'atol': None} if dtype == torch.half else bf16_bwd_thresholds torch.testing.assert_close(native_x.grad, fused_x.grad, **tols, check_dtype=False) @common_utils.parametrize( "dtype, elementwise_affine", list(product(autocast_dtypes, (True, False))) ) def test_autocast_fused_rms_norm(self, dtype, elementwise_affine): bf16_fwd_thresholds = dict(rtol=1.6e-2, atol=3e-4) bf16_bwd_thresholds = dict(rtol=1.6e-2, atol=3e-3) batch_size = 16 normalized_shape = [32, 16] native = FusedRMSNorm( normalized_shape=normalized_shape, elementwise_affine=elementwise_affine ).to(dtype=dtype) fused = FusedRMSNorm( normalized_shape=normalized_shape, elementwise_affine=elementwise_affine ).cuda() native_x, fused_x = _prep_inputs(batch_size, normalized_shape, dtype) expected = native(native_x.cpu()) with torch.cuda.amp.autocast(dtype=dtype): actual = fused(fused_x) tols = {'rtol': None, 'atol': None} if dtype == torch.half else bf16_fwd_thresholds torch.testing.assert_close(actual, expected.detach().clone().cuda(), **tols, check_dtype=False) g_native = torch.rand_like(expected) with torch.no_grad(): g_fused = g_native.detach().clone().cuda() expected.backward(g_native) actual.backward(g_fused) tols = {'rtol': 1e-3, 'atol': 1e-3} if dtype == torch.half else bf16_bwd_thresholds torch.testing.assert_close(native_x.grad.cuda(), fused_x.grad, **tols, check_dtype=False) def _verify_export(self, fused, fused_x): # check that export() is working onnx_str = torch.onnx.export_to_pretty_string(fused, (fused_x,), input_names=['x_in'], opset_version=18, ) assert 'x_in' in onnx_str assert 'ReduceMean' in onnx_str or 'LayerNormalization' in onnx_str def test_rms_export(self): batch_size = 16 normalized_shape = [32, 16] fused = FusedRMSNorm( normalized_shape=normalized_shape, elementwise_affine=True ).cuda() fused_m = MixedFusedRMSNorm( normalized_shape=normalized_shape ).cuda() native_x, fused_x = _prep_inputs(batch_size, normalized_shape, torch.float32) self._verify_export(fused, fused_x) self._verify_export(fused_m, fused_x) def test_layer_norm_export(self): batch_size = 16 normalized_shape = [32, 16] fused = FusedLayerNorm( normalized_shape=normalized_shape, elementwise_affine=True ).cuda() fused_m = MixedFusedLayerNorm( normalized_shape=normalized_shape ).cuda() native_x, fused_x = _prep_inputs(batch_size, normalized_shape, torch.float32) self._verify_export(fused, fused_x) self._verify_export(fused_m, fused_x) instantiate_device_type_tests(TestFusedLayerNorm, globals(), only_for=("cuda",)) if __name__ == "__main__": common_utils.run_tests()
apex-master
tests/L0/run_fused_layer_norm/test_fused_layer_norm.py
import unittest import os import torch from torch.optim import Optimizer import apex from apex.multi_tensor_apply import multi_tensor_applier from itertools import product class RefLAMB(Optimizer): r"""Implements Lamb algorithm. It has been proposed in `Large Batch Optimization for Deep Learning: Training BERT in 76 minutes`_. Arguments: params (iterable): iterable of parameters to optimize or dicts defining parameter groups lr (float, optional): learning rate (default: 1e-3) betas (Tuple[float, float], optional): coefficients used for computing running averages of gradient and its square (default: (0.9, 0.999)) eps (float, optional): term added to the denominator to improve numerical stability (default: 1e-6) weight_decay (float, optional): weight decay (L2 penalty) (default: 0.01) .. _Large Batch Optimization for Deep Learning: Training BERT in 76 minutes: https://arxiv.org/abs/1904.00962 """ def __init__(self, params, lr=1e-3, betas=(0.9, 0.999), eps=1e-6, weight_decay=0.01): if not 0.0 <= lr: raise ValueError("Invalid learning rate: {}".format(lr)) if not 0.0 <= eps: raise ValueError("Invalid epsilon value: {}".format(eps)) if not 0.0 <= betas[0] < 1.0: raise ValueError("Invalid beta parameter at index 0: {}".format(betas[0])) if not 0.0 <= betas[1] < 1.0: raise ValueError("Invalid beta parameter at index 1: {}".format(betas[1])) defaults = dict(lr=lr, betas=betas, eps=eps, weight_decay=weight_decay) super(RefLAMB, self).__init__(params, defaults) if multi_tensor_applier.available: import amp_C self.multi_tensor_l2norm=amp_C.multi_tensor_l2norm # Skip buffer self._dummy_overflow_buf = torch.tensor([0], dtype=torch.int, device=self.param_groups[0]["params"][0].device) self.multi_tensor_lamb = amp_C.multi_tensor_lamb else: raise RuntimeError('apex.optimizers.FusedLAMB requires cuda extensions') def step(self, closure=None): """Performs a single optimization step. Arguments: closure (callable, optional): A closure that reevaluates the model and returns the loss. """ loss = None if closure is not None: loss = closure() # create separate grad lists for fp32, fp16, and bf16 params g_all_32, g_all_16, g_all_bf16 = [], [], [] for group in self.param_groups: for p in group['params']: if p.grad is None: continue if p.dtype == torch.float32: g_all_32.append(p.grad.data) elif p.dtype == torch.float16: g_all_16.append(p.grad.data) elif p.dtype == torch.bfloat16: g_all_bf16.append(p.grad.data) else: raise RuntimeError('FusedLAMB only support fp16, fp32, and bf16.') device = self.param_groups[0]["params"][0].device g_norm_32, g_norm_16, g_norm_bf16 = torch.zeros(1, device=device), torch.zeros(1, device=device), torch.zeros(1, device=device) # compute grad norm for two lists if len(g_all_32) > 0: g_norm_32 = multi_tensor_applier(self.multi_tensor_l2norm, self._dummy_overflow_buf, [g_all_32], False)[0] if len(g_all_16) > 0: g_norm_16 = multi_tensor_applier(self.multi_tensor_l2norm, self._dummy_overflow_buf, [g_all_16], False)[0] if len(g_all_bf16) > 0: g_norm_bf16 = multi_tensor_applier(self.multi_tensor_l2norm, self._dummy_overflow_buf, [g_all_bf16], False)[0] # blend two grad norms to get global grad norm global_grad_norm = multi_tensor_applier(self.multi_tensor_l2norm, self._dummy_overflow_buf, [[g_norm_32, g_norm_16, g_norm_bf16]], False)[0] max_grad_norm = 1.0 clipped_ratio = max_grad_norm / max(global_grad_norm, max_grad_norm) for group in self.param_groups: for p in group['params']: if p.grad is None: continue p.grad.data *= clipped_ratio grad = p.grad.data if grad.is_sparse: raise RuntimeError('Lamb does not support sparse gradients, consider SparseAdam instad.') state = self.state[p] # State initialization if len(state) == 0: state['step'] = 0 # Exponential moving average of gradient values state['m'] = torch.zeros_like(p.data) # Exponential moving average of squared gradient values state['v'] = torch.zeros_like(p.data) m_t, v_t = state['m'], state['v'] beta1, beta2 = group['betas'] state['step'] += 1 # m_t = beta1 * m + (1 - beta1) * g_t m_t.mul_(beta1).add_(grad, alpha=1-beta1) # v_t = beta2 * v + (1 - beta2) * (g_t * g_t) if len(g_all_16) > 0: v_t.mul_(beta2) v_t = v_t.to(torch.float32) grad32 = grad.to(torch.float32) v_t.addcmul_(grad32, grad32, value=1-beta2) else: v_t.mul_(beta2).addcmul_(grad, grad, value=1-beta2) # Debiasing m_t_hat = m_t / (1.0 - beta1 ** state['step']) v_t_hat = v_t / (1.0 - beta2 ** state['step']) update = m_t_hat / v_t_hat.sqrt().add(group['eps']) if group['weight_decay'] != 0: update.add_(p.data, alpha=group['weight_decay']) trust_ratio = 1.0 w_norm = p.data.to(torch.float32).pow(2).sum().sqrt() g_norm = update.pow(2).sum().sqrt() if w_norm > 0 and g_norm > 0: trust_ratio = w_norm / g_norm state['w_norm'] = w_norm state['g_norm'] = g_norm state['trust_ratio'] = trust_ratio step_size = group['lr'] p.data.add_(update, alpha=-step_size*trust_ratio) return loss class TestLamb(unittest.TestCase): def setUp(self, max_abs_diff=1e-3, max_rel_diff=1, iters=7): self.max_abs_diff = max_abs_diff self.max_rel_diff = max_rel_diff self.iters = iters torch.cuda.manual_seed(9876) def tearDown(self): pass def gen_param_optim(self, tensors, lamb_option): ref_param = [] tst_param = [] for tensor in tensors: ref_param.append(torch.nn.Parameter(tensor.clone())) tst_param.append(torch.nn.Parameter(tensor.clone())) ref_optim = self.ref_optim(ref_param, **lamb_option) tst_optim = self.tst_optim(tst_param, use_nvlamb=True, **lamb_option) return (ref_param, tst_param, ref_optim, tst_optim) def gen_grad(self, ref_param, tst_param): for p_ref, p_tst in zip(ref_param, tst_param): p_ref.grad = torch.rand_like(p_ref) p_tst.grad = p_ref.grad def gen_mixed_grad(self, ref_param, tst_param, scale=1.0): half_grads = [] for p_ref, _ in zip(ref_param, tst_param): half_grads.append(torch.rand_like(p_ref).half()) p_ref.grad = half_grads[-1].float() / scale return half_grads def get_max_diff(self, ref_param, tst_param): max_abs_diff = max_rel_diff = 0 for p_ref, p_tst in zip(ref_param, tst_param): max_abs_diff_p = (p_ref - p_tst).abs().max().item() max_rel_diff_p = ((p_ref - p_tst) / p_ref).abs().max().item() if max_abs_diff_p > max_abs_diff: max_abs_diff = max_abs_diff_p if max_rel_diff_p > max_rel_diff: max_rel_diff = max_rel_diff_p return max_abs_diff, max_rel_diff def gen_single_type_test(self, param_type=torch.float, device="cuda"): nelem = 18011 tensor = torch.rand(nelem, dtype=param_type, device=device) weight_decay = [0, 0.01] for wd in weight_decay: lamb_option = {'lr':5e-4, 'betas':(0.9, 0.999), 'eps':1e-08, 'weight_decay':wd} ref_param, tst_param, ref_optim, tst_optim = \ self.gen_param_optim([tensor], lamb_option) if isinstance(tst_optim, apex.optimizers.FusedMixedPrecisionLamb): if param_type != torch.float: # joseli: This parameter is usually passed into the constructor, # but I do not want to change the testing interface. # As long as this parameter is set before the first call to step(), # then it should act normally. tst_optim.reduced_precision_dtype = param_type for i in range(self.iters): self.gen_grad(ref_param, tst_param) ref_optim.step() torch.cuda.synchronize() tst_optim.step() torch.cuda.synchronize() torch.testing.assert_close(tst_param, ref_param) class TestFusedLAMB(TestLamb): def __init__(self, *args, **kwargs): super(TestLamb, self).__init__(*args, **kwargs) self.ref_optim = RefLAMB self.tst_optim = apex.optimizers.FusedLAMB def test_float(self): self.gen_single_type_test(param_type=torch.float) @unittest.skip("PyTorch optimizer is not numerically correct for fp16") def test_half(self): self.gen_single_type_test(param_type=torch.float16) @unittest.skipIf(torch.cuda.device_count()<2, "more than 1 GPU required") def test_multi_device(self): devices = ("cuda:0", "cuda:1") for current_dev, tensor_dev in product(devices, devices): with torch.cuda.device(current_dev): self.gen_single_type_test(param_type=torch.float, device=tensor_dev) def test_multi_params(self): sizes = [[4096, 1024], [4096], [4096, 2048], [32320, 1024], [1]] weight_decay = [0, 0.01] for wd in weight_decay: lamb_option = {'lr':5e-4, 'betas':(0.9, 0.999), 'eps':1e-08, 'weight_decay':wd} tensors = [] for size in sizes: tensors.append(torch.rand(size, dtype=torch.float, device='cuda')) ref_param, tst_param, ref_optim, tst_optim = \ self.gen_param_optim(tensors, lamb_option) for i in range(self.iters): self.gen_grad(ref_param, tst_param) ref_optim.step() tst_optim.step() max_abs_diff, max_rel_diff = self.get_max_diff(ref_param, tst_param) self.assertLessEqual(max_abs_diff, self.max_abs_diff) self.assertLessEqual(max_rel_diff, self.max_rel_diff) def test_lamb_option(self): nelem = 1 tensor = torch.rand(nelem, dtype=torch.float, device='cuda') weight_decay = [0, 0.01] for wd in weight_decay: lamb_option = {'lr':0.01, 'betas':(0.6, 0.9), 'eps':3e-06, 'weight_decay':wd} ref_param, tst_param, ref_optim, tst_optim = \ self.gen_param_optim([tensor], lamb_option) for i in range(self.iters): self.gen_grad(ref_param, tst_param) ref_optim.step() tst_optim.step() max_abs_diff, max_rel_diff = self.get_max_diff(ref_param, tst_param) self.assertLessEqual(max_abs_diff, self.max_abs_diff) self.assertLessEqual(max_rel_diff, self.max_rel_diff) class TestFusedMixedPrecisionLamb(TestLamb): def __init__(self, *args, **kwargs): super(TestLamb, self).__init__(*args, **kwargs) self.ref_optim = RefLAMB self.tst_optim = apex.optimizers.FusedMixedPrecisionLamb def test_float(self): self.gen_single_type_test(param_type=torch.float) def test_bfloat16(self): self.iters = 4 self.gen_single_type_test(param_type=torch.bfloat16) def test_half(self): self.iters = 1 self.gen_single_type_test(param_type=torch.float16) @unittest.skipIf(torch.cuda.device_count()<2, "more than 1 GPU required") def test_multi_device(self): devices = ("cuda:0", "cuda:1") for current_dev, tensor_dev in product(devices, devices): with torch.cuda.device(current_dev): self.gen_single_type_test(param_type=torch.float, device=tensor_dev) def test_multi_params(self): sizes = [[4096, 1024], [4096], [4096, 2048], [32320, 1024], [1]] weight_decay = [0, 0.01] for wd in weight_decay: lamb_option = {'lr':5e-4, 'betas':(0.9, 0.999), 'eps':1e-08, 'weight_decay':wd} tensors = [] for size in sizes: tensors.append(torch.rand(size, dtype=torch.float, device='cuda')) ref_param, tst_param, ref_optim, tst_optim = \ self.gen_param_optim(tensors, lamb_option) for i in range(self.iters): self.gen_grad(ref_param, tst_param) ref_optim.step() tst_optim.step() max_abs_diff, max_rel_diff = self.get_max_diff(ref_param, tst_param) self.assertLessEqual(max_abs_diff, self.max_abs_diff) self.assertLessEqual(max_rel_diff, self.max_rel_diff) def test_lamb_option(self): nelem = 1 tensor = torch.rand(nelem, dtype=torch.float, device='cuda') weight_decay = [0, 0.01] for wd in weight_decay: lamb_option = {'lr':0.01, 'betas':(0.6, 0.9), 'eps':3e-06, 'weight_decay':wd} ref_param, tst_param, ref_optim, tst_optim = \ self.gen_param_optim([tensor], lamb_option) for i in range(self.iters): self.gen_grad(ref_param, tst_param) ref_optim.step() tst_optim.step() max_abs_diff, max_rel_diff = self.get_max_diff(ref_param, tst_param) self.assertLessEqual(max_abs_diff, self.max_abs_diff) self.assertLessEqual(max_rel_diff, self.max_rel_diff) if __name__ == '__main__': script_path = os.path.dirname(os.path.realpath(__file__)) unittest.main()
apex-master
tests/L0/run_optimizers/test_lamb.py
apex-master
tests/L0/run_optimizers/__init__.py
import copy import math import random import unittest import torch import torch.nn.functional as F from torch import nn try: import apex except ImportError as e: HAS_APEX = False else: HAS_APEX = True class Model(torch.nn.Module): def __init__(self): super(Model, self).__init__() self.conv1 = nn.Conv2d(1, 6, 5) self.relu1 = nn.ReLU() self.pool1 = nn.MaxPool2d(2) self.conv2 = nn.Conv2d(6, 16, 5) self.relu2 = nn.ReLU() self.pool2 = nn.MaxPool2d(2) self.fc1 = nn.Linear(256, 120) self.relu3 = nn.ReLU() self.fc2 = nn.Linear(120, 84) self.relu4 = nn.ReLU() self.fc3 = nn.Linear(84, 10) self.relu5 = nn.ReLU() def forward(self, x): y = self.conv1(x) y = self.relu1(y) y = self.pool1(y) y = self.conv2(y) y = self.relu2(y) y = self.pool2(y) y = y.reshape(y.shape[0], -1) y = self.fc1(y) y = self.relu3(y) y = self.fc2(y) y = self.relu4(y) y = self.fc3(y) y = self.relu5(y) return y @unittest.skipIf(not HAS_APEX, "`apex` is not found.") class AdamTest(unittest.TestCase): def setUp(self, seed=0): super().setUp() torch.manual_seed(seed) self.model = Model().cuda() self.model_ = Model().cuda() self.model_.load_state_dict(copy.deepcopy(self.model.state_dict())) self.lr = 0.00001 params = [p for p in self.model.parameters() if p.requires_grad] self.optimizer = torch.optim.Adam(params, lr=self.lr) def testGradScaler(self): params_ = [p for p in self.model_.parameters() if p.requires_grad] optimizer_ = apex.optimizers.FusedAdam(params_, lr=self.lr, capturable=False) scaler = torch.cuda.amp.GradScaler(enabled=True) scaler_ = torch.cuda.amp.GradScaler(enabled=True) for i in range(100): x = torch.rand([32, 1, 28, 28]).cuda().to(memory_format=torch.channels_last) x_ = x.clone() gt = torch.rand([32, 10]).cuda() gt_ = gt.clone() # Reference with torch.cuda.amp.autocast(enabled=True): y = self.model(x) loss = ((gt - y) ** 2).mean() scaler.scale(loss).backward() scaler.step(self.optimizer) scaler.update() # DUT with torch.cuda.amp.autocast(enabled=True): y = self.model_(x) loss_ = ((gt_ - y) ** 2).mean() scaler_.scale(loss_).backward() scaler_.step(optimizer_) scaler_.update() for module in zip(self.model.modules(), self.model_.modules()): m = module[0] m_ = module[1] if isinstance(m, nn.Conv2d) or isinstance(m_, nn.Linear): torch.testing.assert_close(m.weight, m_.weight, atol=1e-3, rtol=1e-3, equal_nan=True) torch.testing.assert_close(m.weight.grad, m_.weight.grad, atol=1e-3, rtol=1e-3, equal_nan=True) # Init for next iteration self.optimizer.zero_grad() optimizer_.zero_grad() self.model_.load_state_dict(copy.deepcopy(self.model.state_dict())) def testGradScalerCapturable(self): params_ = [p for p in self.model_.parameters() if p.requires_grad] optimizer_ = apex.optimizers.FusedAdam(params_, lr=self.lr, capturable=True) scaler = torch.cuda.amp.GradScaler(enabled=True) scaler_ = torch.cuda.amp.GradScaler(enabled=True) for i in range(100): x = torch.rand([32, 1, 28, 28]).cuda().to(memory_format=torch.channels_last) x_ = x.clone() gt = torch.rand([32, 10]).cuda() gt_ = gt.clone() # Reference with torch.cuda.amp.autocast(enabled=True): y = self.model(x) loss = ((gt - y) ** 2).mean() scaler.scale(loss).backward() scaler.step(self.optimizer) scaler.update() # DUT with torch.cuda.amp.autocast(enabled=True): y = self.model_(x) loss_ = ((gt_ - y) ** 2).mean() scaler_.scale(loss_).backward() scaler_.step(optimizer_) scaler_.update() for module in zip(self.model.modules(), self.model_.modules()): m = module[0] m_ = module[1] if isinstance(m, nn.Conv2d) or isinstance(m_, nn.Linear): torch.testing.assert_close(m.weight, m_.weight, atol=1e-3, rtol=1e-3, equal_nan=True) torch.testing.assert_close(m.weight.grad, m_.weight.grad, atol=1e-3, rtol=1e-3, equal_nan=True) # Init for next iteration self.optimizer.zero_grad() optimizer_.zero_grad() self.model_.load_state_dict(copy.deepcopy(self.model.state_dict())) def testGradScalerCapturableMaster(self): # Cast conv layers to FP16 for m in self.model_.modules(): if m.__class__ in [torch.nn.Conv2d]: m.half() params_ = [p for p in self.model_.parameters() if p.requires_grad] optimizer_ = apex.optimizers.FusedAdam(params_, lr=self.lr, capturable=True, master_weights=True) scaler = torch.cuda.amp.GradScaler(enabled=True) scaler_ = torch.cuda.amp.GradScaler(enabled=True) for i in range(100): x = torch.rand([32, 1, 28, 28]).cuda().to(memory_format=torch.channels_last) x_ = x.clone() gt = torch.rand([32, 10]).cuda() gt_ = gt.clone() # Reference with torch.cuda.amp.autocast(enabled=True): y = self.model(x) loss = ((gt - y) ** 2).mean() scaler.scale(loss).backward() scaler.step(self.optimizer) scaler.update() # DUT with torch.cuda.amp.autocast(enabled=True): y = self.model_(x) loss_ = ((gt_ - y) ** 2).mean() scaler_.scale(loss_).backward() scaler_.step(optimizer_) scaler_.update() for module in zip(self.model.modules(), self.model_.modules()): m = module[0] m_ = module[1] if isinstance(m, nn.Conv2d) or isinstance(m_, nn.Linear): torch.testing.assert_close(m.weight, m_.weight.float(), atol=1e-3, rtol=1e-3, equal_nan=True) torch.testing.assert_close(m.weight.grad, m_.weight.grad.float(), atol=1e-3, rtol=1e-3, equal_nan=True) # Init for next iteration self.optimizer.zero_grad() optimizer_.zero_grad() self.model_.load_state_dict(copy.deepcopy(self.model.state_dict())) def testNative(self): params_ = [p for p in self.model_.parameters() if p.requires_grad] optimizer_ = apex.optimizers.FusedAdam(params_, lr=self.lr, capturable=False) for i in range(100): x = torch.rand([32, 1, 28, 28]).cuda().to(memory_format=torch.channels_last) x_ = x.clone() gt = torch.rand([32, 10]).cuda() gt_ = gt.clone() # Reference y = self.model(x) loss = ((gt - y) ** 2).mean() loss.backward() self.optimizer.step() # DUT y = self.model_(x) loss_ = ((gt_ - y) ** 2).mean() loss_.backward() optimizer_.step() for module in zip(self.model.modules(), self.model_.modules()): m = module[0] m_ = module[1] if isinstance(m, nn.Conv2d) or isinstance(m_, nn.Linear): torch.testing.assert_close(m.weight, m_.weight, atol=1e-3, rtol=1e-3, equal_nan=True) torch.testing.assert_close(m.weight.grad, m_.weight.grad, atol=1e-3, rtol=1e-3, equal_nan=True) # Init for next iteration self.optimizer.zero_grad() optimizer_.zero_grad() self.model_.load_state_dict(copy.deepcopy(self.model.state_dict())) if __name__ == '__main__': unittest.main()
apex-master
tests/L0/run_optimizers/test_adam.py
from itertools import product import random import unittest import torch import apex class TestFusedOptimizer(unittest.TestCase): def setUp(self, max_abs_diff=1e-3, max_rel_diff=1, iters=7): self.max_abs_diff = max_abs_diff self.max_rel_diff = max_rel_diff self.iters = iters torch.manual_seed(9876) def tearDown(self): pass def gen_param_optim(self, tensors, options, tst_options=None): # Adding this to make backward compatible with existing tests. Just in # case "tst_options" are not provided, it gets a copy of options # which contains the parameters for the reference optimizer if tst_options == None: tst_options = options ref_param = [] tst_param = [] for tensor in tensors: ref_param.append(torch.nn.Parameter(tensor.clone())) tst_param.append(torch.nn.Parameter(tensor.clone())) ref_optim = self.ref_optim(ref_param, **options) tst_optim = self.fused_optim(tst_param, **tst_options) return (ref_param, tst_param, ref_optim, tst_optim) def gen_grad(self, ref_param, tst_param): for p_ref, p_tst in zip(ref_param, tst_param): p_ref.grad = torch.rand_like(p_ref) p_tst.grad = p_ref.grad def gen_mixed_grad(self, ref_param, tst_param, scale=1.0): half_grads = [] for p_ref, p_tst in zip(ref_param, tst_param): half_grads.append(torch.rand_like(p_ref).half()) p_ref.grad = half_grads[-1].float() / scale return half_grads def get_max_diff(self, ref_param, tst_param): max_abs_diff = max_rel_diff = 0 for p_ref, p_tst in zip(ref_param, tst_param): max_abs_diff_p = (p_ref - p_tst).abs().max().item() max_rel_diff_p = ((p_ref - p_tst) / p_ref).abs().max().item() if max_abs_diff_p > max_abs_diff: max_abs_diff = max_abs_diff_p if max_rel_diff_p > max_rel_diff: max_rel_diff = max_rel_diff_p return max_abs_diff, max_rel_diff def gen_single_type_test(self, param_type=torch.float, device='cuda', *, skip_assert: bool = False): nelem = 278011 # Some ref and test optimizers may require different set of options. # This is a quick workaround to add that functionality while making # minimum changes in existing code. # If there is no "tst_options" field provided, safe to initialize # the test optimizer with the parameters of reference optimizer. if not hasattr(self, 'tst_options'): self.tst_options = self.options tensor = torch.rand(nelem, dtype=param_type, device=device) ref_param, tst_param, ref_optim, tst_optim = \ self.gen_param_optim([tensor], self.options, self.tst_options) for i in range(self.iters): self.gen_grad(ref_param, tst_param) ref_optim.step() tst_optim.step() if skip_assert: return max_abs_diff, max_rel_diff = self.get_max_diff(ref_param, tst_param) self.assertLessEqual(max_abs_diff, self.max_abs_diff) self.assertLessEqual(max_rel_diff, self.max_rel_diff) class TestFusedAdam(TestFusedOptimizer): def setUp(self): super().setUp() self.options = {'lr':5e-4, 'betas':(0.9, 0.999), 'eps':1e-08, 'weight_decay': 0, 'amsgrad': False} self.ref_optim = torch.optim.Adam self.fused_optim = apex.optimizers.FusedAdam def test_float(self): self.gen_single_type_test(param_type=torch.float) # NOTE(mkozuki): Current threshold values look too small for BFloat16. # TODO(mkozuki): Refactor `TestFusedOptimizer` def test_half(self): self.gen_single_type_test(param_type=torch.float16, skip_assert=True) def test_bfloat16(self): self.gen_single_type_test(param_type=torch.bfloat16, skip_assert=True) @unittest.skipIf(torch.cuda.device_count()<2, "more than 1 GPU required") def test_multi_device(self): devices = ("cuda:0", "cuda:1") for current_dev, tensor_dev in product(devices, devices): with torch.cuda.device(current_dev): self.gen_single_type_test(param_type=torch.float, device=tensor_dev) @unittest.skip('Disable until 8/1/2019 adam/adamw upstream picked') def test_multi_params(self): sizes = [[4096, 1024], [4096], [4096, 2048], [32320, 1024], [1]] tensors = [] for size in sizes: tensors.append(torch.rand(size, dtype=torch.float, device='cuda')) ref_param, tst_param, ref_optim, tst_optim = \ self.gen_param_optim(tensors, self.options) for i in range(self.iters): self.gen_grad(ref_param, tst_param) ref_optim.step() tst_optim.step() max_abs_diff, max_rel_diff = self.get_max_diff(ref_param, tst_param) self.assertLessEqual(max_abs_diff, self.max_abs_diff) self.assertLessEqual(max_rel_diff, self.max_rel_diff) @unittest.skip('No longer support fuse scaling') def test_scale(self): nelem = 278011 tensor = torch.rand(nelem, dtype=torch.float, device='cuda') ref_param, tst_param, ref_optim, tst_optim = \ self.gen_param_optim([tensor], self.options) for i in range(self.iters): scale = random.random() * 1000 half_grads = self.gen_mixed_grad(ref_param, tst_param, scale) ref_optim.step() tst_optim.step(grads=half_grads, scale=scale) max_abs_diff, max_rel_diff = self.get_max_diff(ref_param, tst_param) self.assertLessEqual(max_abs_diff, self.max_abs_diff) self.assertLessEqual(max_rel_diff, self.max_rel_diff) @unittest.skip('No longer support output fp16 param') def test_fp16_output(self): nelem = 278011 tensor = torch.rand(nelem, dtype=torch.float, device='cuda') ref_param, tst_param, ref_optim, tst_optim = \ self.gen_param_optim([tensor], self.options) fp16_param = torch.nn.Parameter(tensor.clone().half()) for i in range(self.iters): half_grads = self.gen_mixed_grad(ref_param, tst_param) ref_optim.step() tst_optim.step(grads=half_grads, output_params=[fp16_param]) max_abs_diff, max_rel_diff = self.get_max_diff(ref_param, tst_param) self.assertLessEqual(max_abs_diff, self.max_abs_diff) self.assertLessEqual(max_rel_diff, self.max_rel_diff) max_abs_diff, max_rel_diff = self.get_max_diff(tst_param, \ [fp16_param.float()]) self.assertLessEqual(max_abs_diff, self.max_abs_diff) self.assertLessEqual(max_rel_diff, self.max_rel_diff) def test_adam_option(self): nelem = 1 adam_option = {'lr':0.01, 'betas':(0.6, 0.9), 'eps':3e-06, 'weight_decay':0, 'amsgrad':False} tensor = torch.rand(nelem, dtype=torch.float, device='cuda') ref_param, tst_param, ref_optim, tst_optim = \ self.gen_param_optim([tensor], adam_option) for i in range(self.iters): self.gen_grad(ref_param, tst_param) ref_optim.step() tst_optim.step() max_abs_diff, max_rel_diff = self.get_max_diff(ref_param, tst_param) self.assertLessEqual(max_abs_diff, self.max_abs_diff) self.assertLessEqual(max_rel_diff, self.max_rel_diff) def test_frozen_model(self): nelem = 1 adam_option = {'lr':0.01, 'betas':(0.6, 0.9), 'eps':3e-06, 'weight_decay':0, 'amsgrad':False} tensor = torch.rand(nelem, dtype=torch.float, device='cuda') ref_param, tst_param, ref_optim, tst_optim = \ self.gen_param_optim([tensor], adam_option) #Add an empty param group which may occur for pipeline parallel p-tuning tst_optim.add_param_group({"params": []}) for i in range(self.iters): self.gen_grad(ref_param, tst_param) ref_optim.step() tst_optim.step() max_abs_diff, max_rel_diff = self.get_max_diff(ref_param, tst_param) self.assertLessEqual(max_abs_diff, self.max_abs_diff) self.assertLessEqual(max_rel_diff, self.max_rel_diff) class TestFusedAdagrad(TestFusedOptimizer): def __init__(self, *args, **kwargs): super(TestFusedAdagrad, self).__init__(*args, **kwargs) self.options = {"lr": 5e-4, "eps": 1e-08, "weight_decay": 1.0e-5} self.ref_optim = torch.optim.Adagrad self.fused_optim = apex.optimizers.FusedAdagrad def test_float(self): self.gen_single_type_test(param_type=torch.float) @unittest.skip("PyTorch optimizer is not numerically correct for fp16") def test_half(self): self.gen_single_type_test(param_type=torch.float16) @unittest.skipIf(torch.cuda.device_count()<2, "more than 1 GPU required") def test_multi_device(self): devices = ("cuda:0", "cuda:1") for current_dev, tensor_dev in product(devices, devices): with torch.cuda.device(current_dev): self.gen_single_type_test(param_type=torch.float, device=tensor_dev) def test_multi_params(self): sizes = [[4096, 1024], [4096], [4096, 2048], [32320, 1024], [1]] adagrad_option = {"lr": 5e-4, "eps": 1e-08, "weight_decay": 0} tensors = [] for size in sizes: tensors.append(torch.rand(size, dtype=torch.float, device="cuda")) ref_param, tst_param, ref_optim, tst_optim = self.gen_param_optim( tensors, adagrad_option ) for _ in range(self.iters): self.gen_grad(ref_param, tst_param) ref_optim.step() tst_optim.step() max_abs_diff, max_rel_diff = self.get_max_diff(ref_param, tst_param) self.assertLessEqual(max_abs_diff, self.max_abs_diff) self.assertLessEqual(max_rel_diff, self.max_rel_diff) @unittest.skipIf(torch.cuda.device_count()<2, "more than 1 GPU required") def test_multi_params_different_devices_throws(self): sizes = [[4096, 1024], [4096], [4096, 2048], [32320, 1024], [1]] adagrad_option = {"lr": 5e-4, "eps": 1e-08, "weight_decay": 0} tensors = [] for i, size in enumerate(sizes): tensors.append(torch.rand(size, dtype=torch.float, device="cuda:"+str(i % 2))) ref_param, tst_param, ref_optim, tst_optim = self.gen_param_optim( tensors, adagrad_option ) self.gen_grad(ref_param, tst_param) with self.assertRaisesRegex(RuntimeError, "not on the same device"): tst_optim.step() def test_adagrad_option(self): nelem = 1 adagrad_option = {"lr": 0.01, "eps": 3e-06, "weight_decay": 0} tensor = torch.rand(nelem, dtype=torch.float, device="cuda") ref_param, tst_param, ref_optim, tst_optim = self.gen_param_optim( [tensor], adagrad_option ) for _ in range(self.iters): self.gen_grad(ref_param, tst_param) ref_optim.step() tst_optim.step() max_abs_diff, max_rel_diff = self.get_max_diff(ref_param, tst_param) self.assertLessEqual(max_abs_diff, self.max_abs_diff) self.assertLessEqual(max_rel_diff, self.max_rel_diff) class TestFusedSGD(TestFusedOptimizer): def __init__(self, *args, **kwargs): super(TestFusedSGD, self).__init__(*args, **kwargs) self.options = {"lr": .25, "momentum": .125} self.ref_optim = torch.optim.SGD self.fused_optim = apex.optimizers.FusedSGD def test_float(self): self.gen_single_type_test(param_type=torch.float) def test_half(self): self.gen_single_type_test(param_type=torch.float16) @unittest.skipIf(torch.cuda.device_count()<2, "more than 1 GPU required") def test_multi_device(self): devices = ("cuda:0", "cuda:1") for current_dev, tensor_dev in product(devices, devices): with torch.cuda.device(current_dev): self.gen_single_type_test(param_type=torch.float, device=tensor_dev) if __name__ == '__main__': unittest.main()
apex-master
tests/L0/run_optimizers/test_fused_optimizer.py
import torch from torch.optim import Optimizer import math import apex import unittest from test_fused_optimizer import TestFusedOptimizer from itertools import product class Novograd(Optimizer): """ Implements Novograd algorithm. Args: params (iterable): iterable of parameters to optimize or dicts defining parameter groups lr (float, optional): learning rate (default: 1e-3) betas (Tuple[float, float], optional): coefficients used for computing running averages of gradient and its square (default: (0.95, 0)) eps (float, optional): term added to the denominator to improve numerical stability (default: 1e-8) weight_decay (float, optional): weight decay (L2 penalty) (default: 0) grad_averaging: gradient averaging amsgrad (boolean, optional): whether to use the AMSGrad variant of this algorithm from the paper `On the Convergence of Adam and Beyond`_ (default: False) """ def __init__(self, params, lr=1e-3, betas=(0.95, 0), eps=1e-8, weight_decay=0, grad_averaging=False, amsgrad=False): if not 0.0 <= lr: raise ValueError("Invalid learning rate: {}".format(lr)) if not 0.0 <= eps: raise ValueError("Invalid epsilon value: {}".format(eps)) if not 0.0 <= betas[0] < 1.0: raise ValueError("Invalid beta parameter at index 0: {}".format(betas[0])) if not 0.0 <= betas[1] < 1.0: raise ValueError("Invalid beta parameter at index 1: {}".format(betas[1])) defaults = dict(lr=lr, betas=betas, eps=eps, weight_decay=weight_decay, grad_averaging=grad_averaging, amsgrad=amsgrad) super(Novograd, self).__init__(params, defaults) def __setstate__(self, state): super(Novograd, self).__setstate__(state) for group in self.param_groups: group.setdefault('amsgrad', False) def step(self, closure=None): """Performs a single optimization step. Arguments: closure (callable, optional): A closure that reevaluates the model and returns the loss. """ loss = None if closure is not None: loss = closure() for group in self.param_groups: for p in group['params']: if p.grad is None: continue grad = p.grad.data if grad.is_sparse: raise RuntimeError('Sparse gradients are not supported.') amsgrad = group['amsgrad'] state = self.state[p] # State initialization if len(state) == 0: state['step'] = 0 # Exponential moving average of gradient values state['exp_avg'] = torch.zeros_like(p.data) # Exponential moving average of squared gradient values state['exp_avg_sq'] = torch.zeros([]).to(state['exp_avg'].device) if amsgrad: # Maintains max of all exp. moving avg. of sq. grad. values state['max_exp_avg_sq'] = torch.zeros([]).to(state['exp_avg'].device) exp_avg, exp_avg_sq = state['exp_avg'], state['exp_avg_sq'] if amsgrad: max_exp_avg_sq = state['max_exp_avg_sq'] beta1, beta2 = group['betas'] state['step'] += 1 norm = torch.sum(torch.pow(grad, 2)) if exp_avg_sq == 0: exp_avg_sq.copy_(norm) else: exp_avg_sq.mul_(beta2).add_(norm, alpha=1 - beta2) if amsgrad: # Maintains the maximum of all 2nd moment running avg. till now torch.max(max_exp_avg_sq, exp_avg_sq, out=max_exp_avg_sq) # Use the max. for normalizing running avg. of gradient denom = max_exp_avg_sq.sqrt().add_(group['eps']) else: denom = exp_avg_sq.sqrt().add_(group['eps']) grad.div_(denom) if group['weight_decay'] != 0: grad.add_(p.data, alpha=group['weight_decay']) if group['grad_averaging']: grad.mul_(1 - beta1) exp_avg.mul_(beta1).add_(grad) p.data.add_(exp_avg, alpha=-group['lr']) return loss class TestFusedNovoGrad(TestFusedOptimizer): def __init__(self, *args, **kwargs): super(TestFusedNovoGrad, self).__init__(*args, **kwargs) # The options for NovoGrad and FusedNovoGrad are very specific if they # are expected to behave the same. self.options = {'lr':1e-3, 'betas':(0.95, 0), 'eps':1e-8, 'weight_decay':0, 'grad_averaging':False, 'amsgrad':False} self.tst_options = {'lr':1e-3, 'betas':(0.95, 0), 'eps':1e-8, 'weight_decay':0, 'grad_averaging':False, 'amsgrad':False, 'bias_correction':False, 'reg_inside_moment':True, 'norm_type':2, 'init_zero':False, 'set_grad_none':True} self.ref_optim = Novograd self.fused_optim = apex.optimizers.FusedNovoGrad def test_float(self): self.gen_single_type_test(param_type=torch.float) def test_half(self): self.gen_single_type_test(param_type=torch.float16) @unittest.skipIf(torch.cuda.device_count()<2, "more than 1 GPU required") def test_multi_device(self): devices = ("cuda:1", "cuda:0") for current_dev, tensor_dev in product(devices, devices): with torch.cuda.device(current_dev): torch.cuda.synchronize() self.gen_single_type_test(param_type=torch.float, device=tensor_dev) def test_multi_params(self): sizes = [[4096, 1024], [4096], [4096, 2048], [32320, 1024], [1]] tensors = [] for size in sizes: tensors.append(torch.rand(size, dtype=torch.float, device="cuda")) ref_param, tst_param, ref_optim, tst_optim = self.gen_param_optim( tensors, self.options, self.tst_options ) for _ in range(self.iters): self.gen_grad(ref_param, tst_param) ref_optim.step() tst_optim.step() max_abs_diff, max_rel_diff = self.get_max_diff(ref_param, tst_param) self.assertLessEqual(max_abs_diff, self.max_abs_diff) self.assertLessEqual(max_rel_diff, self.max_rel_diff) if __name__ == '__main__': unittest.main()
apex-master
tests/L0/run_optimizers/test_fused_novograd.py
import logging import unittest import torch from torch.testing._internal import common_utils logging.getLogger("torch").setLevel(logging.WARNING) from apex.transformer import parallel_state from apex.transformer.pipeline_parallel import p2p_communication from apex.transformer.testing.distributed_test_base import NcclDistributedTestBase from apex.transformer.testing.distributed_test_base import UccDistributedTestBase logging.getLogger("apex").setLevel(logging.DEBUG) # [P2P Ops Involved in Pipeline Model Parallel forward/backward] # **forward_backward_pipelining_without_interleaving** # - send_forward / recv_forward # - send_backward / recv_backward # - send_forward_recv_backward # - send_backward_recv_forward # **forward_backward_pipelining_with_interleaving** # - send_backward_recv_backward # - recv_backward # - recv_forward # - send_forward_backward_recv_forward_backward # - send_forward_recv_forward class P2PCommTestBase: numel = 4 shape = (2, 2) dtype = torch.float32 @property def world_size(self): return min(2, torch.cuda.device_count()) def _init_model_parallel(self): parallel_state.initialize_model_parallel( tensor_model_parallel_size_=1, pipeline_model_parallel_size_=self.world_size, virtual_pipeline_model_parallel_size_=None, ) def create_tensor(self, value: int = None): return torch.tensor( [value] * self.numel).view(self.shape).to(device="cuda", dtype=self.dtype) # Brief: Simulate warm-up. # Brief: test `recv_forward` & `send_forward`. def test_no_interleaving_warmup(self): self.assertEqual(self.world_size, 2) self._init_model_parallel() input_tensor = None if parallel_state.is_pipeline_first_stage(): tensor = self.create_tensor(self.rank) print(tensor) p2p_communication.send_forward(output_tensor=tensor, tensor_shape=self.shape, dtype=self.dtype) else: input_tensor = p2p_communication.recv_forward(tensor_shape=self.shape, dtype=self.dtype) if parallel_state.is_pipeline_first_stage(): self.assertIsNone(input_tensor) else: expected_input_tensor = self.create_tensor(self.rank - 1) self.assertEqual(input_tensor, expected_input_tensor) # Brief: test `send_forward`, `send_forward_recv_forward`, and `recv_forward`. def test_send_forward_recv_forward(self): self._init_model_parallel() prev_tensor = None tensor = self.create_tensor(self.rank) if parallel_state.is_pipeline_first_stage(): p2p_communication.send_forward(output_tensor=tensor, tensor_shape=self.shape, dtype=self.dtype) elif parallel_state.is_pipeline_last_stage(): prev_tensor = p2p_communication.recv_forward(tensor_shape=self.shape, dtype=self.dtype) else: prev_tensor = p2p_communication.send_forward_recv_forward( output_tensor=tensor, recv_prev=True, tensor_shape=self.shape, dtype=self.dtype, ) if parallel_state.is_pipeline_first_stage(): self.assertIsNone(prev_tensor) else: expected_prev_tensor = self.create_tensor(self.rank - 1) self.assertEqual(prev_tensor, expected_prev_tensor) # Brief: test `send_backward`, `send_backward_recv_backward`, and `recv_backward`. def test_send_backward_recv_backward(self): self._init_model_parallel() tensor = self.create_tensor(self.rank) next_tensor = None if parallel_state.is_pipeline_first_stage(): next_tensor = p2p_communication.recv_backward(tensor_shape=self.shape, dtype=self.dtype) elif parallel_state.is_pipeline_last_stage(): p2p_communication.send_backward(input_tensor_grad=tensor, tensor_shape=self.shape, dtype=self.dtype) else: next_tensor = p2p_communication.send_backward_recv_backward( input_tensor_grad=tensor, recv_next=True, tensor_shape=self.shape, dtype=self.dtype, ) if parallel_state.is_pipeline_last_stage(): self.assertIsNone(next_tensor) else: expected_next_tensor = self.create_tensor(self.rank + 1) self.assertEqual(next_tensor, expected_next_tensor) # n.b.(mkozuki): Intentionally skip NCCL backend tests as I trust pytorch/pytorch repo. @unittest.skipIf(torch.cuda.device_count() < 2, "Requires >= 2 GPUs") class UccP2PCommTest(P2PCommTestBase, UccDistributedTestBase): pass if __name__ == "__main__": common_utils.run_tests()
apex-master
tests/L0/run_transformer/test_p2p_comm.py
import subprocess import os from apex.transformer.testing.commons import TEST_SUCCESS_MESSAGE def run_gpt(cmd): args = list(cmd.split(" ")) p = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.PIPE) outs, errs = p.communicate() outs = list(str((outs).decode("utf-8")).splitlines()) success = False runtime = 0 num_params = 0 for out in outs: out = str(out) if "Average Iteration Time:" in str(out): slicey = out[out.find(":") + 2 :] try: runtime = float(slicey) except: print(slicey) quit() if "Number of Parameters:" in str(out): slicey = out[out.find(":") + 2 :] try: num_params = int(slicey) except: print(slicey) quit() if str(out) == str(TEST_SUCCESS_MESSAGE): success = True return runtime, round(float(int(num_params)) / 10.0 ** 9, 3), success, errs def plot(runtimes): import matplotlib.pyplot as plt for distributed_setting in runtimes.keys(): plt.scatter( runtimes[distributed_setting].keys(), runtimes[distributed_setting].values(), label=distributed_setting, ) plt.legend() plt.xlabel("Parameters (Billions)") plt.ylabel("Training Iteration time (s)") plt.title(str("GPT Scaling w/ Offloading")) plt.savefig("offload_gpt_scaling.png") plt.close() if not os.path.exists("/my_workspace/"): os.system("mkdir /my_workspace/") os.system("cp *.png /my_workspace/") def main(): runtimes = {} nlist = ( list(range(2000, 10000, 2000)) + list(range(10000, 50000, 5000)) + list(range(50000, 100000, 10000)) ) print("N-List:", nlist) for data_parr, tens_parr, pipe_parr in [(8, 1, 1), (4, 2, 1), (2, 1, 4), (1, 2, 4)]: for offload in [True, False]: dist_setting = ( "ddp=" + str(data_parr) + ", tensor_parr=" + str(tens_parr) + ", pipe_parr=" + str(pipe_parr) + ", offload=" + str(offload) ) runtimes[dist_setting] = {} print("Beginning Testing for", dist_setting) for n in nlist: cmd = "python3 -m torch.distributed.launch --nproc_per_node=8 run_gpt_minimal_test.py" cmd += ( " --micro-batch-size 1 --num-layers " + str(n) + " --hidden-size 128 --num-attention-heads 16" ) cmd += ( " --max-position-embeddings 128 --seq-length 128 --tensor-model-parallel-size " + str(tens_parr) ) cmd += ( " --pipeline-model-parallel-size " + str(pipe_parr) + (" --cpu-offload" if offload else "") ) print(cmd) runtime, bill_params, success, errs = run_gpt(cmd) if success: runtimes[dist_setting][bill_params] = runtime print( str(runtime) + "s per training iter for", str(bill_params) + "B parameter GPT-2", ) if n >= 10000: plot(runtimes) else: print("GPT-2 w/", n, "layers failed using", dist_setting) print("Moving on to the next distributed setting...") print("#" * (25)) print() plot(runtimes) break print(runtimes) plot(runtimes) if __name__ == "__main__": main()
apex-master
tests/L0/run_transformer/gpt_scaling_test.py
from typing import Tuple, List import torch import unittest from apex.transformer import parallel_state from apex.transformer.pipeline_parallel.utils import get_num_microbatches from apex.transformer.pipeline_parallel.schedules.common import ( _get_params_for_weight_decay_optimization, build_model ) from apex.transformer.pipeline_parallel.schedules.fwd_bwd_pipelining_with_interleaving import ( _forward_backward_pipelining_with_interleaving, ) from apex.transformer.pipeline_parallel.utils import ( setup_microbatch_calculator, _reconfigure_microbatch_calculator, update_num_microbatches ) from apex.transformer.testing import global_vars from apex.transformer.testing.commons import ( print_separator, fwd_step_func, model_provider_func ) from apex.transformer.log_util import get_transformer_logger from apex.transformer._data import MegatronPretrainingRandomSampler, MegatronPretrainingSampler from apex.transformer.testing.distributed_test_base import NcclDistributedTestBase from torch.testing._internal import common_utils # note(mkozuki): To see warmup, steady, cooldown iterations, uncomment the line below # set_logging_level("INFO") _logger = get_transformer_logger("pipeline_parallel_test") # note(mkozuki): To see if local batch size increases, uncomment the line below # _logger.setLevel("INFO") NUM_ITERATIONS = 20 NUM_SAMPLES = 16384 // 2 HIDDEN_SIZE = 16 def Dataset(num_samples: int) -> List[Tuple[torch.Tensor, torch.Tensor]]: return [ ( torch.randn(HIDDEN_SIZE, HIDDEN_SIZE), torch.randn(HIDDEN_SIZE // 2, HIDDEN_SIZE // 2), ) for _ in range(num_samples) ] # Run forward & backward with dynamic batch size. def run_interleaved_with_dynamic_batch_size( pipeline_model_parallel_size: int, forward_only: bool, BatchSamplerCls, ) -> None: args = global_vars.get_args() _reconfigure_microbatch_calculator( args.rank, args.rampup_batch_size, args.global_batch_size, args.micro_batch_size, 1, # args.data_parallel_size, ) virtual_pipeline_model_parallel_size = 2 # NOTE (mkozuki): `virtual_pipeline_model_parallel_size` is a requisite for the interleaving scheduling # In megatron, `args.virtual_pipeline_model_parallel_size` is computed in megatron/arguments.py and # used ubiquitously but this test uses custom model so it's safe to abuse. parallel_state.initialize_model_parallel( 1, pipeline_model_parallel_size, virtual_pipeline_model_parallel_size ) pipeline_model_parallel_size = ( parallel_state.get_pipeline_model_parallel_world_size() ) print_separator( f"BatchSamplerCls: {BatchSamplerCls.__name__}, forward_only: {forward_only}" ) model = build_model( model_provider_func, wrap_with_ddp=True, virtual_pipeline_model_parallel_size=virtual_pipeline_model_parallel_size, hidden_size=HIDDEN_SIZE, ) assert isinstance(model, list) assert len(model) == virtual_pipeline_model_parallel_size optimizer = torch.optim.Adam( _get_params_for_weight_decay_optimization(model)) initial_local_minibatch_size = get_num_microbatches() * args.micro_batch_size dataset = Dataset(NUM_SAMPLES) data_loader = torch.utils.data.DataLoader( dataset, batch_sampler=BatchSamplerCls( NUM_SAMPLES, 0, initial_local_minibatch_size, parallel_state.get_data_parallel_rank(), parallel_state.get_data_parallel_world_size(), ), ) data_iter = iter(data_loader) def get_num_samples(batch): if isinstance(batch, torch.Tensor): return len(batch) assert isinstance(batch, (list, tuple)) return [get_num_samples(b) for b in batch] tensor_shape = [args.micro_batch_size, HIDDEN_SIZE, HIDDEN_SIZE] consumed_samples = 0 for i in range(NUM_ITERATIONS): update_num_microbatches(consumed_samples, consistency_check=False) local_batch_size = get_num_microbatches() * args.micro_batch_size data_iter._index_sampler.local_minibatch_size = local_batch_size local_mini_batch = next(data_iter) _logger.info( f"iter: {i} / {NUM_ITERATIONS} " f"local batchsize: {get_num_samples(local_mini_batch)} " f"consumed_samples: {consumed_samples} / {NUM_SAMPLES}" ) _forward_backward_pipelining_with_interleaving( fwd_step_func, local_mini_batch, model, forward_only=forward_only, tensor_shape=tensor_shape, ) consumed_samples += ( parallel_state.get_data_parallel_world_size() * get_num_microbatches() * args.micro_batch_size ) if not forward_only: for m in model: for p in m.parameters(): if p.grad is None: raise RuntimeError("grad not found") else: optimizer.zero_grad(set_to_none=True) torch.cuda.synchronize() class DynamicBatchsizeTestBase: @unittest.skipUnless(torch.cuda.device_count() > 2, "requires at least 3 gpus") def test_dynamic_batchsize(self): n_tests = 0 failures = [] override_args = { "micro_batch_size": 2, "num_layers": 16, "hidden_size": 256, "num_attention_heads": 8, "max_position_embeddings": 512, "seq_length": 512, "global_batch_size": 128, "use_cpu_initialization": True, "world_size": self.world_size, "rank": self.rank, } global_vars.set_global_variables( args_defaults={"global_batch_size": 512, "rampup_batch_size": [64, 64, 1000], }, ignore_unknown_args=True, override_args=override_args, ) args = global_vars.get_args() setup_microbatch_calculator( args.rank, args.rampup_batch_size, args.global_batch_size, args.micro_batch_size, 1, # args.data_parallel_size, ) for BatchSamplerCls in ( MegatronPretrainingSampler, MegatronPretrainingRandomSampler, ): for forward_only in (False, True): n_tests += 1 pipeline_model_parallel_size = self.world_size try: run_interleaved_with_dynamic_batch_size( pipeline_model_parallel_size, forward_only, BatchSamplerCls, ) except Exception as e: msg = ( f"\tforward_only: {forward_only}\n" f"pipeline rank: {parallel_state.get_pipeline_model_parallel_rank()}, " f"virtual pipeline rank: {parallel_state.get_virtual_pipeline_model_parallel_rank()}\n" f"{str(e)}" ) raise RuntimeError(msg) finally: parallel_state.destroy_model_parallel() if failures: print_separator("TEST FAILED:") print("\n".join(failures)) msg = f"{len(failures)} / {n_tests} cases failed" raise RuntimeError(msg) else: if torch.distributed.get_rank() == 0: print_separator("TEST RESULT: ### PASS!") class NcclDynamicBatchsizeTest(DynamicBatchsizeTestBase, NcclDistributedTestBase): pass # TODO: (Fuzzkatt) UCC still doesn't work with fwd_bwd_pipelining_with_interleaving if __name__ == "__main__": torch.backends.cuda.matmul.allow_tf32 = False common_utils.run_tests()
apex-master
tests/L0/run_transformer/test_dynamic_batchsize.py
"""Test for fused softmax functions. Ref: https://github.com/NVIDIA/Megatron-LM/blob/40becfc96c4144985458ac0e0fae45dbb111fbd2/megatron/fused_kernels/tests/test_fused_kernels.py """ # NOQA import itertools import torch from torch.testing._internal import common_utils from apex.transformer import AttnMaskType from apex.transformer.functional import FusedScaleMaskSoftmax def attention_mask_func(attention_scores, attention_mask): return attention_scores.masked_fill(attention_mask, -10000.0) def forward_torch_softmax(input, mask, scale): input = input * scale mask_output = attention_mask_func(input, mask) if mask is not None else input probs = torch.nn.Softmax(dim=-1)(mask_output) all_k_masked = mask.all(axis=-1) zero_attention_mask = (1.0 - all_k_masked.float())[:, :, :, None] probs = probs * zero_attention_mask return probs autocast_dtypes = ( (torch.half, torch.bfloat16) if torch.cuda.is_bf16_supported() else (torch.half,) ) class TestFusedScaleMaskSoftmax(common_utils.TestCase): def _setup_fused_softmax( self, input_in_fp16, input_in_bf16, scale=None, softmax_in_fp32=False, attn_mask_type=AttnMaskType.padding, ): fused_fn = FusedScaleMaskSoftmax( input_in_fp16=input_in_fp16, input_in_bf16=input_in_bf16, mask_func=attention_mask_func, scale=scale, softmax_in_fp32=softmax_in_fp32, attn_mask_type=attn_mask_type, scaled_masked_softmax_fusion=True, ) torch_fn = FusedScaleMaskSoftmax( input_in_fp16=input_in_fp16, input_in_bf16=input_in_bf16, mask_func=attention_mask_func, scale=scale, softmax_in_fp32=softmax_in_fp32, attn_mask_type=attn_mask_type, scaled_masked_softmax_fusion=False, ) return fused_fn, torch_fn def tearDown(self) -> None: torch.cuda.empty_cache() super().tearDown() def test_fused_scale_mask_softmax(self): """ attention_scores.shape = [4, 12, 24, 24] mask.shape = [4, 1, 24, 24] """ for (dtype, scale, softmax_in_fp32, shape) in itertools.product( (torch.half, torch.bfloat16), (None, 2.0), (False, True), ((4, 12, 24, 24), (32, 12, 4, 214)) ): msg = f"{dtype}-{scale}-{softmax_in_fp32}" input_in_fp16 = dtype == torch.half input_in_bf16 = dtype == torch.bfloat16 if not (scale is None or softmax_in_fp32): with self.assertRaises(RuntimeError, msg=msg): self._setup_fused_softmax( input_in_fp16, input_in_bf16, scale, softmax_in_fp32, AttnMaskType.padding, ) return fused_fn, torch_fn = self._setup_fused_softmax( input_in_fp16, input_in_bf16, scale, softmax_in_fp32, AttnMaskType.padding, ) attention_scores_0 = ( torch.randn(shape) .to(device="cuda", dtype=dtype) .requires_grad_(True) ) with torch.no_grad(): attention_scores_1 = attention_scores_0.clone().requires_grad_(True) mask_shape = (shape[0],) + (1,) + shape[2:] mask = torch.randint(0, 2, mask_shape, device="cuda").bool() expected = fused_fn(attention_scores_0, mask) actual = torch_fn(attention_scores_1, mask) self.assertEqual(actual, expected, msg=msg) g0 = torch.rand_like(actual) with torch.no_grad(): g1 = g0.clone() expected.backward(g0) actual.backward(g1) def test_autocast_fused_scale_mask_softmax(self): for dtype in autocast_dtypes: msg = f"dtype: {dtype}" input_in_fp16 = dtype == torch.half input_in_bf16 = dtype == torch.bfloat16 fused_fn, torch_fn = self._setup_fused_softmax( input_in_fp16, input_in_bf16, attn_mask_type=AttnMaskType.padding ) attention_scores_0 = ( torch.randn((4, 12, 24, 24)).cuda().requires_grad_(True) ) with torch.no_grad(): attention_scores_1 = ( attention_scores_0.clone().to(dtype).requires_grad_(True) ) mask = torch.randint(0, 2, (4, 1, 24, 24)).bool().cuda() expected = torch_fn(attention_scores_1, mask) with torch.cuda.amp.autocast(dtype=dtype): actual = fused_fn(attention_scores_0, mask) self.assertEqual(actual.dtype, dtype, msg=msg) self.assertEqual(actual, expected, msg=msg) g0 = torch.rand_like(actual) with torch.no_grad(): g1 = g0.clone() expected.backward(g0) actual.backward(g1) def test_fused_scale_softmax(self): """ attention_scores.shape = [4, 12, 24, 24] mask = None """ for (dtype, scale, softmax_in_fp32, shape) in itertools.product( (torch.half, torch.bfloat16), (None, 2.0), (False, True), ((4, 12, 24, 24), (32, 12, 4, 214)) ): msg = f"{dtype}-{scale}-{softmax_in_fp32}" input_in_fp16 = dtype == torch.half input_in_bf16 = dtype == torch.bfloat16 if not (scale is None or softmax_in_fp32): with self.assertRaises(RuntimeError, msg=msg): self._setup_fused_softmax( input_in_fp16, input_in_bf16, scale, softmax_in_fp32, AttnMaskType.padding, ) return fused_fn, torch_fn = self._setup_fused_softmax( input_in_fp16, input_in_bf16, scale, softmax_in_fp32, AttnMaskType.padding, ) attention_scores_0 = ( torch.randn(shape) .to(device="cuda", dtype=dtype) .requires_grad_(True) ) with torch.no_grad(): attention_scores_1 = attention_scores_0.clone().requires_grad_(True) mask = None expected = fused_fn(attention_scores_0, mask) actual = torch_fn(attention_scores_1, mask) self.assertEqual(actual, expected, msg=msg) g0 = torch.rand_like(actual) with torch.no_grad(): g1 = g0.clone() expected.backward(g0) actual.backward(g1) def test_autocast_fused_scale_softmax(self): for dtype in autocast_dtypes: msg = f"dtype: {dtype}" input_in_fp16 = dtype == torch.half input_in_bf16 = dtype == torch.bfloat16 fused_fn, torch_fn = self._setup_fused_softmax( input_in_fp16, input_in_bf16, attn_mask_type=AttnMaskType.padding ) attention_scores_0 = ( torch.randn((4, 12, 24, 24)).cuda().requires_grad_(True) ) with torch.no_grad(): attention_scores_1 = ( attention_scores_0.clone().to(dtype).requires_grad_(True) ) mask = None expected = torch_fn(attention_scores_1, mask) with torch.cuda.amp.autocast(dtype=dtype): actual = fused_fn(attention_scores_0, mask) self.assertEqual(actual.dtype, dtype, msg=msg) self.assertEqual(actual, expected, msg=msg) g0 = torch.rand_like(actual) with torch.no_grad(): g1 = g0.clone() expected.backward(g0) actual.backward(g1) def test_fused_upper_triangle_mask_softmax(self): """ attn_weights.shape: [4, 12, 24, 24] total_mask.shape: [4, 1, 24, 24] total_mask[0, 0], a 24x24 matrix is like a lower triangular matrix, but upper elements are True and lower elements and diagonal are False. """ for (dtype, scale, softmax_in_fp32) in itertools.product( (torch.half, torch.bfloat16), (None, 2.0), (False, True), ): msg = f"{dtype}-{scale}-{softmax_in_fp32}" input_in_fp16 = dtype == torch.half input_in_bf16 = dtype == torch.bfloat16 if not (scale is None or softmax_in_fp32): with self.assertRaises(RuntimeError, msg=msg): self._setup_fused_softmax( input_in_fp16, input_in_bf16, scale, softmax_in_fp32, AttnMaskType.causal, ) return fused_fn, torch_fn = self._setup_fused_softmax( input_in_fp16, input_in_bf16, scale, softmax_in_fp32, AttnMaskType.causal, ) attn_weights_0 = ( torch.randn((4, 12, 24, 24)) .to(device="cuda", dtype=dtype) .requires_grad_(True) ) with torch.no_grad(): attn_weights_1 = attn_weights_0.clone().requires_grad_(True) total_mask = ( ~(torch.tril(torch.randn((24, 24), device="cuda")).bool()) .unsqueeze(0) .unsqueeze(0) ) total_mask = total_mask.repeat((4, 1, 1, 1)) expected = fused_fn(attn_weights_0, total_mask) actual = torch_fn(attn_weights_1, total_mask) self.assertEqual(actual, expected, msg=msg) g0 = torch.randn_like(actual) with torch.no_grad(): g1 = g0.clone() actual.backward(g0) expected.backward(g1) def test_autocast_fused_upper_triangle_mask_softmax(self): for dtype in autocast_dtypes: msg = f"dtype: {dtype}" input_in_fp16 = dtype == torch.half input_in_bf16 = dtype == torch.bfloat16 fused_fn, torch_fn = self._setup_fused_softmax( input_in_fp16, input_in_bf16, attn_mask_type=AttnMaskType.causal ) attn_weights_0 = ( torch.randn((4, 12, 24, 24)).cuda().requires_grad_(True) ) with torch.no_grad(): attn_weights_1 = ( attn_weights_0.clone().to(dtype).requires_grad_(True) ) total_mask = ( ~(torch.tril(torch.randn((24, 24), device="cuda")).bool()) .unsqueeze(0) .unsqueeze(0) ) with torch.cuda.amp.autocast(dtype=dtype): actual = fused_fn(attn_weights_0, total_mask) self.assertEqual(actual.dtype, dtype, msg=msg) expected = torch_fn(attn_weights_1, total_mask) self.assertEqual(actual, expected, msg=msg) g0 = torch.randn_like(actual) with torch.no_grad(): g1 = g0.clone() actual.backward(g0) expected.backward(g1) class TestGenericFusedSoftmaxKernel(common_utils.TestCase): def setUp(self): super().setUp() self.batch = 2 self.attn = 16 self.scale_t = 1.0 self.dtype = torch.float16 self.device = torch.cuda.current_device() self.thresh = {"atol": 1e-3, "rtol": 1e-3} qlen = [1, 2] klen = [1, 2, 3, 4, 5, 8, 10, 11, 13, 128, 256, 1200, 1234] available_cuda_mem = torch.cuda.memory.mem_get_info(self.device)[0] / (1024 ** 3) if available_cuda_mem > 40: qlen.extend([1234, 2322, 2348]) klen.extend([2048, 3123, 4096, 4128, 7234, 8192]) self.q_k_lens = itertools.product(qlen, klen) def tearDown(self) -> None: torch.cuda.empty_cache() super().tearDown() def test_forward(self, allmasked: bool=False): import generic_scaled_masked_softmax_cuda for qlen, klen in self.q_k_lens: inputs = torch.normal(0, 2, (self.batch, self.attn, qlen, klen), dtype=self.dtype, device=self.device) masks = ( torch.randint(0, 2, (self.batch, 1, qlen, klen), dtype=torch.bool, device=self.device) if not allmasked else torch.ones((self.batch, 1, qlen, klen), dtype=torch.bool, device=self.device) ) softmax_results = generic_scaled_masked_softmax_cuda.forward(inputs, masks, self.scale_t) softmax_results_torch = forward_torch_softmax(inputs, masks, self.scale_t) self.assertEqual( softmax_results_torch.to(self.dtype), softmax_results, **self.thresh, msg=f"(q, k) = ({qlen, klen})") def test_backward(self, allmasked: bool=False): import generic_scaled_masked_softmax_cuda for qlen, klen in self.q_k_lens: inputs = torch.normal(0, 2, (self.batch, self.attn, qlen, klen), dtype=self.dtype, device=self.device) backward = torch.rand_like(inputs, dtype=torch.float16, device=self.device) masks = ( torch.randint(0, 2, (self.batch, 1, qlen, klen), dtype=torch.bool, device=self.device) if not allmasked else torch.ones((self.batch, 1, qlen, klen), dtype=torch.bool, device=self.device) ) softmax_results = generic_scaled_masked_softmax_cuda.forward(inputs, masks, self.scale_t) back_grad = generic_scaled_masked_softmax_cuda.backward(backward, softmax_results, self.scale_t) inputs.requires_grad = True softmax_results_torch = forward_torch_softmax(inputs, masks, self.scale_t) softmax_results_torch.backward(backward) self.assertEqual(back_grad, inputs.grad, **self.thresh, msg=f"(q, k) = ({qlen, klen})") def test_allmasked(self): self.test_forward(True) def test_allmask_backward(self): self.test_backward(True) if __name__ == "__main__": common_utils.run_tests()
apex-master
tests/L0/run_transformer/test_fused_softmax.py
import logging from typing import Tuple import torch import torch.nn.functional as F from torch.testing._internal import common_utils logging.getLogger("torch").setLevel(logging.WARNING) from apex.transformer import parallel_state from apex.transformer import tensor_parallel from apex.transformer.tensor_parallel import cross_entropy from apex.transformer.testing.commons import set_random_seed, IdentityLayer from apex.transformer.testing.distributed_test_base import NcclDistributedTestBase from apex.transformer.testing.distributed_test_base import UccDistributedTestBase logging.getLogger("apex").setLevel(logging.WARNING) def torch_cross_entropy( batch_size: int, seq_length: int, vocab_size: int, logits_scale: float, seed: int, label_smoothing: float = 0.0 ) -> Tuple[torch.Tensor, torch.Tensor]: set_random_seed(seed) identity = IdentityLayer( (batch_size, seq_length, vocab_size), scale=logits_scale ).cuda() logits = identity() target = torch.cuda.LongTensor(size=(batch_size, seq_length)).random_(0, vocab_size) loss = ( F.cross_entropy( logits.view(-1, logits.size()[-1]), target.view(-1), reduction="none", label_smoothing=label_smoothing ) .view_as(target) .mean() ) loss.backward() return loss, identity.weight.grad def tensor_sharded_cross_entropy( batch_size, seq_length, vocab_size, logits_scale, seed, label_smoothing=0.0 ): set_random_seed(seed) identity = IdentityLayer( (batch_size, seq_length, vocab_size), scale=logits_scale ).cuda() logits = identity() logits_parallel = tensor_parallel.scatter_to_tensor_model_parallel_region(logits) target = torch.cuda.LongTensor(size=(batch_size, seq_length)).random_(0, vocab_size) logits_parallel_ = logits_parallel.clone().detach() loss = cross_entropy.vocab_parallel_cross_entropy(logits_parallel, target, label_smoothing=label_smoothing).mean() loss.backward() # check for mutation assert torch.equal(logits_parallel_, logits_parallel) return loss, identity.weight.grad class VocabParallelCrossEntropyTestBase: def test_cross_entropy(self): batch_size, sequence_length, vocab_size_per_partition = 13, 17, 11 logits_scale = 1000.0 seed = 1234 for tensor_model_parallel_world_size in range(1, self.world_size + 1): if self.world_size % tensor_model_parallel_world_size: continue parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_world_size, ) vocab_size = vocab_size_per_partition * tensor_model_parallel_world_size loss_torch, grad_torch = torch_cross_entropy( batch_size, sequence_length, vocab_size, logits_scale, seed ) ( loss_tensor_parallel, grad_tensor_parallel, ) = tensor_sharded_cross_entropy( batch_size, sequence_length, vocab_size, logits_scale, seed ) self.assertEqual( loss_torch, loss_tensor_parallel, msg=f"tensor_model_parallel_size: {tensor_model_parallel_world_size}", ) self.assertEqual( grad_torch, grad_tensor_parallel, msg=f"tensor_model_parallel_size: {tensor_model_parallel_world_size}", ) parallel_state.destroy_model_parallel() class NcclVocabParallelCrossEntropyTest(VocabParallelCrossEntropyTestBase, NcclDistributedTestBase): pass class UccVocabParallelCrossEntropyTest(VocabParallelCrossEntropyTestBase, UccDistributedTestBase): pass if __name__ == "__main__": common_utils.run_tests()
apex-master
tests/L0/run_transformer/test_cross_entropy.py
from functools import partial from typing import List import time import torch import unittest from apex.transformer._ucc_util import HAS_UCC from apex.transformer import parallel_state from apex.transformer.enums import ModelType from apex.transformer.tensor_parallel import model_parallel_cuda_manual_seed from apex.transformer.pipeline_parallel.utils import ( average_losses_across_data_parallel_group, unwrap_model, setup_microbatch_calculator, get_ltor_masks_and_position_ids ) from apex.transformer.pipeline_parallel.schedules.common import ( _get_params_for_weight_decay_optimization, build_model ) from apex.transformer.pipeline_parallel.schedules.fwd_bwd_pipelining_without_interleaving import ( forward_backward_pipelining_without_interleaving, ) from apex.transformer.testing.standalone_gpt import gpt_model_provider from apex.transformer.testing import global_vars from apex.transformer.testing.distributed_test_base import UccDistributedTestBase, NcclDistributedTestBase from torch.testing._internal import common_utils from torch.testing._internal.common_device_type import instantiate_device_type_tests class GptTestBase: def _download_fancy_data(self): text = """ An original sentence not subject to any license restrictions, copyright, or royalty payments. Nothing to see here. Commercial or non-commercial use. Research or non-research purposes. The quick brown fox jumps over the lazy dog. Lorem ipsum. """ text = text * 1024 encoded = text.encode("ascii", "replace") ints = [int(encoded[i]) for i in range(len(encoded))] return torch.tensor(ints) # build a batch given sequence_len and batch size def _generate_fancy_data_labels(self, sequence_len, batch_size): temps = list() for i in range(batch_size): if self.inds is None or self.data_idx >= len(self.inds): # hack as use of RNG will fall out of sync due to pipelines being different model_parallel_cuda_manual_seed(self.MANUAL_SEED) self.inds = torch.randperm(effective_length, device="cuda") self.MANUAL_SEED += 1 self.data_idx = 0 data_idx_ = self.data_idx offset = self.inds[data_idx_] self.data_idx += 1 curr = fancy_data[offset: offset + sequence_len + 1].clone().detach() temps.append(curr) temp = torch.stack(temps, dim=0).cuda() return temp def _get_batch(self, int_tensors: List[torch.Tensor]): data = int_tensors[0] # Unpack. tokens_ = data.long() labels = tokens_[:, 1:].contiguous() tokens = tokens_[:, :-1].contiguous() # Get the masks and position ids. attention_mask, loss_mask, position_ids = get_ltor_masks_and_position_ids( tokens, self.N_VOCAB, # tokenizer.eod, False, # args.reset_position_ids, False, # args.reset_attention_mask, False, # args.eod_mask_loss, ) return tokens, labels, loss_mask, attention_mask, position_ids # Ref: https://github.com/NVIDIA/Megatron-LM/blob/b31e1296354e979722627a6c4dedafe19b51fa97/pretrain_gpt.py#L75 def _loss_func(self, loss_mask, output_tensor): losses = output_tensor.float() loss_mask = loss_mask.view(-1).float() loss = torch.sum(losses.view(-1) * loss_mask) / loss_mask.sum() # Reduce loss for logging. averaged_loss = average_losses_across_data_parallel_group([loss]) return loss, {"lm loss": averaged_loss[0]} # Ref: https://github.com/NVIDIA/Megatron-LM/blob/b31e1296354e979722627a6c4dedafe19b51fa97/pretrain_gpt.py#L86 def _fwd_step_func(self, batch, model): """Forward step.""" tokens, labels, loss_mask, attention_mask, position_ids = self._get_batch( batch) output_tensor = model(tokens, position_ids, attention_mask, labels=labels) return output_tensor, partial(self._loss_func, loss_mask) def _train(self, model, optim, pipeline_model_parallel_size, async_comm): args = global_vars.get_args() fwd_bwd_func = forward_backward_pipelining_without_interleaving tensor_shape = (args.seq_length, args.micro_batch_size, args.hidden_size) runtime = 0 # training loop for i in range(3): since = time.time() if torch.distributed.get_rank() == 0: print("begin iter", i) batch = [ self._generate_fancy_data_labels( args.seq_length, args.global_batch_size) for _ in range(pipeline_model_parallel_size) ] if torch.distributed.get_rank() == 0: print("finished making batch...") optim.zero_grad() fwd_bwd_func( self._fwd_step_func, batch, model, forward_only=False, tensor_shape=tensor_shape, async_comm=async_comm, sequence_parallel_enabled=args.sequence_parallel, ) if torch.distributed.get_rank() == 0: print("finished forward step") # All-reduce layernorm parameters across model parallel nodes # when sequence parallelism is used if parallel_state.get_tensor_model_parallel_world_size() > 1 and global_vars.get_args().sequence_parallel: for model_module in model: unwrapped_model = unwrap_model(model_module) for param in unwrapped_model.parameters(): if getattr(param, 'sequence_parallel_enabled', False): grad = param.grad torch.distributed.all_reduce( grad, group=parallel_state.get_tensor_model_parallel_group()) optim.step() if torch.distributed.get_rank() == 0: print("finished iter", i) runtime += time.time() - since return runtime / 3.0 @unittest.skipUnless(torch.cuda.device_count() > 2, "requires at least 3 gpus") def test_gpt(self): self.MANUAL_SEED = 42 self.inds = None self.data_idx = 0 self.N_VOCAB = 128 init = True tensor_model_parallel_size = 2 if self.world_size % 2 == 0 and self.world_size >= 4 else 1 pipeline_model_parallel_size = self.world_size // tensor_model_parallel_size override_args = { "micro_batch_size": 2, "num_layers": 16, "hidden_size": 256, "num_attention_heads": 8, "max_position_embeddings": 512, "seq_length": 512, "global_batch_size": 128, "pipeline_model_parallel_size": pipeline_model_parallel_size, "tensor_model_parallel_size": tensor_model_parallel_size, "world_size": self.world_size, "rank": self.rank, } global_vars.set_global_variables(override_args=override_args, ignore_unknown_args=True) args = global_vars.get_args() for async_comm in (False,) if args.sequence_parallel else (False, True): global fancy_data global effective_length if init: init = False fancy_data = self._download_fancy_data() args = global_vars.get_args() args.model_type = ModelType.encoder_or_decoder effective_length = fancy_data.size(0) // args.seq_length effective_length = fancy_data.size(0) - args.seq_length args.padded_vocab_size = 128 setup_microbatch_calculator( args.rank, args.rampup_batch_size, args.global_batch_size, args.micro_batch_size, args.data_parallel_size, ) print(args.tensor_model_parallel_size, "MODEL PARALLEL SIZE") parallel_state.initialize_model_parallel( tensor_model_parallel_size_=args.tensor_model_parallel_size, pipeline_model_parallel_size_=args.pipeline_model_parallel_size, default_backend="nccl", p2p_backend=self.DISTRIBUTED_BACKEND, ) model_parallel_cuda_manual_seed(0) model = build_model( gpt_model_provider, wrap_with_ddp=parallel_state.get_data_parallel_world_size() > 1, virtual_pipeline_model_parallel_size=None, cpu_offload=args.cpu_offload, ) assert isinstance(model, list), model _param_groups = _get_params_for_weight_decay_optimization(model) optim = torch.optim.Adam(_param_groups) runtime = self._train( model, optim, args.pipeline_model_parallel_size, async_comm) parallel_state.destroy_model_parallel() torch.cuda.synchronize() class NcclGptTest(GptTestBase, NcclDistributedTestBase): @property def world_size(self) -> int: return min(torch.cuda.device_count(), 8) @unittest.skipUnless(HAS_UCC, "requires pytorch to be built with native ucc") class UccGptTest(GptTestBase, UccDistributedTestBase): @property def world_size(self) -> int: return min(torch.cuda.device_count(), 8) if __name__ == "__main__": common_utils.run_tests()
apex-master
tests/L0/run_transformer/test_gpt_minimal.py
import torch from torch.testing._internal import common_utils from torch.utils.data import Dataset from torch.utils.data import DataLoader from apex.transformer.pipeline_parallel.utils import _split_batch_into_microbatch as split_batch_into_microbatch class MyIterableDataset(Dataset): def __init__(self, start, end): super().__init__() assert end > start, "this example code only works with end >= start" self.start = start self.end = end self.samples = list(range(self.start, self.end)) def __iter__(self): return iter(range(self.start, self.end)) def __getitem__(self, index): return self.samples[index] class MegatronPretrainingRandomSampler: def __init__(self, total_samples, consumed_samples, micro_batch_size, data_parallel_rank, data_parallel_size): # Keep a copy of input params for later use. self.total_samples = total_samples self.consumed_samples = consumed_samples self.micro_batch_size = micro_batch_size self.data_parallel_rank = data_parallel_rank self.data_parallel_size = data_parallel_size self.micro_batch_times_data_parallel_size = \ self.micro_batch_size * data_parallel_size self.last_batch_size = \ self.total_samples % self.micro_batch_times_data_parallel_size # Sanity checks. assert self.total_samples > 0, \ 'no sample to consume: {}'.format(self.total_samples) assert self.micro_batch_size > 0 assert data_parallel_size > 0 assert self.data_parallel_rank < data_parallel_size, \ 'data_parallel_rank should be smaller than data size: {}, ' \ '{}'.format(self.data_parallel_rank, data_parallel_size) def __len__(self): return self.total_samples def __iter__(self): active_total_samples = self.total_samples - self.last_batch_size self.epoch = self.consumed_samples // active_total_samples current_epoch_samples = self.consumed_samples % active_total_samples assert current_epoch_samples % self.micro_batch_times_data_parallel_size == 0 # data sharding and random sampling bucket_size = (self.total_samples // self.micro_batch_times_data_parallel_size) * self.micro_batch_size bucket_offset = current_epoch_samples // self.data_parallel_size start_idx = self.data_parallel_rank * bucket_size g = torch.Generator() g.manual_seed(self.epoch) random_idx = torch.randperm(bucket_size, generator=g).tolist() idx_range = [start_idx + x for x in random_idx[bucket_offset:]] batch = [] # Last batch if not complete will be dropped. for idx in idx_range: batch.append(idx) if len(batch) == self.micro_batch_size: self.consumed_samples += self.micro_batch_times_data_parallel_size yield batch batch = [] # Samples 8 tensors in total. # First sample 4 tensors twice, then sample 2 tensors fourth. class TestBatchSamplerBehavior(common_utils.TestCase): def tearDown(self) -> None: torch.cuda.empty_cache() super().tearDown() def test_batch_sampler_behavior(self): dataset = MyIterableDataset(0, 100) for num_workers in (1, 2, 4): torch.manual_seed(42) loader = DataLoader(dataset, batch_sampler=MegatronPretrainingRandomSampler(100, 0, 4, 0, 1), num_workers=num_workers) samples = [] for i, batch in enumerate(loader): samples.append(batch) if i == 2 - 1: break torch.manual_seed(42) loader = DataLoader(dataset, batch_sampler=MegatronPretrainingRandomSampler(100, 0, 2, 0, 1), num_workers=num_workers) samples2 = [] for i, batch in enumerate(loader): samples2.append(batch) if i == 4 - 1: break self.assertEqual(torch.cat(samples), torch.cat(samples2), msg=f"num_workers={num_workers}") def test_split_batch(self): class MyIterableDataset(Dataset): def __init__(self, start, end): super().__init__() assert end > start, "this example code only works with end >= start" self.start = start self.end = end self.samples = list(range(self.start, self.end)) def __len__(self): return self.end - self.start def __iter__(self): return iter(range(self.start, self.end)) def __getitem__(self, index): return (torch.tensor([index, index]), torch.tensor([index // 2, index // 2])) dataset = MyIterableDataset(0, 100) torch.manual_seed(42) global_batch_size = 16 loader = DataLoader(dataset, batch_sampler=MegatronPretrainingRandomSampler(100, 0, global_batch_size, 0, 1), num_workers=2) batch = next(iter(loader)) for _micro_batch_size in (1, 2, 4, 8): microbatches = list(split_batch_into_microbatch( batch, _micro_batch_size=_micro_batch_size, _global_batch_size=global_batch_size, )) self.assertEqual(len(microbatches), global_batch_size // _micro_batch_size) self.assertEqual(len(microbatches[0][0]), _micro_batch_size) if __name__ == "__main__": common_utils.run_tests()
apex-master
tests/L0/run_transformer/test_batch_sampler.py
import logging import unittest import typing import torch import torch.nn as nn from torch.testing._internal import common_utils from apex.transformer import parallel_state from apex.transformer.tensor_parallel import layers from apex.transformer.testing.commons import set_random_seed from apex.transformer.testing.distributed_test_base import NcclDistributedTestBase from apex.transformer.testing.distributed_test_base import UccDistributedTestBase logging.getLogger("torch").setLevel(logging.WARNING) logging.getLogger("apex").setLevel(logging.WARNING) # N.B.(mkozuki): Disable TF32 matrix multiply. # Matrices used in this test are so small that TF32 matmul # can be less precise so that `self.assertEqual` raises. torch.backends.cuda.matmul.allow_tf32 = False class TensorParallelLayerTestBase: BATCH_SIZE: int = 8 SEQUENCE_LENGTH: int = 128 VOCAB_SIZE: int = 1024 HIDDEN_SIZE: int = 256 INPUT_SIZE_COEFF: int = 256 OUTPUT_SIZE_COEFF: int = 256 SEED: int = 123456 @property def tensor_shape(self) -> typing.Sequence[int]: return [self.SEQUENCE_LENGTH, self.BATCH_SIZE, self.HIDDEN_SIZE] @torch.no_grad() @unittest.skipIf(torch.cuda.device_count() < 2, "Requires >=2 GPUs") def test_all_gather_parity(self) -> None: if self.DISTRIBUTED_BACKEND == "ucc": self.skipTest("torch_ucc does NOT support `torch.distributed._all_gather_base` as of 2022/06/15") from torch.distributed.distributed_c10d import all_gather, _all_gather_base # NOQA for tensor_model_parallel_world_size in range(1, self.world_size + 1): if self.world_size % tensor_model_parallel_world_size: continue parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_world_size, ) tensor_model_parallel_rank = parallel_state.get_tensor_model_parallel_rank() cur_tensor_model_device = torch.device(f"cuda:{tensor_model_parallel_rank}") with torch.no_grad(): tensor = tensor_model_parallel_rank * torch.ones( self.tensor_shape, dtype=torch.float32, device=cur_tensor_model_device) numel = tensor.numel() numel_gathered = tensor_model_parallel_world_size * numel gathered = torch.empty( torch.Size((numel_gathered,)), device=cur_tensor_model_device, dtype=torch.float32, requires_grad=False, ) chunks = [ gathered[i * numel : (i + 1) * numel] for i in range(tensor_model_parallel_world_size) ] all_gather(chunks, tensor, group=parallel_state.get_tensor_model_parallel_group()) gathered_for_base = torch.empty( torch.Size((numel_gathered,)), device=cur_tensor_model_device, dtype=torch.float32, requires_grad=False, ) _all_gather_base( gathered_for_base, tensor, group=parallel_state.get_tensor_model_parallel_group(), ) msg = f"tensor_model_parallel_world_size: {tensor_model_parallel_world_size}" self.assertEqual(gathered, gathered_for_base, msg=msg) parallel_state.destroy_model_parallel() @torch.no_grad() @unittest.skipIf(torch.cuda.device_count() < 2, "Requires >=2 GPUs") def test_reduce_scatter_parity(self) -> None: if self.DISTRIBUTED_BACKEND == "ucc": self.skipTest("torch_ucc does NOT support `torch.distributed._reduce_scatter_base` as of 2022/06/15") from torch.distributed.distributed_c10d import reduce_scatter, _reduce_scatter_base # NOQA for tensor_model_parallel_world_size in range(2, self.world_size + 1): if self.world_size % tensor_model_parallel_world_size: continue parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_world_size, ) tensor_model_parallel_rank = parallel_state.get_tensor_model_parallel_rank() cur_tensor_model_device = torch.device(f"cuda:{tensor_model_parallel_rank}") with torch.no_grad(): input = torch.cat([ i * torch.ones(self.tensor_shape, dtype=torch.float32, device=cur_tensor_model_device) for i in range(tensor_model_parallel_world_size) ]) input_list = [t.clone() for t in input.chunk(tensor_model_parallel_world_size)] output = torch.empty( self.tensor_shape, device=cur_tensor_model_device, dtype=torch.float32, requires_grad=False, ) reduce_scatter( output, input_list, group=parallel_state.get_tensor_model_parallel_group(), ) output_for_base = torch.empty( self.tensor_shape, device=cur_tensor_model_device, dtype=torch.float32, requires_grad=False, ) _reduce_scatter_base( output_for_base, input, group=parallel_state.get_tensor_model_parallel_group(), ) msg = f"tensor_model_parallel_world_size: {tensor_model_parallel_world_size}" self.assertEqual(output, output_for_base, msg=msg) self.assertEqual(input, torch.cat(input_list), msg=msg) parallel_state.destroy_model_parallel() def test_parallel_embedding(self) -> None: for tensor_model_parallel_world_size in range(1, self.world_size + 1): if self.world_size % tensor_model_parallel_world_size: continue parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_world_size, ) set_random_seed(self.SEED + 1) input_tensor = torch.randint( 0, self.VOCAB_SIZE, ( self.BATCH_SIZE, self.SEQUENCE_LENGTH, ), device="cuda", ) loss_weight = torch.randn( ( self.BATCH_SIZE, self.SEQUENCE_LENGTH, self.HIDDEN_SIZE, ), device="cuda", ) set_random_seed(self.SEED) embedding_torch = nn.Embedding( self.VOCAB_SIZE, self.HIDDEN_SIZE, ).cuda() output_torch = embedding_torch(input_tensor) loss_torch = torch.mul(output_torch, loss_weight).sum() loss_torch.backward() # N.B.(mkozuki): With affine weight initialization on GPU, # it's super difficult to keep the consistency with nn.Embedding. # Thus, turning on `use_cpu_initialization`. set_random_seed(self.SEED) embedding_vocab_parallel = layers.VocabParallelEmbedding( self.VOCAB_SIZE, self.HIDDEN_SIZE, init_method=nn.init.normal_, use_cpu_initialization=True, ).cuda() output_vocab_parallel = embedding_vocab_parallel(input_tensor) loss_vocab_parallel = torch.mul( output_vocab_parallel, loss_weight ).sum() loss_vocab_parallel.backward() msg = f"tensor_model_parallel_world_size: {tensor_model_parallel_world_size}" self.assertEqual(output_torch, output_vocab_parallel, msg=msg) self.assertEqual(loss_torch, loss_vocab_parallel, msg=msg) splitted_weight_torch = torch.split( embedding_torch.weight.grad, self.VOCAB_SIZE // tensor_model_parallel_world_size, 0, )[parallel_state.get_tensor_model_parallel_rank()] self.assertEqual( splitted_weight_torch, embedding_vocab_parallel.weight.grad, msg=msg, ) parallel_state.destroy_model_parallel() def _affine_weight_init_test_impl( self, init_device: str, is_column_parallel: bool ) -> None: dim = int(not is_column_parallel) for tensor_model_parallel_world_size in range(1, self.world_size + 1): if self.world_size % tensor_model_parallel_world_size: continue parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_world_size ) input_size: int = self.INPUT_SIZE_COEFF * tensor_model_parallel_world_size output_size: int = self.OUTPUT_SIZE_COEFF * tensor_model_parallel_world_size weight_shape = ( (self.OUTPUT_SIZE_COEFF, input_size) if is_column_parallel else (output_size, self.INPUT_SIZE_COEFF) ) weight = torch.empty(weight_shape) set_random_seed(self.SEED) sharding_dim_size = ( self.OUTPUT_SIZE_COEFF if is_column_parallel else self.INPUT_SIZE_COEFF ) if init_device == "cpu": layers._initialize_affine_weight_cpu( weight, output_size, input_size, sharding_dim_size, dim, nn.init.normal_, params_dtype=torch.float32, ) else: layers._initialize_affine_weight_gpu( weight, torch.nn.init.normal_, dim ) # Target set_random_seed(self.SEED) if init_device == "cpu": main_weight = torch.empty(output_size, input_size) nn.init.normal_(main_weight) curr_weight = torch.split(main_weight, sharding_dim_size, dim=dim)[ parallel_state.get_tensor_model_parallel_rank() ] else: curr_weight = torch.empty(*weight_shape) nn.init.normal_(curr_weight) self.assertEqual( curr_weight, weight, msg=f"tensor_model_parallel_world_size: {tensor_model_parallel_world_size}") parallel_state.destroy_model_parallel() def test_affine_weight_init_column_parallel_cpu(self) -> None: self._affine_weight_init_test_impl(init_device="cpu", is_column_parallel=True) def test_affine_weight_init_column_parallel_gpu(self) -> None: self._affine_weight_init_test_impl(init_device="gpu", is_column_parallel=True) def test_affine_weight_init_row_parallel_cpu(self) -> None: self._affine_weight_init_test_impl(init_device="cpu", is_column_parallel=False) def test_affine_weight_init_row_parallel_gpu(self) -> None: self._affine_weight_init_test_impl(init_device="gpu", is_column_parallel=False) def test_row_parallel_linear(self) -> None: self._row_parallel_linear_test_impl(False, False, False) def test_row_parallel_linear_gradient_accumulation_fusion(self) -> None: self._row_parallel_linear_test_impl(True, False, False) def test_row_parallel_linear_gradient_accumulation_fusion_in_fp16(self) -> None: self._row_parallel_linear_test_impl(True, True, False) # fails on native ucc and torch ucc: ucc does not support reduce scatter @unittest.skipIf(torch.cuda.device_count() < 2, "Sequence Parallel requires >=2 GPUs") def test_row_parallel_linear_sequence_parallel(self) -> None: self._row_parallel_linear_test_impl(False, False, True) # TODO(mkozuki): Merge this with `_column_parallel_linear_test_impl` # Note that `input_is_parallel` is unique to `RowParallelLinear` which could make the merge complicated. def _row_parallel_linear_test_impl( self, gradient_accumulation_fusion: bool, accumulation_in_fp16: bool, sequence_parallel_enabled: bool, ) -> None: tensor_shape = ( self.SEQUENCE_LENGTH, self.BATCH_SIZE, self.HIDDEN_SIZE, ) for tensor_model_parallel_world_size in range( 1 + int(sequence_parallel_enabled), self.world_size + 1 ): if self.world_size % tensor_model_parallel_world_size: continue parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_world_size, ) set_random_seed(self.SEED) linear = layers.RowParallelLinear( self.HIDDEN_SIZE, self.HIDDEN_SIZE, keep_master_weight_for_test=True, params_dtype=torch.float32, use_cpu_initialization=True, gradient_accumulation_fusion=gradient_accumulation_fusion, accumulation_in_fp16=accumulation_in_fp16, sequence_parallel_enabled=sequence_parallel_enabled, # n.b.(mkozuki): RowParallelLinear is constructed with `input_is_parallel=True` # by default, e.g. https://github.com/NVIDIA/NeMo/blob/782b4e1652aaa43c8be390d9\ # db0dc89544afa080/nemo/collections/nlp/modules/common/megatron/transformer.py#L204 input_is_parallel=True, ).cuda() if accumulation_in_fp16: linear = linear.half() # Simulate the situation where fusion of weight grad calculation and gradient accumulation is enabled. if gradient_accumulation_fusion: with torch.no_grad(): linear.weight.main_grad = torch.zeros_like(linear.weight) msg = f"tensor_model_parallel_world_size: {tensor_model_parallel_world_size}" with torch.no_grad(): orig_input_tensor = torch.randn(tensor_shape, requires_grad=True, device="cuda") orig_loss_weight = torch.randn(tensor_shape, device="cuda") input_tensor = orig_input_tensor.chunk( chunks=tensor_model_parallel_world_size, dim=2, )[parallel_state.get_tensor_model_parallel_rank()].contiguous() if sequence_parallel_enabled: loss_weight = orig_loss_weight.chunk( chunks=tensor_model_parallel_world_size, dim=0, )[parallel_state.get_tensor_model_parallel_rank()] else: loss_weight = orig_loss_weight if accumulation_in_fp16: orig_input_tensor = orig_input_tensor.half() input_tensor = input_tensor.half() loss_weight = loss_weight.half() input_tensor.requires_grad_() output, _ = linear(input_tensor) loss = torch.mul(output, loss_weight).sum() loss.backward() self.assertIsNotNone(input_tensor.grad, msg=msg) ref_linear = nn.Linear( in_features=self.HIDDEN_SIZE, out_features=self.HIDDEN_SIZE, bias=False, device="cuda", ) with torch.no_grad(): dldy = orig_loss_weight.clone() x = orig_input_tensor.clone() ref_linear.weight.copy_(linear.master_weight) if accumulation_in_fp16: ref_linear = ref_linear.half() x.requires_grad_() expected_output = ref_linear(x) expected_loss = torch.mul(expected_output, dldy).sum() expected_loss.backward() if not accumulation_in_fp16: if sequence_parallel_enabled: self.assertEqual( x=output, y=expected_output.chunk( chunks=tensor_model_parallel_world_size, dim=0, )[parallel_state.get_tensor_model_parallel_rank()], msg=msg, ) else: self.assertEqual( x=output, y=expected_output, msg=msg, ) grad_attr_name = "main_grad" if gradient_accumulation_fusion else "grad" # NOTE(mkozuki): Numerical errors seems to be enlarged by tensor model parallel. if tensor_model_parallel_world_size == 1: self.assertEqual( x=getattr(linear.weight, grad_attr_name), y=ref_linear.weight.grad.chunk( chunks=tensor_model_parallel_world_size, dim=0, )[parallel_state.get_tensor_model_parallel_rank()], msg=msg, ) parallel_state.destroy_model_parallel() def test_column_parallel_linear(self): self._column_parallel_linear_test_impl(False, False, False, False) def test_column_parallel_linear_async(self): self._column_parallel_linear_test_impl(True, False, False, False) def test_column_parallel_linear_gradient_accumulation_fusion(self): self._column_parallel_linear_test_impl(False, True, False, False) def test_column_parallel_linear_gradient_accumulation_fusion_in_fp16(self): self._column_parallel_linear_test_impl(False, True, True, False) def test_column_parallel_linear_sequence_parallel(self): if self.DISTRIBUTED_BACKEND == "ucc": self.skipTest("Backward's reduce_scatter fails. as of 2022/06/15") self._column_parallel_linear_test_impl(False, False, False, True) @unittest.skipIf(torch.cuda.device_count() < 2, "Sequence Parallel requires >= 2 GPUs") def test_column_parallel_linear_exception(self): with self.assertRaisesRegex( RuntimeError, "`async_tensor_model_parallel_allreduce` and `sequence_parallel_enabled` cannot be enabled at the same time.", ): self._column_parallel_linear_test_impl(True, False, False, True) def _column_parallel_linear_test_impl( self, async_tensor_model_parallel_allreduce: bool, gradient_accumulation_fusion: bool, accumulation_in_fp16: bool, sequence_parallel_enabled: bool, ): for tensor_model_parallel_world_size in range(1, self.world_size + 1): if async_tensor_model_parallel_allreduce and sequence_parallel_enabled: if tensor_model_parallel_world_size == 1: continue if self.world_size % tensor_model_parallel_world_size: continue msg = f"tensor_model_parallel_world_size: {tensor_model_parallel_world_size}" parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_world_size, ) input_tensor_shape = self.tensor_shape expected_output_shape = self.tensor_shape # When sequence parallel, `gather_output` is disabled, i.e., # output of matmul isn't gathered in dimension of feature/hidden (last dim). if sequence_parallel_enabled: expected_output_shape[-1] //= tensor_model_parallel_world_size # tensor's shape is [sequence length, batch size, hidden size] set_random_seed(self.SEED) linear = layers.ColumnParallelLinear( self.HIDDEN_SIZE, self.HIDDEN_SIZE, bias=False, keep_master_weight_for_test=True, params_dtype=torch.float32, use_cpu_initialization=True, gather_output=not sequence_parallel_enabled, no_async_tensor_model_parallel_allreduce=not async_tensor_model_parallel_allreduce, gradient_accumulation_fusion=gradient_accumulation_fusion, accumulation_in_fp16=accumulation_in_fp16, sequence_parallel_enabled=sequence_parallel_enabled, ).cuda() if accumulation_in_fp16: linear = linear.half() # Simulate the situation where fusion of weight grad calculation and gradient accumulation happens. if gradient_accumulation_fusion: with torch.no_grad(): linear.weight.main_grad = torch.zeros_like(linear.weight) orig_input_tensor = torch.randn(input_tensor_shape, device="cuda", requires_grad=True) if accumulation_in_fp16: orig_input_tensor = orig_input_tensor.half() if sequence_parallel_enabled: input_tensor = list( orig_input_tensor.chunk(tensor_model_parallel_world_size, dim=0) )[parallel_state.get_tensor_model_parallel_rank()] else: input_tensor = orig_input_tensor output, _ = linear(input_tensor) # The order of dimension is expected to be (sequence, batch, hidden) self.assertEqual(output.shape, expected_output_shape, msg=msg) orig_loss_weight = torch.randn(input_tensor_shape, device="cuda") if accumulation_in_fp16: orig_loss_weight = orig_loss_weight.half() if sequence_parallel_enabled: loss_weight = orig_loss_weight.chunk( tensor_model_parallel_world_size, dim=2, )[parallel_state.get_tensor_model_parallel_rank()] else: loss_weight = orig_loss_weight loss = torch.mul(output, loss_weight).sum() loss.backward() with torch.no_grad(): dldy = orig_loss_weight.clone() x = orig_input_tensor.clone() ref_linear = nn.Linear( in_features=self.HIDDEN_SIZE, out_features=self.HIDDEN_SIZE, bias=False, device="cuda", ) if accumulation_in_fp16: ref_linear = ref_linear.half() # NOTE(mkozuki): `master_weight` is available because `keep_master_weight_for_test` is set. ref_linear.weight.copy_(linear.master_weight) x.requires_grad_() expected_output = ref_linear(x) if sequence_parallel_enabled: chunk = expected_output.chunk( tensor_model_parallel_world_size, dim=2, )[parallel_state.get_tensor_model_parallel_rank()] self.assertEqual( x=output, y=chunk, msg=msg, ) else: self.assertEqual( x=output, y=expected_output, msg=msg, ) expected_loss = torch.mul(expected_output, dldy).sum() expected_loss.backward() grad_attr_name = "main_grad" if gradient_accumulation_fusion else "grad" # NOTE(mkozuki): Numerical errors seems to be enlarged by tensor model parallel. if tensor_model_parallel_world_size == 1: self.assertEqual( x=getattr(linear.weight, grad_attr_name), y=ref_linear.weight.grad.chunk( chunks=tensor_model_parallel_world_size, dim=0, )[parallel_state.get_tensor_model_parallel_rank()], msg=msg, ) parallel_state.destroy_model_parallel() class NcclTensorParallelLayerTest(TensorParallelLayerTestBase, NcclDistributedTestBase): pass class UccTensorParallelLayerTest(TensorParallelLayerTestBase, UccDistributedTestBase): pass if __name__ == "__main__": common_utils.run_tests()
apex-master
tests/L0/run_transformer/test_layers.py
apex-master
tests/L0/run_transformer/__init__.py
import logging from typing import List, Optional from torch.testing._internal import common_utils logging.getLogger("torch").setLevel(logging.WARNING) from apex.transformer import parallel_state from apex.transformer.pipeline_parallel.utils import ( _reconfigure_microbatch_calculator, get_micro_batch_size, get_num_microbatches, get_current_global_batch_size, update_num_microbatches, ) from apex.transformer.testing.distributed_test_base import NcclDistributedTestBase from apex.transformer.testing.distributed_test_base import UccDistributedTestBase logging.getLogger("apex").setLevel(logging.WARNING) class MicrobatchCalculatorTestBase: GLOBAL_BATCH_SIZE: int = 1024 MICRO_BATCH_SIZE: int = 1 def _test(self, rampup_batch_size: Optional[List[int]]) -> None: for data_parallel_size in range(1, self.world_size + 1): expected_global_batch_size = self.GLOBAL_BATCH_SIZE expected_micro_batch_size = self.MICRO_BATCH_SIZE if rampup_batch_size: expected_global_batch_size = rampup_batch_size[0] num_consumed_samples = 0 step_of_global_batch_size = rampup_batch_size[1] threshold = rampup_batch_size[2] if data_parallel_size > 1 and data_parallel_size % 2 != 0: continue if self.world_size % data_parallel_size != 0: continue msg = f"data_parallel_size: {data_parallel_size}" parallel_state.initialize_model_parallel( tensor_model_parallel_size_=self.world_size // data_parallel_size, pipeline_model_parallel_size_=1, ) self.assertEqual(data_parallel_size, parallel_state.get_data_parallel_world_size(), msg=msg) _reconfigure_microbatch_calculator( self.rank, rampup_batch_size, self.GLOBAL_BATCH_SIZE, self.MICRO_BATCH_SIZE, data_parallel_size, ) self.assertEqual(get_micro_batch_size(), expected_micro_batch_size, msg=msg) self.assertEqual(get_num_microbatches(), expected_global_batch_size / expected_micro_batch_size / data_parallel_size, msg=msg) current_global_batch_size = get_current_global_batch_size() self.assertEqual(current_global_batch_size, expected_global_batch_size, msg=msg) # Make sure `global_batch_size` equals to the final global batch size after # certain number of updates. if rampup_batch_size: update_num_microbatches(current_global_batch_size) for i in range(100): current_global_batch_size = get_current_global_batch_size() update_num_microbatches(current_global_batch_size) current_global_batch_size = get_current_global_batch_size() self.assertEqual(get_current_global_batch_size(), self.GLOBAL_BATCH_SIZE, msg=msg) parallel_state.destroy_model_parallel() def test_constant_microbatch_calculator(self): self._test(rampup_batch_size=None) def test_dynamic_microbatch_calculator(self): self._test(rampup_batch_size=[256, 128, 500]) class NcclMicrobatchCalculatorTest(MicrobatchCalculatorTestBase, NcclDistributedTestBase): pass class UccMicrobatchCalculatorTest(MicrobatchCalculatorTestBase, UccDistributedTestBase): pass if __name__ == "__main__": common_utils.run_tests()
apex-master
tests/L0/run_transformer/test_microbatches.py
import logging import torch.testing from torch.testing._internal import common_utils logging.getLogger("torch").setLevel(logging.WARNING) from apex.transformer import parallel_state from apex.transformer.tensor_parallel import data as data_utils from apex.transformer.testing.distributed_test_base import NcclDistributedTestBase from apex.transformer.testing.distributed_test_base import UccDistributedTestBase logging.getLogger("torch").setLevel(logging.WARNING) class BroadcastDataTestBase: def test_broadcast_data(self): tensor_model_parallel_world_size: int = self.world_size // ( 1 + self.world_size > 1 ) parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_world_size ) target_key_size = { "key1": [7, 11], "key2": [8, 2, 1], "key3": [13], "key4": [5, 1, 2], "key5": [5, 12], } keys = [k for k in target_key_size] data = {} data_t = {} with torch.no_grad(): for key in target_key_size: data[key] = torch.randint(0, 1000, size=target_key_size[key]) data_t[key] = data[key].clone() # "key_x" is supposed to be ignored. data["key_x"] = torch.rand(5) data_t["key_x"] = data["key_x"].clone() if parallel_state.get_tensor_model_parallel_rank() != 0: data = None data_utils._check_data_types(keys, data_t, torch.int64) key_size, _, _ = data_utils._build_key_size_numel_dictionaries(keys, data) for key in keys: self.assertEqual(target_key_size[key], key_size[key]) broadcasted_data = data_utils.broadcast_data(keys, data, torch.int64) for key in keys: self.assertEqual(broadcasted_data[key], data_t[key].cuda()) parallel_state.destroy_model_parallel() class NcclBroadcastDataTest(BroadcastDataTestBase, NcclDistributedTestBase): pass class UccBroadcastDataTest(BroadcastDataTestBase, UccDistributedTestBase): pass if __name__ == "__main__": common_utils.run_tests()
apex-master
tests/L0/run_transformer/test_data.py
import torch import unittest from apex.transformer.testing import global_vars from apex.transformer.testing.standalone_bert import bert_model_provider from apex.transformer.pipeline_parallel.schedules.common import ( _get_params_for_weight_decay_optimization, build_model ) from apex.transformer.pipeline_parallel.schedules import get_forward_backward_func from apex.transformer.pipeline_parallel.utils import ( average_losses_across_data_parallel_group, unwrap_model, setup_microbatch_calculator ) from apex.transformer.log_util import set_logging_level from apex.transformer import tensor_parallel, parallel_state from apex.transformer.enums import ModelType from apex.transformer._ucc_util import HAS_UCC from apex.transformer.testing.distributed_test_base import UccDistributedTestBase, NcclDistributedTestBase import logging from torch.testing._internal import common_utils logging.getLogger("torch").setLevel(logging.WARNING) logging.getLogger("apex").setLevel(logging.WARNING) set_logging_level("WARNING") class BertTestBase: def _download_fancy_data(self): text = """ An original sentence not subject to any license restrictions, copyright, or royalty payments. Nothing to see here. Commercial or non-commercial use. Research or non-research purposes. The quick brown fox jumps over the lazy dog. Lorem ipsum. """ text = text * 1024 encoded = text.encode("ascii", "replace") ints = [int(encoded[i]) for i in range(len(encoded))] return torch.tensor(ints) # build a batch given sequence_len and batch size def _generate_fancy_data_labels(self, sequence_len, batch_size): temps = [] for i in range(batch_size): if self.inds is None or self.data_idx >= len(self.inds): # hack as use of RNG will fall out of sync due to pipelines being different torch.manual_seed(self.MANUAL_SEED) self.inds = torch.randperm( self.effective_length, device="cuda") self.masks = ( torch.rand( len(self.inds) // batch_size + 1, batch_size, sequence_len, device="cuda" ) >= self.MASK_PROB ).long() self.MANUAL_SEED += 1 self.data_idx = 0 if self.rank == 0: print("new epoch", len(self.inds)) print("my start", self.inds[0:5]) print("masks_checksum:", torch.sum(self.masks)) if self.EASY_MODE: data_idx_ = self.data_idx % self.EASY_MODE_SIZ else: data_idx_ = self.data_idx offset = self.inds[data_idx_] # * SEQUENCE_LEN self.data_idx += 1 curr = self.fancy_data[offset: offset + sequence_len].clone().detach() temps.append(curr) temp = torch.stack(temps, dim=0).cuda() mask = self.masks[self.data_idx // batch_size] mask_not = torch.logical_not(mask).long() data = mask * temp + mask_not * 124 label = temp if parallel_state.get_tensor_model_parallel_rank() == 0: data_dict = {"text": data, "label": label, "mask_not": mask_not} else: data_dict = None keys = ["text", "label", "mask_not"] broadcasted_data = tensor_parallel.broadcast_data( keys, data_dict, torch.long) return ( broadcasted_data["text"].long(), broadcasted_data["label"].long(), broadcasted_data["mask_not"], ) def _fwd_step_func(self, batch, model): data, label, loss_mask = batch y = model(data, torch.ones_like(data), lm_labels=label) def loss_func(output_tensor): output_tensor, _ = output_tensor lm_loss_ = output_tensor.float() lm_loss = torch.sum(lm_loss_.view(-1) * loss_mask.reshape(-1)) / loss_mask.sum() averaged_loss = average_losses_across_data_parallel_group([ lm_loss]) if self.data_idx >= 1536: # NOTE (patwang): Loss cutoff might be excessively high but roughly one in five # unlucky random seeds do cause loss to spike to just under 8.0 self.assertLess(averaged_loss, 8.0) return lm_loss, {"avg": averaged_loss} return y, loss_func def _train( self, model, optim, virtual_pipeline_model_parallel_size, pipeline_model_parallel_size, async_comm ): args = global_vars.get_args() sequence_len = args.seq_length micro_batch_size = args.micro_batch_size hidden_size = args.hidden_size global_batch_size = args.global_batch_size forward_backward_func = get_forward_backward_func( virtual_pipeline_model_parallel_size, pipeline_model_parallel_size ) tensor_shape = (sequence_len, micro_batch_size, hidden_size) for _ in range(16): batch = self._generate_fancy_data_labels( sequence_len, global_batch_size) optim.zero_grad() forward_backward_func( self._fwd_step_func, batch, model, forward_only=False, tensor_shape=tensor_shape, async_comm=async_comm, sequence_parallel_enabled=args.sequence_parallel, ) # All-reduce layernorm parameters across model parallel nodes # when sequence parallelism is used if parallel_state.get_tensor_model_parallel_world_size() > 1 and args.sequence_parallel: for model_module in model: unwrapped_model = unwrap_model(model_module) for param in unwrapped_model.parameters(): if getattr(param, 'sequence_parallel_enabled', False): grad = param.grad torch.distributed.all_reduce( grad, group=parallel_state.get_tensor_model_parallel_group()) optim.step() @unittest.skipUnless(torch.cuda.device_count() > 2, "requires at least 3 gpus") def test_bert_without_interleaving(self): self._test_bert(virtual_pipeline_model_parallel_size=None) @unittest.skipUnless(torch.cuda.device_count() > 2, "requires at least 3 gpus") def test_bert_with_interleaving(self): if self.DISTRIBUTED_BACKEND == 'ucc': self.skipTest('skip interleaving with ucc') self._test_bert(virtual_pipeline_model_parallel_size=2) def _test_bert(self, virtual_pipeline_model_parallel_size): self.MANUAL_SEED = 42 self.inds = None self.masks = None self.data_idx = 0 self.MASK_PROB = 0.1 self.EASY_MODE = False self.EASY_MODE_SIZ = 32 tensor_model_parallel_size = 2 if self.world_size % 2 == 0 and self.world_size > 4 else 1 pipeline_model_parallel_size = self.world_size // tensor_model_parallel_size override_args = { "micro_batch_size": 2, "num_layers": 16, "hidden_size": 256, "num_attention_heads": 8, "max_position_embeddings": 512, "seq_length": 512, "global_batch_size": 128, "pipeline_model_parallel_size": pipeline_model_parallel_size, "tensor_model_parallel_size": tensor_model_parallel_size, "bert_binary_head": False, "world_size": self.world_size, "rank": self.rank, } global_vars.set_global_variables(override_args=override_args, ignore_unknown_args=True) args = global_vars.get_args() self.fancy_data = self._download_fancy_data() self.effective_length = self.fancy_data.size(0) // args.seq_length self.effective_length = self.fancy_data.size(0) - args.seq_length if self.rank == 0: print( f'testing backend: {self.DISTRIBUTED_BACKEND} with virtual_pipeline_model_parallel_size: {virtual_pipeline_model_parallel_size}') async_comm = not args.sequence_parallel and virtual_pipeline_model_parallel_size is None self.data_idx = 0 args.padded_vocab_size = 128 # needed in standalone gpt args.model_type = ModelType.encoder_or_decoder setup_microbatch_calculator( args.rank, args.rampup_batch_size, args.global_batch_size, args.micro_batch_size, args.data_parallel_size, ) parallel_state.initialize_model_parallel( args.tensor_model_parallel_size, args.pipeline_model_parallel_size, virtual_pipeline_model_parallel_size, default_backend="nccl", p2p_backend=self.DISTRIBUTED_BACKEND, ) tensor_parallel.random.model_parallel_cuda_manual_seed(0) model = build_model( bert_model_provider, wrap_with_ddp=parallel_state.get_data_parallel_world_size() > 1, virtual_pipeline_model_parallel_size=virtual_pipeline_model_parallel_size, cpu_offload=args.cpu_offload, ) assert isinstance(model, list) assert len(model) == ( 1 if virtual_pipeline_model_parallel_size is None else virtual_pipeline_model_parallel_size ) _param_groups = _get_params_for_weight_decay_optimization(model) optim = torch.optim.Adam(_param_groups) self._train( model, optim, virtual_pipeline_model_parallel_size, args.pipeline_model_parallel_size, async_comm, ) torch.cuda.synchronize() class NcclBertTest(BertTestBase, NcclDistributedTestBase): @property def world_size(self) -> int: return min(torch.cuda.device_count(), 8) @unittest.skipUnless(HAS_UCC, "requires pytorch to be built with native ucc") class UccBertTest(BertTestBase, UccDistributedTestBase): @property def world_size(self) -> int: return min(torch.cuda.device_count(), 8) if __name__ == "__main__": common_utils.run_tests()
apex-master
tests/L0/run_transformer/test_bert_minimal.py
import logging import torch from torch.testing._internal import common_utils logging.getLogger("torch").setLevel(logging.WARNING) from apex.transformer import parallel_state from apex.transformer.tensor_parallel import utils from apex.transformer.testing.distributed_test_base import NcclDistributedTestBase logging.getLogger("apex").setLevel(logging.WARNING) class TransformerUtilsTest(NcclDistributedTestBase): def test_split_tensor_along_last_dim(self): for tensor_model_paralell_world_size in range(1, self.world_size + 1): if self.world_size % tensor_model_paralell_world_size > 0: continue parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_paralell_world_size ) device = "cpu" input_tensor = torch.randn((100, 100, 100), device=device) splits = utils.split_tensor_along_last_dim(input_tensor, 10) last_dim_shapes = torch.tensor( [int(split.size()[-1]) for split in splits] ) self.assertTrue( torch.equal(last_dim_shapes, torch.full((10,), 10),), msg=f"tensor_model_paralell_world_size: {tensor_model_paralell_world_size}", ) parallel_state.destroy_model_parallel() if __name__ == "__main__": common_utils.run_tests()
apex-master
tests/L0/run_transformer/test_transformer_utils.py
import logging import os from torch.testing._internal import common_utils logging.getLogger("torch").setLevel(logging.WARNING) from apex.transformer import parallel_state from apex.transformer.testing.distributed_test_base import NcclDistributedTestBase from apex.transformer.testing.distributed_test_base import UccDistributedTestBase logging.getLogger("apex").setLevel(logging.WARNING) os.environ["BACKEND"] = "NCCL" DATA_PARALLEL_WORLD_SIZE: int = 1 def calc_expected_tensor_model_paralell_rank( rank: int, tensor_model_parallel_world_size: int, ) -> int: return rank % tensor_model_parallel_world_size class ParallelStateTestBase: def test_initialize_model_parallel(self) -> None: self.assertFalse(parallel_state.model_parallel_is_initialized()) for tensor_model_parallel_world_size in range(1, self.world_size + 1): msg = f"tensor_model_parallel_world_siz: {tensor_model_parallel_world_size}" if self.world_size % tensor_model_parallel_world_size: continue pipeline_model_parallel_world_size = ( self.world_size // tensor_model_parallel_world_size ) parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_world_size, pipeline_model_parallel_size_=pipeline_model_parallel_world_size, ) self.assertEqual( tensor_model_parallel_world_size, parallel_state.get_tensor_model_parallel_world_size(), msg=msg, ) expected_tensor_model_parallel_rank = calc_expected_tensor_model_paralell_rank( self.rank, tensor_model_parallel_world_size ) self.assertEqual( expected_tensor_model_parallel_rank, parallel_state.get_tensor_model_parallel_rank(), msg=msg, ) expected_tensor_model_parallel_src_rank = ( self.rank // tensor_model_parallel_world_size ) * tensor_model_parallel_world_size self.assertEqual( expected_tensor_model_parallel_src_rank, parallel_state.get_tensor_model_parallel_src_rank(), msg=msg, ) parallel_state.destroy_model_parallel() self.assertFalse(parallel_state.model_parallel_is_initialized(), msg=msg) def test_initialize_model_parallel_with_virtual_and_split(self) -> None: if self.world_size < 4: self.skipTest("requires >= 4 GPUs") self.assertFalse(parallel_state.model_parallel_is_initialized()) tensor_model_parallel_world_size = 1 + int(self.world_size > 4) pipeline_model_parallel_world_size = ( self.world_size // tensor_model_parallel_world_size ) virtual_pipeline_model_parallel_world_size = 2 pipeline_model_parallel_split_rank = pipeline_model_parallel_world_size // 2 parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_world_size, pipeline_model_parallel_size_=pipeline_model_parallel_world_size, virtual_pipeline_model_parallel_size_=virtual_pipeline_model_parallel_world_size, pipeline_model_parallel_split_rank_=pipeline_model_parallel_split_rank, ) self.assertEqual( calc_expected_tensor_model_paralell_rank( self.rank, tensor_model_parallel_world_size ), parallel_state.get_tensor_model_parallel_rank(), ) self.assertEqual( pipeline_model_parallel_world_size, parallel_state.get_pipeline_model_parallel_world_size(), ) self.assertEqual( virtual_pipeline_model_parallel_world_size, parallel_state.get_virtual_pipeline_model_parallel_world_size(), ) expected_pipeline_rank = ( self.rank - (self.rank % tensor_model_parallel_world_size) ) % pipeline_model_parallel_world_size self.assertEqual( expected_pipeline_rank, parallel_state.get_pipeline_model_parallel_rank(), ) # virtual pipeline model parallel rank is lazily set, i.e., right after the call of # `initialize_model_parallel`, it's set to 0. self.assertEqual( 0, parallel_state.get_virtual_pipeline_model_parallel_rank(), ) self.assertEqual( pipeline_model_parallel_split_rank, parallel_state.get_pipeline_model_parallel_split_rank(), ) fake_split_rank = 77 parallel_state.set_pipeline_model_parallel_split_rank(fake_split_rank) self.assertEqual( fake_split_rank, parallel_state.get_pipeline_model_parallel_split_rank() ) # relative position embedding groups check self.assertEqual( expected_pipeline_rank < pipeline_model_parallel_split_rank, parallel_state.is_rank_in_encoder_relative_position_embedding_group(), ) self.assertEqual( expected_pipeline_rank >= pipeline_model_parallel_split_rank, parallel_state.is_rank_in_decoder_relative_position_embedding_group(), ) parallel_state.destroy_model_parallel() def test_initialize_model_parallel_decoder_only(self) -> None: """Initialize model parallelism for decoder-only Transformers like GPT-3""" self.assertFalse(parallel_state.model_parallel_is_initialized()) for tensor_model_parallel_world_size in range(1, self.world_size + 1): msg = f"tensor_model_parallel_world_size: {tensor_model_parallel_world_size}" if self.world_size % tensor_model_parallel_world_size: continue pipeline_model_parallel_world_size = ( self.world_size // tensor_model_parallel_world_size ) parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_world_size, pipeline_model_parallel_size_=pipeline_model_parallel_world_size, pipeline_model_parallel_split_rank_=0, ) self.assertEqual( tensor_model_parallel_world_size, parallel_state.get_tensor_model_parallel_world_size(), msg=msg, ) expected_tensor_model_parallel_rank = calc_expected_tensor_model_paralell_rank( self.rank, tensor_model_parallel_world_size ) self.assertEqual( expected_tensor_model_parallel_rank, parallel_state.get_tensor_model_parallel_rank(), msg=msg, ) expected_tensor_model_parallel_src_rank = ( self.rank // tensor_model_parallel_world_size ) * tensor_model_parallel_world_size self.assertEqual( expected_tensor_model_parallel_src_rank, parallel_state.get_tensor_model_parallel_src_rank(), msg=msg, ) parallel_state.destroy_model_parallel() self.assertFalse(parallel_state.model_parallel_is_initialized(), msg=msg) class NcclParallelStateTest(ParallelStateTestBase, NcclDistributedTestBase): pass class UccParallelStateTest(ParallelStateTestBase, UccDistributedTestBase): pass if __name__ == "__main__": common_utils.run_tests()
apex-master
tests/L0/run_transformer/test_parallel_state.py
import contextlib import logging import itertools import os from datetime import datetime from packaging.version import parse, Version import re from typing import Optional, Tuple, List import unittest import torch from torch.testing._internal import common_utils from apex._autocast_utils import _get_autocast_dtypes from apex.transformer import parallel_state from apex.transformer.enums import ModelType from apex.transformer.pipeline_parallel import utils as pp_utils from apex.transformer.pipeline_parallel.schedules.common import ( FwdStepFunc, build_model, _get_params_for_weight_decay_optimization, ) from apex.transformer.pipeline_parallel.schedules.fwd_bwd_no_pipelining import ( forward_backward_no_pipelining, ) from apex.transformer.pipeline_parallel.schedules.fwd_bwd_pipelining_with_interleaving import ( _forward_backward_pipelining_with_interleaving, ) from apex.transformer.pipeline_parallel.schedules.fwd_bwd_pipelining_without_interleaving import ( forward_backward_pipelining_without_interleaving, ) from apex.transformer.testing.distributed_test_base import NcclDistributedTestBase from apex.transformer.testing.distributed_test_base import UccDistributedTestBase from apex.transformer.testing.distributed_test_base import HAS_TORCH_UCC_COMPAT_NVIDIA_DRIVER from apex.transformer.testing import commons as testing_utils from apex.transformer._ucc_util import HAS_UCC logging.getLogger("torch").setLevel(logging.WARNING) logging.getLogger("apex").setLevel(logging.WARNING) weight_coeff = 1024 # Guard for https://github.com/pytorch/pytorch/pull/82450 def get_nvidia_pytorch_version(): ver = os.getenv("NVIDIA_PYTORCH_VERSION", "22.08") if "master" in ver: ver = datetime.today().strftime("%y.%m") elif "update_for_" in ver: ver = ver.replace("update_for_", "") return ver CAN_SKIP_SYNC_AFTER_BATCH_ISEND_IRECV = False ngc_container_2209, pytorch_113 = Version("22.09"), Version("1.13") if parse(torch.__version__) >= pytorch_113: CAN_SKIP_SYNC_AFTER_BATCH_ISEND_IRECV = True elif parse(get_nvidia_pytorch_version()) >= ngc_container_2209: CAN_SKIP_SYNC_AFTER_BATCH_ISEND_IRECV = True else: CAN_SKIP_SYNC_AFTER_BATCH_ISEND_IRECV = False def get_init_weights_func(offset: int = 0): @torch.no_grad() def init_weights(m): rank = parallel_state.get_pipeline_model_parallel_rank() if isinstance(m, torch.nn.Linear): m.weight.fill_((rank + offset + 1.0) / weight_coeff) m.bias.fill_(1.0) return init_weights def get_dtype_for_comparison(): if(torch.cuda.get_device_capability() >= (8, 0)): return torch.float64 return torch.float32 def get_target_loss_and_model(global_batch_shape: tuple, hidden_size: int, total_layers: int) -> Tuple[torch.Tensor, List[torch.Tensor]]: model = [] dtype = get_dtype_for_comparison() data = torch.ones(global_batch_shape, dtype=dtype) for i in range(total_layers): w = torch.ones((hidden_size, hidden_size), dtype=dtype) * (i + 1.0) / weight_coeff b = torch.ones(hidden_size, dtype=dtype) w.requires_grad_() b.requires_grad_() # don't need to care about transpose semantics as all values are the same data = torch.matmul(w, data) + b model.append([w, b]) loss = data.sum() / global_batch_shape[0] loss.backward() return loss, model def _get_default_world_sizes_model_parallel_world_size(pipeline_model_parallel_world_size: Optional[int] = None ) -> Tuple[int, int, int]: # TODO: revisit if we can fold this into the class for skip logic / avoid duplication # of world size computation world_size = torch.cuda.device_count() tensor_model_parallel_world_size = 1 data_parallel_size = 1 + (world_size >= 8 and world_size % 2 == 0) if pipeline_model_parallel_world_size is None: pipeline_model_parallel_world_size = world_size // (tensor_model_parallel_world_size * data_parallel_size) else: data_parallel_size = world_size // (tensor_model_parallel_world_size * pipeline_model_parallel_world_size) return tensor_model_parallel_world_size, data_parallel_size, pipeline_model_parallel_world_size class PipelineParallelForwardBackwardTestBase: GLOBAL_BATCH_SIZE = 16 MICRO_BATCH_SIZE = 2 HIDDEN_SIZE = 32 deallocate_options = (True, False) # If :obj:`None`, (torch.float32, torch.float16, torch.bfloat16) are dtype options on Ampere. # You can limit the options by overriding the following `dtypes`. dtypes = None def _forward_backward_test_impl( self, forward_only: bool, fwd_bwd_func: FwdStepFunc, pipeline_model_parallel_world_size: Optional[int], virtual_pipeline_model_parallel_size: Optional[int], async_comm: bool = False, *, default_backend: Optional[str] = None, p2p_backend: Optional[str] = None, sync_batch_comm: bool = True, ) -> None: if fwd_bwd_func == _forward_backward_pipelining_with_interleaving: self.assertIsNotNone(virtual_pipeline_model_parallel_size) self.assertGreater(virtual_pipeline_model_parallel_size, 1) dtype_options = self.dtypes or [torch.float32, torch.double] + _get_autocast_dtypes() for dtype, deallocate_pipeline_outputs in itertools.product( dtype_options, self.deallocate_options, ): grad_scaler = ( torch.cuda.amp.GradScaler(init_scale=4.0) if dtype == torch.half else None ) (tensor_model_parallel_world_size, data_parallel_size, pipeline_model_parallel_world_size) = _get_default_world_sizes_model_parallel_world_size(pipeline_model_parallel_world_size) parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_world_size, pipeline_model_parallel_size_=pipeline_model_parallel_world_size, virtual_pipeline_model_parallel_size_=virtual_pipeline_model_parallel_size, default_backend=default_backend, p2p_backend=p2p_backend, ) pp_utils._reconfigure_microbatch_calculator( rank=parallel_state.get_tensor_model_parallel_rank(), rampup_batch_size=None, global_batch_size=self.GLOBAL_BATCH_SIZE, micro_batch_size=self.MICRO_BATCH_SIZE, data_parallel_size=parallel_state.get_data_parallel_world_size(), ) global_batch_shape = ( self.GLOBAL_BATCH_SIZE // parallel_state.get_data_parallel_world_size(), self.HIDDEN_SIZE, self.HIDDEN_SIZE, ) batch = None if parallel_state.is_pipeline_first_stage(): batch = (torch.ones(global_batch_shape, dtype=dtype).cuda(), ) model = build_model( testing_utils.model_provider_func, # Use DDP only when it's better to have wrap_with_ddp=data_parallel_size > 1, virtual_pipeline_model_parallel_size=virtual_pipeline_model_parallel_size, hidden_size=self.HIDDEN_SIZE, ) offset = pipeline_model_parallel_world_size if virtual_pipeline_model_parallel_size is not None else 0 for idx, model_module in enumerate(model): model_module = model_module.to(dtype) model_module.apply(get_init_weights_func(idx*offset)) _param_groups = _get_params_for_weight_decay_optimization(model) optimizer = torch.optim.Adam(_param_groups, lr=1e-3) pp_utils.update_num_microbatches(0) loss = fwd_bwd_func( testing_utils.fwd_step_func, batch, model, forward_only=forward_only, # `tensor_shape` is the shape of micro batch. tensor_shape=( self.MICRO_BATCH_SIZE, self.HIDDEN_SIZE, self.HIDDEN_SIZE, ), dtype=dtype, async_comm=async_comm, grad_scaler=grad_scaler, deallocate_pipeline_output=deallocate_pipeline_outputs, sync_batch_comm=sync_batch_comm, ) if dtype == get_dtype_for_comparison(): torch.cuda.synchronize() hidden_size = self.HIDDEN_SIZE microbatch_size = self.MICRO_BATCH_SIZE total_layers = pipeline_model_parallel_world_size if virtual_pipeline_model_parallel_size is not None: total_layers *= virtual_pipeline_model_parallel_size target_loss, target_model = get_target_loss_and_model(global_batch_shape, hidden_size, total_layers) for loss_item in loss: x = loss_item['avg'] self.assertEqual(x.item() / microbatch_size, target_loss.item()) if not forward_only: for vm_id, model_module in enumerate(model): params = list(model_module.parameters()) rank = params[0].get_device() offset = pipeline_model_parallel_world_size param_id = rank // data_parallel_size + vm_id * offset target_params = target_model[param_id] self.assertEqual(params[0].cpu(), target_params[0]) self.assertEqual(params[1].cpu(), target_params[1]) self.assertEqual(params[0].grad.cpu() / microbatch_size, target_params[0].grad) self.assertEqual(params[1].grad.cpu() / microbatch_size, target_params[1].grad) if not forward_only: for m in model: for p in m.parameters(): self.assertIsNotNone(p.grad) optimizer.step() optimizer.zero_grad(set_to_none=True) parallel_state.destroy_model_parallel() def test_learning_no_pipelining(self): self._forward_backward_test_impl(False, forward_backward_no_pipelining, 1, None) def test_inference_no_pipelining(self): self._forward_backward_test_impl(True, forward_backward_no_pipelining, 1, None) def test_learning_pipelining_without_interleaving(self, sync_batch_comm: bool = True): self._forward_backward_test_impl( False, forward_backward_pipelining_without_interleaving, None, None, sync_batch_comm=sync_batch_comm, ) def test_inference_pipelining_without_interleaving(self, sync_batch_comm: bool = True): self._forward_backward_test_impl( True, forward_backward_pipelining_without_interleaving, None, None, sync_batch_comm=sync_batch_comm, ) def test_learning_async_pipelining_without_interleaving(self, sync_batch_comm: bool = True): self._forward_backward_test_impl( False, forward_backward_pipelining_without_interleaving, None, None, async_comm=True, sync_batch_comm=sync_batch_comm, ) def test_inference_async_pipelining_without_interleaving(self, sync_batch_comm: bool = True): self._forward_backward_test_impl( True, forward_backward_pipelining_without_interleaving, None, None, async_comm=True, sync_batch_comm=sync_batch_comm, ) # fails on native ucc: times out @unittest.skipUnless(_get_default_world_sizes_model_parallel_world_size()[-1] > 2, "Interleaved schedule requires pipeline_model_parallel_world_size > 2") def test_learning_pipelining_with_interleaving(self, sync_batch_comm: bool = True): self._forward_backward_test_impl( False, _forward_backward_pipelining_with_interleaving, None, virtual_pipeline_model_parallel_size=2, sync_batch_comm=sync_batch_comm, ) # fails on native ucc: times out @unittest.skipUnless(_get_default_world_sizes_model_parallel_world_size()[-1] > 2, "Interleaved schedule requires pipeline_model_parallel_world_size > 2") def test_inference_pipelining_with_interleaving(self, sync_batch_comm: bool = True): self._forward_backward_test_impl( True, _forward_backward_pipelining_with_interleaving, None, virtual_pipeline_model_parallel_size=2, sync_batch_comm=sync_batch_comm, ) # fails on native ucc: times out @unittest.skipUnless(_get_default_world_sizes_model_parallel_world_size()[-1] > 2, "Interleaved schedule requires pipeline_model_parallel_world_size > 2") def test_learning_async_pipelining_with_interleaving(self, sync_batch_comm: bool = True): self._forward_backward_test_impl( False, _forward_backward_pipelining_with_interleaving, None, virtual_pipeline_model_parallel_size=2, async_comm=True, sync_batch_comm=sync_batch_comm, ) # fails on native ucc: times out @unittest.skipUnless(_get_default_world_sizes_model_parallel_world_size()[-1] > 2, "Interleaved schedule requires pipeline_model_parallel_world_size > 2") def test_inference_async_pipelining_with_interleaving(self, sync_batch_comm: bool = True): self._forward_backward_test_impl( True, _forward_backward_pipelining_with_interleaving, None, virtual_pipeline_model_parallel_size=2, async_comm=True, sync_batch_comm=sync_batch_comm, ) class NcclPipelineParallelForwardBackwardTest(NcclDistributedTestBase, PipelineParallelForwardBackwardTestBase): @property def world_size(self) -> int: return min(torch.cuda.device_count(), 8) def _run_hybrid_distributed_backend(self, forward_only: bool) -> None: self._forward_backward_test_impl( forward_only, forward_backward_pipelining_without_interleaving, None, None, default_backend="nccl", p2p_backend="ucc", ) @unittest.skipUnless(HAS_TORCH_UCC_COMPAT_NVIDIA_DRIVER, "Needs driver >= 470.42.01") def _test_hybrid_backends(self, forward_only: bool) -> None: if HAS_UCC: self._run_hybrid_distributed_backend(forward_only) else: with self.assertRaisesRegex( ImportError, re.escape("UCC backend requires pytorch source build with UCC installed and enabled"), ): self._run_hybrid_distributed_backend(forward_only) def test_learning_pipelining_without_interleaving_ucc_for_p2p(self): self._test_hybrid_backends(False) def test_inference_pipelining_without_interleaving_ucc_for_p2p(self): self._test_hybrid_backends(True) @unittest.skipUnless(CAN_SKIP_SYNC_AFTER_BATCH_ISEND_IRECV, "Requires https://github.com/pytorch/pytorch/pull/82450") def test_learning_pipelining_without_interleaving_skyp_sync_after_batch_isend_irecv(self): self.test_learning_pipelining_without_interleaving(sync_batch_comm=False) @unittest.skipUnless(CAN_SKIP_SYNC_AFTER_BATCH_ISEND_IRECV, "Requires https://github.com/pytorch/pytorch/pull/82450") def test_inference_pipelining_without_interleaving_skip_sync_after_batch_isend_irecv(self): self.test_inference_pipelining_without_interleaving(sync_batch_comm=False) @unittest.skipUnless(CAN_SKIP_SYNC_AFTER_BATCH_ISEND_IRECV, "Requires https://github.com/pytorch/pytorch/pull/82450") def test_learning_async_pipelining_without_interleaving_skip_sync_after_batch_isend_irecv(self): self.test_learning_async_pipelining_without_interleaving(sync_batch_comm=False) @unittest.skipUnless(CAN_SKIP_SYNC_AFTER_BATCH_ISEND_IRECV, "Requires https://github.com/pytorch/pytorch/pull/82450") def test_inference_async_pipelining_without_interleaving_skip_sync_after_batch_isend_irecv(self): self.test_inference_async_pipelining_without_interleaving(sync_batch_comm=False) @unittest.skipUnless(CAN_SKIP_SYNC_AFTER_BATCH_ISEND_IRECV, "Requires https://github.com/pytorch/pytorch/pull/82450") def test_learning_pipelining_with_interleaving_skip_sync_after_batch_isend_irecv(self): self.test_learning_pipelining_with_interleaving(sync_batch_comm=False) @unittest.skipUnless(CAN_SKIP_SYNC_AFTER_BATCH_ISEND_IRECV, "Requires https://github.com/pytorch/pytorch/pull/82450") def test_inference_pipelining_with_interleaving_skip_sync_after_batch_isend_irecv(self): self.test_inference_pipelining_with_interleaving(sync_batch_comm=False) @unittest.skipUnless(CAN_SKIP_SYNC_AFTER_BATCH_ISEND_IRECV, "Requires https://github.com/pytorch/pytorch/pull/82450") def test_learning_async_pipelining_with_interleaving_skip_sync_after_batch_isend_irecv(self): self.test_learning_async_pipelining_with_interleaving(sync_batch_comm=False) @unittest.skipUnless(CAN_SKIP_SYNC_AFTER_BATCH_ISEND_IRECV, "Requires https://github.com/pytorch/pytorch/pull/82450") def test_inference_async_pipelining_with_interleaving_skip_sync_after_batch_isend_irecv(self): self.test_inference_async_pipelining_with_interleaving(sync_batch_comm=False) # n.b.(mkozuki): pipeline parallel w/o interleaving with UCX_TLS=tcp,sm fails. class UccPipelineParallelForwardBackwardTest(UccDistributedTestBase, PipelineParallelForwardBackwardTestBase): @property def world_size(self) -> int: return min(torch.cuda.device_count(), 8) deallocate_options = (False,) dtypes = (torch.float32,) # Sanity checking the functionality of `forward_backward_pipelining_without_interleaving` with # `model_type=ModelType.encoder_and_decoder` which is used for pipeline training of transformer # models such as T5. @unittest.skipIf(torch.cuda.device_count() < 4, "Requires >= 4 GPUs") class NcclPipelineParallelWithToyParallelMLP(NcclDistributedTestBase): GLOBAL_BATCH_SIZE: int = 16 MICRO_BATCH_SIZE: int = 2 HIDDEN_SIZE: int = 64 # TODO(mkozuki): Change `DECODER_SEQUENCE_LENGTH` to a value different from `ENCODER_SEQUENCE_LENGTH`. # To test forward_backward_pipelining_without_interleaving with `model_type=ModelType.encoder_and_decoder`, # `decoder_seq_length` is necessary and ideally should be different from `encoder_sequence_length` # but my laziness let me use the same value. # Note that you may have to either update `MyModel` def or define another `MyModel`. # to support different `DECODER_SEQUENCE_LENGTH`. ENCODER_SEQUENCE_LENGTH: int = 32 DECODER_SEQUENCE_LENGTH: int = 32 @property def world_size(self) -> int: return min(torch.cuda.device_count(), 8) # TODO(mkozuki): Set `tensor_model_parallel>1` for encoder_and_decoder as well if there's enough GPUs # in order to let `sequence_parallel_enabled` have an effect on tensor shape logic. def _forward_backward_test_impl( self, *, forward_only: bool, sequence_parallel_enabled: bool, model_type: ModelType, dtype: torch.dtype = torch.float32, ) -> None: # N.B.(mkozuki): It might be better to set `tensor_model_parallel_size` to >1 # if `self.world_size > 5`. Otherwise, `pipeline_model_parallel_split_rank` # can be 1, which can be too far real usecase. tensor_model_parallel_size = 1 + int(self.world_size >= 4) pipeline_model_parallel_world_size = self.world_size // tensor_model_parallel_size if model_type == ModelType.encoder_and_decoder: pipeline_model_parallel_split_rank = pipeline_model_parallel_world_size // 2 else: pipeline_model_parallel_split_rank = None parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_size, pipeline_model_parallel_size_=pipeline_model_parallel_world_size, virtual_pipeline_model_parallel_size_=None, pipeline_model_parallel_split_rank_=pipeline_model_parallel_split_rank, ) testing_utils.set_random_seed(567) pp_utils._reconfigure_microbatch_calculator( rank=parallel_state.get_tensor_model_parallel_rank(), rampup_batch_size=None, global_batch_size=self.GLOBAL_BATCH_SIZE, micro_batch_size=self.MICRO_BATCH_SIZE, data_parallel_size=parallel_state.get_data_parallel_world_size(), ) # TODO(mkozuki): Call `build_model` with `model_type`. model = build_model( testing_utils.mlp_provider_func, wrap_with_ddp=False, virtual_pipeline_model_parallel_size=None, hidden_size=self.HIDDEN_SIZE, sequence_parallel_enabled=sequence_parallel_enabled, ) model = [m.to(dtype=dtype) for m in model] if parallel_state.is_pipeline_first_stage(): batch: Tuple[torch.Tensor] = ( torch.ones( (self.GLOBAL_BATCH_SIZE, self.ENCODER_SEQUENCE_LENGTH, self.HIDDEN_SIZE), dtype=dtype, device="cuda", ), ) else: batch = None forward_backward_pipelining_without_interleaving( forward_step_func=testing_utils.ToyParallelMLPFwdBwdStepFunc( sequence_parallel_enabled=sequence_parallel_enabled, ), batch=batch, model=model, forward_only=forward_only, tensor_shape=( self.ENCODER_SEQUENCE_LENGTH, self.MICRO_BATCH_SIZE, self.HIDDEN_SIZE, ), model_type=model_type, decoder_sequence_length=self.DECODER_SEQUENCE_LENGTH, async_comm=False, grad_scaler=None, deallocate_pipeline_outputs=False, dtype=dtype, sequence_parallel_enabled=sequence_parallel_enabled, ) def test_pipelining_without_interleaving_encoder_and_decoder(self) -> None: self._forward_backward_test_impl(forward_only=False, sequence_parallel_enabled=False, model_type=ModelType.encoder_and_decoder) def test_pipelining_without_interleaving_inferenc_encoder_and_decoder(self) -> None: self._forward_backward_test_impl(forward_only=True, sequence_parallel_enabled=False, model_type=ModelType.encoder_and_decoder) def test_pipelining_without_interleaving_sequence_paralle_encoder_and_decoder(self) -> None: self._forward_backward_test_impl(forward_only=False, sequence_parallel_enabled=True, model_type=ModelType.encoder_and_decoder) def test_pipelining_without_interleaving_inference_sequence_paralle_encoder_and_decoder(self) -> None: self._forward_backward_test_impl(forward_only=True, sequence_parallel_enabled=True, model_type=ModelType.encoder_and_decoder) def test_pipelining_without_interleaving_encoder_or_decoder(self) -> None: self._forward_backward_test_impl(forward_only=False, sequence_parallel_enabled=False, model_type=ModelType.encoder_or_decoder) def test_pipelining_without_interleaving_sequence_parallel_encoder_or_decoder(self) -> None: self._forward_backward_test_impl(forward_only=False, sequence_parallel_enabled=True, model_type=ModelType.encoder_or_decoder) def test_pipelining_without_interleaving_sequence_parallel_encoder_or_decoder_half(self) -> None: self._forward_backward_test_impl(forward_only=False, sequence_parallel_enabled=True, model_type=ModelType.encoder_or_decoder, dtype=torch.half) class NcclPipelineParallelWithCustomSyncContextHandler(NcclDistributedTestBase): GLOBAL_BATCH_SIZE = 32 MICRO_BATCH_SIZE = 1 HIDDEN_SIZE = 1 @property def world_size(self) -> int: return min(torch.cuda.device_count(), 8) @unittest.skipIf(torch.cuda.device_count() < 2 or torch.cuda.device_count() % 2 != 0, "Requires >= 2 GPUs") def test_pipelining_without_interleaving_with_custom_sync_context_handler(self) -> None: # Parallel configuration world_size = torch.cuda.device_count() tensor_model_parallel_world_size = 1 data_parallel_size = 2 if world_size > 2 else 1 pipeline_model_parallel_world_size = world_size // data_parallel_size # Initialize pipeline parallelism parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_world_size, pipeline_model_parallel_size_=pipeline_model_parallel_world_size, ) pp_utils._reconfigure_microbatch_calculator( rank=parallel_state.get_tensor_model_parallel_rank(), rampup_batch_size=None, global_batch_size=self.GLOBAL_BATCH_SIZE, micro_batch_size=self.MICRO_BATCH_SIZE, data_parallel_size=parallel_state.get_data_parallel_world_size(), ) pp_utils.update_num_microbatches(0) # Construct synthetic data dtype = get_dtype_for_comparison() hidden_size = self.HIDDEN_SIZE microbatch_size = self.MICRO_BATCH_SIZE global_batch_shape = ( self.GLOBAL_BATCH_SIZE // parallel_state.get_data_parallel_world_size(), hidden_size, hidden_size, ) batch = None if parallel_state.is_pipeline_first_stage(): batch = (torch.ones(global_batch_shape, dtype=dtype).cuda(), ) # Construct model model = build_model( testing_utils.model_provider_func, wrap_with_ddp=True, hidden_size=hidden_size, )[0] model = model.to(dtype) model.module.apply(get_init_weights_func(0)) # Construct context that destroys all grads on exit has_entered_grad_sync_context = False has_exited_grad_sync_context = False has_called_grad_sync_func = False @contextlib.contextmanager def custom_grad_sync_context(): try: nonlocal has_entered_grad_sync_context has_entered_grad_sync_context = True yield finally: nonlocal has_exited_grad_sync_context has_exited_grad_sync_context = True for param in model.parameters(): param.grad = None def custom_grad_sync_func(): nonlocal has_called_grad_sync_func has_called_grad_sync_func = True # Training step with pipeline parallelism loss = forward_backward_pipelining_without_interleaving( testing_utils.fwd_step_func, batch, model, forward_only=False, tensor_shape=(microbatch_size, hidden_size, hidden_size), dtype=dtype, async_comm=False, grad_scaler=None, deallocate_pipeline_outputs=False, sequence_parallel_enabled=False, custom_sync_context_handler=custom_grad_sync_context, custom_grad_sync_func=custom_grad_sync_func, ) torch.cuda.synchronize() # Check if model has initialized gradients has_any_grads = any(param.grad is not None for param in model.parameters()) has_all_grads = all(param.grad is not None for param in model.parameters()) # Check context behavior self.assertTrue(has_entered_grad_sync_context, 'Has not entered custom sync context') self.assertTrue(has_exited_grad_sync_context, 'Has not exited custom sync context') self.assertEqual( has_any_grads, has_all_grads, 'Expected gradients to all be uninitialized or all be initialized', ) self.assertEqual( has_all_grads, parallel_state.is_pipeline_first_stage(), 'Expected gradients to be initialized only in first pipeline stage', ) # Clean up parallel_state.destroy_model_parallel() @unittest.skipIf(torch.cuda.device_count() < 4 or torch.cuda.device_count() % 2 != 0, "Requires >= 4 GPUs") def test_pipelining_with_interleaving_with_custom_sync_context_handler(self) -> None: # Parallel configuration world_size = torch.cuda.device_count() tensor_model_parallel_world_size = 1 data_parallel_size = 2 if world_size > 4 else 1 pipeline_model_parallel_world_size = world_size // data_parallel_size virtual_pipeline_model_parallel_size = 2 # Initialize pipeline parallelism parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_world_size, pipeline_model_parallel_size_=pipeline_model_parallel_world_size, virtual_pipeline_model_parallel_size_=virtual_pipeline_model_parallel_size, ) pp_utils._reconfigure_microbatch_calculator( rank=parallel_state.get_tensor_model_parallel_rank(), rampup_batch_size=None, global_batch_size=self.GLOBAL_BATCH_SIZE, micro_batch_size=self.MICRO_BATCH_SIZE, data_parallel_size=parallel_state.get_data_parallel_world_size(), ) pp_utils.update_num_microbatches(0) # Construct synthetic data dtype = get_dtype_for_comparison() hidden_size = self.HIDDEN_SIZE microbatch_size = self.MICRO_BATCH_SIZE global_batch_shape = ( self.GLOBAL_BATCH_SIZE // parallel_state.get_data_parallel_world_size(), hidden_size, hidden_size, ) batch = None if parallel_state.is_pipeline_first_stage(): batch = (torch.ones(global_batch_shape, dtype=dtype).cuda(), ) # Construct model model = build_model( testing_utils.model_provider_func, wrap_with_ddp=True, virtual_pipeline_model_parallel_size=virtual_pipeline_model_parallel_size, hidden_size=hidden_size, ) for module in model: module.to(dtype) module.module.apply(get_init_weights_func(0)) # Construct context that keeps track whenever entered/exited grad_sync_context_enter_count = 0 grad_sync_context_exit_count = 0 @contextlib.contextmanager def custom_grad_sync_context(): try: nonlocal grad_sync_context_enter_count grad_sync_context_enter_count += 1 yield finally: nonlocal grad_sync_context_exit_count grad_sync_context_exit_count += 1 for module in model: for param in module.parameters(): param.grad = None # Training step with pipeline parallelism loss = _forward_backward_pipelining_with_interleaving( testing_utils.fwd_step_func, batch, model, forward_only=False, tensor_shape=(microbatch_size, hidden_size, hidden_size), dtype=dtype, async_comm=False, grad_scaler=None, deallocate_pipeline_outputs=False, sequence_parallel_enabled=False, custom_sync_context_handler=custom_grad_sync_context, ) torch.cuda.synchronize() # Check context behavior self.assertTrue( grad_sync_context_enter_count > 0, 'Has not entered custom sync context', ) self.assertEqual( grad_sync_context_enter_count, grad_sync_context_exit_count, 'Has not entered and exited custom sync context ' 'the same number of times', ) self.assertEqual( grad_sync_context_exit_count, virtual_pipeline_model_parallel_size + 1, 'Expected to exit custom sync context once per model chunk ' 'and once at the function end', ) # Clean up parallel_state.destroy_model_parallel() if __name__ == "__main__": common_utils.run_tests()
apex-master
tests/L0/run_transformer/test_pipeline_parallel_fwd_bwd.py
import logging import torch from torch.testing._internal import common_utils from apex.transformer import parallel_state from apex.transformer.tensor_parallel import mappings from apex.transformer.testing.distributed_test_base import NcclDistributedTestBase from apex.transformer.testing.distributed_test_base import UccDistributedTestBase logging.getLogger("torch").setLevel(logging.WARNING) logging.getLogger("apex").setLevel(logging.WARNING) class MappingTestBase: def test_reduce(self): for tensor_model_paralell_world_size in range(1, self.world_size + 1): if self.world_size % tensor_model_paralell_world_size > 0: continue parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_paralell_world_size ) t = torch.full((10, 10, 10, 10), 50, device=f"cuda:{self.rank}") expected = torch.full( (10, 10, 10, 10), 50 * tensor_model_paralell_world_size, device=f"cuda:{self.rank}", ) self.assertTrue( torch.equal(mappings._reduce(t), expected), msg=f"tensor_model_paralell_world_size: {tensor_model_paralell_world_size}", ) parallel_state.destroy_model_parallel() def test_split(self): for tensor_model_paralell_world_size in range(1, self.world_size + 1): if self.world_size % tensor_model_paralell_world_size > 0: continue parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_paralell_world_size ) tensors = [ torch.randn(10, 1) for _ in range(tensor_model_paralell_world_size) ] x = torch.cat(tensors, 1) out = mappings._split_along_last_dim(x) self.assertTrue( torch.equal( out, tensors[parallel_state.get_tensor_model_parallel_rank()] ), msg=f"tensor_model_paralell_world_size: {tensor_model_paralell_world_size}" ) parallel_state.destroy_model_parallel() def test_gather(self): for tensor_model_paralell_world_size in range(1, self.world_size + 1): if self.world_size % tensor_model_paralell_world_size > 0: continue parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_paralell_world_size ) device = f"cuda:{self.rank}" gathered = mappings._gather_along_last_dim( torch.tensor( [parallel_state.get_tensor_model_parallel_rank()], device=device ) ) expected = torch.tensor( [rank for rank in range(tensor_model_paralell_world_size)], device=device, ) self.assertTrue( torch.equal(gathered, expected), msg=f"tensor_model_paralell_world_size: {tensor_model_paralell_world_size}", ) parallel_state.destroy_model_parallel() class NcclMappingTest(MappingTestBase, NcclDistributedTestBase): pass class UccMappingTest(MappingTestBase, UccDistributedTestBase): pass if __name__ == "__main__": common_utils.run_tests()
apex-master
tests/L0/run_transformer/test_mapping.py
import logging import torch from torch.testing._internal import common_utils logging.getLogger("torch").setLevel(logging.WARNING) from apex.transformer import parallel_state from apex.transformer import tensor_parallel from apex.transformer.testing.distributed_test_base import NcclDistributedTestBase from apex.transformer.testing.distributed_test_base import UccDistributedTestBase logging.getLogger("apex").setLevel(logging.WARNING) class TransformerRandomTestBase: def test_set_cuda_rng_state(self): for tensor_model_parallel_world_size in range(1, self.world_size + 1): if self.world_size % tensor_model_parallel_world_size: continue msg = f"tensor_model_parallel_world_size: {tensor_model_parallel_world_size}" parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_world_size ) size, seed = 123, 1234 torch.cuda.manual_seed(seed) tensor = torch.cuda.FloatTensor(size) rng_state = torch.cuda.get_rng_state() rng_state_clone = rng_state.clone() for _ in range(5): torch.randn(size, out=tensor) result_1 = tensor.clone() self.assertEqual(rng_state.sub(rng_state_clone).max(), 0, msg=msg) self.assertGreater( torch.cuda.get_rng_state().sub(rng_state_clone).max(), 0, msg=msg, ) new_rng_state = torch.cuda.get_rng_state() self.assertGreater(new_rng_state.sub(rng_state).max(), 0, msg=msg) tensor_parallel.random._set_cuda_rng_state(rng_state) for _ in range(5): torch.randn(size, out=tensor) tensor_parallel.random._set_cuda_rng_state(rng_state) for _ in range(5): torch.randn(size, out=tensor) result_2 = tensor.clone() self.assertEqual(result_2, result_1, msg=msg) self.assertEqual(rng_state.sub(rng_state_clone).max(), 0, msg=msg) parallel_state.destroy_model_parallel() def test_cuda_rng_tracker(self): for tensor_model_parallel_world_size in range(1, self.world_size + 1): if self.world_size % tensor_model_parallel_world_size: continue msg = f"tensor_model_parallel_world_size: {tensor_model_parallel_world_size}" parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_world_size ) seed_1, seed_2, size = 1234, 4321, [12, 21] tensor = torch.cuda.FloatTensor(size) torch.cuda.manual_seed(seed_1) torch.randn(size, out=tensor) target_11 = tensor.clone() torch.randn(size, out=tensor) target_12 = tensor.clone() torch.cuda.manual_seed(seed_2) torch.randn(size, out=tensor) targt_21 = tensor.clone() torch.randn(size, out=tensor) target_22 = tensor.clone() torch.cuda.manual_seed(seed_1) tensor_parallel.random.get_cuda_rng_tracker().add("test", seed_2) torch.randn(size, out=tensor) result_11 = tensor.clone() with tensor_parallel.random.get_cuda_rng_tracker().fork("test"): torch.randn(size, out=tensor) result_21 = tensor.clone() torch.randn(size, out=tensor) result_12 = tensor.clone() with tensor_parallel.random.get_cuda_rng_tracker().fork("test"): torch.randn(size, out=tensor) result_22 = tensor.clone() self.assertEqual(target_11, result_11, msg=msg) self.assertEqual(target_12, result_12, msg=msg) self.assertEqual(targt_21, result_21, msg=msg) self.assertEqual(target_22, result_22, msg=msg) self.assertNotEqual(result_11, result_21, msg=msg) self.assertNotEqual(result_21, result_22, msg=msg) tensor_parallel.random.get_cuda_rng_tracker().reset() parallel_state.destroy_model_parallel() class NcclTransformerRandomTest(TransformerRandomTestBase, NcclDistributedTestBase): pass class UccTransformerRandomTest(TransformerRandomTestBase, UccDistributedTestBase): pass if __name__ == "__main__": common_utils.run_tests()
apex-master
tests/L0/run_transformer/test_random.py
import unittest import functools as ft import itertools as it from apex import amp from apex.amp import _amp_state import torch from torch import nn import torch.nn.functional as F from torch.nn import Parameter from utils import common_init, HALF, FLOAT,\ ALWAYS_HALF, ALWAYS_FLOAT, MATCH_INPUT class MyModel(torch.nn.Module): def __init__(self, unique): super(MyModel, self).__init__() self.weight0 = Parameter(unique + torch.arange(2, device='cuda', dtype=torch.float32)) self.weight1 = Parameter(1. + unique + torch.arange(2, device='cuda', dtype=torch.float16)) @staticmethod def ops(input, weight0, weight1): return ((input*(weight0.float()))*(weight1.float())).sum() def forward(self, input): return self.ops(input, self.weight0, self.weight1) # Abandon all hope, ye who enter here. # This is hands down the ugliest code I have ever written, but it succeeds in testing # multiple models/optimizers/losses fairly thoroughly. Many of the different test cases # require slightly divergent code in a way that seems near-impossible to genericize into a simple # cross product or nested loops. class TestMultipleModelsOptimizersLosses(unittest.TestCase): def setUp(self): self.x = torch.ones((2), device='cuda', dtype=torch.float32) common_init(self) def tearDown(self): pass def test_2models2losses1optimizer(self): model0 = MyModel(1) model1 = MyModel(2) optimizer = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}, {'params' : model1.parameters(), 'lr' : 0.5}], momentum=0.125) reference_grads = [] for i in range(2): optimizer.zero_grad() loss0 = model0(self.x) loss1 = model1(self.x) loss0.backward() loss1.backward() reference_grads.append([param.grad.data.clone() for param in model0.parameters()] + [param.grad.data.clone() for param in model1.parameters()]) optimizer.step() final_params = [param.data.clone() for param in model0.parameters()] + \ [param.data.clone() for param in model1.parameters()] for opt_level in ("O0", "O1", "O2", "O3"): for how_to_zero in ("none", "model", "optimizer"): for use_multiple_loss_scalers in (True, False): if opt_level == "O1" or opt_level == "O2": inject_inf_iters = (-1, 0, 1) else: inject_inf_iters = (-1,) for inject_inf in inject_inf_iters: if inject_inf >= 0: inject_inf_locs = ("fp16", "fp32") which_backwards = (0, 1) else: inject_inf_locs = ("fdsa",) which_backwards = (None,) for inject_inf_loc in inject_inf_locs: for which_backward in which_backwards: if use_multiple_loss_scalers: num_losses = 2 loss_ids = [0, 1] else: num_losses = 1 loss_ids = [0, 0] if inject_inf >= 0: iters = 3 else: iters = 2 model0 = MyModel(1) model1 = MyModel(2) models = [model0, model1] optimizer = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}, {'params' : model1.parameters(), 'lr' : 0.5}], momentum=0.125) _amp_state.allow_incoming_model_not_fp32 = True [model0, model1], optimizer = amp.initialize( [model0, model1], optimizer, opt_level=opt_level, verbosity=0, cast_model_type=False, num_losses=num_losses) _amp_state.allow_incoming_model_not_fp32 = False _amp_state.loss_scalers[0]._loss_scale = 4.0 if use_multiple_loss_scalers: _amp_state.loss_scalers[1]._loss_scale = 16.0 unskipped = 0 for i in range(iters): if how_to_zero == "none": for model in models: for param in model.parameters(): param.grad = None elif how_to_zero == "model": for model in models: model.zero_grad() else: optimizer.zero_grad() loss0 = model0(self.x) loss1 = model1(self.x) with amp.scale_loss(loss0, optimizer, loss_id=loss_ids[0]) as scaled_loss: scaled_loss.backward() if i == inject_inf and which_backward == 0: if inject_inf_loc == "fp32": model0.weight0.grad[0] = float('inf') elif inject_inf_loc == "fp16": model0.weight1.grad[0] = float('inf') with amp.scale_loss(loss1, optimizer, loss_id=loss_ids[1]) as scaled_loss: scaled_loss.backward() if i == inject_inf and which_backward == 1: if inject_inf_loc == "fp32": model1.weight0.grad[0] = float('inf') elif inject_inf_loc == "fp16": model1.weight1.grad[0] = float('inf') if i != inject_inf: for param, reference_grad in zip(amp.master_params(optimizer), reference_grads[unskipped]): self.assertTrue(torch.allclose(param.grad.float(), reference_grad.float())) unskipped += 1 optimizer.step() model_params = [p for p in model0.parameters()] + [p for p in model1.parameters()] for model, master, reference in zip( model_params, amp.master_params(optimizer), final_params): self.assertTrue(torch.allclose(model, reference)) self.assertTrue(torch.allclose(model, master.to(model.dtype))) if opt_level == "O1": _amp_state.handle._deactivate() def test_3models2losses1optimizer(self): model0 = MyModel(1) model1 = MyModel(2) model2 = MyModel(3) optimizer = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}, {'params' : model1.parameters(), 'lr' : 0.5}, {'params' : model2.parameters(), 'lr' : 0.125}], momentum=0.125) reference_grads = [] for i in range(2): optimizer.zero_grad() loss0 = model0(self.x) + model2(self.x) loss1 = model1(self.x) + model2(self.x) loss0.backward() loss1.backward() reference_grads.append([param.grad.data.clone() for param in model0.parameters()] + [param.grad.data.clone() for param in model1.parameters()] + [param.grad.data.clone() for param in model2.parameters()]) optimizer.step() final_params = [param.data.clone() for param in model0.parameters()] + \ [param.data.clone() for param in model1.parameters()] + \ [param.data.clone() for param in model2.parameters()] for opt_level in ("O0", "O1", "O2", "O3"): for how_to_zero in ("none", "model", "optimizer"): for use_multiple_loss_scalers in (True, False): if opt_level == "O1" or opt_level == "O2": inject_inf_iters = (-1, 0, 1) else: inject_inf_iters = (-1,) for inject_inf in inject_inf_iters: if inject_inf >= 0: inject_inf_locs = ("fp16", "fp32") which_backwards = (0, 1) else: inject_inf_locs = ("fdsa",) which_backwards = (None,) for inject_inf_loc in inject_inf_locs: for which_backward in which_backwards: if use_multiple_loss_scalers: num_losses = 2 loss_ids = [0, 1] else: num_losses = 1 loss_ids = [0, 0] if inject_inf >= 0: iters = 3 if which_backward == 0: which_models = (0, 2) elif which_backward == 1: which_models = (1, 2) else: iters = 2 which_models = (None,) for which_model in which_models: model0 = MyModel(1) model1 = MyModel(2) model2 = MyModel(3) models = [model0, model1, model2] optimizer = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}, {'params' : model1.parameters(), 'lr' : 0.5}, {'params' : model2.parameters(), 'lr' : 0.125}], momentum=0.125) _amp_state.allow_incoming_model_not_fp32 = True [model0, model1, model2], optimizer = amp.initialize( [model0, model1, model2], optimizer, opt_level=opt_level, verbosity=0, cast_model_type=False, num_losses=num_losses) _amp_state.allow_incoming_model_not_fp32 = False _amp_state.loss_scalers[0]._loss_scale = 4.0 if use_multiple_loss_scalers: _amp_state.loss_scalers[1]._loss_scale = 16.0 unskipped = 0 for i in range(iters): if how_to_zero == "none": for model in models: for param in model.parameters(): param.grad = None elif how_to_zero == "model": for model in models: model.zero_grad() else: optimizer.zero_grad() # print("opt_level {} i {} inject_inf {} which_backward {} inject_inf_loc {} which_model {} use_multiple_loss_scalers {}".format(opt_level, i, inject_inf, which_backward, inject_inf_loc, which_model, use_multiple_loss_scalers)) loss0 = model0(self.x) + model2(self.x) loss1 = model1(self.x) + model2(self.x) with amp.scale_loss(loss0, optimizer, loss_id=loss_ids[0]) as scaled_loss: scaled_loss.backward() if i == inject_inf and which_backward == 0: if which_model == 0: inj_model = model0 elif which_model == 2: inj_model = model2 else: raise RuntimeError(which_model + " invalid for loss 0") if inject_inf_loc == "fp32": inj_model.weight0.grad[0] = float('inf') elif inject_inf_loc == "fp16": inj_model.weight1.grad[0] = float('inf') with amp.scale_loss(loss1, optimizer, loss_id=loss_ids[1]) as scaled_loss: scaled_loss.backward() if i == inject_inf and which_backward == 1: if which_model == 1: inj_model = model1 elif which_model == 2: inj_model = model2 else: raise RuntimeError(which_model + " invalid for loss 1 ") if inject_inf_loc == "fp32": inj_model.weight0.grad[0] = float('inf') elif inject_inf_loc == "fp16": inj_model.weight1.grad[0] = float('inf') if i != inject_inf: for param, reference_grad in zip(amp.master_params(optimizer), reference_grads[unskipped]): self.assertTrue(torch.allclose(param.grad.float(), reference_grad.float())) unskipped += 1 optimizer.step() model_params = [p for p in model0.parameters()] + \ [p for p in model1.parameters()] + \ [p for p in model2.parameters()] for model, master, reference in zip( model_params, amp.master_params(optimizer), final_params): self.assertTrue(torch.allclose(model, reference)) self.assertTrue(torch.allclose(model, master.to(model.dtype))) if opt_level == "O1": _amp_state.handle._deactivate() def test_2models2losses2optimizers(self): model0 = MyModel(1) model1 = MyModel(2) optimizer0 = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}], momentum=0.125) optimizer1 = torch.optim.SGD([{'params' : model1.parameters(), 'lr' : 0.5}], momentum=0.25) # Don't do it like this: reference_grads = [[]]*5 # because then it creates a list of 5 references to the same "[]" and appending # to any of them effectively makes you append to all of them, which multiplies # the resulting size of reference_grads by 5x and needless to say makes the test fail. reference_grads = [[], [], [], [], []] final_params = [None, None, None, None, None] for i in range(2): optimizer0.zero_grad() optimizer1.zero_grad() loss0 = model0(self.x) loss1 = model1(self.x) loss0.backward() loss1.backward() reference_grads[0].append([param.grad.data.clone() for param in model0.parameters()] + [param.grad.data.clone() for param in model1.parameters()]) optimizer0.step() optimizer1.step() final_params[0] = [param.data.clone() for param in model0.parameters()] + \ [param.data.clone() for param in model1.parameters()] def what_got_skipped(which_iter, which_backward): if which_iter == 0 and which_backward == 0: return 1 if which_iter == 0 and which_backward == 1: return 2 if which_iter == 1 and which_backward == 0: return 3 if which_iter == 1 and which_backward == 1: return 4 return 0 for which_iter in (0,1): for which_backward in (0,1): model0 = MyModel(1) model1 = MyModel(2) optimizer0 = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}], momentum=0.125) optimizer1 = torch.optim.SGD([{'params' : model1.parameters(), 'lr' : 0.5}], momentum=0.25) for i in range(3): optimizer0.zero_grad() optimizer1.zero_grad() loss0 = model0(self.x) loss1 = model1(self.x) loss0.backward() loss1.backward() if i != which_iter: reference_grads[what_got_skipped(which_iter, which_backward)].append( [param.grad.data.clone() for param in model0.parameters()] + [param.grad.data.clone() for param in model1.parameters()]) if i == which_iter: if which_backward == 0: optimizer1.step() else: optimizer0.step() else: optimizer0.step() optimizer1.step() final_params[what_got_skipped(which_iter, which_backward)] = \ [param.data.clone() for param in model0.parameters()] + \ [param.data.clone() for param in model1.parameters()] for opt_level in ("O0", "O1", "O2", "O3"): for how_to_zero in ("none", "model", "optimizer"): for use_multiple_loss_scalers in (True, False): if opt_level == "O1" or opt_level == "O2": inject_inf_iters = (-1, 0, 1) else: inject_inf_iters = (-1,) for inject_inf in inject_inf_iters: if inject_inf >= 0: inject_inf_locs = ("fp16", "fp32") which_backwards = (0, 1) else: inject_inf_locs = ("fdsa",) which_backwards = (None,) for inject_inf_loc in inject_inf_locs: for which_backward in which_backwards: if use_multiple_loss_scalers: num_losses = 2 loss_ids = [0, 1] else: num_losses = 1 loss_ids = [0, 0] if inject_inf >= 0: iters = 3 else: iters = 2 model0 = MyModel(1) model1 = MyModel(2) models = [model0, model1] optimizer0 = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}], momentum=0.125) optimizer1 = torch.optim.SGD([{'params' : model1.parameters(), 'lr' : 0.5}], momentum=0.25) _amp_state.allow_incoming_model_not_fp32 = True [model0, model1], [optimizer0, optimizer1] = amp.initialize( [model0, model1], [optimizer0, optimizer1], opt_level=opt_level, verbosity=0, cast_model_type=False, num_losses=num_losses) _amp_state.allow_incoming_model_not_fp32 = False _amp_state.loss_scalers[0]._loss_scale = 4.0 if use_multiple_loss_scalers: _amp_state.loss_scalers[1]._loss_scale = 16.0 unskipped = 0 for i in range(iters): if how_to_zero == "none": for model in models: for param in model.parameters(): param.grad = None elif how_to_zero == "model": for model in models: model.zero_grad() else: optimizer0.zero_grad() optimizer1.zero_grad() loss0 = model0(self.x) loss1 = model1(self.x) with amp.scale_loss(loss0, optimizer0, loss_id=loss_ids[0]) as scaled_loss: scaled_loss.backward() if i == inject_inf and which_backward == 0: if inject_inf_loc == "fp32": model0.weight0.grad[0] = float('inf') elif inject_inf_loc == "fp16": model0.weight1.grad[0] = float('inf') with amp.scale_loss(loss1, optimizer1, loss_id=loss_ids[1]) as scaled_loss: scaled_loss.backward() if i == inject_inf and which_backward == 1: if inject_inf_loc == "fp32": model1.weight0.grad[0] = float('inf') elif inject_inf_loc == "fp16": model1.weight1.grad[0] = float('inf') # print("opt_level {} i {} inject_inf {} which_backward {} inject_inf_loc {} use_multiple_loss_scalers {}".format(opt_level, i, inject_inf, which_backward, inject_inf_loc, use_multiple_loss_scalers)) if i != inject_inf: master_params = list(amp.master_params(optimizer0)) + \ list(amp.master_params(optimizer1)) for param, reference_grad in zip(master_params, reference_grads[what_got_skipped(inject_inf, which_backward)][unskipped]): self.assertTrue(torch.allclose(param.grad.float(), reference_grad.float())) unskipped += 1 optimizer0.step() optimizer1.step() model_params = [p for p in model0.parameters()] + [p for p in model1.parameters()] master_params = [p for p in amp.master_params(optimizer0)] + \ [p for p in amp.master_params(optimizer1)] for model, master, reference in zip( model_params, master_params, final_params[what_got_skipped(inject_inf, which_backward)]): self.assertTrue(torch.allclose(model, reference)) self.assertTrue(torch.allclose(model, master.to(model.dtype))) if opt_level == "O1": _amp_state.handle._deactivate() def test_3models2losses2optimizers(self): model0 = MyModel(1) model1 = MyModel(2) model2 = MyModel(3) optimizer0 = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}, {'params' : model1.parameters(), 'lr' : 1.0}], momentum=0.5) optimizer1 = torch.optim.SGD([{'params' : model2.parameters(), 'lr' : 0.5}], momentum=0.25) # Again, can't do this: reference_grads = [[]]*9 reference_grads = [[], [], [], [], [], [], [], [], []] final_params = [None, None, None, None, None, None, None, None, None] for i in range(2): optimizer0.zero_grad() optimizer1.zero_grad() loss0 = model0(self.x) + model1(self.x) loss1 = model2(self.x) + model1(self.x) loss0.backward() loss1.backward() reference_grads[0].append([param.grad.data.clone() for param in model0.parameters()] + [param.grad.data.clone() for param in model1.parameters()]) optimizer0.step() optimizer1.step() final_params[0] = \ [param.data.clone() for param in model0.parameters()] + \ [param.data.clone() for param in model1.parameters()] + \ [param.data.clone() for param in model2.parameters()] def what_got_skipped(which_iter, which_backward, which_model): if which_iter == 0: if which_backward == 0: if which_model == 0: return 1 if which_model == 1: return 2 if which_backward == 1: if which_model == 2: return 3 if which_model == 1: return 4 if which_iter == 1: if which_backward == 0: if which_model == 0: return 5 if which_model == 1: return 6 if which_backward == 1: if which_model == 2: return 7 if which_model == 1: return 8 return 0 for which_iter in (0,1): for which_backward in (0,1): if which_backward == 0: which_models = (0,1) if which_backward == 1: which_models = (2,1) for which_model in which_models: model0 = MyModel(1) model1 = MyModel(2) model2 = MyModel(3) optimizer0 = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}, {'params' : model1.parameters(), 'lr' : 1.0}], momentum=0.5) optimizer1 = torch.optim.SGD([{'params' : model2.parameters(), 'lr' : 0.5}], momentum=0.25) for i in range(3): optimizer0.zero_grad() optimizer1.zero_grad() loss0 = model0(self.x) + model1(self.x) loss1 = model2(self.x) + model1(self.x) loss0.backward() loss1.backward() if i != which_iter: reference_grads[what_got_skipped(which_iter, which_backward, which_model)].append( [param.grad.data.clone() for param in model0.parameters()] + [param.grad.data.clone() for param in model1.parameters()]) if i == which_iter: if which_backward == 0: # if which_model == 0: optimizer1.step() # if which_model == 1: # optimizer1.step() if which_backward == 1: # if which_model == 2: # optimizer0.step() # if which_model == 1: continue else: optimizer0.step() optimizer1.step() final_params[what_got_skipped(which_iter, which_backward, which_model)] = \ [param.data.clone() for param in model0.parameters()] + \ [param.data.clone() for param in model1.parameters()] + \ [param.data.clone() for param in model2.parameters()] for opt_level in ("O0", "O1", "O2", "O3"): for how_to_zero in ("none", "model", "optimizer"): for use_multiple_loss_scalers in (True, False): if opt_level == "O1" or opt_level == "O2": inject_inf_iters = (-1, 0, 1) else: inject_inf_iters = (-1,) for inject_inf in inject_inf_iters: if inject_inf >= 0: inject_inf_locs = ("fp16", "fp32") which_backwards = (0, 1) else: inject_inf_locs = ("fdsa",) which_backwards = (None,) for inject_inf_loc in inject_inf_locs: for which_backward in which_backwards: if use_multiple_loss_scalers: num_losses = 2 loss_ids = [0, 1] else: num_losses = 1 loss_ids = [0, 0] if inject_inf >= 0: iters = 3 if which_backward == 0: which_models = (0, 1) elif which_backward == 1: which_models = (2, 1) else: iters = 2 which_models = (None,) for which_model in which_models: model0 = MyModel(1) model1 = MyModel(2) model2 = MyModel(3) models = [model0, model1, model2] optimizer0 = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}, {'params' : model1.parameters(), 'lr' : 1.0}], momentum=0.5) optimizer1 = torch.optim.SGD([{'params' : model2.parameters(), 'lr' : 0.5}], momentum=0.25) _amp_state.allow_incoming_model_not_fp32 = True [model0, model1, model2], [optimizer0, optimizer1] = amp.initialize( [model0, model1, model2], [optimizer0, optimizer1], opt_level=opt_level, verbosity=0, cast_model_type=False, num_losses=num_losses) _amp_state.allow_incoming_model_not_fp32 = False _amp_state.loss_scalers[0]._loss_scale = 4.0 if use_multiple_loss_scalers: _amp_state.loss_scalers[1]._loss_scale = 16.0 unskipped = 0 for i in range(iters): if how_to_zero == "none": for model in models: for param in model.parameters(): param.grad = None elif how_to_zero == "model": for model in models: model.zero_grad() else: optimizer0.zero_grad() optimizer1.zero_grad() loss0 = model0(self.x) + model1(self.x) loss1 = model2(self.x) + model1(self.x) with amp.scale_loss(loss0, optimizer0, loss_id=loss_ids[0]) as scaled_loss: scaled_loss.backward() if i == inject_inf and which_backward == 0: if which_model == 0: inj_model = model0 elif which_model == 1: inj_model = model1 else: raise RuntimeError(which_model + " invalid for loss 0") if inject_inf_loc == "fp32": inj_model.weight0.grad[0] = float('inf') elif inject_inf_loc == "fp16": inj_model.weight1.grad[0] = float('inf') with amp.scale_loss(loss1, [optimizer0, optimizer1], loss_id=loss_ids[1]) as scaled_loss: scaled_loss.backward() if i == inject_inf and which_backward == 1: if which_model == 2: inj_model = model2 elif which_model == 1: inj_model = model1 else: raise RuntimeError(which_model + " invalid for loss 1 ") if inject_inf_loc == "fp32": inj_model.weight0.grad[0] = float('inf') elif inject_inf_loc == "fp16": inj_model.weight1.grad[0] = float('inf') if i != inject_inf: master_params = list(amp.master_params(optimizer0)) + \ list(amp.master_params(optimizer1)) for param, reference_grad in zip(master_params, reference_grads[what_got_skipped(inject_inf, which_backward, which_model)][unskipped]): self.assertTrue(torch.allclose(param.grad.float(), reference_grad.float())) unskipped += 1 optimizer0.step() optimizer1.step() model_params = [p for p in model0.parameters()] + \ [p for p in model1.parameters()] + \ [p for p in model2.parameters()] master_params = [p for p in amp.master_params(optimizer0)] + \ [p for p in amp.master_params(optimizer1)] # print("opt_level {} i {} inject_inf {} which_backward {} inject_inf_loc {} use_multiple_loss_scalers {} which_model {}".format(opt_level, i, inject_inf, which_backward, inject_inf_loc, use_multiple_loss_scalers, which_model)) for model, master, reference in zip( model_params, master_params, final_params[what_got_skipped(inject_inf, which_backward, which_model)]): self.assertTrue(torch.allclose(model, reference)) self.assertTrue(torch.allclose(model, master.to(model.dtype))) if opt_level == "O1": _amp_state.handle._deactivate() if __name__ == '__main__': unittest.main()
apex-master
tests/L0/run_amp/test_multiple_models_optimizers_losses.py
import unittest import functools as ft import itertools as it from apex import amp import torch from torch import nn import torch.nn.functional as F from utils import common_init, HALF, FLOAT,\ ALWAYS_HALF, ALWAYS_FLOAT, MATCH_INPUT try: import amp_C from amp_C import multi_tensor_l2norm from apex.multi_tensor_apply import MultiTensorApply disabled = False except ImportError as err: print("amp_C fused kernels unavailable, disabling TestMultiTensorApply. ImportError was ", err) disabled = True class TestMultiTensorL2Norm(unittest.TestCase): def setUp(self): common_init(self) self.val = 4.0 self.overflow_buf = torch.cuda.IntTensor(1).zero_() def tearDown(self): pass # The tensor creation here is written for convenience, not speed. def l2norm(self, sizea, sizeb, applier, repeat_tensors, in_type, per_tensor): self.overflow_buf.zero_() a = torch.cuda.FloatTensor(sizea).fill_(self.val) b = torch.cuda.FloatTensor(sizeb).fill_(self.val) in_list = [] for i in range(repeat_tensors): in_list += [a.clone().to(in_type), b.clone().to(in_type)] if per_tensor: norm, norm_per_tensor = applier(multi_tensor_l2norm, self.overflow_buf, [in_list], True) normab = torch.cat((a.norm().view(1), b.norm().view(1))) norm_per_tensor = norm_per_tensor.view(-1, 2) else: norm, _ = applier(multi_tensor_l2norm, self.overflow_buf, [in_list], True) reference = torch.cuda.FloatTensor((sizea + sizeb)*repeat_tensors).fill_(self.val).norm() self.assertTrue(torch.allclose(norm, reference)) if per_tensor: self.assertTrue(torch.allclose(norm_per_tensor, normab)) self.assertTrue(self.overflow_buf.item() == 0) @unittest.skipIf(disabled, "amp_C is unavailable") def test_fuzz(self): input_size_pairs = ( (7777*77, 555*555), (777, 555), (555, 2048*32+1), (2048*32+1, 555), (555, 2048*32), (2048*32, 555), (33333, 555), (555, 33333)) appliers = ( MultiTensorApply(2048*32), MultiTensorApply(333), MultiTensorApply(33333)) repeat_tensors = ( 1, 55) for sizea, sizeb in input_size_pairs: for applier in appliers: for repeat in repeat_tensors: for in_type in (torch.float32, torch.float16): for per_tensor in (False, True): self.l2norm(sizea, sizeb, applier, repeat, in_type, per_tensor) if __name__ == '__main__': unittest.main()
apex-master
tests/L0/run_amp/test_multi_tensor_l2norm.py
import unittest import functools as ft import itertools as it from apex import amp from apex.amp import _amp_state import torch from torch import nn import torch.nn.functional as F from torch.nn import Parameter from utils import common_init, HALF, FLOAT,\ ALWAYS_HALF, ALWAYS_FLOAT, MATCH_INPUT try: import amp_C disabled = False from apex.optimizers import FusedSGD as FusedSGD except ImportError as err: print("amp_C fused kernels unavailable, disabling TestMultiTensorApply. ImportError was ", err) disabled = True class MyModel(torch.nn.Module): def __init__(self, unique): super(MyModel, self).__init__() self.weight0 = Parameter(unique + torch.arange(2, device='cuda', dtype=torch.float32)) self.weight1 = Parameter(1. + unique + torch.arange(2, device='cuda', dtype=torch.float16)) @staticmethod def ops(input, weight0, weight1): return ((input*(weight0.float()))*(weight1.float())).sum() def forward(self, input): return self.ops(input, self.weight0, self.weight1) # Abandon all hope, ye who enter here. # This is hands down the ugliest code I have ever written, but it succeeds in testing # multiple models/optimizers/losses fairly thoroughly. Many of the different test cases # require slightly divergent code in a way that seems near-impossible to genericize into a simple # cross product or nested loops. class TestMultipleModelsOptimizersLosses(unittest.TestCase): def setUp(self): self.x = torch.ones((2), device='cuda', dtype=torch.float32) common_init(self) def tearDown(self): pass @unittest.skipIf(disabled, "amp_C is unavailable") def test_2models2losses1optimizer(self): model0 = MyModel(1) model1 = MyModel(2) optimizer = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}, {'params' : model1.parameters(), 'lr' : 0.5}], momentum=0.125) reference_grads = [] for i in range(2): optimizer.zero_grad() loss0 = model0(self.x) loss1 = model1(self.x) loss0.backward() loss1.backward() reference_grads.append([param.grad.data.clone() for param in model0.parameters()] + [param.grad.data.clone() for param in model1.parameters()]) optimizer.step() final_params = [param.data.clone() for param in model0.parameters()] + \ [param.data.clone() for param in model1.parameters()] for materialize_master_grads in (False, True): for opt_level in ("O0", "O1", "O2", "O3"): for how_to_zero in ("none", "model", "optimizer"): for use_multiple_loss_scalers in (False, True): if opt_level == "O1" or opt_level == "O2": inject_inf_iters = (-1, 0, 1) else: inject_inf_iters = (-1,) for inject_inf in inject_inf_iters: if inject_inf >= 0: inject_inf_locs = ("fp16", "fp32") which_backwards = (0, 1) else: inject_inf_locs = ("fdsa",) which_backwards = (None,) for inject_inf_loc in inject_inf_locs: for which_backward in which_backwards: if use_multiple_loss_scalers: num_losses = 2 loss_ids = [0, 1] else: num_losses = 1 loss_ids = [0, 0] if inject_inf >= 0: iters = 3 else: iters = 2 model0 = MyModel(1) model1 = MyModel(2) models = [model0, model1] optimizer = FusedSGD([{'params' : model0.parameters(), 'lr' : 0.25}, {'params' : model1.parameters(), 'lr' : 0.5}], momentum=0.125, materialize_master_grads=materialize_master_grads) _amp_state.allow_incoming_model_not_fp32 = True [model0, model1], optimizer = amp.initialize( [model0, model1], optimizer, opt_level=opt_level, verbosity=0, cast_model_type=False, num_losses=num_losses) _amp_state.allow_incoming_model_not_fp32 = False _amp_state.loss_scalers[0]._loss_scale = 4.0 if use_multiple_loss_scalers: _amp_state.loss_scalers[1]._loss_scale = 16.0 unskipped = 0 for i in range(iters): if how_to_zero == "none": for model in models: for param in model.parameters(): param.grad = None elif how_to_zero == "model": for model in models: model.zero_grad() else: optimizer.zero_grad() loss0 = model0(self.x) loss1 = model1(self.x) with amp.scale_loss(loss0, optimizer, loss_id=loss_ids[0]) as scaled_loss: scaled_loss.backward() if i == inject_inf and which_backward == 0: if inject_inf_loc == "fp32": model0.weight0.grad[0] = float('inf') elif inject_inf_loc == "fp16": model0.weight1.grad[0] = float('inf') with amp.scale_loss(loss1, optimizer, loss_id=loss_ids[1]) as scaled_loss: scaled_loss.backward() if i == inject_inf and which_backward == 1: if inject_inf_loc == "fp32": model1.weight0.grad[0] = float('inf') elif inject_inf_loc == "fp16": model1.weight1.grad[0] = float('inf') if i != inject_inf: master_params = amp.master_params(optimizer) for param, reference_grad in zip(master_params, reference_grads[unskipped]): if opt_level == "O2" and not materialize_master_grads: continue else: self.assertTrue(torch.allclose(param.grad.float(), reference_grad.float()), "opt_level {} i {} inject_inf {} which_backward {} inject_inf_loc {} use_multiple_loss_scalers {}".format(opt_level, i, inject_inf, which_backward, inject_inf_loc, use_multiple_loss_scalers)) unskipped += 1 optimizer.step() model_params = [p for p in model0.parameters()] + [p for p in model1.parameters()] for model, master, reference in zip( model_params, amp.master_params(optimizer), final_params): self.assertTrue(torch.allclose(model, reference)) self.assertTrue(torch.allclose(model, master.to(model.dtype))) if opt_level == "O1": _amp_state.handle._deactivate() @unittest.skipIf(disabled, "amp_C is unavailable") def test_3models2losses1optimizer(self): model0 = MyModel(1) model1 = MyModel(2) model2 = MyModel(3) optimizer = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}, {'params' : model1.parameters(), 'lr' : 0.5}, {'params' : model2.parameters(), 'lr' : 0.125}], momentum=0.125) reference_grads = [] for i in range(2): optimizer.zero_grad() loss0 = model0(self.x) + model2(self.x) loss1 = model1(self.x) + model2(self.x) loss0.backward() loss1.backward() reference_grads.append([param.grad.data.clone() for param in model0.parameters()] + [param.grad.data.clone() for param in model1.parameters()] + [param.grad.data.clone() for param in model2.parameters()]) optimizer.step() final_params = [param.data.clone() for param in model0.parameters()] + \ [param.data.clone() for param in model1.parameters()] + \ [param.data.clone() for param in model2.parameters()] for materialize_master_grads in (False, True): for opt_level in ("O0", "O1", "O2", "O3"): for how_to_zero in ("none", "model", "optimizer"): for use_multiple_loss_scalers in (False, True): if opt_level == "O1" or opt_level == "O2": inject_inf_iters = (-1, 0, 1) else: inject_inf_iters = (-1,) for inject_inf in inject_inf_iters: if inject_inf >= 0: inject_inf_locs = ("fp16", "fp32") which_backwards = (0, 1) else: inject_inf_locs = ("fdsa",) which_backwards = (None,) for inject_inf_loc in inject_inf_locs: for which_backward in which_backwards: if use_multiple_loss_scalers: num_losses = 2 loss_ids = [0, 1] else: num_losses = 1 loss_ids = [0, 0] if inject_inf >= 0: iters = 3 if which_backward == 0: which_models = (0, 2) elif which_backward == 1: which_models = (1, 2) else: iters = 2 which_models = (None,) for which_model in which_models: model0 = MyModel(1) model1 = MyModel(2) model2 = MyModel(3) models = [model0, model1, model2] optimizer = FusedSGD([{'params' : model0.parameters(), 'lr' : 0.25}, {'params' : model1.parameters(), 'lr' : 0.5}, {'params' : model2.parameters(), 'lr' : 0.125}], momentum=0.125, materialize_master_grads=materialize_master_grads) _amp_state.allow_incoming_model_not_fp32 = True [model0, model1, model2], optimizer = amp.initialize( [model0, model1, model2], optimizer, opt_level=opt_level, verbosity=0, cast_model_type=False, num_losses=num_losses) _amp_state.allow_incoming_model_not_fp32 = False _amp_state.loss_scalers[0]._loss_scale = 4.0 if use_multiple_loss_scalers: _amp_state.loss_scalers[1]._loss_scale = 16.0 unskipped = 0 for i in range(iters): if how_to_zero == "none": for model in models: for param in model.parameters(): param.grad = None elif how_to_zero == "model": for model in models: model.zero_grad() else: optimizer.zero_grad() loss0 = model0(self.x) + model2(self.x) loss1 = model1(self.x) + model2(self.x) with amp.scale_loss(loss0, optimizer, loss_id=loss_ids[0]) as scaled_loss: scaled_loss.backward() if i == inject_inf and which_backward == 0: if which_model == 0: inj_model = model0 elif which_model == 2: inj_model = model2 else: raise RuntimeError(which_model + " invalid for loss 0") if inject_inf_loc == "fp32": inj_model.weight0.grad[0] = float('inf') elif inject_inf_loc == "fp16": inj_model.weight1.grad[0] = float('inf') with amp.scale_loss(loss1, optimizer, loss_id=loss_ids[1]) as scaled_loss: scaled_loss.backward() if i == inject_inf and which_backward == 1: if which_model == 1: inj_model = model1 elif which_model == 2: inj_model = model2 else: raise RuntimeError(which_model + " invalid for loss 1 ") if inject_inf_loc == "fp32": inj_model.weight0.grad[0] = float('inf') elif inject_inf_loc == "fp16": inj_model.weight1.grad[0] = float('inf') if i != inject_inf: master_params = amp.master_params(optimizer) for param, reference_grad in zip(master_params, reference_grads[unskipped]): if opt_level == "O2" and not materialize_master_grads: continue else: self.assertTrue(torch.allclose(param.grad.float(), reference_grad.float()), "opt_level {} i {} inject_inf {} which_backward {} inject_inf_loc {} which_model {} use_multiple_loss_scalers {}".format(opt_level, i, inject_inf, which_backward, inject_inf_loc, which_model, use_multiple_loss_scalers)) unskipped += 1 optimizer.step() model_params = [p for p in model0.parameters()] + \ [p for p in model1.parameters()] + \ [p for p in model2.parameters()] for model, master, reference in zip( model_params, amp.master_params(optimizer), final_params): self.assertTrue(torch.allclose(model, reference)) self.assertTrue(torch.allclose(model, master.to(model.dtype))) if opt_level == "O1": _amp_state.handle._deactivate() @unittest.skipIf(disabled, "amp_C is unavailable") def test_2models2losses2optimizers(self): model0 = MyModel(1) model1 = MyModel(2) optimizer0 = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}], momentum=0.125) optimizer1 = torch.optim.SGD([{'params' : model1.parameters(), 'lr' : 0.5}], momentum=0.25) # Don't do it like this: reference_grads = [[]]*5 # because then it creates a list of 5 references to the same "[]" and appending # to any of them effectively makes you append to all of them, which multiplies # the resulting size of reference_grads by 5x and needless to say makes the test fail. reference_grads = [[], [], [], [], []] final_params = [None, None, None, None, None] for i in range(2): optimizer0.zero_grad() optimizer1.zero_grad() loss0 = model0(self.x) loss1 = model1(self.x) loss0.backward() loss1.backward() reference_grads[0].append([param.grad.data.clone() for param in model0.parameters()] + [param.grad.data.clone() for param in model1.parameters()]) optimizer0.step() optimizer1.step() final_params[0] = [param.data.clone() for param in model0.parameters()] + \ [param.data.clone() for param in model1.parameters()] def what_got_skipped(which_iter, which_backward): if which_iter == 0 and which_backward == 0: return 1 if which_iter == 0 and which_backward == 1: return 2 if which_iter == 1 and which_backward == 0: return 3 if which_iter == 1 and which_backward == 1: return 4 return 0 for which_iter in (0,1): for which_backward in (0,1): model0 = MyModel(1) model1 = MyModel(2) optimizer0 = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}], momentum=0.125) optimizer1 = torch.optim.SGD([{'params' : model1.parameters(), 'lr' : 0.5}], momentum=0.25) for i in range(3): optimizer0.zero_grad() optimizer1.zero_grad() loss0 = model0(self.x) loss1 = model1(self.x) loss0.backward() loss1.backward() if i != which_iter: reference_grads[what_got_skipped(which_iter, which_backward)].append( [param.grad.data.clone() for param in model0.parameters()] + [param.grad.data.clone() for param in model1.parameters()]) if i == which_iter: if which_backward == 0: optimizer1.step() else: optimizer0.step() else: optimizer0.step() optimizer1.step() final_params[what_got_skipped(which_iter, which_backward)] = \ [param.data.clone() for param in model0.parameters()] + \ [param.data.clone() for param in model1.parameters()] for materialize_master_grads in (False, True): for opt_level in ("O0", "O1", "O2", "O3"): for how_to_zero in ("none", "model", "optimizer"): for use_multiple_loss_scalers in (False, True): if opt_level == "O1" or opt_level == "O2": inject_inf_iters = (-1, 0, 1) else: inject_inf_iters = (-1,) for inject_inf in inject_inf_iters: if inject_inf >= 0: inject_inf_locs = ("fp16", "fp32") which_backwards = (0, 1) else: inject_inf_locs = ("fdsa",) which_backwards = (None,) for inject_inf_loc in inject_inf_locs: for which_backward in which_backwards: if use_multiple_loss_scalers: num_losses = 2 loss_ids = [0, 1] else: num_losses = 1 loss_ids = [0, 0] if inject_inf >= 0: iters = 3 else: iters = 2 model0 = MyModel(1) model1 = MyModel(2) models = [model0, model1] optimizer0 = FusedSGD([{'params' : model0.parameters(), 'lr' : 0.25}], momentum=0.125, materialize_master_grads=materialize_master_grads) optimizer1 = FusedSGD([{'params' : model1.parameters(), 'lr' : 0.5}], momentum=0.25, materialize_master_grads=materialize_master_grads) _amp_state.allow_incoming_model_not_fp32 = True [model0, model1], [optimizer0, optimizer1] = amp.initialize( [model0, model1], [optimizer0, optimizer1], opt_level=opt_level, verbosity=0, cast_model_type=False, num_losses=num_losses) _amp_state.allow_incoming_model_not_fp32 = False _amp_state.loss_scalers[0]._loss_scale = 4.0 if use_multiple_loss_scalers: _amp_state.loss_scalers[1]._loss_scale = 16.0 unskipped = 0 for i in range(iters): if how_to_zero == "none": for model in models: for param in model.parameters(): param.grad = None elif how_to_zero == "model": for model in models: model.zero_grad() else: optimizer0.zero_grad() optimizer1.zero_grad() loss0 = model0(self.x) loss1 = model1(self.x) with amp.scale_loss(loss0, optimizer0, loss_id=loss_ids[0]) as scaled_loss: scaled_loss.backward() if i == inject_inf and which_backward == 0: if inject_inf_loc == "fp32": model0.weight0.grad[0] = float('inf') elif inject_inf_loc == "fp16": model0.weight1.grad[0] = float('inf') with amp.scale_loss(loss1, optimizer1, loss_id=loss_ids[1]) as scaled_loss: scaled_loss.backward() if i == inject_inf and which_backward == 1: if inject_inf_loc == "fp32": model1.weight0.grad[0] = float('inf') elif inject_inf_loc == "fp16": model1.weight1.grad[0] = float('inf') # print("opt_level {} i {} inject_inf {} which_backward {} inject_inf_loc {} use_multiple_loss_scalers {}".format(opt_level, i, inject_inf, which_backward, inject_inf_loc, use_multiple_loss_scalers)) if i != inject_inf: master_params = list(amp.master_params(optimizer0)) + \ list(amp.master_params(optimizer1)) for param, reference_grad in zip(master_params, reference_grads[what_got_skipped(inject_inf, which_backward)][unskipped]): if opt_level == "O2" and not materialize_master_grads: continue else: self.assertTrue(torch.allclose(param.grad.float(), reference_grad.float())) unskipped += 1 optimizer0.step() optimizer1.step() model_params = [p for p in model0.parameters()] + [p for p in model1.parameters()] master_params = [p for p in amp.master_params(optimizer0)] + \ [p for p in amp.master_params(optimizer1)] for model, master, reference in zip( model_params, master_params, final_params[what_got_skipped(inject_inf, which_backward)]): self.assertTrue(torch.allclose(model, reference)) self.assertTrue(torch.allclose(model, master.to(model.dtype))) if opt_level == "O1": _amp_state.handle._deactivate() @unittest.skipIf(disabled, "amp_C is unavailable") def test_3models2losses2optimizers(self): model0 = MyModel(1) model1 = MyModel(2) model2 = MyModel(3) optimizer0 = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}, {'params' : model1.parameters(), 'lr' : 1.0}], momentum=0.5) optimizer1 = torch.optim.SGD([{'params' : model2.parameters(), 'lr' : 0.5}], momentum=0.25) # Again, can't do this: reference_grads = [[]]*9 reference_grads = [[], [], [], [], [], [], [], [], []] final_params = [None, None, None, None, None, None, None, None, None] for i in range(2): optimizer0.zero_grad() optimizer1.zero_grad() loss0 = model0(self.x) + model1(self.x) loss1 = model2(self.x) + model1(self.x) loss0.backward() loss1.backward() reference_grads[0].append([param.grad.data.clone() for param in model0.parameters()] + [param.grad.data.clone() for param in model1.parameters()]) optimizer0.step() optimizer1.step() final_params[0] = \ [param.data.clone() for param in model0.parameters()] + \ [param.data.clone() for param in model1.parameters()] + \ [param.data.clone() for param in model2.parameters()] def what_got_skipped(which_iter, which_backward, which_model): if which_iter == 0: if which_backward == 0: if which_model == 0: return 1 if which_model == 1: return 2 if which_backward == 1: if which_model == 2: return 3 if which_model == 1: return 4 if which_iter == 1: if which_backward == 0: if which_model == 0: return 5 if which_model == 1: return 6 if which_backward == 1: if which_model == 2: return 7 if which_model == 1: return 8 return 0 for which_iter in (0,1): for which_backward in (0,1): if which_backward == 0: which_models = (0,1) if which_backward == 1: which_models = (2,1) for which_model in which_models: model0 = MyModel(1) model1 = MyModel(2) model2 = MyModel(3) optimizer0 = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}, {'params' : model1.parameters(), 'lr' : 1.0}], momentum=0.5) optimizer1 = torch.optim.SGD([{'params' : model2.parameters(), 'lr' : 0.5}], momentum=0.25) for i in range(3): optimizer0.zero_grad() optimizer1.zero_grad() loss0 = model0(self.x) + model1(self.x) loss1 = model2(self.x) + model1(self.x) loss0.backward() loss1.backward() if i != which_iter: reference_grads[what_got_skipped(which_iter, which_backward, which_model)].append( [param.grad.data.clone() for param in model0.parameters()] + [param.grad.data.clone() for param in model1.parameters()]) if i == which_iter: if which_backward == 0: # if which_model == 0: optimizer1.step() # if which_model == 1: # optimizer1.step() if which_backward == 1: # if which_model == 2: # optimizer0.step() # if which_model == 1: continue else: optimizer0.step() optimizer1.step() final_params[what_got_skipped(which_iter, which_backward, which_model)] = \ [param.data.clone() for param in model0.parameters()] + \ [param.data.clone() for param in model1.parameters()] + \ [param.data.clone() for param in model2.parameters()] for materialize_master_grads in (False, True): for opt_level in ("O0", "O1", "O2", "O3"): for how_to_zero in ("none", "model", "optimizer"): for use_multiple_loss_scalers in (False, True): if opt_level == "O1" or opt_level == "O2": inject_inf_iters = (-1, 0, 1) else: inject_inf_iters = (-1,) for inject_inf in inject_inf_iters: if inject_inf >= 0: inject_inf_locs = ("fp16", "fp32") which_backwards = (0, 1) else: inject_inf_locs = ("fdsa",) which_backwards = (None,) for inject_inf_loc in inject_inf_locs: for which_backward in which_backwards: if use_multiple_loss_scalers: num_losses = 2 loss_ids = [0, 1] else: num_losses = 1 loss_ids = [0, 0] if inject_inf >= 0: iters = 3 if which_backward == 0: which_models = (0, 1) elif which_backward == 1: which_models = (2, 1) else: iters = 2 which_models = (None,) for which_model in which_models: model0 = MyModel(1) model1 = MyModel(2) model2 = MyModel(3) models = [model0, model1, model2] optimizer0 = FusedSGD([{'params' : model0.parameters(), 'lr' : 0.25}, {'params' : model1.parameters(), 'lr' : 1.0}], momentum=0.5, materialize_master_grads=materialize_master_grads) optimizer1 = FusedSGD([{'params' : model2.parameters(), 'lr' : 0.5}], momentum=0.25, materialize_master_grads=materialize_master_grads) _amp_state.allow_incoming_model_not_fp32 = True [model0, model1, model2], [optimizer0, optimizer1] = amp.initialize( [model0, model1, model2], [optimizer0, optimizer1], opt_level=opt_level, verbosity=0, cast_model_type=False, num_losses=num_losses) _amp_state.allow_incoming_model_not_fp32 = False _amp_state.loss_scalers[0]._loss_scale = 4.0 if use_multiple_loss_scalers: _amp_state.loss_scalers[1]._loss_scale = 16.0 unskipped = 0 for i in range(iters): if how_to_zero == "none": for model in models: for param in model.parameters(): param.grad = None elif how_to_zero == "model": for model in models: model.zero_grad() else: optimizer0.zero_grad() optimizer1.zero_grad() loss0 = model0(self.x) + model1(self.x) loss1 = model2(self.x) + model1(self.x) with amp.scale_loss(loss0, optimizer0, loss_id=loss_ids[0]) as scaled_loss: scaled_loss.backward() if i == inject_inf and which_backward == 0: if which_model == 0: inj_model = model0 elif which_model == 1: inj_model = model1 else: raise RuntimeError(which_model + " invalid for loss 0") if inject_inf_loc == "fp32": inj_model.weight0.grad[0] = float('inf') elif inject_inf_loc == "fp16": inj_model.weight1.grad[0] = float('inf') with amp.scale_loss(loss1, [optimizer0, optimizer1], loss_id=loss_ids[1]) as scaled_loss: scaled_loss.backward() if i == inject_inf and which_backward == 1: if which_model == 2: inj_model = model2 elif which_model == 1: inj_model = model1 else: raise RuntimeError(which_model + " invalid for loss 1 ") if inject_inf_loc == "fp32": inj_model.weight0.grad[0] = float('inf') elif inject_inf_loc == "fp16": inj_model.weight1.grad[0] = float('inf') if i != inject_inf: master_params = list(amp.master_params(optimizer0)) + \ list(amp.master_params(optimizer1)) for param, reference_grad in zip(master_params, reference_grads[what_got_skipped(inject_inf, which_backward, which_model)][unskipped]): if opt_level == "O2" and not materialize_master_grads: continue else: self.assertTrue(torch.allclose(param.grad.float(), reference_grad.float())) unskipped += 1 optimizer0.step() optimizer1.step() model_params = [p for p in model0.parameters()] + \ [p for p in model1.parameters()] + \ [p for p in model2.parameters()] master_params = [p for p in amp.master_params(optimizer0)] + \ [p for p in amp.master_params(optimizer1)] # print("opt_level {} i {} inject_inf {} which_backward {} inject_inf_loc {} use_multiple_loss_scalers {} which_model {}".format(opt_level, i, inject_inf, which_backward, inject_inf_loc, use_multiple_loss_scalers, which_model)) for model, master, reference in zip( model_params, master_params, final_params[what_got_skipped(inject_inf, which_backward, which_model)]): self.assertTrue(torch.allclose(model, reference)) self.assertTrue(torch.allclose(model, master.to(model.dtype))) if opt_level == "O1": _amp_state.handle._deactivate() if __name__ == '__main__': unittest.main()
apex-master
tests/L0/run_amp/test_fused_sgd.py
import unittest import functools as ft import itertools as it from apex import amp from apex.amp import _amp_state import torch from torch import nn import torch.nn.functional as F from torch.nn import Parameter from utils import common_init, HALF, FLOAT,\ ALWAYS_HALF, ALWAYS_FLOAT, MATCH_INPUT class MyModel(torch.nn.Module): def __init__(self, unique): super(MyModel, self).__init__() self.weight0 = Parameter(unique + torch.arange(2, device='cuda', dtype=torch.float32)) self.weight1 = Parameter(1. + unique + torch.arange(2, device='cuda', dtype=torch.float16)) @staticmethod def ops(input, weight0, weight1): return ((input*(weight0.float()))*(weight1.float())).sum() def forward(self, input): return self.ops(input, self.weight0, self.weight1) # Abandon all hope, ye who enter here. class TestAddParamGroup(unittest.TestCase): def setUp(self): self.x = torch.ones((2), device='cuda', dtype=torch.float32) common_init(self) def tearDown(self): pass def zero_grad(self, models, optimizer, how_to_zero): if how_to_zero == "none": for model in models: for param in model.parameters(): param.grad = None elif how_to_zero == "model": for model in models: model.zero_grad() elif how_to_zero == "optimizer": optimizer.zero_grad() def test_add_param_group(self): for opt_level in ("O0", "O1", "O2", "O3"): for zero_before_add in (True, False): for try_accumulation in (True, False): model0 = MyModel(1) model1 = MyModel(2) optimizer = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}], momentum=0.125) optimizer.zero_grad() loss = model0(self.x) loss.backward() optimizer.step() if zero_before_add: optimizer.zero_grad() optimizer.add_param_group({'params' : model1.parameters(), 'lr' : 0.5}) if not zero_before_add: optimizer.zero_grad() loss = model0(self.x) + model1(self.x) loss.backward(retain_graph=try_accumulation) if try_accumulation: loss.backward() optimizer.step() # Once more to make sure the new params pick up momemtums properly optimizer.zero_grad() loss = model0(self.x) + model1(self.x) loss.backward(retain_graph=try_accumulation) if try_accumulation: loss.backward() optimizer.step() reference_params = [param.data.clone() for param in model0.parameters()] + \ [param.data.clone() for param in model1.parameters()] for how_to_zero in "none", "model", "optimizer": model0 = MyModel(1) model1 = MyModel(2) optimizer = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}], momentum=0.125) _amp_state.allow_incoming_model_not_fp32 = True [model0, model1], optimizer = amp.initialize([model0, model1], optimizer, opt_level=opt_level, verbosity=0, cast_model_type=False) _amp_state.allow_incoming_model_not_fp32 = False _amp_state.loss_scalers[0]._loss_scale = 4.0 self.zero_grad([model0, model1], optimizer, how_to_zero) loss = model0(self.x) with amp.scale_loss(loss, optimizer) as scaled_loss: scaled_loss.backward() optimizer.step() if zero_before_add: self.zero_grad([model0, model1], optimizer, how_to_zero) optimizer.add_param_group({'params' : model1.parameters(), 'lr' : 0.5}) if not zero_before_add: self.zero_grad([model0, model1], optimizer, how_to_zero) loss = model0(self.x) + model1(self.x) with amp.scale_loss(loss, optimizer) as scaled_loss: scaled_loss.backward(retain_graph=try_accumulation) if try_accumulation: with amp.scale_loss(loss, optimizer) as scaled_loss: scaled_loss.backward() optimizer.step() # Once more to make sure the new params pick up momentums properly self.zero_grad([model0, model1], optimizer, how_to_zero) loss = model0(self.x) + model1(self.x) with amp.scale_loss(loss, optimizer) as scaled_loss: scaled_loss.backward(retain_graph=try_accumulation) if try_accumulation: with amp.scale_loss(loss, optimizer) as scaled_loss: scaled_loss.backward() optimizer.step() final_params = [param.data.clone() for param in model0.parameters()] + \ [param.data.clone() for param in model1.parameters()] for reference, final in zip(reference_params, final_params): self.assertTrue(torch.allclose(reference.to(final.dtype), final), "opt_level = {}, how_to_zero = {}, zero_before_add = {}".format( opt_level, how_to_zero, zero_before_add)) if __name__ == '__main__': unittest.main()
apex-master
tests/L0/run_amp/test_add_param_group.py
apex-master
tests/L0/run_amp/__init__.py
import unittest import itertools as it from apex import amp import torch from torch import nn import torch.nn.functional as F from utils import common_init, HALF, FLOAT, DTYPES class TestPromotion(unittest.TestCase): def setUp(self): self.handle = amp.init(enabled=True) common_init(self) def tearDown(self): self.handle._deactivate() def run_binary_promote_test(self, fns, input_shape, x_inplace=False): type_pairs = it.product(DTYPES, DTYPES) for fn, (xtype, ytype) in it.product(fns, type_pairs): x = torch.randn(input_shape, dtype=xtype).requires_grad_() x_leaf = x if x_inplace: # We need a non-leaf to call in place on x = x.clone() y = torch.randn(input_shape, dtype=ytype) out = fn(x, y) if x_inplace: # In place: always match xtype self.assertEqual(out.type(), x.type()) else: # Out of place: match widest type if xtype == torch.float or ytype == torch.float: self.assertEqual(out.type(), FLOAT) else: self.assertEqual(out.type(), HALF) out.float().sum().backward() self.assertEqual(x_leaf.grad.dtype, xtype) def test_atan2_matches_widest(self): fns = [lambda x, y : torch.atan2(x, y), lambda x, y : x.atan2(y)] self.run_binary_promote_test(fns, (self.b,)) def test_mul_matches_widest(self): fns = [lambda x, y : torch.mul(x, y), lambda x, y: x.mul(y)] self.run_binary_promote_test(fns, (self.b,)) def test_cat_matches_widest(self): shape = self.b ys = [torch.randn(shape, dtype=torch.half) for _ in range(5)] x_float = torch.randn(shape) out = torch.cat(ys + [x_float]) self.assertEqual(out.type(), FLOAT) x_half = torch.randn(shape, dtype=torch.half) out = torch.cat(ys + [x_half]) self.assertEqual(out.type(), HALF) def test_inplace_exp_is_error_for_half(self): xs = torch.randn(self.b) xs.exp_() self.assertEqual(xs.type(), FLOAT) xs = torch.randn(self.b, dtype=torch.half) with self.assertRaises(NotImplementedError): xs.exp_() def test_inplace_add_matches_self(self): fn = lambda x, y: x.add_(y) self.run_binary_promote_test([fn], (self.b,), x_inplace=True) if __name__ == '__main__': unittest.main()
apex-master
tests/L0/run_amp/test_promotion.py
import unittest import functools as ft import itertools as it from apex import amp import torch from torch import nn import torch.nn.functional as F from math import floor from utils import common_init, HALF, FLOAT,\ ALWAYS_HALF, ALWAYS_FLOAT, MATCH_INPUT try: import amp_C from amp_C import multi_tensor_axpby from apex.multi_tensor_apply import MultiTensorApply disabled = False except ImportError as err: print("amp_C fused kernels unavailable, disabling TestMultiTensorApply. ImportError was ", err) disabled = True TORCH_MAJOR = int(torch.__version__.split('.')[0]) TORCH_MINOR = int(torch.__version__.split('.')[1]) try_nhwc = (TORCH_MAJOR > 1) or (TORCH_MAJOR == 1 and TORCH_MINOR > 4) class TestMultiTensorAxpby(unittest.TestCase): def setUp(self): common_init(self) self.a = 2.0 self.b = 8.0 self.xval = 4.0 self.yval = 16.0 self.overflow_buf = torch.cuda.IntTensor(1).zero_() self.ref = torch.full((1,), 136.0, device="cuda", dtype=torch.float32) def tearDown(self): pass # The tensor creation here is written for convenience, not speed. def axpby(self, sizea, sizeb, applier, repeat_tensors, x_type, y_type, out_type, inplace=False, nhwc=False): self.overflow_buf.zero_() sizea = sizea if isinstance(sizea, tuple) else (sizea,) sizeb = sizeb if isinstance(sizeb, tuple) else (sizeb,) t1 = torch.full(sizea, 1.0, device="cuda", dtype=torch.float32) t2 = torch.full(sizeb, 1.0, device="cuda", dtype=torch.float32) def to_fmt(t, tp): if nhwc: return t.clone().to(tp, memory_format=torch.channels_last) else: return t.clone().to(tp) y_list = [] for i in range(repeat_tensors): y_list += [to_fmt(t1, y_type)*self.yval, to_fmt(t2, y_type)*self.yval] x_list = [to_fmt(x, x_type)*(self.xval/self.yval) for x in y_list] if inplace: out_list = y_list else: out_list = [to_fmt(out, out_type)*3.0 for out in y_list] applier(multi_tensor_axpby, self.overflow_buf, [x_list, y_list, out_list], self.a, self.b, -1) self.assertTrue(all([torch.allclose(out, self.ref.to(out_type)) for out in out_list]), msg="{} {} {} {} {} {} {}".format(sizea, sizeb, repeat_tensors, x_type, y_type, out_type, inplace)) self.assertTrue(self.overflow_buf.item() == 0, msg="{} {} {} {} {} {} {}".format(sizea, sizeb, repeat_tensors, x_type, y_type, out_type, inplace)) # def find_inf(self, sizea, sizeb, applier, repeat_tensors, in_type, out_type, t, ind, val, inplace=False): # self.overflow_buf.zero_() # a = torch.cuda.FloatTensor(sizea).fill_(self.scale) # b = torch.cuda.FloatTensor(sizeb).fill_(self.scale) # out_list = [] # for i in range(repeat_tensors): # out_list += [a.clone().to(out_type), b.clone().to(out_type)] # if inplace: # in_list = out_list # else: # in_list = [out.clone().to(in_type) for out in out_list] # applier(multi_tensor_scale, self.overflow_buf, [in_list, out_list], 1./self.scale) # self.overflow_buf.zero_() # in_list[t][ind] = val # applier(multi_tensor_scale, self.overflow_buf, [in_list, out_list], 1./self.scale) # self.assertTrue(self.overflow_buf.item()) @unittest.skipIf(disabled, "amp_C is unavailable") def test_fuzz(self): input_size_pairs = ( (7777*77, 555*555), (777, 555), (555, 2048*32+1), (2048*32+1, 555), (555, 2048*32), (2048*32, 555), (33333, 555), (555, 33333)) appliers = ( MultiTensorApply(2048*32), MultiTensorApply(333), MultiTensorApply(33333)) repeat_tensors = ( 1, 55) for sizea, sizeb in input_size_pairs: for applier in appliers: for repeat in repeat_tensors: for x_type in (torch.float32, torch.float16): for y_type in (torch.float32, torch.float16): for out_type in (torch.float32, torch.float16): for inplace in (True, False): if inplace is True and (y_type is not out_type): continue else: self.axpby(sizea, sizeb, applier, repeat, x_type, y_type, out_type, inplace=inplace) # self.find_inf(sizea, sizeb, applier, repeat, in_type, out_type, # 0, 0, float('nan'), inplace=inplace) # self.find_inf(sizea, sizeb, applier, repeat, in_type, out_type, # 2*repeat-1, sizeb-1, float('inf'), inplace=inplace) # self.find_inf(sizea, sizeb, applier, repeat, in_type, out_type, # 2*(repeat//2), sizea//2, float('inf'), inplace=inplace) @unittest.skipIf(disabled, "amp_C is unavailable") @unittest.skipIf(not try_nhwc, "torch version is 1.4 or earlier, may not support nhwc") def test_fuzz_nhwc(self): input_size_pairs = ( ((7, 77, 7, 77), (5, 55, 5, 55)), ((1, 1, 777, 1), (1, 1, 555, 1)), ((5, 47, 5, 55), (1, 1, 1, 2048*32 + 1)), ((1, 1, 1, 2048*32 + 1), (55, 47, 5, 55)), ((555, 1, 1, 1), (32, 8, 32, 8)), ((32, 8, 32, 8), (55, 47, 5, 55)), ((1, 1, 33333, 1), (55, 47, 55, 5)), ((55, 47, 55, 5), (1, 1, 33333, 1))) appliers = ( MultiTensorApply(2048*32), MultiTensorApply(333), MultiTensorApply(33333)) repeat_tensors = ( 1, 55) for sizea, sizeb in input_size_pairs: for applier in appliers: for repeat in repeat_tensors: for x_type in (torch.float32, torch.float16): for y_type in (torch.float32, torch.float16): for out_type in (torch.float32, torch.float16): for inplace in (True, False): if inplace is True and (y_type is not out_type): continue else: self.axpby(sizea, sizeb, applier, repeat, x_type, y_type, out_type, inplace=inplace, nhwc=True) # self.find_inf(sizea, sizeb, applier, repeat, in_type, out_type, # 0, 0, float('nan'), inplace=inplace) # self.find_inf(sizea, sizeb, applier, repeat, in_type, out_type, # 2*repeat-1, sizeb-1, float('inf'), inplace=inplace) # self.find_inf(sizea, sizeb, applier, repeat, in_type, out_type, # 2*(repeat//2), sizea//2, float('inf'), inplace=inplace) if __name__ == '__main__': unittest.main()
apex-master
tests/L0/run_amp/test_multi_tensor_axpby.py
import torch HALF = 'torch.cuda.HalfTensor' FLOAT = 'torch.cuda.FloatTensor' DTYPES = [torch.half, torch.float] ALWAYS_HALF = {torch.float: HALF, torch.half: HALF} ALWAYS_FLOAT = {torch.float: FLOAT, torch.half: FLOAT} MATCH_INPUT = {torch.float: FLOAT, torch.half: HALF} def common_init(test_case): test_case.h = 64 test_case.b = 16 test_case.c = 16 test_case.k = 3 test_case.t = 10 torch.set_default_tensor_type(torch.cuda.FloatTensor)
apex-master
tests/L0/run_amp/utils.py
import unittest from apex import amp import random import torch from torch import nn from utils import common_init, HALF class TestRnnCells(unittest.TestCase): def setUp(self): self.handle = amp.init(enabled=True) common_init(self) def tearDown(self): self.handle._deactivate() def run_cell_test(self, cell, state_tuple=False): shape = (self.b, self.h) for typ in [torch.float, torch.half]: xs = [torch.randn(shape, dtype=typ).requires_grad_() for _ in range(self.t)] hidden_fn = lambda: torch.zeros(shape, dtype=typ) if state_tuple: hidden = (hidden_fn(), hidden_fn()) else: hidden = hidden_fn() outputs = [] for i in range(self.t): hidden = cell(xs[i], hidden) if state_tuple: output = hidden[0] else: output = hidden outputs.append(output) for y in outputs: self.assertEqual(y.type(), HALF) outputs[-1].float().sum().backward() for i, x in enumerate(xs): self.assertEqual(x.grad.dtype, x.dtype) def test_rnn_cell_is_half(self): cell = nn.RNNCell(self.h, self.h) self.run_cell_test(cell) def test_gru_cell_is_half(self): cell = nn.GRUCell(self.h, self.h) self.run_cell_test(cell) def test_lstm_cell_is_half(self): cell = nn.LSTMCell(self.h, self.h) self.run_cell_test(cell, state_tuple=True) class TestRnns(unittest.TestCase): def setUp(self): self.handle = amp.init(enabled=True) common_init(self) def tearDown(self): self.handle._deactivate() def run_rnn_test(self, rnn, layers, bidir, state_tuple=False): for typ in [torch.float, torch.half]: x = torch.randn((self.t, self.b, self.h), dtype=typ).requires_grad_() hidden_fn = lambda: torch.zeros((layers + (layers * bidir), self.b, self.h), dtype=typ) if state_tuple: hidden = (hidden_fn(), hidden_fn()) else: hidden = hidden_fn() output, _ = rnn(x, hidden) self.assertEqual(output.type(), HALF) output[-1, :, :].float().sum().backward() self.assertEqual(x.grad.dtype, x.dtype) def test_rnn_is_half(self): configs = [(1, False), (2, False), (2, True)] for layers, bidir in configs: rnn = nn.RNN(input_size=self.h, hidden_size=self.h, num_layers=layers, nonlinearity='relu', bidirectional=bidir) self.run_rnn_test(rnn, layers, bidir) def test_gru_is_half(self): configs = [(1, False), (2, False), (2, True)] for layers, bidir in configs: rnn = nn.GRU(input_size=self.h, hidden_size=self.h, num_layers=layers, bidirectional=bidir) self.run_rnn_test(rnn, layers, bidir) def test_lstm_is_half(self): configs = [(1, False), (2, False), (2, True)] for layers, bidir in configs: rnn = nn.LSTM(input_size=self.h, hidden_size=self.h, num_layers=layers, bidirectional=bidir) self.run_rnn_test(rnn, layers, bidir, state_tuple=True) def test_rnn_packed_sequence(self): num_layers = 2 rnn = nn.RNN(input_size=self.h, hidden_size=self.h, num_layers=num_layers) for typ in [torch.float, torch.half]: x = torch.randn((self.t, self.b, self.h), dtype=typ).requires_grad_() lens = sorted([random.randint(self.t // 2, self.t) for _ in range(self.b)], reverse=True) # `pack_padded_sequence` breaks if default tensor type is non-CPU torch.set_default_tensor_type(torch.FloatTensor) lens = torch.tensor(lens, dtype=torch.int64, device=torch.device('cpu')) packed_seq = nn.utils.rnn.pack_padded_sequence(x, lens) torch.set_default_tensor_type(torch.cuda.FloatTensor) hidden = torch.zeros((num_layers, self.b, self.h), dtype=typ) output, _ = rnn(packed_seq, hidden) self.assertEqual(output.data.type(), HALF) output.data.float().sum().backward() self.assertEqual(x.grad.dtype, x.dtype) if __name__ == '__main__': unittest.main()
apex-master
tests/L0/run_amp/test_rnn.py
import unittest import functools as ft import itertools as it from apex import amp import torch from torch import nn import torch.nn.functional as F from utils import common_init, HALF, FLOAT,\ ALWAYS_HALF, ALWAYS_FLOAT, MATCH_INPUT try: import amp_C from amp_C import multi_tensor_scale from apex.multi_tensor_apply import MultiTensorApply disabled = False except ImportError as err: print("amp_C fused kernels unavailable, disabling TestMultiTensorApply. ImportError was ", err) disabled = True class TestMultiTensorScale(unittest.TestCase): def setUp(self): common_init(self) self.scale = 4.0 self.overflow_buf = torch.cuda.IntTensor(1).zero_() self.ref = torch.cuda.FloatTensor([1.0]) def tearDown(self): pass # The tensor creation here is written for convenience, not speed. def downscale(self, sizea, sizeb, applier, repeat_tensors, in_type, out_type, inplace=False): self.overflow_buf.zero_() a = torch.cuda.FloatTensor(sizea).fill_(self.scale) b = torch.cuda.FloatTensor(sizeb).fill_(self.scale) out_list = [] for i in range(repeat_tensors): out_list += [a.clone().to(out_type), b.clone().to(out_type)] if inplace: in_list = out_list else: in_list = [out.clone().to(in_type) for out in out_list] applier(multi_tensor_scale, self.overflow_buf, [in_list, out_list], 1./self.scale) self.assertTrue(all([torch.allclose(out, self.ref.to(out_type)) for out in out_list])) self.assertTrue(self.overflow_buf.item() == 0) def find_inf(self, sizea, sizeb, applier, repeat_tensors, in_type, out_type, t, ind, val, inplace=False): self.overflow_buf.zero_() a = torch.cuda.FloatTensor(sizea).fill_(self.scale) b = torch.cuda.FloatTensor(sizeb).fill_(self.scale) out_list = [] for i in range(repeat_tensors): out_list += [a.clone().to(out_type), b.clone().to(out_type)] if inplace: in_list = out_list else: in_list = [out.clone().to(in_type) for out in out_list] applier(multi_tensor_scale, self.overflow_buf, [in_list, out_list], 1./self.scale) self.overflow_buf.zero_() in_list[t][ind] = val applier(multi_tensor_scale, self.overflow_buf, [in_list, out_list], 1./self.scale) self.assertTrue(self.overflow_buf.item()) # Currently, the fused kernel gives a hard error if you attempt to downscale # into fp16 output, which imo is the desired behavior. Maybe someday we # will learn otherwise. # @unittest.skipIf(disabled, "amp_C is unavailable") # def test_fp16_to_fp16(self): # self.downscale(self.fp16, self.fp16, self.fp16_ref) # # @unittest.skipIf(disabled, "amp_C is unavailable") # def test_fp32_to_fp16(self): # self.downscale(self.fp32, self.fp16, self.fp16_ref) @unittest.skipIf(disabled, "amp_C is unavailable") def test_fuzz(self): input_size_pairs = ( (7777*77, 555*555), (777, 555), (555, 2048*32+1), (2048*32+1, 555), (555, 2048*32), (2048*32, 555), (33333, 555), (555, 33333)) appliers = ( MultiTensorApply(2048*32), MultiTensorApply(333), MultiTensorApply(33333)) repeat_tensors = ( 1, 55) for sizea, sizeb in input_size_pairs: for applier in appliers: for repeat in repeat_tensors: for in_type in (torch.float32, torch.float16): for out_type in (torch.float32, torch.float16): for inplace in (True, False): if inplace is True and (out_type is not in_type): continue else: self.downscale(sizea, sizeb, applier, repeat, in_type, out_type, inplace=inplace) self.find_inf(sizea, sizeb, applier, repeat, in_type, out_type, 0, 0, float('nan'), inplace=inplace) self.find_inf(sizea, sizeb, applier, repeat, in_type, out_type, 2*repeat-1, sizeb-1, float('inf'), inplace=inplace) self.find_inf(sizea, sizeb, applier, repeat, in_type, out_type, 2*(repeat//2), sizea//2, float('inf'), inplace=inplace) if __name__ == '__main__': unittest.main()
apex-master
tests/L0/run_amp/test_multi_tensor_scale.py
import unittest import torch import torch.nn as nn import torch.nn.functional as F import torch.optim as optim from apex import amp from utils import common_init, FLOAT class MyModel(torch.nn.Module): def __init__(self): super(MyModel, self).__init__() self.conv1 = nn.Conv2d(3, 6, 3, 1, 1) self.bn1 = nn.BatchNorm2d(6) self.param = nn.Parameter(torch.randn(1)) def forward(self, x): x = x * self.param x = F.relu(self.conv1(x)) x = self.bn1(x) return x class TestCheckpointing(unittest.TestCase): def setUp(self): self.initial_lr = 1e-3 self.test_opt_levels = ("O0", "O1", "O2", "O3") def seed(self): torch.manual_seed(2809) torch.backends.cudnn.benchmark = False torch.backends.cudnn.deterministic = True def check_state_dict_fp32(self, state_dict): for key in state_dict: if 'num_batches_tracked' in key: continue param = state_dict[key] self.assertEqual(param.type(), FLOAT, 'Parameter in state_dict not FLOAT') def train_step(self, model, optimizer, data, loss_ids): optimizer.zero_grad() output = model(data) # Call backward for num_losses-1 for idx in loss_ids: loss = output.mean() with amp.scale_loss(loss, optimizer, loss_id=idx) as scaled_loss: scaled_loss.backward(retain_graph=True) optimizer.step() return output def compare_models(self, modelA, modelB, test_setup=''): state_dictA = modelA.state_dict() state_dictB = modelB.state_dict() self.assertEqual(len(state_dictA), len(state_dictB), 'state_dicts have different lengths' + test_setup) for key in state_dictA: paramA = state_dictA[key] paramB = state_dictB[key] self.assertTrue((paramA==paramB).all(), msg='Parameters in state_dices not equal.' + 'key: {}\nparam: {}\nrestored: {}\ndiff: {} for {}'.format( key, paramA, paramB, paramA - paramB, test_setup)) def test_restoring(self): nb_epochs = 10 nb_epochs_restore = nb_epochs // 2 for opt_level in self.test_opt_levels: for res_opt_level in self.test_opt_levels: for amp_before_load in [True, False]: for num_losses in range(1, 3): test_setup = ('#' * 75 + '\n' + \ f'opt_level {opt_level}\n' + \ f'restore_opt_level {res_opt_level}\n' + \ f'amp_before_load {amp_before_load}\n' + \ f'num_losses {num_losses}\n') self.seed() # Create reference model model = MyModel().to('cuda') optimizer = optim.SGD(model.parameters(), lr=self.initial_lr) # Initialize with num_losses*2 for the original model and the restored one model, optimizer = amp.initialize( model, optimizer, opt_level=opt_level, num_losses=num_losses*2, verbosity=0) # Compare training behavior for same restore option # We cannot really generalize it, since a saved model in O0 # would introduce a skipped step in O1, which will raise an error if opt_level == res_opt_level: # train for nb_epochs and restore after nb_epochs_restore for epoch in range(nb_epochs): x = torch.randn(16, 3, 24, 24, device='cuda') output = self.train_step( model, optimizer, x, range(num_losses)) # Initialize model one step before comparing. # Otherwise the batchnorm layers will be updated # additionally in restore_model if epoch == (nb_epochs_restore - 1): # Load model and optimizer checkpoint = { 'model': model.state_dict(), 'optimizer': optimizer.state_dict(), 'amp': amp.state_dict() } # Check state_dict for FP32 tensors self.check_state_dict_fp32(checkpoint['model']) # Restore model restore_model = MyModel().to('cuda') restore_optimizer = optim.SGD( restore_model.parameters(), lr=self.initial_lr) if amp_before_load: restore_model, restore_optimizer = amp.initialize( restore_model, restore_optimizer, opt_level=res_opt_level, num_losses=num_losses*2, verbosity=0) restore_model.load_state_dict(checkpoint['model']) restore_optimizer.load_state_dict(checkpoint['optimizer']) # FIXME: We cannot test the amp.state_dict in the same script # amp.load_state_dict(checkpoint['amp']) if not amp_before_load: restore_model, restore_optimizer = amp.initialize( restore_model, restore_optimizer, opt_level=res_opt_level, num_losses=num_losses*2, verbosity=0) elif epoch >= nb_epochs_restore: restore_output = self.train_step( restore_model, restore_optimizer, x, range(num_losses, num_losses*2)) self.assertTrue( torch.allclose(output.float(), restore_output.float()), 'Output of reference and restored models differ for ' + test_setup) self.compare_models(model, restore_model, test_setup) # if opt_level != res_opt_level else: # skip tests for different opt_levels continue def test_loss_scale_decrease(self): num_losses = 3 nb_decrease_loss_scales = [0, 1, 2] for opt_level in self.test_opt_levels: #print('#' * 75 + f'\n opt_level {opt_level}\n') # Create new tmp copy for this run nb_decrease_loss_scales_tmp = list(nb_decrease_loss_scales) model = MyModel().to('cuda') optimizer = optim.SGD(model.parameters(), lr=self.initial_lr) model, optimizer = amp.initialize( model, optimizer, opt_level=opt_level, num_losses=num_losses, verbosity=0) if amp._amp_state.opt_properties.loss_scale != 'dynamic': #print('Static loss scale set. Skipping opt_level.') continue # force to skip some updates to decrease the loss_scale initial_loss_scales = [] for idx in range(num_losses): initial_loss_scales.append( amp._amp_state.loss_scalers[idx].loss_scale()) for _ in range(len(nb_decrease_loss_scales)): x = torch.randn(16, 3, 24, 24, device='cuda') for idx in range(num_losses): while nb_decrease_loss_scales_tmp[idx] > 0: optimizer.zero_grad() output = model(x * 2**17) loss = output.mean() with amp.scale_loss(loss, optimizer, loss_id=idx) as scaled_loss: scaled_loss.backward(retain_graph=True) optimizer.step() nb_decrease_loss_scales_tmp[idx] -= 1 # Check loss scales afterwards updated_loss_scales = [] for idx in range(num_losses): updated_loss_scales.append( amp._amp_state.loss_scalers[idx].loss_scale()) for factor, update_ls, init_ls in zip(nb_decrease_loss_scales, updated_loss_scales, initial_loss_scales): self.assertEqual(update_ls, init_ls / 2**factor) # Check state dict amp_state_dict = amp.state_dict() for scaler_idx, factor, init_ls in zip(amp_state_dict, nb_decrease_loss_scales, initial_loss_scales): scaler = amp_state_dict[scaler_idx] self.assertEqual(scaler['loss_scale'], init_ls / 2**factor) unskipped_target = 0 self.assertEqual(scaler['unskipped'], unskipped_target) def test_state_dict(self): for opt_level in self.test_opt_levels: # Skip O3 if opt_level == 'O3': continue model = MyModel().to('cuda') optimizer = optim.Adam(model.parameters(), lr=1e-3) model, optimizer = amp.initialize( model, optimizer, opt_level=opt_level, verbosity=0) # Export state_dict and check for Half state_dict = model.state_dict() for key in state_dict: self.assertFalse('Half' in state_dict[key].type()) # Check, if model is still trainable # Create dummy data data = torch.randn(10, 3, 4, 4, device='cuda') target = torch.randn(10, 6, 4, 4, device='cuda') # Get initnial loss optimizer.zero_grad() output = model(data) loss = F.mse_loss(output, target) with amp.scale_loss(loss, optimizer) as scaled_loss: scaled_loss.backward() optimizer.step() last_loss = loss.item() # train for some epochs for epoch in range(10): optimizer.zero_grad() output = model(data) loss = F.mse_loss(output, target) with amp.scale_loss(loss, optimizer) as scaled_loss: scaled_loss.backward() optimizer.step() self.assertTrue(loss.item() < last_loss) last_loss = loss.item() if __name__=='__main__': unittest.main()
apex-master
tests/L0/run_amp/test_checkpointing.py
import unittest import functools as ft import itertools as it from apex import amp from apex.amp import _amp_state import torch from torch import nn import torch.nn.functional as F from utils import common_init, HALF, FLOAT,\ ALWAYS_HALF, ALWAYS_FLOAT, MATCH_INPUT def get_reference_grad(i, w, ops): # Creating new tensors ensures, among other things, that the new tensors are not in the cache. # In fact, they are guaranteed not to use the cache because they are not torch.nn.Parameters. fp32_i = i.detach().clone().float() fp32_w = w.detach().clone().float().requires_grad_() loss = ops(fp32_i, fp32_w) loss.backward() return fp32_w.grad class WhitelistModule(torch.nn.Module): def __init__(self, dtype): super(WhitelistModule, self).__init__() self.weight = torch.nn.Parameter(torch.arange(8*8, device='cuda', dtype=dtype).view(8,8)) @staticmethod def ops(input, weight): return (input.mm(weight)).mm(weight).sum() def forward(self, input): return self.ops(input, self.weight) class BlacklistModule(torch.nn.Module): def __init__(self, dtype): super(BlacklistModule, self).__init__() self.weight = torch.nn.Parameter(torch.arange(2*8, device='cuda', dtype=dtype).view(2,8)) @staticmethod def ops(input, weight): return (input + torch.pow(weight, 2) + torch.pow(weight, 2)).sum() def forward(self, input): return self.ops(input, self.weight) class PromoteModule(torch.nn.Module): def __init__(self, dtype): super(PromoteModule, self).__init__() self.weight = torch.nn.Parameter(torch.arange(2*8, device='cuda', dtype=dtype).view(2,8)) @staticmethod def ops(input, weight): return ((input*weight)*weight).sum() def forward(self, input): return self.ops(input, self.weight) class TestCache(unittest.TestCase): def setUp(self): self.x = torch.ones((2, 8), device='cuda', dtype=torch.float32) common_init(self) def tearDown(self): pass def train_eval_train_test(self, module, t): model = module(t).cuda() optimizer = torch.optim.SGD(model.parameters(), lr=1.0) _amp_state.allow_incoming_model_not_fp32 = True model, optimizer = amp.initialize(model, optimizer, opt_level="O1", verbosity=0) _amp_state.allow_incoming_model_not_fp32 = False def training_step(): for param in model.parameters(): param.grad = None loss = model(self.x).sum() _amp_state.loss_scalers[0]._loss_scale = 4.0 with amp.scale_loss(loss, optimizer) as scaled_loss: scaled_loss.backward() self.assertEqual(len([p.grad for p in model.parameters() if p.grad is not None]), 1) self.assertEqual(model.weight.grad.type(), model.weight.type()) reference_grad = get_reference_grad(self.x, model.weight, model.ops) # Currently there's no difference in the allclose calls, so no need for branching, # but I'm keeping this in case we want different tolerances for fp16 and fp32 checks. if model.weight.grad.type() == "torch.cuda.HalfTensor": self.assertTrue(torch.allclose(model.weight.grad.float(), reference_grad)) elif model.weight.grad.type() == "torch.cuda.FloatTensor": self.assertTrue(torch.allclose(model.weight.grad.float(), reference_grad)) else: raise RuntimeError("model.weight.grad.type = {}".format(model.weight.grad.type())) model.weight.data -= 1. # Simulates first epoch training_step() # Simulates eval with torch.no_grad(): loss = model(self.x).sum() # Simulates resuming training after eval training_step() _amp_state.handle._deactivate() # I could easily have these as a set of for loops in a single test, # instead of going for granularity. def test_whitelist_module_fp16_weight(self): self.train_eval_train_test(WhitelistModule, torch.float16) def test_whitelist_module_fp32_weight(self): self.train_eval_train_test(WhitelistModule, torch.float32) def test_blacklist_module_fp16_weight(self): self.train_eval_train_test(BlacklistModule, torch.float16) def test_blacklist_module_fp32_weight(self): self.train_eval_train_test(BlacklistModule, torch.float32) def test_promote_module_fp16_weight(self): self.train_eval_train_test(PromoteModule, torch.float16) def test_promote_module_fp32_weight(self): self.train_eval_train_test(PromoteModule, torch.float32) if __name__ == '__main__': unittest.main()
apex-master
tests/L0/run_amp/test_cache.py
import unittest import torch from torch import nn from torch.nn import Parameter from apex import amp from apex.parallel.LARC import LARC from utils import common_init class MyModel(torch.nn.Module): def __init__(self, unique): super(MyModel, self).__init__() self.weight0 = Parameter( unique + torch.arange(2, device="cuda", dtype=torch.float32) ) def forward(self, input): return (input * self.weight0).sum() class TestLARC(unittest.TestCase): def setUp(self): self.x = torch.ones((2), device="cuda", dtype=torch.float32) common_init(self) def tearDown(self): pass def test_larc_mixed_precision(self): for opt_level in ["O0", "O1", "O2", "O3"]: model = MyModel(1) optimizer = LARC( torch.optim.SGD( [{"params": model.parameters(), "lr": 0.25}], momentum=0.125 ) ) model, optimizer = amp.initialize( model, optimizer, opt_level=opt_level, verbosity=0 ) optimizer.zero_grad() loss = model(self.x) with amp.scale_loss(loss, optimizer) as scaled_loss: scaled_loss.backward() optimizer.step() if __name__ == "__main__": unittest.main()
apex-master
tests/L0/run_amp/test_larc.py
import unittest import functools as ft import itertools as it from apex import amp import torch from torch import nn import torch.nn.functional as F from utils import common_init, HALF, FLOAT,\ ALWAYS_HALF, ALWAYS_FLOAT, MATCH_INPUT def run_layer_test(test_case, fns, expected, input_shape, test_backward=True): for fn, typ in it.product(fns, expected.keys()): x = torch.randn(input_shape, dtype=typ).requires_grad_() y = fn(x) test_case.assertEqual(y.type(), expected[typ]) if test_backward: y.float().sum().backward() test_case.assertEqual(x.grad.type(), MATCH_INPUT[typ]) class TestBasicCasts(unittest.TestCase): def setUp(self): self.handle = amp.init(enabled=True) common_init(self) def tearDown(self): self.handle._deactivate() def test_linear_is_half(self): m = nn.Linear(self.h, self.h) f = ft.partial(F.linear, weight=m.weight, bias=m.bias) run_layer_test(self, [m, f], ALWAYS_HALF, (self.b, self.h)) def test_conv2d_is_half(self): m = nn.Conv2d(self.c, self.c, self.k) f = ft.partial(F.conv2d, weight=m.weight, bias=m.bias) run_layer_test(self, [m, f], ALWAYS_HALF, (self.b, self.c, self.h, self.h)) def test_softmax_is_float(self): m = nn.Softmax(dim=1) f = ft.partial(F.softmax, dim=1) run_layer_test(self, [m, f], ALWAYS_FLOAT, (self.b, self.h)) def test_group_norm_is_float(self): m = nn.GroupNorm(num_groups=4, num_channels=self.c) run_layer_test(self, [m], ALWAYS_FLOAT, (self.b, self.c, self.h, self.h)) def test_mse_loss_is_float(self): shape = (self.b, self.h) target = torch.randn(shape) mod = nn.MSELoss() m = lambda x: mod(x, target) f = ft.partial(F.mse_loss, target=target) run_layer_test(self, [m], ALWAYS_FLOAT, shape) def test_relu_is_match(self): run_layer_test(self, [nn.ReLU(), F.relu], MATCH_INPUT, (self.b, self.h)) def test_batch_norm_is_match(self): m = nn.BatchNorm2d(num_features=self.c) f = ft.partial(F.batch_norm, running_mean=m.running_mean, running_var=m.running_var, weight=m.weight, bias=m.bias, training=True) run_layer_test(self, [m], MATCH_INPUT, (self.b, self.c, self.h, self.h)) # Test forward-only for BN inference m.eval() f = ft.partial(F.batch_norm, running_mean=m.running_mean, running_var=m.running_var, weight=m.weight, bias=m.bias, training=False) run_layer_test(self, [m, f], MATCH_INPUT, (self.b, self.c, self.h, self.h), test_backward=False) class TestBannedMethods(unittest.TestCase): def setUp(self): self.handle = amp.init(enabled=True) common_init(self) def tearDown(self): self.handle._deactivate() def bce_common(self, assertion): shape = (self.b, self.h) target = torch.rand(shape) mod = nn.BCELoss() m = lambda x: mod(x, target) f = ft.partial(F.binary_cross_entropy, target=target) for fn in [m, f]: x = torch.rand(shape, dtype=torch.half) assertion(fn, x) def test_bce_raises_by_default(self): assertion = lambda fn, x: self.assertRaises(NotImplementedError, fn, x) self.bce_common(assertion) def test_bce_is_float_with_allow_banned(self): self.handle._deactivate() self.handle = amp.init(enabled=True, allow_banned=True) assertion = lambda fn, x: self.assertEqual(fn(x).type(), FLOAT) self.bce_common(assertion) class TestTensorCasts(unittest.TestCase): def setUp(self): self.handle = amp.init(enabled=True) common_init(self) def tearDown(self): self.handle._deactivate() def test_matmul_method_is_half(self): other = torch.randn(self.h, self.h) lhs = lambda x: x.matmul(other) rhs = lambda x: other.matmul(x) run_layer_test(self, [lhs, rhs], ALWAYS_HALF, (self.h, self.h)) def test_matmul_op_is_half(self): other = torch.randn(self.h, self.h) lhs = lambda x: x @ other rhs = lambda x: other @ x run_layer_test(self, [lhs, rhs], ALWAYS_HALF, (self.h, self.h)) def test_pow_method_is_float(self): fn = lambda x: x.pow(2.) run_layer_test(self, [fn], ALWAYS_FLOAT, (self.b, self.h)) def test_pow_op_is_float(self): fn = lambda x: x ** 2. run_layer_test(self, [fn], ALWAYS_FLOAT, (self.b, self.h)) def test_cpu_is_float(self): fn = lambda x: x.cpu() always_cpu_float = {torch.float: 'torch.FloatTensor', torch.half: 'torch.FloatTensor'} run_layer_test(self, [fn], always_cpu_float, (self.b, self.h)) def test_sum_is_float(self): fn = lambda x: x.sum() run_layer_test(self, [fn], ALWAYS_FLOAT, (self.b, self.h)) # TODO: maybe more tests on disabled casting? if __name__ == '__main__': unittest.main()
apex-master
tests/L0/run_amp/test_basic_casts.py
import unittest import torch import torch.nn as nn from apex.fp16_utils import FP16Model class DummyBlock(nn.Module): def __init__(self): super(DummyBlock, self).__init__() self.conv = nn.Conv2d(10, 10, 2) self.bn = nn.BatchNorm2d(10, affine=True) def forward(self, x): return self.conv(self.bn(x)) class DummyNet(nn.Module): def __init__(self): super(DummyNet, self).__init__() self.conv1 = nn.Conv2d(3, 10, 2) self.bn1 = nn.BatchNorm2d(10, affine=False) self.db1 = DummyBlock() self.db2 = DummyBlock() def forward(self, x): out = x out = self.conv1(out) out = self.bn1(out) out = self.db1(out) out = self.db2(out) return out class DummyNetWrapper(nn.Module): def __init__(self): super(DummyNetWrapper, self).__init__() self.bn = nn.BatchNorm2d(3, affine=True) self.dn = DummyNet() def forward(self, x): return self.dn(self.bn(x)) class TestFP16Model(unittest.TestCase): def setUp(self): self.N = 64 self.C_in = 3 self.H_in = 16 self.W_in = 32 self.in_tensor = torch.randn((self.N, self.C_in, self.H_in, self.W_in)).cuda() self.orig_model = DummyNetWrapper().cuda() self.fp16_model = FP16Model(self.orig_model) def test_params_and_buffers(self): exempted_modules = [ self.fp16_model.network.bn, self.fp16_model.network.dn.db1.bn, self.fp16_model.network.dn.db2.bn, ] for m in self.fp16_model.modules(): expected_dtype = torch.float if (m in exempted_modules) else torch.half for p in m.parameters(recurse=False): assert p.dtype == expected_dtype for b in m.buffers(recurse=False): assert b.dtype in (expected_dtype, torch.int64) def test_output_is_half(self): out_tensor = self.fp16_model(self.in_tensor) assert out_tensor.dtype == torch.half
apex-master
tests/L0/run_fp16util/test_fp16util.py
apex-master
tests/L0/run_fp16util/__init__.py
import unittest import torch import apex from apex.transformer.testing.distributed_test_base import NcclDistributedTestBase def init_model_and_optimizer(): model = torch.nn.Linear(1, 1, bias=False).cuda() optimizer = torch.optim.SGD(model.parameters(), 1.0) return model, optimizer @unittest.skipUnless(torch.cuda.is_available(), "") class TestDeprecatedWarning(unittest.TestCase): def test_amp(self): model, optimizer = init_model_and_optimizer() with self.assertWarns(apex.DeprecatedFeatureWarning): _ = apex.amp.initialize(model, optimizer) def test_fp16_model(self): model, _ = init_model_and_optimizer() with self.assertWarns(apex.DeprecatedFeatureWarning): _ = apex.fp16_utils.FP16Model(model) def test_fp16_optimizer(self): _, optimizer = init_model_and_optimizer() with self.assertWarns(apex.DeprecatedFeatureWarning): _ = apex.fp16_utils.FP16_Optimizer(optimizer) def test_fp16_loss_scaler(self): with self.assertWarns(apex.DeprecatedFeatureWarning): apex.fp16_utils.LossScaler() class TestParallel(NcclDistributedTestBase): @property def world_size(self): return min(torch.cuda.device_count(), 2) def test_distributed_data_parallel(self): model, _ = init_model_and_optimizer() with self.assertWarns(apex.DeprecatedFeatureWarning): _ = apex.parallel.DistributedDataParallel(model) def test_convert_syncbn_model(self): model, _ = init_model_and_optimizer() with self.assertWarns(apex.DeprecatedFeatureWarning): _ = apex.parallel.convert_syncbn_model(model) if __name__ == "__main__": unittest.main()
apex-master
tests/L0/run_deprecated/test_deprecated_warning.py
"""Tests for c++ MLP""" from itertools import product from time import time import torch from torch import nn from torch.testing._internal import common_utils from torch.testing._internal.common_device_type import instantiate_device_type_tests from torch.testing._internal.common_device_type import onlyCUDA from apex.mlp import MLP batch_size = 1024 mlp_sizes = [480, 1024, 1024, 512, 256, 1] num_iters = 10 # note(crcrpar): On Ampere, this test should be run without TF32 enabled. class TestMLP(common_utils.TestCase): def test_creation(self): MLP(mlp_sizes) def test_numeric(self): mlp = MLP(mlp_sizes).cuda() mlp_layers = [] for i in range(mlp.num_layers): linear = nn.Linear(mlp_sizes[i], mlp_sizes[i + 1]) with torch.no_grad(): mlp.weights[i].copy_(linear.weight) mlp.biases[i].copy_(linear.bias) mlp_layers.append(linear) mlp_layers.append(nn.ReLU()) ref_mlp = nn.Sequential(*mlp_layers).cuda() test_input = ( torch.empty(batch_size, mlp_sizes[0], device="cuda") .uniform_(-1.0, 1.0) .requires_grad_() ) ref_input = test_input.clone().detach().requires_grad_() mlp_out = mlp(test_input) ref_out = ref_mlp(ref_input) self.assertEqual(mlp_out, ref_out) # Use mean value as scalar loss. Multiply 10 to make it big enough not zero out mlp_out.mean().mul(10.0).backward() ref_out.mean().mul(10.0).backward() self.assertEqual(test_input.grad, ref_input.grad) self.assertEqual(mlp.biases[0].grad, ref_mlp[0].bias.grad) def _test_mlp_impl(self, use_activation: str, bias: bool, enable_autocast: bool): mlp = MLP(mlp_sizes, bias=bias, activation=use_activation).cuda() mlp_layers = [] for i in range(mlp.num_layers): linear = nn.Linear(mlp_sizes[i], mlp_sizes[i + 1], bias=bias) with torch.no_grad(): mlp.weights[i].copy_(linear.weight) if bias: mlp.biases[i].copy_(linear.bias) mlp_layers.append(linear) if use_activation == "relu": mlp_layers.append(nn.ReLU()) if use_activation == "sigmoid": mlp_layers.append(nn.Sigmoid()) ref_mlp = nn.Sequential(*mlp_layers).cuda() test_input = ( torch.empty(batch_size, mlp_sizes[0], device="cuda") .uniform_(-1.0, 1.0) .requires_grad_() ) ref_input = test_input.clone().detach().requires_grad_() with torch.cuda.amp.autocast_mode.autocast(enabled=enable_autocast): mlp_out = mlp(test_input) mlp_loss = mlp_out.mean().mul(10.0) # Use mean value as scalar loss. Multiply 10 to make it big enough not zero out ref_out = ref_mlp(ref_input) ref_loss = ref_out.mean().mul(10.0) mlp_loss.backward() ref_loss.backward() if enable_autocast: self.assertEqual(mlp_out.dtype, torch.float16) self.assertEqual(ref_out.dtype, torch.float16) else: self.assertEqual(mlp_out, ref_out) self.assertEqual(test_input.grad, ref_input.grad) self.assertEqual(mlp.weights[0].grad, ref_mlp[0].weight.grad) @common_utils.parametrize( "use_activation,bias", list(product(("none", "relu", "sigmoid"), (True, False))), ) def test_mlp(self, use_activation: str, bias: bool): self._test_mlp_impl(use_activation, bias, enable_autocast=False) @common_utils.parametrize( "use_activation,bias", list(product(("none", "relu", "sigmoid"), (True, False))), ) def test_mlp_autocast_fp16(self, use_activation: str, bias: bool): self._test_mlp_impl(use_activation, bias, enable_autocast=True) def test_no_grad(self): mlp = MLP(mlp_sizes).cuda() mlp_layers = [] for i in range(mlp.num_layers): linear = nn.Linear(mlp_sizes[i], mlp_sizes[i + 1]) with torch.no_grad(): mlp.weights[i].copy_(linear.weight) mlp.biases[i].copy_(linear.bias) mlp_layers.append(linear) mlp_layers.append(nn.ReLU(inplace=True)) ref_mlp = nn.Sequential(*mlp_layers).cuda() test_input = torch.empty(batch_size, mlp_sizes[0], device="cuda").uniform_(-1.0, 1.0) ref_input = test_input.clone().detach() mlp_out = mlp(test_input) ref_out = ref_mlp(ref_input) self.assertEqual(mlp_out, ref_out) # Use mean value as scalar loss. Multiply 10 to make it big enough not zero out mlp_out.mean().mul(10.0).backward() ref_out.mean().mul(10.0).backward() self.assertEqual(mlp.weights[0].grad, ref_mlp[0].weight.grad) def test_performance_half(self): mlp = MLP(mlp_sizes).cuda().half() mlp_layers = [] for i in range(mlp.num_layers): linear = nn.Linear(mlp_sizes[i], mlp_sizes[i + 1]) mlp.weights[i].data.copy_(linear.weight) mlp.biases[i].data.copy_(linear.bias) mlp_layers.append(linear) mlp_layers.append(nn.ReLU(inplace=True)) ref_mlp = nn.Sequential(*mlp_layers).cuda().half() test_input = ( torch.empty(batch_size, mlp_sizes[0], device="cuda", dtype=torch.half) .fill_(10.0) .requires_grad_() ) ref_input = ( torch.empty(batch_size, mlp_sizes[0], device="cuda", dtype=torch.half) .fill_(10.0) .requires_grad_() ) # Warm up GPU for _ in range(100): ref_out = ref_mlp(ref_input) ref_loss = ref_out.mean() ref_mlp.zero_grad() ref_loss.backward() mlp_out = mlp(test_input) test_loss = mlp_out.mean() mlp.zero_grad() test_loss.backward() torch.cuda.profiler.start() torch.cuda.synchronize() start_time = time() for _ in range(num_iters): ref_out = ref_mlp(ref_input) ref_loss = ref_out.mean() ref_mlp.zero_grad() ref_loss.backward() torch.cuda.synchronize() stop_time = time() ref_time = (stop_time - start_time) * 1000.0 / num_iters print(f"\nPytorch MLP time {ref_time:.4f} ms") torch.cuda.synchronize() start_time = time() for _ in range(num_iters): mlp_out = mlp(test_input) test_loss = mlp_out.mean() mlp.zero_grad() test_loss.backward() torch.cuda.synchronize() stop_time = time() actual_time = (stop_time - start_time) * 1000.0 / num_iters print(f"C++ MLP time {actual_time:.4f} ms") torch.cuda.profiler.stop() self.assertLessEqual( actual_time, ref_time, msg=f"Custom extension took {actual_time:.4f} while PyTorch took {ref_time:.4f}", ) instantiate_device_type_tests(TestMLP, globals(), only_for=("cuda",)) if __name__ == "__main__": common_utils.run_tests()
apex-master
tests/L0/run_mlp/test_mlp.py
import os import logging import itertools from typing import Optional, Tuple, List import unittest import torch from torch.testing._internal import common_utils from torch.testing._internal import common_cuda from torch.testing._internal import common_distributed from apex._autocast_utils import _get_autocast_dtypes from apex.transformer import parallel_state from apex.transformer.pipeline_parallel import utils as pp_utils from apex.transformer.pipeline_parallel.schedules.common import ( FwdStepFunc, build_model, _get_params_for_weight_decay_optimization, ) from apex.transformer.pipeline_parallel.schedules.fwd_bwd_no_pipelining import ( forward_backward_no_pipelining, ) from apex.transformer.pipeline_parallel.schedules.fwd_bwd_pipelining_with_interleaving import ( _forward_backward_pipelining_with_interleaving, ) from apex.transformer.pipeline_parallel.schedules.fwd_bwd_pipelining_without_interleaving import ( forward_backward_pipelining_without_interleaving, ) from apex.transformer.testing.distributed_test_base import UccDistributedTestBase from apex.transformer.testing import commons as testing_utils logging.getLogger("torch").setLevel(logging.WARNING) logging.getLogger("apex").setLevel(logging.WARNING) def _get_default_world_sizes_model_parallel_world_size(pipeline_model_parallel_world_size: Optional[int] = None ) -> Tuple[int, int, int]: # TODO: revisit if we can fold this into the class for skip logic / avoid duplication # of world size computation world_size = torch.cuda.device_count() tensor_model_parallel_world_size = 1 data_parallel_size = 1 + (world_size >= 8 and world_size % 2 == 0) if pipeline_model_parallel_world_size is None: pipeline_model_parallel_world_size = world_size // (tensor_model_parallel_world_size * data_parallel_size) else: data_parallel_size = world_size // (tensor_model_parallel_world_size * pipeline_model_parallel_world_size) return tensor_model_parallel_world_size, data_parallel_size, pipeline_model_parallel_world_size class UccPipelineParallelForwardBackwardProf(UccDistributedTestBase): # The purpose of this class is to test and confirm asynchronous communication via profiling. # Having that in mind, it is safe to skip all the numerical checks. # For unit testing with numerical checks please refer to `tests/L0/run_transformer/test_pipeline_parallel_fwd_bwd.py`. def __init__(self, *args, **kwargs) -> None: super().__init__(*args, **kwargs) self.GLOBAL_BATCH_SIZE = 1024 self.MICRO_BATCH_SIZE = 64 self.HIDDEN_SIZE = 256 self.NUM_FWD_BWD_ITERATIONS = 4 self.deallocate_options = (False,) self.dtypes = (torch.float32,) @property def world_size(self) -> int: return min(torch.cuda.device_count(), 8) def _forward_backward_test_impl( self, forward_only: bool, fwd_bwd_func: FwdStepFunc, pipeline_model_parallel_world_size: Optional[int], virtual_pipeline_model_parallel_size: Optional[int], async_comm: bool = False, *, default_backend: Optional[str] = None, p2p_backend: Optional[str] = None, ) -> None: if fwd_bwd_func == _forward_backward_pipelining_with_interleaving: self.assertIsNotNone(virtual_pipeline_model_parallel_size) self.assertGreater(virtual_pipeline_model_parallel_size, 1) dtype_options = self.dtypes or [torch.float32, torch.double] + _get_autocast_dtypes() for dtype, deallocate_pipeline_outputs in itertools.product( dtype_options, self.deallocate_options, ): grad_scaler = ( torch.cuda.amp.GradScaler(init_scale=4.0) if dtype == torch.half else None ) (tensor_model_parallel_world_size, data_parallel_size, pipeline_model_parallel_world_size) = _get_default_world_sizes_model_parallel_world_size(pipeline_model_parallel_world_size) parallel_state.initialize_model_parallel( tensor_model_parallel_size_=tensor_model_parallel_world_size, pipeline_model_parallel_size_=pipeline_model_parallel_world_size, virtual_pipeline_model_parallel_size_=virtual_pipeline_model_parallel_size, default_backend=default_backend, p2p_backend=p2p_backend, ) pp_utils._reconfigure_microbatch_calculator( rank=parallel_state.get_tensor_model_parallel_rank(), rampup_batch_size=None, global_batch_size=self.GLOBAL_BATCH_SIZE, micro_batch_size=self.MICRO_BATCH_SIZE, data_parallel_size=parallel_state.get_data_parallel_world_size(), ) global_batch_shape = ( self.GLOBAL_BATCH_SIZE // parallel_state.get_data_parallel_world_size(), self.HIDDEN_SIZE, self.HIDDEN_SIZE, ) batch = None if parallel_state.is_pipeline_first_stage(): batch = (torch.ones(global_batch_shape, dtype=dtype).cuda(), ) model = build_model( testing_utils.model_provider_func, # Use DDP only when it's better to have wrap_with_ddp=data_parallel_size > 1, virtual_pipeline_model_parallel_size=virtual_pipeline_model_parallel_size, hidden_size=self.HIDDEN_SIZE, ) offset = pipeline_model_parallel_world_size if virtual_pipeline_model_parallel_size is not None else 0 for idx, model_module in enumerate(model): model_module = model_module.to(dtype) _param_groups = _get_params_for_weight_decay_optimization(model) optimizer = torch.optim.Adam(_param_groups, lr=1e-3) pp_utils.update_num_microbatches(0) for _ in range(self.NUM_FWD_BWD_ITERATIONS): loss = fwd_bwd_func( testing_utils.fwd_step_func, batch, model, forward_only=forward_only, # `tensor_shape` is the shape of micro batch. tensor_shape=( self.MICRO_BATCH_SIZE, self.HIDDEN_SIZE, self.HIDDEN_SIZE, ), dtype=dtype, async_comm=async_comm, grad_scaler=grad_scaler, deallocate_pipeline_output=deallocate_pipeline_outputs, ) parallel_state.destroy_model_parallel() def test_learning_no_pipelining(self): self._forward_backward_test_impl(False, forward_backward_no_pipelining, 1, None) def test_inference_no_pipelining(self): self._forward_backward_test_impl(True, forward_backward_no_pipelining, 1, None) def test_learning_pipelining_without_interleaving(self): self._forward_backward_test_impl( False, forward_backward_pipelining_without_interleaving, None, None ) def test_inference_pipelining_without_interleaving(self): self._forward_backward_test_impl( True, forward_backward_pipelining_without_interleaving, None, None ) def test_learning_async_pipelining_without_interleaving(self): self._forward_backward_test_impl( False, forward_backward_pipelining_without_interleaving, None, None, async_comm=True ) def test_inference_async_pipelining_without_interleaving(self): self._forward_backward_test_impl( True, forward_backward_pipelining_without_interleaving, None, None, async_comm=True ) @unittest.skipUnless(_get_default_world_sizes_model_parallel_world_size()[-1] > 2, "Interleaved schedule requires pipeline_model_parallel_world_size > 2") def test_learning_pipelining_with_interleaving(self): self._forward_backward_test_impl( False, _forward_backward_pipelining_with_interleaving, None, virtual_pipeline_model_parallel_size=2 ) @unittest.skipUnless(_get_default_world_sizes_model_parallel_world_size()[-1] > 2, "Interleaved schedule requires pipeline_model_parallel_world_size > 2") def test_inference_pipelining_with_interleaving(self): self._forward_backward_test_impl( True, _forward_backward_pipelining_with_interleaving, None, virtual_pipeline_model_parallel_size=2 ) @unittest.skipUnless(_get_default_world_sizes_model_parallel_world_size()[-1] > 2, "Interleaved schedule requires pipeline_model_parallel_world_size > 2") def test_learning_async_pipelining_with_interleaving(self): self._forward_backward_test_impl( False, _forward_backward_pipelining_with_interleaving, None, virtual_pipeline_model_parallel_size=2, async_comm=True ) @unittest.skipUnless(_get_default_world_sizes_model_parallel_world_size()[-1] > 2, "Interleaved schedule requires pipeline_model_parallel_world_size > 2") def test_inference_async_pipelining_with_interleaving(self): self._forward_backward_test_impl( True, _forward_backward_pipelining_with_interleaving, None, virtual_pipeline_model_parallel_size=2, async_comm=True ) if __name__ == "__main__": os.environ["UCC_TLS"] = "ucp,cuda" common_distributed.TIMEOUT_DEFAULT = 500 common_utils.run_tests()
apex-master
tests/L1/transformer/pipeline_parallel_fwd_bwd_ucc_async.py
import argparse import os import shutil import time import torch import torch.nn as nn import torch.nn.parallel import torch.backends.cudnn as cudnn import torch.distributed as dist import torch.optim import torch.utils.data import torch.utils.data.distributed import torchvision.transforms as transforms import torchvision.datasets as datasets import torchvision.models as models import numpy as np try: from apex.parallel import DistributedDataParallel as DDP from apex.fp16_utils import * from apex import amp, optimizers from apex.multi_tensor_apply import multi_tensor_applier except ImportError: raise ImportError("Please install apex from https://www.github.com/nvidia/apex to run this example.") model_names = sorted(name for name in models.__dict__ if name.islower() and not name.startswith("__") and callable(models.__dict__[name])) parser = argparse.ArgumentParser(description='PyTorch ImageNet Training') parser.add_argument('data', metavar='DIR', help='path to dataset') parser.add_argument('--arch', '-a', metavar='ARCH', default='resnet18', choices=model_names, help='model architecture: ' + ' | '.join(model_names) + ' (default: resnet18)') parser.add_argument('-j', '--workers', default=4, type=int, metavar='N', help='number of data loading workers (default: 4)') parser.add_argument('--epochs', default=90, type=int, metavar='N', help='number of total epochs to run') parser.add_argument('--start-epoch', default=0, type=int, metavar='N', help='manual epoch number (useful on restarts)') parser.add_argument('-b', '--batch-size', default=256, type=int, metavar='N', help='mini-batch size per process (default: 256)') parser.add_argument('--lr', '--learning-rate', default=0.1, type=float, metavar='LR', help='Initial learning rate. Will be scaled by <global batch size>/256: args.lr = args.lr*float(args.batch_size*args.world_size)/256. A warmup schedule will also be applied over the first 5 epochs.') parser.add_argument('--momentum', default=0.9, type=float, metavar='M', help='momentum') parser.add_argument('--weight-decay', '--wd', default=1e-4, type=float, metavar='W', help='weight decay (default: 1e-4)') parser.add_argument('--print-freq', '-p', default=10, type=int, metavar='N', help='print frequency (default: 10)') parser.add_argument('--resume', default='', type=str, metavar='PATH', help='path to latest checkpoint (default: none)') parser.add_argument('-e', '--evaluate', dest='evaluate', action='store_true', help='evaluate model on validation set') parser.add_argument('--pretrained', dest='pretrained', action='store_true', help='use pre-trained model') parser.add_argument('--prof', dest='prof', action='store_true', help='Only run 10 iterations for profiling.') parser.add_argument('--deterministic', action='store_true') parser.add_argument("--local_rank", default=0, type=int) parser.add_argument('--sync_bn', action='store_true', help='enabling apex sync BN.') parser.add_argument('--has-ext', action='store_true') parser.add_argument('--opt-level', type=str) parser.add_argument('--keep-batchnorm-fp32', type=str, default=None) parser.add_argument('--loss-scale', type=str, default=None) parser.add_argument('--fused-adam', action='store_true') parser.add_argument('--prints-to-process', type=int, default=10) cudnn.benchmark = True def fast_collate(batch): imgs = [img[0] for img in batch] targets = torch.tensor([target[1] for target in batch], dtype=torch.int64) w = imgs[0].size[0] h = imgs[0].size[1] tensor = torch.zeros( (len(imgs), 3, h, w), dtype=torch.uint8 ) for i, img in enumerate(imgs): nump_array = np.asarray(img, dtype=np.uint8) if(nump_array.ndim < 3): nump_array = np.expand_dims(nump_array, axis=-1) nump_array = np.rollaxis(nump_array, 2) tensor[i] += torch.from_numpy(nump_array) return tensor, targets best_prec1 = 0 args = parser.parse_args() # Let multi_tensor_applier be the canary in the coalmine # that verifies if the backend is what we think it is assert multi_tensor_applier.available == args.has_ext print("opt_level = {}".format(args.opt_level)) print("keep_batchnorm_fp32 = {}".format(args.keep_batchnorm_fp32), type(args.keep_batchnorm_fp32)) print("loss_scale = {}".format(args.loss_scale), type(args.loss_scale)) print("\nCUDNN VERSION: {}\n".format(torch.backends.cudnn.version())) if args.deterministic: cudnn.benchmark = False cudnn.deterministic = True torch.manual_seed(args.local_rank) torch.set_printoptions(precision=10) def main(): global best_prec1, args args.distributed = False if 'WORLD_SIZE' in os.environ: args.distributed = int(os.environ['WORLD_SIZE']) > 1 args.gpu = 0 args.world_size = 1 if args.distributed: args.gpu = args.local_rank % torch.cuda.device_count() torch.cuda.set_device(args.gpu) torch.distributed.init_process_group(backend='nccl', init_method='env://') args.world_size = torch.distributed.get_world_size() assert torch.backends.cudnn.enabled, "Amp requires cudnn backend to be enabled." # create model if args.pretrained: print("=> using pre-trained model '{}'".format(args.arch)) model = models.__dict__[args.arch](pretrained=True) else: print("=> creating model '{}'".format(args.arch)) model = models.__dict__[args.arch]() if args.sync_bn: import apex print("using apex synced BN") model = apex.parallel.convert_syncbn_model(model) model = model.cuda() # Scale learning rate based on global batch size args.lr = args.lr*float(args.batch_size*args.world_size)/256. if args.fused_adam: optimizer = optimizers.FusedAdam(model.parameters()) else: optimizer = torch.optim.SGD(model.parameters(), args.lr, momentum=args.momentum, weight_decay=args.weight_decay) model, optimizer = amp.initialize( model, optimizer, # enabled=False, opt_level=args.opt_level, keep_batchnorm_fp32=args.keep_batchnorm_fp32, loss_scale=args.loss_scale ) if args.distributed: # By default, apex.parallel.DistributedDataParallel overlaps communication with # computation in the backward pass. # model = DDP(model) # delay_allreduce delays all communication to the end of the backward pass. model = DDP(model, delay_allreduce=True) # define loss function (criterion) and optimizer criterion = nn.CrossEntropyLoss().cuda() # Optionally resume from a checkpoint if args.resume: # Use a local scope to avoid dangling references def resume(): if os.path.isfile(args.resume): print("=> loading checkpoint '{}'".format(args.resume)) checkpoint = torch.load(args.resume, map_location = lambda storage, loc: storage.cuda(args.gpu)) args.start_epoch = checkpoint['epoch'] best_prec1 = checkpoint['best_prec1'] model.load_state_dict(checkpoint['state_dict']) optimizer.load_state_dict(checkpoint['optimizer']) print("=> loaded checkpoint '{}' (epoch {})" .format(args.resume, checkpoint['epoch'])) else: print("=> no checkpoint found at '{}'".format(args.resume)) resume() # Data loading code traindir = os.path.join(args.data, 'train') valdir = os.path.join(args.data, 'val') if(args.arch == "inception_v3"): crop_size = 299 val_size = 320 # I chose this value arbitrarily, we can adjust. else: crop_size = 224 val_size = 256 train_dataset = datasets.ImageFolder( traindir, transforms.Compose([ transforms.RandomResizedCrop(crop_size), transforms.RandomHorizontalFlip(), # transforms.ToTensor(), Too slow # normalize, ])) val_dataset = datasets.ImageFolder(valdir, transforms.Compose([ transforms.Resize(val_size), transforms.CenterCrop(crop_size), ])) train_sampler = None val_sampler = None if args.distributed: train_sampler = torch.utils.data.distributed.DistributedSampler(train_dataset) val_sampler = torch.utils.data.distributed.DistributedSampler(val_dataset) train_loader = torch.utils.data.DataLoader( train_dataset, batch_size=args.batch_size, shuffle=(train_sampler is None), num_workers=args.workers, pin_memory=True, sampler=train_sampler, collate_fn=fast_collate) val_loader = torch.utils.data.DataLoader( val_dataset, batch_size=args.batch_size, shuffle=False, num_workers=args.workers, pin_memory=True, sampler=val_sampler, collate_fn=fast_collate) if args.evaluate: validate(val_loader, model, criterion) return for epoch in range(args.start_epoch, args.epochs): if args.distributed: train_sampler.set_epoch(epoch) # train for one epoch train(train_loader, model, criterion, optimizer, epoch) if args.prof: break # evaluate on validation set prec1 = validate(val_loader, model, criterion) # remember best prec@1 and save checkpoint if args.local_rank == 0: is_best = prec1 > best_prec1 best_prec1 = max(prec1, best_prec1) save_checkpoint({ 'epoch': epoch + 1, 'arch': args.arch, 'state_dict': model.state_dict(), 'best_prec1': best_prec1, 'optimizer' : optimizer.state_dict(), }, is_best) class data_prefetcher(): def __init__(self, loader): self.loader = iter(loader) self.stream = torch.cuda.Stream() self.mean = torch.tensor([0.485 * 255, 0.456 * 255, 0.406 * 255]).cuda().view(1,3,1,1) self.std = torch.tensor([0.229 * 255, 0.224 * 255, 0.225 * 255]).cuda().view(1,3,1,1) # With Amp, it isn't necessary to manually convert data to half. # if args.fp16: # self.mean = self.mean.half() # self.std = self.std.half() self.preload() def preload(self): try: self.next_input, self.next_target = next(self.loader) except StopIteration: self.next_input = None self.next_target = None return with torch.cuda.stream(self.stream): self.next_input = self.next_input.cuda(non_blocking=True) self.next_target = self.next_target.cuda(non_blocking=True) # With Amp, it isn't necessary to manually convert data to half. # if args.fp16: # self.next_input = self.next_input.half() # else: self.next_input = self.next_input.float() self.next_input = self.next_input.sub_(self.mean).div_(self.std) def next(self): torch.cuda.current_stream().wait_stream(self.stream) input = self.next_input target = self.next_target self.preload() return input, target def train(train_loader, model, criterion, optimizer, epoch): batch_time = AverageMeter() data_time = AverageMeter() losses = AverageMeter() top1 = AverageMeter() top5 = AverageMeter() # switch to train mode model.train() end = time.time() run_info_dict = {"Iteration" : [], "Loss" : [], "Speed" : []} prefetcher = data_prefetcher(train_loader) input, target = prefetcher.next() i = -1 while input is not None: i += 1 # No learning rate warmup for this test, to expose bitwise inaccuracies more quickly # adjust_learning_rate(optimizer, epoch, i, len(train_loader)) if args.prof: if i > 10: break # measure data loading time data_time.update(time.time() - end) # compute output output = model(input) loss = criterion(output, target) # measure accuracy and record loss prec1, prec5 = accuracy(output.data, target, topk=(1, 5)) if args.distributed: reduced_loss = reduce_tensor(loss.data) prec1 = reduce_tensor(prec1) prec5 = reduce_tensor(prec5) else: reduced_loss = loss.data losses.update(to_python_float(reduced_loss), input.size(0)) top1.update(to_python_float(prec1), input.size(0)) top5.update(to_python_float(prec5), input.size(0)) # compute gradient and do SGD step optimizer.zero_grad() with amp.scale_loss(loss, optimizer) as scaled_loss: scaled_loss.backward() # for param in model.parameters(): # print(param.data.double().sum().item(), param.grad.data.double().sum().item()) # torch.cuda.synchronize() torch.cuda.nvtx.range_push("step") optimizer.step() torch.cuda.nvtx.range_pop() torch.cuda.synchronize() # measure elapsed time batch_time.update(time.time() - end) end = time.time() # If you decide to refactor this test, like examples/imagenet, to sample the loss every # print_freq iterations, make sure to move this prefetching below the accuracy calculation. input, target = prefetcher.next() if i % args.print_freq == 0 and i > 1: if args.local_rank == 0: print('Epoch: [{0}][{1}/{2}]\t' 'Time {batch_time.val:.3f} ({batch_time.avg:.3f})\t' 'Speed {3:.3f} ({4:.3f})\t' 'Data {data_time.val:.3f} ({data_time.avg:.3f})\t' 'Loss {loss.val:.10f} ({loss.avg:.4f})\t' 'Prec@1 {top1.val:.3f} ({top1.avg:.3f})\t' 'Prec@5 {top5.val:.3f} ({top5.avg:.3f})'.format( epoch, i, len(train_loader), args.world_size * args.batch_size / batch_time.val, args.world_size * args.batch_size / batch_time.avg, batch_time=batch_time, data_time=data_time, loss=losses, top1=top1, top5=top5)) run_info_dict["Iteration"].append(i) run_info_dict["Loss"].append(losses.val) run_info_dict["Speed"].append(args.world_size * args.batch_size / batch_time.val) if len(run_info_dict["Loss"]) == args.prints_to_process: if args.local_rank == 0: torch.save(run_info_dict, str(args.has_ext) + "_" + str(args.opt_level) + "_" + str(args.loss_scale) + "_" + str(args.keep_batchnorm_fp32) + "_" + str(args.fused_adam)) quit() def validate(val_loader, model, criterion): batch_time = AverageMeter() losses = AverageMeter() top1 = AverageMeter() top5 = AverageMeter() # switch to evaluate mode model.eval() end = time.time() prefetcher = data_prefetcher(val_loader) input, target = prefetcher.next() i = -1 while input is not None: i += 1 # compute output with torch.no_grad(): output = model(input) loss = criterion(output, target) # measure accuracy and record loss prec1, prec5 = accuracy(output.data, target, topk=(1, 5)) if args.distributed: reduced_loss = reduce_tensor(loss.data) prec1 = reduce_tensor(prec1) prec5 = reduce_tensor(prec5) else: reduced_loss = loss.data losses.update(to_python_float(reduced_loss), input.size(0)) top1.update(to_python_float(prec1), input.size(0)) top5.update(to_python_float(prec5), input.size(0)) # measure elapsed time batch_time.update(time.time() - end) end = time.time() if args.local_rank == 0 and i % args.print_freq == 0: print('Test: [{0}/{1}]\t' 'Time {batch_time.val:.3f} ({batch_time.avg:.3f})\t' 'Speed {2:.3f} ({3:.3f})\t' 'Loss {loss.val:.4f} ({loss.avg:.4f})\t' 'Prec@1 {top1.val:.3f} ({top1.avg:.3f})\t' 'Prec@5 {top5.val:.3f} ({top5.avg:.3f})'.format( i, len(val_loader), args.world_size * args.batch_size / batch_time.val, args.world_size * args.batch_size / batch_time.avg, batch_time=batch_time, loss=losses, top1=top1, top5=top5)) input, target = prefetcher.next() print(' * Prec@1 {top1.avg:.3f} Prec@5 {top5.avg:.3f}' .format(top1=top1, top5=top5)) return top1.avg def save_checkpoint(state, is_best, filename='checkpoint.pth.tar'): torch.save(state, filename) if is_best: shutil.copyfile(filename, 'model_best.pth.tar') class AverageMeter(object): """Computes and stores the average and current value""" def __init__(self): self.reset() def reset(self): self.val = 0 self.avg = 0 self.sum = 0 self.count = 0 def update(self, val, n=1): self.val = val self.sum += val * n self.count += n self.avg = self.sum / self.count def adjust_learning_rate(optimizer, epoch, step, len_epoch): """LR schedule that should yield 76% converged accuracy with batch size 256""" factor = epoch // 30 if epoch >= 80: factor = factor + 1 lr = args.lr*(0.1**factor) """Warmup""" if epoch < 5: lr = lr*float(1 + step + epoch*len_epoch)/(5.*len_epoch) # if(args.local_rank == 0): # print("epoch = {}, step = {}, lr = {}".format(epoch, step, lr)) for param_group in optimizer.param_groups: param_group['lr'] = lr def accuracy(output, target, topk=(1,)): """Computes the precision@k for the specified values of k""" maxk = max(topk) batch_size = target.size(0) _, pred = output.topk(maxk, 1, True, True) pred = pred.t() correct = pred.eq(target.view(1, -1).expand_as(pred)) res = [] for k in topk: correct_k = correct[:k].view(-1).float().sum(0, keepdim=True) res.append(correct_k.mul_(100.0 / batch_size)) return res def reduce_tensor(tensor): rt = tensor.clone() dist.all_reduce(rt, op=dist.reduce_op.SUM) rt /= args.world_size return rt if __name__ == '__main__': main()
apex-master
tests/L1/common/main_amp.py
import argparse import torch parser = argparse.ArgumentParser(description='Compare') parser.add_argument('--opt-level', type=str) parser.add_argument('--keep-batchnorm-fp32', type=str, default=None) parser.add_argument('--loss-scale', type=str, default=None) parser.add_argument('--fused-adam', action='store_true') parser.add_argument('--use_baseline', action='store_true') args = parser.parse_args() base_file = str(args.opt_level) + "_" +\ str(args.loss_scale) + "_" +\ str(args.keep_batchnorm_fp32) + "_" +\ str(args.fused_adam) file_e = "True_" + base_file file_p = "False_" + base_file if args.use_baseline: file_b = "baselines/True_" + base_file dict_e = torch.load(file_e) dict_p = torch.load(file_p) if args.use_baseline: dict_b = torch.load(file_b) torch.set_printoptions(precision=10) print(file_e) print(file_p) if args.use_baseline: print(file_b) # ugly duplication here... if not args.use_baseline: for n, (i_e, i_p) in enumerate(zip(dict_e["Iteration"], dict_p["Iteration"])): assert i_e == i_p, "i_e = {}, i_p = {}".format(i_e, i_p) loss_e = dict_e["Loss"][n] loss_p = dict_p["Loss"][n] assert loss_e == loss_p, "Iteration {}, loss_e = {}, loss_p = {}".format(i_e, loss_e, loss_p) print("{:4} {:15.10f} {:15.10f} {:15.10f} {:15.10f}".format( i_e, loss_e, loss_p, dict_e["Speed"][n], dict_p["Speed"][n])) else: for n, (i_e, i_p) in enumerate(zip(dict_e["Iteration"], dict_p["Iteration"])): assert i_e == i_p, "i_e = {}, i_p = {}".format(i_e, i_p) loss_e = dict_e["Loss"][n] loss_p = dict_p["Loss"][n] loss_b = dict_b["Loss"][n] assert loss_e == loss_p, "Iteration {}, loss_e = {}, loss_p = {}".format(i_e, loss_e, loss_p) assert loss_e == loss_b, "Iteration {}, loss_e = {}, loss_b = {}".format(i_e, loss_e, loss_b) print("{:4} {:15.10f} {:15.10f} {:15.10f} {:15.10f} {:15.10f} {:15.10f}".format( i_e, loss_b, loss_e, loss_p, dict_b["Speed"][n], dict_e["Speed"][n], dict_p["Speed"][n]))
apex-master
tests/L1/common/compare.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- # # PyTorch documentation build configuration file, created by # sphinx-quickstart on Fri Dec 23 13:31:47 2016. # # This file is execfile()d with the current directory set to its # containing dir. # # Note that not all possible configuration values are present in this # autogenerated file. # # All configuration values have a default; values that are commented out # serve to show the default. # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the # documentation root, use os.path.abspath to make it absolute, like shown here. # import os import sys sys.path.insert(0, os.path.abspath('.')) # sys.path.insert(0, os.path.abspath('../../apex/parallel/')) import apex # import multiproc import sphinx_rtd_theme # -- General configuration ------------------------------------------------ # If your documentation needs a minimal Sphinx version, state it here. # # needs_sphinx = '1.0' # Add any Sphinx extension module names here, as strings. They can be # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom # ones. extensions = [ 'sphinx.ext.autodoc', 'sphinx.ext.autosummary', 'sphinx.ext.doctest', 'sphinx.ext.intersphinx', 'sphinx.ext.todo', 'sphinx.ext.coverage', 'sphinx.ext.mathjax', 'sphinx.ext.napoleon', 'sphinx.ext.viewcode', 'sphinx.ext.extlinks', ] napoleon_use_ivar = True # Add any paths that contain templates here, relative to this directory. templates_path = ['_templates'] # The suffix(es) of source filenames. # You can specify multiple suffix as a list of string: # # source_suffix = ['.rst', '.md'] source_suffix = '.rst' # The master toctree document. master_doc = 'index' # General information about the project. project = 'Apex' copyright = '2018' author = 'Christian Sarofeen, Natalia Gimelshein, Michael Carilli, Raul Puri' # The version info for the project you're documenting, acts as replacement for # |version| and |release|, also used in various other places throughout the # built documents. # # The short X.Y version. # TODO: change to [:2] at v1.0 # version = 'master (' + torch.__version__ + ' )' version = '0.1' # The full version, including alpha/beta/rc tags. # TODO: verify this works as expected release = '0.1.0' # The language for content autogenerated by Sphinx. Refer to documentation # for a list of supported languages. # # This is also used if you do content translation via gettext catalogs. # Usually you set "language" from the command line for these cases. language = None # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. # This pattern also affects html_static_path and html_extra_path exclude_patterns = [] # The name of the Pygments (syntax highlighting) style to use. pygments_style = 'sphinx' # If true, `todo` and `todoList` produce output, else they produce nothing. todo_include_todos = True # -- Options for HTML output ------------------------------------------------- # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. # html_theme = 'sphinx_rtd_theme' html_theme_path = [sphinx_rtd_theme.get_html_theme_path()] # Theme options are theme-specific and customize the look and feel of a theme # further. For a list of options available for each theme, see the # documentation. # html_theme_options = { 'collapse_navigation': False, 'display_version': True, 'logo_only': True, } # html_logo = '_static/img/nv-pytorch2.png' # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". html_static_path = ['_static'] # html_style_path = 'css/pytorch_theme.css' html_context = { 'css_files': [ 'https://fonts.googleapis.com/css?family=Lato', '_static/css/pytorch_theme.css' ], } # -- Options for HTMLHelp output --------------------------------------------- # Output file base name for HTML help builder. htmlhelp_basename = 'PyTorchdoc' # -- Options for LaTeX output ------------------------------------------------ latex_elements = { # The paper size ('letterpaper' or 'a4paper'). # # 'papersize': 'letterpaper', # The font size ('10pt', '11pt' or '12pt'). # # 'pointsize': '10pt', # Additional stuff for the LaTeX preamble. # # 'preamble': '', # Latex figure (float) alignment # # 'figure_align': 'htbp', } # Grouping the document tree into LaTeX files. List of tuples # (source start file, target name, title, # author, documentclass [howto, manual, or own class]). latex_documents = [ (master_doc, 'apex.tex', 'Apex Documentation', 'Torch Contributors', 'manual'), ] # -- Options for manual page output ------------------------------------------ # One entry per manual page. List of tuples # (source start file, name, description, authors, manual section). man_pages = [ (master_doc, 'Apex', 'Apex Documentation', [author], 1) ] # -- Options for Texinfo output ---------------------------------------------- # Grouping the document tree into Texinfo files. List of tuples # (source start file, target name, title, author, # dir menu entry, description, category) texinfo_documents = [ (master_doc, 'Apex', 'Apex Documentation', author, 'Apex', 'One line description of project.', 'Miscellaneous'), ] # Example configuration for intersphinx: refer to the Python standard library. intersphinx_mapping = { 'python': ('https://docs.python.org/', None), 'numpy': ('http://docs.scipy.org/doc/numpy/', None), } # -- A patch that prevents Sphinx from cross-referencing ivar tags ------- # See http://stackoverflow.com/a/41184353/3343043 from docutils import nodes from sphinx.util.docfields import TypedField from sphinx import addnodes def patched_make_field(self, types, domain, items, **kw): # `kw` catches `env=None` needed for newer sphinx while maintaining # backwards compatibility when passed along further down! # type: (List, unicode, Tuple) -> nodes.field def handle_item(fieldarg, content): par = nodes.paragraph() par += addnodes.literal_strong('', fieldarg) # Patch: this line added # par.extend(self.make_xrefs(self.rolename, domain, fieldarg, # addnodes.literal_strong)) if fieldarg in types: par += nodes.Text(' (') # NOTE: using .pop() here to prevent a single type node to be # inserted twice into the doctree, which leads to # inconsistencies later when references are resolved fieldtype = types.pop(fieldarg) if len(fieldtype) == 1 and isinstance(fieldtype[0], nodes.Text): typename = u''.join(n.astext() for n in fieldtype) typename = typename.replace('int', 'python:int') typename = typename.replace('long', 'python:long') typename = typename.replace('float', 'python:float') typename = typename.replace('type', 'python:type') par.extend(self.make_xrefs(self.typerolename, domain, typename, addnodes.literal_emphasis, **kw)) else: par += fieldtype par += nodes.Text(')') par += nodes.Text(' -- ') par += content return par fieldname = nodes.field_name('', self.label) if len(items) == 1 and self.can_collapse: fieldarg, content = items[0] bodynode = handle_item(fieldarg, content) else: bodynode = self.list_type() for fieldarg, content in items: bodynode += nodes.list_item('', handle_item(fieldarg, content)) fieldbody = nodes.field_body('', bodynode) return nodes.field('', fieldname, fieldbody) TypedField.make_field = patched_make_field
apex-master
docs/source/conf.py
from __future__ import print_function import argparse import os import random import torch import torch.nn as nn import torch.nn.parallel import torch.backends.cudnn as cudnn import torch.optim as optim import torch.utils.data import torchvision.datasets as dset import torchvision.transforms as transforms import torchvision.utils as vutils try: from apex import amp except ImportError: raise ImportError("Please install apex from https://www.github.com/nvidia/apex to run this example.") parser = argparse.ArgumentParser() parser.add_argument('--dataset', default='cifar10', help='cifar10 | lsun | mnist |imagenet | folder | lfw | fake') parser.add_argument('--dataroot', default='./', help='path to dataset') parser.add_argument('--workers', type=int, help='number of data loading workers', default=2) parser.add_argument('--batchSize', type=int, default=64, help='input batch size') parser.add_argument('--imageSize', type=int, default=64, help='the height / width of the input image to network') parser.add_argument('--nz', type=int, default=100, help='size of the latent z vector') parser.add_argument('--ngf', type=int, default=64) parser.add_argument('--ndf', type=int, default=64) parser.add_argument('--niter', type=int, default=25, help='number of epochs to train for') parser.add_argument('--lr', type=float, default=0.0002, help='learning rate, default=0.0002') parser.add_argument('--beta1', type=float, default=0.5, help='beta1 for adam. default=0.5') parser.add_argument('--ngpu', type=int, default=1, help='number of GPUs to use') parser.add_argument('--netG', default='', help="path to netG (to continue training)") parser.add_argument('--netD', default='', help="path to netD (to continue training)") parser.add_argument('--outf', default='.', help='folder to output images and model checkpoints') parser.add_argument('--manualSeed', type=int, help='manual seed') parser.add_argument('--classes', default='bedroom', help='comma separated list of classes for the lsun data set') parser.add_argument('--opt_level', default='O1', help='amp opt_level, default="O1"') opt = parser.parse_args() print(opt) try: os.makedirs(opt.outf) except OSError: pass if opt.manualSeed is None: opt.manualSeed = 2809 print("Random Seed: ", opt.manualSeed) random.seed(opt.manualSeed) torch.manual_seed(opt.manualSeed) cudnn.benchmark = True if opt.dataset in ['imagenet', 'folder', 'lfw']: # folder dataset dataset = dset.ImageFolder(root=opt.dataroot, transform=transforms.Compose([ transforms.Resize(opt.imageSize), transforms.CenterCrop(opt.imageSize), transforms.ToTensor(), transforms.Normalize((0.5, 0.5, 0.5), (0.5, 0.5, 0.5)), ])) nc=3 elif opt.dataset == 'lsun': classes = [ c + '_train' for c in opt.classes.split(',')] dataset = dset.LSUN(root=opt.dataroot, classes=classes, transform=transforms.Compose([ transforms.Resize(opt.imageSize), transforms.CenterCrop(opt.imageSize), transforms.ToTensor(), transforms.Normalize((0.5, 0.5, 0.5), (0.5, 0.5, 0.5)), ])) nc=3 elif opt.dataset == 'cifar10': dataset = dset.CIFAR10(root=opt.dataroot, download=True, transform=transforms.Compose([ transforms.Resize(opt.imageSize), transforms.ToTensor(), transforms.Normalize((0.5, 0.5, 0.5), (0.5, 0.5, 0.5)), ])) nc=3 elif opt.dataset == 'mnist': dataset = dset.MNIST(root=opt.dataroot, download=True, transform=transforms.Compose([ transforms.Resize(opt.imageSize), transforms.ToTensor(), transforms.Normalize((0.5,), (0.5,)), ])) nc=1 elif opt.dataset == 'fake': dataset = dset.FakeData(image_size=(3, opt.imageSize, opt.imageSize), transform=transforms.ToTensor()) nc=3 assert dataset dataloader = torch.utils.data.DataLoader(dataset, batch_size=opt.batchSize, shuffle=True, num_workers=int(opt.workers)) device = torch.device("cuda:0") ngpu = int(opt.ngpu) nz = int(opt.nz) ngf = int(opt.ngf) ndf = int(opt.ndf) # custom weights initialization called on netG and netD def weights_init(m): classname = m.__class__.__name__ if classname.find('Conv') != -1: m.weight.data.normal_(0.0, 0.02) elif classname.find('BatchNorm') != -1: m.weight.data.normal_(1.0, 0.02) m.bias.data.fill_(0) class Generator(nn.Module): def __init__(self, ngpu): super(Generator, self).__init__() self.ngpu = ngpu self.main = nn.Sequential( # input is Z, going into a convolution nn.ConvTranspose2d( nz, ngf * 8, 4, 1, 0, bias=False), nn.BatchNorm2d(ngf * 8), nn.ReLU(True), # state size. (ngf*8) x 4 x 4 nn.ConvTranspose2d(ngf * 8, ngf * 4, 4, 2, 1, bias=False), nn.BatchNorm2d(ngf * 4), nn.ReLU(True), # state size. (ngf*4) x 8 x 8 nn.ConvTranspose2d(ngf * 4, ngf * 2, 4, 2, 1, bias=False), nn.BatchNorm2d(ngf * 2), nn.ReLU(True), # state size. (ngf*2) x 16 x 16 nn.ConvTranspose2d(ngf * 2, ngf, 4, 2, 1, bias=False), nn.BatchNorm2d(ngf), nn.ReLU(True), # state size. (ngf) x 32 x 32 nn.ConvTranspose2d( ngf, nc, 4, 2, 1, bias=False), nn.Tanh() # state size. (nc) x 64 x 64 ) def forward(self, input): if input.is_cuda and self.ngpu > 1: output = nn.parallel.data_parallel(self.main, input, range(self.ngpu)) else: output = self.main(input) return output netG = Generator(ngpu).to(device) netG.apply(weights_init) if opt.netG != '': netG.load_state_dict(torch.load(opt.netG)) print(netG) class Discriminator(nn.Module): def __init__(self, ngpu): super(Discriminator, self).__init__() self.ngpu = ngpu self.main = nn.Sequential( # input is (nc) x 64 x 64 nn.Conv2d(nc, ndf, 4, 2, 1, bias=False), nn.LeakyReLU(0.2, inplace=True), # state size. (ndf) x 32 x 32 nn.Conv2d(ndf, ndf * 2, 4, 2, 1, bias=False), nn.BatchNorm2d(ndf * 2), nn.LeakyReLU(0.2, inplace=True), # state size. (ndf*2) x 16 x 16 nn.Conv2d(ndf * 2, ndf * 4, 4, 2, 1, bias=False), nn.BatchNorm2d(ndf * 4), nn.LeakyReLU(0.2, inplace=True), # state size. (ndf*4) x 8 x 8 nn.Conv2d(ndf * 4, ndf * 8, 4, 2, 1, bias=False), nn.BatchNorm2d(ndf * 8), nn.LeakyReLU(0.2, inplace=True), # state size. (ndf*8) x 4 x 4 nn.Conv2d(ndf * 8, 1, 4, 1, 0, bias=False), ) def forward(self, input): if input.is_cuda and self.ngpu > 1: output = nn.parallel.data_parallel(self.main, input, range(self.ngpu)) else: output = self.main(input) return output.view(-1, 1).squeeze(1) netD = Discriminator(ngpu).to(device) netD.apply(weights_init) if opt.netD != '': netD.load_state_dict(torch.load(opt.netD)) print(netD) criterion = nn.BCEWithLogitsLoss() fixed_noise = torch.randn(opt.batchSize, nz, 1, 1, device=device) real_label = 1 fake_label = 0 # setup optimizer optimizerD = optim.Adam(netD.parameters(), lr=opt.lr, betas=(opt.beta1, 0.999)) optimizerG = optim.Adam(netG.parameters(), lr=opt.lr, betas=(opt.beta1, 0.999)) [netD, netG], [optimizerD, optimizerG] = amp.initialize( [netD, netG], [optimizerD, optimizerG], opt_level=opt.opt_level, num_losses=3) for epoch in range(opt.niter): for i, data in enumerate(dataloader, 0): ############################ # (1) Update D network: maximize log(D(x)) + log(1 - D(G(z))) ########################### # train with real netD.zero_grad() real_cpu = data[0].to(device) batch_size = real_cpu.size(0) label = torch.full((batch_size,), real_label, device=device) output = netD(real_cpu) errD_real = criterion(output, label) with amp.scale_loss(errD_real, optimizerD, loss_id=0) as errD_real_scaled: errD_real_scaled.backward() D_x = output.mean().item() # train with fake noise = torch.randn(batch_size, nz, 1, 1, device=device) fake = netG(noise) label.fill_(fake_label) output = netD(fake.detach()) errD_fake = criterion(output, label) with amp.scale_loss(errD_fake, optimizerD, loss_id=1) as errD_fake_scaled: errD_fake_scaled.backward() D_G_z1 = output.mean().item() errD = errD_real + errD_fake optimizerD.step() ############################ # (2) Update G network: maximize log(D(G(z))) ########################### netG.zero_grad() label.fill_(real_label) # fake labels are real for generator cost output = netD(fake) errG = criterion(output, label) with amp.scale_loss(errG, optimizerG, loss_id=2) as errG_scaled: errG_scaled.backward() D_G_z2 = output.mean().item() optimizerG.step() print('[%d/%d][%d/%d] Loss_D: %.4f Loss_G: %.4f D(x): %.4f D(G(z)): %.4f / %.4f' % (epoch, opt.niter, i, len(dataloader), errD.item(), errG.item(), D_x, D_G_z1, D_G_z2)) if i % 100 == 0: vutils.save_image(real_cpu, '%s/real_samples.png' % opt.outf, normalize=True) fake = netG(fixed_noise) vutils.save_image(fake.detach(), '%s/amp_fake_samples_epoch_%03d.png' % (opt.outf, epoch), normalize=True) # do checkpointing torch.save(netG.state_dict(), '%s/netG_epoch_%d.pth' % (opt.outf, epoch)) torch.save(netD.state_dict(), '%s/netD_epoch_%d.pth' % (opt.outf, epoch))
apex-master
examples/dcgan/main_amp.py
import torch import argparse import os from apex import amp # FOR DISTRIBUTED: (can also use torch.nn.parallel.DistributedDataParallel instead) from apex.parallel import DistributedDataParallel parser = argparse.ArgumentParser() # FOR DISTRIBUTED: Parse for the local_rank argument, which will be supplied # automatically by torch.distributed.launch. parser.add_argument("--local_rank", default=0, type=int) args = parser.parse_args() # FOR DISTRIBUTED: If we are running under torch.distributed.launch, # the 'WORLD_SIZE' environment variable will also be set automatically. args.distributed = False if 'WORLD_SIZE' in os.environ: args.distributed = int(os.environ['WORLD_SIZE']) > 1 if args.distributed: # FOR DISTRIBUTED: Set the device according to local_rank. torch.cuda.set_device(args.local_rank) # FOR DISTRIBUTED: Initialize the backend. torch.distributed.launch will provide # environment variables, and requires that you use init_method=`env://`. torch.distributed.init_process_group(backend='nccl', init_method='env://') torch.backends.cudnn.benchmark = True N, D_in, D_out = 64, 1024, 16 # Each process receives its own batch of "fake input data" and "fake target data." # The "training loop" in each process just uses this fake batch over and over. # https://github.com/NVIDIA/apex/tree/master/examples/imagenet provides a more realistic # example of distributed data sampling for both training and validation. x = torch.randn(N, D_in, device='cuda') y = torch.randn(N, D_out, device='cuda') model = torch.nn.Linear(D_in, D_out).cuda() optimizer = torch.optim.SGD(model.parameters(), lr=1e-3) model, optimizer = amp.initialize(model, optimizer, opt_level="O1") if args.distributed: # FOR DISTRIBUTED: After amp.initialize, wrap the model with # apex.parallel.DistributedDataParallel. model = DistributedDataParallel(model) # torch.nn.parallel.DistributedDataParallel is also fine, with some added args: # model = torch.nn.parallel.DistributedDataParallel(model, # device_ids=[args.local_rank], # output_device=args.local_rank) loss_fn = torch.nn.MSELoss() for t in range(500): optimizer.zero_grad() y_pred = model(x) loss = loss_fn(y_pred, y) with amp.scale_loss(loss, optimizer) as scaled_loss: scaled_loss.backward() optimizer.step() if args.local_rank == 0: print("final loss = ", loss)
apex-master
examples/simple/distributed/distributed_data_parallel.py
import argparse import os import shutil import time import torch import torch.nn as nn import torch.nn.parallel import torch.backends.cudnn as cudnn import torch.distributed as dist import torch.optim import torch.utils.data import torch.utils.data.distributed import torchvision.transforms as transforms import torchvision.datasets as datasets import torchvision.models as models import numpy as np try: from apex.parallel import DistributedDataParallel as DDP from apex.fp16_utils import * from apex import amp, optimizers from apex.multi_tensor_apply import multi_tensor_applier except ImportError: raise ImportError("Please install apex from https://www.github.com/nvidia/apex to run this example.") def fast_collate(batch, memory_format): imgs = [img[0] for img in batch] targets = torch.tensor([target[1] for target in batch], dtype=torch.int64) w = imgs[0].size[0] h = imgs[0].size[1] tensor = torch.zeros( (len(imgs), 3, h, w), dtype=torch.uint8).contiguous(memory_format=memory_format) for i, img in enumerate(imgs): nump_array = np.asarray(img, dtype=np.uint8) if(nump_array.ndim < 3): nump_array = np.expand_dims(nump_array, axis=-1) nump_array = np.rollaxis(nump_array, 2) tensor[i] += torch.from_numpy(nump_array) return tensor, targets def parse(): model_names = sorted(name for name in models.__dict__ if name.islower() and not name.startswith("__") and callable(models.__dict__[name])) parser = argparse.ArgumentParser(description='PyTorch ImageNet Training') parser.add_argument('data', metavar='DIR', help='path to dataset') parser.add_argument('--arch', '-a', metavar='ARCH', default='resnet18', choices=model_names, help='model architecture: ' + ' | '.join(model_names) + ' (default: resnet18)') parser.add_argument('-j', '--workers', default=4, type=int, metavar='N', help='number of data loading workers (default: 4)') parser.add_argument('--epochs', default=90, type=int, metavar='N', help='number of total epochs to run') parser.add_argument('--start-epoch', default=0, type=int, metavar='N', help='manual epoch number (useful on restarts)') parser.add_argument('-b', '--batch-size', default=256, type=int, metavar='N', help='mini-batch size per process (default: 256)') parser.add_argument('--lr', '--learning-rate', default=0.1, type=float, metavar='LR', help='Initial learning rate. Will be scaled by <global batch size>/256: args.lr = args.lr*float(args.batch_size*args.world_size)/256. A warmup schedule will also be applied over the first 5 epochs.') parser.add_argument('--momentum', default=0.9, type=float, metavar='M', help='momentum') parser.add_argument('--weight-decay', '--wd', default=1e-4, type=float, metavar='W', help='weight decay (default: 1e-4)') parser.add_argument('--print-freq', '-p', default=10, type=int, metavar='N', help='print frequency (default: 10)') parser.add_argument('--resume', default='', type=str, metavar='PATH', help='path to latest checkpoint (default: none)') parser.add_argument('-e', '--evaluate', dest='evaluate', action='store_true', help='evaluate model on validation set') parser.add_argument('--pretrained', dest='pretrained', action='store_true', help='use pre-trained model') parser.add_argument('--prof', default=-1, type=int, help='Only run 10 iterations for profiling.') parser.add_argument('--deterministic', action='store_true') parser.add_argument("--local_rank", default=os.getenv('LOCAL_RANK', 0), type=int) parser.add_argument('--sync_bn', action='store_true', help='enabling apex sync BN.') parser.add_argument('--opt-level', type=str) parser.add_argument('--keep-batchnorm-fp32', type=str, default=None) parser.add_argument('--loss-scale', type=str, default=None) parser.add_argument('--channels-last', type=bool, default=False) args = parser.parse_args() return args def main(): global best_prec1, args args = parse() print("opt_level = {}".format(args.opt_level)) print("keep_batchnorm_fp32 = {}".format(args.keep_batchnorm_fp32), type(args.keep_batchnorm_fp32)) print("loss_scale = {}".format(args.loss_scale), type(args.loss_scale)) print("\nCUDNN VERSION: {}\n".format(torch.backends.cudnn.version())) cudnn.benchmark = True best_prec1 = 0 if args.deterministic: cudnn.benchmark = False cudnn.deterministic = True torch.manual_seed(args.local_rank) torch.set_printoptions(precision=10) args.distributed = False if 'WORLD_SIZE' in os.environ: args.distributed = int(os.environ['WORLD_SIZE']) > 1 args.gpu = 0 args.world_size = 1 if args.distributed: args.gpu = args.local_rank torch.cuda.set_device(args.gpu) torch.distributed.init_process_group(backend='nccl', init_method='env://') args.world_size = torch.distributed.get_world_size() assert torch.backends.cudnn.enabled, "Amp requires cudnn backend to be enabled." if args.channels_last: memory_format = torch.channels_last else: memory_format = torch.contiguous_format # create model if args.pretrained: print("=> using pre-trained model '{}'".format(args.arch)) model = models.__dict__[args.arch](pretrained=True) else: print("=> creating model '{}'".format(args.arch)) model = models.__dict__[args.arch]() if args.sync_bn: import apex print("using apex synced BN") model = apex.parallel.convert_syncbn_model(model) model = model.cuda().to(memory_format=memory_format) # Scale learning rate based on global batch size args.lr = args.lr*float(args.batch_size*args.world_size)/256. optimizer = torch.optim.SGD(model.parameters(), args.lr, momentum=args.momentum, weight_decay=args.weight_decay) # Initialize Amp. Amp accepts either values or strings for the optional override arguments, # for convenient interoperation with argparse. model, optimizer = amp.initialize(model, optimizer, opt_level=args.opt_level, keep_batchnorm_fp32=args.keep_batchnorm_fp32, loss_scale=args.loss_scale ) # For distributed training, wrap the model with apex.parallel.DistributedDataParallel. # This must be done AFTER the call to amp.initialize. If model = DDP(model) is called # before model, ... = amp.initialize(model, ...), the call to amp.initialize may alter # the types of model's parameters in a way that disrupts or destroys DDP's allreduce hooks. if args.distributed: # By default, apex.parallel.DistributedDataParallel overlaps communication with # computation in the backward pass. # model = DDP(model) # delay_allreduce delays all communication to the end of the backward pass. model = DDP(model, delay_allreduce=True) # define loss function (criterion) and optimizer criterion = nn.CrossEntropyLoss().cuda() # Optionally resume from a checkpoint if args.resume: # Use a local scope to avoid dangling references def resume(): if os.path.isfile(args.resume): print("=> loading checkpoint '{}'".format(args.resume)) checkpoint = torch.load(args.resume, map_location = lambda storage, loc: storage.cuda(args.gpu)) args.start_epoch = checkpoint['epoch'] global best_prec1 best_prec1 = checkpoint['best_prec1'] model.load_state_dict(checkpoint['state_dict']) optimizer.load_state_dict(checkpoint['optimizer']) print("=> loaded checkpoint '{}' (epoch {})" .format(args.resume, checkpoint['epoch'])) else: print("=> no checkpoint found at '{}'".format(args.resume)) resume() # Data loading code traindir = os.path.join(args.data, 'train') valdir = os.path.join(args.data, 'val') if(args.arch == "inception_v3"): raise RuntimeError("Currently, inception_v3 is not supported by this example.") # crop_size = 299 # val_size = 320 # I chose this value arbitrarily, we can adjust. else: crop_size = 224 val_size = 256 train_dataset = datasets.ImageFolder( traindir, transforms.Compose([ transforms.RandomResizedCrop(crop_size), transforms.RandomHorizontalFlip(), # transforms.ToTensor(), Too slow # normalize, ])) val_dataset = datasets.ImageFolder(valdir, transforms.Compose([ transforms.Resize(val_size), transforms.CenterCrop(crop_size), ])) train_sampler = None val_sampler = None if args.distributed: train_sampler = torch.utils.data.distributed.DistributedSampler(train_dataset) val_sampler = torch.utils.data.distributed.DistributedSampler(val_dataset) collate_fn = lambda b: fast_collate(b, memory_format) train_loader = torch.utils.data.DataLoader( train_dataset, batch_size=args.batch_size, shuffle=(train_sampler is None), num_workers=args.workers, pin_memory=True, sampler=train_sampler, collate_fn=collate_fn) val_loader = torch.utils.data.DataLoader( val_dataset, batch_size=args.batch_size, shuffle=False, num_workers=args.workers, pin_memory=True, sampler=val_sampler, collate_fn=collate_fn) if args.evaluate: validate(val_loader, model, criterion) return for epoch in range(args.start_epoch, args.epochs): if args.distributed: train_sampler.set_epoch(epoch) # train for one epoch train(train_loader, model, criterion, optimizer, epoch) # evaluate on validation set prec1 = validate(val_loader, model, criterion) # remember best prec@1 and save checkpoint if args.local_rank == 0: is_best = prec1 > best_prec1 best_prec1 = max(prec1, best_prec1) save_checkpoint({ 'epoch': epoch + 1, 'arch': args.arch, 'state_dict': model.state_dict(), 'best_prec1': best_prec1, 'optimizer' : optimizer.state_dict(), }, is_best) class data_prefetcher(): def __init__(self, loader): self.loader = iter(loader) self.stream = torch.cuda.Stream() self.mean = torch.tensor([0.485 * 255, 0.456 * 255, 0.406 * 255]).cuda().view(1,3,1,1) self.std = torch.tensor([0.229 * 255, 0.224 * 255, 0.225 * 255]).cuda().view(1,3,1,1) # With Amp, it isn't necessary to manually convert data to half. # if args.fp16: # self.mean = self.mean.half() # self.std = self.std.half() self.preload() def preload(self): try: self.next_input, self.next_target = next(self.loader) except StopIteration: self.next_input = None self.next_target = None return # if record_stream() doesn't work, another option is to make sure device inputs are created # on the main stream. # self.next_input_gpu = torch.empty_like(self.next_input, device='cuda') # self.next_target_gpu = torch.empty_like(self.next_target, device='cuda') # Need to make sure the memory allocated for next_* is not still in use by the main stream # at the time we start copying to next_*: # self.stream.wait_stream(torch.cuda.current_stream()) with torch.cuda.stream(self.stream): self.next_input = self.next_input.cuda(non_blocking=True) self.next_target = self.next_target.cuda(non_blocking=True) # more code for the alternative if record_stream() doesn't work: # copy_ will record the use of the pinned source tensor in this side stream. # self.next_input_gpu.copy_(self.next_input, non_blocking=True) # self.next_target_gpu.copy_(self.next_target, non_blocking=True) # self.next_input = self.next_input_gpu # self.next_target = self.next_target_gpu # With Amp, it isn't necessary to manually convert data to half. # if args.fp16: # self.next_input = self.next_input.half() # else: self.next_input = self.next_input.float() self.next_input = self.next_input.sub_(self.mean).div_(self.std) def next(self): torch.cuda.current_stream().wait_stream(self.stream) input = self.next_input target = self.next_target if input is not None: input.record_stream(torch.cuda.current_stream()) if target is not None: target.record_stream(torch.cuda.current_stream()) self.preload() return input, target def train(train_loader, model, criterion, optimizer, epoch): batch_time = AverageMeter() losses = AverageMeter() top1 = AverageMeter() top5 = AverageMeter() # switch to train mode model.train() end = time.time() prefetcher = data_prefetcher(train_loader) input, target = prefetcher.next() i = 0 while input is not None: i += 1 if args.prof >= 0 and i == args.prof: print("Profiling begun at iteration {}".format(i)) torch.cuda.cudart().cudaProfilerStart() if args.prof >= 0: torch.cuda.nvtx.range_push("Body of iteration {}".format(i)) adjust_learning_rate(optimizer, epoch, i, len(train_loader)) # compute output if args.prof >= 0: torch.cuda.nvtx.range_push("forward") output = model(input) if args.prof >= 0: torch.cuda.nvtx.range_pop() loss = criterion(output, target) # compute gradient and do SGD step optimizer.zero_grad() if args.prof >= 0: torch.cuda.nvtx.range_push("backward") with amp.scale_loss(loss, optimizer) as scaled_loss: scaled_loss.backward() if args.prof >= 0: torch.cuda.nvtx.range_pop() # for param in model.parameters(): # print(param.data.double().sum().item(), param.grad.data.double().sum().item()) if args.prof >= 0: torch.cuda.nvtx.range_push("optimizer.step()") optimizer.step() if args.prof >= 0: torch.cuda.nvtx.range_pop() if i%args.print_freq == 0: # Every print_freq iterations, check the loss, accuracy, and speed. # For best performance, it doesn't make sense to print these metrics every # iteration, since they incur an allreduce and some host<->device syncs. # Measure accuracy prec1, prec5 = accuracy(output.data, target, topk=(1, 5)) # Average loss and accuracy across processes for logging if args.distributed: reduced_loss = reduce_tensor(loss.data) prec1 = reduce_tensor(prec1) prec5 = reduce_tensor(prec5) else: reduced_loss = loss.data # to_python_float incurs a host<->device sync losses.update(to_python_float(reduced_loss), input.size(0)) top1.update(to_python_float(prec1), input.size(0)) top5.update(to_python_float(prec5), input.size(0)) torch.cuda.synchronize() batch_time.update((time.time() - end)/args.print_freq) end = time.time() if args.local_rank == 0: print('Epoch: [{0}][{1}/{2}]\t' 'Time {batch_time.val:.3f} ({batch_time.avg:.3f})\t' 'Speed {3:.3f} ({4:.3f})\t' 'Loss {loss.val:.10f} ({loss.avg:.4f})\t' 'Prec@1 {top1.val:.3f} ({top1.avg:.3f})\t' 'Prec@5 {top5.val:.3f} ({top5.avg:.3f})'.format( epoch, i, len(train_loader), args.world_size*args.batch_size/batch_time.val, args.world_size*args.batch_size/batch_time.avg, batch_time=batch_time, loss=losses, top1=top1, top5=top5)) if args.prof >= 0: torch.cuda.nvtx.range_push("prefetcher.next()") input, target = prefetcher.next() if args.prof >= 0: torch.cuda.nvtx.range_pop() # Pop range "Body of iteration {}".format(i) if args.prof >= 0: torch.cuda.nvtx.range_pop() if args.prof >= 0 and i == args.prof + 10: print("Profiling ended at iteration {}".format(i)) torch.cuda.cudart().cudaProfilerStop() quit() def validate(val_loader, model, criterion): batch_time = AverageMeter() losses = AverageMeter() top1 = AverageMeter() top5 = AverageMeter() # switch to evaluate mode model.eval() end = time.time() prefetcher = data_prefetcher(val_loader) input, target = prefetcher.next() i = 0 while input is not None: i += 1 # compute output with torch.no_grad(): output = model(input) loss = criterion(output, target) # measure accuracy and record loss prec1, prec5 = accuracy(output.data, target, topk=(1, 5)) if args.distributed: reduced_loss = reduce_tensor(loss.data) prec1 = reduce_tensor(prec1) prec5 = reduce_tensor(prec5) else: reduced_loss = loss.data losses.update(to_python_float(reduced_loss), input.size(0)) top1.update(to_python_float(prec1), input.size(0)) top5.update(to_python_float(prec5), input.size(0)) # measure elapsed time batch_time.update(time.time() - end) end = time.time() # TODO: Change timings to mirror train(). if args.local_rank == 0 and i % args.print_freq == 0: print('Test: [{0}/{1}]\t' 'Time {batch_time.val:.3f} ({batch_time.avg:.3f})\t' 'Speed {2:.3f} ({3:.3f})\t' 'Loss {loss.val:.4f} ({loss.avg:.4f})\t' 'Prec@1 {top1.val:.3f} ({top1.avg:.3f})\t' 'Prec@5 {top5.val:.3f} ({top5.avg:.3f})'.format( i, len(val_loader), args.world_size * args.batch_size / batch_time.val, args.world_size * args.batch_size / batch_time.avg, batch_time=batch_time, loss=losses, top1=top1, top5=top5)) input, target = prefetcher.next() print(' * Prec@1 {top1.avg:.3f} Prec@5 {top5.avg:.3f}' .format(top1=top1, top5=top5)) return top1.avg def save_checkpoint(state, is_best, filename='checkpoint.pth.tar'): torch.save(state, filename) if is_best: shutil.copyfile(filename, 'model_best.pth.tar') class AverageMeter(object): """Computes and stores the average and current value""" def __init__(self): self.reset() def reset(self): self.val = 0 self.avg = 0 self.sum = 0 self.count = 0 def update(self, val, n=1): self.val = val self.sum += val * n self.count += n self.avg = self.sum / self.count def adjust_learning_rate(optimizer, epoch, step, len_epoch): """LR schedule that should yield 76% converged accuracy with batch size 256""" factor = epoch // 30 if epoch >= 80: factor = factor + 1 lr = args.lr*(0.1**factor) """Warmup""" if epoch < 5: lr = lr*float(1 + step + epoch*len_epoch)/(5.*len_epoch) # if(args.local_rank == 0): # print("epoch = {}, step = {}, lr = {}".format(epoch, step, lr)) for param_group in optimizer.param_groups: param_group['lr'] = lr def accuracy(output, target, topk=(1,)): """Computes the precision@k for the specified values of k""" maxk = max(topk) batch_size = target.size(0) _, pred = output.topk(maxk, 1, True, True) pred = pred.t() correct = pred.eq(target.view(1, -1).expand_as(pred)) res = [] for k in topk: correct_k = correct[:k].reshape(-1).float().sum(0, keepdim=True) res.append(correct_k.mul_(100.0 / batch_size)) return res def reduce_tensor(tensor): rt = tensor.clone() dist.all_reduce(rt, op=dist.reduce_op.SUM) rt /= args.world_size return rt if __name__ == '__main__': main()
apex-master
examples/imagenet/main_amp.py
# Copyright (c) 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from setuptools import setup setup( entry_points={ "modulus.models": [ "AFNO = modulus.models.afno:AFNO", "DLWP = modulus.models.dlwp:DLWP", "FNO = modulus.models.fno:FNO", "GraphCastNet = modulus.models.graphcast:GraphCastNet", "MeshGraphNet = modulus.models.meshgraphnet:MeshGraphNet", "FullyConnected = modulus.models.mlp:FullyConnected", "Pix2Pix = modulus.models.pix2pix:Pix2Pix", "One2ManyRNN = modulus.models.rnn:One2ManyRNN", #'SphericalFourierNeuralOperatorNet = modulus.models.sfno:SphericalFourierNeuralOperatorNet', "SRResNet = modulus.models.srrn:SRResNet", ], } )
modulus-main
setup.py
# Copyright (c) 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ constant values used by Modulus """ import torch import numpy as np # string used to determine derivatives diff_str: str = "__" def diff(y: str, x: str, degree: int = 1) -> str: """Function to apply diff string""" return diff_str.join([y] + degree * [x]) # for changing to float16 or float64 tf_dt = torch.float32 np_dt = np.float32 # tensorboard naming TF_SUMMARY = False # Pytorch Version for which JIT will be default on # Torch version of NGC container 22.08 JIT_PYTORCH_VERSION = "1.13.0a0+d321be6" # No scaling is needed if using NO_OP_SCALE NO_OP_SCALE = (0.0, 1.0) # If using NO_OP_NORM, it is effectively doing no normalization NO_OP_NORM = (-1.0, 1.0)
modulus-main
modulus/constants.py
# Copyright (c) 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from .models.module import Module from .models.meta import ModelMetaData from .datapipes.datapipe import Datapipe from .datapipes.meta import DatapipeMetaData from .datapipes.datapipe import Datapipe __version__ = "0.3.0a0"
modulus-main
modulus/__init__.py
# Copyright (c) 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
modulus-main
modulus/metrics/__init__.py
# Copyright (c) 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # TODO(Dallas) Introduce Ensemble RMSE and MSE routines. import torch from typing import Union Tensor = torch.Tensor def mse(pred: Tensor, target: Tensor, dim: int = None) -> Union[Tensor, float]: """Calculates Mean Squared error between two tensors Parameters ---------- pred : Tensor Input prediction tensor target : Tensor Target tensor dim : int, optional Reduction dimension. When None the losses are averaged or summed over all observations, by default None Returns ------- Union[Tensor, float] Mean squared error value(s) """ return torch.mean((pred - target) ** 2, dim=dim) def rmse(pred: Tensor, target: Tensor, dim: int = None) -> Union[Tensor, float]: """Calculates Root mean Squared error between two tensors Parameters ---------- pred : Tensor Input prediction tensor target : Tensor Target tensor dim : int, optional Reduction dimension. When None the losses are averaged or summed over all observations, by default None Returns ------- Union[Tensor, float] Root mean squared error value(s) """ return torch.sqrt(mse(pred, target, dim=dim))
modulus-main
modulus/metrics/general/mse.py
# Copyright (c) 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from torch import Tensor from abc import ABC class WeightedStatistic(ABC): """A convenience class for computing weighted statistics of some input Parameters ---------- weights : Tensor Weight tensor """ def __init__(self, weights: Tensor): assert torch.all(weights > 0.0).item(), "Expected all weights to be positive." self.weights = self._normalize(weights) def __call__(self, x: Tensor, dim: int): """ Convenience method to make sure weights have appropriate shapes. """ w = self.weights if w.ndim == 1: assert x.shape[dim] == len(w), ( "Expected inputs and weights to have the same size along the reduction dimension but have dimensions" + str(len(x[dim])) + " and " + str(len(w)) + "." ) if dim < 0: dim = x.ndim + dim for i in range(x.ndim): if i < dim: w = w.unsqueeze(0) elif i > dim: w = w.unsqueeze(-1) else: assert (x.ndim == w.ndim) and ( x.shape[dim] == w.shape[dim] ), "Expected inputs and weights to have compatible shapes." return w def _normalize(self, weights: Tensor) -> Tensor: """Normalize unnormalized weights, for convenience Parameters ---------- weights : Tensor Unnormalized weights Returns ------- Tensor Normalized weights """ return weights / torch.sum(weights) class WeightedMean(WeightedStatistic): """ Compute weighted mean of some input. Parameters ---------- weights : Tensor Weight tensor """ def __init__(self, weights: Tensor): super().__init__(weights) def __call__(self, x: Tensor, dim: int, keepdims: bool = False) -> Tensor: """Compute weighted mean Parameters ---------- x : Tensor Input data dim : int Dimension to take aggregate keepdims : bool, optional Keep aggregated dimension, by default False Returns ------- Tensor Weighted mean """ w = super().__call__(x, dim) return torch.sum(x * w, dim=dim, keepdims=keepdims) class WeightedVariance(WeightedStatistic): """ Compute weighted variance of some input. Parameters ---------- weights : Tensor Weight tensor """ def __init__(self, weights: Tensor): super().__init__(weights) self.wm = WeightedMean(self.weights) def __call__(self, x: Tensor, dim: int, keepdims: bool = False): """Compute weighted variance Parameters ---------- x : Tensor Input data dim : int Dimension to take aggregate keepdims : bool, optional Keep aggregated dimension, by default False Returns ------- Tensor Weighted variance """ w = super().__call__(x, dim) # Compute weighted mean wm = self.wm(x, dim, keepdims=True) # Computing scaling for standard deviation number_of_non_zero_weights = torch.sum(w > 0.0) scale = (number_of_non_zero_weights - 1.0) / number_of_non_zero_weights return torch.sum(w * (x - wm) ** 2, dim=dim, keepdims=keepdims) / scale
modulus-main
modulus/metrics/general/reduction.py
# Copyright (c) 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch import numpy import torch.distributed as dist from typing import Union, Tuple from modulus.distributed.manager import DistributedManager from .ensemble_metrics import EnsembleMetrics Tensor = torch.Tensor @torch.jit.script def linspace(start: Tensor, stop: Tensor, num: int) -> Tensor: """Element-wise multi-dimensional linspace Replicates the bahaviour of numpy.linspace over all elements of multi-dimensional tensors in PyTorch. Parameters ---------- start : Tensor Starting input Tensor stop : Tensor Ending input Tensor, should be of same size a input num : int Number of steps between start and end values between each element Returns ------- Tensor Tensor of evenly spaced numbers over defined interval [num, *start.shape] """ # create a tensor of 'num' steps from 0 to 1 steps = torch.arange(num + 1, dtype=torch.float32, device=start.device) / (num) # reshape the 'steps' tensor to [-1, *([1]*start.ndim)] to allow for broadcastings # - using 'steps.reshape([-1, *([1]*start.ndim)])' would be nice here but # torchscript "cannot statically infer the expected size of a list in this contex", # hence the code below for i in range(start.ndim): steps = steps.unsqueeze(-1) # the output starts at 'start' and increments until 'stop' in each dimension out = start[None] + steps * (stop - start)[None] return out @torch.jit.script def _low_memory_bin_reduction_counts( inputs: Tensor, bin_edges: Tensor, counts: Tensor, number_of_bins: int ): """Computes low-memory bin counts This function calculates a low-memory bin count of the inputs tensor and adding the result to counts, in place. The low-memory usage comes at the cost of iterating over the batched dimension. This bin count is done with respect to computing a pmf/pdf. Parameters ---------- inputs : Tensor Inputs to be binned, has dimension [B, ...] where B is the batch dimension that the binning is done over bin_edges : Tensor Bin edges with dimension [N+1, ...] where N is the number of bins counts : Tensor Existing bin count tensor with dimension [N, ...] where N is the number of bins number_of_bins : int Number of bins Returns ------- Tensor PDF bin count tensor [N, ...] """ for j in range(inputs.shape[0]): counts[0] += (inputs[j] < bin_edges[1]).int() for i in range(1, number_of_bins - 1): for j in range(inputs.shape[0]): counts[i] += (inputs[j] < bin_edges[i + 1]).int() - ( inputs[j] < bin_edges[i] ).int() for j in range(inputs.shape[0]): counts[number_of_bins - 1] += ( 1 - (inputs[j] < bin_edges[number_of_bins - 1]).int() ) return counts @torch.jit.script def _high_memory_bin_reduction_counts( inputs: Tensor, bin_edges: Tensor, counts: Tensor, number_of_bins: int ) -> Tensor: """Computes high-memory bin counts This function calculates a high-memory bin count of the inputs tensor and adding the result to counts, in place. The high-memory usage comes from computing the entire reduction in memory. See _low_memory_bin_reduction for an alternative. This bin count is done with respect to computing a pmf/pdf. Parameters ---------- inputs : Tensor Inputs to be binned, has dimension [B, ...] where B is the batch dimension that the binning is done over. bin_edges : Tensor Bin edges with dimension [N+1, ...] where N is the number of bins. counts : Tensor Existing bin count tensor with dimension [N, ...] where N is the number of bins. number_of_bins : int Number of bins Returns ------- Tensor PDF bin count tensor [N, ...] """ counts[0] += torch.count_nonzero(inputs < bin_edges[1], dim=0) for i in range(1, number_of_bins - 1): counts[i] += torch.count_nonzero( inputs < bin_edges[i + 1], dim=0 ) - torch.count_nonzero(inputs < bin_edges[i], dim=0) counts[number_of_bins - 1] += inputs.shape[0] - torch.count_nonzero( inputs < bin_edges[number_of_bins - 1], dim=0 ) return counts @torch.jit.script def _low_memory_bin_reduction_cdf( inputs: Tensor, bin_edges: Tensor, counts: Tensor, number_of_bins: int ) -> Tensor: """Computes low-memory cumulative bin counts This function calculates a low-memory cumulative bin count of the inputs tensor and adding the result to counts, in place. The low-memory usage comes at the cost of iterating over the batched dimension. This bin count is done with respect to computing a cmf/cdf. Parameters ---------- inputs : Tensor Inputs to be binned, has dimension [B, ...] where B is the batch dimension that the binning is done over bin_edges : Tensor Bin edges with dimension [N+1, ...] where N is the number of bins counts : Tensor Existing bin count tensor with dimension [N, ...] where N is the number of bins number_of_bins : int Number of bins Returns ------- Tensor CDF bin count tensor [N, ...] """ for i in range(number_of_bins - 1): for j in range(inputs.shape[0]): counts[i] += (inputs[j] < bin_edges[i + 1]).int() counts[number_of_bins - 1] += inputs.shape[0] return counts @torch.jit.script def _high_memory_bin_reduction_cdf( inputs: torch.Tensor, bin_edges: torch.Tensor, counts: torch.Tensor, number_of_bins: int, ) -> Tensor: """Computes high-memory cumulative bin counts This function calculates a high-memory cumulative bin countof the inputs tensor and adding the result to counts, in place. The high-memory usage comes from computing the entire reduction in memory. This bin count is done with respect to computing a cmf/cdf. Parameters ---------- inputs : torch.Tensor Inputs to be binned, has dimension [B, ...] where B is the batch dimension that the binning is done over. bin_edges : torch.Tensor Bin edges with dimension [N+1, ...] where N is the number of bins counts : torch.Tensor Existing bin count tensor with dimension [N, ...] where N is the number of bins number_of_bins : int Number of bins Returns ------- Tensor CDF bin count tensor [N, ...] """ for i in range(number_of_bins - 1): counts[i] += torch.count_nonzero(inputs < bin_edges[i + 1], dim=0) counts[number_of_bins - 1] = inputs.shape[0] return counts def _count_bins( input: torch.Tensor, bin_edges: torch.Tensor, counts: Union[torch.Tensor, None] = None, cdf: bool = False, ) -> Tensor: """Computes (un)Cumulative bin counts of input tensor This function calculates the bin count of the inputs tensor and adding the result to counts, in place. Attempts to use a _high_memory_bin_reduction for performance reasons, but will fall back to less performant, less memory intensive routine. Parameters ---------- input : Tensor Inputs to be binned, has dimension [B, ...] where B is the batch dimension that the binning is done over bin_edges : Tensor Bin edges with dimension [N+1, ...] where N is the number of bins counts : Union[torch.Tensor, None] Existing bin count tensor with dimension [N, ...] where N is the number of bins. If no counts is passed then we construct an empty counts, by default None cdf : bool, optional Compute a counts or cumulative counts; will calculate unnormalized counts function otherwise, by default False Returns ------- Tensor CDF bin count tensor [N, ...] """ bins_shape = bin_edges.shape number_of_bins = bins_shape[0] - 1 assert ( bins_shape[1:] == input.shape[1:] ), "Expected bin_edges and inputs to have compatible non-leading dimensions." if counts is None: counts = torch.zeros( (number_of_bins, *bins_shape[1:]), dtype=torch.int64, device=input.device ) else: assert ( bins_shape[1:] == counts.shape[1:] ), "Expected bin_edges and counts to have compatible non-leading dimensions." if cdf: try: counts = _high_memory_bin_reduction_cdf( input, bin_edges, counts, number_of_bins ) except RuntimeError: counts = _low_memory_bin_reduction_cdf( input, bin_edges, counts, number_of_bins ) else: try: counts = _high_memory_bin_reduction_counts( input, bin_edges, counts, number_of_bins ) except RuntimeError: counts = _low_memory_bin_reduction_counts( input, bin_edges, counts, number_of_bins ) return counts def _get_mins_maxs(*inputs: Tensor, axis: int = 0) -> Tuple[Tensor, Tensor]: """Get max and min value across specified dimension Parameters ---------- inputs : (Tensor ...) Input tensor(s) axis : int, optional Axis to calc min/max values with, by default 0 Returns ------- Tuple[Tensor, Tensor] (Minimum, Maximum) values of inputs """ assert len(inputs) > 0, "At least one tensor much be provided" input = inputs[0] inputs = list(inputs)[1:] # Check shape consistency s = input.shape for x in inputs: assert s[1:] == x.shape[1:] # Compute low and high for input low = torch.min(input, axis=axis)[0] high = torch.max(input, axis=axis)[0] # Iterative over any and all args, storing the latest inf/sup. for x in inputs: low0 = torch.min(x, axis=axis)[0] high0 = torch.max(x, axis=axis)[0] low = torch.where(low < low0, low, low0) high = torch.where(high > high0, high, high0) return low, high def _update_bins_counts( input: Tensor, bin_edges: Tensor, counts: Tensor, cdf: bool = False, tol: float = 1e-2, ) -> Tuple[Tensor, Tensor]: """Utility for updating an existing histogram with new inputs Parameters ---------- input : Tensor Input tensor of updated data for binning bin_edges : Tensor Current bin range tensor [N+1, ...] where N is the number of bins counts : Tensor Existing bin count tensor with dimension [N, ...] where N is the number of bins cdf : bool, optional Compute a histogram or a cumulative density function; will calculate probability density function otherwise, by default False tol : float, optional Binning tolerance, by default 1e-4 Returns ------- Tuple[Tensor, Tensor] Updated (bin, count) tensors """ # Compute new lows and highs, compare against old bins low, high = _get_mins_maxs(input) # If in distributed environment, reduce to get extrema min and max if DistributedManager.is_initialized() and dist.is_initialized(): dist.all_reduce(low, op=dist.ReduceOp.MIN) dist.all_reduce(high, op=dist.ReduceOp.MAX) low = torch.where(low < bin_edges[0], low, bin_edges[0]) high = torch.where(high > bin_edges[-1], high, bin_edges[-1]) # Test if bin_edges is a superset and do not recompute bin_edges. if ~(torch.all(low == bin_edges[0]) & torch.all(high == bin_edges[-1])): # There are extrema in inputs/args that are outside of bin_edges and we must recompute bin_edges and counts. ## Need to make sure that the new bin_edges are consistent with the old bin_edges. # Need to compute dbin_edges = bin_edges[1] - bin_edges[0] # find minimum k s.t. bin_edges[0] - k*dbin_edges < low # set start = bin_edges[0] - k*dbin_edges # find minimum k s.t. bin_edges[-1] + k*dbin_edges > high # set end = bin_edges[-1] + k*dbin_edges dbin_edges = bin_edges[1] - bin_edges[0] old_number_of_bins = bin_edges.shape[0] - 1 number_of_bins = old_number_of_bins lk = torch.max(torch.ceil((bin_edges[0] - low) / dbin_edges)).int().item() start = bin_edges[0] - lk * dbin_edges number_of_bins += lk uk = torch.max(torch.ceil((high - bin_edges[-1]) / dbin_edges)).int().item() end = bin_edges[-1] + uk * dbin_edges number_of_bins += uk bin_edges = linspace(start, end, number_of_bins) new_counts = torch.zeros( (number_of_bins, *bin_edges.shape[1:]), dtype=torch.int64, device=bin_edges.device, ) new_counts[lk : lk + old_number_of_bins] += counts counts = new_counts # Count inputs to bins partial_counts = _count_bins(input, bin_edges, counts=None, cdf=cdf) # If in distributed environment, reduce to get extrema min and max if DistributedManager.is_initialized() and dist.is_initialized(): dist.all_reduce(partial_counts, op=dist.ReduceOp.SUM) counts += partial_counts # Finally, combine the new partial counts with the existing counts return bin_edges, counts def _compute_counts_cdf( *inputs: Tensor, bins: Union[int, Tensor] = 10, counts: Union[None, Tensor] = None, cdf: bool = False, tol: float = 1e-2, verbose: bool = False, ) -> Tuple[Tensor, Tensor]: """Computes the (un)Cumulative histograms of a tensor(s) over the leading dimension Parameters ---------- inputs : (Tensor ...) Input data tensor(s) [B, ...] bins : Union[int, Tensor], optional Either the number of bins, or a tensor of bin edges with dimension [N+1, ...] where N is the number of bins. If counts is passed, then bins is interpreted to be the bin edges for the counts tensor, by default 10 counts : Union[None, Tensor], optional Existing count tensor to combine results with. Must have dimensions [N, ...] where N is the number of bins. Passing a tensor may also require recomputing the passed bins to make sure inputs and bins are compatible., by default None cdf : bool, optional Compute a histogram or a cumulative density function; will calculate probability density function otherwise, by default False tol : float, optional Binning tolerance, by default 1e-4 verbose : bool, optional Verbose printing, by default False Returns ------- Tuple[Tensor, Tensor] The calculated (bin edges [N+1, ...], count [N, ...]) tensors """ # Check shapes of inputs s = inputs[0].shape for input in inputs[1:]: assert s[1:] == input.shape[1:] if type(bins) is int: if verbose: print("Bins is passed as an int.") # Compute largest bins needed low, high = _get_mins_maxs(*inputs) number_of_bins = bins bin_edges = linspace(low, high, number_of_bins) # Bin inputs counts = None for input in inputs: counts = _count_bins(input, bin_edges, counts=counts, cdf=cdf) return bin_edges, counts elif type(bins) is torch.Tensor: bin_edges = bins if verbose: print("Bins is passed as a tensor") if counts is None: # Do not need to update counts if verbose: print("No counts are passed.") number_of_bins = bin_edges.shape[0] - 1 # Get largest bin edges needed from input/args low, high = _get_mins_maxs(*inputs) # Compare against existing bin_edges low = torch.where(low < bin_edges[0], low, bin_edges[0]) high = torch.where(high > bin_edges[-1], high, bin_edges[-1]) # Update, if necessary if torch.any(low != bin_edges[0]) | torch.any(high != bin_edges[-1]): bin_edges = linspace(low, high, number_of_bins) # Bin inputs counts = None for input in inputs: counts = _count_bins(input, bin_edges, counts=counts, cdf=cdf) return bin_edges, counts else: # Counts do need to be update if verbose: print("Counts are being updated.") for input in inputs: bin_edges, counts = _update_bins_counts( input, bin_edges, counts, cdf=cdf ) return bin_edges, counts else: raise ValueError("Input bin type is not supported.") def histogram( *inputs: Tensor, bins: Union[int, Tensor] = 10, counts: Union[None, Tensor] = None, verbose: bool = False, ) -> Tuple[Tensor, Tensor]: """Computes the histogram of a set of tensors over the leading dimension This function will compute bin edges and bin counts of given input tensors. If existing bin edges or count tensors are supplied, this function will update these existing statistics with the new data. Parameters ---------- inputs : (Tensor ...) Input data tensor(s) [B, ...] bins : Union[int, Tensor], optional Either the number of bins, or a tensor of bin edges with dimension [N+1, ...] where N is the number of bins. If counts is passed, then bins is interpreted to be the bin edges for the counts tensor, by default 10 counts : Union[None, Tensor], optional Existing count tensor to combine results with. Must have dimensions [N, ...] where N is the number of bins. Passing a tensor may also require recomputing the passed bins to make sure inputs and bins are compatible, by default None verbose : bool, optional Verbose printing, by default False Returns ------- Tuple[Tensor, Tensor] The calculated (bin edges [N+1, ...], count [N, ...]) tensors """ return _compute_counts_cdf( *inputs, bins=bins, counts=counts, cdf=False, verbose=verbose ) def cdf( *inputs: Tensor, bins: Union[int, Tensor] = 10, counts: Union[None, Tensor] = None, verbose: bool = False, ) -> Tuple[Tensor, Tensor]: """Computes the cumulative density function of a set of tensors over the leading dimension This function will compute CDF bins of given input tensors. If existing bins or count tensors are supplied, this function will update these existing statistics with the new data. Parameters ---------- inputs : (Tensor ...) Input data tensor(s) [B, ...] bins : Union[int, Tensor], optional Either the number of bins, or a tensor of bin edges with dimension [N+1, ...] where N is the number of bins. If counts is passed, then bins is interpreted to be the bin edges for the counts tensor, by default 10 counts : Union[None, Tensor], optional Existing count tensor to combine results with. Must have dimensions [N, ...] where N is the number of bins. Passing a tensor may also require recomputing the passed bins to make sure inputs and bins are compatible, by default None verbose : bool, optional Verbose printing, by default False Returns ------- Tuple[Tensor, Tensor] The calculated (bin edges [N+1, ...], cdf [N, ...]) tensors """ bin_edges, counts = _compute_counts_cdf( *inputs, bins=bins, counts=counts, cdf=True, verbose=verbose ) cdf = counts / counts[-1] # Normalize return bin_edges, cdf class Histogram(EnsembleMetrics): """ Convenience class for computing histograms, possibly as a part of a distributed or iterative environment Parameters ---------- input_shape : Tuple[int] Input data shape bins : Union[int, Tensor], optional Initial bin edges or number of bins to use, by default 10 tol : float, optional Bin edge tolerance, by default 1e-3 """ def __init__( self, input_shape: Tuple[int], bins: Union[int, Tensor] = 10, tol: float = 1e-2, **kwargs, ): super().__init__(input_shape, **kwargs) if type(bins) is int: self.number_of_bins = bins else: self.number_of_bins = bins.shape[0] - 1 assert self.input_shape[1:] == bins.shape[1:] self.counts_shape = self.input_shape self.counts_shape[0] = self.number_of_bins self.tol = tol # Initialize bins start = -1.0 * torch.ones( self.input_shape[1:], device=self.device, dtype=self.dtype ) end = torch.ones(self.input_shape[1:], device=self.device, dtype=self.dtype) self.bin_edges = linspace(start, end, self.number_of_bins) self.counts = torch.zeros( self.counts_shape, device=self.device, dtype=torch.int64 ) def __call__(self, input: Tensor) -> Tuple[Tensor, Tensor]: """Calculate histogram Parameters ---------- inputs : Tensor Input data tensor [B, ...] Returns ------- Tuple[Tensor, Tensor] The calculated (bin edges [N+1, ...], counts [N, ...]) tensors """ # Build bin_edges if DistributedManager.is_initialized() and dist.is_initialized(): start, _ = torch.min(input, axis=0) end, _ = torch.max(input, axis=0) # We assume that the start/end across the distributed environments # need to be combined. dist.all_reduce(start, op=dist.ReduceOp.MIN) dist.all_reduce(end, op=dist.ReduceOp.MAX) self.bin_edges = linspace(start, end, self.number_of_bins) self.counts = _count_bins(input, self.bin_edges) dist.all_reduce(self.counts, op=dist.ReduceOp.SUM) return self.bin_edges, self.counts else: self.bin_edges, self.counts = histogram(input, bins=self.number_of_bins) return self.bin_edges, self.counts def update(self, input: Tensor) -> Tuple[Tensor, Tensor]: """Update current histogram with new data Parameters ---------- inputs : Tensor Input data tensor [B, ...] Returns ------- Tuple[Tensor, Tensor] The calculated (bin edges [N+1, ...], counts [N, ...]) tensors """ self._check_shape(input) # TODO(Dallas) Move distributed calls into finalize. self.bin_edges, self.counts = _update_bins_counts( input, self.bin_edges, self.counts ) self.number_of_bins = self.bin_edges.shape[0] return self.bin_edges, self.counts def finalize(self, cdf: bool = False) -> Tuple[Tensor, Tensor]: """Finalize the histogram, which computes the pdf and cdf Parameters ---------- cdf : bool, optional Compute a cumulative density function; will calculate probability density function otherwise, by default False Returns ------- Tuple[Tensor, Tensor] The calculated (bin edges [N+1, ...], PDF or CDF [N, ...]) tensors """ # Normalize counts hist_norm = self.counts.sum(dim=0) self.pdf = self.counts / hist_norm if cdf: self.cdf = torch.cumsum(self.pdf, dim=0) return self.bin_edges, self.cdf else: return self.bin_edges, self.pdf def normal_pdf( mean: Tensor, std: Tensor, bin_edges: Tensor, grid: str = "midpoint" ) -> Tensor: """Computes the probability density function of a normal variable with given mean and standard deviation. This PDF is given at the locations given by the midpoint of the bin_edges. This function uses the standard formula: .. math: \\frac{1}{\\sqrt{2*\\pi} std } \\exp( -\\frac{1}{2} (\\frac{x-mean}{std})^2 ) where erf is the error function. Parameters ---------- mean : Tensor mean tensor std : Tensor standard deviation tensor bins : Tensor The tensor of bin edges with dimension [N+1, ...] where N is the number of bins. grid : str A string that indicates where in the bins should the cdf be defined. Can be one of {"midpoint", "left", "right"}. Returns ------- Tensor The calculated cdf tensor with dimension [N, ...] """ if grid == "midpoint": bin_mids = 0.5 * (bin_edges[:-1] + bin_edges[1:]) elif grid == "right": bin_mids = bin_edges[1:] elif grid == "left": bin_mids = bin_edges[:-1] else: assert ValueError( "This type of grid is not defined. Choose one of {'mids', 'right', 'left'}." ) return ( torch.exp(-0.5 * ((bin_mids - mean[None, ...]) / std[None, ...]) ** 2) / std[None, ...] / torch.sqrt(torch.as_tensor(2.0 * torch.pi)) ) def normal_cdf( mean: Tensor, std: Tensor, bin_edges: Tensor, grid: str = "midpoint", ) -> Tensor: """Computes the cumulative density function of a normal variable with given mean and standard deviation. This CDF is given at the locations given by the midpoint of the bin_edges. This function uses the standard formula: .. math: \\frac{1}{2} ( 1 + erf( \\frac{x-mean}{std \\sqrt{2}}) ) ) where erf is the error function. Parameters ---------- mean : Tensor mean tensor std : Tensor standard deviation tensor bins : Tensor The tensor of bin edges with dimension [N+1, ...] where N is the number of bins. grid : str A string that indicates where in the bins should the cdf be defined. Can be one of {"mids", "left", "right"}. Returns ------- Tensor The calculated cdf tensor with dimension [N, ...] """ if grid == "midpoint": bin_mids = 0.5 * (bin_edges[:-1] + bin_edges[1:]) elif grid == "right": bin_mids = bin_edges[1:] elif grid == "left": bin_mids = bin_edges[:-1] else: assert ValueError( "This type of grid is not defined. Choose one of {'mids', 'right', 'left'}." ) return 0.5 + 0.5 * torch.erf( (bin_mids - mean[None, ...]) / (torch.sqrt(torch.as_tensor([2.0], device=mean.device)) * std[None, ...]) )
modulus-main
modulus/metrics/general/histogram.py
# Copyright (c) 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
modulus-main
modulus/metrics/general/__init__.py
# Copyright (c) 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch import torch.distributed as dist from typing import Union, Tuple, List from abc import ABC, abstractmethod from modulus.distributed.manager import DistributedManager from warnings import warn Tensor = torch.Tensor class EnsembleMetrics(ABC): """Abstract class for ensemble performance related metrics Can be helpful for distributed and sequential computations of metrics. Parameters ---------- input_shape : Union[Tuple[int,...], List] Shape of input tensors without batched dimension. device : torch.device, optional Pytorch device model is on, by default 'cpu' dtype : torch.dtype, optional Standard dtype to initialize any tensor with """ def __init__( self, input_shape: Union[Tuple[int, ...], List[int]], device: Union[str, torch.device] = "cpu", dtype: torch.dtype = torch.float32, ): super().__init__() self.input_shape = list(input_shape) self.device = torch.device(device) self.dtype = dtype if DistributedManager.is_initialized() and not (dist.is_initialized()): warn( "DistributedManager is detected and initialized but torch distributed \ process group is not initialized. In order to use this class, please initialize \ torch process group, see https://pytorch.org/docs/stable/distributed.html" ) def _check_shape(self, inputs: Tensor) -> None: """ Check input shapes for non-batched dimension. """ assert [i == s for (i, s) in zip(inputs.shape[1:], self.input_shape)], ( "Expected new input to have compatible shape with existing shapes but got" + str(inputs.shape) + "and" + str(self.input_shape) + "." ) def __call__(self, *args): """ Initial calculation for stored information. Will also overwrite previously stored data. """ raise NotImplementedError("Class member must implement a __call__ method.") def update(self, *args): """ Update initial or stored calculation with additional information. """ raise NotImplementedError("Class member must implement an update method.") def finalize(self, *args): """ Marks the end of the sequential calculation, used to finalize any computations. """ raise NotImplementedError("Class member must implement a finalize method.") def _update_mean( old_sum: Tensor, old_n: Union[int, Tensor], inputs: Tensor, batch_dim: Union[int, None] = 0, ) -> Tuple[Tensor, Union[int, Tensor]]: """Updated mean sufficient statistics given new data This method updates a running sum and number of samples with new (possibly batched) inputs Parameters ---------- old_sum : Tensor Current, or old, running sum old_n : Union[int, Tensor] Current, or old, number of samples input : Tensor New input to add to current/old sum. May be batched, in which case the batched dimension must be flagged by passing an int to batch_dim. batch_dim : Union[int, None], optional Whether the new inputs contain a batch of new inputs and what dimension they are stored along. Will reduce all elements if None, by default 0. Returns ------- Tuple[Tensor, Union[int, Tensor]] Updated (rolling sum, number of samples) """ if batch_dim is None: inputs = torch.unsqueeze(inputs, 0) batch_dim = 0 new_sum = old_sum + torch.sum(inputs, dim=batch_dim) new_n = old_n + inputs.shape[batch_dim] return new_sum, new_n class Mean(EnsembleMetrics): """Utility class that computes the mean over a batched or ensemble dimension This is particularly useful for distributed environments and sequential computation. Parameters ---------- input_shape : Union[Tuple, List] Shape of broadcasted dimensions """ def __init__(self, input_shape: Union[Tuple, List], **kwargs): super().__init__(input_shape, **kwargs) self.sum = torch.zeros(self.input_shape, dtype=self.dtype, device=self.device) self.n = torch.zeros([1], dtype=torch.int32, device=self.device) def __call__(self, inputs: Tensor, dim: int = 0) -> Tensor: """Calculate an initial mean Parameters ---------- inputs : Tensor Input data dim : Int Dimension of batched data Returns ------- Tensor Mean value """ assert ( inputs.device == self.device ), f"Input device, {inputs.device}, and Module device, {self.device}, must be the same." self.sum = torch.sum(inputs, dim=dim) self.n = torch.as_tensor([inputs.shape[dim]], device=self.device) # TODO(Dallas) Move distributed calls into finalize. if DistributedManager.is_initialized() and dist.is_initialized(): dist.all_reduce(self.sum, op=dist.ReduceOp.SUM) dist.all_reduce(self.n, op=dist.ReduceOp.SUM) return self.sum / self.n def update(self, inputs: Tensor, dim: int = 0) -> Tensor: """Update current mean and essential statistics with new data Parameters ---------- inputs : Tensor Inputs tensor dim : int Dimension of batched data Returns ------- Tensor Current mean value """ self._check_shape(inputs) assert ( inputs.device == self.device ), f"Inputs device, {inputs.device}, and Module device, {self.device}, must be the same." # TODO(Dallas) Move distributed calls into finalize. if DistributedManager.is_initialized() and dist.is_initialized(): # Collect local sums, n sums = torch.sum(inputs, batch_dim=dim) n = torch.as_tensor([inputs.shape[dim]], device=self.device) # Reduce dist.all_reduce(sums, op=dist.ReduceOp.SUM) dist.all_reduce(n, op=dist.ReduceOp.SUM) # Update self.sum += sums self.n += n else: self.sum, self.n = _update_mean(self.sum, self.n, inputs, batch_dim=dim) return self.sum / self.n def finalize( self, ) -> Tensor: """Compute and store final mean Returns ------- Tensor Final mean value """ self.mean = self.sum / self.n return self.mean def _update_var( old_sum: Tensor, old_sum2: Tensor, old_n: Union[int, Tensor], inputs: Tensor, batch_dim: Union[int, None] = 0, ) -> Tuple[Tensor, Tensor, Union[int, Tensor]]: """Updated variance sufficient statistics given new data This method updates a running running sum, squared sum, and number of samples with new (possibly batched) inputs Parameters ---------- old_sum : Tensor Current, or old, running sum old_sum2 : Tensor Current, or old, running squared sum old_n : Union[int, Tensor] Current, or old, number of samples inputs : Tensor New input to add to current/old sum. May be batched, in which case the batched dimension must be flagged by passing an int to batch_dim. batch_dim : Union[int, None], optional Whether the new inputs contain a batch of new inputs and what dimension they are stored along. Will reduce all elements if None, by default 0. Returns ------- Tuple[Tensor, Tensor, Union[int, Tensor]] Updated (rolling sum, rolling squared sum, number of samples) Note ---- See "Updating Formulae and a Pairwise Algorithm for Computing Sample Variances" by Chan et al. http://i.stanford.edu/pub/cstr/reports/cs/tr/79/773/CS-TR-79-773.pdf for details. """ if batch_dim is None: inputs = torch.unsqueeze(inputs, 0) batch_dim = 0 temp_n = inputs.shape[batch_dim] temp_sum = torch.sum(inputs, dim=batch_dim) temp_sum2 = torch.sum((inputs - temp_sum / temp_n) ** 2, dim=batch_dim) delta = old_sum * temp_n / old_n - temp_sum new_sum = old_sum + temp_sum new_sum2 = old_sum2 + temp_sum2 new_sum2 += old_n / temp_n / (old_n + temp_n) * delta**2 new_n = old_n + temp_n return new_sum, new_sum2, new_n class Variance(EnsembleMetrics): """Utility class that computes the variance over a batched or ensemble dimension This is particularly useful for distributed environments and sequential computation. Parameters ---------- input_shape : Union[Tuple, List] Shape of broadcasted dimensions Note ---- See "Updating Formulae and a Pairwise Algorithm for Computing Sample Variances" by Chan et al. http://i.stanford.edu/pub/cstr/reports/cs/tr/79/773/CS-TR-79-773.pdf for details. """ def __init__(self, input_shape: Union[Tuple, List], **kwargs): super().__init__(input_shape, **kwargs) self.n = torch.zeros([1], dtype=torch.int32, device=self.device) self.sum = torch.zeros(self.input_shape, dtype=self.dtype, device=self.device) self.sum2 = torch.zeros(self.input_shape, dtype=self.dtype, device=self.device) def __call__(self, inputs: Tensor, dim: int = 0) -> Tensor: """Calculate an initial variance Parameters ---------- inputs : Tensor Input data dim : Int Dimension of batched data Returns ------- Tensor Unbiased variance values """ assert ( inputs.device == self.device ), f"Input device, {inputs.device}, and Module device, {self.device}, must be the same." self.sum = torch.sum(inputs, dim=dim) self.n = torch.as_tensor([inputs.shape[0]], device=self.device) if DistributedManager.is_initialized() and dist.is_initialized(): # Compute mean and send around. dist.all_reduce(self.sum, op=dist.ReduceOp.SUM) dist.all_reduce(self.n, op=dist.ReduceOp.SUM) self.sum2 = torch.sum((inputs - self.sum / self.n) ** 2, dim=dim) dist.all_reduce(self.sum2, op=dist.ReduceOp.SUM) else: self.sum2 = torch.sum((inputs - self.sum / self.n) ** 2, dim=dim) if self.n < 2.0: return self.sum2 else: return self.sum2 / (self.n - 1.0) def update(self, inputs: Tensor) -> Tensor: """Update current variance value and essential statistics with new data Parameters ---------- inputs : Tensor Input data Returns ------- Tensor Unbiased variance tensor """ self._check_shape(inputs) assert ( inputs.device == self.device ), f"Input device, {inputs.device}, and Module device, {self.device}, must be the same." new_n = torch.as_tensor([inputs.shape[0]], device=self.device) new_sum = torch.sum(inputs, dim=0) # TODO(Dallas) Move distributed calls into finalize. if DistributedManager.is_initialized() and dist.is_initialized(): dist.all_reduce(new_n, op=dist.ReduceOp.SUM) dist.all_reduce(new_sum, op=dist.ReduceOp.SUM) new_sum2 = torch.sum((inputs - new_sum / new_n) ** 2, dim=0) dist.all_reduce(new_sum2, op=dist.ReduceOp.SUM) else: # Calculate new statistics new_sum2 = torch.sum((inputs - new_sum / new_n) ** 2, dim=0) delta = self.sum * new_n / self.n - new_sum # Update self.sum += new_sum self.sum2 += new_sum2 self.sum2 += self.n / new_n / (self.n + new_n) * (delta) ** 2 self.n += new_n if self.n < 2.0: return self.sum2 else: return self.sum2 / (self.n - 1.0) @property def mean(self) -> Tensor: """Mean value""" return self.sum / self.n def finalize(self, std: bool = False) -> Tuple[Tensor, Tensor]: """Compute and store final mean and unbiased variance / std Parameters ---------- std : bool, optional Compute standard deviation, by default False Returns ------- Tensor Final (mean, variance/std) value """ assert ( self.n > 1.0 ), "Error! In order to finalize, there needs to be at least 2 samples." self.var = self.sum2 / (self.n - 1.0) if std: self.std = torch.sqrt(self.var) return self.std else: return self.var
modulus-main
modulus/metrics/general/ensemble_metrics.py
# Copyright (c) 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch Tensor = torch.Tensor def wasserstein(bin_edges: Tensor, cdf_x: Tensor, cdf_y: Tensor) -> Tensor: """1-Wasserstein distance between two discrete CDF functions This norm is typically used to compare two different forecast ensembles (for X and Y). Creates a map of distance and does not accumulate over lat/lon regions. Computes W(F_X, F_Y) = int[ |F_X(x) - F_Y(x)| ] dx where F_X is the empirical cdf of X and F_Y is the empirical cdf of Y. Parameters ---------- bin_edges : Tensor Tensor containing bin edges. The leading dimension must represent the N+1 bin edges. cdf_x : Tensor Tensor containing a CDF one, defined over bins. The non-zeroth dimensions of bins and cdf must be compatible. cdf_y : Tensor Tensor containing a CDF two, defined over bins. Must be compatible with cdf_x in terms of bins and shape. Returns ------- Tensor The 1-Wasserstein distance between cdf_x and cdf_y """ return torch.sum( torch.abs(cdf_x - cdf_y) * (bin_edges[1, ...] - bin_edges[0, ...]), dim=0 )
modulus-main
modulus/metrics/general/wasserstein.py
# Copyright (c) 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch import numpy as np from typing import Union from .histogram import cdf as cdf_function Tensor = torch.Tensor def _crps_gaussian(mean: Tensor, std: Tensor, obs: Union[Tensor, np.ndarray]) -> Tensor: """Computes the local Continuous Ranked Probability Score (CRPS) using assuming that the forecast distribution is normal. Creates a map of CRPS and does not accumulate over lat/lon regions. Computes: .. math: CRPS(mean, std, y) = std * [ \\frac{1}{\\sqrt{\\pi}}} - 2 \\phi ( \\frac{x-mean}{std} ) - ( \\frac{x-mean}{std} ) * (2 \\Phi(\\frac{x-mean}{std}) - 1) ] where \\phi and \\Phi are the normal gaussian pdf/cdf respectively. Parameters ---------- mean : Tensor Tensor of mean of forecast distribution. std : Tensor Tensor of standard deviation of forecast distribution. obs : Union[Tensor, np.ndarray] Tensor or array containing an observation over which the CRPS is computed with respect to. Broadcasting dimensions must be compatible with the non-zeroth dimensions of bins and cdf. Returns ------- Tensor Map of CRPS """ if isinstance(obs, np.ndarray): obs = torch.from_numpy(obs).to(mean.device) # Check shape compatibility assert mean.shape == std.shape, ( "Mean and standard deviation must have" + "compatible shapes but found" + str(mean.shape) + " and " + str(std.shape) + "." ) assert mean.shape == obs.shape, ( "Mean and obs must have" + "compatible shapes but found" + str(mean.shape) + " and " + str(obs.shape) + "." ) d = (obs - mean) / std phi = torch.exp(-0.5 * d**2) / torch.sqrt(torch.as_tensor(2 * torch.pi)) # Note, simplified expression below is not exactly Gaussian CDF Phi = torch.erf(d / torch.sqrt(torch.as_tensor(2.0))) return std * (2 * phi + d * Phi - 1.0 / torch.sqrt(torch.as_tensor(torch.pi))) def _crps_from_cdf( bin_edges: Tensor, cdf: Tensor, obs: Union[Tensor, np.ndarray] ) -> Tensor: """Computes the local Continuous Ranked Probability Score (CRPS) using a cumulative distribution function. Creates a map of CRPS and does not accumulate over lat/lon regions. Computes: CRPS(X, y) = int[ (F(x) - 1[x - y])^2 ] dx where F is the empirical cdf of X. Parameters ---------- bins_edges : Tensor Tensor [N+1, ...] containing bin edges. The leading dimension must represent the N+1 bin edges. cdf : Tensor Tensor [N, ...] containing a cdf, defined over bins. The non-zeroth dimensions of bins and cdf must be compatible. obs : Union[Tensor, np.ndarray] Tensor or array containing an observation over which the CRPS is computed with respect to. Broadcasting dimensions must be compatible with the non-zeroth dimensions of bins and cdf. Returns ------- Tensor Map of CRPS """ if isinstance(obs, np.ndarray): obs = torch.from_numpy(obs).to(cdf.device) assert bin_edges.shape[1:] == cdf.shape[1:], ( "Expected bins and cdf to have compatible non-zeroth dimensions but have shapes" + str(bin_edges.shape[1:]) + " and " + str(cdf.shape[1:]) + "." ) assert bin_edges.shape[1:] == obs.shape, ( "Expected bins and observations to have compatible broadcasting dimensions but have shapes" + str(bin_edges.shape[1:]) + " and " + str(obs.shape) + "." ) assert bin_edges.shape[0] == cdf.shape[0] + 1, ( "Expected zeroth dimension of cdf to be equal to the zeroth dimension of bins + 1 but have shapes" + str(bin_edges.shape[0]) + " and " + str(cdf.shape[0]) + "+1." ) dbins = bin_edges[1, ...] - bin_edges[0, ...] bin_mids = 0.5 * (bin_edges[1:] + bin_edges[:-1]) obs = torch.ge(bin_mids, obs).int() return torch.sum(torch.abs(cdf - obs) ** 2 * dbins, dim=0) def _crps_from_counts( bin_edges: Tensor, counts: Tensor, obs: Union[Tensor, np.ndarray] ) -> Tensor: """Computes the local Continuous Ranked Probability Score (CRPS) using a histogram of counts. Creates a map of CRPS and does not accumulate over lat/lon regions. Computes: CRPS(X, y) = int[ (F(x) - 1[x - y])^2 ] dx where F is the empirical cdf of X. Parameters ---------- bins_edges : Tensor Tensor [N+1, ...] containing bin edges. The leading dimension must represent the N+1 bin edges. counts : Tensor Tensor [N, ...] containing counts, defined over bins. The non-zeroth dimensions of bins and counts must be compatible. obs : Union[Tensor, np.ndarray] Tensor or array containing an observation over which the CRPS is computed with respect to. Broadcasting dimensions must be compatible with the non-zeroth dimensions of bins and counts. Returns ------- Tensor Map of CRPS """ if isinstance(obs, np.ndarray): obs = torch.from_numpy(obs).to(counts.device) assert bin_edges.shape[1:] == counts.shape[1:], ( "Expected bins and cdf to have compatible non-zeroth dimensions but have shapes" + str(bin_edges.shape[1:]) + " and " + str(counts.shape[1:]) + "." ) assert bin_edges.shape[1:] == obs.shape, ( "Expected bins and observations to have compatible broadcasting dimensions but have shapes" + str(bin_edges.shape[1:]) + " and " + str(obs.shape) + "." ) assert bin_edges.shape[0] == counts.shape[0] + 1, ( "Expected zeroth dimension of cdf to be equal to the zeroth dimension of bins + 1 but have shapes" + str(bin_edges.shape[0]) + " and " + str(counts.shape[0]) + "+1." ) cdf_hat = torch.cumsum(counts / torch.sum(counts, dim=0), dim=0) return _crps_from_cdf(bin_edges, cdf_hat, obs) def crps( pred: Tensor, obs: Union[Tensor, np.ndarray], bins: Union[int, Tensor] = 10 ) -> Tensor: """ Computes the local Continuous Ranked Probability Score (CRPS) by computing a histogram and CDF of the predictions. Creates a map of CRPS and does not accumulate over lat/lon regions. Computes: CRPS(X, y) = int[ (F(x) - 1[x - y])^2 ] dx where F is the empirical cdf of X. Parameters ---------- pred : Tensor Tensor [B, ...] containing the ensemble predictions. The leading dimension must represent the ensemble dimension. obs : Union[Tensor, np.ndarray] Tensor or array containing an observation over which the CRPS is computed with respect to. Broadcasting dimensions must be compatible with the non-zeroth dimensions of bins and cdf. bins : Union[int, Tensor], optional Either the number of bins, or a tensor of bin edges with dimension [N+1, ...] where N is the number of bins, by default 10. Returns ------- Tensor Map of CRPS """ bin_edges, cdf = cdf_function(pred, bins=bins) return _crps_from_cdf(bin_edges, cdf, obs)
modulus-main
modulus/metrics/general/crps.py
# Copyright (c) 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # TODO(Dallas) Introduce Distributed Class for computation. import torch import numpy as np from .histogram import histogram from typing import Union Tensor = torch.Tensor def _entropy_from_counts(p: Tensor, bin_edges: Tensor, normalized=True) -> Tensor: """Computes the Statistical Entropy of a random variable using a histogram. Uses the formula: .. math: Entropy(X) = \\int p(x) * \\log( p(x) ) dx Parameters ---------- p : Tensor Tensor [N, ...] containing counts/pdf, defined over bins. The non-zeroth dimensions of bin_edges and p must be compatible. bins_edges : Tensor Tensor [N+1, ...] containing bin edges. The leading dimension must represent the N+1 bin edges. normalized : Bool, Optional Boolean flag determining whether the returned statistical entropy is normalized. Normally the entropy for a compact bounded probability distribution is bounded between a pseudo-dirac distribution, ent_min, and a uniform distribution, ent_max. This normalization transforms the entropy from [ent_min, ent_max] to [0, 1] Returns ------- Tensor Tensor containing the Information/Statistical Entropy """ assert bin_edges.shape[1:] == p.shape[1:], ( "Expected bins and pdf to have compatible non-zeroth dimensions but have shapes" + str(bin_edges.shape[1:]) + " and " + str(p.shape[1:]) + "." ) assert bin_edges.shape[0] == p.shape[0] + 1, ( "Expected zeroth dimension of cdf to be equal to the zeroth dimension of bins + 1 but have shapes" + str(bin_edges.shape[0]) + " and " + str(p.shape[0]) + "+1." ) dbins = bin_edges[1:] - bin_edges[:-1] bin_mids = 0.5 * (bin_edges[1:] + bin_edges[:-1]) p = p / torch.trapz(p, bin_mids, dim=0) + 1e-8 ent = torch.trapz(-1.0 * p * torch.log(p), bin_mids, dim=0) if normalized: max_ent = torch.log(bin_edges[-1] - bin_edges[0]) min_ent = 0.5 + 0.5 * torch.log(2 * torch.pi * dbins[0] ** 2) return (ent - min_ent) / (max_ent - min_ent) else: return ent def _relative_entropy_from_counts( p: Tensor, q: Tensor, bin_edges: Tensor, ) -> Tensor: """Computes the Relative Statistical Entropy, or KL Divergence of two random variables using their histograms. Uses the formula: .. math: Entropy(X) = \\int p(x) * \\log( p(x)/q(x) ) dx Parameters ---------- p : Tensor Tensor [N, ...] containing counts/pdf, defined over bins. The non-zeroth dimensions of bin_edges and p must be compatible. q : Tensor Tensor [N, ...] containing counts/pdf, defined over bins. The non-zeroth dimensions of bin_edges and q must be compatible. bins_edges : Tensor Tensor [N+1, ...] containing bin edges. The leading dimension must represent the N+1 bin edges. Returns ------- Tensor Map of Statistical Entropy """ assert bin_edges.shape[1:] == p.shape[1:], ( "Expected bins and pdf to have compatible non-zeroth dimensions but have shapes" + str(bin_edges.shape[1:]) + " and " + str(p.shape[1:]) + "." ) assert bin_edges.shape[0] == p.shape[0] + 1, ( "Expected zeroth dimension of cdf to be equal to the zeroth dimension of bins + 1 but have shapes" + str(bin_edges.shape[0]) + " and " + str(p.shape[0]) + "+1." ) assert p.shape == q.shape, ( "Expected p and q to have compatible shapes but have shapes" + str(p.shape) + " and " + str(q.shape) + "." ) dbins = bin_edges[1:] - bin_edges[:-1] bin_mids = 0.5 * (bin_edges[1:] + bin_edges[:-1]) p = p / torch.trapz(p, bin_mids, dim=0) + 1e-8 q = q / torch.trapz(q, bin_mids, dim=0) + 1e-8 return torch.trapz(p * torch.log(p / q), bin_mids, dim=0)
modulus-main
modulus/metrics/general/entropy.py
# Copyright (c) 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch import numpy as np from modulus.metrics.general.histogram import histogram, linspace from typing import Union Tensor = torch.Tensor def find_rank( bin_edges: Tensor, counts: Tensor, obs: Union[Tensor, np.ndarray] ) -> Tensor: """Finds the rank of the observation with respect to the given counts and bins. Parameters ---------- bins_edges : Tensor Tensor [N+1, ...] containing bin edges. The leading dimension must represent the N+1 bin edges. counts : Tensor Tensor [N, ...] containing counts, defined over bins. The non-zeroth dimensions of bins and counts must be compatible. obs : Union[Tensor, np.ndarray] Tensor or array containing an observation over which the CRPS is computed with respect to. Broadcasting dimensions must be compatible with the non-zeroth dimensions of bins and counts. Returns ------- Tensor Tensor of rank for eac of the batched dimensions [...] """ if isinstance(obs, np.ndarray): obs = torch.from_numpy(obs).to(counts.device) assert bin_edges.shape[1:] == counts.shape[1:], ( "Expected bins and counts to have compatible non-zeroth dimensions but have shapes" + str(bin_edges.shape[1:]) + " and " + str(counts.shape[1:]) + "." ) assert bin_edges.shape[1:] == obs.shape, ( "Expected bins and observations to have compatible broadcasting dimensions but have shapes" + str(bin_edges.shape[1:]) + " and " + str(obs.shape) + "." ) assert bin_edges.shape[0] == counts.shape[0] + 1, ( "Expected zeroth dimension of counts to be equal to the zeroth dimension of bins + 1 but have shapes" + str(bin_edges.shape[0]) + " and " + str(counts.shape[0]) + "+1." ) n = torch.sum(counts, dim=0)[0] bin_mids = 0.5 * (bin_edges[1:] + bin_edges[:-1]) right = torch.sum(counts * (bin_mids <= obs[None, ...]), dim=0) return right / n def _rank_probability_score_from_counts( rank_bin_edges: Tensor, rank_counts: Tensor ) -> Tensor: """Finds the rank of the observation with respect to the given counts and bins. Computes .. math: 3 * \int_0^1 (F_X(x) - F_U(x))^2 dx where F represents a cumulative distribution function, X represents the rank distribution and U represents a Uniform distribution. Parameters ---------- rank_bins_edges : Tensor Tensor [N+1, ...] containing rank bin edges. The leading dimension must represent the N+1 bin edges. rank_counts : Tensor Tensor [N, ...] containing rank counts, defined over bins. The non-zeroth dimensions of bin edges and counts must be compatible. Returns ------- Tensor Tensor of the Ranked Probability Score for each batched dimension of the input. """ cdf = torch.cumsum(rank_counts, dim=0) cdf = cdf / cdf[-1] normalization = torch.sum((1.0 - rank_bin_edges[1:]) ** 2, dim=0) return torch.sum((cdf - rank_bin_edges[1:]) ** 2, dim=0) / normalization def rank_probability_score(ranks: Tensor) -> Tensor: """Computes the Rank Probability Score for the passed ranks. Internally, this creates a histogram for the ranks and computes the Rank Probability Score (RPS) using the histogram. With the histogram the RPS is computed as .. math: \int_0^1 (F_X(x) - F_U(x))^2 dx where F represents a cumulative distribution function, X represents the rank distribution and U represents a Uniform distribution. For computation of the ranks, use _find_rank. Parameters ---------- ranks : Tensor Tensor [B, ...] containing ranks, where the leading dimension represents the batch, or ensemble, dimension. The non-zeroth dimensions are batched over. Returns ------- Tensor Tensor of RPS for each of the batched dimensions [...] """ start = 0.0 * ranks[0, ...] end = start + 1.0 bins = linspace(start, end, 10) bin_edges, bin_counts = histogram(ranks, bins=bins) return _rank_probability_score_from_counts(bin_edges, bin_counts)
modulus-main
modulus/metrics/general/calibration.py
# Copyright (c) 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # TODO(Dallas) Introduce Distributed Class for computation. import torch from modulus.metrics.general.histogram import normal_pdf, normal_cdf, histogram from modulus.metrics.general.entropy import _entropy_from_counts Tensor = torch.Tensor def efi( pred_cdf: Tensor, bin_edges: Tensor, climatology_mean: Tensor, climatology_std: Tensor, ) -> Tensor: """Calculates the Extreme Forecast Index (EFI) for an ensemble forecast against a climatological distribution. Parameters ---------- pred_cdf : Tensor Cumulative distribution function of predictions of shape [N, ...] where N is the number of bins. This cdf must be defined over the passed bin_edges. bin_edges : Tensor Tensor of bin edges with shape [N+1, ...] where N is the number of bins. climatology_mean : Tensor Tensor of climatological mean with shape [...] climatology_std : Tensor Tensor of climatological std with shape [...] Returns ------- Tensor EFI values of each of the batched dimensions. Note ---- Reference: https://www.atmos.albany.edu/daes/atmclasses/atm401/spring_2016/ppts_pdfs/ECMWF_EFI.pdf """ clim_cdf = normal_cdf(climatology_mean, climatology_std, bin_edges, grid="right") return torch.trapz( (clim_cdf - pred_cdf) / torch.sqrt(1e-8 + clim_cdf * (1.0 - clim_cdf)), clim_cdf, dim=0, ) def normalized_entropy( pred_pdf: Tensor, bin_edges: Tensor, climatology_mean: Tensor, climatology_std: Tensor, ) -> Tensor: """Calculates the relative entropy, or surprise, of using the prediction distribution as opposed to the climatology distribution. Parameters ---------- pred_cdf : Tensor Cumulative distribution function of predictions of shape [N, ...] where N is the number of bins. This cdf must be defined over the passed bin_edges. bin_edges : Tensor Tensor of bin edges with shape [N+1, ...] where N is the number of bins. climatology_mean : Tensor Tensor of climatological mean with shape [...] climatology_std : Tensor Tensor of climatological std with shape [...] Returns ------- Tensor Relative Entropy values of each of the batched dimensions. Note ---- Reference: https://www.atmos.albany.edu/daes/atmclasses/atm401/spring_2016/ppts_pdfs/ECMWF_EFI.pdf """ clim_pdf = normal_pdf(climatology_mean, climatology_std, bin_edges, grid="right") return 1.0 - _entropy_from_counts(pred_pdf, bin_edges) / _entropy_from_counts( clim_pdf, bin_edges )
modulus-main
modulus/metrics/climate/efi.py
# Copyright (c) 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from torch import Tensor from abc import ABC from modulus.metrics.general.reduction import WeightedMean, WeightedVariance def _compute_lat_weights(lat: Tensor) -> Tensor: """Computes weighting for latitude reduction Parameters ---------- lat : Tensor A one-dimension tensor [H] representing the latitudes at which the function will return weights for Returns ------- Tensor Latitude weight tensor [H] """ nlat = len(lat) lat_weight = torch.abs(torch.cos(torch.pi * (lat / 180))) lat_weight = lat_weight / lat_weight.sum() return lat_weight def zonal_mean(x: Tensor, lat: Tensor, dim: int = -2, keepdims: bool = False) -> Tensor: """Computes zonal mean, weighting over the latitude direction that is specified by dim Parameters ---------- x : Tensor The tensor [..., H, W] over which the mean will be computed lat : Tensor A one-dimension tensor representing the latitudes at which the function will return weights for dim : int, optional The int specifying which dimension of x the reduction will occur, by default -2 keepdims : bool, optional Keep aggregated dimension, by default False Returns ------- Tensor Zonal mean tensor of x over the latitude dimension """ weights = _compute_lat_weights(lat) wm = WeightedMean(weights) return wm(x, dim=dim, keepdims=keepdims) def zonal_var( x: Tensor, lat: Tensor, std: bool = False, dim: int = -2, keepdims: bool = False, ) -> Tensor: """Computes zonal variance, weighting over the latitude direction Parameters ---------- x : Tensor The tensor [..., H, W] over which the variance will be computed lat : Tensor A one-dimension tensor [H] representing the latitudes at which the function will return weights for std : bool, optional Return zonal standard deviation, by default False dim : int, optional The int specifying which dimension of x the reduction will occur, by default -2 keepdims : bool, optional Keep aggregated dimension, by default False Returns ------- Tensor The variance (or standard deviation) of x over the latitude dimension """ weights = _compute_lat_weights(lat) ws = WeightedVariance(weights) var = ws(x, dim=dim, keepdims=keepdims) if std: return torch.sqrt(var) else: return var def global_mean(x: Tensor, lat: Tensor, keepdims: bool = False) -> Tensor: """Computes global mean This function computs the global mean of a lat/lon grid by weighting over the latitude direction and then averaging over longitude Parameters ---------- x : Tensor The lat/lon tensor [..., H, W] over which the mean will be computed lat : Tensor A one-dimension tensor [H] representing the latitudes at which the function will return weights for keepdims : bool, optional Keep aggregated dimension, by default False Returns ------- Tensor Global mean tensor """ assert ( x.ndim > 2 ), "Expected x to have at least two dimensions, with the last two dimensions representing lat and lon respectively" # Mean out the latitudes lat_reduced = zonal_mean(x, lat, dim=-2, keepdims=keepdims) # Return after reduction across longitudes return torch.mean(lat_reduced, dim=-1, keepdims=keepdims) def global_var( x: Tensor, lat: Tensor, std: bool = False, keepdims: bool = False, ) -> Tensor: """Computes global variance This function computs the global variance of a lat/lon grid by weighting over the latitude direction and then averaging over longitude Parameters ---------- x : Tensor The lat/lon tensor [..., H, W] over which the variance will be computed lat : Tensor A one-dimension tensor [H] representing the latitudes at which the function will return weights for std : bool, optional Return global standard deviation, by default False keepdims : bool, optional Keep aggregated dimension, by default False Returns ------- Tensor Global variance tensor """ assert ( x.ndim > 2 ), "Expected x to have at least two dimensions, with the last two dimensions representing lat and lon respectively" # Take global mean, incorporated weights gm = global_mean(x, lat, keepdims=True) # Take var of lat lat_reduced = zonal_mean((x - gm) ** 2, lat, dim=-2, keepdims=keepdims) # Take var over longitude long_reduce = torch.mean(lat_reduced, dim=-1, keepdims=keepdims) if std: return torch.sqrt(long_reduce) else: return long_reduce
modulus-main
modulus/metrics/climate/reduction.py
# Copyright (c) 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
modulus-main
modulus/metrics/climate/__init__.py
# Copyright (c) 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # TODO(Dallas) Introduce Distributed Class for computation. import torch from modulus.metrics.climate.reduction import _compute_lat_weights Tensor = torch.Tensor def acc(pred: Tensor, target: Tensor, climatology: Tensor, lat: Tensor) -> Tensor: """Calculates the Anomaly Correlation Coefficient Parameters ---------- pred : Tensor [..., H, W] Predicted tensor on a lat/long grid target : Tensor [..., H, W] Target tensor on a lat/long grid climatology : Tensor [..., H, W] climatology tensor lat : Tensor [H] latitude tensor Returns ------- Tensor ACC values for each field Note ---- Reference: https://www.atmos.albany.edu/daes/atmclasses/atm401/spring_2016/ppts_pdfs/ECMWF_ACC_definition.pdf """ assert ( pred.ndim > 2 ), "Expected predictions to have at least two dimensions (lat, lon)" assert ( target.ndim > 2 ), "Expected targets to have at least two dimensions (lat, lon)" assert ( climatology.ndim > 2 ), "Expected climatology to have at least two dimensions (lat, lon)" # subtract climate means pred_hat = pred - climatology target_hat = target - climatology # Get aggregator lat_weight = _compute_lat_weights(lat) # Weighted mean pred_bar = torch.sum( lat_weight[:, None] * pred_hat, dim=(-2, -1), keepdim=True ) / torch.sum( lat_weight[:, None] * torch.ones_like(pred_hat), dim=(-2, -1), keepdim=True ) target_bar = torch.sum( lat_weight[:, None] * target_hat, dim=(-2, -1), keepdim=True ) / torch.sum( lat_weight[:, None] * torch.ones_like(target_hat), dim=(-2, -1), keepdim=True ) pred_diff = pred_hat - pred_bar target_diff = target_hat - target_bar p1 = torch.sum(lat_weight[:, None] * pred_diff * target_diff, dim=(-2, -1)) p2 = torch.sum(lat_weight[:, None] * pred_diff * pred_diff, dim=(-2, -1)) p3 = torch.sum(lat_weight[:, None] * target_diff * target_diff, dim=(-2, -1)) m = p1 / torch.sqrt(p2 * p3) return m
modulus-main
modulus/metrics/climate/acc.py