INSTRUCTION
stringlengths
1
46.3k
RESPONSE
stringlengths
75
80.2k
Plots the images which correspond to the selected class (y) and to the specific case (prediction is correct - is_true=True, prediction is wrong - is_true=False) Arguments: y (int): the selected class is_correct (boolean): a boolean flag (True, False) which specify the what t...
def plot_by_correct(self, y, is_correct): """ Plots the images which correspond to the selected class (y) and to the specific case (prediction is correct - is_true=True, prediction is wrong - is_true=False) Arguments: y (int): the selected class is_correct (boolean):...
Extracts the predicted classes which correspond to the selected class (y) and have probabilities nearest to 1/number_of_classes (eg. 0.5 for 2 classes, 0.33 for 3 classes) for the selected class. Arguments: y (int): the selected class Returns: idxs (numpy.ndarra...
def most_by_uncertain(self, y): """ Extracts the predicted classes which correspond to the selected class (y) and have probabilities nearest to 1/number_of_classes (eg. 0.5 for 2 classes, 0.33 for 3 classes) for the selected class. Arguments: y (int): the selected class ...
PyTorch distributed training launch helper that spawns multiple distributed processes
def main( gpus:Param("The GPUs to use for distributed training", str)='all', script:Param("Script to run", str, opt=False)='', args:Param("Args to pass to script", nargs='...', opt=False)='' ): "PyTorch distributed training launch helper that spawns multiple distributed processes" # Loosely based on...
Add the metrics names to the `Recorder`.
def on_train_begin(self, **kwargs): "Add the metrics names to the `Recorder`." self.names = ifnone(self.learn.loss_func.metric_names, []) if not self.names: warn('LossMetrics requested but no loss_func.metric_names provided') self.learn.recorder.add_metric_names(self.names)
Initialize the metrics for this epoch.
def on_epoch_begin(self, **kwargs): "Initialize the metrics for this epoch." self.metrics = {name:0. for name in self.names} self.nums = 0
Update the metrics if not `train`
def on_batch_end(self, last_target, train, **kwargs): "Update the metrics if not `train`" if train: return bs = last_target.size(0) for name in self.names: self.metrics[name] += bs * self.learn.loss_func.metrics[name].detach().cpu() self.nums += bs
Finish the computation and sends the result to the Recorder.
def on_epoch_end(self, last_metrics, **kwargs): "Finish the computation and sends the result to the Recorder." if not self.nums: return metrics = [self.metrics[name]/self.nums for name in self.names] return {'last_metrics': last_metrics+metrics}
Create the various optimizers.
def on_train_begin(self, **kwargs): "Create the various optimizers." self.G_A,self.G_B = self.learn.model.G_A,self.learn.model.G_B self.D_A,self.D_B = self.learn.model.D_A,self.learn.model.D_B self.crit = self.learn.loss_func.crit self.opt_G = self.learn.opt.new([nn.Sequential(*f...
Steps through the generators then each of the critics.
def on_batch_end(self, last_input, last_output, **kwargs): "Steps through the generators then each of the critics." self.G_A.zero_grad(); self.G_B.zero_grad() fake_A, fake_B = last_output[0].detach(), last_output[1].detach() real_A, real_B = last_input self._set_trainable(D_A=Tru...
Put the various losses in the recorder.
def on_epoch_end(self, last_metrics, **kwargs): "Put the various losses in the recorder." return add_metrics(last_metrics, [s.smooth for k,s in self.smootheners.items()])
Prepare file with metric names.
def on_train_begin(self, **kwargs: Any) -> None: "Prepare file with metric names." self.path.parent.mkdir(parents=True, exist_ok=True) self.file = self.path.open('a') if self.append else self.path.open('w') self.file.write(','.join(self.learn.recorder.names[:(None if self.add_time ...
Add a line with `epoch` number, `smooth_loss` and `last_metrics`.
def on_epoch_end(self, epoch: int, smooth_loss: Tensor, last_metrics: MetricsList, **kwargs: Any) -> bool: "Add a line with `epoch` number, `smooth_loss` and `last_metrics`." last_metrics = ifnone(last_metrics, []) stats = [str(stat) if isinstance(stat, int) else '#na#' if stat is None else f'{s...
Return two lists, one for the model parameters in FP16 and one for the master parameters in FP32.
def get_master(layer_groups:ModuleList, flat_master:bool=False) -> Tuple[List[List[Tensor]], List[List[Tensor]]]: "Return two lists, one for the model parameters in FP16 and one for the master parameters in FP32." split_params = split_no_wd_params(layer_groups) model_params = [[param for param in pg if para...
Copy the `model_params` gradients to `master_params` for the optimizer step.
def model_g2master_g(model_params:Sequence[Tensor], master_params:Sequence[Tensor], flat_master:bool=False)->None: "Copy the `model_params` gradients to `master_params` for the optimizer step." if flat_master: for model_group,master_group in zip(model_params,master_params): if len(master_gro...
Copy `master_params` to `model_params`.
def master2model(model_params:Sequence[Tensor], master_params:Sequence[Tensor], flat_master:bool=False)->None: "Copy `master_params` to `model_params`." if flat_master: for model_group,master_group in zip(model_params,master_params): if len(model_group) != 0: for model, maste...
Prepare the master model.
def on_train_begin(self, **kwargs:Any)->None: "Prepare the master model." #Get a copy of the model params in FP32 self.model_params, self.master_params = get_master(self.learn.layer_groups, self.flat_master) #Changes the optimizer so that the optimization step is done in FP32. ne...
Scale gradients up by `self.loss_scale` to prevent underflow.
def on_backward_begin(self, last_loss:Rank0Tensor, **kwargs:Any) -> Rank0Tensor: "Scale gradients up by `self.loss_scale` to prevent underflow." #To avoid gradient underflow, we scale the gradients ret_loss = last_loss * self.loss_scale return {'last_loss': ret_loss}
Convert the gradients back to FP32 and divide them by the scale.
def on_backward_end(self, **kwargs:Any)->None: "Convert the gradients back to FP32 and divide them by the scale." if self.dynamic and grad_overflow(self.model_params) and self.loss_scale > 1: self.loss_scale /= 2 self.noskip = 0 #The step will be skipped since we don'...
Update the params from master to model and zero grad.
def on_step_end(self, **kwargs:Any)->None: "Update the params from master to model and zero grad." #Zeros the gradients of the model since the optimizer is disconnected. self.learn.model.zero_grad() #Update the params from master to model. master2model(self.model_params, self.mas...
Scale the image so that the smallest axis is of size targ. Arguments: im (array): image targ (int): target size
def scale_min(im, targ, interpolation=cv2.INTER_AREA): """ Scale the image so that the smallest axis is of size targ. Arguments: im (array): image targ (int): target size """ r,c,*_ = im.shape ratio = targ/min(r,c) sz = (scale_to(c, ratio, targ), scale_to(r, ratio, targ)) re...
Zoom the center of image x by a factor of z+1 while retaining the original image size and proportion.
def zoom_cv(x,z): """ Zoom the center of image x by a factor of z+1 while retaining the original image size and proportion. """ if z==0: return x r,c,*_ = x.shape M = cv2.getRotationMatrix2D((c/2,r/2),0,z+1.) return cv2.warpAffine(x,M,(c,r))
Stretches image x horizontally by sr+1, and vertically by sc+1 while retaining the original image size and proportion.
def stretch_cv(x,sr,sc,interpolation=cv2.INTER_AREA): """ Stretches image x horizontally by sr+1, and vertically by sc+1 while retaining the original image size and proportion. """ if sr==0 and sc==0: return x r,c,*_ = x.shape x = cv2.resize(x, None, fx=sr+1, fy=sc+1, interpolation=interpolation) nr...
Perform any of 8 permutations of 90-degrees rotations or flips for image x.
def dihedral(x, dih): """ Perform any of 8 permutations of 90-degrees rotations or flips for image x. """ x = np.rot90(x, dih%4) return x if dih<4 else np.fliplr(x)
Adjust image balance and contrast
def lighting(im, b, c): """ Adjust image balance and contrast """ if b==0 and c==1: return im mu = np.average(im) return np.clip((im-mu)*c+mu+b,0.,1.).astype(np.float32)
Return a squared resized image
def no_crop(im, min_sz=None, interpolation=cv2.INTER_AREA): """ Return a squared resized image """ r,c,*_ = im.shape if min_sz is None: min_sz = min(r,c) return cv2.resize(im, (min_sz, min_sz), interpolation=interpolation)
Return a center crop of an image
def center_crop(im, min_sz=None): """ Return a center crop of an image """ r,c,*_ = im.shape if min_sz is None: min_sz = min(r,c) start_r = math.ceil((r-min_sz)/2) start_c = math.ceil((c-min_sz)/2) return crop(im, start_r, start_c, min_sz)
Randomly crop an image with an aspect ratio and returns a squared resized image of size targ References: 1. https://arxiv.org/pdf/1409.4842.pdf 2. https://arxiv.org/pdf/1802.07888.pdf
def googlenet_resize(im, targ, min_area_frac, min_aspect_ratio, max_aspect_ratio, flip_hw_p, interpolation=cv2.INTER_AREA): """ Randomly crop an image with an aspect ratio and returns a squared resized image of size targ References: 1. https://arxiv.org/pdf/1409.4842.pdf 2. https://arxiv.org/pdf/18...
Cut out n_holes number of square holes of size length in image at random locations. Holes may overlap.
def cutout(im, n_holes, length): """ Cut out n_holes number of square holes of size length in image at random locations. Holes may overlap. """ r,c,*_ = im.shape mask = np.ones((r, c), np.int32) for n in range(n_holes): y = np.random.randint(0, r) x = np.random.randint(0, c) y1 ...
Calculate dimension of an image during scaling with aspect ratio
def scale_to(x, ratio, targ): '''Calculate dimension of an image during scaling with aspect ratio''' return max(math.floor(x*ratio), targ)
crop image into a square of size sz,
def crop(im, r, c, sz): ''' crop image into a square of size sz, ''' return im[r:r+sz, c:c+sz]
Convert mask YY to a bounding box, assumes 0 as background nonzero object
def to_bb(YY, y="deprecated"): """Convert mask YY to a bounding box, assumes 0 as background nonzero object""" cols,rows = np.nonzero(YY) if len(cols)==0: return np.zeros(4, dtype=np.float32) top_row = np.min(rows) left_col = np.min(cols) bottom_row = np.max(rows) right_col = np.max(cols) ...
Transforming coordinates to pixels. Arguments: y : np array vector in which (y[0], y[1]) and (y[2], y[3]) are the the corners of a bounding box. x : image an image Returns: Y : image of shape x.shape
def coords2px(y, x): """ Transforming coordinates to pixels. Arguments: y : np array vector in which (y[0], y[1]) and (y[2], y[3]) are the the corners of a bounding box. x : image an image Returns: Y : image of shape x.shape """ ...
Apply a collection of transformation functions :fns: to images
def compose(im, y, fns): """ Apply a collection of transformation functions :fns: to images """ for fn in fns: #pdb.set_trace() im, y =fn(im, y) return im if y is None else (im, y)
Given the statistics of the training image sets, returns separate training and validation transform functions
def tfms_from_stats(stats, sz, aug_tfms=None, max_zoom=None, pad=0, crop_type=CropType.RANDOM, tfm_y=None, sz_y=None, pad_mode=cv2.BORDER_REFLECT, norm_y=True, scale=None): """ Given the statistics of the training image sets, returns separate training and validation transform functions """ ...
Generate a standard set of transformations Arguments --------- normalizer : image normalizing function denorm : image denormalizing function sz : size, sz_y = sz if not specified. tfms : iterable collection of transformation functions max_zoom : floa...
def image_gen(normalizer, denorm, sz, tfms=None, max_zoom=None, pad=0, crop_type=None, tfm_y=None, sz_y=None, pad_mode=cv2.BORDER_REFLECT, scale=None): """ Generate a standard set of transformations Arguments --------- normalizer : image normalizing function denorm : ...
Returns separate transformers of images for training and validation. Transformers are constructed according to the image statistics given by the model. (See tfms_from_stats) Arguments: f_model: model, pretrained or not pretrained
def tfms_from_model(f_model, sz, aug_tfms=None, max_zoom=None, pad=0, crop_type=CropType.RANDOM, tfm_y=None, sz_y=None, pad_mode=cv2.BORDER_REFLECT, norm_y=True, scale=None): """ Returns separate transformers of images for training and validation. Transformers are constructed according to th...
Return list of files in `c` that are images. `check_ext` will filter to `image_extensions`.
def get_image_files(c:PathOrStr, check_ext:bool=True, recurse=False)->FilePathList: "Return list of files in `c` that are images. `check_ext` will filter to `image_extensions`." return get_files(c, extensions=(image_extensions if check_ext else None), recurse=recurse)
Open a COCO style json in `fname` and returns the lists of filenames (with maybe `prefix`) and labelled bboxes.
def get_annotations(fname, prefix=None): "Open a COCO style json in `fname` and returns the lists of filenames (with maybe `prefix`) and labelled bboxes." annot_dict = json.load(open(fname)) id2images, id2bboxes, id2cats = {}, collections.defaultdict(list), collections.defaultdict(list) classes = {} ...
Function that collect `samples` of labelled bboxes and adds padding with `pad_idx`.
def bb_pad_collate(samples:BatchSamples, pad_idx:int=0) -> Tuple[FloatTensor, Tuple[LongTensor, LongTensor]]: "Function that collect `samples` of labelled bboxes and adds padding with `pad_idx`." if isinstance(samples[0][1], int): return data_collate(samples) max_len = max([len(s[1].data[1]) for s in sample...
Normalize `x` with `mean` and `std`.
def normalize(x:TensorImage, mean:FloatTensor,std:FloatTensor)->TensorImage: "Normalize `x` with `mean` and `std`." return (x-mean[...,None,None]) / std[...,None,None]
Denormalize `x` with `mean` and `std`.
def denormalize(x:TensorImage, mean:FloatTensor,std:FloatTensor, do_x:bool=True)->TensorImage: "Denormalize `x` with `mean` and `std`." return x.cpu().float()*std[...,None,None] + mean[...,None,None] if do_x else x.cpu()
`b` = `x`,`y` - normalize `x` array of imgs and `do_y` optionally `y`.
def _normalize_batch(b:Tuple[Tensor,Tensor], mean:FloatTensor, std:FloatTensor, do_x:bool=True, do_y:bool=False)->Tuple[Tensor,Tensor]: "`b` = `x`,`y` - normalize `x` array of imgs and `do_y` optionally `y`." x,y = b mean,std = mean.to(x.device),std.to(x.device) if do_x: x = normalize(x,mean,std) if...
Create normalize/denormalize func using `mean` and `std`, can specify `do_y` and `device`.
def normalize_funcs(mean:FloatTensor, std:FloatTensor, do_x:bool=True, do_y:bool=False)->Tuple[Callable,Callable]: "Create normalize/denormalize func using `mean` and `std`, can specify `do_y` and `device`." mean,std = tensor(mean),tensor(std) return (partial(_normalize_batch, mean=mean, std=std, do_x=do_x,...
Make channel the first axis of `x` and flatten remaining axes
def channel_view(x:Tensor)->Tensor: "Make channel the first axis of `x` and flatten remaining axes" return x.transpose(0,1).contiguous().view(x.shape[1],-1)
Size to resize to, to hit `targ_sz` at same aspect ratio, in PIL coords (i.e w*h)
def resize_to(img, targ_sz:int, use_min:bool=False): "Size to resize to, to hit `targ_sz` at same aspect ratio, in PIL coords (i.e w*h)" w,h = img.size min_sz = (min if use_min else max)(w,h) ratio = targ_sz/min_sz return int(w*ratio),int(h*ratio)
Check if the image in `file` exists, maybe resize it and copy it in `dest`.
def verify_image(file:Path, idx:int, delete:bool, max_size:Union[int,Tuple[int,int]]=None, dest:Path=None, n_channels:int=3, interp=PIL.Image.BILINEAR, ext:str=None, img_format:str=None, resume:bool=False, **kwargs): "Check if the image in `file` exists, maybe resize it and copy it in `dest`." ...
Check if the images in `path` aren't broken, maybe resize them and copy it in `dest`.
def verify_images(path:PathOrStr, delete:bool=True, max_workers:int=4, max_size:Union[int]=None, recurse:bool=False, dest:PathOrStr='.', n_channels:int=3, interp=PIL.Image.BILINEAR, ext:str=None, img_format:str=None, resume:bool=None, **kwargs): "Check if the images in `path` are...
Call `train_tfm` and `valid_tfm` after opening image, before converting from `PIL.Image`
def _db_pre_transform(self, train_tfm:List[Callable], valid_tfm:List[Callable]): "Call `train_tfm` and `valid_tfm` after opening image, before converting from `PIL.Image`" self.train_ds.x.after_open = compose(train_tfm) self.valid_ds.x.after_open = compose(valid_tfm) return self
Call `train_tfm` and `valid_tfm` after opening image, before converting from `PIL.Image`
def _ll_pre_transform(self, train_tfm:List[Callable], valid_tfm:List[Callable]): "Call `train_tfm` and `valid_tfm` after opening image, before converting from `PIL.Image`" self.train.x.after_open = compose(train_tfm) self.valid.x.after_open = compose(valid_tfm) return self
Resize images to `size` using `RandomResizedCrop`, passing along `kwargs` to train transform
def _presize(self, size:int, val_xtra_size:int=32, scale:Tuple[float]=(0.08, 1.0), ratio:Tuple[float]=(0.75, 4./3.), interpolation:int=2): "Resize images to `size` using `RandomResizedCrop`, passing along `kwargs` to train transform" return self.pre_transform( tvt.RandomResizedCrop(size, sc...
Create an `ImageDataBunch` from `LabelLists` `lls` with potential `ds_tfms`.
def create_from_ll(cls, lls:LabelLists, bs:int=64, val_bs:int=None, ds_tfms:Optional[TfmList]=None, num_workers:int=defaults.cpus, dl_tfms:Optional[Collection[Callable]]=None, device:torch.device=None, test:Optional[PathOrStr]=None, collate_fn:Callable=data_collate, size:int=None, no_che...
Create from imagenet style dataset in `path` with `train`,`valid`,`test` subfolders (or provide `valid_pct`).
def from_folder(cls, path:PathOrStr, train:PathOrStr='train', valid:PathOrStr='valid', valid_pct=None, classes:Collection=None, **kwargs:Any)->'ImageDataBunch': "Create from imagenet style dataset in `path` with `train`,`valid`,`test` subfolders (or provide `valid_pct`)." path=Path(p...
Create from a `DataFrame` `df`.
def from_df(cls, path:PathOrStr, df:pd.DataFrame, folder:PathOrStr=None, label_delim:str=None, valid_pct:float=0.2, fn_col:IntsOrStrs=0, label_col:IntsOrStrs=1, suffix:str='', **kwargs:Any)->'ImageDataBunch': "Create from a `DataFrame` `df`." src = (ImageList.from_df(df, path=path, folde...
Create from a csv file in `path/csv_labels`.
def from_csv(cls, path:PathOrStr, folder:PathOrStr=None, label_delim:str=None, csv_labels:PathOrStr='labels.csv', valid_pct:float=0.2, fn_col:int=0, label_col:int=1, suffix:str='', delimiter:str=None, header:Optional[Union[int,str]]='infer', **kwargs:Any)->'ImageDataBunch': "Cr...
Create from list of `fnames` in `path`.
def from_lists(cls, path:PathOrStr, fnames:FilePathList, labels:Collection[str], valid_pct:float=0.2, item_cls:Callable=None, **kwargs): "Create from list of `fnames` in `path`." item_cls = ifnone(item_cls, ImageList) fname2label = {f:l for (f,l) in zip(fnames, labels)} ...
Create from list of `fnames` in `path` with `label_func`.
def from_name_func(cls, path:PathOrStr, fnames:FilePathList, label_func:Callable, valid_pct:float=0.2, **kwargs): "Create from list of `fnames` in `path` with `label_func`." src = ImageList(fnames, path=path).split_by_rand_pct(valid_pct) return cls.create_from_ll(src.label_from_func(label_func),...
Create from list of `fnames` in `path` with re expression `pat`.
def from_name_re(cls, path:PathOrStr, fnames:FilePathList, pat:str, valid_pct:float=0.2, **kwargs): "Create from list of `fnames` in `path` with re expression `pat`." pat = re.compile(pat) def _get_label(fn): if isinstance(fn, Path): fn = fn.as_posix() res = pat.search(st...
Create an empty `ImageDataBunch` in `path` with `classes`. Typically used for inference.
def single_from_classes(path:Union[Path, str], classes:Collection[str], ds_tfms:TfmList=None, **kwargs): "Create an empty `ImageDataBunch` in `path` with `classes`. Typically used for inference." warn("""This method is deprecated and will be removed in a future version, use `load_learner` after ...
Grab a batch of data and call reduction function `func` per channel
def batch_stats(self, funcs:Collection[Callable]=None, ds_type:DatasetType=DatasetType.Train)->Tensor: "Grab a batch of data and call reduction function `func` per channel" funcs = ifnone(funcs, [torch.mean,torch.std]) x = self.one_batch(ds_type=ds_type, denorm=False)[0].cpu() return [fu...
Add normalize transform using `stats` (defaults to `DataBunch.batch_stats`)
def normalize(self, stats:Collection[Tensor]=None, do_x:bool=True, do_y:bool=False)->None: "Add normalize transform using `stats` (defaults to `DataBunch.batch_stats`)" if getattr(self,'norm',False): raise Exception('Can not call normalize twice') if stats is None: self.stats = self.batch_stats(...
Open image in `fn`, subclass and overwrite for custom behavior.
def open(self, fn): "Open image in `fn`, subclass and overwrite for custom behavior." return open_image(fn, convert_mode=self.convert_mode, after_open=self.after_open)
Get the list of files in `path` that have an image suffix. `recurse` determines if we search subfolders.
def from_folder(cls, path:PathOrStr='.', extensions:Collection[str]=None, **kwargs)->ItemList: "Get the list of files in `path` that have an image suffix. `recurse` determines if we search subfolders." extensions = ifnone(extensions, image_extensions) return super().from_folder(path=path, extens...
Get the filenames in `cols` of `df` with `folder` in front of them, `suffix` at the end.
def from_df(cls, df:DataFrame, path:PathOrStr, cols:IntsOrStrs=0, folder:PathOrStr=None, suffix:str='', **kwargs)->'ItemList': "Get the filenames in `cols` of `df` with `folder` in front of them, `suffix` at the end." suffix = suffix or '' res = super().from_df(df, path=path, cols=cols, **kwargs...
Get the filenames in `path/csv_name` opened with `header`.
def from_csv(cls, path:PathOrStr, csv_name:str, header:str='infer', **kwargs)->'ItemList': "Get the filenames in `path/csv_name` opened with `header`." path = Path(path) df = pd.read_csv(path/csv_name, header=header) return cls.from_df(df, path=path, **kwargs)
Show the `xs` (inputs) and `ys` (targets) on a figure of `figsize`.
def show_xys(self, xs, ys, imgsize:int=4, figsize:Optional[Tuple[int,int]]=None, **kwargs): "Show the `xs` (inputs) and `ys` (targets) on a figure of `figsize`." rows = int(np.ceil(math.sqrt(len(xs)))) axs = subplots(rows, rows, imgsize=imgsize, figsize=figsize) for x,y,ax in zip(xs, ys,...
Show `xs` (inputs), `ys` (targets) and `zs` (predictions) on a figure of `figsize`.
def show_xyzs(self, xs, ys, zs, imgsize:int=4, figsize:Optional[Tuple[int,int]]=None, **kwargs): "Show `xs` (inputs), `ys` (targets) and `zs` (predictions) on a figure of `figsize`." if self._square_show_res: title = 'Ground truth\nPredictions' rows = int(np.ceil(math.sqrt(len(xs...
Generate classes from unique `items` and add `background`.
def generate_classes(self, items): "Generate classes from unique `items` and add `background`." classes = super().generate_classes([o[1] for o in items]) classes = ['background'] + list(classes) return classes
Show the `xs` (inputs) and `ys`(targets) on a figure of `figsize`.
def show_xys(self, xs, ys, imgsize:int=4, figsize:Optional[Tuple[int,int]]=None, **kwargs): "Show the `xs` (inputs) and `ys`(targets) on a figure of `figsize`." axs = subplots(len(xs), 2, imgsize=imgsize, figsize=figsize) for i, (x,y) in enumerate(zip(xs,ys)): x.show(ax=axs[i,0], **...
Show `xs` (inputs), `ys` (targets) and `zs` (predictions) on a figure of `figsize`.
def show_xyzs(self, xs, ys, zs, imgsize:int=4, figsize:Optional[Tuple[int,int]]=None, **kwargs): "Show `xs` (inputs), `ys` (targets) and `zs` (predictions) on a figure of `figsize`." title = 'Input / Prediction / Target' axs = subplots(len(xs), 3, imgsize=imgsize, figsize=figsize, title=title, w...
get total, used and free memory (in MBs) for gpu `id`. if `id` is not passed, currently selected torch device is used
def gpu_mem_get(id=None): "get total, used and free memory (in MBs) for gpu `id`. if `id` is not passed, currently selected torch device is used" if not use_gpu: return GPUMemory(0, 0, 0) if id is None: id = torch.cuda.current_device() try: handle = pynvml.nvmlDeviceGetHandleByIndex(id) ...
get [gpu_id, its_free_ram] for the first gpu with highest available RAM
def gpu_with_max_free_mem(): "get [gpu_id, its_free_ram] for the first gpu with highest available RAM" mem_all = gpu_mem_get_all() if not len(mem_all): return None, 0 free_all = np.array([x.free for x in mem_all]) id = np.argmax(free_all) return id, free_all[id]
A decorator that runs `GPUMemTrace` w/ report on func
def gpu_mem_trace(func): "A decorator that runs `GPUMemTrace` w/ report on func" @functools.wraps(func) def wrapper(*args, **kwargs): with GPUMemTrace(ctx=func.__qualname__, on_exit_report=True): return func(*args, **kwargs) return wrapper
iterate through all the columns of a dataframe and modify the data type to reduce memory usage.
def reduce_mem_usage(df): """ iterate through all the columns of a dataframe and modify the data type to reduce memory usage. """ start_mem = df.memory_usage().sum() / 1024**2 print('Memory usage of dataframe is {:.2f} MB'.format(start_mem)) #Removed from debugging columns = df.columns ...
Return ' (ctx: subctx)' or ' (ctx)' or ' (subctx)' or '' depending on this and constructor arguments
def _get_ctx(self, subctx=None): "Return ' (ctx: subctx)' or ' (ctx)' or ' (subctx)' or '' depending on this and constructor arguments" l = [] if self.ctx is not None: l.append(self.ctx) if subctx is not None: l.append(subctx) return '' if len(l) == 0 else f" ({': '.j...
Put `learn` on distributed training with `cuda_id`.
def _learner_distributed(learn:Learner, cuda_id:int, cache_dir:PathOrStr='tmp'): "Put `learn` on distributed training with `cuda_id`." learn.callbacks.append(DistributedTrainer(learn, cuda_id)) learn.callbacks.append(DistributedRecorder(learn, cuda_id, cache_dir)) return learn
Constructs a XResNet-18 model. Args: pretrained (bool): If True, returns a model pre-trained on ImageNet
def xresnet18(pretrained=False, **kwargs): """Constructs a XResNet-18 model. Args: pretrained (bool): If True, returns a model pre-trained on ImageNet """ model = XResNet(BasicBlock, [2, 2, 2, 2], **kwargs) if pretrained: model.load_state_dict(model_zoo.load_url(model_urls['xresnet18'])) ...
Constructs a XResNet-50 model. Args: pretrained (bool): If True, returns a model pre-trained on ImageNet
def xresnet50_2(pretrained=False, **kwargs): """Constructs a XResNet-50 model. Args: pretrained (bool): If True, returns a model pre-trained on ImageNet """ model = XResNet(Bottleneck, [3, 4, 6, 3], **kwargs) if pretrained: model.load_state_dict(model_zoo.load_url(model_urls['xresnet50'])) ...
Calculate loss and metrics for a batch, call out to callbacks as necessary.
def loss_batch(model:nn.Module, xb:Tensor, yb:Tensor, loss_func:OptLossFunc=None, opt:OptOptimizer=None, cb_handler:Optional[CallbackHandler]=None)->Tuple[Union[Tensor,int,float,str]]: "Calculate loss and metrics for a batch, call out to callbacks as necessary." cb_handler = ifnone(cb_handler, Ca...
Tuple of predictions and targets, and optional losses (if `loss_func`) using `dl`, max batches `n_batch`.
def get_preds(model:nn.Module, dl:DataLoader, pbar:Optional[PBar]=None, cb_handler:Optional[CallbackHandler]=None, activ:nn.Module=None, loss_func:OptLossFunc=None, n_batch:Optional[int]=None) -> List[Tensor]: "Tuple of predictions and targets, and optional losses (if `loss_func`) using `dl`, max batc...
Calculate `loss_func` of `model` on `dl` in evaluation mode.
def validate(model:nn.Module, dl:DataLoader, loss_func:OptLossFunc=None, cb_handler:Optional[CallbackHandler]=None, pbar:Optional[PBar]=None, average=True, n_batch:Optional[int]=None)->Iterator[Tuple[Union[Tensor,int],...]]: "Calculate `loss_func` of `model` on `dl` in evaluation mode." model.eval(...
Simple training of `model` for 1 epoch of `dl` using optim `opt` and loss function `loss_func`.
def train_epoch(model:nn.Module, dl:DataLoader, opt:optim.Optimizer, loss_func:LossFunction)->None: "Simple training of `model` for 1 epoch of `dl` using optim `opt` and loss function `loss_func`." model.train() for xb,yb in dl: loss = loss_func(model(xb), yb) loss.backward() opt.ste...
Fit the `model` on `data` and learn using `loss_func` and `opt`.
def fit(epochs:int, learn:BasicLearner, callbacks:Optional[CallbackList]=None, metrics:OptMetrics=None)->None: "Fit the `model` on `data` and learn using `loss_func` and `opt`." assert len(learn.data.train_dl) != 0, f"""Your training dataloader is empty, can't train a model. Use a smaller batch size (ba...
Load a `Learner` object saved with `export_state` in `path/file` with empty data, optionally add `test` and load on `cpu`. `file` can be file-like (file or buffer)
def load_learner(path:PathOrStr, file:PathLikeOrBinaryStream='export.pkl', test:ItemList=None, **db_kwargs): "Load a `Learner` object saved with `export_state` in `path/file` with empty data, optionally add `test` and load on `cpu`. `file` can be file-like (file or buffer)" source = Path(path)/file if is_pathli...
Initialize recording status at beginning of training.
def on_train_begin(self, pbar:PBar, metrics_names:Collection[str], **kwargs:Any)->None: "Initialize recording status at beginning of training." self.pbar = pbar self.names = ['epoch', 'train_loss'] if self.no_val else ['epoch', 'train_loss', 'valid_loss'] self.metrics_names = metrics_nam...
Record learning rate and momentum at beginning of batch.
def on_batch_begin(self, train, **kwargs:Any)->None: "Record learning rate and momentum at beginning of batch." if train: self.lrs.append(self.opt.lr) self.moms.append(self.opt.mom)
Record the loss before any other callback has a chance to modify it.
def on_backward_begin(self, smooth_loss:Tensor, **kwargs:Any)->None: "Record the loss before any other callback has a chance to modify it." self.losses.append(smooth_loss) if self.pbar is not None and hasattr(self.pbar,'child'): self.pbar.child.comment = f'{smooth_loss:.4f}'
Save epoch info: num_batch, smooth_loss, metrics.
def on_epoch_end(self, epoch:int, num_batch:int, smooth_loss:Tensor, last_metrics=MetricsList, **kwargs:Any)->bool: "Save epoch info: num_batch, smooth_loss, metrics." self.nb_batches.append(num_batch) if last_metrics is not None: self.val_losses.append(last_metrics[0]) ...
Format stats before printing.
def format_stats(self, stats:TensorOrNumList)->None: "Format stats before printing." str_stats = [] for name,stat in zip(self.names,stats): str_stats.append('#na#' if stat is None else str(stat) if isinstance(stat, int) else f'{stat:.6f}') if self.add_time: str_stats.append(f...
Add `names` to the inner metric names.
def add_metric_names(self, names): "Add `names` to the inner metric names." if hasattr(self, '_added_met_names'): self._added_met_names += names else: self._added_met_names = names
Plot learning rate, `show_moms` to include momentum.
def plot_lr(self, show_moms=False, skip_start:int=0, skip_end:int=0, return_fig:bool=None)->Optional[plt.Figure]: "Plot learning rate, `show_moms` to include momentum." lrs = self._split_list(self.lrs, skip_start, skip_end) iterations = self._split_list(range_of(self.lrs), skip_start, skip_end) ...
Plot learning rate and losses, trimmed between `skip_start` and `skip_end`. Optionally plot and return min gradient
def plot(self, skip_start:int=10, skip_end:int=5, suggestion:bool=False, return_fig:bool=None, **kwargs)->Optional[plt.Figure]: "Plot learning rate and losses, trimmed between `skip_start` and `skip_end`. Optionally plot and return min gradient" lrs = self._split_list(self.lrs, skip_start, ...
Plot training and validation losses.
def plot_losses(self, skip_start:int=0, skip_end:int=0, return_fig:bool=None)->Optional[plt.Figure]: "Plot training and validation losses." fig, ax = plt.subplots(1,1) losses = self._split_list(self.losses, skip_start, skip_end) iterations = self._split_list(range_of(self.losses), skip_s...
Plot metrics collected during training.
def plot_metrics(self, skip_start:int=0, skip_end:int=0, return_fig:bool=None)->Optional[plt.Figure]: "Plot metrics collected during training." assert len(self.metrics) != 0, "There are no metrics to plot." fig, axes = plt.subplots(len(self.metrics[0]),1,figsize=(6, 4*len(self.metrics[0]))) ...
Look at params (annotated with `Param`) in func and return an `ArgumentParser`
def anno_parser(func): "Look at params (annotated with `Param`) in func and return an `ArgumentParser`" p = ArgumentParser(description=func.__doc__) for k,v in inspect.signature(func).parameters.items(): param = func.__annotations__.get(k, Param()) kwargs = param.kwargs if v.default ...
Decorator to create a simple CLI from `func` using `anno_parser`
def call_parse(func): "Decorator to create a simple CLI from `func` using `anno_parser`" name = inspect.currentframe().f_back.f_globals['__name__'] if name == "__main__": args = anno_parser(func).parse_args() func(**args.__dict__) else: return func
Decorator to create a simple CLI from `func` using `plac`
def call_plac(f): "Decorator to create a simple CLI from `func` using `plac`" name = inspect.currentframe().f_back.f_globals['__name__'] if name == '__main__': import plac res = plac.call(f) if callable(res): res() else: return f
Takes in text tokens and returns int2tok and tok2int converters Arguments: tokens(list): List of tokens. Can be a list of strings, or a list of lists of strings. max_vocab(int): Number of tokens to return in the vocab (sorted by frequency) min_freq(int): Minimum number of instances a to...
def numericalize_tok(tokens, max_vocab=50000, min_freq=0, unk_tok="_unk_", pad_tok="_pad_", bos_tok="_bos_", eos_tok="_eos_"): """Takes in text tokens and returns int2tok and tok2int converters Arguments: tokens(list): List of tokens. Can be a list of strings, or a list of lists of strings. ...
If your convolutional window is greater than 1 and you save previous xs, you must reset at the beginning of each new sequence.
def reset(self): "If your convolutional window is greater than 1 and you save previous xs, you must reset at the beginning of each new sequence." for layer in self.layers: layer.reset() if self.bidirectional: for layer in self.layers_bwd: layer.reset()
Start a new kernel, and return its Manager and Client
def start_new_kernel(startup_timeout=60, kernel_name='python', **kwargs): """Start a new kernel, and return its Manager and Client""" logger.debug('Starting new kernel: "%s"' % kernel_name) km = KernelManager(kernel_name=kernel_name, kernel_spec_manager=NbvalKernelspecManager()) k...
Returns a :class:`KernelSpec` instance for the given kernel_name. Raises :exc:`NoSuchKernel` if the given kernel name is not found.
def get_kernel_spec(self, kernel_name): """Returns a :class:`KernelSpec` instance for the given kernel_name. Raises :exc:`NoSuchKernel` if the given kernel name is not found. """ if kernel_name == CURRENT_ENV_KERNEL_NAME: return self.kernel_spec_class( resour...