python_code
stringlengths
0
229k
""" Support TorchDynamo(https://github.com/facebookresearch/torchdynamo) backends """ import argparse import contextlib import distutils.util from typing import List import torch import torch._dynamo as torchdynamo from torchbenchmark.util.model import is_staged_train_test def parse_torchdynamo_args(dynamo_args: List[...
""" Utils for managing backends """ import functools BACKENDS = dict() def create_backend(fn): @functools.wraps(fn) def inner(model: 'torchbenchmark.util.model.BenchmarkModel', **kwargs): if model is None: return None try: return fn(model, **kwargs) except Keyb...
import torch from torchbenchmark.util.backends import create_backend from typing import List WARMUP_ITER = 3 @create_backend def cudagraph(model: 'torchbenchmark.util.model.BenchmarkModel', backend_args: List[str]): cudagraph_func_name = f"cudagraph_{model.test}" assert hasattr(model, cudagraph_func_name), f"...
from typing import List import torch import argparse from torchbenchmark.util.backends import create_backend from torchbenchmark.util.env_check import is_hf_model def parse_torch_trt_args(backend_args: List[str]): """Parses CLI-provided backend arguments to extract Torch-TRT keywords Returns kwargs dictiona...
""" Utilities to measure metrics of a model. """ import torch import time import dataclasses from torchbenchmark.util.model import BenchmarkModel from torchbenchmark.util.experiment.instantiator import TorchBenchModelConfig from torchbenchmark import ModelTask from typing import List, Union, Tuple, Optional WARMUP_ROU...
""" Utilities to instantiate TorchBench models in the same process or child process. Functions in this file don't handle exceptions. They expect callers handle all exceptions. """ import os import importlib import dataclasses from typing import Optional, List, Dict from torchbenchmark.util.model import BenchmarkModel f...
from torchbenchmark.tasks import NLP from torchbenchmark.util.framework.huggingface.model_factory import HuggingFaceModel class Model(HuggingFaceModel): task = NLP.LANGUAGE_MODELING DEFAULT_TRAIN_BSIZE = 2 DEFAULT_EVAL_BSIZE = 1 def __init__(self, test, device, batch_size=None, extra_args=[]): ...
import subprocess import sys import os from torchbenchmark.util.framework.huggingface.patch_hf import patch_transformers, cache_model def pip_install_requirements(): subprocess.check_call([sys.executable, '-m', 'pip', 'install', '-q', '-r', 'requirements.txt']) if __name__ == '__main__': pip_install_requirem...
import dataclasses @dataclasses.dataclass class DRQConfig: env = "cartpole_swingup" # IMPORTANT: if action_repeat is used the effective number of env steps needs to be # multiplied by action_repeat in the result graphs. # This is a common practice for a fair comparison. # See the 2nd paragraph in A...
import numpy as np import torch import torch.nn as nn import torch.nn.functional as F import copy import math from . import drqutils class Encoder(nn.Module): """Convolutional encoder for image-based observations.""" def __init__(self, obs_shape, feature_dim): super().__init__() assert len(ob...
import math import os import random from collections import deque import numpy as np import scipy.linalg as sp_la import gym import torch import torch.nn as nn import torch.nn.functional as F from skimage.util.shape import view_as_windows from torch import distributions as pyd class eval_mode: def __init__(self...
import copy import math import pickle as pkl import numpy as np import torch import os import sys import torch.nn as nn from typing import Tuple import torch.nn.functional as F from gym import spaces from ...util.model import BenchmarkModel from torchbenchmark.tasks import REINFORCEMENT_LEARNING from .drqutils import...
import numpy as np import kornia import torch import torch.nn as nn import torch.nn.functional as F class ReplayBuffer: """Buffer to store environment transitions.""" def __init__(self, obs_shape, action_shape, capacity, image_pad, device): self.capacity = capacity self.device = device ...
import subprocess import sys from utils import s3_utils def pip_install_requirements(): subprocess.check_call([sys.executable, '-m', 'pip', 'install', '-q', '-r', 'requirements.txt']) if __name__ == '__main__': pip_install_requirements() s3_utils.checkout_s3_data("MODEL_PKLS", "drq/obs.pkl", decompress=Fa...
from torchbenchmark.util.framework.timm.model_factory import TimmModel from torchbenchmark.tasks import COMPUTER_VISION class Model(TimmModel): task = COMPUTER_VISION.CLASSIFICATION DEFAULT_TRAIN_BSIZE = 32 DEFAULT_EVAL_BSIZE = 32 def __init__(self, test, device, batch_size=None, extra_args=[]): ...
from torchbenchmark.util.framework.vision.model_factory import TorchVisionModel from torchbenchmark.tasks import COMPUTER_VISION import torchvision.models as models class Model(TorchVisionModel): task = COMPUTER_VISION.CLASSIFICATION DEFAULT_TRAIN_BSIZE = 32 DEFAULT_EVAL_BSIZE = 32 def __init__(self, ...
# Generated by gen_torchvision_benchmark.py import torch import torch.optim as optim import torchvision.models as models from torch.quantization import quantize_fx from ...util.model import BenchmarkModel from torchbenchmark.tasks import COMPUTER_VISION from typing import Tuple class Model(BenchmarkModel): task ...
import os from torchbenchmark.tasks import COMPUTER_VISION from torchbenchmark.util.framework.detectron2.model_factory import Detectron2Model MODEL_NAME = os.path.basename(os.path.dirname(os.path.abspath(__file__))) MODEL_DIR = os.path.abspath(os.path.dirname(__file__)) class Model(Detectron2Model): task = COMPUT...
import os from torchbenchmark.util.framework.detectron2 import install_detectron2 MODEL_NAME = os.path.basename(os.path.dirname(os.path.abspath(__file__))) MODEL_DIR = os.path.abspath(os.path.dirname(__file__)) if __name__ == '__main__': install_detectron2(MODEL_NAME, MODEL_DIR)
""" https://github.com/phlippe/uvadlc_notebooks_benchmarking/blob/main/PyTorch/Tutorial5_Inception_ResNet_DenseNet.py """ from types import SimpleNamespace from torchbenchmark.tasks import COMPUTER_VISION from torchbenchmark.util.model import BenchmarkModel import torch.nn as nn import torch import torch.optim as optim...
import torch import os import itertools import random import itertools from pathlib import Path from typing import Tuple from detectron2.checkpoint import DetectionCheckpointer # TorchBench imports from torchbenchmark.util.model import BenchmarkModel from torchbenchmark.tasks import COMPUTER_VISION MODEL_NAME = os.p...
import os from torchbenchmark.util.framework.detectron2 import install_detectron2 MODEL_NAME = os.path.basename(os.path.dirname(os.path.abspath(__file__))) MODEL_DIR = os.path.abspath(os.path.dirname(__file__)) if __name__ == '__main__': install_detectron2(MODEL_NAME, MODEL_DIR)
"""General-purpose training script for image-to-image translation. This script works for various models (with option '--model': e.g., pix2pix, cyclegan, colorization) and different datasets (with option '--dataset_mode': e.g., aligned, unaligned, single, colorization). You need to specify the dataset ('--dataroot'), e...
#!/usr/bin/env python import torch import os from pathlib import Path torch.backends.cudnn.deterministic = True torch.backends.cudnn.benchmark = False from ...util.model import BenchmarkModel from torchbenchmark.tasks import COMPUTER_VISION from typing import Tuple from torchbenchmark import DATA_PATH from .train_cyc...
"""General-purpose test script for image-to-image translation. Once you have trained your model with train.py, you can use this script to test the model. It will load a saved model from '--checkpoints_dir' and save the results to '--results_dir'. It first creates model and dataset given the option. It will hard-code ...
import subprocess import sys from utils import s3_utils def pip_install_requirements(): subprocess.check_call([sys.executable, '-m', 'pip', 'install', '-q', '-r', 'requirements.txt']) if __name__ == '__main__': s3_utils.checkout_s3_data("INPUT_TARBALLS", "pytorch_CycleGAN_and_pix2pix_inputs.tar.gz", decompre...
from .base_options import BaseOptions class TestOptions(BaseOptions): """This class includes test options. It also includes shared options defined in BaseOptions. """ def initialize(self, parser): parser = BaseOptions.initialize(self, parser) # define shared options parser.add_argum...
from .base_options import BaseOptions class TrainOptions(BaseOptions): """This class includes training options. It also includes shared options defined in BaseOptions. """ def initialize(self, parser): parser = BaseOptions.initialize(self, parser) # add torchbench options par...
"""This package options includes option modules: training options, test options, and basic options (used in both training and test)."""
import argparse import os from ..util import util import torch from ..models import get_option_setter from ..data import get_option_setter as get_option_setter_data class BaseOptions(): """This class defines options used during both training and test time. It also implements several helper functions such as ...
import random import torch class ImagePool(): """This class implements an image buffer that stores previously generated images. This buffer enables us to update discriminators using a history of generated images rather than the ones produced by the latest generators. """ def __init__(self, pool_...
"""This module contains simple helper functions """ from __future__ import print_function import torch import numpy as np from PIL import Image import os def tensor2im(input_image, imtype=np.uint8): """"Converts a Tensor array into a numpy image array. Parameters: input_image (tensor) -- the input i...
import dominate from dominate.tags import meta, h3, table, tr, td, p, a, img, br import os class HTML: """This HTML class allows us to save images and write texts into a single HTML file. It consists of functions such as <add_header> (add a text header to the HTML file), <add_images> (add a row of imag...
"""This package includes a miscellaneous collection of useful helper functions."""
from __future__ import print_function import os import tarfile import requests from warnings import warn from zipfile import ZipFile from bs4 import BeautifulSoup from os.path import abspath, isdir, join, basename class GetData: """A Python script for downloading CycleGAN or pix2pix datasets. Parameters: ...
import numpy as np import os import sys import ntpath import time from . import util, html from subprocess import Popen, PIPE if sys.version_info[0] == 2: VisdomExceptionBase = Exception else: VisdomExceptionBase = ConnectionError def save_images(webpage, visuals, image_path, aspect_ratio=1.0, width=256): ...
from .base_model import BaseModel from . import networks class TestModel(BaseModel): """ This TesteModel can be used to generate CycleGAN results for only one direction. This model will automatically set '--dataset_mode single', which only loads the images from one collection. See the test instruction fo...
"""Model class template This module provides a template for users to implement custom models. You can specify '--model template' to use this model. The class name should be consistent with both the filename and its model option. The filename should be <model>_dataset.py The class name should be <Model>Dataset.py It im...
import torch import itertools from ..util.image_pool import ImagePool from .base_model import BaseModel from . import networks class CycleGANModel(BaseModel): """ This class implements the CycleGAN model, for learning image-to-image translation without paired data. The model training requires '--dataset_...
import torch from .base_model import BaseModel from . import networks class Pix2PixModel(BaseModel): """ This class implements the pix2pix model, for learning a mapping from input images to output images given paired data. The model training requires '--dataset_mode aligned' dataset. By default, it uses ...
"""This package contains modules related to objective functions, optimizations, and network architectures. To add a custom model class called 'dummy', you need to add a file called 'dummy_model.py' and define a subclass DummyModel inherited from BaseModel. You need to implement the following five functions: -- <__...
from .pix2pix_model import Pix2PixModel import torch from skimage import color # used for lab2rgb import numpy as np class ColorizationModel(Pix2PixModel): """This is a subclass of Pix2PixModel for image colorization (black & white image -> colorful images). The model training requires '-dataset_model color...
import os import torch from collections import OrderedDict from abc import ABC, abstractmethod from . import networks class BaseModel(ABC): """This class is an abstract base class (ABC) for models. To create a subclass, you need to implement the following five functions: -- <__init__>: ...
import torch import torch.nn as nn from torch.nn import init import functools from torch.optim import lr_scheduler ############################################################################### # Helper Functions ############################################################################### class Identity(nn.Modu...
# Simple script to make sure basic usage # such as training, testing, saving and loading # runs without errors. import os def run(command): print(command) exit_status = os.system(command) if exit_status > 0: exit(1) if __name__ == '__main__': # download mini datasets if not os.path.exist...
# The following code is modified from https://github.com/shelhamer/clockwork-fcn import sys import os import glob import numpy as np from PIL import Image class cityscapes: def __init__(self, data_path): # data_path something like /data2/cityscapes self.dir = data_path self.classes = ['roa...
# The following code is modified from https://github.com/shelhamer/clockwork-fcn import numpy as np def get_out_scoremap(net): return net.blobs['score'].data[0].argmax(axis=0).astype(np.uint8) def feed_net(net, in_): """ Load prepared input into net. """ net.blobs['data'].reshape(1, *in_.shape) ...
import os import caffe import argparse import numpy as np import scipy.misc from PIL import Image from util import segrun, fast_hist, get_scores from cityscapes import cityscapes parser = argparse.ArgumentParser() parser.add_argument("--cityscapes_dir", type=str, required=True, help="Path to the original cityscapes da...
# HED batch processing script; modified from https://github.com/s9xie/hed/blob/master/examples/hed/HED-tutorial.ipynb # Step 1: download the hed repo: https://github.com/s9xie/hed # Step 2: download the models and protoxt, and put them under {caffe_root}/examples/hed/ # Step 3: put this script under {caffe_root}/exampl...
import os.path from data.base_dataset import BaseDataset, get_transform from data.image_folder import make_dataset from skimage import color # require skimage from PIL import Image import numpy as np import torchvision.transforms as transforms class ColorizationDataset(BaseDataset): """This dataset class can loa...
"""This module implements an abstract base class (ABC) 'BaseDataset' for datasets. It also includes common transformation functions (e.g., get_transform, __scale_width), which can be later used in subclasses. """ import random import numpy as np import torch.utils.data as data from PIL import Image import torchvision....
"""Dataset class template This module provides a template for users to implement custom datasets. You can specify '--dataset_mode template' to use this dataset. The class name should be consistent with both the filename and its dataset_mode option. The filename should be <dataset_mode>_dataset.py The class name should...
"""This package includes all the modules related to data loading and preprocessing To add a custom dataset class called 'dummy', you need to add a file called 'dummy_dataset.py' and define a subclass 'DummyDataset' inherited from BaseDataset. You need to implement four functions: -- <__init__>: ...
"""A modified image folder class We modify the official PyTorch image folder (https://github.com/pytorch/vision/blob/master/torchvision/datasets/folder.py) so that this class can load images from both current directory and its subdirectories. """ import torch.utils.data as data from PIL import Image import os import...
from .base_dataset import BaseDataset, get_transform from .image_folder import make_dataset from PIL import Image class SingleDataset(BaseDataset): """This dataset class can load a set of images specified by the path --dataroot /path/to/data. It can be used for generating CycleGAN results only for one side w...
import os.path from data.base_dataset import BaseDataset, get_params, get_transform from data.image_folder import make_dataset from PIL import Image class AlignedDataset(BaseDataset): """A dataset class for paired image dataset. It assumes that the directory '/path/to/data/train' contains image pairs in the ...
import os.path from .base_dataset import BaseDataset, get_transform from .image_folder import make_dataset from PIL import Image import random class UnalignedDataset(BaseDataset): """ This dataset class can load unaligned/unpaired datasets. It requires two directories to host training images from domain ...
from torchbenchmark.util.framework.timm.model_factory import TimmModel from torchbenchmark.tasks import COMPUTER_VISION class Model(TimmModel): task = COMPUTER_VISION.GENERATION DEFAULT_TRAIN_BSIZE = 32 DEFAULT_EVAL_BSIZE = 32 def __init__(self, test, device, batch_size=None, extra_args=[]): ...
from torchbenchmark.tasks import NLP from torchbenchmark.util.framework.huggingface.model_factory import HuggingFaceModel class Model(HuggingFaceModel): task = NLP.LANGUAGE_MODELING # Original train batch size per device: 8 # Source: https://github.com/huggingface/transformers/blob/master/examples/flax/lan...
import subprocess import sys import os from torchbenchmark.util.framework.huggingface.patch_hf import patch_transformers, cache_model def pip_install_requirements(): subprocess.check_call([sys.executable, '-m', 'pip', 'install', '-q', '-r', 'requirements.txt']) if __name__ == '__main__': pip_install_requirem...
from torchbenchmark.util.framework.timm.model_factory import TimmModel from torchbenchmark.tasks import COMPUTER_VISION class Model(TimmModel): task = COMPUTER_VISION.CLASSIFICATION # Original train batch size 128, hardware Nvidia rtx 3090 # Source: https://gist.github.com/rwightman/bb59f9e245162cee0e38bd...
import torch import torch.optim as optim import torch.nn as nn import torch.nn.functional as F from functorch import make_functional_with_buffers, vmap, grad import functools from pathlib import Path from typing import Tuple from ...util.model import BenchmarkModel from torchbenchmark.tasks import OTHER def loss_for...
import subprocess import sys def pip_install_requirements(): subprocess.check_call([sys.executable, '-m', 'pip', 'install', '-q', '-r', 'requirements.txt']) if __name__ == '__main__': pip_install_requirements()
import torch from dalle2_pytorch import DALLE2, Unet, Decoder, DiffusionPriorNetwork, DiffusionPrior, OpenAIClipAdapter torch.backends.cudnn.deterministic = False torch.backends.cudnn.benchmark = False from ...util.model import BenchmarkModel from torchbenchmark.tasks import COMPUTER_VISION class Model(BenchmarkMod...
import os import patch import subprocess import sys def patch_dalle2(): import dalle2_pytorch current_dir = os.path.dirname(os.path.abspath(__file__)) dalle2_dir = os.path.dirname(dalle2_pytorch.__file__) dalle2_patch = patch.fromfile(os.path.join(current_dir, "dalle2_pytorch.patch")) if not dalle2...
from torchbenchmark.util.framework.vision.model_factory import TorchVisionModel from torchbenchmark.tasks import COMPUTER_VISION import torchvision.models as models class Model(TorchVisionModel): task = COMPUTER_VISION.CLASSIFICATION DEFAULT_TRAIN_BSIZE = 32 DEFAULT_EVAL_BSIZE = 32 def __init__(self, ...
from torchbenchmark.tasks import NLP from torchbenchmark.util.framework.huggingface.model_factory import HuggingFaceModel class Model(HuggingFaceModel): task = NLP.LANGUAGE_MODELING DEFAULT_TRAIN_BSIZE = 8 DEFAULT_EVAL_BSIZE = 1 def __init__(self, test, device, batch_size=None, extra_args=[]): ...
import subprocess import sys import os from torchbenchmark.util.framework.huggingface.patch_hf import patch_transformers, cache_model def pip_install_requirements(): subprocess.check_call([sys.executable, '-m', 'pip', 'install', '-q', '-r', 'requirements.txt']) if __name__ == '__main__': pip_install_requirem...
from torchbenchmark.util.framework.timm.model_factory import TimmModel from torchbenchmark.tasks import COMPUTER_VISION class Model(TimmModel): task = COMPUTER_VISION.CLASSIFICATION DEFAULT_TRAIN_BSIZE = 32 DEFAULT_EVAL_BSIZE = 64 def __init__(self, test, device, batch_size=None, extra_args=[]): ...
""" Maskrcnn model from torchvision """ import torch import os import itertools import random import numpy as np from ...util.model import BenchmarkModel from torchbenchmark.tasks import COMPUTER_VISION from pathlib import Path from typing import Tuple # Model specific imports import torchvision from .coco_utils impo...
import sys import subprocess from utils import s3_utils def pip_install_requirements(): subprocess.check_call([sys.executable, '-m', 'pip', 'install', '-q', '-r', 'requirements.txt']) if __name__ == '__main__': s3_utils.checkout_s3_data("INPUT_TARBALLS", "coco2017-minimal.tar.gz", decompress=True) pip_ins...
import torch from pycocotools import mask as coco_mask from torchvision.transforms import functional as F def convert_coco_poly_to_mask(segmentations, height, width): masks = [] for polygons in segmentations: rles = coco_mask.frPyObjects(polygons, height, width) mask = coco_mask.decode(rles) ...
import os import logging import torch from pathlib import Path from contextlib import suppress # TorchBench imports from torchbenchmark.util.model import BenchmarkModel from torchbenchmark.tasks import COMPUTER_VISION # effdet imports from effdet import create_model, create_loader from effdet.data import resolve_inpu...
from effdet.data import resolve_input_config, SkipSubset from effdet import create_loader, create_dataset, create_evaluator from effdet.anchors import Anchors, AnchorLabeler from effdet.data.dataset_config import CocoCfg from dataclasses import dataclass, field from typing import Dict @dataclass class Coco2017Minima...
import torch from collections import OrderedDict from contextlib import suppress from timm.utils import AverageMeter, reduce_tensor def train_epoch( epoch, model, loader, optimizer, args, lr_scheduler=None, saver=None, output_dir='', amp_autocast=suppress, loss_scaler=None, model_ema=None, num_...
import yaml import argparse from timm.utils import add_bool_arg def get_args(config_file=None): def _parse_args(): if config_file: with open(config_file, 'r') as f: cfg = yaml.safe_load(f) parser.set_defaults(**cfg) # There may be remaining unrecognized ...
import os import sys import patch from pathlib import Path import subprocess from utils import s3_utils def patch_effdet(): import effdet current_dir = os.path.dirname(os.path.abspath(__file__)) patch_file = os.path.join(current_dir, "effdet.patch") target_dir = os.path.dirname(effdet.__file__) p =...
import os import torch from torch.distributed._tensor import DeviceMesh from torch.distributed.tensor.parallel import parallelize_module from torch.distributed.tensor.parallel.style import ColwiseParallel, RowwiseParallel from torchbenchmark.tasks import NLP from ...util.model import BenchmarkModel from .model import...
"""Full definition of a LLaMA Language Model, all of it in this single file. Based on the nanoGPT implementation: https://github.com/karpathy/nanoGPT. """ # mypy: ignore-errors import math from dataclasses import dataclass from typing import List, Optional, Tuple, Union from typing_extensions import Self import torch...
from ...util.model import BenchmarkModel from torchbenchmark.tasks import NLP import torch from .model import SequenceGenerator, create_model import torch class Model(BenchmarkModel): task = NLP.LANGUAGE_MODELING DEFAULT_EVAL_BSIZE = 1 def __init__(self, test, device, batch_size=None, extra_args=[]): ...
# Copyright (c) Meta Platforms, Inc. and affiliates. # Portions of this code are derived from https://github.com/facebookresearch/metaseq import torch import torch.nn as nn import torch.nn.functional as F import torch.utils.benchmark as benchmark from torch import Tensor from typing import Optional, Dict, Any from tqd...
from torchbenchmark.util.framework.timm.model_factory import TimmModel from torchbenchmark.tasks import COMPUTER_VISION class Model(TimmModel): task = COMPUTER_VISION.GENERATION DEFAULT_TRAIN_BSIZE = 32 DEFAULT_EVAL_BSIZE = 32 def __init__(self, test, device, batch_size=None, extra_args=[]): ...
from torchbenchmark.tasks import NLP from torchbenchmark.util.framework.huggingface.model_factory import HuggingFaceModel class Model(HuggingFaceModel): task = NLP.LANGUAGE_MODELING DEFAULT_TRAIN_BSIZE = 4 DEFAULT_EVAL_BSIZE = 1 def __init__(self, test, device, batch_size=None, extra_args=[]): ...
import subprocess import sys import os from torchbenchmark.util.framework.huggingface.patch_hf import patch_transformers, cache_model def pip_install_requirements(): subprocess.check_call([sys.executable, '-m', 'pip', 'install', '-q', '-r', 'requirements.txt']) if __name__ == '__main__': pip_install_requirem...
import numpy as np import random import time import torch from argparse import Namespace from .meta import Meta from pathlib import Path from typing import Tuple from ...util.model import BenchmarkModel from torchbenchmark.tasks import OTHER torch.backends.cudnn.deterministic = True torch.backends.cudnn.benchmark = Fa...
import torch from torch import nn from torch import optim from torch.nn import functional as F from torch.utils.data import TensorDataset, DataLoader from torch import optim import numpy as np from .learner import Learner from copy import deepcopy class Meta(nn.Module): """ Meta Learn...
import torch from torch import nn from torch.nn import functional as F import numpy as np from typing import List class Learner(nn.Module): """ """ def __init__(self, config, imgc, imgsz): """ :param config: network config file, type:list of (string, list) :param imgc: ...
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. import numpy as np import torch from typing import Optional, Tuple from .sam import Sam from .transforms import ResizeLo...
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. import torch from functools import partial from .image_encoder import ImageEncoderViT from .mask_decoder import MaskDeco...
# Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. import numpy as np import torch from torch.nn import functional as F from torchvision.transforms.functional import resize,...