python_code
stringlengths
0
229k
import subprocess import sys import os from torchbenchmark.util.framework.huggingface.patch_hf import patch_transformers, cache_model if __name__ == '__main__': model_name = os.path.basename(os.path.dirname(os.path.abspath(__file__))) cache_model(model_name)
import torch import sys import os from torchbenchmark import REPO_PATH from typing import Tuple # Import FAMBench model path class add_path(): def __init__(self, path): self.path = path def __enter__(self): sys.path.insert(0, self.path) def __exit__(self, exc_type, exc_value, traceback): ...
import os import sys import torch import subprocess from torchbenchmark import REPO_PATH def update_fambench_submodule(): "Update FAMBench submodule of the benchmark repo" update_command = ["git", "submodule", "update", "--init", "--recursive", os.path.join("submodules","FAMBench")] ...
import os from torchbenchmark.tasks import COMPUTER_VISION from torchbenchmark.util.framework.detectron2.model_factory import Detectron2Model MODEL_NAME = os.path.basename(os.path.dirname(os.path.abspath(__file__))) MODEL_DIR = os.path.abspath(os.path.dirname(__file__)) class Model(Detectron2Model): task = COMPUT...
import os from torchbenchmark.util.framework.detectron2 import install_detectron2 MODEL_NAME = os.path.basename(os.path.dirname(os.path.abspath(__file__))) MODEL_DIR = os.path.abspath(os.path.dirname(__file__)) if __name__ == '__main__': install_detectron2(MODEL_NAME, MODEL_DIR)
from torchbenchmark.tasks import NLP from torchbenchmark.util.framework.huggingface.model_factory import HuggingFaceModel class Model(HuggingFaceModel): task = NLP.LANGUAGE_MODELING DEFAULT_TRAIN_BSIZE = 4 DEFAULT_EVAL_BSIZE = 1 def __init__(self, test, device, batch_size=None, extra_args=[]): ...
import subprocess import sys import os from torchbenchmark.util.framework.huggingface.patch_hf import patch_transformers, cache_model def pip_install_requirements(): subprocess.check_call([sys.executable, '-m', 'pip', 'install', '-q', '-r', 'requirements.txt']) if __name__ == '__main__': pip_install_requirem...
import argparse import random import torch import numpy as np from torchbenchmark.util.env_check import set_random_seed from .bert_pytorch import parse_args from .bert_pytorch.trainer import BERTTrainer from .bert_pytorch.dataset import BERTDataset, WordVocab from .bert_pytorch.model import BERT from torch.utils.dat...
import unittest from bert_pytorch import BERT class BERTVocabTestCase(unittest.TestCase): pass
import subprocess import sys def setup_install(): subprocess.check_call([sys.executable, 'setup.py', 'develop']) if __name__ == '__main__': setup_install()
import argparse from .model import BERT def parse_args(args=None): parser = argparse.ArgumentParser() parser.add_argument("--script", required=False, action="store_true") parser.add_argument("-d", "--debug", required=False, type=str, default=None) parser.add_argument("-c", "--train_dataset", required=...
import argparse from torch.utils.data import DataLoader from bert_pytorch import parse_args from .model import BERT from .trainer import BERTTrainer from .dataset import BERTDataset, WordVocab import random import torch import numpy as np def train(): # Make all randomness deterministic random.seed(1337) ...
import pickle from collections import Counter class TorchVocab(object): """Defines a vocabulary object that will be used to numericalize a field. Attributes: freqs: A collections.Counter object holding the frequencies of tokens in the data used to build the Vocab. stoi: A collectio...
from .dataset import BERTDataset from .vocab import WordVocab
from torch.utils.data import Dataset import torch import random QUIET = True class BERTDataset(Dataset): def __init__(self, corpus_path, vocab, seq_len, encoding="utf-8", corpus_lines=None, on_memory=True, generator = None): self.vocab = vocab self.seq_len = seq_len self.on_memory = on_...
import torch.nn as nn from .bert import BERT class BERTLM(nn.Module): """ BERT Language Model Next Sentence Prediction Model + Masked Language Model """ def __init__(self, bert: BERT, vocab_size): """ :param bert: BERT model which should be trained :param vocab_size: tota...
from .bert import BERT from .language_model import BERTLM
import torch import torch.nn as nn from .attention import MultiHeadedAttention from .utils import SublayerConnection, PositionwiseFeedForward from .utils.tensor2tensor import TensorToTensor class LambdaModule(torch.nn.Module): def __init__(self, att): super().__init__() self.attention = att ...
import torch import torch.nn as nn from .transformer import TransformerBlock from .embedding import BERTEmbedding class BERT(nn.Module): """ BERT model : Bidirectional Encoder Representations from Transformers. """ def __init__(self, vocab_size, hidden=768, n_layers=12, attn_heads=12, dropout=0.1): ...
import torch.nn as nn import torch.nn.functional as F import torch import math from ..utils.tensor2tensor import TensorToTensor from typing import Optional class Attention(nn.Module): """ Compute 'Scaled Dot Product Attention """ def forward(self, query, key, value, dropout: TensorToTensor, mask: Opt...
from .multi_head import MultiHeadedAttention from .single import Attention
import torch import torch.nn as nn from .single import Attention from typing import Optional class DropoutWrapper(nn.Module): def __init__(self, p): super().__init__() self.dropout = nn.Dropout(p=p) def forward(self, x: torch.Tensor) -> torch.Tensor: return self.dropout(x) class Mult...
import torch.nn as nn class TokenEmbedding(nn.Embedding): def __init__(self, vocab_size, embed_size=512): super().__init__(vocab_size, embed_size, padding_idx=0)
from .bert import BERTEmbedding
import torch.nn as nn import torch import math class PositionalEmbedding(nn.Module): def __init__(self, d_model, max_len=512): super().__init__() # Compute the positional encodings once in log space. pe = torch.zeros(max_len, d_model).float() # Changed from upstream, see https://...
import torch.nn as nn class SegmentEmbedding(nn.Embedding): def __init__(self, embed_size=512): super().__init__(3, embed_size, padding_idx=0)
import torch import torch.nn as nn from .token import TokenEmbedding from .position import PositionalEmbedding from .segment import SegmentEmbedding class BERTEmbedding(nn.Module): """ BERT Embedding which is consisted with under features 1. TokenEmbedding : normal embedding matrix 2. Position...
import torch import torch.nn as nn from .layer_norm import LayerNorm from .tensor2tensor import TensorToTensor class SublayerConnection(nn.Module): """ A residual connection followed by a layer norm. Note for code simplicity the norm is first as opposed to last. """ def __init__(self, size, dropou...
import torch @torch.jit.interface class TensorToTensor(torch.nn.Module): def forward(self, x: torch.Tensor) -> torch.Tensor: pass
import torch import torch.nn as nn class PositionwiseFeedForward(nn.Module): "Implements FFN equation." def __init__(self, d_model, d_ff, dropout=0.1): super(PositionwiseFeedForward, self).__init__() self.w_1 = nn.Linear(d_model, d_ff) self.w_2 = nn.Linear(d_ff, d_model) self....
from .feed_forward import PositionwiseFeedForward from .layer_norm import LayerNorm from .sublayer import SublayerConnection
import torch.nn as nn import torch class LayerNorm(nn.Module): "Construct a layernorm module (See citation for details)." def __init__(self, features, eps=1e-6): super(LayerNorm, self).__init__() self.a_2 = nn.Parameter(torch.ones(features)) self.b_2 = nn.Parameter(torch.zeros(feature...
'''A wrapper class for optimizer ''' import numpy as np class ScheduledOptim(): '''A simple wrapper class for learning rate scheduling''' def __init__(self, optimizer, d_model, n_warmup_steps): self._optimizer = optimizer self.n_warmup_steps = n_warmup_steps self.n_current_steps = 0 ...
import torch import torch.nn as nn from torch.optim import Adam from torch.utils.data import DataLoader from ..model import BERTLM, BERT from .optim_schedule import ScheduledOptim class BERTTrainer: """ BERTTrainer make the pretrained BERT model with two LM training method. 1. Masked Language Model ...
from .pretrain import BERTTrainer
"Doctr detection model" from doctr.models import ocr_predictor import numpy as np import torch # TorchBench imports from torchbenchmark.util.model import BenchmarkModel from torchbenchmark.tasks import COMPUTER_VISION from typing import Tuple class Model(BenchmarkModel): task = COMPUTER_VISION.DETECTION DEFA...
import os import warnings import subprocess import sys def pip_install_requirements(): try: subprocess.check_call(["conda", "install", "-y", "expecttest", "libglib", "pango", "-c", "conda-forge"]) except: warnings.warn("The doctr_det_predictor model requires conda binary libaries to be installe...
import torch import torch.optim as optim import torch.nn as nn import torchvision.models as models from functorch import make_functional_with_buffers, vmap, grad from typing import Tuple from ...util.model import BenchmarkModel from torchbenchmark.tasks import OTHER def compute_norms(sample_grads): batch_size = ...
import subprocess import sys def pip_install_requirements(): subprocess.check_call([sys.executable, '-m', 'pip', 'install', '-q', '-r', 'requirements.txt']) if __name__ == '__main__': pip_install_requirements()
from torchvision.utils import save_image import torch import torch.nn.functional as F import numpy as np import os import time import datetime from .model import Generator from .model import Discriminator class Solver: """Solver for training and testing StarGAN.""" def __init__(self, celeba_loader, rafd_load...
from torch.utils import data from torchvision import transforms as T from torchvision.datasets import ImageFolder from PIL import Image import torch import os import random class CelebA(data.Dataset): """Dataset class for the CelebA dataset.""" def __init__(self, image_dir, attr_path, selected_attrs, transfo...
#!/usr/bin/env python import os import torch import random import numpy as np from .solver import Solver from .data_loader import get_loader from .main import parse_config, makedirs from typing import Tuple from ...util.model import BenchmarkModel from torchbenchmark.tasks import COMPUTER_VISION from torchbenchmark imp...
import tensorflow as tf class Logger: """Tensorboard logger.""" def __init__(self, log_dir): """Initialize summary writer.""" self.writer = tf.summary.create_file_writer(log_dir) def scalar_summary(self, tag, value, step): """Add scalar summary.""" with self.writer.as_de...
import torch import torch.nn as nn import torch.nn.functional as F import numpy as np class ResidualBlock(nn.Module): """Residual Block with instance normalization.""" def __init__(self, dim_in, dim_out): super(ResidualBlock, self).__init__() self.main = nn.Sequential( nn.Conv2d(di...
import subprocess import sys from utils import s3_utils def pip_install_requirements(): subprocess.check_call([sys.executable, '-m', 'pip', 'install', '-q', '-r', 'requirements.txt']) if __name__ == '__main__': s3_utils.checkout_s3_data("INPUT_TARBALLS", "pytorch_stargan_inputs.tar.gz", decompress=True) p...
import os import argparse import torch from torch.backends import cudnn from .solver import Solver from .data_loader import get_loader def str2bool(v): return v.lower() in ('true') def makedirs(config): "Create directories if not exist." if not os.path.exists(config.log_dir): os.makedirs(config....
import os from torchbenchmark.tasks import COMPUTER_VISION from torchbenchmark.util.framework.detectron2.model_factory import Detectron2Model MODEL_NAME = os.path.basename(os.path.dirname(os.path.abspath(__file__))) MODEL_DIR = os.path.abspath(os.path.dirname(__file__)) class Model(Detectron2Model): task = COMPUT...
import os from torchbenchmark.util.framework.detectron2 import install_detectron2 MODEL_NAME = os.path.basename(os.path.dirname(os.path.abspath(__file__))) MODEL_DIR = os.path.abspath(os.path.dirname(__file__)) if __name__ == '__main__': install_detectron2(MODEL_NAME, MODEL_DIR)
#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved import argparse import builtins import math import os import random import shutil import time import warnings import torch import torch.nn as nn import torch.nn.parallel import torch.backends.cudnn as cudnn import torch.distri...
#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved from argparse import Namespace import random import torch import torch.nn as nn import torch.nn.parallel import torch.backends.cudnn as cudnn import torch.distributed as dist import torch.optim import torch.utils.data import t...
# only needs torch and torchvision if __name__ == '__main__': pass
#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved import argparse import builtins import os import random import shutil import time import warnings import torch import torch.nn as nn import torch.nn.parallel import torch.backends.cudnn as cudnn import torch.distributed as dis...
# Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved
# Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved import torch import torch.nn as nn class MoCo(nn.Module): """ Build a MoCo model with: a query encoder, a key encoder, and a queue https://arxiv.org/abs/1911.05722 """ def __init__(self, base_encoder, dim=128, K=65536, m=0.999,...
# Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved from PIL import ImageFilter import random class TwoCropsTransform: """Take two random crops of one image as the query and key.""" def __init__(self, base_transform): self.base_transform = base_transform def __call__(self, x):...
#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved import pickle as pkl import sys import torch if __name__ == "__main__": input = sys.argv[1] obj = torch.load(input, map_location="cpu") obj = obj["state_dict"] newmodel = {} for k, v in obj.items(): ...
#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved import os from detectron2.checkpoint import DetectionCheckpointer from detectron2.config import get_cfg from detectron2.engine import DefaultTrainer, default_argument_parser, default_setup, launch from detectron2.evaluation i...
import os from torchbenchmark.tasks import COMPUTER_VISION from torchbenchmark.util.framework.detectron2.model_factory import Detectron2Model MODEL_NAME = os.path.basename(os.path.dirname(os.path.abspath(__file__))) MODEL_DIR = os.path.abspath(os.path.dirname(__file__)) class Model(Detectron2Model): task = COMPUT...
import os from torchbenchmark.util.framework.detectron2 import install_detectron2 MODEL_NAME = os.path.basename(os.path.dirname(os.path.abspath(__file__))) MODEL_DIR = os.path.abspath(os.path.dirname(__file__)) if __name__ == '__main__': install_detectron2(MODEL_NAME, MODEL_DIR)
import cv2 import torch import random import numpy as np from .baseline.Renderer.model import FCN from .baseline.DRL.evaluator import Evaluator from .baseline.utils.util import * from .baseline.DRL.ddpg import DDPG from .baseline.DRL.multi import fastenv from ...util.model import BenchmarkModel from typing import Tupl...
import subprocess import sys from utils import s3_utils def pip_install_requirements(): subprocess.check_call([sys.executable, '-m', 'pip', 'install', '-q', '-r', 'requirements.txt']) if __name__ == '__main__': s3_utils.checkout_s3_data("INPUT_TARBALLS", "Super_SloMo_inputs.tar.gz", decompress=True) pip_...
import sys import json import torch import numpy as np import argparse import torchvision.transforms as transforms import cv2 from DRL.ddpg import decode from utils.util import * from PIL import Image from torchvision import transforms, utils device = torch.device("cuda" if torch.cuda.is_available() else "cpu") aug = ...
import os import cv2 import torch import numpy as np import argparse import torch.nn as nn import torch.nn.functional as F from DRL.actor import * from Renderer.stroke_gen import * from Renderer.model import * device = torch.device("cuda" if torch.cuda.is_available() else "cpu") width = 128 parser = argparse.Argumen...
#!/usr/bin/env python3 import cv2 import random import numpy as np import argparse from DRL.evaluator import Evaluator from utils.util import * from utils.tensorboard import TensorBoard import time exp = os.path.abspath('.').split('/')[-1] writer = TensorBoard('../train_log/{}'.format(exp)) os.system('ln -sf ../train_...
import cv2 import torch import numpy as np import torch.nn as nn import torch.nn.functional as F from utils.tensorboard import TensorBoard from Renderer.model import FCN from Renderer.stroke_gen import * writer = TensorBoard("../train_log/") import torch.optim as optim criterion = nn.MSELoss() net = FCN() optimizer ...
import cv2 import numpy as np def normal(x, width): return (int)(x * (width - 1) + 0.5) def draw(f, width=128): x0, y0, x1, y1, x2, y2, z0, z2, w0, w2 = f x1 = x0 + (x2 - x0) * x1 y1 = y0 + (y2 - y0) * y1 x0 = normal(x0, width * 2) x1 = normal(x1, width * 2) x2 = normal(x2, width * 2) ...
import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.utils.weight_norm as weightNorm class FCN(nn.Module): def __init__(self): super(FCN, self).__init__() self.fc1 = (nn.Linear(10, 512)) self.fc2 = (nn.Linear(512, 1024)) self.fc3 = (nn.Linear(1024, 2048...
import torch import torch.nn as nn import numpy as np from torch.optim import Adam, SGD from torch import autograd from torch.autograd import Variable import torch.nn.functional as F from torch.autograd import grad as torch_grad import torch.nn.utils.weight_norm as weightNorm from utils.util import * device = torch.de...
import cv2 import torch import numpy as np from env import Paint from utils.util import * from DRL.ddpg import decode device = torch.device("cuda" if torch.cuda.is_available() else "cpu") class fastenv(): def __init__(self, max_episode_length=10, env_batch=64, \ writer=None): ...
import numpy as np import torch import torch.nn as nn import torch.nn.functional as F from torch.optim import Adam, SGD from Renderer.model import * from DRL.rpm import rpm from DRL.actor import * from DRL.critic import * from DRL.wgan import * from utils.util import * device = torch.device("cuda" if torch.cuda.is_avai...
# from collections import deque import numpy as np import random import torch import pickle as pickle class rpm: # replay memory def __init__(self, buffer_size): self.buffer_size = buffer_size self.buffer = [] self.index = 0 def append(self, obj): if self.size() > s...
import numpy as np import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.utils.weight_norm as weightNorm from torch.autograd import Variable import sys def conv3x3(in_planes, out_planes, stride=1): return (nn.Conv2d(in_planes, out_planes, kernel_size=3, stride=stride, padding=1, bias...
import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.utils.weight_norm as weightNorm from torch.autograd import Variable import sys def conv3x3(in_planes, out_planes, stride=1): return weightNorm(nn.Conv2d(in_planes, out_planes, kernel_size=3, stride=stride, padding=1, bias=True)) d...
import numpy as np from utils.util import * class Evaluator: def __init__(self, args, writer): self.validate_episodes = args.validate_episodes self.max_step = args.max_step self.env_batch = args.env_batch self.writer = writer self.log = 0 def __call__(self, env, po...
import os import torch from torch.autograd import Variable USE_CUDA = torch.cuda.is_available() def prRed(prt): print("\033[91m {}\033[00m" .format(prt)) def prGreen(prt): print("\033[92m {}\033[00m" .format(prt)) def prYellow(prt): print("\033[93m {}\033[00m" .format(prt)) def prLightPurple(prt): print("\033[94m {}\...
import PIL import scipy.misc from io import BytesIO import tensorboardX as tb from tensorboardX.summary import Summary class TensorBoard: def __init__(self, model_dir): self.summary_writer = tb.FileWriter(model_dir) def add_image(self, tag, img, step): summary = Summary() bio = BytesIO...
import sys import json import torch import numpy as np import argparse import torchvision.transforms as transforms import cv2 from .DRL.ddpg import decode from .utils.util import * from PIL import Image # device = torch.device("cuda" if torch.cuda.is_available() else "cpu") aug = transforms.Compose([transforms.ToPILI...
import os import cv2 import torch import numpy as np import argparse import torch.nn as nn import torch.nn.functional as F from DRL.actor import * from Renderer.stroke_gen import * from Renderer.model import * device = torch.device("cuda" if torch.cuda.is_available() else "cpu") width = 128 parser = argparse.Argumen...
#!/usr/bin/env python3 import cv2 import random import numpy as np import argparse from DRL.evaluator import Evaluator from utils.util import * from utils.tensorboard import TensorBoard import time exp = os.path.abspath('.').split('/')[-1] writer = TensorBoard('../train_log/{}'.format(exp)) os.system('ln -sf ../train_...
import cv2 import torch import numpy as np import sys import torch.nn as nn import torch.nn.functional as F from utils.tensorboard import TensorBoard from Renderer.model import FCN from Renderer.stroke_gen import * #writer = TensorBoard("../train_log/") import torch.optim as optim import argparse parser = argparse.A...
import cv2 import numpy as np def normal(x, width): return (int)(x * (width - 1) + 0.5) def draw(f, width=128): x0, y0, x1, y1, x2, y2, z0, z2, w0, w2 = f x1 = x0 + (x2 - x0) * x1 y1 = y0 + (y2 - y0) * y1 x0 = normal(x0, width * 2) x1 = normal(x1, width * 2) x2 = normal(x2, width * 2) ...
import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.utils.weight_norm as weightNorm class FCN(nn.Module): def __init__(self): super(FCN, self).__init__() self.fc1 = (nn.Linear(10, 512)) self.fc2 = (nn.Linear(512, 1024)) self.fc3 = (nn.Linear(1024, 2048...
import torch import torch.nn as nn import numpy as np from torch.optim import Adam, SGD from torch import autograd from torch.autograd import Variable import torch.nn.functional as F from torch.autograd import grad as torch_grad import torch.nn.utils.weight_norm as weightNorm from ..utils.util import * device = torch....
import cv2 import torch import numpy as np from ..env import Paint from ..utils.util import * # device = torch.device("cuda" if torch.cuda.is_available() else "cpu") class fastenv(): def __init__(self, max_episode_length=10, env_batch=64, writer=None, images=None, device="cpu", ...
import numpy as np import torch import torch.nn as nn import torch.nn.functional as F from torch.optim import Adam, SGD from ..Renderer.model import * from .rpm import rpm from .actor import * from .critic import * from .wgan import * from ..utils.util import * # device = torch.device("cuda" if torch.cuda.is_available...
# from collections import deque import numpy as np import random import torch import pickle as pickle class rpm(object): # replay memory def __init__(self, buffer_size): self.buffer_size = buffer_size self.buffer = [] self.index = 0 def append(self, obj): if self.si...
import numpy as np import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.utils.weight_norm as weightNorm from torch.autograd import Variable import sys def conv3x3(in_planes, out_planes, stride=1): return (nn.Conv2d(in_planes, out_planes, kernel_size=3, stride=stride, padding=1, bias...
import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.utils.weight_norm as weightNorm from torch.autograd import Variable import sys def conv3x3(in_planes, out_planes, stride=1): return weightNorm(nn.Conv2d(in_planes, out_planes, kernel_size=3, stride=stride, padding=1, bias=True)) c...
import numpy as np from ..utils.util import * class Evaluator(object): def __init__(self, args, env_batch, writer): self.validate_episodes = args.validate_episodes self.max_step = args.max_step self.env_batch = env_batch self.writer = writer self.log = 0 def __call__(...
import os import torch from torch.autograd import Variable USE_CUDA = torch.cuda.is_available() def prRed(prt): print("\033[91m {}\033[00m" .format(prt)) def prGreen(prt): print("\033[92m {}\033[00m" .format(prt)) def prYellow(prt): print("\033[93m {}\033[00m" .format(prt)) def prLightPurple(prt): print("\033[94m {}\...
import PIL import scipy.misc from io import BytesIO import tensorboardX as tb from tensorboardX.summary import Summary class TensorBoard(object): def __init__(self, model_dir): self.summary_writer = tb.FileWriter(model_dir) def add_image(self, tag, img, step): summary = Summary() bio =...
from torchbenchmark.tasks import NLP from torchbenchmark.util.framework.huggingface.model_factory import HuggingFaceModel class Model(HuggingFaceModel): task = NLP.LANGUAGE_MODELING # Original train batch size per device: 8 # Source: https://github.com/huggingface/transformers/blob/master/examples/flax/lan...
import subprocess import sys import os from torchbenchmark.util.framework.huggingface.patch_hf import patch_transformers, cache_model def pip_install_requirements(): subprocess.check_call([sys.executable, '-m', 'pip', 'install', '-q', '-r', 'requirements.txt']) if __name__ == '__main__': pip_install_requirem...
from torchbenchmark.util.framework.vision.model_factory import TorchVisionModel from torchbenchmark.tasks import COMPUTER_VISION import torchvision.models as models class Model(TorchVisionModel): task = COMPUTER_VISION.CLASSIFICATION # Train batch size: use the smallest example batch of 128 (assuming only 1 wo...
from torchbenchmark.util.model import BenchmarkModel from torchbenchmark.tasks import COMPUTER_VISION import torch.nn as nn import torch from types import SimpleNamespace import torch.utils.data as data class DenseLayer(nn.Module): def __init__(self, c_in, bn_size, growth_rate, act_fn): """ Input...
from torchbenchmark.util.framework.timm.model_factory import TimmModel from torchbenchmark.tasks import COMPUTER_VISION class Model(TimmModel): task = COMPUTER_VISION.DETECTION DEFAULT_TRAIN_BSIZE = 32 DEFAULT_EVAL_BSIZE = 32 def __init__(self, test, device, batch_size=None, extra_args=[]): s...