id
int64
0
190k
prompt
stringlengths
21
13.4M
docstring
stringlengths
1
12k
184,614
import math import torch import torch.nn as nn from functools import partial from modeling_finetune import Block, _cfg, PatchEmbed, RelativePositionBias from timm.models.registry import register_model from timm.models.layers import trunc_normal_ as __call_trunc_normal_ class VisionTransformerForMaskedImageModeling(nn.M...
null
184,615
import math import torch import torch.nn as nn from functools import partial from modeling_finetune import Block, _cfg, PatchEmbed, RelativePositionBias from timm.models.registry import register_model from timm.models.layers import trunc_normal_ as __call_trunc_normal_ class VisionTransformerForMaskedImageModelingCLS(V...
null
184,616
import math import torch import torch.nn as nn from functools import partial from modeling_finetune import Block, _cfg, PatchEmbed, RelativePositionBias from timm.models.registry import register_model from timm.models.layers import trunc_normal_ as __call_trunc_normal_ class VisionTransformerForMaskedImageModeling(nn.M...
null
184,617
import math import torch import torch.nn as nn from functools import partial from modeling_finetune import Block, _cfg, PatchEmbed, RelativePositionBias from timm.models.registry import register_model from timm.models.layers import trunc_normal_ as __call_trunc_normal_ class VisionTransformerForMaskedImageModelingCLS(V...
null
184,618
import math import torch import torch.nn as nn from functools import partial from modeling_finetune import Block, _cfg, PatchEmbed, RelativePositionBias from timm.models.registry import register_model from timm.models.layers import trunc_normal_ as __call_trunc_normal_ class VisionTransformerForMaskedImageModeling(nn.M...
null
184,619
import math import sys from typing import Iterable import torch import torch.nn as nn import utils def train_one_epoch(model: torch.nn.Module, data_loader: Iterable, optimizer: torch.optim.Optimizer, device: torch.device, ...
null
184,620
import math import sys from typing import Iterable import torch import torch.nn as nn import utils def evaluate(data_loader, model, device, log_writer=None, epoch=None, args=None): metric_logger = utils.MetricLogger(delimiter=" ") header = 'Validation:' # switch to evaluation mode model.eval() ...
null
184,621
import math import sys from typing import Iterable import torch import torch.nn as nn import utils def calculate_codebook_usage(data_loader, model, device, log_writer=None, epoch=None, args=None): metric_logger = utils.MetricLogger(delimiter=" ") header = 'Calculating codebook usage:' # switch to evalua...
null
184,622
from cgitb import enable import math import sys from typing import Iterable import torch import torch.nn as nn import torch.nn.functional as F import utils def train_one_epoch(model: torch.nn.Module, vqkd: torch.nn.Module, data_loader: Iterable, optimizer: torch.optim.Optimizer, ...
null
184,623
import argparse import datetime import numpy as np import time import torch import torch.backends.cudnn as cudnn import json import os from pathlib import Path from timm.models import create_model from optim_factory import create_optimizer from datasets import build_beit_pretraining_dataset from engine_for_pretraining ...
null
184,624
import argparse import datetime import numpy as np import time import torch import torch.backends.cudnn as cudnn import json import os from pathlib import Path from timm.models import create_model from optim_factory import create_optimizer from datasets import build_beit_pretraining_dataset from engine_for_pretraining ...
null
184,625
import argparse import datetime import numpy as np import time import torch import torch.backends.cudnn as cudnn import json import os from pathlib import Path from timm.models import create_model from optim_factory import create_optimizer from datasets import build_beit_pretraining_dataset from engine_for_pretraining ...
null
184,626
import torch import numpy as np from torch import nn, einsum import torch.nn.functional as F import math from collections import OrderedDict from functools import partial, reduce from einops import rearrange from timm.models.layers import trunc_normal_ from timm.data.constants import IMAGENET_DEFAULT_MEAN, IMAGENET_DEF...
null
184,627
import torch import numpy as np from torch import nn, einsum import torch.nn.functional as F import math from collections import OrderedDict from functools import partial, reduce from einops import rearrange from timm.models.layers import trunc_normal_ from timm.data.constants import IMAGENET_DEFAULT_MEAN, IMAGENET_DEF...
null
184,628
import torch import numpy as np from torch import nn, einsum import torch.nn.functional as F import math from collections import OrderedDict from functools import partial, reduce from einops import rearrange from timm.models.layers import trunc_normal_ from timm.data.constants import IMAGENET_DEFAULT_MEAN, IMAGENET_DEF...
null
184,629
import logging import os import sys from dataclasses import dataclass, field from typing import Optional import numpy as np from datasets import ClassLabel, load_dataset, load_metric import layoutlmft.data.datasets.funsd import transformers from layoutlmft.data import DataCollatorForKeyValueExtraction from layoutlmft.d...
null
184,630
import torch from detectron2.data.detection_utils import read_image from detectron2.data.transforms import ResizeTransform, TransformList def normalize_bbox(bbox, size): return [ int(1000 * bbox[0] / size[0]), int(1000 * bbox[1] / size[1]), int(1000 * bbox[2] / size[0]), int(1000 * ...
null
184,631
import torch from detectron2.data.detection_utils import read_image from detectron2.data.transforms import ResizeTransform, TransformList def simplify_bbox(bbox): return [ min(bbox[0::2]), min(bbox[1::2]), max(bbox[2::2]), max(bbox[3::2]), ]
null
184,632
import torch from detectron2.data.detection_utils import read_image from detectron2.data.transforms import ResizeTransform, TransformList def merge_bbox(bbox_list): x0, y0, x1, y1 = list(zip(*bbox_list)) return [min(x0), min(y0), max(x1), max(y1)]
null
184,633
import torch from detectron2.data.detection_utils import read_image from detectron2.data.transforms import ResizeTransform, TransformList def load_image(image_path): image = read_image(image_path, format="BGR") h = image.shape[0] w = image.shape[1] img_trans = TransformList([ResizeTransform(h=h, w=w, n...
null
184,634
import os import re import numpy as np from transformers.utils import logging _re_checkpoint = re.compile(r"^" + PREFIX_CHECKPOINT_DIR + r"\-(\d+)$") def get_last_checkpoint(folder): content = os.listdir(folder) checkpoints = [ path for path in content if _re_checkpoint.search(path) is ...
null
184,635
import os import re import numpy as np from transformers.utils import logging logger = logging.get_logger(__name__) The provided code snippet includes necessary dependencies for implementing the `re_score` function. Write a Python function `def re_score(pred_relations, gt_relations, mode="strict")` to solve the follow...
Evaluate RE predictions Args: pred_relations (list) : list of list of predicted relations (several relations in each sentence) gt_relations (list) : list of list of ground truth relations rel = { "head": (start_idx (inclusive), end_idx (exclusive)), "tail": (start_idx (inclusive), end_idx (exclusive)), "head_type": ent...
184,636
import math import torch import torch.nn.functional as F import torch.utils.checkpoint from torch import nn from torch.nn import CrossEntropyLoss import detectron2 from detectron2.modeling import META_ARCH_REGISTRY from transformers import PreTrainedModel from transformers.modeling_outputs import ( BaseModelOutputW...
null
184,637
import math import torch import torch.nn.functional as F import torch.utils.checkpoint from torch import nn from torch.nn import CrossEntropyLoss import detectron2 from detectron2.modeling import META_ARCH_REGISTRY from transformers import PreTrainedModel from transformers.modeling_outputs import ( BaseModelOutputW...
null
184,638
def add_layoutlmv2_config(cfg): _C = cfg # ----------------------------------------------------------------------------- # Config definition # ----------------------------------------------------------------------------- _C.MODEL.MASK_ON = True # When using pre-trained models in Detectron1 or...
null
184,639
import random import numpy as np import torch import os import shutil import sys The provided code snippet includes necessary dependencies for implementing the `set_seed` function. Write a Python function `def set_seed(args)` to solve the following problem: Set seed for reproducibility Here is the function: def set_...
Set seed for reproducibility
184,640
import random import numpy as np import torch import os import shutil import sys The provided code snippet includes necessary dependencies for implementing the `set_exp_folder` function. Write a Python function `def set_exp_folder(args)` to solve the following problem: Create a folder to store experimental results e.g...
Create a folder to store experimental results e.g., checkpoints or log
184,641
import random import numpy as np import torch import os import shutil import sys The provided code snippet includes necessary dependencies for implementing the `check_screen` function. Write a Python function `def check_screen()` to solve the following problem: Check whether the experiment is in screen Here is the fu...
Check whether the experiment is in screen
184,642
from genericpath import exists import os import torch.nn as nn import torch import logging from tqdm import tqdm, trange import timeit import collections import json import math from bs4 import BeautifulSoup from copy import deepcopy import string import re from torch.utils.tensorboard import SummaryWriter from torch.u...
null
184,643
import os import torch import collections import logging from tqdm import tqdm, trange import json import bs4 from os import path as osp from bs4 import BeautifulSoup as bs from torch.utils.data import Dataset import networkx as nx from lxml import etree import pickle from transformers import BertTokenizer import argpa...
null
184,644
import os import torch import collections import logging from tqdm import tqdm, trange import json import bs4 from os import path as osp from bs4 import BeautifulSoup as bs from torch.utils.data import Dataset import networkx as nx from lxml import etree import pickle from transformers import BertTokenizer import argpa...
null
184,645
import logging import os import sys from dataclasses import dataclass, field from typing import Optional import datasets from datasets import load_dataset, load_metric import transformers from trainer_qa import QuestionAnsweringTrainer from transformers import ( AutoConfig, AutoModelForQuestionAnswering, Au...
null
184,648
from __future__ import absolute_import, division, print_function import json import logging import math import collections from io import open from os import path as osp from tqdm import tqdm import bs4 from bs4 import BeautifulSoup as bs from transformers.models.bert.tokenization_bert import BasicTokenizer, whitespace...
r""" replace the special expressions in the html file for specific punctuation.
184,649
from __future__ import absolute_import, division, print_function import json import logging import math import collections from io import open from os import path as osp from tqdm import tqdm import bs4 from bs4 import BeautifulSoup as bs from transformers.models.bert.tokenization_bert import BasicTokenizer, whitespace...
null
184,650
from __future__ import absolute_import, division, print_function import argparse import logging import os import random import glob import timeit import numpy as np import torch from torch.utils.data import (DataLoader, RandomSampler, SequentialSampler) from torch.utils.data.distributed import DistributedSampler from t...
r""" Train the model
184,651
import argparse import collections import json import os import re import string import sys from copy import deepcopy from bs4 import BeautifulSoup import sys sys.path.append(os.getcwd()) def parse_args(): parser = argparse.ArgumentParser() parser.add_argument('data_file', metavar='data.json', help='Input dat...
null
184,652
import argparse import collections import json import os import re import string import sys from copy import deepcopy from bs4 import BeautifulSoup The provided code snippet includes necessary dependencies for implementing the `make_pages_list` function. Write a Python function `def make_pages_list(dataset)` to solve ...
r""" Record all the pages which appears in the dataset and return the list.
184,653
import argparse import collections import json import os import re import string import sys from copy import deepcopy from bs4 import BeautifulSoup The provided code snippet includes necessary dependencies for implementing the `make_qid_to_has_ans` function. Write a Python function `def make_qid_to_has_ans(dataset)` t...
r""" Pick all the questions which has answer in the dataset and return the list.
184,654
import argparse import collections import json import os import re import string import sys from copy import deepcopy from bs4 import BeautifulSoup def normalize_answer(s): """Lower text and remove punctuation, articles and extra whitespace.""" def remove_articles(text): regex = re.compile(r'\b(a|an|the...
r""" Calculate all the three matrix (exact match, f1, POS) for each question. Arguments: dataset (dict): the dataset in use. preds (dict): the answer text prediction for each question in the dataset. tag_preds (dict): the answer tags prediction for each question in the dataset. root_dir (str): the base directory for th...
184,655
import argparse import collections import json import os import re import string import sys from copy import deepcopy from bs4 import BeautifulSoup The provided code snippet includes necessary dependencies for implementing the `make_eval_dict` function. Write a Python function `def make_eval_dict(exact_scores, f1_scor...
r""" Make the dictionary to show the evaluation results.
184,656
import argparse import collections import json import os import re import string import sys from copy import deepcopy from bs4 import BeautifulSoup def merge_eval(main_eval, new_eval, prefix): for k in new_eval: main_eval['%s_%s' % (prefix, k)] = new_eval[k]
null
184,657
import csv import json import argparse import os.path as osp import os from operator import itemgetter def parse_args(): parser = argparse.ArgumentParser() parser.add_argument("--root_dir", default=None, type=str, required=True, help="The root directory of the raw WebSRC dataset; The ou...
null
184,658
import csv import json import argparse import os.path as osp import os from operator import itemgetter def convert_csv_to_dict(args): dir_list = os.walk(args.root_dir) print('Start Converting') data, websites, qas, answers = [], [], [], [] last_domain = None for d, _, fs in dir_list: for ...
null
184,659
import csv import json import argparse import os.path as osp import os from operator import itemgetter def dataset_split(args, dataset): def count(last, curr): if last is None: return False if last != curr: return False return True split = json.load(open(osp.joi...
null
184,660
from __future__ import absolute_import from __future__ import division from __future__ import print_function import os import pickle import sys from absl import app from absl import flags import tqdm import constants The provided code snippet includes necessary dependencies for implementing the `pack_swde_data` functi...
Packs the swde dataset to a single file. Args: swde_path: The path to SWDE dataset pages (http://shortn/_g22KuARPAi). pack_path: The path to save packed SWDE dataset file. cut_off: To shorten the list for testing. Returns: None
184,661
from __future__ import absolute_import, division, print_function import argparse import logging import os import random import glob import numpy as np from torch.utils.data import (DataLoader, RandomSampler, SequentialSampler) from torch.utils.data.distributed import DistributedSampler from tensorboardX import SummaryW...
null
184,662
from __future__ import absolute_import, division, print_function import argparse import logging import os import random import glob import numpy as np from torch.utils.data import (DataLoader, RandomSampler, SequentialSampler) from torch.utils.data.distributed import DistributedSampler from tensorboardX import SummaryW...
r""" Load and process the raw data.
184,663
from __future__ import absolute_import, division, print_function import argparse import logging import os import random import glob import numpy as np from torch.utils.data import (DataLoader, RandomSampler, SequentialSampler) from torch.utils.data.distributed import DistributedSampler from tensorboardX import SummaryW...
null
184,664
from __future__ import absolute_import from __future__ import division from __future__ import print_function import collections import os import pickle import random import re import sys import unicodedata from absl import app from absl import flags import lxml from lxml import etree from lxml.html.clean import Cleaner...
Extracts all the xpaths and labels the nodes for all the pages.
184,665
import math import os import torch import torch.utils.checkpoint from torch import nn from torch.nn import CrossEntropyLoss from transformers.activations import ACT2FN from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward, \ replace_return_docstrings from transformers.model...
Replace non-padding symbols with their position numbers. Position numbers begin at padding_idx+1. Padding symbols are ignored. This is modified from fairseq's `utils.make_positions`. Args: x: torch.Tensor x: Returns: torch.Tensor
184,666
from __future__ import absolute_import from __future__ import division from __future__ import print_function import os import json import glob import logging import argparse import math from tqdm import tqdm import numpy as np import torch import random import pickle from s2s_ft.modeling_decoding import BertForSeq2SeqD...
null
184,667
from __future__ import absolute_import from __future__ import division from __future__ import print_function import os import json import glob import logging import argparse import math from tqdm import tqdm import numpy as np import torch import random import pickle from s2s_ft.modeling_decoding import BertForSeq2SeqD...
null
184,668
from __future__ import absolute_import, division, print_function import argparse import logging import os import json import random import numpy as np import torch from torch.utils.data import (DataLoader, SequentialSampler) from torch.utils.data.distributed import DistributedSampler try: from torch.utils.tensorboa...
Train the model
184,669
from __future__ import absolute_import, division, print_function import argparse import logging import os import json import random import numpy as np import torch from torch.utils.data import (DataLoader, SequentialSampler) from torch.utils.data.distributed import DistributedSampler import tqdm from s2s_ft.modeling im...
null
184,670
from __future__ import absolute_import, division, print_function import argparse import logging import os import json import random import numpy as np import torch from torch.utils.data import (DataLoader, SequentialSampler) from torch.utils.data.distributed import DistributedSampler try: from torch.utils.tensorboa...
null
184,671
from __future__ import absolute_import, division, print_function import argparse import logging import os import json import random import numpy as np import torch from torch.utils.data import (DataLoader, SequentialSampler) from torch.utils.data.distributed import DistributedSampler import tqdm from s2s_ft.modeling im...
null
184,672
import torch import logging from transformers.modeling_utils import cached_path, WEIGHTS_NAME, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME logger = logging.getLogger(__name__) def get_checkpoint_from_transformer_cache( archive_file, pretrained_model_name_or_path, pretrained_model_archive_map, cache_dir, force_do...
null
184,673
import torch import logging from transformers.modeling_utils import cached_path, WEIGHTS_NAME, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME logger = logging.getLogger(__name__) def hf_roberta_to_hf_bert(state_dict): logger.info(" * Convert Huggingface RoBERTa format to Huggingface BERT format * ") new_state_dict = {} ...
null
184,674
import torch import logging from transformers.modeling_utils import cached_path, WEIGHTS_NAME, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME logger = logging.getLogger(__name__) def hf_electra_to_hf_bert(state_dict): logger.info(" * Convert Huggingface ELECTRA format to Huggingface BERT format * ") new_state_dict = {} ...
null
184,675
import torch import logging from transformers.modeling_utils import cached_path, WEIGHTS_NAME, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME def hf_bert_to_hf_bert(state_dict): # keep no change return state_dict
null
184,676
import torch import logging from transformers.modeling_utils import cached_path, WEIGHTS_NAME, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME logger = logging.getLogger(__name__) def unilm_to_hf_bert(state_dict): logger.info(" * Convert Fast QKV format to Huggingface BERT format * ") new_state_dict = {} for key in st...
null
184,677
from __future__ import absolute_import, division, print_function import logging import os import json import random import glob import torch import tqdm import array import collections import torch.utils.data from transformers.file_utils import WEIGHTS_NAME def deserialize_str(x): return x.decode('ascii')
null
184,678
from __future__ import absolute_import, division, print_function import logging import os import json import random import glob import torch import tqdm import array import collections import torch.utils.data from transformers.file_utils import WEIGHTS_NAME try: import lmdb except: pass logger = logging.getLogg...
null
184,679
import numpy as np from random import randint, shuffle, choice from random import random as rand import math import logging import torch import torch.utils.data def get_random_word(vocab_words): i = randint(0, len(vocab_words)-1) return vocab_words[i]
null
184,680
import numpy as np from random import randint, shuffle, choice from random import random as rand import math import logging import torch import torch.utils.data def batch_list_to_batch_tensors(batch): batch_tensors = [] for x in zip(*batch): if x[0] is None: batch_tensors.append(None) ...
null
184,681
import numpy as np from random import randint, shuffle, choice from random import random as rand import math import logging import torch import torch.utils.data def _get_word_split_index(tokens, st, end): split_idx = [] i = st while i < end: if (not tokens[i].startswith('##')) or (i == st): ...
null
184,682
import numpy as np from random import randint, shuffle, choice from random import random as rand import math import logging import torch import torch.utils.data def _expand_whole_word(tokens, st, end): new_st, new_end = st, end while (new_st >= 0) and tokens[new_st].startswith('##'): new_st -= 1 wh...
null
184,683
from __future__ import absolute_import, division, print_function, unicode_literals import logging import math import os import torch from torch import nn from torch.nn.modules.loss import _Loss import torch.nn.functional as F from transformers.modeling_bert import \ BertPreTrainedModel, BertSelfOutput, BertIntermed...
Adapted from Mesh Tensorflow: https://github.com/tensorflow/mesh/blob/0cb87fe07da627bf0b7e60475d59f95ed6b5be3d/mesh_tensorflow/transformer/transformer_layers.py#L593
184,684
from __future__ import absolute_import, division, print_function, unicode_literals import logging import math import os import torch from torch import nn from torch.nn.modules.loss import _Loss import torch.nn.functional as F from transformers.modeling_bert import \ BertPreTrainedModel, BertSelfOutput, BertIntermed...
null
184,685
from __future__ import absolute_import from __future__ import division from __future__ import print_function import os import copy import json import math import logging import tarfile import tempfile import shutil import numpy as np from functools import partial import torch from torch import nn from torch.nn import C...
Implementation of the gelu activation function. For information: OpenAI GPT's gelu is slightly different (and gives slightly different results): 0.5 * x * (1 + torch.tanh(math.sqrt(2 / math.pi) * (x + 0.044715 * torch.pow(x, 3))))
184,686
from __future__ import absolute_import from __future__ import division from __future__ import print_function import os import copy import json import math import logging import tarfile import tempfile import shutil import numpy as np from functools import partial import torch from torch import nn from torch.nn import C...
null
184,687
from __future__ import absolute_import from __future__ import division from __future__ import print_function import os import copy import json import math import logging import tarfile import tempfile import shutil import numpy as np from functools import partial import torch from torch import nn from torch.nn import C...
Adapted from Mesh Tensorflow: https://github.com/tensorflow/mesh/blob/0cb87fe07da627bf0b7e60475d59f95ed6b5be3d/mesh_tensorflow/transformer/transformer_layers.py#L593
184,688
from __future__ import absolute_import from __future__ import division from __future__ import print_function import os import copy import json import math import logging import tarfile import tempfile import shutil import numpy as np from functools import partial import torch from torch import nn from torch.nn import C...
null
184,689
from __future__ import absolute_import from __future__ import division from __future__ import print_function import os import logging import glob import json import argparse import math import string from multiprocessing import Pool, cpu_count from tqdm import tqdm, trange from pathlib import Path import numpy as np im...
null
184,690
from __future__ import absolute_import from __future__ import division from __future__ import print_function import os import logging import glob import json import argparse import math import string from multiprocessing import Pool, cpu_count from tqdm import tqdm, trange from pathlib import Path import numpy as np im...
null
184,691
from __future__ import absolute_import from __future__ import division from __future__ import print_function import os import logging import glob import json import argparse import math import string from multiprocessing import Pool, cpu_count from tqdm import tqdm, trange from pathlib import Path import numpy as np im...
null
184,694
from __future__ import absolute_import from __future__ import division from __future__ import print_function import os import logging import glob import json import argparse import math import string from multiprocessing import Pool, cpu_count from tqdm import tqdm, trange from pathlib import Path import numpy as np im...
null
184,697
from __future__ import absolute_import from __future__ import division from __future__ import print_function import os import logging import glob import json import argparse import math import string from multiprocessing import Pool, cpu_count from tqdm import tqdm, trange from pathlib import Path import numpy as np im...
null
184,698
import pickle import math import argparse import glob import logging from pathlib import Path from tqdm import tqdm import unicodedata from transformers import BertTokenizer, RobertaTokenizer, XLMRobertaTokenizer from s2s_ft.tokenization_unilm import UnilmTokenizer from s2s_ft.tokenization_minilm import MinilmTokenizer...
null
184,699
import pickle import math import argparse import glob import logging from pathlib import Path from tqdm import tqdm import unicodedata from transformers import BertTokenizer, RobertaTokenizer, XLMRobertaTokenizer from s2s_ft.tokenization_unilm import UnilmTokenizer from s2s_ft.tokenization_minilm import MinilmTokenizer...
null
184,700
import pickle import math import argparse import glob import logging from pathlib import Path from tqdm import tqdm import unicodedata from transformers import BertTokenizer, RobertaTokenizer, XLMRobertaTokenizer from s2s_ft.tokenization_unilm import UnilmTokenizer from s2s_ft.tokenization_minilm import MinilmTokenizer...
null
184,701
import pickle import math import argparse import glob import logging from pathlib import Path from tqdm import tqdm import unicodedata from transformers import BertTokenizer, RobertaTokenizer, XLMRobertaTokenizer from s2s_ft.tokenization_unilm import UnilmTokenizer from s2s_ft.tokenization_minilm import MinilmTokenizer...
null
184,702
import math import numpy as np import logging import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import Parameter from torch import Tensor from typing import Any, Dict, List, Tuple, Callable, Optional def module_name_fordropout(module_name: str) -> str: if module_name == "TransformerE...
null
184,703
import math import numpy as np import logging import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import Parameter from torch import Tensor from typing import Any, Dict, List, Tuple, Callable, Optional The provided code snippet includes necessary dependencies for implementing the `utils_ma...
Replace non-padding symbols with their position numbers. Position numbers begin at padding_idx+1. Padding symbols are ignored.
184,704
import math import numpy as np import logging import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import Parameter from torch import Tensor from typing import Any, Dict, List, Tuple, Callable, Optional def utils_item(tensor): # tpu-comment: making this a no-op for xla devices. if t...
null
184,705
import math import numpy as np import logging import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import Parameter from torch import Tensor from typing import Any, Dict, List, Tuple, Callable, Optional The provided code snippet includes necessary dependencies for implementing the `fsdp_wra...
Helper to wrap layers/modules in FSDP. This falls back to a no-op if fairscale is not available. Args: module (nn.Module): module to (maybe) wrap min_num_params (int, Optional): minimum number of layer params to wrap
184,706
import math import numpy as np import logging import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import Parameter from torch import Tensor from typing import Any, Dict, List, Tuple, Callable, Optional The provided code snippet includes necessary dependencies for implementing the `quant_no...
Wraps modules and applies quantization noise to the weights for subsequent quantization with Iterative Product Quantization as described in "Training with Quantization Noise for Extreme Model Compression" Args: - module: nn.Module - p: amount of Quantization Noise - block_size: size of the blocks for subsequent quantiz...
184,707
import math import numpy as np import logging import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import Parameter from torch import Tensor from typing import Any, Dict, List, Tuple, Callable, Optional logger = logging.getLogger(__name__) def relu_squared(x: torch.Tensor): return F.relu...
Returns the activation function corresponding to `activation`
184,708
import math import numpy as np import logging import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import Parameter from torch import Tensor from typing import Any, Dict, List, Tuple, Callable, Optional def softmax(x, dim: int, onnx_trace: bool = False): if onnx_trace: return F....
null
184,709
import math import numpy as np import logging import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import Parameter from torch import Tensor from typing import Any, Dict, List, Tuple, Callable, Optional The provided code snippet includes necessary dependencies for implementing the `compute_...
Computes random mask spans for a given shape Args: shape: the the shape for which to compute masks. should be of size 2 where first element is batch size and 2nd is timesteps padding_mask: optional padding mask of the same size as shape, which will prevent masking padded elements mask_prob: probability for each token t...
184,710
import math import numpy as np import logging import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import Parameter from torch import Tensor from typing import Any, Dict, List, Tuple, Callable, Optional class MultiheadAttention(nn.Module): """Multi-headed attention. See "Attention Is...
Initialize the weights specific to the BERT Model. This overrides the default initializations depending on the specified arguments. 1. If normal_init_linear_weights is set then weights of linear layer will be initialized using the normal distribution and bais will be set to the specified value. 2. If normal_init_embed_...
184,711
import math import numpy as np import logging import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import Parameter from torch import Tensor from typing import Any, Dict, List, Tuple, Callable, Optional def pad_to_multiple(x, multiple, dim=-1, value=0): # Inspired from https://github.co...
null
184,712
import math import numpy as np import logging import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import Parameter from torch import Tensor from typing import Any, Dict, List, Tuple, Callable, Optional def is_xla_tensor(tensor): return torch.is_tensor(tensor) and tensor.device.type == "...
null
184,713
import math import numpy as np import logging import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import Parameter from torch import Tensor from typing import Any, Dict, List, Tuple, Callable, Optional class LearnedPositionalEmbedding(nn.Embedding): """ This module learns positional...
null
184,714
import math import numpy as np import logging import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import Parameter from torch import Tensor from typing import Any, Dict, List, Tuple, Callable, Optional try: from apex.normalization import FusedLayerNorm as _FusedLayerNorm has_fused_l...
null
184,715
import math from typing import Any, Dict, List, Optional import torch import torch.nn as nn from fairseq import utils from fairseq.distributed import fsdp_wrap from fairseq.models import FairseqIncrementalDecoder from fairseq.models.transformer import TransformerConfig from fairseq.modules import ( AdaptiveSoftmax,...
null
184,716
import math from typing import Any, Dict, List, Optional import torch import torch.nn as nn from fairseq import utils from fairseq.distributed import fsdp_wrap from fairseq.models import FairseqIncrementalDecoder from fairseq.models.transformer import TransformerConfig from fairseq.modules import ( AdaptiveSoftmax,...
null
184,717
import math from typing import Dict, List, Optional import torch import torch.nn as nn import torch.nn.functional as F from fairseq import utils from fairseq.distributed import fsdp_wrap from fairseq.models import FairseqEncoder from fairseq.modules import ( FairseqDropout, LayerDropModuleList, LayerNorm, ...
null
184,718
import math from typing import Dict, List, Optional import torch import torch.nn as nn import torch.nn.functional as F from fairseq import utils from fairseq.distributed import fsdp_wrap from fairseq.models import FairseqEncoder from fairseq.modules import ( FairseqDropout, LayerDropModuleList, LayerNorm, ...
null
184,719
import logging import numpy as np import torch import os import itertools from fairseq.data import FairseqDataset, data_utils from fairseq.data import ( AppendTokenDataset, ConcatDataset, PrependTokenDataset, data_utils, indexed_dataset, ) logger = logging.getLogger(__name__) class LanguageTripleDat...
null