code
stringlengths
82
53.2k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase = {"""configuration_fnet""": ["""FNET_PRETRAINED_CONFIG_A...
88
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) UpperCAmelCase = { """configuration_vision_encoder_decoder""": ["""VisionEncoderDeco...
88
1
"""simple docstring""" import math def lowerCAmelCase_ ( lowercase_ : int ): '''simple docstring''' __SCREAMING_SNAKE_CASE : List[str] = [] __SCREAMING_SNAKE_CASE : Tuple = 2 __SCREAMING_SNAKE_CASE : Optional[int] = int(ma...
709
"""simple docstring""" _lowerCamelCase = ''' # Installazione di Transformers ! pip install transformers datasets # Per installare dalla fonte invece dell\'ultima versione rilasciata, commenta il comando sopra e # rimuovi la modalità commento al comando seguente. # ! pip install git+https://github.co...
401
0
from __future__ import annotations A : Union[str, Any] = [-10, -5, 0, 5, 5.1, 11, 13, 21, 3, 4, -21, -10, -5, -1, 0] A : Union[str, Any] = [-5, 0, 5, 5.1, 11, 13, 21, -1, 4, -1, -10, -5, -1, 0, -1] def _lowerCAmelCase ( _lowerCAmelCase ) -> li...
371
import argparse import json from dataclasses import dataclass, field from functools import partial from pathlib import Path from typing import List import timm import torch import torch.nn as nn from huggingface_hub import hf_hub_download from torch import Tensor from transformers import AutoImageP...
371
1
'''simple docstring''' import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING fro...
700
'''simple docstring''' def A_ ( SCREAMING_SNAKE_CASE_ ) ->int: if not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): raise ValueError("""multiplicative_persistence() only accepts integral values""" ) if num < 0: raise ValueError("""multiplicative_persistence() does not accept n...
603
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.se...
175
'''simple docstring''' import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random...
330
0
from __future__ import annotations import numpy as np from numpy import floataa from numpy.typing import NDArray def lowercase_ ( _A : NDArray[floataa] , _A : NDArray[floataa] , _A : list[int] , _A : int , ): """simple docstring""...
5
def lowercase_ ( _A : int ): """simple docstring""" if not isinstance(_A , _A ): lowerCamelCase__ : List[str] = F"Input value of [number={number}] must be an integer" raise TypeError(_A ) if number < 0: retu...
5
1
import argparse import math import traceback import dateutil.parser as date_parser import requests def snake_case (__lowercase ) -> str: '''simple docstring''' _snake_case : List[Any] = {} _snake_case : Any = job['started_at'] _snake_case : List[str] ...
670
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ : Optional[int] = logging.get_logger(__name__) UpperCAmelCase_ : List[str] = { "google/vivit-b-16x2-kinetics400": ( "https://huggingface.co/google...
120
0
import math import os import unittest from transformers import MegatronBertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import Conf...
706
import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask snake_case = logging.getLogger(__name__) class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' def __i...
488
0
"""simple docstring""" import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) class snake_case_ ( lowerCamelCase_ ): """simple docstring""" def __init__( self...
34
"""simple docstring""" from typing import Callable, Dict, Optional, Tuple import torch from torch import nn from torch.distributions import ( AffineTransform, Distribution, Independent, NegativeBinomial, Normal, StudentT, TransformedDistribution, ) class snake_case_ ( ...
34
1
import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE : Tuple = logging.get_logger(__name__) SCREAMING_SNA...
719
# Function to print upper half of diamond (pyramid) def __A ( _A ): """simple docstring""" for i in range(0 , _A ): for _ in range(0 , n - i - 1 ): # printing spaces print(" " , end="" ) for _ in range(0 , i + 1 ): # printing ...
525
0
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.alt_diffusion.modeling_robert...
313
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging A : Optional[int] = logging.get_logger(__na...
128
0
"""simple docstring""" import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class __a (UpperCamelCase_): '''simple docstring''' def _a ( self , _a ) -> Union[str, Any]: """simple docstring""" ...
12
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) a :Any = { "configuration_roberta_prelayernorm": [ "ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_A...
12
1
'''simple docstring''' from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_avail...
421
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) snake_case_ = { ...
421
1
'''simple docstring''' import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_tor...
711
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaInpaintPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import flo...
465
0
def SCREAMING_SNAKE_CASE ( __UpperCamelCase , __UpperCamelCase) -> int: return x if y == 0 else greatest_common_divisor(__UpperCamelCase , x % y) def SCREAMING_SNAKE_CASE ( __UpperCamelCase , __UpperCamelCase) -> int: return (x * y) // greatest_common_divisor(_...
515
import fire from utils import calculate_rouge, save_json def SCREAMING_SNAKE_CASE ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None , **__UpperCamelCase) -> Any: a = [x.strip() for x in open(__UpperCamelCase).readlines()] a = [x.strip() fo...
515
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = {'vocab_file': ...
711
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase_ = {'configuration_ibert': ['IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'IBertConfig', 'IBertOnnxConfig']} try: if not is_torch_available(): raise OptionalDepen...
596
0
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipe...
138
'''simple docstring''' import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer...
138
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowerCAmelCase_ : Optional[int] = { 'configuration_roc_bert': ['ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RoCBertConf...
521
'''simple docstring''' import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six #...
521
1
'''simple docstring''' def A ( UpperCamelCase_ : int ) -> str: '''simple docstring''' lowerCAmelCase__ = int(UpperCamelCase_ ) if decimal in (0, 1): # Exit cases for the recursion return str(UpperCamelCase_ ) lowerCAmelCase__ ,lowerCAmelCase__...
48
'''simple docstring''' from __future__ import annotations a__ : Optional[int] = list[tuple[int, int]] a__ : List[Any] = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], ...
368
0
import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import numpy as np import pandas as pd from datasets import load_dataset import transformers from transformers import ( AutoConfig, BartForSequenceClassification, DataColl...
709
from typing import Any import numpy as np def _snake_case ( lowerCAmelCase : np.ndarray ): """simple docstring""" return np.array_equal(lowerCAmelCase , matrix.conjugate().T ) def _snake_case ( lowerCAmelCase : np.ndarray , lowerCAmelCase ...
316
0
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_...
372
'''simple docstring''' from manim import * class _UpperCamelCase ( SCREAMING_SNAKE_CASE): '''simple docstring''' def a__ ( self ) -> List[str]: lowercase : List[Any] = Rectangle(height=0.5 , width=0.5 ) lowercase : str = ...
372
1
"""simple docstring""" from copy import deepcopy class lowerCAmelCase_ : '''simple docstring''' def __init__( self , snake_case_ = None , snake_case_ = None ) -> None: if arr is None and size is not None: _...
573
"""simple docstring""" import gc import unittest from transformers import CTRLConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_mod...
573
1
from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import KarrasVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class lowercase_ ( A ): __lowerCamelCase...
443
"""simple docstring""" import argparse import json import os from collections import OrderedDict import numpy as np import tensorflow as tf import torch def __lowerCamelCase ( a_ : Dict ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE :Optional[int] = os.p...
498
0
"""simple docstring""" import argparse import os from accelerate.utils import ComputeEnvironment from .cluster import get_cluster_input from .config_args import cache_dir, default_config_file, default_yaml_config_file, load_config_from_file # noqa: F401 from .config_utils import _ask_fi...
718
"""simple docstring""" import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor _UpperCamelCase = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE_ ( snake_case__ ): """simple docstring""" def _...
363
0
from __future__ import annotations def lowerCamelCase__ (__lowerCamelCase ): _SCREAMING_SNAKE_CASE : List[str] = len(__lowerCamelCase ) # We need to create solution object to save path. _SCREAMING_SNAKE_CASE : Any = [[0 for _ in ...
249
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase__ =logging.get_logger(__name__) UpperCamelCase__ ={ 'xlm-mlm-en-2048': 'https://huggingface.co/xlm-mlm-en-20...
249
1
'''simple docstring''' def snake_case__ ( _A: list ) -> list: '''simple docstring''' def merge(_A: list , _A: list ) -> list: def _merge(): while left and right: yield (left if left[0] <= right[0] else right).pop(0 ) yield from left...
605
'''simple docstring''' from collections.abc import Generator from math import sin def snake_case__ ( _A: bytes ) -> bytes: '''simple docstring''' if len(_A ) != 32: raise ValueError("""Input must be of length 32""" ) lowerCAmelCase = b"""""" f...
605
1
"""simple docstring""" import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import Mode...
636
"""simple docstring""" import importlib import os import sys # This is required to make the module import works (when the python process is running from the root of the repo) sys.path.append('''.''') def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_C...
682
0
'''simple docstring''' import copy import os from typing import TYPE_CHECKING, List, Union if TYPE_CHECKING: pass from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ =logging.get_logger(__name__) lowercase__ ={ 'kakaobrain/align-base': ...
720
'''simple docstring''' from __future__ import annotations def UpperCamelCase_ ( A__ , A__ , A__ , A__ ): if (direction == 1 and array[indexa] > array[indexa]) or ( direction == 0 and array[indexa] < array[indexa] ): a_ , a_ = array[indexa], array[indexa] ...
511
0
'''simple docstring''' from torch import nn def __lowerCAmelCase ( a_ ) -> Optional[Any]: '''simple docstring''' if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn ==...
251
'''simple docstring''' _lowerCAmelCase :Union[str, Any] = { """meter""": """m""", """kilometer""": """km""", """megametre""": """Mm""", """gigametre""": """Gm""", """terametre""": """Tm""", """petametre""": """Pm""", """exametre""": """Em""", """zettametre""":...
251
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase: Tuple =logging.get_logger(__name__) _UpperCamelCase: Any ={ 'transfo-xl-wt103': 'https://huggingface.co/transfo-xl-wt103/resolve/main/config.json', } class __lowercase( SCREAMING_SNAKE_CASE ...
585
import argparse import csv import logging import os import random import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset from tqdm import tqdm, trange from transformers import ( CONFIG_NAME, WEIGHTS_NAME, AdamW, OpenAIGPTDoubleHeadsMod...
585
1
from unittest.mock import Mock, patch from file_transfer.send_file import send_file @patch('''socket.socket''' ) @patch('''builtins.open''' ) def lowerCamelCase_ ( UpperCamelCase_ , UpperCamelCase_ ): # ===== initialization ===== _a : List[str] = Mock() ...
471
def lowerCamelCase_ ( UpperCamelCase_ , UpperCamelCase_ ): if a < 0 or b < 0: raise ValueError('''the value of both inputs must be positive''' ) _a : str = str(bin(UpperCamelCase_ ) )[2:] # remove the leading "0b" _a : Dict = ...
471
1
import unittest from parameterized import parameterized from transformers import AutoTokenizer, GPTNeoXConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import C...
230
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device lowercase_ = False class A_ ( unittest.TestCase ): '''simple docst...
230
1
"""simple docstring""" from typing import Dict, Optional import numpy as np import datasets __snake_case : Optional[Any] = '\nIoU is the area of overlap between the predicted segmentation and the ground truth divided by the area of union\nbetween the predicted segmentation and ...
293
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case : Tuple = logging.get_logger(__name__) __snake_case : Any = { 'facebook/xglm-564M': 'https://huggingface.co/facebook/xglm-564M/resolve/main/config.j...
293
1
import unittest from transformers import is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from tensorflow.python.eager import context from tensorflow.python.framework import ops from transformers import GradientAccumulator, ...
710
"""simple docstring""" import os import sys import unittest __A : Optional[Any] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, 'utils')) import check_dummies # noqa: E402 from check_dummies import create_dummy_files,...
281
0
def _snake_case (__lowercase , __lowercase , __lowercase , __lowercase): UpperCamelCase_ , UpperCamelCase_ = len(__lowercase), len(grid[0]) if ( min(__lowercase , __lowercase) < 0 or row == row_length or col == col_length or...
23
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, StableDiffusionSAGPipeline, UNetaDConditionModel, ) from diffusers.utils import slow, torch_d...
23
1
import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaV...
516
from __future__ import annotations from collections.abc import Callable from typing import Any, Generic, TypeVar _lowerCamelCase : Optional[int] = TypeVar('''T''') class lowerCAmelCase__ ( Generic[T] ): '''simple docstring''' def __init__( self , ...
516
1
from scipy.stats import spearmanr import datasets __lowerCAmelCase = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations i...
147
from maths.prime_factors import prime_factors def lowercase_ ( _UpperCamelCase ): '''simple docstring''' if not isinstance(_UpperCamelCase , _UpperCamelCase ): __lowercase = F'Input value of [number={number}] must be an integer' raise TypeError(_UpperCamelCase ) ...
639
0
'''simple docstring''' import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() __A = logging.get_logger(__name__)...
705
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __A = { '''configuration_biogpt''': ['''BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BioGptConfig'''], '...
61
0
"""simple docstring""" import re from filelock import FileLock try: import nltk A: Optional[int] = True except (ImportError, ModuleNotFoundError): A: List[Any] = False if NLTK_AVAILABLE: with FileLock(".lock") as lock: nltk.download("punkt", qui...
160
"""simple docstring""" import argparse import torch from transformers import FunnelBaseModel, FunnelConfig, FunnelModel, load_tf_weights_in_funnel from transformers.utils import logging logging.set_verbosity_info() def _snake_case ( UpperCamelCase : List[str] , UpperCamelCase : Any ...
160
1
'''simple docstring''' import os import tempfile import unittest from transformers import NezhaConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...generation.test_utils import Generatio...
119
'''simple docstring''' from math import ceil def UpperCAmelCase_ ( __lowercase : Any , __lowercase : int ) -> Any: '''simple docstring''' _UpperCAmelCase = list(range(0 , __lowercase ) ) _UpperCAmelCase = [item for sublist in l...
119
1
'''simple docstring''' import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py _a : Dict = '.' # Internal Tensor...
447
'''simple docstring''' import argparse from collections import defaultdict import yaml a : List[Any] = 'docs/source/en/_toctree.yml' def __magic_name__ ( __UpperCAmelCase ) -> str: '''simple docstring''' snake_case_ = defaultdict(__Upp...
640
0
import pytest import datasets.config from datasets.utils.info_utils import is_small_dataset @pytest.mark.parametrize('''dataset_size''' , [None, 4_00 * 2**20, 6_00 * 2**20] ) @pytest.mark.parametrize('''input_in_memory_max_size''' , ['''default''', 0, 1_00 * 2**20, 9_00 * 2**20] ) def _...
429
import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging,...
429
1
from collections import defaultdict from math import ceil, sqrt def _lowercase( __a : int = 100_0000 , __a : int = 10 ): a__ =defaultdict(__a ) for outer_width in range(3 , (t_limit // 4) + 2 ): if outer_width * outer_width > t_l...
20
from __future__ import annotations from collections import namedtuple def __lowerCamelCase ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> tuple: _UpperCAmelCase = namedtuple("result" , "name value" ) if (voltage, current, power).count(0 ) != 1: ...
684
0
from argparse import ArgumentParser from .env import EnvironmentCommand def lowerCAmelCase__ ( ): """simple docstring""" __a = ArgumentParser("""Diffusers CLI tool""" , usage="""diffusers-cli <command> [<args>]""" ) __a = parser.add_subparsers(help="""diffusers-...
717
def lowerCAmelCase__ ( _SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : Tuple=False ): """simple docstring""" if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and isinstance(_SCREAMING_...
547
0
import warnings from ...utils import logging from .image_processing_videomae import VideoMAEImageProcessor __lowercase : Optional[Any] =logging.get_logger(__name__) class A ( __lowercase ): def __init__( self: int , *_lowerCAmelCase: Optional[Any] ...
54
def a__ ( lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' if len(lowercase__ ) != len(lowercase__ ): raise ValueError("The length of profit and weight must be same." ) if max_weight <= 0: raise ValueError("max_weight mu...
54
1
'''simple docstring''' import unittest from pathlib import Path from shutil import copyfile from transformers import SPIECE_UNDERLINE, is_sentencepiece_available from transformers.models.speech_to_text import SpeechaTextTokenizer from transformers.models.speech_to_text.tokenization_speech_to_text import VOCAB...
719
'''simple docstring''' from __future__ import annotations import numpy as np from numpy import floataa from numpy.typing import NDArray def __snake_case ( lowercase : NDArray[floataa] , lowercase : NDArray[floataa] , lowercase : list[int] , lowercase : ...
420
0
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_tf_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_tf_available(): import tensorflow as tf __A : Dict = logging.get_logger(__nam...
16
import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.p...
243
0
from __future__ import annotations def SCREAMING_SNAKE_CASE_ ( __A : list[int | float] , __A : int , __A : int ) -> int | float: """simple docstring""" if len(__A ) == 0: raise ValueError('find_max() arg is an empty sequence'...
720
import random class SCREAMING_SNAKE_CASE__ : @staticmethod def SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE__ : str ) -> tuple[list[int], list[int]]: a_ : int = [ord(SCREAMING_SNAKE_CASE__ ) for i in text] a_ : Any = ...
443
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = { """sh...
92
import argparse import math import os from copy import deepcopy import torch from audio_diffusion.models import DiffusionAttnUnetaD from diffusion import sampling from torch import nn from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel lowerCAmelCase_ = { '''...
39
0
def __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> int: """simple docstring""" if len(_lowerCAmelCase ) != len(_lowerCAmelCase ): raise ValueError("""The length of profit and weight must be same.""" ) if max_weight <= 0: raise Va...
704
import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils impo...
520
0
"""simple docstring""" def lowerCamelCase__ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ )-> List[Any]: """simple docstring""" global f # a global dp table for knapsack if f[i][j] < 0: if j < wt[i - 1]:...
554
"""simple docstring""" import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class SCREAMIN...
180
0
'''simple docstring''' from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
708
'''simple docstring''' import argparse import os import re import packaging.version __lowerCAmelCase : Optional[int] = "examples/" __lowerCAmelCase : Dict = { "examples": (re.compile(r"^check_min_version\(\"[^\"]+\"\)\s*$", re.MULTILINE), "check_min_version(\"VER...
654
0
"""simple docstring""" import unittest import numpy as np from transformers import BertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): ...
265
'''simple docstring''' from collections.abc import Iterator, MutableMapping from dataclasses import dataclass from typing import Generic, TypeVar lowerCAmelCase__ = TypeVar('''KEY''') lowerCAmelCase__ = TypeVar('''VAL''') @dataclass(frozen=lowerCamelCase__ , slots=lowerCamelCase...
41
0
'''simple docstring''' from __future__ import annotations def __snake_case (__UpperCAmelCase , __UpperCAmelCase ): """simple docstring""" if nth_term == "": return [""] lowerCamelCase_ : List[str] = int(__UpperCAmelCase ) lowerCamelCase_ : List[str] ...
708
'''simple docstring''' def __snake_case (__UpperCAmelCase = 3 , __UpperCAmelCase = 7 , __UpperCAmelCase = 1000000 ): """simple docstring""" lowerCamelCase_ : Any = 0 lowerCamelCase_ : Tuple = 1 for current_denominator in range(1 , limit + ...
418
0
'''simple docstring''' from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging __SCR...
244
'''simple docstring''' from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin...
244
1
"""simple docstring""" import gc import unittest import numpy as np import torch from diffusers import StableDiffusionKDiffusionPipeline from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() @slow @requir...
132
"""simple docstring""" import sys def lowerCamelCase_ ( __lowerCAmelCase ) -> List[Any]: '''simple docstring''' lowerCamelCase__ =len(__lowerCAmelCase ) lowerCamelCase__ =[[0 for x in range(__lowerCAmelCase )] for x in range(__lowerCAmel...
132
1
import argparse import collections import json from pathlib import Path import requests import torch import yaml from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTImageProcessor, MobileViTVaConfig, MobileViTVaForImageClassification, MobileViTVaForS...
300
from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class a ( UpperCAmelCase ): def _UpperCAmelCase ( self , A_ ): '''simple docstring''' return 0.0 ...
300
1
import argparse import json from typing import List from ltp import LTP from transformers import BertTokenizer def __lowerCamelCase ( _lowercase ) -> int: # This defines a "chinese character" as anything in the CJK Unicode block: # https://en.wikipedia.org/wiki/CJK_Unified_Ideog...
170
def __lowerCamelCase ( _lowercase ) -> int: assert ( isinstance(_lowercase , _lowercase ) and number_of_steps > 0 ), F'number_of_steps needs to be positive integer, your input {number_of_steps}' if number_of_steps == 1: return 1 UpperCamelCase , Uppe...
170
1
"""simple docstring""" import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import re...
4
'''simple docstring''' # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip ins...
172
0
"""simple docstring""" import unittest import torch from diffusers import VQModel from diffusers.utils import floats_tensor, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() ...
709
"""simple docstring""" import random import timeit from functools import wraps from typing import Callable, Optional from ..configuration_utils import PretrainedConfig from ..models.auto.modeling_tf_auto import TF_MODEL_MAPPING, TF_MODEL_WITH_LM_HEAD_MAPPING from ..utils import is_pyanvml_available, is_tf_...
529
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, is_vision_available, ) __A : Optional[int] = {'''processing_layoutxlm''': ...
499
"""simple docstring""" import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision ...
499
1
'''simple docstring''' import os def UpperCAmelCase_ ( ): """simple docstring""" with open(os.path.dirname(lowerCamelCase_ ) + "/grid.txt" ) as f: lowerCAmelCase__ : Optional[Any] = [] # noqa: E741 for _ in range(2_0 ): l.append([int(lowerCamelCase_ ) for x in f.readline(...
568
'''simple docstring''' import logging import re import pytorch_quantization import pytorch_quantization.nn as quant_nn import torch from pytorch_quantization import calib from pytorch_quantization.tensor_quant import QuantDescriptor snake_case = logging.getLogger(__name__) snake_case = ...
568
1
'''simple docstring''' import argparse import tensorflow as tf import torch from transformers import BertConfig, BertForMaskedLM from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertPooler, BertSelfAttention, BertSelfOutput, ) from transformer...
210
'''simple docstring''' from json import JSONDecodeError # Workaround for requests.exceptions.JSONDecodeError import requests def lowerCamelCase__ ( A : str = "isbn/0140328726" ): '''simple docstring''' UpperCAmelCase = olid.strip().strip('''/''' ) # Remove leading/tr...
210
1
'''simple docstring''' lowerCamelCase = 0 # The first color of the flag. lowerCamelCase = 1 # The second color of the flag. lowerCamelCase = 2 # The third color of the flag. lowerCamelCase = (red, white, blue) def a ( lowerCamelCase_...
710
'''simple docstring''' import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging lowerCamelCase :Union[str, Any] = logging.get_logger(__name__) class _lowerCA...
686
0
'''simple docstring''' import heapq as hq import math from collections.abc import Iterator class snake_case__ : def __init__( self : Optional[int] , __a : List[str] ) -> Any: '''simple docstring''' __snake_case : Dict = st...
286
from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, ...
202
0
import argparse import os import torch from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) __UpperCamelCase : List[Any] = { 'sample_size': 32, 'in_channels': 3, 'out_channels': 3, 'layers_per_block': 2...
641
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class lowerCAmelCase__( snake_case__ ...
641
1
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import TransformeraDModel, VQDiffusionPipeline, VQDiffusionScheduler, VQModel from diffusers.pipelines.vq_diffusion.pipeline_vq_diffusion im...
38
import argparse import json from pathlib import Path import torch import torchaudio from datasets import load_dataset from huggingface_hub import hf_hub_download from transformers import ASTConfig, ASTFeatureExtractor, ASTForAudioClassification from transformers.utils import logging logging.set_verbosity_info() __...
472
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available A : Union[str, Any] = { 'configuration_roc_bert': ['ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RoCBertConfig'], 'tokenization_roc_bert': ['R...
700
def _lowerCAmelCase ( _lowerCAmelCase = "The quick brown fox jumps over the lazy dog" , ) -> bool: '''simple docstring''' __snake_case = set() # Replace all the whitespace in our sentence __snake_case = input_str.replace(" " , "" ) ...
473
0
def __snake_case ( _UpperCamelCase ) -> Union[str, Any]: # noqa: E741 _a = len(lowercase_ ) _a = 0 _a = [0] * n _a = [False] * n _a = [False] * n def dfs(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): if parent ...
487
"""simple docstring""" import json import os import shutil import tempfile import unittest import numpy as np from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES, BertTokenizer from transformers.testing_utils import require_tokenizers, require...
674
0
import collections import os from typing import List, Optional, Tuple from transformers.utils import is_jieba_available, requires_backends if is_jieba_available(): import jieba from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __lowerCAmelCase : Any = ...
76
from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCAmelCase : Tuple = logging.get_logger(__name__) __lowerCAmelCase : Union[str, Any] = { 'microsoft/cvt-13': 'https://huggingface.co/microsoft/cvt-13/resolve/main/config.json', # See all...
76
1
import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_availabl...
67
'''simple docstring''' import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class snake_case__ ( SCREAMING_SNAKE_CASE_ ): A__ = ['''image_processor''', '''tokenizer'''] A__ = '''CLIPImageProcessor''' A__ = ('''CLIPTokeni...
286
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase : Optional[int] ={"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): ...
702
import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) fr...
504
0
# Function to print upper half of diamond (pyramid) def SCREAMING_SNAKE_CASE__ ( lowerCAmelCase_ : Tuple ) -> Optional[Any]: """simple docstring""" for i in range(0 ,lowerCAmelCase_ ): for _ in range(0 ,n - i - 1 ): # printing spaces...
220
from __future__ import annotations class lowerCAmelCase_ : '''simple docstring''' def __init__( self , __UpperCAmelCase ): SCREAMING_SNAKE_CASE_ : Any =order # a_{0} ... a_{k} SCREAMING_SNAKE_CASE_ : List[str] =[1.0] ...
220
1
def UpperCamelCase_ ( a_ , a_ ) ->int: A =1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): A =n - k # Calculate C(n,k) for i in range(a_ ): result *= n - i result //= i + 1 return result def UpperCamelCase_ ( a_ ) ->int: r...
689
import json import os from typing import Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __a = logging.get_logger(__name__) __a = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", } __a ...
689
1
"""simple docstring""" import copy from typing import Dict, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING from ..detr import DetrConfig from ..swin import SwinConfig A_ : List[str] ={ """facebook/maskformer-swin-base-a...
650
"""simple docstring""" import argparse import torch from transformers import ( WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForAudioFrameClassification, WavaVecaForSequenceClassification, WavaVecaForXVector, logging, ) logging.set_verbosity_info() A_ : int =logging.get_...
650
1
import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester fro...
718
def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> bool: SCREAMING_SNAKE_CASE_ : Union[str, Any] = len(SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : Optional[Any] = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each...
311
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case__ = { "configuration_clipseg": [ "CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP", "CLIPSegConfig", "CLIPSegTextConfig", "CLIPSegVisionConfig", ...
395
import sacrebleu as scb from packaging import version from sacrebleu import CHRF import datasets SCREAMING_SNAKE_CASE : Tuple = "\\n@inproceedings{popovic-2015-chrf,\n title = \"chr{F}: character n-gram {F}-score for automatic {MT} evaluation\",\n author = \"Popovi{\'c}, Maja\",\n bookti...
635
0
'''simple docstring''' import torch from transformers import AutoModel class _snake_case ( torch.nn.Module ): def __init__( self , _SCREAMING_SNAKE_CASE="sayef/fsner-bert-base-uncased" ): '''simple docstring''' super(_SCREAMING_SNAKE_CASE , self ).__init__() ...
514
'''simple docstring''' import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, to...
514
1
import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration lowercase_ = 5_00_00 lowercase_ = 50_00 lowercase_ , lowercase_ = os.path.split(__file__) lowercase_ = os.path.join(RESULTS_BASEPATH, 'results', RESULTS_FILENAME.r...
562
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = {'vocab_file': 'sentencepi...
562
1
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamem...
706
import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature fro...
167
0
'''simple docstring''' import itertools import os from collections import Counter, defaultdict from concurrent.futures import ThreadPoolExecutor, as_completed import numpy as np import datasets from .execute import check_correctness UpperCAmelCase_ : Any = '''\ @misc{chen2021evaluating, t...
24
"""simple docstring""" import flax.linen as nn import jax.numpy as jnp from .attention_flax import FlaxTransformeraDModel from .resnet_flax import FlaxDownsampleaD, FlaxResnetBlockaD, FlaxUpsampleaD class lowercase( nn.Module ): '''simple docstring''' ...
609
0
from __future__ import annotations _snake_case = [] def lowerCamelCase_ ( A : list[list[int]] , A : int , A : int ): """simple docstring""" for i in range(len(A ) ): if board[row][i] == 1: return False for i in range(l...
718
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { "sail/pool...
413
0
'''simple docstring''' import argparse import collections import os import re import tempfile import pandas as pd from datasets import Dataset from huggingface_hub import hf_hub_download, upload_folder from transformers.utils import direct_transformers_import # All paths are set with the intent you should ru...
98
'''simple docstring''' from __future__ import annotations import queue class __lowerCAmelCase : """simple docstring""" def __init__( self : str , lowerCAmelCase__ : Optional[int] ) -> str: '''simple docstring''' _UpperCamelC...
98
1
import math lowerCAmelCase : List[str] =10 lowerCAmelCase : List[Any] =7 lowerCAmelCase : Dict =BALLS_PER_COLOUR * NUM_COLOURS def A__ ( __A = 20 ): '''simple docstring''' _lowerCamelCase : Optional[Any] = math.co...
15
from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class __snake_case ( __lowerCAmelCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ...
15
1
'''simple docstring''' import numpy as np import torch import tqdm from ...models.unet_ad import UNetaDModel from ...pipelines import DiffusionPipeline from ...utils import randn_tensor from ...utils.dummy_pt_objects import DDPMScheduler class SCREAMING_SNAKE_CASE__ ( a__ ): def __init__( ...
685
import argparse import os from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_task_guides.py __UpperCAmelCase = '''src/transformers''' __UpperCAmelCase = '''docs/source...
40
0
import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here t...
337
def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> list[int]: if length <= 0 or not isinstance(lowerCamelCase__ , lowerCamelCase__ ): raise ValueError('Length must be a positive integer.' ) return [n * (2 * n - 1) for n in range(lowerCamelCase__ )] if __name__ == "__ma...
337
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging A_ = logging.get_logger(__name__) A_ = {"openai-gpt": "https://huggingface.co/openai-gpt/resolve/main/config.json"} class UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' S...
42
'''simple docstring''' import numpy as np from cva import COLOR_BGR2GRAY, CV_8UC3, cvtColor, filteraD, imread, imshow, waitKey def _UpperCamelCase ( __UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> np.ndarray: #...
42
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) A : int = { 'configuration_swiftformer': [ 'SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SwiftFormerConfig', ...
719
'''simple docstring''' from ..utils import DummyObject, requires_backends class lowerCamelCase ( metaclass=__UpperCAmelCase ): _SCREAMING_SNAKE_CASE = ["torch", "scipy"] def __init__( self : Tuple , *__snake_case : List[Any] , **__snake_c...
273
0
"""simple docstring""" import argparse from transformers import BigBirdConfig, BigBirdForPreTraining, BigBirdForQuestionAnswering, load_tf_weights_in_big_bird from transformers.utils import logging logging.set_verbosity_info() def lowercase ( lowerCAmelCase__ : str , lo...
695
import os def SCREAMING_SNAKE_CASE ( ) -> Union[str, Any]: __UpperCAmelCase =os.path.dirname(os.path.realpath(snake_case__ ) ) __UpperCAmelCase =os.path.join(snake_case__ , '''triangle.txt''' ) with open(snake_case__ ) as f: __UpperCAmelCase =f....
132
0
'''simple docstring''' def UpperCamelCase_ ( __SCREAMING_SNAKE_CASE ): """simple docstring""" snake_case_ : List[str] = current_set.copy() for row_index, row in enumerate(__SCREAMING_SNAKE_CASE ): snake_case_ : int = row[0] for colum...
92
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a_ = {"configuration_wavlm": ["WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "WavLMConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailab...
92
1
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import ...
663
import io import json import unittest from parameterized import parameterized from transformers import FSMTForConditionalGeneration, FSMTTokenizer from transformers.testing_utils import get_tests_dir, require_torch, slow, torch_device from utils import calculate_bleu SCREAMING_SNAKE_CASE__ : Any ...
85
0
from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, Fla...
303
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() lowerC...
303
1
'''simple docstring''' import math import os import unittest from transformers import MegatronBertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configur...
18
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available A : Union[str, Any] = { """configuration_m2m_100""": ["""M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP""", """M2M100Config""...
349
0
'''simple docstring''' def _UpperCamelCase ( __A ) -> Optional[int]: '''simple docstring''' if not head: return True # split the list to two parts UpperCamelCase__ , UpperCamelCase__ = head.next, head while fast and fast.next: ...
714
'''simple docstring''' import argparse from tax import checkpoints from transformers import AutoConfig, FlaxAutoModelForSeqaSeqLM def _UpperCamelCase ( __A , __A , __A ) -> Tuple: '''simple docstring''' UpperCamelCase__ = AutoConfig.from_pretrain...
223
0
'''simple docstring''' import argparse import json import torch from diffusers import DDPMScheduler, LDMPipeline, UNetaDModel, VQModel def lowerCAmelCase_ ( snake_case_ : Dict , snake_case_ : str=1 ) -> List[str]: '''simple docstring''' if n_shave_...
78
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowerCAmelCase_ = { """configuration_nezha""": ["""NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NezhaConfig"""], } try: if not is_torch_available():...
678
0
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE...
386
"""simple docstring""" import argparse from tax import checkpoints from transformers import AutoConfig, FlaxAutoModelForSeqaSeqLM def lowerCamelCase_( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Union[str, Any]: '''simple docstring''' _lowerCamelCase : ...
386
1
import argparse import json import os import torch from torch import nn from transformers import NllbMoeConfig, NllbMoeModel from transformers.modeling_utils import dtype_byte_size from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME def _a ( SCREAMING_SNAKE_CASE ): """simpl...
43
lowerCAmelCase = { 'a': 'AAAAA', 'b': 'AAAAB', 'c': 'AAABA', 'd': 'AAABB', 'e': 'AABAA', 'f': 'AABAB', 'g': 'AABBA', 'h': 'AABBB', 'i': 'ABAAA', 'j': 'BBBAA', 'k': 'ABAAB', 'l': 'ABABA', 'm': 'ABABB', 'n': 'ABBAA', 'o': 'ABBAB', ...
43
1
'''simple docstring''' import argparse import ast import logging import os import sys import pandas as pd import torch from tqdm import tqdm from transformers import BartForConditionalGeneration, RagRetriever, RagSequenceForGeneration, RagTokenForGeneration from transformers import logging as transformers_logging ...
703
'''simple docstring''' import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( HubertConfig, HubertForCTC, HubertModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) ...
330
0
"""simple docstring""" # XXX: we want transformers master here - in the absense of conftest manipulating sys.path: # hack it in for now: import sys from pathlib import Path A : Optional[Any] = Path(__file__).resolve().parents[3] / 'src' sys.path.insert(1, str(git_repo_path)) impo...
516
"""simple docstring""" import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SegformerConfig, SegformerForImageClassification, ...
516
1
"""simple docstring""" import re def _UpperCamelCase ( UpperCamelCase ) -> list: """simple docstring""" return [char.split() for char in re.split(R"[^ a-z A-Z 0-9 \s]" , str_ )] def _UpperCamelCase ( UpperCamelCase ) -> str: """simple docst...
711
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Tuple: """simple docstring""" global f # a global dp table for knapsack if f[i][j] < 0: if j < wt[i - 1]: __UpperCAmelCa...
487
0