code
stringlengths
82
53.2k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
from __future__ import annotations def A__ ( lowerCamelCase ) -> int: # preprocessing the first row for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 , len(lowerCamelCase )...
548
lowerCamelCase_ : Tuple = { """meter""": """m""", """kilometer""": """km""", """megametre""": """Mm""", """gigametre""": """Gm""", """terametre""": """Tm""", """petametre""": """Pm""", """exametre""": """Em""", """zettametre""": """Zm""", """yottametre""": """Ym""", } # Expo...
548
1
"""simple docstring""" import unittest import numpy as np import torch from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad class A_(unittest.TestCase ): """simple docstring""" def _lowerCAmelCase ( self ): _low...
349
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cach...
349
1
"""simple docstring""" import baseaa def _snake_case ( lowercase__ ): return baseaa.baaencode(string.encode('utf-8' ) ) def _snake_case ( lowercase__ ): return baseaa.baadecode(a_ ).decode('utf-8' ) i...
630
'''simple docstring''' import numpy as np # Importing the Keras libraries and packages import tensorflow as tf from tensorflow.keras import layers, models if __name__ == "__main__": # Initialising the CNN # (Sequential- Building the model layer by layer) UpperCamelCase =models.Sequ...
208
0
from math import sqrt def UpperCAmelCase ( lowerCAmelCase__ ): '''simple docstring''' assert isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) and ( number >= 0 ), "'number' must been an int and positive" __A = True # 0 and 1 are none prime...
205
from __future__ import annotations import collections import pprint from pathlib import Path def UpperCAmelCase ( lowerCAmelCase__ ): '''simple docstring''' return "".join(sorted(lowerCAmelCase__ ) ) def UpperCAmelCase ( lowerCAmelCase__ ): '''simple d...
205
1
'''simple docstring''' import math import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from .attention_processor import Attention from .embeddings import get_timestep_embedding from .modeling_utils import ModelMixin class __lowercase ( _lowercase , _low...
422
'''simple docstring''' import importlib import torch import yaml from omegaconf import OmegaConf from taming.models.vqgan import VQModel def lowercase_ ( _lowercase , _lowercase=False ) -> Dict: '''simple docstring''' lowerCamelCase_ : Tuple = OmegaConf.load(_lowerca...
422
1
from __future__ import annotations from typing import Any class lowercase_ : def __init__( self: List[str], _lowercase: Optional[Any] = 6): '''simple docstring''' __lowerCAmelCase = None __lowerCAmelCase = None self.create_l...
709
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available __A : Optional[Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Tuple ...
334
0
import unittest from parameterized import parameterized from transformers import AutoTokenizer, GPTNeoXConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common i...
583
from __future__ import annotations def UpperCamelCase__ ( _A: float , _A: float , _A: float ): '''simple docstring''' if days_between_payments <= 0: raise ValueError("""days_between_payments must be > 0""" ) if daily_interest_rate...
479
0
import argparse import os from accelerate.test_utils import execute_subprocess_async def __snake_case ( lowercase : Optional[int]=None ): if subparsers is not None: snake_case_ = subparsers.add_parser("test" ) else: snake_case_ = argparse.Argumen...
708
'''simple docstring''' import argparse import tensorflow as tf import torch from transformers import BertConfig, BertForMaskedLM from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertPooler, BertSelfAttention, BertSelfOutput, ) from transfor...
420
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase : List[Any] = { 'configuration_pegasus_x': ['PEGASUS_X_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PegasusXConfig'], } try: if not is...
50
"""simple docstring""" import json import os import sys import tempfile import unittest from pathlib import Path from shutil import copyfile from huggingface_hub import HfFolder, Repository, create_repo, delete_repo from requests.exceptions import HTTPError import transformers from transformers import ( ...
177
0
'''simple docstring''' from __future__ import annotations import csv import requests from bsa import BeautifulSoup def __magic_name__( lowerCamelCase = ""): __lowerCAmelCase = url or '''https://www.imdb.com/chart/top/?ref_=nv_mv_250''' __lowerCAmelCase ...
717
'''simple docstring''' import json import sys def __magic_name__( lowerCamelCase, lowerCamelCase): with open(lowerCamelCase, encoding='''utf-8''') as f: __lowerCAmelCase = json.load(lowerCamelCase) __lowerCAmelCase = ['''<detail...
474
0
'''simple docstring''' import inspect import unittest from transformers import ConvNextVaConfig from transformers.models.auto import get_values from transformers.models.auto.modeling_auto import MODEL_FOR_BACKBONE_MAPPING_NAMES, MODEL_MAPPING_NAMES from transformers.testing_utils import require_torch, require_v...
92
import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() __lowerCamelCase : int = logging.get_logger(__name__) __lowerCamelCase : List[str...
323
0
'''simple docstring''' import argparse import json import os from collections import OrderedDict import numpy as np import tensorflow as tf import torch def __UpperCAmelCase ( a_): snake_case_ = os.path.join(args.tf_model_dir , 'parameters.json') snake_case_ ...
703
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, Enc...
607
0
'''simple docstring''' from __future__ import annotations from collections import deque class SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : Dict , snake_case : str ): """simple docstring""" _snake_case : O...
517
'''simple docstring''' import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, ...
400
0
'''simple docstring''' import argparse import os from pathlib import Path import torch from bark.generation import _load_model as _bark_load_model from huggingface_hub import hf_hub_download from transformers import EncodecConfig, EncodecModel, set_seed from transformers.models.bark.configuration_bark import (...
717
'''simple docstring''' import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECK...
88
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a : List[Any] = logging.get_logger(__name__) a : List[str] = { ""...
639
'''simple docstring''' def UpperCamelCase_ ( _UpperCAmelCase : int ) -> int: """simple docstring""" if not isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise ValueError("Input must be an integer" ) if input_num <= 0:...
244
0
import logging import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, BertEncode...
105
import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING lowercase : Optional[int] = logging.get_logge...
105
1
import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow _lowerCAmelCase: List[str] = logging.getLogger() @unittest.skip('Temporarily disable the d...
20
from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, ...
20
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ = logging.get_logger(__name__) snake_case_ = { 'facebook/vit-mae-base': 'https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json', # See all...
68
'''simple docstring''' from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ): ...
68
1
'''simple docstring''' import inspect import unittest import numpy as np from transformers import ViTConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, float...
185
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase_ : List[Any] = { '''configuration_git''': ['''GIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GitConfig''', '''GitVisionConfig'''], '''...
185
1
def SCREAMING_SNAKE_CASE ( __lowerCAmelCase = 6008_5147_5143 ) -> Dict: try: snake_case__ = int(__SCREAMING_SNAKE_CASE ) except (TypeError, ValueError): raise TypeError('''Parameter n must be int or castable to int.''' ) if n <= 0: raise ...
704
import importlib import inspect import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowerCamelCase__ : int = """src/transformers""" # This is to make sure the tra...
208
0
import logging import os import sys from pathlib import Path from unittest.mock import patch from parameterized import parameterized from run_eval import run_generate from run_eval_search import run_search from transformers.testing_utils import CaptureStdout, TestCasePlus, slow from utils import ROUGE_KEYS l...
198
from __future__ import annotations import sys from collections import deque from typing import Generic, TypeVar lowercase = TypeVar("T") class UpperCamelCase_ ( Generic[T] ): '''simple docstring''' lowerCAmelCase = 42 # Cache store of keys lowerCAmelCase = ...
198
1
import pytest import requests from datasets.utils.file_utils import http_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline @pytest.mark.integration def A_ ( ): """simple docstring""" with offline(OfflineSimulationMode.CONNECTION_TIMES_OUT ): ...
716
import gc import unittest import numpy as np import torch from diffusers import ( AudioDiffusionPipeline, AutoencoderKL, DDIMScheduler, DDPMScheduler, DiffusionPipeline, Mel, UNetaDConditionModel, UNetaDModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.test...
353
0
from queue import Queue from typing import TYPE_CHECKING, Optional if TYPE_CHECKING: from ..models.auto import AutoTokenizer class __UpperCamelCase : """simple docstring""" def UpperCAmelCase__ ( self : Tuple , _A : Optional[int] ): ...
74
from typing import Dict import numpy as np from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline, PipelineException if is_tf_available(): import tensorflow as tf from ..tf_utils import stable_softmax if i...
171
0
import argparse import json import re from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileNetVaConfig, MobileNetVaForImageClassification, MobileNetVaImageProcessor, load_tf_weights_in_mobilenet_va, ) ...
715
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Value from .base import TaskTemplate @dataclass(frozen=__A ) class lowerCAmelCase_ ( __A ): '''simple docstring''' _lowercase = field...
153
0
def a_ ( __magic_name__ = 1_000 ) -> int: """simple docstring""" snake_case : Any = -1 snake_case : Any = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N elimina...
598
"""simple docstring""" import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _UpperCAmelCase ( __a , unittest.TestCase): __...
238
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logg...
127
from __future__ import annotations from typing import Any class lowercase__ : """simple docstring""" def __init__( self : str , __a : int ): snake_case__ : Any = num_of_nodes snake_case__ : list[list[int]] ...
127
1
import argparse import torch from transformers import GPTaConfig, GPTaModel, load_tf_weights_in_gpta from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def __magic_name__ ( lowercase , lowercase , lowercase ) -> ...
458
'''simple docstring''' def _UpperCamelCase (_lowerCamelCase : int )-> int: '''simple docstring''' __snake_case = abs(_lowerCamelCase ) __snake_case = 0 while n > 0: res += n % 10 n //= 10 return res def ...
24
0
"""simple docstring""" import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py _A = """src/transformers""" # This is to ma...
507
"""simple docstring""" import inspect import os import unittest import torch import accelerate from accelerate import debug_launcher from accelerate.test_utils import ( execute_subprocess_async, require_cpu, require_huggingface_suite, require_multi_gpu, require_single_gpu, ) from accelerate...
507
1
from __future__ import annotations def _lowerCamelCase ( lowerCamelCase_: list[int] , lowerCamelCase_: int ): '''simple docstring''' A : list[list[int]] = [] A : list[int] = [] A : Tuple = 0 A : Tuple = su...
256
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING UpperCamelCase_ = logging.get_logger(__name__) class _SCREAMING_SNAKE_CASE ( snake_case ): lowerCamelCase_ = 'upernet' ...
256
1
import unittest from knapsack import greedy_knapsack as kp class lowerCAmelCase__ ( unittest.TestCase ): def A_ ( self ) -> Tuple: '''simple docstring''' _UpperCamelCase = [10, 20, 30, 40, 50, 60] _UpperCamelCase = [2, 4, 6, 8,...
202
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_...
202
1
import torch from torch import nn from transformers import CLIPPreTrainedModel, CLIPVisionModel from ...models.attention import BasicTransformerBlock from ...utils import logging a_ = logging.get_logger(__name__) # pylint: disable=invalid-name class lowercase__ ( _UpperCAmelCase ): de...
339
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class lowercase__ ( unittest.TestCase ): def Up...
339
1
from __future__ import annotations from collections.abc import Iterator from typing import Any class _UpperCAmelCase : """simple docstring""" def __init__( self : Union[str, Any], lowerCamelCase : Any ): '''simple docstring''' lowercase__ = d...
715
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): impo...
671
0
def UpperCAmelCase_ ( __UpperCAmelCase : str , __UpperCAmelCase : Union[str, Any] ) -> Any: SCREAMING_SNAKE_CASE_ = [1] for i in range(2 , __UpperCAmelCase ): factorials.append(factorials[-1] * i ) assert 0 <= k < factorials[...
31
import unittest from typing import Tuple import torch from diffusers.utils import floats_tensor, randn_tensor, torch_all_close, torch_device from diffusers.utils.testing_utils import require_torch @require_torch class lowerCamelCase_ : '''simple docstring''' @property def lowerCAmelCase_ ...
31
1
import os from collections.abc import Iterator def _lowercase ( _UpperCAmelCase = "." ) -> Iterator[str]: for dir_path, dir_names, filenames in os.walk(_UpperCAmelCase ): lowerCamelCase =[d for d in dir_names if d != """scripts""" and d[0] not in """._"""] for...
713
import math from collections.abc import Iterator from itertools import takewhile def _lowercase ( _UpperCAmelCase ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, al...
269
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase = { """configuration_convbert""": ["""CONVBERT_PRETRAINED_CONFIG_ARCH...
474
"""simple docstring""" a__ : Optional[int] = [ 999, 800, 799, 600, 599, 500, 400, 399, 377, 355, 333, 311, 288, 266, 244, 222, 200, 199, 177, 155, 133, 111, 88, 66, 44, 22, 0, ] ...
589
0
from __future__ import annotations from collections.abc import Callable from typing import Any, Generic, TypeVar __A = TypeVar('T') class SCREAMING_SNAKE_CASE ( Generic[T] ): """simple docstring""" def __init__( self: Tuple , __A: list[T] , __A: Ca...
716
import flax.linen as nn import jax import jax.numpy as jnp class SCREAMING_SNAKE_CASE ( nn.Module ): """simple docstring""" A_ = 42 A_ = jnp.floataa def __A ( self: Tuple ) -> Tuple: _A = nn.Conv( ...
62
0
def __lowerCamelCase ( _lowercase = 600851475143 ) -> int: try: UpperCamelCase = int(_lowercase ) except (TypeError, ValueError): raise TypeError('Parameter n must be int or castable to int.' ) if n <= 0: raise ValueError('Paramet...
282
'''simple docstring''' import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from t...
92
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase : str =logging.get_logger(__name__) lowerCAmelCase : Dict ={ "microsoft/cvt-13": "https://huggingface.co/microsoft/cvt-13/resolve/main/config.json", # See all Cvt models at https://...
15
from math import log from scipy.constants import Boltzmann, physical_constants lowerCAmelCase : List[Any] =300 # TEMPERATURE (unit = K) def A__ ( __A , __A , __A , ): '''simple docstring''' if donor_conc <= 0: raise ValueError("""Dono...
15
1
'''simple docstring''' import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common...
614
'''simple docstring''' import logging import re import pytorch_quantization import pytorch_quantization.nn as quant_nn import torch from pytorch_quantization import calib from pytorch_quantization.tensor_quant import QuantDescriptor UpperCamelCase__ : Optional[Any] = logging.getLogger(__name__) Uppe...
614
1
"""simple docstring""" import argparse import torch from transformers import ( SpeechTaConfig, SpeechTaFeatureExtractor, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaProcessor, SpeechTaTokenizer, logging, ) from transformers.tokenizati...
719
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { '''shi-labs/n...
401
0
'''simple docstring''' import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis from lavis.models import load_model_and_preprocess from PIL import Image from transformer...
13
'''simple docstring''' A__ : dict[tuple[int, int, int], int] = {} def UpperCAmelCase__ ( UpperCAmelCase_ : int , UpperCAmelCase_ : int , UpperCAmelCase_ : int ) -> int: # if we are absent twice, or late 3 consecutive days, ...
13
1
'''simple docstring''' import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __magic_name__ (__low...
700
import gc import random import tempfile import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.stable_diffusion_sa...
226
0
import unittest from transformers import RoFormerTokenizer, RoFormerTokenizerFast from transformers.testing_utils import require_rjieba, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_rjieba @require_tokenizers class __A ( A_ , unittest.TestCase ): ...
157
import unittest from transformers import PegasusConfig, PegasusTokenizer, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import...
157
1
'''simple docstring''' def _UpperCAmelCase ( __A : str , __A : str ): def get_matched_characters(__A : str , __A : str ) -> str: a_ : Union[str, Any] = [] a_ : int = mi...
666
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional from packaging import version if TYPE_CHECKING: from ... import PreTrainedTokenizer, TensorType from ...configuration_utils import PretrainedConfig from ...
666
1
'''simple docstring''' import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever _UpperCamelCase = logging.getLogger(__name__) class __magic_name__ ...
111
'''simple docstring''' from __future__ import annotations from statistics import mean def lowerCAmelCase__ ( lowerCamelCase : list[int] ,lowerCamelCase : list[int] ,lowerCamelCase : int ): _A : Optional[Any] = [0] * no_of_processes _A : Li...
128
0
import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class _SCREAMING_SNAKE_CASE (unittest.TestCase ): def __snake_case ( self : List[Any] )->None: __SCREAMING_SNAKE_CASE : int ...
447
from diffusers.utils.testing_utils import require_onnxruntime @require_onnxruntime class _SCREAMING_SNAKE_CASE : pass
447
1
"""simple docstring""" import re from filelock import FileLock try: import nltk lowercase__ = True except (ImportError, ModuleNotFoundError): lowercase__ = False if NLTK_AVAILABLE: with FileLock(""".lock""") as lock: nltk.download("""punkt""", quiet=True) def ...
610
"""simple docstring""" from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be chec...
610
1
def _SCREAMING_SNAKE_CASE ( __snake_case ) -> bool: return credit_card_number.startswith(("""34""", """35""", """37""", """4""", """5""", """6""") ) def _SCREAMING_SNAKE_CASE ( __snake_case ) -> bool: _UpperCAmelCase = credit_card_number _Up...
708
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(UpperCAmelCase ) , '''Tatoeba d...
402
0
class UpperCAmelCase_ : """simple docstring""" def __init__( self , _a , _a , _a ) -> List[str]: _a : List[Any] = name _a : List[str] = value _a : List[str...
14
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a__ = logging.get_logger(__name__) a__ = { '''google/mobi...
14
1
import math from typing import Any, Callable, List, Optional, Tuple, Union import numpy as np import torch from ...models import TaFilmDecoder from ...schedulers import DDPMScheduler from ...utils import is_onnx_available, logging, randn_tensor if is_onnx_available(): from ..onnx_utils import OnnxRuntimeMode...
315
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def lowercase ( __A : bytes , __A : int ) -> np.array: '''simple docstring''' snake_case : List[str] = f"""{sampling_rate}""" snake_...
315
1
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokeniz...
131
"""simple docstring""" from __future__ import annotations from fractions import Fraction from math import gcd, sqrt def _snake_case ( lowercase__ ): _lowerCamelCase : int = int(number**0.5 ) return number == sq * sq ...
630
0
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING A__ : Any =logging.get_logger(__name__)...
499
'''simple docstring''' import argparse import re import requests import torch # git clone https://github.com/salesforce/BLIP.git from models.blip import blip_decoder from models.blip_itm import blip_itm from models.blip_vqa import blip_vqa from PIL import Image from torchvision import transforms from t...
499
1
from .integrations import ( is_optuna_available, is_ray_available, is_sigopt_available, is_wandb_available, run_hp_search_optuna, run_hp_search_ray, run_hp_search_sigopt, run_hp_search_wandb, ) from .trainer_utils import ( HPSearchBackend, default_hp_spa...
228
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) lowerCamelCase : List[Any] ={ '''configuration_mega''': ['''MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MegaConfig''', '''MegaOnnxConfig'''], ...
228
1
from unittest.mock import patch import pyspark from datasets.packaged_modules.spark.spark import ( Spark, SparkExamplesIterable, _generate_iterable_examples, ) from ..utils import ( require_dill_gt_0_3_2, require_not_windows, ) def __lowerCAmelCase ( UpperCamelCase , Upper...
706
lowerCAmelCase_ = 6_55_21 def __lowerCAmelCase ( UpperCamelCase ) -> int: lowerCAmelCase__ : List[str] = 1 lowerCAmelCase__ : List[Any] = 0 for plain_chr in plain_text: lowerCAmelCase__ : Union[str, Any] = (a + ord(UpperCamelC...
470
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _lowerca...
659
"""simple docstring""" from __future__ import annotations from math import ceil, floor, sqrt def lowerCAmelCase_( lowercase_ : int = 2_00_00_00 ) -> int: _lowerCamelCase = [0] _lowerCamelCase = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ...
661
0
import copy import os import cva import numpy as np from matplotlib import pyplot as plt class _lowercase : def __init__( self ): snake_case__ : Optional[Any] ="""""" snake_case__ : List[str] ="""""" snake_case__ : Dict ...
448
def A__ ( _a : list ): '''simple docstring''' if len(_a ) <= 1: return [tuple(_a )] snake_case__ : Optional[int] =[] def generate(_a : int , _a : list ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , _a ) for i i...
448
1
import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversati...
2
'''simple docstring''' import copy import random from transformers import CLIPTokenizer class SCREAMING_SNAKE_CASE ( lowercase_ ): '''simple docstring''' def __init__( self : int , *snake_case : Optional[Any] , **snake_case : Optional[int]...
517
0
"""simple docstring""" import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, Au...
717
"""simple docstring""" import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class _UpperCAme...
141
0
import functools def __UpperCAmelCase ( a_ , a_): snake_case_ = len(__SCREAMING_SNAKE_CASE) snake_case_ = len(__SCREAMING_SNAKE_CASE) @functools.cache def min_distance(a_ , a_) -> int: # if first word index is overflow - delete...
198
"""simple docstring""" def lowercase__( __SCREAMING_SNAKE_CASE : int = 2_00 ): lowercase_ : str = [1, 2, 5, 10, 20, 50, 1_00, 2_00] lowercase_ : Dict = [0] * (pence + 1) lowercase_ : List[Any] = 1 # base case: 1 way to make ...
425
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging a__ : Dict = logging.get_logger(__name__) a__ : Any = {'''vocab_file''': '''s...
333
import argparse from pathlib import Path from transformers import AutoConfig, AutoTokenizer, RagConfig, RagSequenceForGeneration, RagTokenForGeneration def UpperCAmelCase_( a__ , a__ , a__ , a__ , a__ = None , a__ = None , a__ = None , ): """simple docstring""" if config...
333
1
def __SCREAMING_SNAKE_CASE ( lowercase_ = 1000 ) -> int: '''simple docstring''' return sum(2 * a * ((a - 1) // 2) for a in range(3 , n + 1 ) ) if __name__ == "__main__": print(solution())
462
import inspect import unittest from transformers import DPTConfig from transformers.file_utils import is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common im...
462
1
from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import *...
46
from ..utils import DummyObject, requires_backends class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docst...
46
1
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_funnel import FunnelTokenizer A_ = logging.get_logger(__name__) A_ = {"vocab_file": "v...
42
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py SCREAMING_SNAKE_CASE = 'src/transformers' # This i...
485
0
"""simple docstring""" from sklearn.metrics import mean_squared_error import datasets __lowercase = '\\n@article{scikit-learn,\n title={Scikit-learn: Machine Learning in {P}ython},\n author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V.\n and Thirion, B. an...
711
"""simple docstring""" from timeit import timeit def lowercase ( A_ )-> int: '''simple docstring''' if number < 0: raise ValueError("the value of input must not be negative" ) a : Dict = 0 while number: number &= number - 1 res...
135
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __A : int = { 'configuration_resnet': ['RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ResNe...
394
'''simple docstring''' class _UpperCamelCase : '''simple docstring''' def __init__( self , _a ): """simple docstring""" # we need a list not a string, so do something to change the type a__ = arr.split(',' ) ...
394
1
"""simple docstring""" import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def ...
261
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler,...
261
1
from typing import Dict from .base import GenericTensor, Pipeline class __UpperCAmelCase ( __A ): """simple docstring""" def snake_case_ ( self , __A=None , __A=None , __A=None , **__A ): if tokenize_kwargs is None: ...
99
"""simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] ={ 'a': 'AAAAA', 'b': 'AAAAB', 'c': 'AAABA', 'd': 'AAABB', 'e': 'AABAA', 'f': 'AABAB', 'g': 'AABBA', 'h': 'AABBB', 'i': 'ABAAA', 'j': 'BBBAA', 'k': 'ABAAB', 'l': 'ABABA', 'm': 'AB...
434
0
"""simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] ='0.21.0' from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch,...
558
"""simple docstring""" from typing import Callable, Dict, Optional, Tuple import torch from torch import nn from torch.distributions import ( AffineTransform, Distribution, Independent, NegativeBinomial, Normal, StudentT, TransformedDistribution, ) class _UpperCAmelC...
558
1
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bert import BertTokenizer a_ = logging.get_logger(__name__) a_ = {""...
177
"""simple docstring""" from collections import namedtuple a_ = namedtuple("""from_to""", """from_ to""") a_ = { """cubicmeter""": from_to(1, 1), """litre""": from_to(0.001, 1000), """kilolitre""": from_to(1, 1), """gallon""": from_to(0.00454, 264.172), """cubicyard""": fro...
177
1
import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class lowerCAmelCase_ ( unittest.TestCase ): def a_ ( self : Optional[Any...
707
def _A ( _UpperCamelCase , _UpperCamelCase ): _UpperCAmelCase : Tuple = len(_UpperCamelCase ) _UpperCAmelCase : Tuple = len(_UpperCamelCase ) _UpperCAmelCase : Dict = [[False for _ in range(m + 1 )] for _ in range(n + 1 )] _UpperCAmelCase : List[A...
416
0
'''simple docstring''' from math import isqrt def snake_case__ ( UpperCamelCase ) -> Optional[int]: _UpperCamelCase : Dict = [True] * max_number for i in range(2 ,isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 ,UpperCamelCas...
683
__lowercase = { "km/h": 1.0, "m/s": 3.6, "mph": 1.6_0_9_3_4_4, "knot": 1.8_5_2, } __lowercase = { "km/h": 1.0, "m/s": 0.2_7_7_7_7_7_7_7_8, "mph": 0.6_2_1_3_7_1_1_9_2, "knot": 0.5_3_9_9_5_6_8_0_3, } def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_S...
167
0
import argparse import logging import os import re import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, DataCollatorForLanguageModeling, PushToHubCallback, TFAutoModelForMaskedLM, create_optimizer, ) a = logging.getLogger(__name_...
718
from math import sqrt def UpperCamelCase_( __magic_name__ : int = 1000000 ): """simple docstring""" _lowerCAmelCase :int = 0 _lowerCAmelCase :int = 0 _lowerCAmelCase :int while num_cuboids <= limit: max_cu...
382
0
"""simple docstring""" import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_configurati...
139
"""simple docstring""" from collections.abc import Callable import numpy as np def lowerCamelCase ( _UpperCamelCase : Callable , _UpperCamelCase : float , _UpperCamelCase : float , _UpperCamelCase : float , _UpperCamelCase : float ) -> np.array: ...
139
1
'''simple docstring''' import json import os from typing import Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging A : Tuple = logging.get_logger(__name__) A : List[Any] = { """vocab_file""": ...
163
'''simple docstring''' import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class SCREAMING_SNAKE_CASE( datasets.BeamBasedBuilder ): ...
163
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a ={ 'configuration_time_series_transformer': [ 'TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TimeSeriesTransformerConfig', ], } try: ...
530
def _A ( _lowercase , _lowercase ) -> int: """simple docstring""" return (pointa[0] - pointa[0]) ** 2 + (pointa[1] - pointa[1]) ** 2 def _A ( _lowercase , _lowercase=0 ) -> Dict: """simple docstring""" return sorted(_lowercase , k...
1
0
"""simple docstring""" import tempfile import unittest from make_student import create_student_by_copying_alternating_layers from transformers import AutoConfig from transformers.file_utils import cached_property from transformers.testing_utils import require_torch __A = """sshleifer/bart-tiny...
173
"""simple docstring""" __A = 6_55_21 def UpperCamelCase ( _lowerCAmelCase : str ): __a = 1 __a = 0 for plain_chr in plain_text: __a = (a + ord(_lowerCAmelCase )) % MOD_ADLER __a = (b + a) % MOD_ADLER return (b << 16) | a
173
1
'''simple docstring''' import math def UpperCamelCase__ ( __magic_name__ : int ) -> int: '''simple docstring''' if not isinstance(__magic_name__ , __magic_name__ ): snake_case__ : Union[str, Any] = f"Input value of [number={number}] must be an i...
38
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase_ : int = logging.get_logger(__name__) lowerCAmelCase_ : Any = { '''facebook/wav2vec2-base-960h''': '''https:...
673
0
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational ...
580
import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from accelerate.test_utils import require_mult...
580
1
import os def __snake_case ( ) -> List[Any]: with open(os.path.dirname(UpperCamelCase__ ) + '''/p022_names.txt''' ) as file: _a = str(file.readlines()[0] ) _a = names.replace('''"''' , '''''' ).split(''',''' ) names.sort() _a = 0 _a ...
487
import argparse import os import re import torch from flax.traverse_util import flatten_dict from tax import checkpoints from transformers import ( AutoTokenizer, PixaStructConfig, PixaStructForConditionalGeneration, PixaStructImageProcessor, PixaStructProcessor, PixaStructTextCo...
240
0
"""simple docstring""" def lowerCamelCase__ ( UpperCAmelCase_ = 10_00 )-> int: """simple docstring""" UpperCamelCase , UpperCamelCase = 1, 1 UpperCamelCase = [] for i in range(1 , n + 1 ): UpperCamelCase ...
556
"""simple docstring""" import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer...
556
1
import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartaaTokenizer, MBartaaTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, requ...
550
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transform...
550
1
'''simple docstring''' import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudi...
701
'''simple docstring''' from __future__ import annotations import unittest from transformers import EsmConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_te...
323
0
"""simple docstring""" import unittest from queue import Empty from threading import Thread from transformers import AutoTokenizer, TextIteratorStreamer, TextStreamer, is_torch_available from transformers.testing_utils import CaptureStdout, require_torch, torch_device from ..test_modeling_common import ids_tenso...
594
"""simple docstring""" from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(...
594
1
'''simple docstring''' class UpperCAmelCase_ : """simple docstring""" def __init__( self : Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' lowercase : Optional[Any] ={} def A__ ( self : Any ) -> None: ''...
8
'''simple docstring''' def lowercase_ ( __A : int , __A : int ) -> str: """simple docstring""" if number < 0 or shift_amount < 0: raise ValueError('''both inputs must be positive integers''' ) lowercase : List[Any] =str(bin(__A ) ) ...
8
1
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer, TFAutoModelForSeque...
137
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BlipaProcessor, BlipImageProce...
137
1
import multiprocessing import os from typing import BinaryIO, Optional, Union import fsspec from .. import Dataset, Features, NamedSplit, config from ..formatting import query_table from ..packaged_modules.json.json import Json from ..utils import logging from ..utils.typing import NestedDataStructureLike, ...
83
def snake_case ( snake_case__ :int = 1_000_000) -> int: _A = set(range(3 , snake_case__ , 2)) primes.add(2) for p in range(3 , snake_case__ , 2): if p not in primes: continue primes.difference...
83
1
import math def lowercase__( A , A ): snake_case__ : str = len(lowerCAmelCase_ ) snake_case__ : Dict = int(math.floor(math.sqrt(lowerCAmelCase_ ) ) ) snake_case__ : Optional[int] = 0 while arr[min(lowerCAme...
170
"""simple docstring""" import os from pathlib import Path def snake_case ( ) -> Tuple: from torch.utils.cpp_extension import load _snake_case = Path(lowerCAmelCase_ ).resolve().parent.parent.parent / '''kernels''' / '''deformable_detr''' _snake_case = ...
103
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __lowerCamelCase : Tuple = { '''configuration_blenderbot_sma...
718
import collections import json import math import os import re import time from fnmatch import fnmatch from typing import Dict import requests from slack_sdk import WebClient __lowerCamelCase : Dict = WebClient(token=os.environ['''CI_SLACK_BOT_TOKEN''']) def lowercase__ ( __A: List[Any] ...
501
0
import logging from dataclasses import dataclass, field from typing import Optional from seqaseq_trainer import arg_to_scheduler from transformers import TrainingArguments UpperCamelCase__ : Union[str, Any] = logging.getLogger(__name__) @dataclass class lowerCAmelCase_ ( lower...
105
'''simple docstring''' from __future__ import annotations from collections import deque class __lowerCAmelCase : '''simple docstring''' def __init__(self : str , UpperCamelCase : list[str] ): '''simple docstring''' lowercas...
460
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required ...
721
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, ...
62
0
'''simple docstring''' import inspect import os import unittest from dataclasses import dataclass import torch from accelerate import Accelerator, DistributedDataParallelKwargs, GradScalerKwargs from accelerate.state import AcceleratorState from accelerate.test_utils import execute_subprocess_async, require_cu...
120
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_lxmert import LxmertTokenizer UpperCAmelCase_ : Optional[Any] = {"vocab_file": "vocab.txt", "tokeniz...
120
1
"""simple docstring""" import unittest import numpy as np from diffusers import LMSDiscreteScheduler, OnnxStableDiffusionInpaintPipeline from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onn...
706
"""simple docstring""" def __lowerCAmelCase ( __UpperCamelCase : int , __UpperCamelCase : int ): '''simple docstring''' snake_case_ : Any = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): ...
21
0
from typing import Any def _UpperCamelCase ( lowercase__ ): if not input_list: return [] __SCREAMING_SNAKE_CASE : Dict = [input_list.count(snake_case__ ) for value in input_list] __SCREAMING_SNAKE_CASE : Any = max(snake_case__ ) # Gets the maximu...
696
from __future__ import annotations import math def SCREAMING_SNAKE_CASE__ ( snake_case__ :int ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples ...
67
0
"""simple docstring""" import argparse import json from typing import List from ltp import LTP from transformers.models.bert.tokenization_bert import BertTokenizer def a__ ( __lowercase ) -> Optional[Any]: # This defines a "chinese character" as anything in the CJK Unicode...
704
"""simple docstring""" def a__ ( __lowercase , __lowercase , __lowercase , __lowercase ) -> str: # Return True if there is node that has not iterated. _A = [False] * len(__lowercase ) _A = [] queue.append(__lowercase ) _A = True...
621
0
'''simple docstring''' import torch from diffusers import EulerDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class lowercase_ ( _A ): a_ = (EulerDiscreteScheduler,) a_ = 10 def lowerCamelCase_ ( s...
660
'''simple docstring''' import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_availabl...
660
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase_: List[Any] = logging.get_logger(__name__) lowerCAmelCase_: Optional[Any] = { "google/canine-s": "https://huggingface.co/google/canine-s/resolve/main/config.json", ...
668
"""simple docstring""" from sympy import diff, lambdify, symbols from sympy.functions import * # noqa: F403 def __a ( A , A , A = "x" , A = 10**-10 , A = 1 , ): '''simple docstring''' lowercase__ = symbols(A ) lowercase__ = ...
668
1
'''simple docstring''' import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor snake_case = logging.get_logger(__name__) class lowerCAmelCase ( UpperCamelCase_ ): def __init__( self : List[Any] , *a__ : int , **a__ :...
378
'''simple docstring''' def UpperCAmelCase_ ( lowerCamelCase_ ): """simple docstring""" lowerCAmelCase__ : str = len(lowerCamelCase_ ) lowerCAmelCase__ : Optional[Any] = len(matrix[0] ) lowerCAmelCase__ : Any = min(lowerCamelCase_ , lowerCamelCase_ ...
378
1
"""simple docstring""" import math def lowerCAmelCase__ ( _UpperCamelCase : float , _UpperCamelCase : float ) -> float: """simple docstring""" if initial_intensity < 0: raise ValueError('The value of intensity cannot be ...
104
"""simple docstring""" import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_CASE__ = logging.get_logger(__name__) S...
104
1
'''simple docstring''' from __future__ import annotations import numpy as np def lowerCamelCase__ ( SCREAMING_SNAKE_CASE : np.ndarray ): UpperCAmelCase = np.shape(SCREAMING_SNAKE_CASE ) if rows != columns: UpperCAmelCase = ( "'table' has to be of square sh...
447
from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class a ( UpperCAmelCase ): def _UpperCAmelCase ( self , A_ ): '''simple docstring''' return 0.0 ...
300
0
'''simple docstring''' from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCamelCase : Optional[Any] = { '''configuration_cpmant''': ...
707
'''simple docstring''' from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def __UpperCAmelCase ( )-> int: """simple docstring""" snake_case_ : Any = { ...
656
0