code
stringlengths
82
53.2k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available SCREAMING_SNAKE_CASE :Tuple = { '''configuration_ctrl''': ['''CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''CTRLConfig'''], '''tokenization_ctrl''': ['''CTRL...
283
from __future__ import annotations from math import gcd def _lowerCAmelCase ( lowerCAmelCase_ :int , lowerCAmelCase_ :int = 2 , lowerCAmelCase_ :int = 1 , lowerCAmelCase_ :int = 3 , )->int | None: '''simple docstring''' if num < 2: raise...
283
1
from .imports import is_rich_available if is_rich_available(): from rich.traceback import install install(show_locals=False) else: raise ModuleNotFoundError('''To use the rich extension, install rich with `pip install rich`''')
77
import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester fr...
77
1
'''simple docstring''' import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, PLBartTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...t...
131
'''simple docstring''' from torch import nn class A ( nn.Module ): def __init__( self , snake_case_ , snake_case_ ) -> List[Any]: super().__init__() _a = class_size _a = embed_size # self.mlp1 = nn.Linear(embed_size, emb...
131
1
"""simple docstring""" from decimal import Decimal, getcontext from math import ceil, factorial def lowerCamelCase_(__SCREAMING_SNAKE_CASE )-> str: if not isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): raise TypeError("""Undefined for non-integers""" ) ...
720
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.hugg...
635
0
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from tra...
266
from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import BaseOutput, is_torch_available, is_transformers_available @dataclass class __lowerCAmelCase ( _UpperCamelC...
266
1
import flax.linen as nn import jax.numpy as jnp from .attention_flax import FlaxTransformeraDModel from .resnet_flax import FlaxDownsampleaD, FlaxResnetBlockaD, FlaxUpsampleaD class lowerCamelCase_ ( nn.Module ): _lowerCamelCase : int _lowerCamelCase : int _lowerC...
403
import unittest import numpy as np import torch from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad class lowerCamelCase_ ( unittest.TestCase ): def __magic_name__ ( self ): a_ = 10 def __...
403
1
import pytest UpperCAmelCase_ = """__dummy_dataset1__""" UpperCAmelCase_ = """ import json import os import datasets REPO_URL = \"https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/\" URLS = {\"train\": REPO_URL + \"wikiann-bn-train.jsonl\", \"validation\": REPO_...
2
"""simple docstring""" # limitations under the License. # NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from .pipelines import DiffusionPipeline, ImagePipelineOutput # noqa: F401 from ...
512
0
def __snake_case ( _UpperCAmelCase = 10_00 ): """simple docstring""" lowercase = 2**power lowercase = 0 while n: lowercase = r + n % 10, n // 10 return r if __name__ == "__main__": print(solution(int(str(input()).strip()))) ...
719
from __future__ import annotations from fractions import Fraction from math import gcd, sqrt def __snake_case ( _UpperCAmelCase ): """simple docstring""" lowercase = int(number**0.5 ) return number == sq * sq def __snake_case...
314
0
def snake_case ( lowerCamelCase ): '''simple docstring''' return str(lowerCamelCase ) == str(lowerCamelCase )[::-1] def snake_case ( lowerCamelCase ): '''simple docstring''' return int(lowerCamelCase ) + int(str(lowerCamelCase )[::-1]...
80
"""simple docstring""" from typing import TYPE_CHECKING from ..utils import _LazyModule UpperCamelCase_ : Dict = { '''config''': [ '''EXTERNAL_DATA_FORMAT_SIZE_LIMIT''', '''OnnxConfig''', '''OnnxConfigWithPast''', '''OnnxSeq2SeqConfigWithPast''', ...
115
0
"""simple docstring""" import heapq as hq import math from collections.abc import Iterator class _UpperCAmelCase : def __init__( self , lowercase_ ) -> Tuple: UpperCAmelCase = str(id_ ) UpperCAmelCase = No...
183
"""simple docstring""" import itertools import os import random import tempfile import unittest import numpy as np from transformers import TvltFeatureExtractor, is_datasets_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from trans...
183
1
from __future__ import annotations import math import random from typing import Any class UpperCAmelCase__ : """simple docstring""" def __init__( self: str ) -> None: '''simple docstring''' __UpperCAmelCase = [] __UpperCAmelCase = 0 ...
221
def __lowerCAmelCase ( A_ : int ) -> int: __UpperCAmelCase = 0 while num > 0: digit_sum += num % 10 num //= 10 return digit_sum def __lowerCAmelCase ( A_ : int = 1_00 ) -> int: __UpperCAmelCase = 1 __UpperCAmelCase = ...
221
1
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, ...
713
'''simple docstring''' import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging _UpperCamelCase : List[Any] = "\\n\n" _UpperCamelCase : List[Any] = "\...
514
0
'''simple docstring''' from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def _snake_case ( _SCREAMING_SNAKE_CASE : int ) -> Dict: """simple docstring""" # A local function to see if a...
433
'''simple docstring''' from math import factorial def _snake_case ( _SCREAMING_SNAKE_CASE : int = 100 ) -> int: """simple docstring""" return sum(map(_SCREAMING_SNAKE_CASE , str(factorial(_SCREAMING_SNAKE_CASE ) ) ) ) if __name__ =...
433
1
'''simple docstring''' from ...processing_utils import ProcessorMixin class __A ( a ): """simple docstring""" A_ = ['image_processor', 'feature_extractor'] A_ = 'TvltImageProcessor' A_ = 'TvltFeatureExtr...
318
'''simple docstring''' def _lowerCAmelCase ( lowercase : int , lowercase : int ) ->str: """simple docstring""" if a < 0 or b < 0: raise ValueError('''the value of both inputs must be positive''' ) lowercase__ = st...
318
1
import warnings from typing import List import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import is_flax_available, is_tf_available, is_torch_available class _a ( _UpperCamelCase ): '''simple docstring''' l...
520
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def a_ ( ): __lowerCAmelCase = ArgumentParser( description=( 'PyTorch TPU distributed training launch ' ...
53
0
import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils impor...
527
import numpy as np import torch from torch.utils.data import DataLoader from accelerate.utils.dataclasses import DistributedType class lowerCamelCase_ : '''simple docstring''' def __init__( self , snake_case_=2 , snake_case_=3 , snake_case_=6_4 , snake_case_=Non...
527
1
"""simple docstring""" import operator as op def lowercase__ ( snake_case_ :Union[str, Any] ): __UpperCAmelCase = [] __UpperCAmelCase = lambda snake_case_ , snake_case_ : int(x / y ) # noqa: E731 integer division operation __UpperCAmelCase ...
49
"""simple docstring""" import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class UpperCamelCase...
95
0
"""simple docstring""" from ...utils import logging from ..ta.modeling_tf_ta import TFTaEncoderModel, TFTaForConditionalGeneration, TFTaModel from .configuration_mta import MTaConfig _lowercase = logging.get_logger(__name__) _lowercase = "T5Config" class __SCREAMING_SNAKE_CASE ( lowerCAmelC...
700
"""simple docstring""" import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_a...
625
0
"""simple docstring""" import unittest from transformers import DonutProcessor __A : Dict = '''naver-clova-ix/donut-base''' class _UpperCAmelCase ( unittest.TestCase ): def A ( self : Optional[int] ) -> Tuple: ...
231
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __A : Union[str, Any] = logging.get_logger(__name__) __A : Optional[int] = { '''google/realm-cc-news-pretrained-embedder''': ( '...
231
1
"""simple docstring""" import unittest from transformers import DonutProcessor A = 'naver-clova-ix/donut-base' class UpperCAmelCase__ ( unittest.TestCase ): def A_ ( self : int ) -> str: '''simple docstring''' A = Donu...
109
"""simple docstring""" import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational ...
109
1
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tok...
93
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_torch_available, is_torch_tpu_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_torch_available(): import torch if is_torch_tpu_available(check_device=False)...
162
0
import math def A_ ( snake_case : int ) -> int: '''simple docstring''' if not isinstance(snake_case , snake_case ): __UpperCamelCase = f"Input value of [number={number}] must be an integer" raise TypeError(snake_case ) if number < 1...
701
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING lowercase__ : Tuple = logging.get_logger(__name__) lo...
451
0
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowerCAmelCase_ : Union[str, Any] = logging.get_l...
692
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE__ = { "configuration_roformer": ["ROFORMER_PRET...
532
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { """configuration_nllb_moe""": [ """NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NllbMoeConfig""", ] } try: if not is_torch_available...
719
import copy import tempfile import unittest from huggingface_hub import HfFolder, delete_repo from parameterized import parameterized from requests.exceptions import HTTPError from transformers import AutoConfig, GenerationConfig from transformers.testing_utils import TOKEN, USER, is_staging_test class __magi...
226
0
import argparse import random import joblib import numpy as np import torch from igf.igf import ( SecondaryLearner, collect_objective_set, compute_perplexity, generate_datasets, load_gpta, recopy_gpta, set_seed, train_secondary_learner, ) from torch.utils.data impo...
43
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCAmelCase = { 'configuration_convbert': ['CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ConvBertConfig',...
43
1
from __future__ import annotations __SCREAMING_SNAKE_CASE : Optional[Any] = list[list[int]] # assigning initial values to the grid __SCREAMING_SNAKE_CASE : Matrix = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0...
149
def snake_case_ ( lowercase__ : list[int] ): '''simple docstring''' _lowerCAmelCase =[] if len(lowercase__ ) == 1: return [nums.copy()] for _ in range(len(lowercase__ ) ): _lowerCAmelCase =nums.pop(0 ) _lowerCAmelCase ...
149
1
import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): ...
491
import unittest import numpy as np import torch from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad class UpperCamelCase ( unittest.TestCase ): def __A ( self ): A__ = 10 def __A ( self ): ...
491
1
import json import os import shutil import warnings from argparse import ArgumentParser, Namespace from pathlib import Path from typing import List from ..utils import logging from . import BaseTransformersCLICommand try: from cookiecutter.main import cookiecutter __snake_case :str = True except ImportEr...
60
from unittest import TestCase from datasets import Sequence, Value from datasets.arrow_dataset import Dataset class _A ( __UpperCAmelCase ): def _lowerCamelCase ( self : int): '''simple docstring''' return [ {"col_1": 3, "col_2": "a"}, ...
60
1
def lowerCAmelCase ( UpperCamelCase__ : int , UpperCamelCase__ : int ) -> str: """simple docstring""" if a < 0 or b < 0: raise ValueError('''the value of both inputs must be positive''' ) __SCREAMING_SNAKE_CASE: Any = str(bin(U...
202
from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, ...
202
1
from __future__ import annotations from scipy.special import comb # type: ignore class lowercase__ : """simple docstring""" def __init__( self : Union[str, Any] , __a : list[tuple[float, float]] ): snake_case__ : int = list_of_...
127
import re import string import numpy as np import datasets lowercase_: Optional[Any] = '\nReturns the rate at which the input predicted strings exactly match their references, ignoring any strings input as part of the regexes_to_ignore list.\n' lowercase_: Optional[int] = '\...
127
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available a : Optional[Any] = { '''configuration_altclip''': [ '''ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''AltCLI...
555
"""simple docstring""" def _UpperCamelCase ( _A ) -> int: """simple docstring""" if not isinstance(_A , _A ) or number < 0: raise ValueError("""Input must be a non-negative integer""" ) _UpperCAmelCase = 0 while number: # This way we arrive ...
555
1
import json import os import unittest from transformers import BatchEncoding, MvpTokenizer, MvpTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ......
108
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from ...utils import BaseOutput, OptionalDependencyNotAvailable, is_torch_available, is_transformers_available from .timesteps import ( fastaa_timesteps, smartaa_timesteps, smartaa_timesteps, sma...
108
1
"""simple docstring""" from ..utils import DummyObject, requires_backends class __lowerCamelCase ( metaclass=A__ ): '''simple docstring''' a_ : int = ["""sentencepiece"""] def __init__( self : str , *a_ : Any , **a_ : ...
610
"""simple docstring""" import json import os import unittest from transformers import AutoTokenizer, GPTaTokenizer, GPTaTokenizerFast from transformers.models.gpta.tokenization_gpta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import Token...
610
1
UpperCamelCase__ = '''Alexander Joslin''' import operator as op from .stack import Stack def UpperCAmelCase__ ( _A ): """simple docstring""" a_ = {'''*''': op.mul, '''/''': op.truediv, '''+''': op.add, '''-''': op.sub} a_ ...
143
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCamelCase__ = '''▁''' UpperCamelCase__ = {'''vocab_file''': '''...
143
1
"""simple docstring""" import argparse import json from typing import List from ltp import LTP from transformers import BertTokenizer def UpperCamelCase__ ( lowercase__ : Tuple ): # This defines a "chinese character" as anything in the CJK Unicode block: # https://en.wikipedia.or...
134
"""simple docstring""" import argparse import requests import torch from PIL import Image from transformers import CLIPProcessor, GroupViTConfig, GroupViTModel def UpperCamelCase__ ( lowercase__ : List[Any] ): # vision encoder if "img_encoder.pos_embed" in name: snake_case : ...
134
1
"""simple docstring""" import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' def __init__( self : Dict, lowerCamelCase ...
625
"""simple docstring""" import argparse import json import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( VideoMAEConfig, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEImageProcessor, ) def snake_case__ ...
625
1
import re import time from typing import Optional import IPython.display as disp from ..trainer_callback import TrainerCallback from ..trainer_utils import IntervalStrategy, has_length def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> str: _lowercase : Any = int(_SCRE...
66
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( ...
186
0
import string def a ( A__ : str ) -> str: """simple docstring""" _lowercase ='' for i in sequence: _lowercase =ord(A__ ) if 65 <= extract <= 90: output += chr(155 - extract ) elif 97 <= extrac...
380
import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs...
380
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _A = { 'configuration_upernet': ['UperNetConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvai...
159
"""simple docstring""" class lowerCamelCase : '''simple docstring''' def __init__( self : str , _snake_case : list[int] ) -> None: SCREAMING_SNAKE_CASE__ = len(_snake_case ) SCREAMING_SNAKE_CASE__ = [0] * len_arr...
159
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE : Union[str, Any] = { 'configuration_bridgetower': [ 'BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Br...
703
import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def lowerCAmelCase_ ( _SCREAMING_SNAKE_CASE : List[str] , _SCREAMING_SNAKE_CASE : Any , _SCREA...
138
0
"""simple docstring""" def lowercase (_snake_case ) -> list[int]: '''simple docstring''' if num <= 0: raise ValueError("Input must be a positive integer" ) __UpperCamelCase = [True] * (num + 1) __UpperCamelCase = 2 while p * p <= num: if prime...
505
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailabl...
505
1
def a__ ( lowercase__ ): '''simple docstring''' UpperCAmelCase_ =0 while len(lowercase__ ) > 1: UpperCAmelCase_ =0 # Consider two files with minimum cost to be merged for _ in range(2 ): UpperCAmelCase_ =file...
709
import sys import turtle def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' return (pa[0] + pa[0]) / 2, (pa[1] + pa[1]) / 2 def a__ ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , ): ...
550
0
from __future__ import annotations def __UpperCamelCase ( lowercase__ : list[int] , lowercase__ : int ) -> list[int]: '''simple docstring''' lowerCAmelCase_ : Dict = 0 lowerCAmelCase_ : Union[str, Any] = len(lower...
600
def __UpperCamelCase ( lowercase__ : int , lowercase__ : int , lowercase__ : int ) -> float: '''simple docstring''' lowerCAmelCase_ : Tuple = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) ...
600
1
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin ...
122
"""simple docstring""" import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models...
122
1
'''simple docstring''' import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class SCREAMING_SNAKE_CASE ( datasets.BeamBasedBuilder ): """simple docstring""" ...
309
'''simple docstring''' import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import ja...
497
0
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dim...
260
import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput def _UpperCamelCase ( ...
260
1
'''simple docstring''' from collections.abc import Callable import numpy as np def lowercase__ ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase )-> np.array: UpperCamelCase = ...
301
'''simple docstring''' import argparse import copy def lowercase__ ( __UpperCamelCase )-> Union[str, Any]: UpperCamelCase = {} with open(__UpperCamelCase ) as f: for line in f: if line.split()[0] not ...
301
1
import json import os import re import sys import urllib.request import requests from bsa import BeautifulSoup a__ : Dict = { "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36" " (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582" } def _lowerCAm...
642
import math import sys def _lowerCAmelCase ( A__ ): lowercase__ = '' try: with open(A__ , 'rb' ) as binary_file: lowercase__ = binary_file.read() for dat in data: lowercase__ = F'''{dat:08b}''' r...
642
1
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-...
495
'''simple docstring''' import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_a...
495
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowerCAmelCase : str = { 'configuration_table_transformer': [ 'TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TableTransformerConfig', 'TableTransform...
662
import argparse import os import re import numpy as np import PIL import torch from timm import create_model from torch.optim.lr_scheduler import OneCycleLR from torch.utils.data import DataLoader, Dataset from torchvision.transforms import Compose, RandomResizedCrop, Resize, ToTensor from accelerate import Acce...
662
1
def _snake_case ( __snake_case ): _UpperCamelCase = len(__snake_case ) _UpperCamelCase = sum(__snake_case ) _UpperCamelCase = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): _Upp...
10
"""simple docstring""" import argparse import os import re import packaging.version __UpperCamelCase : Union[str, Any] = '''examples/''' __UpperCamelCase : str = { '''examples''': (re.compile(R'''^check_min_version\("[^"]+"\)\s*$''', re.MULTILINE), '''check_min_v...
4
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _lowercase: Optional[int] = { '''configuration_vision_encoder_decoder''': ['''VisionEncoderDecoderConfig''', '''VisionEncoderDe...
225
import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, BertEmbeddings, BertLayer, ...
225
1
def UpperCAmelCase__ ( lowerCamelCase_ : int = 1_0_0_0_0_0_0 ): __a : Optional[Any] = 1 __a : Tuple = 1 __a : Dict = {1: 1} for inputa in range(2 , lowerCamelCase_ ): __a : str = ...
47
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from ...utils import logging, randn_tensor from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline lowercase = logging.get_logger(__name__) # pylint: disable=invalid-name cl...
573
0
A_ : Any = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] A_ : Tuple = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] A_ : Optional[int] = { 0: '''Sunday''', 1: '''Monday''', 2: '''Tuesday''', 3: '''Wednesday''', 4: '''Thursday''', 5: '''Friday''', ...
32
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class _lowercase ( UpperCAmelCase__ ): def A ( self : Optional[int] , __lowerCAmelCase : str ) -> Union[str, Any]: """s...
32
1
'''simple docstring''' import argparse import os import transformers from .convert_slow_tokenizer import SLOW_TO_FAST_CONVERTERS from .utils import logging logging.set_verbosity_info() lowercase = logging.get_logger(__name__) lowercase = {name: getattr(transformers, name + ...
211
'''simple docstring''' def __A ( _SCREAMING_SNAKE_CASE : int ): """simple docstring""" if number > 0: raise ValueError("input must be a negative integer" ) __SCREAMING_SNAKE_CASE : Tuple = len(bin(_SCREAMING_SNAKE_C...
211
1
'''simple docstring''' def SCREAMING_SNAKE_CASE ( a_ : str , a_ : str ): __a = len(a_ ) __a = len(a_ ) __a = ( first_str_length if first_str_length > second_str_length else second_str_length ) __a = [] for char_cou...
490
'''simple docstring''' def SCREAMING_SNAKE_CASE ( a_ : int , a_ : int ): return int((input_a, input_a).count(0 ) != 0 ) def SCREAMING_SNAKE_CASE ( ): assert nand_gate(0 , 0 ) == 1 assert nand_gate(0 , 1 ) == 1 a...
490
1
'''simple docstring''' from __future__ import annotations def _SCREAMING_SNAKE_CASE ( lowerCamelCase__ : List[Any] ): '''simple docstring''' create_state_space_tree(lowerCamelCase__ , [] , 0 , [0 for i in range(len(lowerCamelCase__ ) )] ...
135
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configu...
162
0
"""simple docstring""" import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, log...
251
"""simple docstring""" import json import os import unittest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_tokenizers from ...test_tokenization_common import To...
251
1
'''simple docstring''' import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, ClassLabel, Features from .base import TaskTemplate @dataclass(frozen=__magic_name__ ) class __lowerCAmelCase ( __magic_name__ ): """simpl...
98
'''simple docstring''' from __future__ import annotations def _A ( A__ , A__ ): """simple docstring""" if b == 0: return (1, 0) ((__lowercase) , (__lowercase)) = extended_euclid(A__ , a % b ) __lowercase = a // b return (y, x - k * y) def...
41
0
"""simple docstring""" import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class __snake_case ( _lowercase): snake_case__ : Union[str, Any] = (UnCLIPScheduler,) def SCREAMING_SNAKE_CASE ( self : Tuple , *...
709
"""simple docstring""" lowerCAmelCase__ = {'''a''': ['''c''', '''b'''], '''b''': ['''d''', '''e'''], '''c''': [], '''d''': [], '''e''': []} lowerCAmelCase__ = ['''a''', '''b''', '''c''', '''d''', '''e'''] def snake_case_ ( A_ : Any, A_ : Optional[Any...
598
0
from __future__ import annotations from math import pi, sqrt def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): if inductance <= 0: raise ValueError("Inductance cannot be 0 or negative" ) elif capacitance <= 0: raise ValueError("Capacitance cannot be 0 or negative"...
413
'''simple docstring''' import argparse import os import torch from transformers import FlavaConfig, FlavaForPreTraining from transformers.models.flava.convert_dalle_to_flava_codebook import convert_dalle_checkpoint def lowerCamelCase__ ( SCREAMING_SNAKE_CASE : str ): # encoder.embeddings are dou...
447
0
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TextGenerationPipeline, logging, pipeline, ) from transformers.testing_utils import ( CaptureLogger, is_pipeline_test, require_...
517
'''simple docstring''' import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _lower...
517
1
'''simple docstring''' from __future__ import annotations from math import pow, sqrt def __UpperCAmelCase ( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_) -> Dict: if (resistance, reactance, impedance).count(0) != 1: raise ValueError('One an...
596
'''simple docstring''' def A_ ( _lowerCAmelCase : int ): """simple docstring""" _lowerCamelCase : int = (1 + 24 * n) ** 0.5 return ((1 + root) / 6) % 1 == 0 def A_ ( _lowerCAmelCase : int = 5000 ): """simple docstring""" ...
44
0
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_f...
718
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase : List[str] = logging.get_logger(__name__) lowercase ...
423
0
'''simple docstring''' import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from t...
51
from __future__ import annotations import inspect import unittest from math import floor import numpy as np from transformers import CvtConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from...
166
0
import logging import os from typing import Dict, List, Optional, Union import torch import torch.nn as nn from accelerate.utils.imports import ( is_abit_bnb_available, is_abit_bnb_available, is_bnb_available, ) from ..big_modeling import dispatch_model, init_empty_weights from .datac...
714
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available a_ : str = { "configuration_data2vec_audio": ["DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP", "Data2VecAudioCon...
532
0
"""simple docstring""" import argparse import os import torch from transformers import FlavaConfig, FlavaForPreTraining from transformers.models.flava.convert_dalle_to_flava_codebook import convert_dalle_checkpoint def __magic_name__ ( _lowerCamelCase: int ) -> List[Any]: '''simple docstri...
535
"""simple docstring""" class lowercase : def __init__(self : Dict ) -> List[Any]: """simple docstring""" lowerCAmelCase = {} def UpperCAmelCase (self : Union[str, Any] ) -> None: """simple docstring""" print(self.vert...
535
1
from __future__ import annotations def _lowercase ( lowerCamelCase__ : str, lowerCamelCase__ : list[str] | None = None ): _a = word_bank or [] # create a table _a = len(_lowerCamelCase ) + 1 _a = [] for _ in range(_lowerCamelCase ...
700
'''simple docstring''' import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require...
691
0
# XXX: we want transformers master here - in the absense of conftest manipulating sys.path: # hack it in for now: import sys from pathlib import Path _lowerCamelCase : Tuple = Path(__file__).resolve().parents[3] / """src""" sys.path.insert(1, str(git_repo_path)) import dataclasses # noqa impor...
87
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase_ = {"""configuration_xgl...
384
0
"""simple docstring""" import argparse import os import re import tensorflow as tf import torch from transformers import BertConfig, BertModel from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase = logging.get_logger(__name__) def lowerCAmelCase_ ( lowe...
716
"""simple docstring""" from __future__ import annotations def lowerCAmelCase_ ( lowercase_ : int , lowercase_ : int ): '''simple docstring''' if b == 0: return (1, 0) ((__SCREAMING_SNAKE_CASE) , (__SCREAMING_SNAKE_CASE)) : Tuple = ext...
401
0
def lowerCamelCase__ ( __lowerCamelCase : int ): __UpperCAmelCase : List[str] = [1] __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Any = 0, 0, 0 __UpperCAmelCase : Any = ugly_nums[ia] * 2 __UpperCAmelCase : O...
63
from sympy import diff, lambdify, symbols from sympy.functions import * # noqa: F403 def lowerCAmelCase_ ( lowercase: str , lowercase: complex , lowercase: str = "x" , lowercase: float = 10**-10 , lowercase: int = 1 , ) -> complex: '''simple docstri...
271
0
"""simple docstring""" A = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] A = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] A = { 0: """Sunday""", 1: """Monday""", 2: """Tuesday""", 3: """Wednesday""", 4: """Thursday""", 5: """Friday""", 6: """Saturday""", } def UpperCamelCase_ ( lowerCame...
147
"""simple docstring""" from __future__ import annotations import unittest from transformers import MobileBertConfig, is_tf_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf...
147
1
import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase = logging.get_logger(__name__) lowerCAmelCase = {'''vocab_file''': '''vocab.json'''} lowerCAmelCase = { '''voca...
230
import copy import os from typing import TYPE_CHECKING, List, Union if TYPE_CHECKING: pass from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase = logging.get_logger(__name__) lowerCAmelCase = { '''kakaobrain/align-base''': '''https://h...
230
1
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging lowercase : Dict = logging.get_logger(__name__) def A_ ( A__ , ...
392
from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase : List[str] = logging.get_logger(__name__) lowercase : Any = { """huggingface/time-series-transformer-tourism-monthly""": ( """https://hugging...
392
1
import importlib.util import json import os import warnings from dataclasses import dataclass, field import torch from ..training_args import TrainingArguments from ..utils import cached_property, is_sagemaker_dp_enabled, logging A_ : Tuple = logging.get_logger(__n...
57
import argparse import torch from torch import nn from transformers import SpeechaTextConfig, SpeechaTextForConditionalGeneration def lowerCAmelCase_ ( _lowercase : Dict) -> List[str]: """simple docstring""" a__ : Union[str, Any] = [ """encod...
136
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : List[str] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : str = { 'SCUT-DLVCLab/lilt-roberta-en-base': ( 'https...
697
'''simple docstring''' import numpy as np from PIL import Image def _snake_case ( lowercase , lowercase , lowercase ) -> np.ndarray: __a : Any = np.array(lowercase ) if arr.shape[0] != arr.shape[1]: raise ValueError("""The input array i...
697
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) _SCREAMING_SNAKE_CASE : List[Any] = { "configuration_longformer": [ "LONGFORMER_PR...
550
import os from distutils.util import strtobool def UpperCAmelCase__ (UpperCamelCase_ ,UpperCamelCase_ ): """simple docstring""" for e in env_keys: snake_case = int(os.environ.get(UpperCamelCase_ ,-1 ) ) if val >= 0...
550
1
import argparse import json import os import tensorstore as ts import torch from flax import serialization from flax.traverse_util import flatten_dict, unflatten_dict from tensorflow.io import gfile from transformers.modeling_utils import dtype_byte_size from transformers.models.switch_transformers.c...
627
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transform...
627
1
import sys UpperCamelCase = ( '73167176531330624919225119674426574742355349194934' '96983520312774506326239578318016984801869478851843' '85861560789112949495459501737958331952853208805511' '12540698747158523863050715693290963295227443043557' '66896648950445244523161731856403...
61
"""simple docstring""" def A_ (__a ): '''simple docstring''' A_ = len(__a ) while cur > 1: # Find the maximum number in arr A_ = arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi A_ = arr[mi::-1] + ar...
115
0
from collections.abc import Iterable from typing import Any class a : """simple docstring""" def __init__( self : Union[str, Any] , lowerCamelCase : int | None = None ) -> Optional[int]: __snake_case : int = value ...
203
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LDMTextToImagePipeline, UNetaDConditionModel from diffusers.utils.testing_utils import ( enable_full_determinism, load_numpy, ...
203
1
import os import zipfile import pytest from datasets.utils.extract import ( BzipaExtractor, Extractor, GzipExtractor, LzaExtractor, SevenZipExtractor, TarExtractor, XzExtractor, ZipExtractor, ZstdExtractor, ) from .utils import require_lza, require_pyazr, require_zstandard ...
66
'''simple docstring''' import numpy # List of input, output pairs lowercase : Any = ( ((5, 2, 3), 15), ((6, 5, 9), 25), ((11, 12, 13), 41), ((1, 1, 1), 8), ((11, 12, 13), 41), ) lowercase : str = (((5_15, 22, 13), 5_55), ((61,...
634
0
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from trans...
720
import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TextClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipeli...
259
0
"""simple docstring""" import argparse from tax import checkpoints from transformers import AutoConfig, FlaxAutoModelForSeqaSeqLM def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = Auto...
636
'''simple docstring''' import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyV...
404
0
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common im...
315
import time import warnings from abc import ABC from copy import deepcopy from typing import Optional import torch from ..utils import add_start_docstrings, logging __lowercase : Union[str, Any] = logging.get_logger(__name__) __lowercase : Optional[Any] = r''' Args: i...
315
1
from ..utils import DummyObject, requires_backends class A__ ( metaclass=A__ ): """simple docstring""" _lowercase = ['transformers', 'torch', 'note_seq'] def __init__( self : Optional[int] , *lowerCamelCase__ : int , **lowerCamelCase__ : int ): require...
37
'''simple docstring''' import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput de...
78
0
"""simple docstring""" import gc import unittest import numpy as np import torch import torch.nn.functional as F from transformers import ( ClapTextConfig, ClapTextModelWithProjection, RobertaTokenizer, SpeechTaHifiGan, SpeechTaHifiGanConfig, ) from diffusers import ( AudioLDMPipeline...
703
"""simple docstring""" import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import Backbon...
299
0
def __lowerCamelCase ( UpperCAmelCase_ : int = 1000 ): """simple docstring""" a :Union[str, Any] = 2**power a :Union[str, Any] = 0 while n: a , a :List[str] = r + n % 10, n // 10 return r if __name__ == ...
445
from dataclasses import dataclass from typing import Optional, Tuple import torch from torch import nn from transformers import RobertaPreTrainedModel, XLMRobertaConfig, XLMRobertaModel from transformers.utils import ModelOutput @dataclass class _snake_case ( _snake_case ): SCREAMING_SNAKE_CASE_...
445
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available __snake_case : int = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except Optio...
703
"""simple docstring""" import argparse import logging import os from pathlib import Path from typing import Any, Dict import pytorch_lightning as pl from pytorch_lightning.utilities import rank_zero_info from transformers import ( AdamW, AutoConfig, AutoModel, AutoMod...
615
0
'''simple docstring''' import random import timeit from functools import wraps from typing import Callable, Optional from ..configuration_utils import PretrainedConfig from ..models.auto.modeling_tf_auto import TF_MODEL_MAPPING, TF_MODEL_WITH_LM_HEAD_MAPPING from ..utils import is_pyanvml_available, ...
433
'''simple docstring''' import copy import tempfile import unittest from transformers import MaMaaaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from transformers.utils import cached_property from ...generation.test_...
399
0
import numpy class __SCREAMING_SNAKE_CASE : def __init__( self, _a, _a ) -> None: __SCREAMING_SNAKE_CASE = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second a...
214
from datasets.utils.patching import _PatchedModuleObj, patch_submodule from . import _test_patching def _A ( ) -> Optional[int]: """simple docstring""" import os as original_os from os import path as original_path from os import rename as original_rename ...
214
1
'''simple docstring''' import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.text import TextDatasetReader from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def __A ( lowerCamelCase_ , lowerCamelCase_ ): """s...
379
from timeit import timeit def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE ) -> int: if number < 0: raise ValueError('the value of input must not be negative' ) SCREAMING_SNAKE_CASE_ : Tuple = 0 while number: number &= number - 1 result += 1 return result def _...
345
0
"""simple docstring""" import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def lowercase_ ( ) -> Tuple: lowerCAmelCase__ : Any = ArgumentParser( description=( ...
711
"""simple docstring""" from math import atan, cos, radians, sin, tan from .haversine_distance import haversine_distance _A = 6_378_137.0 _A = 6_356_752.314_245 _A = 6_3_7_8_1_3_7 def lowercase_ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ...
507
0
"""simple docstring""" from ..utils import DummyObject, requires_backends class lowercase__ ( metaclass=A_ ): __UpperCAmelCase = ['''torch''', '''scipy'''] def __init__( self , *SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE) -> str: ...
88
"""simple docstring""" from math import isqrt, loga def _snake_case ( __snake_case : int ): """simple docstring""" _lowerCamelCase : List[str] = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]:...
88
1
"""simple docstring""" import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class UpperCAmelCase_ ( ...
717
"""simple docstring""" def __UpperCAmelCase ( _snake_case : list ): _lowercase = len(_snake_case ) for _ in range(_snake_case ): for i in range(_ % 2, arr_size - 1, 2 ): if arr[i + 1] < arr[i]: _lowercase , _lowercase ...
227
0
import importlib import shutil import threading import warnings from typing import List import fsspec import fsspec.asyn from . import compression from .hffilesystem import HfFileSystem __lowerCamelCase : Tuple = importlib.util.find_spec("""s3fs""") is not None if _has_safs: from .safilesystem impo...
297
import argparse import os import re __lowerCamelCase : int = """src/diffusers""" # Pattern that looks at the indentation in a line. __lowerCamelCase : List[str] = re.compile(R"""^(\s*)\S""") # Pattern that matches `"key":" and puts `key` in group 0. __lowerCamelCase : Union[str, Any] =...
297
1
"""simple docstring""" import gc import unittest import numpy as np import torch from diffusers import StableDiffusionKDiffusionPipeline from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() @slow ...
222
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...feature_extraction_utils import FeatureExtractionMixin from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimensio...
222
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case : str =logging.get_logger(__name__) __snake_case : Tuple ={ 'microsoft/biogpt': 'https://huggingface.co/microsoft/biogpt/resolve/main/config.json', # See all BioGPT models at http...
647
def lowerCAmelCase__ ( lowerCamelCase_ : Dict): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = len(lowerCamelCase_) while cur > 1: # Find the maximum number in arr lowerCAmelCase__ : Tuple = arr.index(max(arr[0:cur])) ...
647
1
import argparse import json import torch from diffusers import DDPMScheduler, LDMPipeline, UNetaDModel, VQModel def snake_case_ ( snake_case , snake_case=1 ) -> Tuple: if n_shave_prefix_segments >= 0: return ".".join(path.split('.' )[n_shave_...
335
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) __lowerCAmelCase = {} try: if not is_sentencepiece_available(): raise OptionalDependencyN...
335
1
'''simple docstring''' import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @req...
24
from __future__ import annotations def UpperCAmelCase_ ( UpperCAmelCase__ ): if len(UpperCAmelCase__ ) == 0: return [] lowercase_ , lowercase_ = min(UpperCAmelCase__ ), max(UpperCAmelCase__ ) lowercase_ = int(max_value - min_value ) + 1 lowercase...
412
0
import json import sys def __lowerCAmelCase ( A , A ): with open(__snake_case , encoding="utf-8" ) as f: UpperCAmelCase_ = json.load(__snake_case ) UpperCAmelCase_ = ["<details>", "<summary>Show updated benchmarks!</summary>", " "] fo...
712
import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTes...
268
0